Category: gpu

  • The Sandy Bridge Review: Intel Core i7-2600K – AnandTech

    Quick Sync is just awesome. Its simply the best way to get videos onto your smartphone or tablet. Not only do you get most if not all of the quality of a software based transcode, you get performance thats better than what high-end discrete GPUs are able to offer. If you do a lot of video transcoding onto portable devices, Sandy Bridge will be worth the upgrade for Quick Sync alone.

    For everyone else, Sandy Bridge is easily a no brainer. Unless you already have a high-end Core i7, this is what youll want to upgrade to.

    via The Sandy Bridge Review: Intel Core i7-2600K, i5-2500K and Core i3-2100 Tested – AnandTech :: Your Source for Hardware Analysis and News.

    Previously in this blog I have recounted stories from Tom’s Hardware and Anandtech.com surrounding the wicked cool idea of tapping the vast resources contained within your GPU while you’re not playing video games. Producers of GPUs like nVidia and AMD both wanted to market their products to people who not only gamed but occasionally ripped video from DVDs and played them back on ipods or other mobile devices. The amount of time sunk into doing these kinds of conversions were made somewhat less of a pain due to the ability to run the process on a dual core Wintel computer, browsing web pages  while re-encoding the video in the background. But to get better speeds one almost always needs to monopolize all the cores on the machine and free software like HandBrake and others will take advantage of those extra cores, thus slowing your machine, but effectively speeding up the transcoding process. There was hope that GPUs could accelerate the transcoding process beyond what was achievable with a multi-core cpu from Intel. An example is also Apple’s widespread adoption of OpenCL as a pipeline to the GPU to send rendering requests for any video frames or video processing that may need to be done in iTunes, QuickTime or the iLife applications. And where I work, we get asked to do a lot of transcoding of video to different formats for customers. Usually someone wants a rip from a DVD that they can put on a flash drive and take with them into a classroom.

    However, now it appears there is a revolution in speed in the works where Intel is giving you faster transcodes for free. I’m talking about Intel’s new Quick Sync technology using the integrated graphics core as a video transcode accelerator. The speeds of transcoding are amazingly fast and given the speed, trivial to do for anyone including the casual user. In the past everyone seemed to complain about how slow their computer was especially for ripping DVDs or transcoding the rips to smaller more portable formats. Now, it takes a few minutes to get an hour of video into the right format. No more blue Monday. Follow the link to the story and analysis from Anandtech.com as they ran head to head comparisons of all the available techniques of re-encoding/transcoding a Blue-ray video release into a smaller .mp4 file encoded in as h.264. They did comparisons of Intel four-core cpus (which took the longest and got pretty good quality) versus GPU accelerated transcodes, versus the new Intel QuickSync technology coming out soon on the Sandy Bridge gen Intel i7 cpus. It is wicked cool how fast these transcodes are and it will make the process of transcoding trivial compared to how long it takes to actually ‘watch’ the video you spent all that time converting.

    Links to older GPU accelerated video articles:

    https://carpetbomberz.com/2008/06/25/gpu-accelerated-h264-encoding/
    https://carpetbomberz.com/2009/06/12/anandtech-avivo/
    https://carpetbomberz.com/2009/06/23/vreveal-gpu/
    https://carpetbomberz.com/2010/10/18/microsoft-gpu-video-encoding-patent/

  • Announcing the first free software Blu-ray encoder

    Diary Of An x264 Developer » (4/25/2010)

    For many years it has been possible to make your own DVDs with free software tools.  Over the course of the past decade, DVD creation evolved from the exclusive domain of the media publishing companies to something basically anyone could do on their home computer.

    The move towards Blu-ray encoding is very encouraging. In reading the article I don’t see a mention of CUDA or OpenCL acceleration of the encoding process. As was the case for MPEG-2 a glaring need for acceleration of the process was painfully obvious once people started converting long form videos. I know x264 encoding can be accelerated by splitting threads across CPUs on a multi-core processor. But why not unleash the floodgates and get some extra horsepower from the ATI or nVidia graphics card too. We’re talking large frames and large frame rates and the only way to guarantee adoption of the new format is to make the encoding process fast, fast, fast.

  • Apple A4 processor really stripped-down Cortex A8? | Electronista

    The custom A4 processor in the iPad is in reality a castrated Cortex A8 ARM design, say several sources.

    via Apple A4 processor really stripped-down Cortex A8? | Electronista.

    This is truly interesting, and really shows some attempt to optimize the chips with ‘known’ working designs. Covering the first announcement of the A4 chip by Brightside of News, I tried to argue that customizing a chip by licensing a core design from ARM Holdings Inc. isn’t all that custom. Following this Ashlee Vance wrote in the NYTimes the cost of development for the A4 ‘could be’ upwards of $1Billion. And now just today MacNN/Electronista is saying Apple used the ARM A8. By this I mean the ARM Cortex A8 is a licensed core already being used in the Apple iPhone 3GS. It is a proven, known cpu core that engineers are familiar with at Apple. Given the level of familiarity, it’s a much smaller step to optimize that same CPU core for speed and integration with other functions. Like for instance the GPU or memory controllers can be tightly bound into the final CPU. Add a dose of power management and you got good performance and good battery life. It’s not cutting edge to be sure, but it is more guaranteed to work right out of the gate. That’s a bloodthirsty step in the right direction of market domination. However, the market hasn’t quite yet shown itself to be so large and self sustaining that slate devices are a sure thing in the casual/auxiliary/secondary computing device market. You may have an iPhone and you may have a laptop, bu this device is going to be purchased IN ADDITION not INSTEAD OF those two existing device markets. So anyone who can afford a third device is probably going to be the target market for iPad as opposed to creating a new platform for people that want to substitute an iPad for either the iPhone or laptop.

  • Intel Gets Graphic with Chip Delay – Bits Blog – NYTimes.com

    Intel’s executives were quite brash when talking about Larrabee even though most of its public appearances were made on PowerPoint slides. They said that Larrabee would roar onto the scene and outperform competing products.

    via Intel Gets Graphic with Chip Delay – Bits Blog – NYTimes.com.

    And so now finally the NY Times nails the coffin shut on Intel’s Larrabee saga. To refresh your memory this is the second attempt by Intel to create a graphics processor. The first failed attempt was some years ago in the late 1990s when 3dfx (bought by nVidia) was tearing up the charts with their Voodoo 1 and Voodoo 2 PCI-based 3D accelerator cards. The age of Quake, Quake 2 were upon us and everyone wanted smoother frame rates. Intel wanted to show its prowess in the design of a low cost graphics card running on the brand new AGP slot which Intel had just invented (remember AGP?). What turned out was a similar set of delays and poor performance as engineering samples came out of the development labs. Given the torrid pace of products released by nVidia and eventually ATI, Intel couldn’t keep up. Their benchmark was surpassed by the time their graphics card saw the light of day, and they couldn’t give them away. (see Wikipedia: Intel  i740)

    Intel i740 AGP graphics card
    1998 saw the failure of the Intel i740 AGP graphics card

    The Intel740, or i740, is a graphics processing unit using an AGP interface released by Intel in 1998. Intel was hoping to use the i740 to popularize the AGP port, while most graphics vendors were still using PCI. Released with enormous fanfare, the i740 proved to have disappointing real-world performance, and sank from view after only a few months on the market

    Enter Larrabee, a whole new ball game at Intel, right?! The trend toward larger numbers of parallel processors on GPUs from nVidia and ATI/AMD led Intel to believe they might leverage some of their production lines to make a graphics card again. But this time it was different, nVidia had moved from single purpose GPUs to General Purpose GPUs in order to create a secondary market using their cards as compute intensive co-processor cards. They called it CUDA and provided a few development tools at the early stages. Intel latched onto this idea of the General Purpose GPU and decided they could do better. What’s more general purpose than an Intel x86 processor right? And what if you could provided the libraries and Hardware Abstraction Layer that could turn a larger number of processor cores into something that looked and smelled like a GPU?

    For Intel it seemed like a win/win/win everybody wins. The manufacturing lines using older design rules at the 45nm size could be utilized for production, making the graphics card pure profit. They could put 32 processors on a card and program them to do multi duties for the OS (graphics for games, co-processor for transcoding videos to MP4). But each time they did a demo a product white paper and demo at a trade show it became obvious the timeline and schedule was slipping. They had benchmarks to show, great claims to make, future projections of performance to declare. Roadmaps were the order of the day. But just last week rumors started to set in.

    Similar to the graphics card foray of the past Intel couldn’t beat it’s time to market demons. The Larrabee project was going to be so late and still was using 45nm manufacturing design rules. Given Intel’s top of the line production lines moved to 32nm this year, and nVidia and AMD are doing design process shrinks on their current products, Intel was at a disadvantage. Rather than scrap the thing and lose face again, they decided to recover somewhat and put Larrabee out there as a free software/hardware development kit and see if that was enough to get people to bite. I don’t know what if any benefit any development on this platform would bring. It would rank right up there with the Itanium and i740 as hugely promoted dead-end products with zero to negative market share. Big Fail – Do Not Want.

    And for you armchair Monday morning technology quarter backs here are some links to enjoy leading up to the NYTimes article today:

    Tim Sweeney Laments Intel Larrabee Demise (Tom’s Hardware Dec. 7)

    Intel Kills Consumer Larrabee Plans (Slashdot Dec. 4)

    Intel delays Larrabee GPU, aims for developer “kit” in 2010 (MacNN Dec. 4)

    Intel condemns tardy Larrabee to dev purgatory (The Register Dec.4)

  • Acrossair on the iPhone

    It looks like the iPhone OS 3.1 is going to do nothing more the open up the video feed on the camera so that you can overlay data on top of that video. In essence, the Augmented Reality is using your iPhone’s video as a “desktop” picture and placing items on top of that. Acrossair’s iPhone App, Nearest Tube uses the OpenGL libraries to skew and distort that data as you point the camera in different directions, thus providing a little more of a 3D perspective than say something like Layar which I have talked about previously on this blog. Chetan Demani, one of the founders of Acrossair also points out going forward any company making AR type apps will need to utilize existing location information and pre-load all the data they want to display. So the nirvana of just-in-time downloads of location data to overlay on your iPhone video image is not here,… and may not be for a while. What will differentiate the software producers though is the relevancy, and accuracy of their location information. So there will be some room for competition for a quite some time.

    He went on to say that it’s pretty simple to do AR applications using the new 3.1 APIs, due out in September. ” It’s a pretty straightforward API. There’s no complexity in there. All it does is it just switches on the video feed at the background. That’s the only API that’s published. All we’re doing is using that video feed at the back. It just displays the video feed as if it’s a live camera feed.

    via Augmenting Reality with the iPhone – O’Reilly Broadcast.

  • vReveal uses GPU to accelerate video fixes

    Before and After

    There’s a new video trend in personal home video. Companies are lining up to provide aftermarket tools to process and provide corrections to camera phone video. Pure Digital’s Flip! camera line has some tools available to do some minor cutting to video clips and publish it to sharing websites. All of which presents an entrepreurial opporunity to provide pay for tools to help improve poorly shot video.

    Some tools are provided within video editing suites like Apples iMovie (it corrects camera shake). Now on the PC there are two new products, one of which is designed to take advantage of the nVidia GPU acceleration of parallel programming. The product is called vReveal

    While vReveal works with Windows XP or Vista (and not with Macs), it will make its enhancements much faster if the machine contains a recent graphics processing card from Nvidia, Dr. Varah said. Nvidia is an investor and a marketing partner with vReveal; a specific list of cards is at vReveal’s Web site.

    via Novelties – Making a Fuzzy Video Come Into Focus – NYTimes.com.