Category: mobile

  • Apple A5 from the Apple TV 3 – and an iPad 2! » Chipworks

    Pictures of the two different A5 chips
    From: 9to5mac, the two Apple A5 cpus in question

    Not only did Apple roll out a new processor that was not what it was advertised to be, but it also snuck in a new process technology for the manufacturing of this new A5. The previous generation A5, part number APL0498, was manufactured on Samsung Semiconductors’ 45 nm LP CMOS process. This new A5 processor is manufactured on Samsung’s new 32 nm high-k metal gate, gate first, LP CMOS process technology.

    via Update – Apple A5 from the Apple TV 3 – and an iPad 2! » Technology Blog » Chipworks.

    Check out the article at the Chipworks website, just follow the link above. They have a great rundown of what they discovered in their investigation of the most recent Apple A5 chips. These chips are appearing in a newly revised AppleTV but have also appeared in more recently manufactured Apple iPad 2 as well. There was some amount of surprise that Apple didn’t adopt a shrunk down die ruling for the A5X used in the iPad 3. Most of the work went into the integrated graphics of the A5X as it was driving a much higher rez ‘Retina’-like display.

    Very, very sneaky of Apple to slip in the next generation smaller die size on a ‘hobby’ product like the Apple TV. This is proof positive that when someone says something is a hobby, it isn’t necessarily so. I for one am both heartened and intrigued that Apple is attempting to get a 32nm processor out there on their ‘low power’ low cost products. Now that this part has also been discovered in the more recently constructed Apple iPad 2 units, I wonder what kind of heat, battery life differences there are versus an early model iPad 2 using the A5 part number APL0498?

    Keeping up with the Samsungs is all important these days and Apple has got to keep its CPU die rulings in step with the next generation of of chip fabrication giants. Intel is pushing 22nm, Samsung has been on 32nm for a while and then there’s Apple sitting 1 or 2 generations behind the cutting edge. I fear this may have resulted in some of the heat issues that were first brought to people’s attention by Consumer Reports weeks after the introduction of the iPad 3. With any luck and process engineering speed, the A5X can jump ship to the 32nm fabrication line at Samsung sooner rather than later.

  • Apple A5X CPU in Review

    Apple Inc.
    Apple Inc. (Photo credit: marcopako )

    A meta-analysis of the Apple A5X system on chip

    (from the currently shipping 3rd Gen iPad)

    New Ipad’s A5X beats NIVIDIA Tegra 3 in some tests (MacNN|Electronista)

    Apple’s A5X Die (and Size?) Revealed (Anandtech.com)

    Chip analysis reveals subtle changes to new iPad innards (AppleInsider-quoting Anandtech)

    Apple A5X Die Size Measured: 162.94mm^2, Samsung 45nm LP Confirmed (Update from Anandtech based on a more technical analysis of the chip)

    Reading through all the hubbub and hand-waving from the technology ‘teardown’ press outlets, one would have expected a bigger leap from Apple’s chip designers. A fairly large chip sporting an enormous graphics processor integrated into the die is what Apple came up with to help boost itself to the next higher rez display (so-called Retina Display). The design rule is still a pretty conservative 45nm (rather than try to push the envelope by going with 32nm or thinner to bring down the power requirements). Apple similarly had to boost its battery capacity to make up for this power hungry pixel demon by almost 2X more than the first gen iPad. So for almost the ‘same’ amount of battery capacity (10 hours of reserve power), you get the higher rez display. But a bigger chip and higher rez display will add up to some extra heat being generated, generally speaking. Which leads us to a controversy.

    Given this knowledge there has been a recent back and forth argument over thermal design point for iPad 3rd generation. Consumer Reports published an online article saying the power/heat dissipation was much higher than previous generation iPads. They included some thermal photographs indicating the hot spots on the back of the device and relative temperatures. While the iPad doesn’t run hotter than a lot of other handheld devices (say Android tablets). It does run hotter than say an iPod Touch. But as Apple points out that has ALWAYS been the case. So you gain some things you give up some things and still Apple is the market leader in this form factor, years ahead of the competition. And now the tempest in the teapot is winding down as Consumer Reports (via LA Times.com)has rated the 3rd Gen iPad as it’s no. 1 tablet on the market (big surprise). So while they aren’t willing to retract their original claim of high heat, they are willing to say it doesn’t count as ’cause for concern’. So you be the judge when you try out the iPad in the Apple Store. Run it through its paces, a full screen video or 2 should heat up the GPU and CPU enough to get the electrons really racing through the device.

    A picture of the Apple A5X
    This is the new System on Chip used by the Apple 3rd generation iPad
  • ARM Wants to Put the Internet in Your Umbrella | Wired Enterprise | Wired.com

    Image representing Wired Magazine as depicted ...
    Image via CrunchBase

    On Tuesday, the company unveiled its new ARM Cortex-M0+ processor, a low-power chip designed to connect non-PC electronics and smart sensors across the home and office.

    Previous iterations of the Cortex family of chips had the same goal, but with the new chip, ARM claims much greater power savings. According to the company, the 32-bit chip consumes just nine microamps per megahertz, an impressively low amount even for an 8- or 16-bit chip.

    via ARM Wants to Put the Internet in Your Umbrella | Wired Enterprise | Wired.com.

    Lower power means a very conservative power budget especially for devices connected to the network. And 32 bits is nothing to sneeze at considering most manufacturers would pick a 16 or 8-bit chip to bring down the cost and power budget too. According to this article the degree of power savings is so great in fact that in sleep mode the chip consumes almost no power at all. For this market Moore’s Law is paying off big benefits especially given the bonus of a 32bit core. So not only will you get a very small lower power cpu, you’ll have a much more diverse range of software that could run on it and take advantage of a larger memory address space as well. I think non-PC electronics could include things as simple as web cams or cellphone cameras. Can you imagine a CMOS camera chip with a whole 32bit cpu built in? Makes you wonder no just what it could do, but what ELSE it could do, right?

    The term ‘Internet of Things‘ is bandied about quite a bit as people dream about cpus and networks connecting ALL the things. And what would be the outcome if your umbrella was connected to the Internet? What if ALL the umbrellas were connected? You could log all kinds of data, whether it was opened or close, what the ambient temperature is. It would be like a portable weather station for anyone aggregating all the logged data potentially. And the list goes on and on. Instead of Tire pressure monitors, why not also capture video of the tire as it is being used commuting to work. It could help measure the tire wear and setup and appointment when you need to get a wheel alignment. It could determine how many times you hit potholes and suggest smoother alternate routes. That’s the kind of blue sky wide open conjecture that is enabled by a 32-bit low/no power cpu.

    Moore's Law, The Fifth Paradigm.
    Moore’s Law, The Fifth Paradigm. (Photo credit: Wikipedia)
  • Raspberry Pi Hits a Slight Manufacturing Delay

    English: Extract from Raspberry Pi board at Tr...
    Image via Wikipedia

    The $35 Raspberry Pi “Model B” is board of choice to ship out to consumers first. It contains two USB ports, 256 MB of RAM, an Ethernet port and a 700 MHz Broadcom BCM2835 SoC. The Videocore 4 GPU within the SoC is roughly the equivalent to the original Xboxs level of performance, providing Open GL ES 2.0, hardware-accelerated OpenVG, and 1080p30 H.264 high-profile decode.

    via Raspberry Pi Hits a Slight Manufacturing Delay.

    Raspberry Pi boards are on the way and the components list is still pretty impressive for $35 USD. Not bad, given they had a manufacturing delay. The re-worked boards should ship out as a second batch once they have been tested fully. It also appears all the other necessary infrastructure is slowly falling into place to help create a rich environment for curious and casually interested purchasers of the Raspberry Pi. For instance let’s look at the Fedora remixes for Raspberry Pi.

    A remix in the Open source software community refers to a distribution of an OS that can run without compiling on a particular chip architecture whether it be the Raspberry Pi Broadcom chip or an Intel x86 variety. In addition to the OS a number of other pre-configured applications will be included so that you can start using the computer right away instead of having to download lots of apps. The best part of this is not only the time savings but the lowering of the threshold to less technical users. Also of note is the particular Fedora OS distributions chosen LXDE and XFCE both noted for being less resource intensive and smaller in physical size. The documentation on the Fedora website indicates these two distros are geared for older less capable, less powerful computers that you would still like to use around the house. And for a Raspberry Pi user, getting a tuned OS specifically compiled for your CPU and ready to go is a big boon.

    What’s even more encouraging is the potential for a Raspberry Pi community to begin optimizing and developing a new range of apps specifically geared towards this new computer architecture. I know the Fedora Yum project is a great software package manager using the RPM format for adding and removing software components as things change. And having a Yum app geared specifically for Raspberry Pi users might give a more App store like experience for the more casual users interested in dabbling. Right now there’s a group at Seneca College in Toronto, CA doing work on an app store-like application that would facilitate the process off discovering, downloading and

    Logo for Raspberry Pi
    Raspberry Pi project

    trying out different software pre-compiled for the Raspberry Pi computer.

    Broadcom
    Image via Wikipedia
  • AnandTech – Microsoft Provides Windows on ARM Details

    7 by Steven Sinofsky (President of Windows Div...
    Image via Wikipedia

    As reported by Andrew Cunningham for Anandtech: Weve known that Microsoft has been planning an ARM-compatible version of Windows since well before we knew anything else about Windows 8, but the particulars have often been obscured both by unclear signals from Microsoft itself and subsequent coverage of those unclear signals by journalists. Steven Sinofsky has taken to the Building Windows blog today to clear up some of this ambiguity, and in doing so has drawn a clearer line between the version of Windows that will run on ARM, and the version of Windows that will run on x86 processors.

    via AnandTech – Microsoft Provides Windows on ARM Details.

    That’s right ARM cpus are in the news again this time info for the planned version of Windows 8 for the mobile CPU. And it is a separate version of Windows OS not unlike Windows CE or Windows Mobile or Windows Embedded. They are all called Windows, but are very different operating systems. The product will be called Windows on ARM (WOA) and is only just now being tested internally at Microsoft with a substantial development and release to developers still to be announced.

    One upshot of this briefing from Sinofsky was the mobile-centric Metro interface will not be the only desktop available on WOA devices. You will also be able to use the traditional looking Windows desktop and not incur a big battery power performance hit. Which makes it a little more palatable to a wider range of users no doubt who might consider buying a phone or tablet or Ultrabook running an ARM cpu running the new Windows 8 OS. Along the same lines there will be a version of Office apps that will also run on WOA devices including the big three Word, Excel and Powerpoint. These versions will be optimized for mobile devices with touch interfaces which means you should buy the right version of Office for your device (if it doesn’t come pre-installed).

    Lastly the optimization and linking to specially built Windows on ARM devices means you won’t be able to install the OS on just ‘any’ hardware you like. Similar to Windows Mobile, you will need to purchase a device designed for the OS and most likely with a version pre-installed from the factory. This isn’t like a desktop OS built to run on many combos of hardware with random devices installed, it’s going to be much more specific and refined than that. Microsoft wants to really constrain and coordinate the look and feel of the OS on many mobile devices so that an average person can expect it to work similarly and look similar no matter who the manufacturer of the device will be. One engineering choice that is going to assist with this goal is an attempt to address the variations in devices by using so-called “Class Drivers” to support the chipsets and interfaces in a WOA device. This is a less device specific way of support say a display panel, keyboard without having to know every detail. A WOA device will have to be designed and built to a spec provided by Microsoft for which then it will provide a generic ‘class driver’ for that keyboard, display panel, USB 3.0 port, etc. So unlike Apple it won’t just be a limited set of hardware components necessarily, but they will have to meet the specs to be supported by the Windows on ARM OS. This no doubt will make it much easier for Microsoft to keep it’s OS up to date as compared to say in the Google Android universe where the device manufacturers have to provide the OS updates (which in fact is not often as they prefer people to upgrade their device to get the new OS releases).

  • Buzzword: Augmented Reality

    Augmented Reality in the Classroom Craig Knapp
    Augmented Reality in the Classroom Craig Knapp (Photo credit: caswell_tom)

    What it means. “Augmented reality” sounds very “Star Trek,” but what is it, exactly? In short, AR is defined as “an artificial environment created through the combination of real-world and computer-generated data.”

    via Buzzword: Augmented Reality.

    Nice little survey from the people at Consumer Reports, with specific examples given from the Consumer Electronics Show this past January. Whether it’s software or hardware there’s a lot of things that can be labeled and marketed as ‘Augmented Reality’. On this blog I’ve concentrated more on the apps running on smartphones with integrated cameras, acclerometers and GPS. Those pieces are important building blocks for an integrated Augmented Reality-like experience. But as this article from CR shows, your experience may vary quite a bit.

    In my commentary on stories posted by others on the Internet, I have covered mostly just the examples of AR apps on mobile phones. Specifically I’ve concentrated on the toolkit provided by Layar to add metadata to existing map points of interest. The idea of ‘marking up’ the existing landscape for me holds a great deal of promise as the workload is shifted off the creator of the 3D world to the people traveling within it. The same could hold true for Massively Multiplayer Games and some worlds do allow the members to do that kind of building and marking up of the environment itself. But Layar provides a set of data that you can call up while merely pointing the cell phone camera at a compass direction and then bring up the associated data.

    It’s a sort of hunt for information, sometimes it’s well done if the metadata mark-up is well done. But like many crowd-sourced efforts some amount of lower quality work or worse vandalism occurs. But this should keep anyone from trying to enhance the hidden data that can be discovered through a Layar enhanced Real World. I’m hoping the mobile phone based AR applications grow and find a niche if not a killer app. It’s still early days and mobile phone AR is not being adopted very quickly but I think there’s still a lot of untapped resources there. I don’t think we have discovered all the possible applications of mobile phone AR.

  • ARM Pitches Tri-gate Transistors for 20nm and Beyond

    English: I am the author of this image.
    Image via Wikipedia

    . . . 20 nm may represent an inflection point in which it will be necessary to transition from a metal-oxide semiconductor field-effect transistor MOSFET to Fin-Shaped Field Effect Transistors FinFET or 3D transistors, which Intel refers to as tri-gate designs that are set to debut with the companys 22 nm Ivy Bridge product generation.

    via ARM Pitches Tri-gate Transistors for 20nm and Beyond.

    Three Dimensional transistors in the news again. Previously Intel announced they were adopting a new design for their next generation next smaller design rule for the Ivy Bridge generation Intel CPUs. Now ARM is also doing work to integrate similar technology into their ARM cpu cores as well. No doubt in order to lower Thermal Design Point and maintain clock speed as well are both driving this move to refine and narrow the design rules for the ARM architecture. Knowing Intel is still the top research and development outfit for silicon semi-conductors would give pause to anyone directly competing with them, but ARM is king of the low power semi-conductor and keeping pace with Intel’s design rules is an absolute necessity.

    I don’t know how quickly ARM is going to be able to get a licensee to jump onboard and adopt the new design. Hopefully a large operation like Samsung can take this on and get the chip into it’s design, development, production lines at a chip fabrication facility as soon as possible. Likewise other contract manufacturers like Taiwan Semiconductor Manufacturing Company (TSMC) should also try to get this chip into their facilities quickly too. That way the cell-phone and tablet markets can benefit too as they use a lot of ARM licensed cpu cores and similar intellectual property in their shipping products. And my interest is not so much invested in the competition between Intel and ARM for low power computing but more the overall performance of any single ARM design once it’s been in production for a while and optimized the way Apple designs its custom CPUs using ARM licensed cpu cores. The single most outstanding achievement of Apple in their design and production of the iPad is the battery charge duration of 10 hours. Which to date, is an achievement that has not been beaten, even by other manufacturers and products who also license ARM intellectual property. So if  the ARM design is good and can be validated and proto-typed with useful yields quickly, Apple will no doubt be the first to benefit, and by way of Apple so will the consumer (hopefully).

    Schematic view (L) and SEM view (R) of Intel t...
    Image via Wikipedia
  • Xen hypervisor ported to ARM chips • The Register

    Deutsch: Offizielles Logo der ARM-Prozessorarc...
    Image via Wikipedia

    You can bet that if ARM servers suddenly look like they will be taking off that Red Hat and Canonical will kick in some help and move these Xen and KVM projects along. Server maker HP, which has launched the “Redstone” experimental server line using Calxedas new quad-core EnergyCore ARM chips, might also help out. Dell has been playing around with ARM servers, too, and might help with the hypervisor efforts as well.

    via Xen hypervisor ported to ARM chips • The Register.

    This is an interesting note, some open source Hypervisor projects are popping up now that the ARM Cortex A15 has been announced and some manufacturers are doling out development boards. What it means longer term is hard to say other than it will potentially be a boon to manufacturers using the ARM15 in massively parallel boxes like Calxeda. Or who are trying to ‘roll their own’ ARM based server farms and want to have the flexibility of virtual machines running under a hypervisor environment. However, the argument remains, “Why use virtual servers on massively parallel cpu architectures when a  1:1 cpu core to app ratio is more often preferred?”

    However, I would say old habits of application and hardware consolidation die hard and virtualization is going to be expected because that’s what ‘everyone’ does in their data centers these days. So knowing that a hypervisor is available will help foster some more hardware sales of what will most likely be a niche products for very specific workloads (ie. Calxeda, Qanta SM-2, SeaMicro). And who knows maybe this will foster more manufacturers or even giant data center owners (like Apple, Facebook and Google) to attempt experiments of rolling their own ARM15 environments knowing there’s a ready made hypervisor out there that they can compile on the new ARM chip.

    However, I think all eyes are really still going to be on the next generation ARM version 8 with the full 64bit memory and instruction set. Toolsets nowadays are developed in house by a lot of the datacenters and the dominant instruction set is Intel x64 (IA64) which means the migration to 64bits has already happened. Going back to 32bits just to gain the advantage of the lower power ARM architecture is far to costly for most. Whereas porting from IA64 to 64bit ARM architecture is something more datacenters might be willing to do if the potential cost/benefit ratio is high enough to cross-compile and debug. So legacy management software toolsets are really going to drive a lot of testing and adoption decisions by data centers looking at their workloads and seeing if ARM cpus fit their longer term goals of saving money by using less power.

  • Disruptions: Wearing Your Computer on Your Sleeve – NYTimes.com

    English: This depicts the evolution of wearabl...
    Image via Wikipedia: Bad old days of Wearable Computers

    Wearable computing is a broad term. Technically, a fancy electronic watch is a wearable computer. But the ultimate version of this technology is a screen that would somehow augment our vision with information and media.

    via Disruptions: Wearing Your Computer on Your Sleeve – NYTimes.com.

    Augmented Reality in the news, only this time it’s Google so it’s like for rilz, yo! Just kidding, it will be very interesting given Google’s investment in the Android OS and power-saving mobile computing what kind of wearable computers they will develop. No offense to MIT Media Lab, but getting something into the hands of end-users is something Google is much more accomplished at doing (but One Laptop Per Child however is the counter-argument of course). I think mobile phones are already kind of like a wearable computer. Think back to the first iPod arm bands right? Essentially now just scale the ipod up to the size of an Android and it’s no different. It’s practically wearable today (as Bilton says in his article).

    What’s different then with this effort is the accessorizing of the real wearable computer (the smart phone) giving it the augmentation role we’ve seen with products like Layar. But maybe not just limited to cameras, video screens and information overlays, the next wave would have auxiliary wearable sensors communicating back to the smartphone like the old Nike accelerometer that would fit into special Nike Shoes. And also consider the iPod Nano ‘wrist watch’ fad as it exists today. It may not run the Apple iOS, but it certainly could transmit data to your smartphone if need be. Which leads to the hints and rumors of attempts by Apple to create ‘curved glass’.

    This has been an ongoing effort by Apple, without being tied to any product or feature in their current product line. Except maybe the iPhone. Most websites I’ve read to date speculate the curvature is not very pronounced and a styling cue to further help marketing and sales of the iPhone. But in this article the curvature Bilton is talking about would be more like the face of a bracelet around the wrist, much more pronounced. Thus the emphasis on curved glass might point to more work being done on wearable computers.

    Lastly Bilton’s article goes into a typical futuristic projection of what form the video display will take. No news to report on this topic specifically as it’s a lot of hand-waving and make believe where contact lenses potentially can become display screens. As for me, the more pragmatic approach of companies like Layar creating iPhone/Augmented Reality software hybrids is going to ship sooner and prototype faster than the make believe video contact lenses of the Future.The takeaway I get from Bilton’s article is there’s more of a defined move to create more functions with the smartphone as more of a computer. Though MIT Media Lab have labeled this ‘wearable computing’ think of it more generally as Ubiquitous Computing where the smartphone and its data connection are with you wherever you go.

  • The PC is dead. Why no angry nerds? :: The Future of the Internet — And How to Stop It

    Famously proprietary Microsoft never dared to extract a tax on every piece of software written by others for Windows—perhaps because, in the absence of consistent Internet access in the 1990s through which to manage purchases and licenses, there’d be no realistic way to make it happen.

    via The PC is dead. Why no angry nerds? :: The Future of the Internet — And How to Stop It.

    While true that Microsoft didn’t tax Software Developers who sold product running on the Windows OS, a kind of a tax levy did exist for hardware manufacturers creating desktop pc’s with Intel chips inside. But message received I get the bigger point, cul-de-sacs don’t make good computers. They do however make good appliances. But as the author Jonathan Zittrain points out we are becoming less aware of the distinction between a computer and an applicance, and have lowered our expectation accordingly.

    In fact this points to the bigger trend of not just computers becoming silos of information/entertainment consumption no, not by a long shot. This trend was preceded by the wild popularity of MySpace, followed quickly by Facebook and now Twitter. All platforms as described by their owners with some amount of API publishing and hooks allowed to let in 3rd party developers (like game maker Zynga). But so what if I can play Scrabble or Farmville with my ‘friends’ on a social networking ‘platform’? Am I still getting access to the Internet? Probably not, as you are most likely reading what ever filters into or out of the central all-encompassing data store of the Social Networking Platform.

    Like the old World Maps in the days before Columbus, there be Dragons and the world ends HERE even though platform owners might say otherwise. It is an Intranet pure and simple, a gated community that forces unique identities on all participants. Worse yet it is a big brother-like panopticon where each step and every little movement monitored and tallied. You take quizzes, you like, you share, all these things are collection points, check points to get more data about you. And that is the TAX levied on anyone who voluntarily participates in a social networking platform.

    So long live the Internet, even though it’s frontier, wild-catting days are nearly over. There will be books and movies like How the Cyberspace was Won, and the pioneers will all be noted and revered. We’ll remember when we could go anywhere we wanted and do lots of things we never dreamed. But those days are slipping as new laws get passed under very suspicious pretenses all in the name of Commerce. As for me I much prefer Freedom over Commerce, and you can log that in your stupid little database.

    Cover of "The Future of the Internet--And...
    Cover via Amazon