5 Comments

Summary:

Intel this week announced a $12 million investment into a visual computing research program focused on using three-dimensional imaging for entertainment, data analysis, medical imaging and scientific research. The Intel Visual Computing Institute is located at Saarland University in Saarbrücken, Germany, and will receive the $12 […]

intelIntel this week announced a $12 million investment into a visual computing research program focused on using three-dimensional imaging for entertainment, data analysis, medical imaging and scientific research. The Intel Visual Computing Institute is located at Saarland University in Saarbrücken, Germany, and will receive the $12 million over the next five years. The investment is both a nod to the company’s interest in better imaging tools and a path to selling faster, high-margin processors.

Intel is also on the defensive. Its push into 3D is the chip giant’s attempt to find ways to get folks to buy faster, high-margin chips, be they the current generation of Core i7 CPUs or eventually Intel’s Larrabee specialty graphics chip. But it also needs a high-demand chip that can help pad its margins from the success of its low-cost Atom chip.

It can take a lot of computing power to process and display 3D imagery. It also takes a lot of processing power to collect visual data in real time, be it from cameras tracking a person’s movements or merely plucking it from a stream of data and rendering that data as a 3D image. The benefits of visual computing are not limited to better gaming, but can create realistic motion-capture that can be translated to an avatar online, or offer more realistic artificial intelligence that can enable a computer to react to visual cues.

But the key for Intel is the fact that chips used to process such data and to render the graphics are ones with high margins. As video and imagery have become more common, Intel has seen graphics-focused chipmakers such as Nvidia and AMD (which bought GPU maker ATI in 2005) get more attention from mainstream consumers and infringe on its PC market. For example, Apple is now using Nvidia’s graphics chips in its MacBooks after bumping Intel’s integrated graphics last year, and personal tech guru Walt Mossberg has told consumers to consider GPU upgrades rather than x86 upgrades to boost performance on their PCs.

Intel has its own fabs that are able to produce more and more chips every few years, which means the company needs to sell a lot of them. And because it’s making high-end chips for servers, it needs to keep abreast of cutting-edge research and invest continually in the business. In 2008 Intel invested $5.2 billion, or 15 percent of its revenue, in R&D. However, those R&D costs are one of the things that eat away at Intel’s margins, which have shrunk to 45.3 percent in the first quarter of 2009 from 53.8 percent in the first quarter of 2008.

Another drain on Intel’s margins are its low-power Atom processors, which sell for about $250 less than the company’s high-end PC chips. While Intel notes that the Atom chips are high-margin, it looks like they’re also cannibalizing sales of more powerful Intel processors in laptops. The 45-nm Atom costs about $29, compared with $279 for the company’s Penryn Core 2 processor for standard notebook computers, according to Robert Castellano, president of New Tripoli, Pa.-based market research firm The Information Network. He says Intel has had to shift some production of Atom chips to a foundry to keep its margins intact.

But if it can’t count on Atom to fill its fabs, Intel must find other mass-market chips that it can charge dearly for. Graphics may offer that chance, although it’s a more competitive market than Intel is used to.

This article also appeared on BusinessWeek.com.

  1. Wesley Faulkner Friday, May 15, 2009

    This is a smart move for Intel, but $12 million over the next five years sounds more like a NFL player’s salary than a real investment. With the amount of money that company makes I am not sure if they have faith that area of research will pay off anytime soon.

  2. IT Management » Blog Archive » Friday Links: Cyberattacks, 3D Imaging, Cisco Friday, May 15, 2009

    [...] is working on 3D imaging. Gigaom wonders if that can keep the company on [...]

  3. Jason Lackey Friday, May 15, 2009

    I suspect that everything Intel needs to learn about 3D can be learned down the street on San Thomas Road (where Nvidia is) rather than some place in Germany.

  4. Indeed 12M is peanuts. Taking some capital expenditure into account, this can feed a team of ~8 to 10 researchers per year for 5 years. Wish Intel’s future was so easy to secure !
    On the other hand, the 12M can be well spent, if it results in vital technologies and patents being created. Clearly 3D is becoming mainstream, and who knows, there could be a lot of interesting stuff waiting to be invented..and not necessarily waiting only for some smart folks on San Thomas Road to stumble upon. I’d wish luck to the Saarland folks!

  5. Chip Industry to See Slow Recovery in 2010 Friday, June 5, 2009

    [...] will be dampened by firms selling cheaper chips, such as Intel’s Atom, which can cost almost 10 times less than Intel’s processors for notebooks. That’s good for consumers, who get cheaper gear, [...]

Comments have been disabled for this post