Stay on Top of Enterprise Technology Trends
Get updates impacting your industry from our GigaOm Research Community
We’ve wrapped up another Structure: Europe conference and, just like last year’s inaugural event, it was full of insights into the state of cloud computing in Europe and around the world. The fact that this was our first event since learning the full scale of the NSA’s internet-spying operations certainly played a role in many of the discussions — both onstage and in the hallways — as did a seemingly general consensus that Amazon(s amzn) Web Services cannot and will not be the only cloud computing provider that matters.
But what struck me most of all was the ridiculous amount of data that many of the speakers are dealing with. We’re not talking about a few terabytes here and there; rather, we’re talking about petabytes of storage and single companies — heck, single events — accounting for significant portions of all the world’s web traffic.
- Netflix(s nflx) streams 114,000 years of video every month and has built custom boxes for its content-delivery network that contain between 100 and 150 terabytes of storage apiece. In the United States alone, Netflix accounts for 35 percent of all web traffic, according to some estimates.
- Felix Baumgartner’s record-breaking jump last year generated 8 million concurrent views on YouTube(s goog) and consumed almost a quarter of the world’s bandwidth.
- An unnamed social network recently purchased 85 petabytes of storage from EMC.
- Blogging platform WordPress now underpins nearly 20 percent of the world’s blogs, which combine to be bigger than Yahoo in terms of web traffic.
- Facebook(s fb) grew so big so fast that it had started building its own servers to keep its hardware costs and power bills from spiraling out of control. And whereas its heterogeneous data centers have a 3 percent rate of server failure, its Swedish data center full of all custom-built servers has only a 1 percent rate.
- By 2018, BMW estimates it will have 10 million connected vehicles on the road, generating some 100 million requests and 1 terabyte of data volume per day. Those numbers actually seem small.
And then there’s CERN. The Swiss research institution is generating, processing and storing staggering amounts of data. Here are the highlights, but you’ll really want to watch Tim Bell’s entire presentation for the full and fascinating story behind why it’s growing so fast and how it’s doing so without crashing and burning:
- CERN’s 100-megapixel cameras take 40 million pictures a second of proton collisions — creating 1 petabyte of data per second that needs to be filtered down to reasonable levels for analysis and retention.
- CERN currently keeps about 35 petabytes of data per year, which scientists want to keep for 20 years. Its archival system consists of 45,000 tape drives.
- CERN’s current cloud environment consists of 50,000 cores right now and is expected to grow to about 300,000 cores by 2015.
[protected-iframe id=”ba56168c298440676942cacd63d729ac-14960843-6578147″ info=”http://new.livestream.com/accounts/74987/events/2361507/videos/30326858/player?autoPlay=false&height=360&mute=false&width=640″ width=”640″ height=”360″ frameborder=”0″ scrolling=”no”]
A transcript of Bell’s talk is available here.
All told, however, Structure: Europe was a great conference and a few statistics don’t really do it justice. Check out the live coverage page, read the blog posts and, if you have time, watch the sessions. Whether your interests lie in anything from big data to European cloud strategy, from privacy to webscale infrastructure, I think it’ll be worth the time.