Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
A Rare Tour of Microsoft’s Hyperscale Datacenters (nextplatform.com)
153 points by Katydid on Sept 26, 2016 | hide | past | favorite | 48 comments


"What came out of this was a realization that we were really building large air conditioners, that we were not in the IT business but in the industrial air conditioning business when you build a facility like this."

This was one of the big secrets that Google learned early on. Every bit of air you cool that isn't going into a computer is wasted. The issue is that co-location facilities have to be ready for any kind of user equipment, but if you own the entire data center and all the equipment inside you can design it differently and much more efficiently. It stops being computers in a building and starts being a building sized computer.


Funny story about datacenters, I worked on optimizing algorithmic performance for 4 years at one the big 4 but I was never allowed in a datacenter because I didn't have a green card. So my team would go on and explore whilst i had beers in the lobby.


That is funny. When I toured the supernap data center in Las Vegas (Blekko was thinking about locating a cluster there) they had a very elaborate security theatre sort of thing going on. Presumably they wanted to impress me at how secure it was. Clearly it was amusing as one of the sysadmins from a company there pulled me aside and said not to worry, once you had a contract it was a lot simpler to go in and out.


Can you elaborate on that? Wasn't it a private facility? Why is your green card status relevant to that? Were they hosting government assets and that was a requirement on the government's part? In that case odd that it's "green card" and not "citizenship".


Yes, they had DoD and DoE assets. They are very specific about who can go in there.


maybe by "green card" they mean a green light didn't light up when he scanned his access card. <badum-ching>


He likely means the "green card" as in the green card.

Datacenters are sensitive facilities. It's not surprising that they block entrance to many people.

From what I've seen in Europe, I can remember a couple of places in military, datacenters and national research centers where it's clearly written "European Only" on the jobs.


Whats the green card?


https://www.usa.gov/green-cards

"A Green Card (Permanent Resident Card): Gives you official immigration status in the United States. Entitles you to certain rights and responsibilities. Is required if you wish to naturalize as a U.S. Citizen"


My guess is he was a contractor and not a FT employee?


It will be interesting to see how these improvements trickle down.

At the moment, a few providers have the scale and skills to run datacenters much more efficiently. But I'm guessing that within a few years there will be some generic datacenter-in-a-container available, with efficiency not much inferior to the big four.

At that point, we go back to the hosting market of 15 years ago. Everybody can offer a datacenter without deep technical knowledge, and sell compute cycles on an open cloud market.

It's like all the tiny hosting providers, except that it requires more capital. So it becomes financialised -- if you can get cheap power, low temperatures and good connectivity, and borrow a few million dollars cheaply, then you're in business. But margins collapse precisely because nobody can do it.

In the end, once we get over the transition period of this move to cloud everything, datacenters end up like utilities


One issue with data centers in a box and scaling is physical security. It's possible to pack these data centers into containers, but physical access controls and site security is something in which there are cost benefits to grouping lots of them together. Also there are big benefits to locating these things next to cheap power. So it ends up being more economical for a variety of reasons to have these things a bit less tiny and geographically distributed using today's technologies.


Sun tried the data center in a shipping container. I believe it was named Project Blackbox.


What really struck me from this is the complete move over to software based networking. If all the big cloud players do this, and if you think a lot of infrastructure will be moving over to the cloud (because of cost pressures) over the next few years, what does that mean for Cisco and other big sellers of hardware based networking gear? Is that whole business going to go away?


Beyond using overlay networks from the hypervisor and doing things like load balancing in software with a bit of extra special sauce, there isn't really as much "software defined networking" going on in these places. Certainly no magic as articles like this make out.

In the underlay it's all still very much basic, old fashioned networking using existing protocols: BGP in the DC at FB and MSFT, and BGP, ISIS, MPLS in the WANs. (AMZN likes to pretend they're special so I won't comment on them.)

Google is a bit different, in part I think this is due to their culture and early scale they were forced to drive a lot of developments wrt merchant silicon use, and this naturally led them to their own network OS solutions, with a simplified semi-centralized IGP solution for their DCs, and centralized-TE solution for their inter-DC WAN. It's not entirely clear to me if they would still feel the need to do this if they were starting again.


>Is that whole business going to go away?

No. It's just going to look different, with the most successful vendors selling flexible, nearly white-box hardware with good long-term maintenance terms.

> if you think a lot of infrastructure will be moving over to the cloud (because of cost pressures) over the next few years

There's certainly going to be a TON of infrastructure moving to 'the cloud' in the next 5-10 years, but cost isn't necessarily the strongest driving factor pushing cloud adoption. Big, established Enterprise-with-a-capital-E type businesses are often quite capable of continuing to run infrastructure in-house with better bang-for-buck in terms of raw capacity than what is presently capable with public/hosted private cloud solutions.


I'm guessing the gross margin on "nearly white-box hardware" (even with the maintenance contracts) is a lot less than the 64% Cisco is doing now!


I guess they'll still have their office gear like SMB, firewall, WAN, etc. to lean for a while.


> they'll still have their office gear like SMB, firewall, WAN, etc. to lean for a while.

A while? More like forever. Short of corporations abandoning the notion of shared physical workspaces (which isn't likely to happen in the foreseeable future), corporate MDF, IDF, and edge gear will always be around in some form or another.


People are already starting to talk about whiteboxing the branch office by running all the "value" as VMs (VNFs) on a generic x86 box.


Even in that scenario they still have a bit of an edge because they can still take advantage of their lead on software by turning their stuff into virtual appliances.



If you pay attention to Cisco's communications and moves you will see what they are doing - moving from hardware to software (an operating system and related services) and letting go thousands of hardware engineers.


Cisco's software has been utter rubbish for the entire lifetime of the company. Without the hardware why would anybody go to Cisco for software? CIOs are just stuck in a mental rut?


Because you're trained, certified and experienced on Cisco's proprietary and cruddy command line. Why wouldn't you push for that and a high salary, instead of something generic for a low salary?


Cisco is one of the top most beloved b2b brands. I'm talking Apple level attachment here.


I was also surprised by this. I thought networking was an area where dedicated hardware had an obvious advantage, because you need to do simple, well-specified things very fast.


Cisco is big into SDN. We've just finished a rollout of ACI where I work.


1.02 Winter PUE is very impressive.

I took a class on design for low PUE implementation and some comments from government data center technicians who said, in order to comply with the federally mandated PUE requirements, that people were leaving on or turning on zombie boxes to up their IT load.


>I took a class on design for low PUE implementation and some comments from government data center technicians who said, in order to comply with the federally mandated PUE requirements, that people were leaving on or turning on zombie boxes to up their IT load.

AFAIK, the DCOI sets a target PUE of 1.5 or less, so running unnecessary workloads to meet the target PUE doesn't make much sense. I would bet there was some other kind of tomfoolery going on (hiding the fact that the DC overspent on efficiency when building out/upgrading the facility, or something along those lines).


I took the parent's comment to mean: the datacenter had a relatively fixed amount of power going toward the infrastructure, so they'd turn on additional servers to add more power to the compute side of the ratio. They'd be wasting power but the ratio of power spent on servers to power spent on infrastructure would look better.


Power electronics losses (in watts loss per watts delivered) can be well-approximated as having a constant term + linear term + quadratic term. If the load is very low, then the power delivery equipment will still be spending its constant term losses, which may be large compared to the total actual load.

But the PUE measurement system doesn't know how the various severs are spending power, just how much they are spending. So a busy wait, computing pi, dynamic language hash table lookups, or doing useful work all look the same.



I immediately thought of that xkcd with this quote.

My OpEx with the new datacenters is that I have to change the filters, and that is really the only maintenance I have. And we have moved to a resiliency configuration where I put more servers in each box than I need and if one breaks, I just turn it off and wait for the next refresh cycle. The whole OpEx changes with the delivery model of the white box. So we learned quite a bit there, but now we have got to really scale.”


What do they use for SDN or is it proprietary? My impression is that this is a major part of each large cloud provider's secret sauce.



Microsoft doesn't use OpenDaylight.

I've never heard of any serious non-research uses of it, and I've spent a good bit of time looking. Every time I've heard a rumor, it had turned out to be false.


Microsoft is putting much of their secret sauce in Windows Server, Azure Stack, and SONIC. It will be interesting to see whether this will be successful for them or their customers.


>Microsoft shifted from outside air cooling to adiabatic cooling, where air is blown over screens soaked with water in the walls of the datacenter to create cool air through evaporation of the water.

I was wondering about that sentence, isn't humidifying air a bit problematic in a datacenter? Computers and water usually don't get along that well..


Is this what gives Microsoft that infinite scalability they are advertising now? I'm still not sure how that works.


Nothing is infinitely scalable. Someone is lying if they tell you otherwise.


It would be interesting to know what MSFT's cost to run apps per users is. That would be a better measurement that captures total cost, rather than just the PUE.


That metric primarily captures efficiency and activity of the app, rather than efficiency of the DC. A hello world app would look amazingly efficient by this measure, because it doesn't do anything, so you're basically just capped by how many connections a box can support. A video processing app would look amazingly inefficient just because it's doing so much work.

An infrequently-used app also looks better by this metric than a frequently used app. The service behind the mobile weather app you use might look more efficient than Facebook, just because you use it once a day instead if a dozen.

Disclosure: Microsoft employee, not involved in our data center designs.


interesting, i live near Quincy and know a number of people that are helping build their new DCs and a couple that work there. seen many pictures of the inside and how the layout works, cool stuff. what amazes me the most if the crap hardware they run on.


I find it funny that we can now have a super-computer in our pockets, but we still build huge expensive data-centers. I wonder if computation is like roads: The more we can do, the more we need.


It's a different thing reacting to one person than reacting to a million people, especially when those million people's actions are tightly coupled with each other in a social graph. But if you could fit a million people's state into memory AND also handle a very fast flow of messages at one physical box, well that's probably good enough for government work! But we love stateless, wasteful architectures (wasting server CPU time, memory, and network bandwidth) so that means you need 1000 machines to support 1 million people.

(That's not entirely fair since locality also requires some overhead. One machine that supported a million users might be great if all of those people were in one city. But usually that's not the case, so at the very least you need a box in the top 100 cities (by whatever measure, the simplest being population), at least.)


Interesting remark. I make the same calculation as yours and I also find that there currently is 1000 servers per million inhabitants on Earth today (between Amazon, Rackspace, MS and DO). In other words, one server serves 1,000 inhabitants. But given that not everyone has access to technology, let alone pay for Cloud services, I've estimated that we've provisioned 1 servers to serve all online services (bank, electricity, govt, OSS, Netflix, Volkswagen software, etc) for 100 citizen in developed areas.

It's an extremely bad, resource-intensive architecture ;)

The number of servers in private companies might be around 1 per citizen, and the number of processors per human around 100x (incl. mobile phone, tv, smart lamps). Given a proc has 5m transistors and humans have 100m neurons... our architecture is so bad that we're already outnumbered by machines by a factor of 20 at least.


It's more like we have a terminal for a super computer in our pockets than an actual supercomputer. A lot of a phone's tasks are offloaded on servers.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: