Visiting Facebook's Hardware Labs

We visited Facebook's hardware labs in September, an experience resembling entering the chocolate factory from Charlie and the Chocolate Factory; though the machinery was far less enjoyable to chew on. More importantly though, we were already familiar with the 'chocolate', in that by reading the specifications and following OCP related news, most of the systems present in their labs we could point out and name.


Wannes, Johan, and Matt Corddry, director of hardware of engineering in the Facebook Hardware labs

This symbolizes one of the ultimate goals for the Open Compute project: complete standardization of the datacenter out of commodity components that can be sourced from multiple vendors. And when the standards do not fit your exotic workload, you have a solid foundation to start from. This approach has some pleasant side effects: when working in an OCP powered datacenter, you could switch jobs to another OCP DC and just carry on doing sysadmin tasks -- you know the system, you have your tools. When migrating from a Dell to HP environment for example, the switch will be a larger hurdle due to differentiation by marketing.

Microsoft's Open Cloud Server v2 spec actually goes the extra mile by supplying you with an API specification and implementation in the Chassis controller, giving devops a REST API to manage the hardware.

Intel Decathlete v2, AMD Open 3.0, and Microsoft OCS

Facebook is not the only vendor to contribute open server hardware to open compute project either; Intel and AMD joined pretty soon after the OCP was founded, and last year Microsoft joined the party as well in a big way. The Intel Decathlete is currently in its second incarnation with Haswell support. Intel uses its Decathlete motherboards, which are compatible with ORv1, to build its reference 1U/2U 19" server implementations. These systems are seen in critical environments, like High Frequency Trading systems, where the customers want a server built by the same people who built the CPU and chipset, just so it all ought to work well together.

AMD has its Open 3.0 platform, which we detailed in 2013. This server platform is AMD's way of getting its foot in the door of OCP hyperscale datacenters, certainly when considering price. AMD seems to be taking a bit of a break improving its regular Opteron x86 CPUs, and we wonder if we might see the company bring its AMD Opteron-A ARM64 platform (dubbed 'Seattle') into the fold.

Microsoft brought us its Open Cloud Server (v2), systems that basically power all of Microsoft's cloud services (e.g. Azure), which is a high-density blade-like solution for standard 19" racks.


(Image Courtesy Microsoft)

A 12U chassis, equipped with 6 large 140x140mm fans, 6 power supplies, and a chassis manager module carries 24 nodes. Similar to Facebook's servers, there are two node types: one for compute, one for storage. A major difference however is that the chassis provides network connectivity at the back using a 40 QSFP+ port and a 10 SFP+ port for each node. The compute nodes mate with the connectors inside the chassis, the actual network cabling can remain fixed. The same principle is applied to the storage nodes, where the actual SAS connectors are found on the chassis, eliminating the need for cabling runs to connect the compute and JBOD nodes.

A V2 compute node comes with up two Intel Haswell CPUs, with a 120 Watt maximum thermal allowance, paired to the C610 chipset and with 16 DIMM DDR4 slots to share, for a total memory capacity of 512GB. Storage can be provided through one of the 10 SATA ports or via NVMe flash storage. The enclosure provides space for four 3.5" hard disks, four 2.5" SSDs (though space is shared between two of the bottom SSD slots), and a NVMe card. A mezzanine header allows you to plug in a network controller or a SAS controller card. Management of the node can be done through the AST1050 BMC providing standard IPMI functionality, in addition a serial console of each node is available at the chassis manager as well.

The storage node is a JBOD in which then 3.5" SATA III hard disks can be placed, all connected to a SAS expander board. The expander board then connects to the SAS connectors on the tray backplane, where they can be linked to a compute node.

Benchmark Results Networking
Comments Locked

26 Comments

View All Comments

  • SuperVeloce - Wednesday, April 29, 2015 - link

    From Mass storage: "Compared to hard disks optical media touts greater reliability, with Blu-ray discs having a life expectancy of 50 years and some discs could even be able to live on for a century."

    Yeah sure. Like my expensive gold color cd's from different vendors, baked on different high quality writers, now mostly not working anymore after some 15-20 years. Despite being held in almost perfect environment all these years
  • Uplink10 - Wednesday, April 29, 2015 - link

    Someday they are going to figure out that:
    -SAS HDDs are costlier but if you are using RAID it does not matter, they should use consumer drives and not overpriced enterprise drives
    -I calculated sometimes back if Bluray cold storage is cheaper than HDDs but it is not and more so you cannot change the data once you write it, it is better to go with HDDs
  • toyotabedzrock - Wednesday, April 29, 2015 - link

    You have to wonder what these networking chip vendors are hiding in the firmware that makes them so resistant to open sourcing the code.
  • Casper42 - Monday, May 4, 2015 - link

    Johan, some of the HP info at the end was interesting, but incomplete.
    If you (or anyone reading this) plan to talk to HP, they will also talk about their relatively new CloudLine "CL" type machines as well.
    They come in standard 1RU/2RU designs as well as OpenRack designs coming soon.
    And the SL line is all being morphed over to Project Apollo which uses the XL prefix.
    Apollo 2500 is now live, 4X00 will replace SL4500, 6000 has already replaced S6500, and the 8000 was a net-new add for Gen9 focused on big HPC farms.
    So anything SL is, or soon will be, a dead platform. (The SLs you mention could be an exception since they are not widely commercially available)
  • Netpower - Tuesday, June 2, 2015 - link

    One general problem with this design is how to take care of power line disturbances entering the power shelves via the 277V AC lines. The 48V DC is filtered via the 48V battery but you must add a filter/power line conditioner somewhere to make sure that transients and sags doesn't kill your power shelves. The 380V DC approach by (http://www.emergealliance.org) is much more reliable and still have all the advantages with higher efficiency, lower cable losses etc.
  • Astana - Wednesday, September 16, 2020 - link

    It is believed that the nicotine addiction of a smoker prevents you from quitting. But modern research has shown that the quitting process is complicated by psychological addiction. That's why many people can't cope with the habit of smoking. But there are results with https://heroindetoxeurope.com/ drvorobjev hospital they're all successful.
    The problem of smoking is not only the problem of men, but also increasingly - women. Many women are wondering: "How can you quit smoking and not get better? Psychologists have proven that when a woman quits smoking, she subconsciously tries to replace, to compensate for the process of smoking with food. This feeling is also familiar to men who quit smoking, but they do not attach much importance to it. As a result, very soon an extra 3-4 kg or more will appear.

Log in

Don't have an account? Sign up now