Slashdot is powered by your submissions, so send in your scoop

 



Forgot your password?
typodupeerror
×
Facebook Hardware

Will "Group Hug" Commoditize the Hardware Market? 72

Will the Open Compute Project’s Common Slot specification and Facebook’s Group Hug board commoditize the data center hardware market even further? Analyst opinions vary widely, indicating that time and additional development work may be necessary before any sort of consensus is reached. At the Open Compute Summit last week, Frank Frankovsky, director of hardware design and supply chain operations at Facebook, announced both the Open Slot specification and Facebook’s prototype Open Slot board, known as “Group Hug.” Group Hug’s premise is simple: disaggregate the CPU in a way that allows virtually any processor to be linked to the motherboard. This has never been done before with a CPU, which has traditionally required its own socket, its own chipset, and thus its own motherboard. Group Hug is designed to accommodate CPUs from AMD, Intel, and even ARM vendors such as Applied Micro and Calxeda.
This discussion has been archived. No new comments can be posted.

Will "Group Hug" Commoditize the Hardware Market?

Comments Filter:
  • if only it came true. Even if politics weren't involed, it still wouldn't be easy at all I imagine. It'd require processors to either rely on a standard memory controller, or to implement their own, along with all sorts of other similar challenges with performance vs. compatibility.
    • by Jeng ( 926980 )

      I would figure the memory would be on the daughter card with the processor. That way the main motherboard wouldn't need to be compatible with all the different memory choices, just have to be compatible with the daughter card.

      • by h4rr4r ( 612664 )

        If memory and CPU on are the daughter card, how is this any different than a blade chassis?

        Seems like doing that remove pretty much all the value from the project.

        • It's not any different from blades. Actually group hug is not hotswap, so it's worse than blades but probably cheaper.

        • by MBCook ( 132727 )
          It's basically a return to the backplane days of the 8080/8086, except that memory has to stay on the CPU card for speed reasons.
          • by Amouth ( 879122 )

            just what i was thinking, looked over at an old self and wondered if i could sell them an old backplane box as an example of how to make it work.. it was so nice to just drop in another CPU card as you needed.

          • by swb ( 14022 )

            Why not make memory its own card type and have optical interconnects for memory? That should allow enough speed for memory access and with a common interface standard you could design your CPU to do it natively or have a translation controller on your CPU card.

      • I think the is the basic idea, which is why the whole idea won't work. Basically, they are sawing the motherboard in 2, where the CPU and memory are on the daugterboard, and the rest of the components (SATA,USB3, PCIe slots, sound, video outputs) are all that remain on the motherboard. I think it might provide for some interesting ideas where you could have a CPU board that accepted a new chip without having to go out and re-buy the part that the peripheral plug into. Or you might be able to get a mothe
        • by mlts ( 1038732 )

          I wouldn't mind a system of going to a completely passive backplane architecture, although with electrical signal distances, this likely wouldn't be really doable until we have the ability to get optical signals onto the fiber from the chip die itself (which means a lot of muxing/de-muxing since having tons of optical connections would be a lot harder than solder pads.)

        • I think the is the basic idea, which is why the whole idea won't work. Basically, they are sawing the motherboard in 2, where the CPU and memory are on the daugterboard, and the rest of the components (SATA,USB3, PCIe slots, sound, video outputs) are all that remain on the motherboard

          Why would it work any less than a graphics card? Isn't that the same? GPU and memory on a daughterboard with a fast interface to the motherboard.

        • by butlerm ( 3112 )

          Basically, they are sawing the motherboard in 2, where the CPU and memory are on the daugterboard, and the rest of the components (SATA,USB3, PCIe slots, sound, video outputs) are all that remain on the motherboard

          It is actually a micro-server architecture. Think small form-factor blade servers with an optional PCIe interconnect, optional remote SATA devices, and one mandatory ethernet interface, all running through what looks like an ordinary PCIe slot, but isn't.
          http://www.opencompute.org/wp/wp-content/up [opencompute.org]

  • by caseih ( 160668 ) on Monday January 21, 2013 @03:18PM (#42649789)

    All I see are links to other slashdot articles. Are we going for a new record here? First the ridiculous post about Microsoft welling their entertainment division, now this. And the same style of headline too, which of course is answered with, "No."

    Mr Editor, can you at least post a link to some information, like maybe the site where this specification is detailed? Maybe the project web site itself?

    • Mr Editor, can you at least post a link to some information, like maybe the site where this specification is detailed? Maybe the project web site itself?

      The editors have been outsourced. Now, a team of twenty people who have english as an eleventh language review every submission and green light only those that meet the criterion spelled out in the three ring binder. The three ring binder itself was created from a 7 line Perl script, written by a subcontractor from China, who was hired by a contractor for Dice, who recently acquired the Slashdot brand identity, who shows up once every two weeks to collect his paycheck and update the seed in the random numbe

    • If they have actual links to real articles, then it isn't nearly the electronic masturbatory exercise you see in front of you here.

      You can't link to your own shit if you have real information to link to...

    • by butlerm ( 3112 )

      Here is a link to an actual specification. If you read it, you will see that about half of what has been written about this announcement is wildly off base. We are talking micro-servers here - complete with on board cpu, ram, boot eeprom, flash storage, and ethernet. PCIe and SATA connections to the backplane are optional. Think small form factor blade server.
      http://www.opencompute.org/wp/wp-content/uploads/2013/01/Open_Compute_Project_Micro-Server_Card_Specification_v0.5.pdf [opencompute.org]

  • S100 anyone? (Score:4, Insightful)

    by cdrguru ( 88047 ) on Monday January 21, 2013 @03:19PM (#42649809) Homepage

    One architecture that supported "variable CPUs" was S100 where it is was typical to have a CPU card, one or more memory cards, and multiple I/O cards all plugged into a backplane. There were CPU cards for the Apple ][, but these were complete computers on a card that simply allowed use of the Apple ][ I/O.

    Given today's multi-gigahertz processors with gigahertz memory access, I would think it would be difficult, if not impossible to effectively separate the CPU and the memory by very much. Similarly, it gets pretty complicated with high speed DMA I/O when you move it away from the memory it is accessing. I'm sure it could be done, but the performance is going to suffer just from the physical distances. Add in connector resistance and noise and you have ample justification for putting the CPU, chipset and RAM in a very small module that then plugs into the rest of the computer for I/O.

    • That's a motherboard. CPU, RAM, and chipset are on the motherboard. There's some PCI slots to plug other shit into.
    • by vlm ( 69642 )

      You would probably find googling for N8VEM SBC v2 to be very interesting. S100 lives! as does a eurocard connector-ized version of the same idea, more or less.

      http://n8vem-sbc.pbworks.com/ [pbworks.com]

      I have the partially assembled system on my workbench. I need a nice blizzard to keep me inside soldering, that'll take care of it. Its all antique thru-hole instead of modern SMD which I find harder to work with and certainly much bigger but its no big deal.

      Add in connector resistance

      At least the n8vem design has a standard pc molex on the ecb

    • by Anonymous Coward

      Yeah, I loved the "This has never been done before with a CPU" as I have a couple of S100 systems sitting in my Shed. I am constantly amazed by todays "youth" who know so little (read nothing) of computing's past.

      The late 70's and early 80's was probably the era of greatest diversity of computing ideas there has ever been, and perhaps it was even more "open" than today as users could buy complete service manuals for their computers (and have a good chance of fixing them!), there was a ton of info about the

      • almost everything we see in consumer devices has been done before in some market, or at the NSA (the latter of which will not talk about it, but we know because of James Bamford)

      • I came along just a bit later, but this part especially is true:

        "users could buy complete service manuals for their computers (and have a good chance of fixing them!), there was a ton of info about the OS's"

        The virtual complete absence of true user manuals to this day baffles/angers me.

        When I took 'computer class' in the mid-90s we still learned mostly in versions of DOS and we used 5 1/4 and 3 1/2 floppys (mostly the latter).

        We could afford 2 computers that could run the current version of Windows.

      • by mlts ( 1038732 )

        Probably one of the better magazines I bought was the old Computer Shopper, before it shrunk into a "regular" size magazine. Stain Veit's articles were always a treat, and even the ads were useful, back when there were tons of white-box makers (Arche, Bell, Austin PC, etc.)

        The early Mac magazines were like this as well. If you had a special device that could scan, you could actually scan a page out of the magazine and have a couple useful applications each month.

        I do miss the good magazines that just don'

    • by Agripa ( 139780 )

      Not all of the CPU cards for the Apple ][ had on board memory. The popular Z80 Softcard used the motherboard memory which made it slower than other Z80 expansion cards.

    • Given today's multi-gigahertz processors with gigahertz memory access, I would think it would be difficult, if not impossible to effectively separate the CPU and the memory by very much. Similarly, it gets pretty complicated with high speed DMA I/O when you move it away from the memory it is accessing. I'm sure it could be done, but the performance is going to suffer just from the physical distances. Add in connector resistance and noise and you have ample justification for putting the CPU, chipset and RAM in a very small module that then plugs into the rest of the computer for I/O.

      If they were just moving the CPU to a card then yes, but apparently they aren't:

      Intel, another key member of the Open Compute Project, announced it would release to the group a silicon-based optical system that enables the data and computing elements in a rack of computer servers to communicate at 100 gigabits a second.
      More important, it means that elements of memory and processing that now must be fixed closely together can be separated within a rack, and used as needed for different kinds of tasks.

      htt [nytimes.com]

      • by butlerm ( 3112 )

        More important, it means that elements of memory and processing that now must be fixed closely together can be separated within a rack, and used as needed for different kinds of tasks.

        This statement is in reference to Intel's proposal, which is still vaporware. I seriously doubt they are talking about locating main memory away from the processors. That would more or less be suicidal.

        Facebook's design certainly does no such thing.
        http://www.opencompute.org/wp/wp-content/uploads/2013/01/Open_Compute_Project [opencompute.org]

        • Well yes, like I said the concern would be there only if they were just moving the CPU to an external card - which they aren't doing in the server card spec (they are proposing memory goes on the card as well) and they wouldn't be doing with the Intel technology which provides additional highspeed optical links, either one of those solutions mitigates the issues in the first post i responded to.
    • by Anonymous Coward

      Add to that CompactPCI, VME, VME64, VME64x, PXI, VXI, VXS, VPI, OpenVPI.

      I probably forgot some, but it seems there is more computer-bus/formfactors that don't call for a specific CPU than do.

    • by ncc74656 ( 45571 ) *

      There were CPU cards for the Apple ][, but these were complete computers on a card that simply allowed use of the Apple ][ I/O.

      Most of these were just a CPU (usually a Z80) and the minimal logic necessary to take over from the 6502 on the motherboard. A relatively small handful of cards included their own RAM; it was far cheaper to use what was already in the computer.

      The only Apple II expansion card that comes to mind that really was a complete computer on a card was the Applied Engineering PC Transporte [applearchives.com]

  • Also see wire-wrapped and bit-sliced.

  • The CPU is a small part of the cost of the server

    What is the point in doing this? Where is the return on investment?

    • I'm not sure what the idea is, but it seems they hope it will open up the market. Suppose there's a company putting a new CPU on the market. Today, they have to come up with a motherboard as well, or convince one of the big boys to design one around their new CPU. With this architecture they only need to design the much simpler CPU card around their CPU. This lowers the barrier to entry and means more competition, which is nice for big datacenters like Amazon and Facebook, who buy servers by the boatloa
      • by Wolfrider ( 856 )

        --Yeah, I don't know if this is really going to take off. (In general) Universal = generic = NOT optimized for speed/efficiency, etc...

    • Well, this is to show those scumbags at Intel and AMD that refuse to create products for their competition who's boss! I mean, why wouldn't they spend extra time and money to create a bunch of connections that their customers aren't going to use, and probably make their products perform worse by introducing unneeded complexity?

      Never mind that we did already have a "universal" CPU socket, or at least one as close as it mattered. It was called Socket 7, and it fit Intel / AMD / VIA CPUs. And it was abandon

    • by butlerm ( 3112 )

      It is not a CPU slot specification. It is a micro-server slot specification, which is much more practical. Think small form factor blade server. The PCIe part is actually optional.

  • I do no think it means what you think it means. Something that is a commodity product is fungible, meaning any indiviual product from any of various vendors is effectively interchangable with any product of the same kind from any other vendor. Computer hardware has been commoditized for a long time. While processors are not wholely interchangable (AMD vs Intel), the motherboard/CPU combo generally is. Everything else in a computer can more or less be swapped out with a different brand with the same or sim
  • by Junta ( 36770 ) on Monday January 21, 2013 @04:01PM (#42650323)
    If you want it to be, server hardware is already commoditized. All the priciest components are interchangeable (you *can* buy DIMMs from wherever and cram them in your server from a technology standpoint). Apart from Intel and AMD playing this game where the IOH is generally affine to some particular CPU generation, things are already there (and the IOH is a pretty inexpensive part however you slice it, and could already be subbed out for a PCI-e constructed device if they saw fit). Now the catch is that for *most* traditional IT shops, there continues to be value in well-integrated systems. This will not change that picture.

    I see this as another step toward two goals:
    -Getting ARM into the datacenter in some reputable fashion (which may or may not make sense, depending on whether a compelling performance per watt case can be made that offsets the energy/manufacturing gap that might be incurred from requiring more packages to get to performance desired).
    -Rebranding 'whitebox'. White box vendors are viewed as the low cost alternative to HP/Dell/IBM, but image wise they are viewed anywhere between 'unacceptably bad' to, at best, 'just as good' from a select portion of the market. Putting cost aside, no one thinks of white box as 'better' than the expensive names. A lot of open compute at the system level is the same thing that has been the reality for the last decade with a shiny new name. The same standards that everyone already followed are getting highlighted more explicitly. This is the opportunity, through marketing, to change minds to say 'better' in some cases or at least make the 'unacceptable' segment of the market take another look.
  • You could probably have an ARM, low load, low energy comsumption processor and a nice High-Performace processor on the same board. You'd just then manage when the high-performance activates, and you could probably switch any (assuming hot-plug) without taking it offline.... It's nice to dream, isn't it?

  • by Joe_Dragon ( 2206452 ) on Monday January 21, 2013 @04:07PM (#42650381)

    do the cards have room for 4-8 / 6-12 ram slots each? and yes that's full size ram.

    • In a setup like this you wouldn't put the RAM on the CPU card. It'd go on the backplane interconnect, independent of the CPU. Think the PDP-11 Unibus or the VAX-11 Synchronous Backplane Interconnect, which are where I first encountered the concept of a backplane and independent CPU, memory, co-processor and I/O processor modules. I doubt they originated there, though, my guess is the concept goes back to the IBM mainframes of the 60s. It was an amusing cycle: external modules would migrate onto the CPU boar

      • by butlerm ( 3112 )

        The memory is going to stay on the processor cards. It would be somewhere between slow and ridiculously slow (by modern standards) to do anything else. The slot interface is PCIe x8. An I/O interconnect. Not memory, certainly not SMP. More like a tightly coupled cluster.

    • by butlerm ( 3112 )

      No. These are micro-servers we are talking about. Two RAM slots typical. Low power energy efficient CPUs too. ARM to start with.

      Take a look:
      http://www.opencompute.org/wp/wp-content/uploads/2013/01/Open_Compute_Project_Micro-Server_Card_Specification_v0.5.pdf [opencompute.org]

  • pci-e X8 is limited IO why not at least X16?

    X8 can be used up by 1 video card on it's own.

    • by butlerm ( 3112 )

      This is intended for server use. No video output required. The PCIe part is actually optional. I wouldn't expect to see this in workstations anytime soon, not without a major redesign at any rate. The form factor is designed for small, low power processors. The interface is not designed for SMP or anything like that either.

    • by dacut ( 243842 )

      They're only using the PCIe x8 physical connectors; the electrical signals do not resemble PCIe at all.

      Presumably, they're also relocating the actual slot location to avoid stupid errors (like plugging one of these into an actual PCIe x8 slot or vice-versa).

  • I'll wait for Happy Ending.
  • by niks42 ( 768188 ) on Monday January 21, 2013 @04:36PM (#42650643)
    Looking at the photos of the backplane, it looks like S100 era technology. Where is the trendy stuff? I want to see hermetically sealed illuminated glass-like blocks, changing color and sliding out automatically on detection of failure, a high-bandwidth optical interface on each edge, power inductively coupled to avoid metal connectors, an eerie surround sound voice saying "Dave ... Stop" ..
    • In other words, an iPhone. Minus the data connectors and inductive power. And the rack.

      Ya know, I think you're on to something. If Intel can take their optical interconnect out of the lab, it just might be possible. At a reasonable price. It's possible now, for an unreasonable price.

      And that's the machine SGI would be building, if SGI were anything but a shadow of its former self.

  • And add in some optical links so we can finally scale motherboards to something awesome.

    Being limited to certain designs / lengths because of electrical circuitry...madness.

  • What I don't see in TFA is something that describes how the one big hurdle of this type of design will be overcome- the IMMENSE costs! The speeds that processors and RAM runs is so high you can't just drop it down on a board and expect it to work- you're in a long loop of simulate, build, test, repeat, and each iteration is extremely costly- we're not talking Arduino here- in reality, these boards (populated) in mass production will cost hundreds in just BOM costs, not counting the assembly. If the biggies

Without life, Biology itself would be impossible.

Working...