[HN Gopher] Oxide builds servers as they should be [audio]
___________________________________________________________________
 
Oxide builds servers as they should be [audio]
 
Author : tosh
Score  : 129 points
Date   : 2022-07-09 17:15 UTC (5 hours ago)
 
web link (changelog.com)
w3m dump (changelog.com)
 
| alberth wrote:
| Who's the target buyer of Oxide?
| 
| I ask genuinely ... because I don't understand who will spend the
| premium for a rack server that is easier to maintain?
| 
| In a world of cloud and dedicated hosted servers (where the users
| of the server is not the buyer of the server) - who would buy
| Oxide?
| 
| It seems like whoever can build the cheapest server wins now.
| 
|  _Said differently, cloud has completely changed the value-chain
| in server buyers_
 
  | walrus01 wrote:
  | ultimately somebody has to buy and own and operate the
  | underlying hardware of the "cloud".
  | 
  | obviously AWS, azure, etc do their own hardware procurement and
  | totally bespoke stuff in house.
  | 
  | so this would be targeted at any smaller players in hosting
  | companies that are not at that massive size of azure, but need
  | vast quantity of servers.
 
| brianzelip wrote:
| Slightly off topic, but hey Jerod if you're reading, here's a
| vote to keep up the weekly brief Changelog episodes!
 
| fio_ini wrote:
| Awesome product. How much $$$?
 
| nimbius wrote:
| according to the website for oxide theyre "Hardware, with the
| software baked in" so I'm worried this is going to be more like
| Cisco's hamfisted UCS platform or dell/HP lock in with drac and
| ilo and "hyperconverged" embedded Java fuckery that can never be
| patched.
 
  | sokoloff wrote:
  | I've been relatively satisfied with HP's iLO offering, in that
  | it's been sufficient for my needs, avoids many pointless
  | fly/drives to the colo (or dealing with the comic misnomer that
  | is "smart hands"), and is still getting the occasional update
  | on 8-year old servers.
  | 
  | Even for servers in my building, I find it's more convenient
  | than a crash cart for most things (and of course my desk is a
  | better work environment).
 
  | mlindner wrote:
  | All of the source will be open and re-flashable from what I've
  | read. So I assume that won't be a problem.
 
  | steveklabnik wrote:
  | I wrote a while back about the difference between
  | "hyperconverged" and "hyperscalar":
  | https://news.ycombinator.com/item?id=30688865
 
| smabie wrote:
| Seems like this company has gotten a lot of hype but what so they
| actually do? Cloud hosting? Or selling physical servers? Both of
| those things sound like a terrible business tho..
 
  | stingraycharles wrote:
  | From what I gather, it's extremely beefy, physical servers with
  | DPUs built-in. I.e. perfect for large on-premise deployments
  | while maintaining the high performance that DPUs enable, all as
  | a single appliance.
  | 
  | It's the kind of thing I would expect Nvidia or Dell to become
  | big at as well, and Bryan Cantrill has enough leverage to pique
  | my interest significantly here.
 
    | wmf wrote:
    | I don't think they're using DPUs.
 
      | jjtheblunt wrote:
      | what's a DPU?
 
        | unixhero wrote:
        | Datacenter processing unit, it's a rather new thing
 
        | gertrunde wrote:
        | Also known as: SmartNIC, and a bunch of other names,
        | basically a NIC with an embedded CPU (usually ARM of some
        | sort), so can offload some workload into the NIC.
        | 
        | An example is nvida's bluefield, there are others
        | 
        | That sort of thing is being used in a few areas, can even
        | run ESXi on them.
        | 
        | There was a fun article on servethehome a while back
        | using them to build an arm cluster (Link:
        | https://www.servethehome.com/building-the-
        | ultimate-x86-and-a... )
 
  | qbasic_forever wrote:
  | Even a cursory glance of their website or listen to the episode
  | pretty clearly explains they're building server hardware for
  | hyperscalers. I.e. they want to build and sell the hardware
  | that your hosting provider would buy to host your applications.
  | I would guess they see quite a market for on-premises
  | computing, particularly with companies that through legal or
  | competitive reasons cannot or do not want to lose control of
  | their data by hosting it on a cloud service.
 
    | ec109685 wrote:
    | Haven't fully grocked how this will get huge. If there is a
    | big market here, won't the AWS Outposts of the world win this
    | market?
 
      | jbnorth wrote:
      | Disclaimer: I work for AWS
      | 
      | I think Outpost is a great hybrid-ish solution for those
      | companies with workloads in AWS that want to supplement
      | that with on-premise workloads using the same APIs and
      | tools they are already used to.
      | 
      | The use cases I see for Oxide are, like other have said,
      | are for hyperscalers or hosting providers who want to have
      | their own on-premise infrastructure that isn't tied to
      | another entity like AWS. Whether that's because you have
      | the in-house talent to manage it or for compliance reasons
      | it's required it does have a niche.
 
        | swyx wrote:
        | ok so.. what exactly counts as a hyperscaler? and how
        | many of them are there?
 
      | spamizbad wrote:
      | Fwiw I work with an engineer that came from a shop that
      | used them. Long story short: they're not remotely cost
      | effective, are quite limited, and existing AWS tool chains
      | run into all sorts of hair-pulling snags.
      | 
      | My gut is that they are for places that need on-Prem for
      | compliance purposes for certain parts of their stack
 
      | chucky_z wrote:
      | AWS Outposts are tragically limiting in so many ways. If
      | you are a small shop they are great but if you don't fit
      | their model exactly you are SOL.
      | 
      | Same problem with Anthos.
      | 
      | You can work around the limitations by having a million
      | points of presence with them and splitting up your
      | workloads but there's real costs associated with that
      | model. Running 1 rack in 100 DCs is a lot harder than 10
      | racks in 10 DCs.
 
  | wmf wrote:
  | It's a private cloud rack aka hyperconverged infrastructure
  | rack but they're allergic to standard terminology for some
  | reason. I assume the market is cloud repatriation for startups.
 
    | mlindner wrote:
    | Wikipedia says that hyperconverged infrastructure typically
    | uses commercial off-the-shelf computers. So that's likely why
    | they're allergic to it. As they're designing their own
    | computers so the term isn't accurate.
    | 
    | https://en.wikipedia.org/wiki/Hyper-converged_infrastructure
    | 
    | Oxide racks won't be able to use anything other than Oxide
    | computers (or maybe eventually third party computers that
    | follow Oxide standards).
 
  | [deleted]
 
| bombas wrote:
| I don't understand how Oxide is any different than what VCE
| attempted to accomplish from a converged infra perspective.
 
  | mlindner wrote:
  | Didn't VCE still use consumer rack computers? All they were was
  | basically a hardware reseller with their own racks. They re-
  | sold Cisco stuff. They just repackaged everyone else's stuff as
  | a value added service.
 
    | rtp4me wrote:
    | Disclaimer: I worked for VCE (now Dell/EMC)
    | 
    | VCE used Cisco C2xx and C4xx rackmount and BLxxx blade
    | servers (before the EMC/Dell acquisition). Now, I believe
    | they use all Dell hardware.
 
  | Tuna-Fish wrote:
  | They designed and built their own hardware and wrote all their
  | own firmware.
 
| brodo wrote:
| Here is Bryans application video for YC120 2019:
| https://youtu.be/px9OjW7GB0Q
 
  | nynx wrote:
  | Definitely worth watching
 
| labrador wrote:
| Cranston made a pretty good case for Oxide in a talk he gave -
| his argument as I understood it is that legacy BIOS is
| proprietary and insecure
 
  | gruturo wrote:
  | > Cranston made a pretty good case for Oxide in a talk he gave
  | 
  | Hmmm wrong Bryan? Breaking Bad wouldn't look the same if you
  | swapped them, I'm afraid.
 
    | gfodor wrote:
    | Breaking Bad where Bryan Cantrill runs an illegal Bitcoin
    | mining operation on Oxide servers under a NYC laundromat
 
      | labrador wrote:
      | His nemesis is El Salvadoran volcano crypto cartels, but
      | Cantrill mounts a 51% attack and destroys their Magmacoin
 
      | hinkley wrote:
      | Their driers are really slow because they use the exhaust
      | heat from the server room.
 
      | zja wrote:
      | "Jesse, we have to compile"
 
        | mtlmtlmtlmtl wrote:
        | printf("%s\n", my_name);
 
        | tialaramex wrote:
        | They're a Rust shop, so these days it would be:
        | 
        | println!("{my_name}");
 
      | eduction wrote:
      | The NSA tries to shut down Bryan Cranston so he goes
      | underground, building illegally powerful servers in a
      | commandeered Pixar storage room at night and sleeping under
      | the stage of youth music venue 924 Gilman by day.
 
        | elijahwright wrote:
        | This is the funniest thing I've read today. Thank you!
 
    | labrador wrote:
    | There was another commenter who made the same mistake. I was
    | going to correct them but thought better of it, then made the
    | same mistake myself!
 
| walrus01 wrote:
| what does this accomplish that buying some open compute platform
| based x86-64 servers in whole-rack quantity does not?
| 
| Or a bunch of the the 4-servers-in-a-single-2U-package setup from
| Supermicro?
| 
| And then interconnecting them by your own choice of 100GbE
| switches at top of rack.
 
| Klasiaster wrote:
| What is missing is a middle ground: an improvement over the IPMI
| broken-state-of-things that is aimed for automated deployments,
| without demanding a certain network environment for PXE booting
| and configuration of IP addresses for management.
| 
| I haven't used RedFish but it seems it does not address the
| problems of IPMI in this regard: To be fully automatable there
| shouldn't be an assumption about the network environment - rather
| I would like that to have an L2-based protocol (one has to use
| the MAC address as identifier anyway) or if IPv4/v6 is used then
| at least demand mandatory link-local auto-addressing by default.
| However, RedFish seems to support emulating a CD-ROM drive for
| booting an installation media, this is a nice idea but what I
| really would like to see is directly writing an image to the hard
| disk. Then of coure the hardware manufacturers would need to
| write high-quality firmware and not the regular quirk- and bug-
| ridden stuff that we know from IPMI implementations.
 
  | deivid wrote:
  | I have put our implementation of _exactly_ this in production
  | last week.
  | 
  | Boot a "CD-ROM" via IPMI into a custom installer, flash image
  | (generated from a Dockerfile!) to disk, reboot. Install takes
  | ~15 seconds, full process (going via POST once) takes ~4
  | minutes. This also allows us to have a "hardware validation"
  | image (that one doesn't get persisted to disk).
  | 
  | Not sure if there's any plans to make it public right now, but
  | I'll ask around. Feel free to contact me via email (in profile)
 
    | yjftsjthsd-h wrote:
    | > flash image (generated from a Dockerfile!)
    | 
    | Any hints about how to do this? I'd love to use something as
    | nice to work with as a Dockerfile to build bare metal
    | installs.
 
      | deivid wrote:
      | I wrote down the kernel of the idea quite a while ago
      | here[0], however, the actual version we run nowadays is
      | using UEFI and written in Go (installer) + Python (API that
      | generates the ISO on demand)
      | 
      | Fly.io[1] also does this (although they boot the result in
      | a VM, the concepts are the same)
      | 
      | [0]: https://blog.davidv.dev/docker-based-images-on-
      | baremetal.htm...
      | 
      | [1]: https://fly.io/blog/docker-without-docker/
 
    | Klasiaster wrote:
    | Yes, automation with IPMI can be done (for a particular
    | hardware due to quirks), and I recommend doing it (and I have
    | done it) with a separate DHCP environment for the BMC NICs to
    | be at least agnostic about the OS NIC's network environment.
    | 
    | So, I share your excitement when it works but I'm not happy
    | with the out-of-the box experience of IPMI. I wish there was
    | something better but unfortunately this is just one case of a
    | split/mismatch of what HW people provide and what software
    | people need.
 
    | justinclift wrote:
    | > but I'll ask around.
    | 
    | Please do. There's a need for it. :)
 
  | bcantrill wrote:
  | Your last sentence reveals what you probably already know to be
  | true: there isn't a middle ground. We tried for years to get
  | hyperscaler-class software running on commodity-class servers
  | -- and came to the conclusion that the hyperscalers came to
  | (and as Alan Kay famously observed): to be really serious about
  | elastic infrastructure, one needs to do one's own hardware.
  | Now, ~2.5 years into Oxide, I can substantiate that with more
  | detail that I could have ever imagined: there is no middle
  | ground -- and indeed, this is why the infrastructure world has
  | become so acutely bifurcated!
 
    | mwcampbell wrote:
    | Do you think there's at least a place for a holistic system
    | that's not a whole rack? Or do you think that those of us
    | with smaller computing needs will just have to either rent
    | from a hyperscaler or put up with PC servers?
 
| simonw wrote:
| https://oxide.computer/product                   CPU: 2,048 x86
| cores (AMD Milan)         Memory: Up to 32TB DRAM
| Storage: 1024TB of raw storage in NVMe         Network switch:
| Intel Tofino 2         Network speed: 100 Gbps
| 
| That's a pretty meaty server!
 
  | [deleted]
 
  | boulos wrote:
  | That's for the rack.
 
    | bogota wrote:
    | Was going to say.. if they fit than in a 4U I'm about to make
    | my datacenter 8 times smaller
 
      | pclmulqdq wrote:
      | I saw an IBM-funded research project that fit 1,024 ARM
      | cores and 2-4 TB of memory into a 3U box. The box didn't
      | include the liquid cooling system you needed to keep the
      | thing running.
 
        | wmf wrote:
        | You could get that today with a 2U 4N Altra Max system.
 
  | littlestymaar wrote:
  | Noob here: any idea of the approximate price tag that should be
  | expected for such a monster?
 
    | mjgerm wrote:
    | Looking at the hardware, I'd wag somewhere in the $2-3M
    | range. Depends on how big of a profit margin they want to
    | make or how scrappy they feel like being (could probably go
    | as low as ~$1M if they can make it up in volume).
    | 
    | If they're following the typical 1/5x pricing model for
    | support, that'd be roughly $500k/yr/rack. But it's also hard
    | to do that while simultaneously describing Dell as
    | "rapacious".
 
      | littlestymaar wrote:
      | Thanks!
 
| chairmanwow1 wrote:
| Listened to 18 min before stopping, not really sure companies
| values warranted so much time.
| 
| Furthermore pretty leading question to compare Brian to Steve
| Jobs forced me out.
 
  | trhway wrote:
  | Back from Sun times I remember Brian loving the limelight.
  | Somebody comparing you to Jobs is the pinnacle of a limelight
  | search in our industry.
 
  | sgt wrote:
  | A lot of people have been compared to Jobs. At the very least,
  | Cranston has some substance.
 
    | phpnode wrote:
    | * Cantrill
 
      | swyx wrote:
      | i like that this is a running joke in this thread now
 
        | unixhero wrote:
        | Fork yeah!
 
    | c7DJTLrn wrote:
    | More substance than Elizabeth Holmes at least.
 
  | 1123581321 wrote:
  | He demurred to the comparison and made a solid book
  | recommendation. Seems better to just skip through uninteresting
  | segments with the +45s button.
 
  | [deleted]
 
  | [deleted]
 
| uthinter wrote:
| If anyone from Oxide is here, could you guys restart the "On the
| metal" podcast . It was the only CS related podcast I ever liked.
 
  | JoachimSchipper wrote:
  | "On the metal" was indeed lots of fun; I understand that it's
  | been "succeeded" by Oxide's Twitter spaces,
  | https://github.com/oxidecomputer/twitter-spaces
 
    | mlindner wrote:
    | Twitter spaces are awful however and are also not easily
    | recorded and can't be used easily on any device. (You need to
    | go to some kind of youtube link from one of the participants
    | to listen to a recording.) Also their twitter talks are more
    | unfocused rants by the participants with little focus. So
    | it's hard to listen to. I'm sure it's more fun for the
    | participants, but not so much for the listeners. You also
    | lose the comfort of consistency of participants.
    | 
    | Some of them have even been delving into some of the
    | participants personal politics as well which is just
    | something I'm not interested in hearing.
 
      | sam_bristow wrote:
      | The Oxide Twitter spaces are recorded and distributed as a
      | podcast [1]. If you're not a fan of the content it won't
      | help but it's at least more convenient to consume than
      | through Twitter.
      | 
      | [1] https://feeds.transistor.fm/oxide-and-friends
 
      | bcantrill wrote:
      | We've turned it into a podcast as well, so you should be
      | able to enjoy it wherever you consume podcasts.[0]
      | 
      | While there may be some unfocused rants in there (sorry, I
      | guess?), there's also a lot of extraordinary technical
      | content -- certainly, if anyone else has described their
      | board bringup experiences in as explicit technical detail
      | as we have in [1] and [2], I would love to be pointed to
      | it!
      | 
      | [0] https://podbay.fm/p/oxide-and-friends
      | 
      | [1] Tales from the Bringup Lab: https://podbay.fm/p/oxide-
      | and-friends/e/1638838800
      | 
      | [2] More Tales from the Bringup Lab:
      | https://podbay.fm/p/oxide-and-friends/e/1650326400
 
        | mlindner wrote:
        | Thanks for the link for a podcast version. That's
        | helpful. I did listen to those two talks earlier and
        | those were probably the most interesting. I did dislike
        | that there seemed to be a constant process of
        | interrupting each other in the middle of someone telling
        | a story so the story became quite fragmented as it took
        | some time to come back to the topic (or the topic was
        | forgotten entirely and never elaborated on after the
        | interruption).
        | 
        | Another thing that ended up being frustrating is repeated
        | references to some image and I had to pause and go
        | digging through people's personal twitter accounts to try
        | and find the image that was being talked about. There
        | seemed to be an assumption that the listener follows all
        | employees personal twitter accounts.
        | 
        | There's also audio quality issues in general as it's not
        | nearly as good a quality of media equipment as was used
        | for the original on the metal podcast. (Maybe it's
        | similarly cheap equipment, but it was better quality
        | before regardless.)
 
        | yardie wrote:
        | I love those little interruptions. Like when a guest is
        | speaking on a topic and they bring up a name or company
        | of the past. And that segues into another bit of SV lore.
 
    | uthinter wrote:
    | Thanks. I'll check it out. On the last podcast they had Ken
    | Shirriff and it was what got me hooked.
 
      | bcantrill wrote:
      | We're (obviously!) huge fans of _On the Metal_ [0] too (and
      | the episode you cite with Ken was indeed extraordinary[1]!)
      | -- but we have come to like our _Oxide and Friends_ [2]
      | Twitter Space even more. The reasons why are manifold, and
      | really merit their own long-form piece, but as a few
      | examples of episodes that show why we find it so
      | compelling, see "Tales from the Bringup Lab"[3], "Theranos,
      | Silicon Valley, and the March Madness of Tech Fraud"[4],
      | "Another LPC55 Vulnerability"[5], "The Sidecar Switch"[6],
      | "The Pragmatism of Hubris"[7], "Debugging
      | Methodologies"[8], or "The Books in the Box"[9].
      | 
      | There's tons more where that came from; if you are a fan of
      | _On the Metal_ , I don't think you'll be disappointed --
      | and it has the added advantage that you join a future
      | conversation!
      | 
      | [0] On The Metal: https://podbay.fm/p/on-the-metal
      | 
      | [1] Ken Shirriff: https://podbay.fm/p/on-the-
      | metal/e/1611669600
      | 
      | [2] Friends of Oxide: https://podbay.fm/p/oxide-and-friends
      | 
      | [3] Tales from the Bringup Lab: https://podbay.fm/p/oxide-
      | and-friends/e/1638838800
      | 
      | [4] Theranos, Silicon Valley, and the March Madness of Tech
      | Fraud: https://podbay.fm/p/oxide-and-friends/e/1632182400
      | 
      | [5] Another LPC55 Vulnerability: https://podbay.fm/p/oxide-
      | and-friends/e/1649116800
      | 
      | [6] The Pragmatism of Hubris: https://podbay.fm/p/oxide-
      | and-friends/e/1639443600
      | 
      | [7] The Sidecar Switch: https://podbay.fm/p/oxide-and-
      | friends/e/1638234000
      | 
      | [8] Debugging Methodologies: https://podbay.fm/p/oxide-and-
      | friends/e/1652745600
      | 
      | [9] The Books in the Box: https://podbay.fm/p/oxide-and-
      | friends/e/1632787200
 
        | newsclues wrote:
        | Oxide and Friends is good despite Twitter Spaces, but On
        | the Metal was great. I hope to see it return.
 
        | sams99 wrote:
        | Could you include a link to the new podcastified Twitter
        | space in the footer of https://oxide.computer/ ?
        | 
        | And maybe do an addendum 5 minute announcement episode on
        | the "on the metal" podcast?
        | 
        | Love your excellent work, thank you all
 
        | bcantrill wrote:
        | Yes, that's in the works: we have a website redesign
        | coming up, and this was included in it. So stay tuned!
        | 
        | And the 5 minute announcement on the _On the Metal_ feed
        | is a great idea; thanks for the idea -- and for the kind
        | words.
        | 
        | Finally, I hasten to add: we have a really exciting Space
        | coming up on Monday[0], where we'll be joined by Jon
        | Masters to talk about the importance of integrating
        | hardware and software teams; join us!
        | 
        | [0]
        | https://twitter.com/bcantrill/status/1545441245853495296
 
    | mtlmtlmtlmtl wrote:
    | That's such a shame. I've tried listening to those recordings
    | and it's hard to follow without the context of the "Twitter
    | Space" which I'm not even sure what is because I don't use
    | Twitter. The audio quality is also quite grating to listen
    | to.
    | 
    | Would love recommendations for in depth, tech related
    | podcasts that stand on their on, like On The Metal.
 
      | mwcampbell wrote:
      | IMO the recordings stand just fine on their own. Even when
      | I'm in the Twitter space live, I'm certainly not looking at
      | tweets or anything else that's happening on the screen,
      | unless I want to talk.
 
  | zozbot234 wrote:
  | They've been posting their Twitter Spaces chats.
  | https://github.com/oxidecomputer/twitter-spaces Not as focused
  | as the podcast, but still very high quality content.
 
  | Toxide wrote:
  | Other have posted the twitter link, but here is the youtube
  | link: https://www.youtube.com/channel/UCFn4S3OexFT9YhxJ8GWdUYQ
 
___________________________________________________________________
(page generated 2022-07-09 23:00 UTC)