Hacker Newsnew | past | comments | ask | show | jobs | submit | KenoFischer's commentslogin

Congratulations to the Oxide team! It's a tough market out there :)! I'm still personally frustrated that I don't get to play with the hardware (too expensive for our internal server needs; not the right fit for our datacenter partners/customers), but I'm excited to see that they're successful and hopefully they'll come around to my use case eventually :). In the meantime, I appreciate that they're building largely in the open - every once in a while I'll glance at their issue tracker for light bedtime reading. Just recently we had some fun internally throwing our controls software at their thermal loop as a usage example - it's often hard to find compelling real-world systems to use as openly sharable examples (of course we have interesting customer problems, but that's all NDA'd), so having companies build real stuff in the open is fantastic. Great company, wish them the best.


> too expensive for our internal server needs; not the right fit for our datacenter partners/customers

You and me both. They're doing neat stuff, but I wonder how many other potential customers feel that way too.

What is Oxide's market? It feels a bit like advanced alien technology that is ultimately a little too weird and expensive for most enterprises to adopt.


I always thought a company like Railway would be an Oxide customer. But Railway is building their own servers in their own datacenters. So I am really curious who is small enough to buy Oxide, but large enough to need Oxide?


The same sorts of customers that SGI used to sell to in the pre-cloud era. DoD. Oil and gas. Finance.

People with deep pockets and good reasons to want to keep certain parts of their infra very close to home. Also the kind of people that expect very highly skilled people to show up and get their in-house app running.

(I was an SGI HPC customer once. I still miss the old SGI. Sigh.)


Maybe DigitalOcean?


how does it compare to Nutanix?


Of topic, but where the hell did Nutanix come from? I've never heard of them until recently and all of a sudden, they are being marketed as a serious competitor to VMware etc.


They have been around for a long time and were one of the first to have a hyper-converged solution where all storage in the nodes is pooled and usable by any node. They also have their own hypervisor. You can get 4 nodes per 2U so pretty dense. In the datacenter my company uses a company had dozens of Nutanix boxes sitting in the hallway for months before they finally installed them. They are pretty notoriously expensive so only really used by companies with big IT budgets.


OK, must be one of those weird things where I just never noticed them. Super strange because I've been very aware of this space for decades.


They have been around for nearly 20 years. I viewed them as an also-ran until Broadcom decided they didn’t need any of us as VMware customers anymore. Now Nutanix seems like a viable path for on-prem VM workloads that need a new home for those who don’t want to part with an arm and a leg on licensing but can’t move to public cloud either. I’m not sure how much of that market Oxide can capture. Not sure Nutanix is still doing the hyperconverged hardware themselves anymore.


Nutanix / Oxide have a VERY different market / customer base.


I've been curious about Oxide for a year or two without fully understanding their product. People talking about the "hyperconverged" market in this thread gave me an understanding for the first time.

Given this, can you help me understand in what ways they are different?

When I went to the Nutanix website yesterday, the link showed that'd I'd previously visited them (not a surprise, I look up lots of things I see mentioned in discussions) but their website does an extremely poor job of explaining their business to someone who lacks foundational understanding, even once I'd started reading about "hyperconverged" just before.


If you want to KNOW the chain of custody for all of your OS and software, from the bootloader to the switch chip, and you want to run this virtualization platform airgapped, buying at rack-scale, you want Oxide. They are making basically everything in-house. That's government, energy, finance, etc. Customers that need descretion, security, performance, and something that works very reliably in a high-trust environment, with a pretty high level of performance.

Also check this out: https://www.linkedin.com/posts/bryan-cantrill-b6a1_unbeknown...

If you need a basic "vm platform", VMware, Proxmox, Nutanix, etc. all fit the bill with varying levels of feature and cost. Nutanix has also been making some fairly solid kubernetes plays, which is nice on hyperconverged infrastructure.

Then if you need a container platform, you go the opposite direction - Kubernetes/OpenShift and run your VMs from your container platform instead of running your containers from your VM platform.

As far as "hyperconverged"...

"Traditionally" with something like VMware, you ran a 3-tier infrastructure: compute, a storage array, and network switching. If you needed to expand compute, you just threw in another 1U-4U on the shelf. Then you wire it up to the switch, provision the network to it, provision the storage, add it to the cluster, etc. This model has some limitations but it scales fairly well with mid-level performance. Those storage arrays can be expensive though!

As far as "hyperconverged", you get bigger boxes with better integration. One-click firmware upgrades for the hardware fleet, if desired. Add a node, it gets discovered, automatically provisions to the rest of the configuration options you've set. The network switching fabric is built into the box, as is the storage. This model brings everything local (with a certain amount of local redundancy in the hardware itself), which makes many workloads blazing fast. You may still on occasion need to connect to massive storage arrays somewhere if you have very large datasets, but it really depends on the application workloads your organization runs. Hyperconverged doesn't scale compute as cheaply, but in return you get much faster performance.


https://news.ycombinator.com/item?id=30688865

Here is an answer by steveklabnik about this topic.


theyve been marketed as a serious competitor to vmware for 15 years. their sales reps mightve just not found you until recently. but we did a poc with them 10 years ago and i dont believe much has changed since


Is that what DO is using at the moment? Never heard about it

Yeah, me too. I know that they have already explained why they can't, but I'd love for them to build a mini box that we could try it out on.


> Do they have "leggo my eggo" itself trademarked?

As a matter of fact, they do:

https://tsdr.uspto.gov/#caseNumber=77021301&caseType=SERIAL_...

The full complaint linked above has a full list of trademarks. There's also a claim for trade dress infringement, since the food truck uses the same font and red-yellow-white color scheme.


However, that particular phrase appears to be trademarked for: waffles, pancakes, french toast


Skimming the complaint, Kellogg looks to be arguing it is a well-known mark,[1] and is also making a trade dress claim.

[1] https://www.uspto.gov/ip-policy/trademark-policy/well-known-...


You're supposed to generate a random one, but the only consequence of not doing so is that you won't be able to register your package if someone else already took the UUID (which is a pain if you have registered versions in a private registry). That said, "vanity" UUIDs are a bad look, so we'd probably reject them if someone tried that today, but there isn't any actual issue with them.


Funny to see this come back and see my write-up linked. I did this 8 years ago and think I was the first on this particular board (although others had done similar on other boards). I still have a pile of them sitting on my desk because I accidentally kept bricking them by being ... not careful. That said, even at the time this board was already old, so I guess it's positively prehistoric at this point. I eventually stopped working on this because I thought that others were making sufficient progress. It hasn't really fully materialized yet, but between openbmc, opensil, DC-SCM and the work the oxide folks are doing, I'm still hopeful that we'll get out of server firmware hell eventually.


Out of curiosity: how "bricked" are these boards? Is there irreversible hardware damage (and, if so, how?), or has some firmware just gotten overwritten?


One of them I managed to fry the pcie root complex somehow, not sure exactly how. One I damaged the traces to the BMC SPI flash. Two others I think just have bricked firmware, but it's been years, so I don't remember for sure.


Tried it out.

1. First authentication didn't work on my headless system, because it wants an oauth redirect to localhost - sigh.

2. Next, WebFetch isn't able to navigate github, so I had to manually dig out some references for it.

3. About 2 mins in, I just got ``` ℹ Rate limiting detected. Automatically switching from gemini-2.5-pro to gemini-2.5-flash for faster responses for the remainder of this session. ``` in a loop with no more progress.

I understand the tool is new, so not drawing too many conclusions from this yet, but it does seem like it was rushed out a bit.


Similar. Yesterday things seemed to be going okay. It was trucking along, making code changes.

Then I hit the rate limit. - Fine, no worries, it'll be interesting to see if the quality changes.

Then it starts getting stuck and taking forever to complete anything. So I shut it down for the day.

Today, I start it back up and ask it to pickup where it left off and it starts spinning. I forget about it and come back 7.5 hours later and it' still spinning. When I kill it it said: 1 Turn, 90k input tokens, 6.5 hours of API time... WTH?

And now I'm just totally rate limited - `status: 429, statusText: 'Too Many Requests'` - every time. Also, I can't find any kind of usage data anywhere!


I got the same experience. I ran `gemini --debug` and it spit out the authentication URL at least. However I got the same `Rate limiting detected` for a simple refactoring task in a repo with 2 python files after a few minutes.


not working for me either. getting "too many requests". my own CLI tool i wrote for gemini last week works better lol. maybe i have something configured wrong? though that really shouldn't be the case. my api key env var is correct.


What's the motivation for restricting to Pro+ if billing is via premium requests? I have a (free, via open source work) Pro subscription, which I occasionally use. I would have been interested in trying out the coding agent, but how do I know if it's worth $40 for me without trying it ;).


Great question!

We started with Pro+ and Enterprise first because of the higher number of premium requests included with the monthly subscription.

Whilst we've seen great results within GitHub, we know that Copilot won't get it right every time, and a higher allowance of free usage means that a user can play around and experiment, rather than running out of credits quickly and getting discouraged.

We do expect to open this up to Pro and Business subscribers - and we're also looking at how we can extend access to open source maintainers like yourself.


It's linked from the main website if you hit the "Log In" button and there was communication to customers about this, though I had the same initial reaction, which is why I looked around for corroboration before posting this.


Ah thanks, that's what I was looking for.


My understanding is that the TinyTapeout people were using efabless as a service provider and efabless was also providing some sponsorship, but that they are institutionally distinct. There's a LinkedIn post from the TinyTapeout folks that they're looking into alternatives.


That's a relief! And Tiny Tapeout has already done a beta with IHP's open-source 130nm BiCMOS SiGe PDK.

The IHP PDK is really a lot more exciting to me than the Skywater stuff because it's aimed at submillimeter analog things (450GHz fₜ, 650GHz fastest oscillator) and why would you fab a digital design in 130nm instead of just programming an FPGA?


> why would you fab a digital design in 130nm instead of just programming an FPGA?

That’s an interesting concept. So an fpga implemented on a current 7nm process is more performant (clock speed and energy use) than an asic on a 130nm process? How about 40nm process? I feel like there’s a graph of intersecting lines here.


I think perf is usually relatively close between an optimized design in a 7 nm FPGA and an optimized design in ~40 nm CMOS, but it's not 1:1. The FPGAs are usually higher-performance than 130 nm, but there are certain things that are easier in ASICs (eg analog-related stuff).


Speaking as a newbie - FPGAs can't get anywhere near the same clock speed, though, right? So the equivalence only applies if the work is parallelizable?


With the exception of the highest clock speed chips (eg Intel CPUs), clock speeds can actually be comparable. 45 nm CPUs got to 2.5 GHz, and if you tickle a 7 nm FPGA just right it can get to ~800 MHz to a GHz. Things like microcontrollers and chips that are generally less optimized than the old Intel CPUs (which were mostly drawn at the transistor level and use a speed-optimized process) are much closer in speed. A 3-stage RISC-V at 45 nm is probably also running at 400 MHz or less, and the FPGA is capable of a 3 stage RISC-V at that speed.

But yes, in general, FPGAs on certain computational tasks will need deeper pipelines or the use of parallelism. Usually, pipeline depth works. Actually, if you look at the Intel front side bus (less optimized than the core), that's about the speed you can get from a 7 nm FPGA.


The Sky130 IO pads can't go faster than 33Mhz (at least the ones in the open source PDK), and the OpenLane flow isn't yet timing driven, so anything internal isn't going to break more than 100Mhz. These aren't fast chips or fast processes, Skywater is mostly for pedagogical and niche military and research tapeouts.


$4600 on ebay for a 7/10nm xilinx versal. So is 130nm/40nm ASIC cheaper than $4600?


A few sq mm at 40 nm is about $20k, and you can only configure it once. I think the Versal also gives you more useful gates at that size (thanks to block RAMs and hard multipliers).


What about power consumption?


The FPGA will have higher static power (running all the overheads) but probably lower dynamic power for the same design. 40 nm is old at this point for high-performance chips.


The static power might also depend on whether the FPGA is an SRAM type or a floating-gate type, I'd think. Does Lattice have any parts fabbed in relatively new processes?


You're right. Pretty much everything is SRAM now, though. Even the MAX 10 is an SRAM FPGA with a flash-backed storage memory.


That price is probably $4600 for one Xilinx Versal.

For the MPW run you would get ~100 parts. When everything is said and done, and you pay for packaging etc., on a MPW run you'll likely pay something like $50K. So ~$500ea

The eFabless price of $10K for a full run, including a packaged part, was an unparalleled deal.



This is a single chip. At scale, the ASIC is absolutely cheaper.


Radiation tolerance is one case. For the price of a tiny tapeout run you could count on one hand how many qualified radiation tolerant ICs you could buy. There's some sauce involved with process choices for radiation tolerance, but one of critical things to do is use large features.


IHP is excitinybut their PDK is horrible compared to major fabs like TSMC or GF. Anyone using it for products hate it.


Hmm, that's interesting. What are the major pain points? As you can probably guess, I don't have access to TSMC's or GF's PDKs.


Inaccuracies with respect to the silicon (active and passove parsitics, mechanical stress effects, temperature dependencies etc) mainly, user friendliness and proper documentation as secondary .


Thank you!


> and why would you fab a digital design in 130nm instead of just programming an FPGA?

Because you need some analog features with your digital design.


If you need some analog features, that's conventionally called a "mixed-signal design", not a "digital design". I wasn't talking about mixed-signal designs, for which it's obvious that an FPGA is unlikely to work.


You should really look into summaries on how deep sub-micron adds more problems as processes shrink. It's crazy that 28nm and under even work at all. They also break faster in more ways than larger, mature nodes.

Far as 130nm, I'll give you a few reasons I'd use one over a 7nm FPGA. This is a non-HW guy saying what he's heard from pro's at different times. HW people, feel free to correct me about whatever I get wrong.

1. Unit prices. If you can take the upfront cost (NRE), the per unit price will be much lower than FPGA's. You might charge plenty per unit depending on the market. This can be a source of profit.

2. Older, larger nodes are said to be better for analog. Lots of designs are mixed-signal to use analog for it's lower power, extra performance, or how it doesn't blink (no rise/fall with clock).

3. ASIC's can't be reprogrammed like FPGA's. The custom design might be more secure like Sandia Secure Processor (Score) or CHERI RISC-V. FPGA's can only do one of these except for antifuse FPGA's.

4. Larger nodes are easier to visually inspect for backdoor with cheaper, teardown hardware. Who knows what's in the FPGA's.

5. Larger nodes are easier to synthesize, P&R, and auto-inspect (eg Calibre). That means open-source tools have a better chance of working or even being developed.

6. If not too power hungry (or power is cheap), some applications can let you outperform 7nm parts with parallel use of 130nm parts which are much cheaper or highly-optimized. An example what media wanting to do distributed, massively-parallel design for doing NN training maybe with 8-bitters and on-board, analog accelerators. My inspiration, aside from old MPP clusters (eg Thinking Machines), was a wafer-scale, analog NN done before Cerebras.

7. Improved reliability in general. In trusted checkers or fault-tolerant configuration, I feel like the 130nm parts are less likely to have a double failure or fail before the 7nm nodes.

8. If there's a business case, saying you built your own hardware is cool. It might even attract talent who benefit the company in other ways.

That's off the top of my head. Again, I just read a lot of stuff on ASIC's.

On a side note, you might find eASIC's Nextreme's interesting. They're Structured ASIC's that work like FPGA's in that design gets put on something with pre-made blocks to save money. Except, instead of software programmed, some via or metal layers get customized for the routing. While that reduces NRE cost, doing the routing in hardware supposedly reduces unit prices and energy maybe with a performance boost. They used to sample chips out quickly and relatively cheaply. Also, I think Triad Semiconductor had S-ASIC's with analog stuff.


eASIC Nextreme sounds like a good ol' fashioned ULA (uncommitted logic array), the sort of thing that's at least as old as the Sinclair ZX81 (where it drove the per-unit cost through the floor).


I hadn't heard of that. Looking it up, it's a type of gate array which I believe inspired both S-ASIC's and devices like FPGA's. Here's an intro to each for those following along:

https://en.m.wikipedia.org/wiki/Gate_array

http://eda.ee.ucla.edu/EE201A-04Spring/ASICslides.ppt

I also found a link with the pricing of one. It was $45,000 for 45 prototypes on 45nm through eASIC.

https://www.design-reuse.com/news/25107/easic-45nm-asic-valu...

That put having chips made into the realm of possibilities for even a small business. Other costs might prevent that but I could see more stuff opening up. I also envisioned hard blocks done on those nodes for common components so the S-ASIC was used for custom logic (eg differentiators).


Thanks! Yeah, for analog the case is obvious, both because there's no such thing as an analog FPGA and because smaller feature size comes with big drawbacks for analog; that's why I said, "why would you fab a digital design in 130nm". The others I'm less sure about, but they do sound plausible.



Yeah, there have been a dozen different attempts to make "an FPGA, but analog". This is one of them. They all failed. You'll note the page hasn't been updated since 02006: https://web.archive.org/web/20060715013941/http://www.anadig.... Analog circuits aren't fungible the way digital circuits are.

I'm not saying it's not a worthwhile research direction, just that it hasn't borne commercially viable fruit so far, despite decades of attempts.


wait, does 130nm imply i can send them verilog and receive ASICs in the mail?


Basically yes, but you have to generate the GDS-II from your Verilog yourself, you have to pay them several thousand dollars, the turnaround time is nearly a year, and your first and maybe second and third tapeout will probably have bugs that keep it from working at all.


This comports with my - admittedly decade-and-a-half old - understanding of the code to silicon pipeline/flow.


There are open tool chains that will compile your design using the cells defined in the outreach programs fab specific standards. However, it will not necessarily function like your simulated hardware design.

Getting the hardware cell simulation working is not trivial, and Synopsys charges more per seat than most startups spend on labor in a year.

YMMV =3


I really love Oxide to an unhealthy amount (it's become a bit of a meme among my colleagues), but sometimes I do wonder whether they went about their go-to-market the right way. They really tried to do everything at once - custom servers, custom router, custom rack, everything. Their accomplishments are technologically impressive, but, as somebody who is in a position to make purchasing decisions, not economically attractive. They're 3x more expensive than our existing hardware, two generations behind (I'm aware they're on track for a refresh) and don't have any GPUs. E.g. what I would have loved to see is just an after-market BMC/NIC/firmware solution using their stack. Plug it into a cheap Gigabyte system (their BMC is pluggable and NIC is OCP) and just have the control plane manage it as a whole box. I'd have easily paid serveral thousand $ per server just for that. All the rack scale integration, virtualization, migration, network storage, etc stuff is cool, but not everyone needs it. Get your foot in the door at customers, build up some volume for better deals with AMD, and then start building the custom rack stuff ... Of course it's easy to be a critic from the side lines. As I said, I do really love what the Oxide folks are doing, I just really hope it'll become possible for me to buy their gear at some point.


First, thanks for the love -- it's deeply appreciated! Our go-to-market is not an accident: we spent a ton of time (too much time?) looking at how every company had endeavored (and failed) in this space, and then considering a bunch of other options besides. Plugging into a "cheap Gigabyte" system wouldn't actually allow us to build what we've built, and we know this viscerally: before we had our system built, we had to have hardware to build our software on -- which was... a bunch of cheap Gigabyte systems. We had the special pain of relearning all of the reasons why we took the approach we've taken: these systems are a non-starter with respect to foundation.

You may very well not need the system that we have built, but lots of people do -- and the price point versus the alternatives (public cloud or on-prem commodity HW + pretty price SW) has proven to be pretty compelling. I don't know if we'll ever have a product that hits your price point (which sounds like... the cost of Gigabyte plus a few thousand bucks?), but at least the software is all open source!


Please forgive my tergiversation. I fully trust that you know your path and I know how annoying it is to be why-dont-they-just'd. As I said, I'm rooting for you.


> The meaning of TERGIVERSATION is evasion of straightforward action or clear-cut statement : equivocation


There's two dictionary definitions of tergiversate. One is the one you quoted, the other is one of desertion. Both meanings of the word are pejorative in the sense that the word comes with a connotation of betrayal of a cause. What I wanted to express was an acknowledgement that I understood the feeling that you get when someone who's clearly a fan of your work nevertheless does not provide a clear endorsement. It's easy emotionally to dismiss people who "just don't get it". But when someone does get it but chooses to equivocate, that can feel like an emotional betrayal. So I was looking for a word that covered that with the right connotation. I originally used apostasy, but it didn't feel quite right, because I wasn't really renouncing, more failing to fully endorse, so tergiversation it was. Of course having to write an entire paragraph to explain your word choice kind of defeats the purpose of choosing a single well fitting word over just writing a sentence of simple words that explains what you mean. But hey, I write enough technical writing, documentation, reports, grants, etc. all day where clarity is paramount that I feel like I get to have a little vocabulary treat in my personal writing ;).


+1 for use of tergiversation


So my question: any Arm-based system or GPU-based system on the horizon?


You just described why commodity servers won over engineered systems that came before Oxide (like Nutanix, Sun / Oracle Exa*, VCE etc).

So I totally agree with your go-to-market comment, because it’s also a bet against cloud.

I wish them luck though.


And yet, non of the hyperscalers use commodity server. They are buying parts from the OCP but those are hardly 'commodity' servers. So did they win?


I kinda feel that their focus is more on building a great technology (& culture?) than a great business.

Not necessarily a bad choice; after all, for what shall it profit a man, if he shall gain the whole world, and lose his own soul?


We are definitely very much building a business! We have the iconoclastic belief that you can build a business by having a terrific team building a great product that customers love. And we're getting there![0]

[0] https://www.theregister.com/2024/11/18/llnl_oxide_compute/


Oxide are doing great work. Hoping they can probe the market a bit more for us out on the sidelines preparing to drop in and compete with some similar tech.


Id also wish I could get to play around with a cheaper version of their tech, but they probably havw enough customers that really want a large-scale solution that is completely customizable


I'm curious what their burn rate is.


Yes. Linked press release says they borrowed equipment from NARA to play it.


Relevant Animats comment at the time.

https://news.ycombinator.com/item?id=40958494

Related:

The NSA Is Defeated by a 1950s Tape Recorder. Can You Help Them? -https://news.ycombinator.com/item?id=40957026 - July 2024 (24 comments)

Admiral Grace Hopper's landmark lecture is found, but the NSA won't release it -https://news.ycombinator.com/item?id=40926428 - July 2024 (9 comments)


https://www.nsa.gov/Press-Room/Press-Releases-Statements/Pre...

> able to retrieve the footage contained on two 1’ APEX tapes

I'm no expert, but I think they meant 1-inch AMPEX tape.

Also, perhaps they should record it in doubly. #ThisIsSpinalTap #Stonehenge #InchsToFeet


The page now says AMPEX. Did they read your comment?

p.s. I'm imagining ECHELON flagging your comment and someone from the NSA quickly modifying the document on the sly...


Or someone in the NSA reads Hacker News.

Hi, anonymous NSA employee!


haha now we're all on a list, aren't we? shit.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: