←back to thread

1071 points kentonv | 1 comments | | HN request time: 0.214s | source

I wasn't quite sure if this qualified as "Show HN" given you can't really download it and try it out. However, dang said[0]:

> If it's hardware or something that's not so easy to try out over the internet, find a different way to show how it actually works—a video, for example, or a detailed post with photos.

Hopefully I did that?

Additionally, I've put code and a detailed guide for the netboot computer management setup on GitHub:

https://github.com/kentonv/lanparty

Anyway, if this shouldn't have been Show HN, I apologize!

[0] https://news.ycombinator.com/item?id=22336638

Show context
RulerOf ◴[] No.42159254[source]
> I've never heard of anyone else having done anything like this. This surprises me! But, surely, if someone else did it, someone would have told me about it? If you know of another, please let me know!

I never had the tenacity to consider my build "finished," and definitely didn't have your budget, but I built a 5-player room[1] for DotA 2 back in 2013.

I got really lucky with hardware selection and ended up fighting with various bugs over the years... diagnosing a broken video card was an exercise in frustration because the virtualization layer made BSODs impossible to see.

I went with local disk-per-VM because latency matters more than throughput, and I'd been doing iSCSI boot for such a long time that I was intimately familiar with the downsides.

I love your setup (thanks for taking the time to share this BTW) and would love to know if you ever get the local CoW working.

My only tech-related comment is that I will also confirm that those 10G cards are indeed trash, and would humbly suggest an Intel-based eBay special. You could still load iPXE (I assume you're using it) from the onboard NIC, continue using it for WoL, but shift the netboot over to the add-in card via a script, and probably get better stability and performance.

[1]: https://imgur.com/a/4x4-four-desktops-one-system-kWyH4

replies(1): >>42159408 #
kentonv ◴[] No.42159408[source]
Hah, you really did the VM thing? A lot of people have suggested that to me but I didn't think it'd actually work. Pretty cool!

Yeah I'm pretty sure my onboard 10G Marvell AQtion ethernet is the source of most of my stability woes. About half the time any of these machines boot up, Windows bluescreens within the first couple minutes, and I think it has something to do with the iSCSI service crashing. Never had trouble in the old house where the machines had 1G network -- but load times were painful.

Luckily if the machines don't crash in the first couple minutes, then they settle down and work fine...

Yeah I could get higher-quality 10G cards and put them in all the machines but they seem expensive...

replies(9): >>42159687 #>>42160035 #>>42160606 #>>42160613 #>>42161388 #>>42162551 #>>42162914 #>>42163053 #>>42163249 #
1. amluto ◴[] No.42163249[source]
Just buy used 10G hardware from an HFT firm :). Seriously, though, 10G gear is cheap these days.

I bet one could put an unreasonable amount of effort into convincing an Nvidia Bluefield card to pretend to be a disk well enough to get Windows to mount it. I imagine that AWS is doing something along those lines too, but with more cheap chips and less Nvidia markup…

There has got to be a way to convince Windows to do an overlay block device that involves magic words like “thin provisioning”. But two seconds of searching didn’t find it. Every self-respecting OS (Linux, FreeBSD, etc) has had this capability for decades, of course. Amusingly, AFAICT, major clouds also mostly lack this capability — performance of the obvious solution in AWS (boot everything off an AMI) is notoriously poorly performing.