Hacker Newsnew | past | comments | ask | show | jobs | submit | more _fjb4's commentslogin

Yeah as a long-time user I was really hoping for that myself, or at least have some sort of compromise. Certainly not that trainwreck of an AMA.


Wouldn't that be Hetzner?


Hetzner lacks a lot of features to count as a proper cloud. Still wishing for managed kubernetes.


Managed k8s probably wouldn't count as a bare bones cloud.


Upcloud, Vultr, Scaleway, Linode, DigitalOcean, etc have all included managed k8s. I think the standards have moved up and it's pretty much a standard feature of most bare bones cloud now.


It's not easy to move a group to a new platform, and there's very little people using Lemmy currently. The thing about Reddit despite its flaws is that it allows access to many large communities with a single account.


OpenLLAMA will be released soon and it's 100% compatible with the original LLAMA.

https://github.com/openlm-research/open_llama


The fact that this is commodity hardware makes ggml extremely impressive and puts the tech in the hands of everyone. I recently reported my experience running 7B llama.cpp on a 15 year old Core 2 Quad [1] - when that machine came out it was a completely different world and I certainly never imagined how AI would look like today. This was around when the first iPhone was released and everyone began talking about how smartphones would become the next big thing. We saw what happened 15 years later...

Today with the new k-quants users are reporting that 30B models are working with 2-bit quantization on 16GB CPUs and GPUs [2]. That's enabling access to millions of consumers and the optimizations will only improve from there.

[1] https://old.reddit.com/r/LocalLLaMA/comments/13q6hu8/7b_perf...

[2] https://github.com/ggerganov/llama.cpp/pull/1684, https://old.reddit.com/r/LocalLLaMA/comments/141bdll/moneros...


Likely you can use a Bluetooth keyboard and mouse with it.


It also doesn't have ECC as well which was a staple of the previous Mac Pro line.


I wonder if on-die RAM is less susceptible to memory errors?

I suspect that it is. Feels like less can go wrong. You have physically shorter interconnects, and the RAM is perhaps more of a known quantity relative to $SOME_RANDOM_MANUFACTURERS_DIMMS. But that is only a guess.

However, I don't know if that's true. I guess it's not necessarily more resistant to random cosmic rays or whatever.


I personally like the Adminforge instance https://teddit.adminforge.de. It's much quicker than the original teddit.net.


Theoretically the ARC should work with llama.cpp using OpenCL, but I haven't seen benchmarks or even a confirmation that it works.


I think the tinybox is meant to be a training/inference server meant for tinygrad and filled with those AMD cards. Very likely it will run Linux.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: