Hacker Newsnew | past | comments | ask | show | jobs | submit | more lagrange77's commentslogin

I'll steal that.


This is comparing two orthogonal properties.

SaaS is a business model while malleable vs. rigid is a property of the software itself.


You have forgotten the most important part: Lay off 90% of those devs.


But automating isn't a programming paradigm.

> They are young and inexperienced today, but won't stay that way for long.

I doubt that. For me this is the real dilemma with a generation of LLM-native developers. Does a worker in a fully automated watch factory become better at the craft of watchmaking with time?


I think the idea that LLMs are just good at "automating" is the old curmudgeon idea that young people won't have.

I think the fundamental shift is something like having ancillary awareness of code at all but high capability to architect and drill down into product details. In other words, fresh-faced LLM programmers will come out the gate looking like really good product managers.

Similar to how C++ programmers looked down on web developers for not knowing all about malloc and pointers. Why dirty your mind with details that are abstracted away? Someone needs to know the underlying code at some point, but that may be reserved for the wizards making "core libraries" or something.

But the real advancement will be not being restricted by what used to be impossible. Why not a UI that is generated on the fly on every page load? Or why even have a webform that people have to fill out, just have the website ask users for the info it needs?


Yeah i agree with most of what you say.

> looking like really good product managers.

Exactly and that's a different field with a different skillset than developer/programmer.

And that's the purpose of technology in the first place tbh, to make the hard/tedious work easier.


But do those watches tell time better? or harder? or louder? Once you have the quartz crystal and have digital watches, mechanical movements became obsolete. Rolex and Patek Philippe are still around, but it's more of a curiosity than anything.


Absolutely agree. The watches do tell time better. But the factory worker does not become better at the craft of watchmaking or EE.


Is it my browser, or does the video in the readme not have sound?


No sound! YouTube video in README does.

I was tempted to put Erik Satie in the README video. Didn’t want to risk copyright issues


Does any of you use one of these (Dokploy, CapRover, Dokku, Coolify) like Netlify, as advertised by some?

For me, the core feature of Netlify is building and deploying static websites quickly, with minimal configuration and triggered by git commits.

Does any of these really resemble that experience (except for the CDN Netlify uses, of course)?


I use Coolify for my own personal static site and it’s just like that. Git pushes redeploy my site and I get a discord notification once’s it’s done. The only manual thing I did was use a cloudlfare tunnel So it’s available to the public, since I am using my homelab to host Coolify.

I host maybe 8 different side projects on Coolify like this. Most don’t even have a Dockerfile in the repo. I use the standard nix packs option, and builds, rolling deployments etc are auto handled.


https://docs.dokploy.com/docs/core/comparison

Dokploy vs. CapRover, Dokku, Coolify


This seems like an unfair comparison for Dokku. I haven’t used the rest, but I have used Dokploy and Dokku. Dokku has had every single feature I could want or need, even accounting for weird edge cases. It just doesn’t have a UI.

With Dokploy, on the other hand, I found the UI difficult to navigate, which would be fine if the documentation was good but it was lacking.

But for many of the features their comparison claims Dokku doesn’t have, it actually does: database support, scheduled jobs, docker compose support. It has some form of monitoring. Overall Dokku has been a pretty robust solution for me and anything it might be missing, like in monitoring for instance, I can just add at the system level.

To be clear, I’m not anti-Dokploy and I think the more these tools improve the better. Just wanted to share my experience in defense of Dokku. Being able to spin up your apps on a cheap VPS is incredibly empowering over having to pay 10x more for managed services like Heroku or Render.


> I became the math consultant for A Beautiful Mind in part because I was such an Apollo 13 buff.

Cool, awesome job, as far as i can tell as a fan of the movie!

So you did what was best for yourself... and the group.


While you're here..

Do you guys know a website that clearly shows which OS LLM models run on / fit into a specific GPU(setup)?

The best heuristic i could find for the necessary VRAM is Number of Parameters × (Precision / 8) × 1.2 from here [0].

[0] https://medium.com/@lmpo/a-guide-to-estimating-vram-for-llms...


Yeah we have tried to build calculators before it just depends so much.

Your equation is roughly correct, but I tend to multiply by a factor of 2 not 1.2 to allow for highly concurrent traffic.


huggingface has this built in if you care to fill out your software and hardware profile here:

https://huggingface.co/settings/local-apps

Then on the model pages, it will show you whether you can use it.


Interesting, never knew about that! I filled out my details, then went to https://huggingface.co/openai/gpt-oss-120b but I'm not sure if I see any difference? Where is it supposed to show if I can run it or not?


You’ll see green check next to models you can use on the model card.

https://huggingface.co/unsloth/gpt-oss-20b-GGUF


Ah, it only works for GGUF, not for .safetensors (which the format HuggingFace themselves came up with :P ) ? I see the checks at https://huggingface.co/unsloth/gpt-oss-20b-GGUF but nothing at https://huggingface.co/openai/gpt-oss-120b, seems a bit backwards.


For those kind of models, you know if you can run them. :D

Also most of the times they are split up and, sometimes, you’ll get an indicator on the splits.

It’s still a work in progress to check all hardware and model format compatibility but it’s a great start until GGUF becomes the standard.


Maybe I'm spoiled by having great internet connection, but I usually download the weights and try to run them via various tools (llama.cpp, LM Studio, vLLM and SGLang typically) and see what works. There seems to be so many variables involved (runners, architectures, implementations, hardware and so on) that none of the calculators I've tried so far been accurate, both in the way that they've over-estimated and under-estimated what I could run.

So in the end, trying to actually run them seems to be the only fool-proof way of knowing for sure :)


Thanks for your answers!

While it is seemingly hard to calculate it, maybe one should just make a database website that tracks specific setups (model, exact variant / quantisation, runner, hardware) where users can report, which combination they got running (or not) along with metrics like tokens/s.

Visitors could then specify their runner and hardware and filter for a list of models that would run on that.


Yeah, what you're suggesting sounds like it could be more useful than the "generalized calculators" people are currently publishing and using.


That's why i despise advertising and marketing so much.

It forcefully tries to hack you and steal your scarcest resource.


Have to practice attention hygiene. Add blockers, switching to something else if add or woke crap pops up. Annoying audio adds in shops.. I start singing to myself to keep verbal part of the brain busy.


Oh i do.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: