Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

in the case of imagen, I suppose the cost is at least two orders of magnitude over 500k.


... no it definitely wasn't. that's $50m. read the paper, they tell you how long it took on a v4-256, which you know the public rental price for.


Pretty much everybody who trains models trains tens to hundreds of models before their final run. But a better way to think about this is: Google spent money to develop the TPUv4, then paid money to have a bunch built, and they're sitting in data centers, either consuming electricity or not. Google has clearly decided the value of these results exceeds the amortized cost of running the TPU fleet (which is an enormous expense).


While I don't agree with gp's price estimate of $50M, you are also forgetting that to train the final model, they had to iterate on the model for research and development. This imposes a significant multiple over the cost to just train the final model.


they're talking about the total cost of the project, including the salaries of the humans, their health care, their office space, etc.


And where did you think the tagged dataset and software came from?


Marginal cost of using that is basically $0. The internal dataset is a sunk cost that's already been paid for (presumably for their other, revenue generating products like Google Images). Half of their dataset is a publicly available one.


One order of magnitude over 500k it's 5M, two orders is 50M




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: