Hacker Newsnew | past | comments | ask | show | jobs | submit | numeri's commentslogin

but Taalas had to quantize Llama 3.1 8B to death to get it to fit. It can't produce coherent non-English text at all.

and if I was to guess, the latest generation of models (Claude Opus 4.6, GPT-5.3-codex, etc.) differ from Opus 4.5, GPT 5.2 primarily in the addition of deeper, more difficult (most likely agentic and coding-based, like Terminal Bench) tasks to their RLVR training.

I could be completely off, as my intuition here is fully based on public research papers, but it seems to explain the current state of things fairly well.


No, Python or units[1] is always a better choice if I'm near a computer (and I nearly always am these days, unfortunately, I suppose). I do have three wonderful slide rules, though.

[1]: https://www.gnu.org/software/units/


Introducing a solid zero-knowledge age verification option is the opposite direction of ending anonymity in the Internet, which other parts of the same governments are also working on.

So yeah, I'll gladly trust and cheer on the part working in the right direction.


I'll just throw in support for gaming on Linux – it's pretty nice feeling these days! I still have the occasional (once every 5–8 months?) update cause a short-lived bug, but it's a very justifiable trade-off to avoid Windows these days.


This is written by someone who's not an AI researcher, working with tiny models on toy datasets. It's at the level of a motivated undergraduate student in their first NLP course, but not much more.


If one can easily reach parity with a motivated undergrad by leveraging LLMs I will still consider it impressive.

While the 5-minutes model will never be useful in itself it lays the groundwork for amateurs and small groups to getting into developing small models. There's at the moment another HN headline hyping up a tiny model that scores impressively at the arc-agi benchmarks so it's clearly not a dead end to explore what is "household-affordable" models.

Though an approach that doesn't lean on the authors $200/month OAI sub would've been more interesting to follow.


You can also reach research parity by downloading a Github repository. Is that impressive too?


Downloading a file is not equivalent to having high level abstractified control over running software.

And if it is then I'm a farmer because I bought potatoes from the store.


One sign would be occasionally changing course in response to overwhelming employee feedback. If that never or almost never happens, the feedback is being ignored, not taken constructively and not followed.


This isn't right – calibration (informally, the degree to which certainty in the model's logits correlates with its chance of getting an answer correct) is well studied in LLMs of all sizes. LLMs are not (generally) well calibrated.


I'm a little shocked at Simon's conclusion here. We have a man who bought an social media website so he could control what's said, and founded an AI lab so he could get a bot that agrees with him, and who has publicly threatened said AI with being replaced if it doesn't change its political views/agree with him.

His company has also been caught adding specific instructions in this vein to its prompt.

And now it's searching for his tweets to guide its answers on political questions, and Simon somehow thinks it could be unintended, emergent behavior? Even if it were, calling this unintended would be completely ignoring higher order system dynamics (a behavior is still intended if models are rejected until one is found that implements the behavior) and the possibility of reinforcement learning to add this behavior.


Elon obviously wants Grok to reflect his viewpoints, and has said so multiple times.

I do not think he wants it to openly say "I am now searching for tweets from:elonmusk in order to answer this question". That's plain embarrassing for him.

That's what I meant by "I think there is a good chance this behavior is unintended".


I really like your posts, and they're generally very clearly written. Maybe this one's just the odd duck out, as it's hard for me to find what you actually meant (as clarified in your comment here) in this paragraph:

> This suggests that Grok may have a weird sense of identity—if asked for its own opinions it turns to search to find previous indications of opinions expressed by itself or by its ultimate owner. I think there is a good chance this behavior is unintended!

I'd say it's far more likely that:

1. Elon ordered his research scientists to "fix it" – make it agree with him

2. They did RL (probably just basic tool use training) to encourage checking for Elon's opinions

3. They did not update the UI (for whatever reason – most likely just because research scientists aren't responsible for front-end, so they forgot)

4. Elon is likely now upset that this is shown so obviously

The key difference is that I think it's incredibly unlikely that this is emergent behavior due to an "sense of identity", as opposed to direct efforts of the xAI research team. It's likely also a case of https://en.wiktionary.org/wiki/anticipatory_obedience.


That's why I said "I think there is a good chance" - I think what you describe here (anticipatory obedience) is possible too, but I honestly wouldn't be surprised to hear that the from:elonmusk searches genuinely were unintended behavior.

I find this as accidental behavior almost more interesting than a deliberate choice.


Willison's razor: Never dismiss behaviors as either malice or stupidity when there's a much more interesting option that can be explored.


I side with Occam's razor here, and with another commenter in this thread. People are construing entire conspiracy theories to explain fake replies when asked for system prompt, lying in Github repos, etc.


What if searching for Elon's tweets was indeed intended, but it wasn't supposed to show up in the UI?


Occam's razor would seem to apply here.


> That's plain embarrassing for him

You think that's the tipping point of him being embarrassed?


It seems as if the buzz around AI is so intoxicating that people forgo basic reasoning about the world around them. The recent Grok video where Elon is giddy about Grok’s burgeoning capabilities. Altman’s claims that AI will usher in a new utopia. This singularity giddiness is infectious yet denies the worsening world around us - exacerbated by AI - mass surveillance, authoritarianism, climate change.

Psychologically I wonder if these half-baked hopes provide a kind of escapist outlet. Maybe for some people it feels safer to hide your head in the sand where you can no longer see the dangers around you.


I think cognitive dissonance explains much of it. Assuming Altman isn’t a sociopath (not unheard of in CEOs) he must feel awful about himself on some level. He may be many things, but he is certainly not naive about the impact ai will have on labor and need for ubi. The mind flips from the uncomfortable feeling of “I’m getting rich by destroying society as we know it” to “I am going to save the world with my super important ai innovations!”

Cognitive dissonance drives a lot “save the world” energy. People have undeserved wealth they might feel bad about, given prevailing moral traditions, if they weren’t so busy fighting for justice or saving the planet or something that allows them to feel more like a super hero than just another sinful human.


On top of all of that, he demonstrates that Grok has an egregious and intentional bias but then claims it's inexplainable happenstance due to some sort of self-awareness? How do you think it became self-aware Simon?


That's the thing, some people do see things in their mind that clearly. It's about as rare as full aphantasia, but it's absolutely a spectrum.


There's really no way to know this, as it's all based on subjective experiences in which two people could easily describe the same sensation differently.


That's a bold claim! Actually, there are plenty of scientific experiments that show actual differences between people who report aphantasia and those who don't, including different stress responses to frightening non-visual descriptions, different susceptibility to something called image priming, lower "cortical excitability in the primary visual cortex", and more: https://en.wikipedia.org/wiki/Aphantasia

So we know that at least the people who claim to see nothing act differently. Could it just be that people who act differently describe the sensation differently, you might ask?

No, because there are actual cases of acquired aphantasia after neurological damage. These people used to belong to the group that claimed to be able to imagine visual images, got sick, then sought medical help when they could no longer visualize. For me, at least, that's pretty cut and dry evidence that it's not just differing descriptions of the same (or similar) sensations.


If you recall, I prefaced my original comment with "Half the time,"


I really don't think so. I can't visualize with perfect clarity, but I can do pretty well, especially if I try. It tends to shift, so "count the stripes on the tiger" doesn't quite work, but I can do the exercise of visualizing a ball on a table and then saying what color it is.

There is no possible way that anyone could honestly describe this experience as "I don't visualize," any more than someone with working ears could describe their experience as "I don't hear anything."


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: