Hacker Newsnew | past | comments | ask | show | jobs | submit | SamDc73's commentslogin

The Stargate money didn’t show up I guess, and now the whole gridlock is collapsing?

I find https://github.com/steipete/gogcli a bit easier (but still confusing to setup)

Google Workspace API(s) keys and Roles was always confusing to me at so many levels .. and they just seem to keeping topping that confusion, no one is addressing the core (honestly not sure if that is even possible at this point)


pre-ai if I had to include Google search queries in a commit, I’d be so embarrassed I’d probably never commit code like ever

Didn't Dario Amodei ask for more government intervention regarding AI?

Not a contradiction with this post

I highly suspect he might even consider Anthropic since they enforced restrictions at some point on OpenClaw form using there APIs


yes that's the blunder I'm talking about


I switched from using YouTube to invidious mainly because they don't support shorts and blocked YouTube on the DNS level, it's a bit slower, but I know I won't be sucked into doom-scrolling


I mean they are only running a small version of codex can they run the full one? Or the technology isn't there yet?


1000 tokens/sec for a highly specialised model is where we are going to see agents requiring.

Dedicated knowledge, fast output, rapid iteration.

I have been trying out SMOL models as coding models don't need to the full corpus of human history.

My most recent build was good but too small.

I am thinking of a model that is highly tuned to coding and agentic loops.


This is model 12188, which claims to rival SOTA models while not even being in the same league.

In terms of intelligence per compute, it’s probably the best model I can realistically run locally on my laptop for coding. It’s solid for scripting and small projects.

I tried it on a mid-size codebase (~50k LOC), and the context window filled up almost immediately, making it basically unusable unless you’re extremely explicit about which files to touch. I tested it with a 8k context window but will try again with 32k and see if it becomes more practical.

I think the main blocker for using local coding models more is the context window. A lot of work is going into making small models “smarter,” but for agentic coding that only gets you so far. No matter how smart the model is, an agent will blow through the context as soon as it reads a handful of files.


The small context window has been a recognized problem for a while now. Really only Google has the ability to use a good long context window


you should look into using subagents, which each have their own context window and don't pollute the main one


What are you talking about? Qwen3-Coder-Next supports 256k context. Did you wanted to say that you don't have enough memory to run it locally yourself?


Yes!

I tried to go as far as 32k on the context window but beyond that it won't be usable on my laptop (Ryzen AI 365, 32gb RAM and 6gb of VRAM)


You need minimum ie. 2x 24G GPUs for this model (you need 46GB minimum).


I was waiting for someone to say "this is what happens when you vibe code"



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: