Hacker Newsnew | past | comments | ask | show | jobs | submit | explosion-s's commentslogin

Just curious, is there any smaller version of this model capable of running on edge devices? Even my Mac M1 with 8gb ram couldn't run the C version.


This semi-quantized version targets the Jetson Orin Nano, but only comes with a simple inference engine.

https://huggingface.co/Teaspoon-AI/Voxtral-Mini-4B-INT4-Jets...


https://kyutai.org/stt has an implementation for MLX and mentions iPhones, so it should work on edge devices, Macs and iPhones.


I'm curious to see if you are able to run the model now from the CLI?


This is an interesting problem, one I've thought a lot about myself. On one hand, LLMs have the capacity to greatly help people, and I think, especially in the realm of gradually learning how to program, on the other hand, the non-determinism is such a difficult problem to work around.

One current idea of mine, is to iteratively make things more and more specific, this is the approach I take with psuedocode-expander ([0]) and has proven generally useful. I think there's a lot of value in the LLM instead of one shot generating something linearly, building from the top down with human feedback, for instance. I give a lot more examples on the repo for this project, and encourage any feedback or thoughts on LLM driven code generation in a more sustainable then vibe-coding way.

[0]: https://github.com/explosion-Scratch/psuedocode-expander/


> on the other hand, the non-determinism is such a difficult problem to work around.

Well, you can always set temperature to 0, but that doesn't remove hallucinations.


I just made it delete 45,000 lines of code I told it to ruthlessly remove everything lmao


vie*t*nam?


And what is "this book"? - To me this rings heavily of being written by a LLM as well. Sad.


while I did have an LLM help me edit my notes/thoughts, "this book" is in reference to the book I am attempting to write. This passage is an excerpt!


Having no knowledge about this isn't there a possibility that this is innocuous, that is that there is some difference in how each provider, such as T-Mobile or Verizon, etc., send the data about how many bars are available such that it needs to be bumped by one?


I looked at the usernames of the people that it mentioned and I don't think that it was actually mentioned on Reddit. It it seems like the front end is mostly AI generated.


I made something similar to this with WebGL shaders (the benefit being it works across browsers): https://real-glass.vercel.app - The tricky thing for me was making it refract real HTML elements behind


I considered WebGL, and I agree—a shader is more performant for real-time effects.

But WebGL comes with drawbacks:

- You need JS code running before anything shows up.

- Shaders can’t directly manipulate the DOM render. To make refraction work, you’d have to re-render everything into a canvas—which isn’t really “the web” anymore.

With the SVG/CSS approach, you can pre-render the displacement map (at build time or on the backend) and get the refraction visible on the very first frame. Plus, it integrates cleanly with existing, traditional UIs.

That said, this approach could definitely be improved. Ideally we’d have shader-like features in the SVG Filter spec (there was a proposal, but it seems abandoned). There are some matrix operations available in SVG Filters, but they’re limited—and for my first blog post I wanted to focus more on pedagogy, art, and technique than heavy optimization.


I'm on mobile and your site works much better than the WebGL one.


What is causing the ghosting/delay when moving the glass over text?


probably shaders are compiling and initialising on every drag movement


Cool this looks like it even has dispersion, i.e. colors separate at the edge of the glass element.



Yes chromatic abberation is caused by dispersion.


Damn, outnerded.


Looks nice! It's too slow to actually use though. Op's is much smoother.


It's the opposite on my macbook pro/chrome computer... the OP is unusable, but the webGL version is super smooth


it's extremely quick on my M3 Mac too.


Not over here. As far as I understand Op's solution does not utilize a gpu.


I actually see gpu utilisation in OP's website when I move things (m3 pro), but this other solution shows much less gpu utilisation (prob more efficient?).


Impressive!


Okay it turned out to be octocat - and that company offers a subscription service to open models as well. That explains my confusion because the image looks very amphibian like but it wasn't actually a newt or salamander or anything. I eventually found this because I recalled that it was a subscription, not just an open router s service


Octocat? You have a link? I get Github Github and more Github if I search.


I think they mean Octofriend (https://github.com/synthetic-lab/octofriend) from Syntethci Lab (https://synthetic.new/landing/home).


Yeah sorry that's what I meant. My apologies


I googled so much and found nothing I literally spent a full half an hour googling


Maybe do an "image" search with Google or DDG instead of searching for text, e.g.,

https://www.google.com/search?q=AI+CLI+TOOL+LOGO+NEWT+SALAMA...

or

https://duckduckgo.com/?t=ffab&q=AI%2BCLI%2BTOOL%2BLOGO%2BNE...

or, as busymom0 suggested below, use the Hacker News search at the bottom of each HN page.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: