Hacker Newsnew | past | comments | ask | show | jobs | submit | Flomolok's commentslogin

It's not a hype when it's delivers and I'm also not seeing a ceiling yet

Yet again interesting progress.

Also I like the idea of using the pose model to generate not a NPC but a avatar living in my phone or glas cube as a hologram. That would be quite scifi futuristic


I tried Claude.

If hardware continues it's evolution of speed in the next 10 years I can have Claude but local + running constantly and yeah that would change certain things fundamentaly


Try llama 3.3 70B. On groq or something. Runs on a 64GB macbook (4bit quantized, which seems to not impact quality much). Things have come a long way. Compare to llama 2 70b. It's wild


Llama 3.3 70B 8-bit MLX runs on Macbook 128GB at 7+ tokens per second while running a full suite of other tools, even at the 130k tokens size, and behaves with surprising coherence. Reminded me of this time last year, first trying Mixtral 8x22 — which still offers a distinctive je ne sais quoi!


Qwen 2.5 32B Coder is actually a viable locally-hosted alternative to Claude 3.5 Sonnet.

It's not better, but if I couldn't access Claude for some reason, I would definitely use it.


When ChatGPT was down a few days back, I locally booted up Codestral. It was decent and usable.


I don't think so. The amount of features something like Google cloud has + experience is not cheap to just build.

Plenty of companies try in Germany the schwarz group aka Lidl and the German Telekom too.

Also normal companies are not fast. It takes time for them to learn about something new, understanding it and then implementing it before they finally can us it.

If Ms tells you they will do it in 1-2 years everyone is fine with that. And GitHub announced GitHub spark. Google has this already.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: