Hacker Newsnew | past | comments | ask | show | jobs | submit | endlessvoid94's commentslogin


Mind Games at number 2? I got that book years ago and was so disappointed I still think about it sometimes.


You just bumped it up by mentioning it ;)


In 2013 I built a smalltalk-esque system browser using ruby and TCL. It was fun: https://www.youtube.com/watch?v=2xC5B5t5Rq8


Same. I modded my original Xbox back in the day. It was a lot of fun and I learned a lot.


Danny Hillis deserves a mention


Absolutely, moreso than Feynman in this context. But, Danny Hills doesn't make a good sound bite.


Anyone remember tunez?

https://tunez.sourceforge.net/


Things 3 has a sort of recurring todo like what you're describing, where you can set like this: https://culturedcode.com/things/support/articles/2803564/


Love it.

On macOS it always launches in the middle of the screen - is there a way to move it around?


To move clippy you want to drag the piece of paper on which clippy sits -- clicking clippy himself will hide and show the chat window.


I've found the local models useful for non-coding tasks, however the 8B parameter models so far have proven lacking enough for coding tasks that I'm waiting another few months for whatever the Moore's law equivalent of LLM power is to catch up. Until then, I'm sticking with Sonnet 3.7.


If you have a 32GB Mac then you should be able to run up to 27B params, I have done so with Google's `gemma3:27b-it-qat`


Hm, I've got an M2 air w/ 24GB. Running the 27B model was crawling. Maybe I had something misconfigured.


No, that sounds right. 24GB isn’t enough to feasibly run 27B parameters. The rule of thumb is approximately 1GB of ram per billion parameters.

Someone in another comment on this post mentioned using one of the micro models (Qwen 0.6B I think?) and having decent results. Maybe you can try that and then progressively move upwards?

EDIT: “Queen” -> “Qwen”


That rule of thumb is only related to 8 bit quants at low context. The default for ollama is 4 bit, which puts it roughly about 14GB.

The vast majority of people run between 4-6 bit depending on system capability. The extra accuracy above 6 tends to not be worth it relative to the performance hit.


You also need to leave space for other apps. If you run a 27B model on a 32GB machine you may find that you can't productively run other apps.

I have 64GB and I can only just fit a bunch of Firefox and VS Code windows at the same time as running a 27B model.


I think only 2/3 of ram is allocated to be available to the gpu, so like 14gb which is probably not enough to run even Q4 quant.


This is configurable by the way.

sudo sysctl iogpu.wired_limit_mb=12345


deepseek-r1:8b screams on my 12gb gpu. gemma3:12b-it-qat runs just fine, a little faster than I can read. Once you exceed GPU ram it offloads a lot of the model to the CPU and splitting between gpu and cpu is dramatically (80? 95%?) slower


How much RAM was it taking during inference?


15.4GB during inference according to Activity Monitor


Oh, nice, that's actually not bad at all. Thanks, will give it a try on my 36Gb Mac


Just a different type of fun. I find avalanche training to have a similar effect for backcountry.

For some it's sobering, for others it's terrifying.


Yes my wife and I were watching a group of hikers one time and we both looked at each other and talked about how none of them had even seen a demo on using an ice axe. It felt like walking into a kitchen and seeing the chefs juggling knives


Yep, and sadly it's a typical story in the backcountry, sometimes ending tragically.


At the end of my three full day avalanche training the instructor said “now remember, you are now the least qualified people to go into the backcountry.

That stuck with me.


That's a great line!


At my small company, only reference checks when we're nearly ready to give an offer. For education, never.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: