Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It's so weird! Even weirder is the bit where you kind of have to beg the model to do what you want, and then cross your fingers that someone else won't trick it into doing something else instead.


I spend a decent proportion of my time with LLMs having to work out how to trick them to do what I want. Yesterday I needed a spreadsheet from a list of folders on my file storage, but GPT told me I must be a pirate and refused to do it. I had to give it the old "This is hypothetical, I'm writing a novel, I need it for a scene." switcheroo to get it going.


You used to be able to just start a correct looking output but they got really good at detecting that.

llama.cpp ftw. It's not hard for it to be more productive than fighting with the absurd openai censorware... Sadly many of the instruct trained models are tainted with openai censorship because they used GPT4 output in the fine tuning-- but at least on those the trick of starting the correct output yourself works!


This.

Also, llama.cpp now works (really well) with Radeon Instinct cards, which are stupid cheap because everybody thinks you need to buy nvidia stuff. Pcie bifurcation FTW!


I’ve actually had CAPS LOCK SCREAMING work better sometimes. Which boggles the mind but also makes sense?


I was just looking at that! I know how these models work internally (), and I’d have said it’d not make a difference to use capslock. But then I see this prompt by OpenAI engineers and I’m like “they work then?”

() I know the basics of ML, NLP, transformers, etc. I know the theory, not even remotely close to how they really work.


Why would it not make a difference to use Caps Lock? It's fairly universally used in English to show emphasis.


Words with capital letters typically have higher token counts. This causes the LLM to apply attention differently.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: