Hacker Newsnew | past | comments | ask | show | jobs | submit | sergiomattei's commentslogin

Zipped up tightly? This has been all over mainstream news outlets, social media, everywhere.

We are commenting on a submission linked to the New York Times.


Compared to, say, the coverage from Ukraine during February 2022, actual information getting out from the ground is sparser. Or the opening "shock and awe" campaign in Iraq in 2003, there were Western and international media in Baghdad reporting on it in real time, shooting video from their hotels:

https://youtu.be/m8KimNtB9HI

The reason why isn't really a mystery: Iran has never been exactly welcoming to Western media, and internet access there was intentionally shut off after the recent protests. There's plenty of coverage- it's front page everywhere- but a paucity of information.

It's all over social media, but hardly any of that is from Iranians in Iran, it's just people outside it like you and me mostly just yapping. Occasionally you'll hear something second-hand from someone with family in Iran who managed some brief connectivity.


They can't really not talk about it, it's a world war unfolding. It's going to affect every person alive. But as much as it can be, it is absolutely being mitigated in traditional media.

I built my own harness on Elixir/Erlang[0]. It's very nice, but I see why TypeScript is a popular choice.

No serialization/JSON-RPC layer between a TS CLI and Elixir server. TS TUI libraries utilities are really nice (I rewrote the Elixir-based CLI prototype as it was slowing me down). Easy to extend with custom tools without having to write them in Elixir, which can be intimidating.

But you're right that Erlang's computing vision lends itself super well to this problem space.

[1]: https://github.com/matteing/opal


No shade, I think it looks cool and will likely use it, but next time maybe disclose that you’re the founder?


Good point and I will keep that in mind next time.

I am not a founder of this though. This is not a business. It is an open-source project.


This looks nice, congrats on the launch. I'm trying this at work on Monday, I suffer from this problem for tons of MCP tools that are MCP only, not CLI, and completely fill my context window.

First thoughts: it seems the broader community is moving towards Agent Skills as a "replacement" for MCPs to tackle the context pollution problem.

Agent harnesses like Pi don't ship with MCP support as an intentional design choice. MCP servers[0] are being rewritten as pure CLIs in order to support this new scenario.

Thoughts on this?

[0]: https://github.com/microsoft/playwright-cli


Problem with skills is that agents have to load the entire skill for them to use it, or you could break down big skills into multiple smaller ones, which is just a hassle compared to cmcp.

With cmcp you get to use all thr available mcps, while still not bloating context.


Papers like these are much needed bucket of ice water. We antropomorphize these systems too much.

Skimming through conclusions and results, the authors conclude that LLMs exhibit failures across many axes we'd find to be demonstrative of AGI. Moral reasoning, simple things like counting that a toddler can do, etc. They're just not human and you can reasonably hypothesize most of these failures stem from their nature as next-token predictors that happen to usually do what you want.

So. If you've got OpenClaw running and thinking you've got Jarvis from Iron Man, this is probably a good read to ground yourself.

Note there's a GitHub repo compiling these failures from the authors: https://github.com/Peiyang-Song/Awesome-LLM-Reasoning-Failur...


Isn't it strange that we expect them to act like humans even though after a model was trained it remains static? How is this supposed to be even close to "human like" anyway


> Isn't it strange that we expect them to act like humans even though after a model was trained it remains static?

An LLM is more akin to interacting with a quirky human that has anterograde amnesia because it can't form long-term memories anymore, it can only follow you in a long-ish conversation.


If we could reset a human to a prior state after a conversation then would conversations with them not still be "human like"?

I'm not arguing that LLMs are human here, just that your reasoning doesn't make sense.


Henry Molaison was exactly this.


I mean you can continue to evolve the model weights but the performance would suck so we don't do it. Models are built to an optimal state for a general set of benchmarks, and weights are frozen in that state.


> We antropomorphize these systems too much.

They're sold as AGI by the cloud providers and the whole stock market scam will collapse if normies are allowed to peek behind the curtain.


The stock market being built on conjecture? Surely not sir.


> conclude that LLMs exhibit failures across many axes we'd find to be demonstrative of AGI.

Which LLMs? There's tons of them and more powerful ones appear every month.


True but the fundamental architecture tends not to be radically different, it's more about the training/RL regime


But the point is that to even start to claim that a limitation holds for all LLMs you can't use empirical results that have been demonstrated only for a few old models. You either have a theoretical proof, or you have empirical results that hold for all existing models, including the latest ones.


Most of the claims are likely falsified using current models. I wouldn’t take many of them seriously.


I wouldn't take baseless "likely" claims or the people who make them seriously.


I falsified it on another thread


https://en.wikipedia.org/wiki/List_of_cognitive_biases

Specifically, the idea that LLMs fail to solve some tasks correctly due to fundamental limitations where humans also fail periodically well may be an instance of the fundamental attribution error.


My small agent harness[0] does this as well.

The tasks tool is designed to validate a DAG as input, whose non-blocked tasks become cheap parallel subagent spawns using Erlang/OTP.

It works quite well. The only problem I’ve faced is getting it to break down tasks using the tool consistently. I guess it might be a matter of experimenting further with the system prompt.

[1]: https://github.com/matteing/opal


This is probably the most beautiful homepage + docs combo I’ve ever seen. The copy is awesome too. It feels human.

Great work.


Much appreciated. Definitely A LOT of back and forth on the copy, design was actually very few iterations. Docs of course have their own rule in AGENTS.md, since Claude seems only slightly less likely than your average software engineer to forget about documentation. :)


I feel the same way.

> That includes code outside of the happy path, like error handling and input validation. But also other typing exercises like processing an entity with 10 different types, where each type must be handled separately. Or propagating one property through the system on 5 different types in multiple layers.

With AI, I feel I'm less caught up in the minutia of programming and have more cognitive space for the fun parts: engineering systems, designing interfaces and improving parts of a codebase.

I don't mind this new world. I was never too attached to my ability to pump out boilerplate at a rapid pace. What I like is engineering and this new AI world allows me to explore new approaches and connect ideas faster than I've ever been able to before.


> explore new approaches and connect ideas faster

This is the hidden super power of LLM - prototyping without attachment to the outcome.

Ten years ago, if you wanted to explore a major architectural decision, you would be bogged down for weeks in meetings convincing others, then a few more weeks making it happen. Then if it didn't work out, it feels like failure and everyone gets frustrated.

Now it's assumed you can make it work fast - so do it four different ways and test it empirically. LLMs bring us closer to doing actual science, so we can do away with all the voodoo agile rituals and high emotional attachment that used to dominate the decision process.


I basically just _accidentally_ added a major new feature to one of my projects this week.

In the sense that, I was trying to explain what I wanted to do to a coworker and my manager, and we kept going back and forth trying to understand the shape of it and what value it would add and how much time it would be worth spending and what priority we should put on it.

And I was like -- let me just spend like an hour putting together a partially working prototype for you, and claude got _so close_ to just completely one-shotting the entire feature in my first prompt, that I ended up spending 3 hours just putting the finishing touches on it and we shipped it before we even wrote a user story. We did all that work after it was already done. Claude even mocked up a fully interactive UI for our UI designer to work from.

It's literally easier and faster to just tell claude to do something than to explain why you want to do it to a coworker.


That's only because no one understood agile or XP and they've become a "no one actually does that stuff" joke to many. I have first hand experience with prototyping full features in a day or two and throwing the result away. It comes with the added benefit of getting your hands dirty and being able to make more informed decisions when doing the actual implementation. It has always been possible, just most people didn't want to do it.


Are you not concerned that this world is deeply tied to you having an internet connection to one of a couple companies' servers? They can jack up the price, cut you off, etc.


Seeing how things are moving, I'm expecting for compute requirements to go down over a longer time horizon, as most technologies do.

I'd rather spend my time preparing for this new world now.


Not going to last long though, at least not professionally. AI will do the spec and architecture too. The LLM will do the entire pipeline between customer or market research to deployment. This is already possible with bug fixes pretty much. And many features too depending on the business.


It AI gets to that level generally, there won’t be a customer, a market research department, or a software company at all.

But if AI is capable of that it’s not a big step to being capable of doing any white collar job, and we’ll either reorganize our economy completely or collapse.


I don't know. LLMs are great at writing code; but you have to have the right ideas to get decent output.

I spend tons of time handholding LLMs--they're not a replacement for thinking. If you give them a closed-loop problem where it's easy to experiment and check for correctness, then sure. But many problems are open-loop where there's no clear benchmark.

LLMs are powerful if you have the right ideas. Input = output. Otherwise you get slop that breaks often and barely gets the job done, full of hallucinations and incorrect reasoning. Because they can't think for you.


Dude, this is awesome. El Nuevo Dia on the Wii is peak bori brain. :)


Thanks, Sergio! Appreciate it. Was definitely fun getting it all working and seeing that familiar logo pop up on the Wii of all places!


Real courageous from that guy calling someone a "fat uncle" on a Twitter thread. Could've applied that same energy IRL and told him to tone it down.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: