For the best experience on desktop, install the Chrome extension to track your reading on news.ycombinator.com
Hacker Newsnew | past | comments | ask | show | jobs | submit | history | more tobr's commentsregister

I wonder why they fail this specific way. If you just let them do stuff everything quickly turns spaghetti. They seem to overlook obvious opportunities to simplify things or see a pattern and follow through. The default seems to be to add more, rather than rework or adjust what’s already in place.


I suspect it has something to do with a) the average quality of code in open source repos and b) the way the reward signal is applied in RL post-training - does the model face consequences of a brittle implementation for a task?

I wonder if these RL runs can extend over multiple sequential evaluations, where poor design in an early task hampers performance later on, as measured by amount of tokens required to add new functionality without breaking existing functionality.


Yeah I've been wondering if the increasing coding RL is going to draw models towards very short term goals relative to just learning from open source code in the wild


To me this seems like a natural consequence of the next-token prediction model. In one particular prompt you can’t “backtrack” once you’ve emitted a token. You can only move forwards. You can iteratively refine (e.g the agent can one shot itself repeatedly), but the underlying mechanism is still present.

I can’t speak for all humans, but I tend to code “nonlinearly”, jumping back and forth and typically going from high level (signatures, type definitions) to low level (fill in function bodies). I also do a lot of deletion as I decide that actually one function isn’t needed or if I find a simpler way to phrase a particular section.

Edit: in fact thinking on this more, code is _much_ closer to a tree than sequence of tokens. Not sure what to do with that, except maybe to try a tree based generator which iteratively adds child nodes.


This would make sense to me as an explanation when it only outputs code. (And I think it explains why code often ends up subtly mangled when moved in a refactoring, where a human would copy paste, the agent instead has to ”retype” it and often ends up slightly changing formatting, comments, identifiers, etc.)

But for the most part, it’s spending more tokens on analysis and planning than pure code output, and that’s where these problems need to be caught.


I feel like planning is also inherently not sequential. Typically you plan in broad strokes, then recursively jump in and fill in the details. On the surface it doesn’t seem to be all that much different than codegen. Code is just more highly specified planning. Maybe I’m misunderstanding your point?


All it does is generate soup. Some of which may taste good.

There is no thinking, no matter what marketing tells you.


LLMs are next token predictors. Their core functionality boils down to simply adding more stuff.


They do what you tell them to. If you regularly tell them to look for opportunities to clean up/refactor the code, they will.


I also expected hardware to be involved. But in the context of a list of tutorials on how to use this live coding tool the title makes sense though.


And at the same time, the fastest growing consumer product of all time is called ”ChatGPT”.


Perhaps if the product is compelling enough, the name doesn’t matter - and conversely, if the product is borderline, it had better have a great name.


Chat gpt is a great name though — you “chat” with the “GPT” so its self informing (even if you dont know what a GPT is), it’s 4 syllables that roll off the tongue well together.

RSS, has no vowels, no information, and looks like an alphabet term you might see at the doctor’s office or in an HR onboarding form at a corpo.


Randos are just calling it "Chat" now.

"I'll ask Chat about x!"


In Japan it's now known colloquially as 「チャッピー」 ("Chappy" or "Chappie"). High praise that it has received such shortened and personified version so quickly.


It’s the new ”I looked it up on wiki”.


I've heard 'just ai it' from high schoolers.


As a European, my impression is that things named something something ”Euro” tend to be cheap and low quality. I don’t think it’s possible to build a positive consumer brand around ”Eurosky”. I support the cause though - we probably need to find a catchy word like ”Brexit” or ”enshittification” to make it salient.


This is almost universally true for every national identity (or however we want to widen the term to include Euro).

If you have a good product, you usually lead with that. "Made in X" becomes one bullet point in the list of things that make you great. If you lead with "made in X" or even make that your entire brand, that's a sign that you probably don't have much else to bring to the table.

The only real exception are foods and beverages. And even there it's questionable


> Eurosky is a pan-European initiative spearheaded by a coalition of entrepreneurs, technologists and civil society organizations

A brit, a belgian and a german by the looks of their profiles, which are just their linkedin pages.

Posting this to HN feels like some guys trying to do "growth hacking" with Brusselian characteristics.

Honestly I even propose this conjecture: If you are in Europe you will learn about any truly European social media from some other source long before it appears on HN.


Elevator pitch could be "Wirecard for Social Media".


When I read "Eurosky", Skyshield immediately came to mind. Sounds like a military project.


For the record, now it has changed again, to ’Meta’s AI smart glasses and data privacy concerns’, which is even more milquetoast.

Parent and another comment reacting to this change have also been (artificially, I must assume) sunk from top to below gems like ’Too funny that the subcontractor working for meta is “sama”’.


> Be the change, my man. Try to make a podcast.

This might be the funniest thing I’ve read today.


Are you asking what I do or are you asking for advice on what you should do?


Yes


Sounds like neither. More like throw a dice and press the up button.


I had been considering ditching everyday ChatGPT use in favor of Claude anyway, but hadn’t gotten around to it mostly out of habit. Now I have a good reason to do it.


Same, I had put Claude in my metaphorical shopping cart about two weeks ago but I already had some inertia with ChatGPT + Codex and figured it wouldn't be better enough to justify changing.

That has changed, so I canceled my ChatGPT membership and signed up for Claude. I still have five bucks of credit I bought a year ago for the OpenAI API that I do not believe I can have refunded back, so some of my apps are going to have to stick to OpenAI until those credits run out since I'm not going to just donate five bucks to them.

Playing with it now, I honestly can't tell too much of a difference, which as far as I am concerned is a good thing.


With this amount of competition it's almost weird to be paying anyone anything when one can just switch between free tiers of GPT, Claude, Gemini, Kimi, Qwen, Deepseek, Le Chat and an endless firehose of local models. The more your usage is randomly spread out, the less each provider can presumably profile you too as nobody has the full picture.


You should also consider ollama and local models.


Consider carefully the usage limits of both services before deleting your account (as you cannot create a new one later with the same email). Claude's €20/month sub offers very little and this has unfortunately kept me from switching when I tried earlier this month.


I get a fair amount of use out of it. I'm not using it for professional software development, just hobby stuff that I don't to write the boring parts of. For 20 bucks a month that seems pretty reasonable.


I had been using both, ChatGPT mostly for chats and Claude mostly for code. Now I cancelled the ChatGPT subscription and turned on extra usage in Claude instead.


Consider that "little" is very subjective here, I find it to offer more than enough.


Say more. What did you see?


Apple is the only place I've ever worked where I really feel I'll get summarily fired for saying the wrong thing during a meeting or in my pod (if a manager overhears) - the culture is that draconian. So don't hold your breath waiting for someone to tell you things they're under NDA about (or just general litigation pressure).


Their hardware and integration between hardware and software is extremely good, still. The software on its own has gone downhill for a long time.


Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search:

HN For You