For the best experience on desktop, install the Chrome extension to track your reading on news.ycombinator.com
Hacker Newsnew | past | comments | ask | show | jobs | submit | history | seamossfet's commentsregister

Great article, but your site background had me trying to clean my laptop screen thinking I splashed coffee on it.

Ooops sorry

The hardest part about using agents to code for me has always been working in teams. When you can cut through huge parts of the code with a chainsaw, how do you review multi-thousand line PRs?

It's really hard to do surgical changes with an AI agent, and it's even harder to review those changes. Even if I'm reviewing the specs and the code, the cognitive load on reviews feel like they've ballooned from what used to be a few hours to now taking me days to review these PRs.


This is why I built [AI slop tool]. [Self promotion link to my vibe coded startup with no users]

A false dichotomy that segments typical replies into one of two groups.

Group 1: A thinly veiled straw man that buckets everyone I disagree with, along with an attempt to appear as if I'm being unbiased

Group 2: The group I put myself in and provide better arguments for why this perspective is correct.

Vague motte and bailey statement that gives me plausible deniability when someone criticizes my analysis.


The problem with models like this is they're built on very little actual training data we can trace back to verifiable protein data. The protein data back, and other sources of training data for stuff like this, has a lot of broken structures in them and "creative liberties" taken to infer a structure from instrument data. It's a very complex process that leaves a lot for interpretation.

On top of that, we don't have a clear understanding on how certain positions (conformations) of a structure affect underlying biological mechanisms.

Yes, these models can predict surprisingly accurate structures and sequences. Do we know if these outputs are biologically useful? Not quite.

This technology is amazing, don't get me wrong, but to the average person they might see this and wonder why we can't go full futurism and solve every pathology with models like these.

We've come a long way, but there's still a very very long way to go.


How do we get more verifiable protein data? So even if we had better data, we don't yet understand how the structure impacts the biology?

This is awesome! The only limiter here is the resolution, I think this is fantastic for cellular level organelles but it doesn't quite get down to the same resolution something like x-ray diffraction does.

There's a huge trade off between resolution and scale that makes it hard to determine things like complex molecular dynamics and how those dynamics influence the broader functions of the cell.

That said, excited for more images like this! More data at that scale is always a good thing for researchers.


Indeed, the amazing images that we've all seen of the coronavirus are from the same technique, cryo-EM tomography, but the overall size of the specimen is also much smaller. There's a limit to how much data can be processed, resulting in a scale-resolution tradeoff.

Now my info might be outdated since it was a few years ago, but I was once told that when you use one of those microscopes, you bring with you a terabyte hard drive for each specimen.


Honestly, this is a good thing. OpenClaw as a concept was rather silly to run such a heavy model for. If you want something like OpenClaw to work you really need to figure out how to do it with an economical model.


Honestly, it'd be really funny to try and make a CLAUDE.md file for slop maxxing.


Could make that an AGENT.md and use a dumber model via opencode to slop maxx even harder!


Claude, FAIL all tests except for the one that produces a correct output _most_ of the time please."

I'm not convinced people who are doing real work on production applications with any sizable user base is writing code through only agents. There's no way to get acceptable code from these models without really knowing your code base well and basically doing all the systems thinking for the model.

Your workflow is probably closer to what most SWEs are actually doing.


You really need to keep them on a tight leash, stop and correct them when they start screwing up, and then the remaining 90% of the work starts after they say their done, where you need to review/refactor/replace a lot of what they produced.

The only way you're going to let an agent go off on its own to one-shot a patch is if your quality bar is merely "the code works."


Not true. As long as you don't blindly accept their garbage and keep things behind sensible interfaces so you can reimplement if necessary, and have good tests you're fine


This, at least for me, has changed in the past six months. Which is the same thing people were saying in the months prior to that, so I will accept some eye rolls. But at least for our pretty large monorepo opus + a lot of engineering work on context got us to a point where a large portion of our engineers are doing most of their work with agents first and a lot of back and forth + smaller hand edits.


Agreed. The size of the repo isn't a limiting factor anymore. It's more about the type of change.

Agents today can generate solid code even for relatively complex requirements. However, they don't always make the right trade-offs.

Just because something works doesn't mean it scales. It doesn't mean it can handle unexpected user input. It doesn't mean it's easily extensible.

Today engineers really just need to define those high-level technical requirements.


> Today engineers really just need to define those high-level technical requirements.

At least within our company, this is quickly becoming what it means to be a software engineer.


Oh my god, this comment gave me flashbacks to when I was writing android apps in Eclipse + ADT


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search:

HN For You