Hacker Newsnew | past | comments | ask | show | jobs | submit | aix1's commentslogin

My mildly amusing anecdote is that, whenever Claude Code produces something particularly egregious, I often find it sufficient to reply with just "wtf?" for it to present a much more sensible version of the code (which often needs further refinement, but that's another story...)

But we don't evolve IL or assembly code as the system evolves. We regenerate it from scratch every time.

It is therefore not important whether some intermediate version of that low-level code was completely impossible to understand.

It is not so with LLM-written high-level code. More often than not, it does need to be understood and maintained by someone or something.

These days, I mainly focus on two things in LLM code reviews:

1. Making sure unit tests have good coverage of expected behaviours.

2. Making sure the model is making sound architectural decisions, to avoid accumulating tech debt that'll need to be paid back later. It's very hard to check this with unit tests.


We get stuck reviewing the output assembly when it's broken, and that does happen from time to time. The reason that doesn't happen often is that generation of assembly follows strict rules, which people have tried their best to test. That's not the behavior we're going to get out of a LLM.

Yes, prompts aren't analogous to higher-level code, they're analogous to wizards or something like that which were always rightly viewed with suspicion.

Part of it observability bias: longer, more widespread outages are more likely to draw signficant attention. This doesn't mean that there aren't also shorter, smaller-scope outages, it's just that we're much less likely to know about them.

For example, if there's a problem that gets caught at the 1% stage of a staged rollout, we're probably not going to find ourselves discussing it on HN.


And how is it going, in terms of finding those limit? It would be very interesting to hear about areas where the actual experience turned out to be wildly different from your expectations, in either direction.

This looks cool, but what I'd really like is a self-hosted version that I could use to auto-subtitle videos I already have locally. This would help my language learning a great deal.

If any of you have already figured out a tool/workflow for this, I'd love to learn from your experience.


This thread prompted me to look into this. It seems that all I need is a thin wrapper around whisper-ctranslate2. So I wrote one and am playing with it right now.

I'm finding language auto-detection to be a bit wonky (for example, it repeatedly identified Ladykracher audio as English instead of German). I ended up having to force a language instead. The only show in my library where this approach doesn't work is Parlement[1], but I can live with that.

On the whole this is looking quite promising. Thanks for the idea.

[1] https://en.wikipedia.org/wiki/Parlement_(TV_series)


Another potential factor at play is the accuracy of delivery. It is generally easier to accurately deliver one quick dose vs daily doses over multiple weeks (due to patient positioning errors, the patient losing weight, soft tissues moving around etc).

The 42 -> 137 also jumped out at me. On the face of it, the associated improvement sure does sound like overfitting to the eval set.

Would love to hear more, if you are happy sharing!

I have a Metabo vacuum (ASA 30 H PC) and absolutely love it. What's bad about the ergonomics of yours?

I have a Sebo. The primary thing I dislike is the weight. You'd think something this heavy would have some kind of performance advantage, but it doesn't. I've seen battery powered shit from Walmart suck harder than this machine does.

Matt Levine put it really well: "We will create God and then ask it for money."

The answer is easy: more corruption. No AI God needed

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: