Hacker Newsnew | past | comments | ask | show | jobs | submit | lacoolj's commentslogin

What percentage of this code was written by LLM/AI?

For myself, I'd estimate ~50%

Not useful for things it hadn't been trained on before. But now I have the core functionality in place - it's been of great help.


Hey mathematician, how much of this formula did you calculate with an abacus instead of a calculator?

Hey 'software engineer', how much of the output of an LLM it's actually reproducible vs the one from a calculator or any programming language with the same input in different sessions?

A lot vs a human? I bet the LLM with the same prompt will write same code as before more often than I would (given I don’t remember what I wrote in the past).

Not really related to this 'discussion' but this is an interesting problem in the AI space. It's essentially a well understood problem in unreliable distributed systems - if you have a series of steps that might not respond with the same answer every time (because one might fail usually) then how do you get to a useful and reliable outcome? I've been experimenting with running a prompt multiple times and having an agent diff the output to find parts that some runs missed, or having it vote on which run resulted in the best response, with a modicum of success. If you're concerned about having another layer of AI in there then getting the agents to return some structured output that you can just run through a deterministic function is an alternative.

Non-determinism is a problem that you can mitigate to some extent with a bit of effort, and is important if your AI is running without a human-in-the-loop step. If you're there prompting it though then it doesn't actually matter. If you don't get a good result just try again.


Don’t know if this is an annoying response… but how about just going through the code and check and grade the quality yourself?

I could do, but the end goal is to scale this to 100x what I can do myself, and there isn't time to review all those changes. By attempting to answer the problem when it's tiny and I can still keep it in my head then I'll end up building something that works at scale.

Maybe. The point is that this is all new, and looking forwards I think it's worth figuring out this stuff early.


Why are you so concerned about the LLM producing the exact same code across different sessions? Seems like a really weird thing to focus on. Why aren't you focused on things like security, maintainability, UI/UX, performance?

Agreed. It's not like humans can produce the same output given the same input for anything more than trivial inputs.

I'd argue that it's actually a benefit; I like that I can do several generations and compare them and pick the best result. HP, for example, used to do this with software teams, and that's how we got Rocky Mountain BASIC (AIUI the competing team was East Coast BASIC).


This is probably an adjacent result of this (from anthropic launch post):

> In Claude Code, we’ve raised the default effort level to xhigh for all plans.

Try changing your effort level and see what results you get


effort level is separate from tokenization. Tokenization impacts you the same regardless.

I find 5 thinking levels to be super confusing - I dont really get why they went from 3 -> 5


I would love more information.

What exactly did the request for information say from DHS? What exactly was the reason for them to look for you specifically (certainly there are many others protesting)? Following up on that, how do others avoid something like this? What red flags should be avoided and how?

There may or may not be a solid answer for any of this. But this article feels like it's made for awareness, when it could also be made for action, with the right details included.


So it's the age of AI. And this seems like a great new benchmark! Lots of text, structured but each item a separate "task". Each thing requiring its own new image + textual representation.

I copy + pasted the whole article (minus the few included images) and added this prompt in Gemini 3 Pro:

> Take each of the following and add an image representing the act being described. The image should be very basic. Think of signs in buildings - exit signs, bathroom door signs, no smoking signs, etc. That style of simplicity. Just simple, flat, elegant vector graphic lines for the chopsticks, hands, bowls, etc.

Google Gemini output: https://gemini.google.com/share/11df1bc53e3d

I think this is pretty dang good for a one-shot run. I also ran this through Claude Opus 4.6 Extended (doesn't generate images directly, so it made an HTML page and some vector icons). Not as good as Gemini IMO. See here if curious: https://claude.ai/public/artifacts/8b6589b3-4da4-4fd5-b862-c...

Anyone able to do this better with a different prompt or model (or both)?


Nice that you discovered LLM, welcome.

But next time, keep your findings for a thread related to the topic of LLM wonders, not when it's unrelated, such as chopsticks.


> But I think that’s too long. The patent has already served its purpose well, and I believe that holding on to it any longer benefits nobody.

Damn dude didn't you pay like ... over $10k for that patent?


I took care of the whole thing myself without lawyers, so I ended up paying something like $950 in various filing fees.


How did you get that diploma/plaque, is that something every patent author will have?


You can buy them from various manufacturers that make them; you often get unsolicited mail from them as your name and address is on the patent filings.


The plaque was a personal order from one of the many companies that make them. What you actually get from the USPTO looks like this: https://x.com/EricLengyel/status/1159917092331642880/photo/1


Using Claude for code you use yourself or at your own company internally is one thing, but when you start injecting it into widely-shared projects like this (or, the linux kernel, or Debian, etc) there will always be a lingering feeling of the project being tainted.

Just my opinion, probably not a popular one. But I will be avoiding an upgrade to Node.js after 24.14 for a while if this is becoming an acceptable precedent.


I still think everyone is trying to run away from the copyright problems with AI, and suspect it's going to come back to bite them. Eventually. (No I'm not willing to bet on exactly when because I'm sure it'll be a lot longer than I'd like).


Owning copyright of something and writing it are very different things


Not in the US. Copyright exists from the moment the work is created.

Source: https://www.copyright.gov/help/faq/faq-general.html


This page is a good example of when Typescript is both unnecessary and off-putting, based on the content's purpose.

The author is trying to demonstrate a problem and the proposed solution, using example code that the reader then needs to read, formulate a visual "structure" in their mind, and then apply that structure to the subsequent code.

Typescript is not in any way invalid here, and as a tool when programming something that will run in the real-world, can be invaluable.

However, when writing anything you want someone to deeply comprehend, you want to use the least amount of text needed to get your point across. Adding types don't serve any purpose here. The types used are generics, which tell you nothing specific, hence the name, and the names given to the functions and object properties are enough to convey what this code is doing.


Good luck finding a phone with a headphone jack anymore though :(

I love my wired headphones though. They support BT but I've used that maybe twice. Ever. Obviously was only because I was using my phone with them, which again don't have a port for the cord.


Can you get a support ticket in to Anthropic and post the results here?

Would like to see their take on this


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: