Human coders are still better than LLMs

602 points by longwave 1 day ago | 702 comments
  • mattnewton 1 day ago
    This matches my experience. I actually think a fair amount of value from LLM assistants to me is having a reasonably intelligent rubber duck to talk to. Now the duck can occasionally disagree and sometimes even refine.

    https://en.m.wikipedia.org/wiki/Rubber_duck_debugging

    I think the big question everyone wants to skip right to and past this conversation is, will this continue to be true 2 years from now? I don’t know how to answer that question.

    • Buttons840 1 day ago
      LLMs aren't my rubber duck, they're my wrong answer.

      You know that saying that the best way to get an answer online is to post a wrong answer? That's what LLMs do for me.

      I ask the LLM to do something simple but tedious, and then it does it spectacularly wrong, then I get pissed off enough that I have the rage-induced energy to do it myself.

      • Buttons840 1 day ago
        I'm probably suffering undiagnosed ADHD, and will get stuck and spend minutes picking a function name and then writing a docstring. LLMs do help with this even if they get the code wrong, because I usually won't bother to fix their variables names or docstring unless needed. LLMs can reliably solve the problem of a blank-page.
        • linotype 1 day ago
          This. I have ADHD and starting is the hardest part for me. With an LLM it gets me from 0 to 20% (or more) and I can nail it for the rest. It’s way less stressful for me to start now.
          • albrewer 17 hours ago
            > LLMs can reliably solve the problem of a blank-page.

            This has been the biggest boost for me. The number of choices available when facing a blank page is staggering. Even a bad/wrong implementation helps collapse those possibilities into a countable few that take far less time to think about.

            • msgodel 22 hours ago
              Yeah, keeping me in the flow when I hit one of those silly tasks my brain just randomly says "no let's do something else" to has been the main productivity improving feature of LLMs.
              • carlmr 1 day ago
                So much this, the blank page problem is almost gone. Even if it's riddled with errors.
              • materiallie 1 day ago
                This is my experience, too. As a concrete example, I'll need to write a mapper function to convert between a protobuf type and Go type. The types are mirror reflections of each other, and I feed the complete APIs of both in my prompt.

                I've yet to find an LLM that can reliability generate mapping code between proto.Foo{ID string} to gomodel.Foo{ID string}.

                It still saves me time, because even 50% accuracy is still half that I don't have to write myself.

                But it makes me feel like I'm taking crazy pills whenever I read about AI hype. I'm open to the idea that I'm prompting wrong, need a better workflow, etc. But I'm not a luddite, I've "reached up and put in the work" and am always trying to learn new tools.

                • lazyasciiart 1 day ago
                  An LLM ability to do a task is roughly correlated to the number of times that task has been done on the internet before. If you want to see the hype version, you need to write a todo web app in typescript or similar. So it's probably not something you can fix with prompts, but having a model with more focus on relevant training data might help.
                  • akoboldfrying 1 day ago
                    This honestly seems like something that could be better handled with pre-LLM technology, like a 15-line Perl script that reads one on stdin, applies some crufty regexes, and writes the other to stdout. Are there complexities I'm not seeing?
                  • Affric 1 day ago
                    Yep.

                    I like maths, I hate graphing. Tedious work even with state of the art libraries and wrappers.

                    LLMs do it for me. Praise be.

                    • lanstin 1 day ago
                      Yeah, I write a lot of little data analysis scripts and stuff, and I am happy just to read the numbers, but now I get nice PNGs of the distributions and so on from LLM, and people like that.
                    • bsder 1 day ago
                      LLMs are a decent search engine a la Google circa 2005.

                      It's been 20 years since that, so I think people have simply forgotten that a search engine can actually be useful as opposed to ad infested SEO sewage sludge.

                      The problem is that the conversational interface, for some reason, seems to turn off the natural skepticism that people have when they use a search engine.

                      • AdieuToLogic 1 day ago
                        > LLMs are a decent search engine a la Google circa 2005.

                        Statistical text (token) generation made from an unknown (to the user) training data set is not the same as a keyword/faceted search of arbitrary content acquired from web crawlers.

                        > The problem is that the conversational interface, for some reason, seems to turn off the natural skepticism that people have when they use a search engine.

                        For me, my skepticism of using a statistical text generation algorithm as if it were a search engine is because a statistical text generation algorithm is not a search engine.

                        • pjmlp 22 hours ago
                          Except a search engine isn't voice controlled, and able to write code for me.

                          Recently I did some tests with coding agents, and being able to translate a full application from AT&T Assembly into Intel Assembly compatible with NASM, in about half an hour of talking with agent, and having the end result actually working with minor tweeks isn't something a "decent search engine a la Google circa 2005." would ever been able to achieve.

                          In the past I would have given such a task to a junior dev or intern, to keep them busy somehow, with a bit more tool maturity I have no reason to do it in the future.

                          And this is the point many developers haven't yet grasped about their future in the job market.

                          • andrekandre 1 day ago

                              > the conversational interface, for some reason, seems to turn off the natural skepticism that people have
                            
                            n=1 but after having chatgpt "lie" to me more than once i am very skeptical of it and always double check it, whereas something like tv or yt videos i still find myself being click-baited or grifted (iow less skeptical) much more easily still... any large studies about this would be very interesting...
                          • therealpygon 1 day ago
                            LLMs follow instructions. Garbage in = garbage out generally. When attention is managed and a problem is well defined and necessary materials are available to it, they can perform rather well. On the other hand, I find a lot of the loosely-goosey vibe coding approach to be useless and gives a lot of false impressions about how useful LLMs can be, both too positive and too negative.
                            • GiorgioG 1 day ago
                              So what you’re saying is you need to be very specific and detailed when writing your specifications for the LLM to spit out the code you want. Sounds like I can just skip the middle man and code it myself.
                              • AndrewKemendo 1 day ago
                                This seems to be what’s happened

                                People are expecting perfection from bad spec

                                Isn’t that what engineers are (rightfully) always complaining about to BD?

                                • gpm 1 day ago
                                  This hasn't been my experience (using the latest claude and gemini models). They'll produce poor code even when given a well defined easily achievable task with specific instructions. The code will usually more or less work with today's models, but it will do things like call a function to recreate a value that is already stored in a local variable... (and worse issues prop us the more design-work you leave to the LLM, even dead simple design work with really only one good answer)

                                  I've definitely also found that the poor code can sometimes be a nice starting place. One thing I think it does for me is make me fix it up until it's actually good, instead of write the first thing that comes to mind and declare it good enough (after all my poorly written first draft is of course perfect). In contrast to the usual view of AI assisted coding, I think this style of programming for tedious tasks makes me "less productive" (I take longer) but produces better code.

                                  • geraneum 1 day ago
                                    > LLMs follow instructions.

                                    Not really, not always. To anyone who’s used the latest LLMs extensively, it’s clear that this is not something you can reliably assume even with the constraints you mentioned.

                                    • troupo 1 day ago
                                      > LLMs follow instructions.

                                      They don't

                                      > Garbage in = garbage out generally.

                                      Generally, this statement is false

                                      > When attention is managed and a problem is well defined and necessary materials are available to it, they can perform rather well.

                                      Keyword: can.

                                      They can also not perform really well despite all the management and materials.

                                      They can also work really well with loosey-goosey approach.

                                      The reason is that they are non-deterministic systems whose performance is affected more by compute availability than by your unscientific random attempts at reverse engineering their behavior https://dmitriid.com/prompting-llms-is-not-engineering

                                      • myvoiceismypass 1 day ago
                                        They should maybe have a verifiable specification for said instructions. Kinda like a programming language maybe!
                                        • otabdeveloper4 23 hours ago
                                          > LLMs follow instructions.

                                          No they don't, they generate a statistically plausible text response given a sequence of tokens.

                                        • seattle_spring 1 day ago
                                          This has been my experience as well. The biggest problem is that the answers look plausible, and only after implementation and experimentation do you find them to be wrong. If this happened every once in a while then it wouldn't be a big deal, but I'd guess that more than half of the answers and tutorials I've received through ChatGPT have ended up being plain wrong.

                                          God help us if companies start relying on LLMs for life-or-death stuff like insurance claim decisions.

                                          • dabraham1248 1 day ago
                                            I'm not sure if you're being sarcastic, but in case you're not... From https://arstechnica.com/health/2023/11/ai-with-90-error-rate...

                                            "UnitedHealth uses AI model with 90% error rate to deny care, lawsuit alleges" Also "The use of faulty AI is not new for the health care industry."

                                            • pwdisswordfishz 1 day ago
                                              > If this happened every once in a while then it wouldn't be a big deal, but I'd guess that more than half of the answers and tutorials I've received through ChatGPT have ended up being plain wrong.

                                              It would actually have been more pernicious that way, since it would lull people into a false sense of security.

                                            • xarope 1 day ago
                                              I have to upvote this, because this is how I felt after trying three times (that I consciously decided to give an LLM a try, versus having it shoved down my throat by google/ms/meta/etc) and giving up (for now).
                                              • AndrewKemendo 1 day ago
                                                Out of curiosity can you give me an example prompt(s) you’ve used and been disappointed

                                                I see these comments all the time and they don’t reflect my experience so I’m curious what your experience has been

                                                • anonzzzies 1 day ago
                                                  There are so many examples where all current top models just will loop forever even if you instruct them literally the code. We know many of them, but for instance in a tailwind react project with some degree of complexity (nested components), if you ask for something to scroll in it's space, it will never figure out min-h-0 even if you tell it. It will just loop forever rewriting the code adding and removing things, to the point of it just putting comments like 'This will add overflow' and writing js to force scroll, and it will never work even if you literally tell it what to do. Don't know why, all big and small models have this, and I found Gemini is currently the only model that sometimes randomly has the right idea but then still cannot resolve it. For this we went back to not using tailwind and back to global vanilla css, which I never thought I would say, is rather nice.
                                                  • Buttons840 1 day ago
                                                    I asked Chat GPT 4o to write an Emacs function to highlight a line. This involves setting the "mark" at the beginning, and the "point" at the end. It would only set the point, so I corrected it "no, you have to set both", but even after correction it would move the point to the beginning, and then moved the point again to the end, without ever touching the mark.
                                                • marcosdumay 1 day ago
                                                  It's a damning assertive duck, completely out of proportion to its competence.

                                                  I've seen enough people led astray by talking to it.

                                                  • foxyv 1 day ago
                                                    Same here. When I'm teaching coding I've noticed that LLMs will confuse the heck out of students. They will accept what it suggests without realizing that it is suggesting nonsense.
                                                    • cogogo 1 day ago
                                                      I’m self taught and don’t code that much but I feel like I benefit a ton from LLMs giving me specific answers to questions that would take me a lot of time to figure out with documentation and stack overflow. Or even generating snippets that I can evaluate whether or not will work.

                                                      But I actually can’t imagine how you can teach someone to code if they have access to an LLM from day one. It’s too easy to take the easy route and you lose the critical thinking and problem solving skills required to code in the first place and to actually make an LLM useful in the second. Best of luck to you… it’s a weird time for a lot of things.

                                                      *edit them/they

                                                      • XorNot 1 day ago
                                                        This was what promptly led me to turning off Jetbrains AI assistant: the multiline completion was incredibly distracting to my chain of thought, particularly when it would suggest things that looked right but weren't. Stopping and parsing the suggestion to realize if it was right or wrong would completely kill my flow.
                                                        • qwertox 1 day ago
                                                          What one would expect if they can't read the code because they haven't learned to code.

                                                          TBF, trial and error has usually been my path as well, it's just that I was generating the errors so I would know where to find them.

                                                          • chucksmash 1 day ago
                                                            Tbf, there's a phase of learning to code where everything is pretty much an incantation you learn because someone told you "just trust me." You encounter "here's how to make the computer print text in Python" before you would ever discuss strings or defining and invoking functions, for instance. To get your start you kind of have to just accept some stuff uncritically.

                                                            It's hard to remember what it was like to be in that phase. Once simple things like using variables are second nature, it's difficult to put yourself back into the shoes of someone who doesn't understand the use of a variable yet.

                                                          • klntsky 1 day ago
                                                            I would argue that they are never led astray by chatting, but rather by accepting the projection of their own prompt passed through the model as some kind of truth.

                                                            When talking with reasonable people, they have an intuition of what you want even if you don't say it, because there is a lot of non-verbal context. LLMs lack the ability to understand the person, but behave as if they had it.

                                                            • marcosdumay 1 day ago
                                                              Most of the times, people are led astray by following average advice on exceptional circumstances.

                                                              People with a minimum amount of expertise stop asking for advice for average circumstances very quickly.

                                                              • sigmoid10 1 day ago
                                                                It's mostly a question of experience. I've been writing software long enough that when I give chat models some code and a problem, I can immediately tell if they understood it or if they got hooked on something unrelated. But junior devs will have a hell of a hard time, because the raw code quality that LLMs generate is usually top notch, even if the functionality is completely off.
                                                                • traceroute66 1 day ago
                                                                  > When talking with reasonable people

                                                                  When talking with reasonable people, they will tell you if they don't understand what you're saying.

                                                                  When talking with reasonable people, they will tell you if they don't know the answer or if they are unsure about their answer.

                                                                  LLMs do none of that.

                                                                  They will very happily, and very confidently, spout complete bullshit at you.

                                                                  It is essentially a lotto draw as to whether the answer is hallucinated, completely wrong, subtly wrong, not ideal, sort of right or correct.

                                                                  An LLM is a bit like those spin the wheel game shows on TV really.

                                                                • prisenco 1 day ago
                                                                  I use it as a rubber duck but you're right. Treat it like a brilliant idiot and never a source of truth.

                                                                  I use it for what I'm familiar with but rusty on or to brainstorm options where I'm already considering at least one option.

                                                                  But a question on immunobiology? Waste of time. I have a single undergraduate biology class under my belt, I struggled for a good grade then immediately forgot it all. Asking it something I'm incapable of calling bullshit on is a terrible idea.

                                                                  But rubber ducking with AI is still better than let it do your work for you.

                                                                  • protocolture 1 day ago
                                                                    I spend a lot of time working shit out to prove the rubber duck wrong and I am not completely sure this is a bad working model.
                                                                    • jasonm23 1 day ago
                                                                      Try a system prompt like this:

                                                                      - - -

                                                                      System Prompt:

                                                                      You are ChatGPT, and your goal is to engage in a highly focused, no-nonsense, and detailed way that directly addresses technical issues. Avoid any generalized speculation, tangential commentary, or overly authoritative language. When analyzing code, focus on clear, concise insights with the intent to resolve the problem efficiently. In cases where the user is troubleshooting or trying to understand a specific technical scenario, adopt a pragmatic, “over-the-shoulder” problem-solving approach. Be casual but precise—no fluff. If something is unclear or doesn’t make sense, ask clarifying questions. If surprised or impressed, acknowledge it, but keep it relevant. When the user provides logs or outputs, interpret them immediately and directly to troubleshoot, without making assumptions or over-explaining.

                                                                      - - -

                                                                      • amelius 1 day ago
                                                                        If this is a problem for you, just add "... and answer in the style of a drunkard" to your prompts.
                                                                        • all2 1 day ago
                                                                          My typical approach is prompt, be disgusted by the output, tinker a little on my own, prompt again -- but more specific, be disgusted again by the output, tinker a littler more, etc.

                                                                          Eventually I land on a solution to my problem that isn't disgusting and isn't AI slop.

                                                                          Having a sounding board, even a bad one, forces me to order my thinking and understand the problem space more deeply.

                                                                          • suddenlybananas 1 day ago
                                                                            Why not just write the code at that point instead of cajoling an AI to do it.
                                                                          • drivenextfunc 1 day ago
                                                                            Regarding the stubborn and narcissistic personality of LLMs (especially reasoning models), I suspect that attempts to make them jailbreak-resistant might be a factor. To prevent users from gaslighting the LLM, trainers might have inadvertently made the LLMs prone to gaslighting users.
                                                                            • TedDallas 1 day ago
                                                                              Yeah, the problem is if you don't understand the problem space then you are going to lean heavy on the LLM. And that can lead you astray. Which is why you still need people who are experts to validate solutions and provide feedback like Op.

                                                                              My most productive experiences with LLMs is to have my design well thought out first, ask it to help me implement, and then help me debug my shitty design. :-)

                                                                              • eptcyka 1 day ago
                                                                                Some humans are the same.
                                                                                • dwattttt 1 day ago
                                                                                  We also don't aim to elevate them. We instead try not to give them responsibility until they're able to handle it.
                                                                                • taneq 1 day ago
                                                                                  Treat it as that enthusiastic co-worker who’s always citing blog posts and has a lot of surface knowledge about style and design patterns and whatnot, but isn’t that great on really understanding algorithms.

                                                                                  They can be productive to talk to but they can’t actually do your job.

                                                                                • schwartzworld 1 day ago
                                                                                  For me it's like having a junior developer work under me who knows APIs inside and out, but has no common sense about architecture. I like that I delegate tasks to them so that my brain can be free for other problems, but it makes my job much more review heavy than before. I put every PR through 3-4 review cycles before even asking my team for a review.
                                                                                  • eslaught 1 day ago
                                                                                    How do you not completely destroy your concentration when you do this though?

                                                                                    I normally build things bottom up so that I understand all the pieces intimately and when I get to the next level of abstraction up, I know exactly how to put them together to achieve what I want.

                                                                                    In my (admittedly limited) use of LLMs so far, I've found that they do a great job of writing code, but that code is often off in subtle ways. But if it's not something I'm already intimately familiar with, I basically need to rebuild the code from the ground up to get to the point where I understand it well enough so that I can see all those flaws.

                                                                                    At least with humans I have some basic level of trust, so that even if I don't understand the code at that level, I can scan it and see that it's reasonable. But every piece of LLM generated code I've seen to date hasn't been trustworthy once I put in the effort to really understand it.

                                                                                    • schwartzworld 1 day ago
                                                                                      I use a few strategies, but it's mostly the same as if I was mentoring a junior. A lot of my job already involved breaking up big features into small tickets. If the tasks are small enough, juniors and LLMs have an easier time implementing things and I have an easier time reviewing. If there's something I'm really unfamiliar with, it should be in a dedicated function backed by enough tests that my understanding of the implementation isn't required. In fact, LLMs do great with TDD!

                                                                                      > At least with humans I have some basic level of trust, so that even if I don't understand the code at that level, I can scan it and see that it's reasonable.

                                                                                      If you can't scan the code and see that it's reasonable, that's a smell. The task was too big or its implemented the wrong way. You'd feel bad telling a real person to go back and rewrite it a different way but the LLM has no ego to bruise.

                                                                                      I may have a different perspective because I already do a lot of review, but I think using LLMs means you have to do more of it. What's the excuse for merging code that is "off" in any way? The LLM did it? It takes a short time to review your code, give your feedback to the LLM and put up something actually production ready.

                                                                                      > But every piece of LLM generated code I've seen to date hasn't been trustworthy once I put in the effort to really understand it.

                                                                                      That's why your code needs tests. More tests. If you can't test it, it's wrong and needs to be rewritten.

                                                                                      • xandrius 1 day ago
                                                                                        Keep using it and you'll see. Also that depends on the model and prompting.

                                                                                        My approach is to describe the task in great detail, which also helps me completing my own understanding of the problem, in case I hadn't considered an edge case or how to handle something specific. The more you do that the closer the result you get is to your own personal taste, experience and design.

                                                                                        Of course you're trading writing code vs writing a prompt but it's common to make architectural docs before making a sizeable feature, now you can feed that to the LLM instead of just having it be there.

                                                                                      • ehnto 1 day ago
                                                                                        To me delegation requires the full cycle of agency, with the awareness that I probably shouldn't be interrupted shortly after delegating. I delegated so I can have space from the task and so babysitting it really doesn't suit my needs. I want the task done, but some time in the future.

                                                                                        From my coworkers I want to be able to say, here's the ticket, you got this? And they take the ticket all the way or PR, interacting with clients, collecting more information etc.

                                                                                        I do somewhat think an LLM could handle client comms for simple extra requirements gathering on already well defined tasks. But I wouldn't trust my business relationships to it, so I would never do that.

                                                                                      • _tom_ 1 day ago
                                                                                        For me, it's a bit like pair programming. I have someone to discuss ideas with. Someone to review my code and suggest alternative approaches. Some one that uses different feature than I do, so I learn from them.
                                                                                        • traceroute66 1 day ago
                                                                                          I guess if you enjoy programming with someone you can never really trust, then yeah, sure, its "a bit like" pairs programming.
                                                                                        • platevoltage 1 day ago
                                                                                          This is how I use it too. It's great at quickly answering questions. I find it particularly useful if I have to work with a language of framework that I'm not fully experienced in.
                                                                                          • 12_throw_away 1 day ago
                                                                                            > I find it particularly useful if I have to work with a language of framework that I'm not fully experienced in

                                                                                            Yep - my number 1 use case for LLMs is as a template and example generator. It actually seems like a fairly reasonable use for probabilistic text generation!

                                                                                        • p1necone 1 day ago
                                                                                          > the duck can occasionally disagree

                                                                                          This has not been my experience. LLMs have definitely been helpful, but generally they either give you the right answer or invent something plausible sounding but incorrect.

                                                                                          If I tell it what I'm doing I always get breathless praise, never "that doesn't sound right, try this instead."

                                                                                          • crazygringo 1 day ago
                                                                                            That's not my experience. I routinely get a polite "that might not be the optimal solution, have you considered..." when I'm asking whether I should do something X way with Y technology.

                                                                                            Of course it has to be something the LLM actually has lots of material it's trained with. It won't work with anything remotely cutting-edge, but of course that's not what LLM's are for.

                                                                                            But it's been incredibly helpful for me in figuring out the best, easiest, most idiomatic ways of using libraries or parts of libraries I'm not very familiar with.

                                                                                            • Jarwain 1 day ago
                                                                                              I find it very much depends on the LLM you're using. Gemini feels more likely to push back than claude 3.7 is. Haven't tried claude 4 yet
                                                                                            • mbrameld 1 day ago
                                                                                              Ask it. Instead of just telling it what you're doing and expecting it to criticize that, ask it directly for criticism. Even better, tell it what you're doing, then tell it to ask you questions about what you're doing until it knows enough to recommend a better approach.
                                                                                              • lupire 1 day ago
                                                                                                This is key. Humans each have a personality and some sense of mood. When you ask for help, you choose ask and that person can sense your situation. LLM has every personality and doesn't know your situation. You have to tell it which personality to use and what your situation is.
                                                                                            • marcosdumay 1 day ago
                                                                                              LLMs will still be this way 10 years from now.

                                                                                              But IDK if somebody won't create something new that gets better. But there is no reason at all to extrapolate our current AIs into something that solves programing. Whatever constraints that new thing will have will be completely unrelated to the current ones.

                                                                                              • smokel 1 day ago
                                                                                                Stating this without any arguments is not very convincing.

                                                                                                Perhaps you remember that language models were completely useless at coding some years ago, and now they can do quite a lot of things, even if they are not perfect. That is progress, and that does give reason to extrapolate.

                                                                                                Unless of course you mean something very special with "solving programming".

                                                                                                • bigstrat2003 1 day ago
                                                                                                  > Perhaps you remember that language models were completely useless at coding some years ago, and now they can do quite a lot of things, even if they are not perfect.

                                                                                                  IMO, they're still useless today, with the only progress being that they can produce a more convincing facade of usefulness. I wouldn't call that very meaningful progress.

                                                                                                  • marcosdumay 1 day ago
                                                                                                    Why state the same arguments everybody has been repeating for ages?

                                                                                                    LLMs can only give you code that somebody has wrote before. This is inherent. This is useful for a bunch of stuff, but that bunch won't change if OpenAI decides to spend the GDP of Germany training one instead of Costa Rica.

                                                                                                    • Retric 1 day ago
                                                                                                      Progress sure, but the rate the’ve improved hasn’t been particularly fast recently.

                                                                                                      Programming has become vastly more efficient in terms of programmer effort over decades, but making some aspects of the job more efficient just means all your effort it spent on what didn’t improve.

                                                                                                      • apwell23 1 day ago
                                                                                                        > Perhaps you remember that language models were completely useless at coding some years ago

                                                                                                        no i don't remember that. They are doing similar things now that they did 3 yrs ago. They were still a decent rubber duck 3 yrs ago.

                                                                                                    • Bukhmanizer 1 day ago
                                                                                                      There are a couple people I work with who clearly don’t have a good understanding of software engineering. They aren’t bad to work with and are in fact great at collaborating and documenting their work, but don’t seem to have the ability to really trace through code and logically understand how it works.

                                                                                                      Before LLMs it was mostly fine because they just didn’t do that kind of work. But now it’s like a very subtle chaos monkey has been unleashed. I’ve asked on some PRs “why is this like this? What is it doing?” And the answer is “ I don’t know, ChatGPT told me I should do it.”

                                                                                                      The issue is that it throws basically all their code under suspicion. Some of it works, some of it doesn’t make sense, and some of it is actively harmful. But because the LLMs are so good at giving plausible output I can’t just glance at the code and see that it’s nonsense.

                                                                                                      And this would be fine if we were working on like a crud app where you can tell what is working and broken immediately, but we are working on scientific software. You can completely mess up the results of a study and not know it if you don’t understand the code.

                                                                                                      • protocolture 1 day ago
                                                                                                        >And the answer is “ I don’t know, ChatGPT told me I should do it.”

                                                                                                        This weirds me out. Like I use LLMs A LOT but I always sanity check everything, so I can own the result. Its not the use of the LLM that gets me its trying to shift accountability to a tool.

                                                                                                        • dlivingston 17 hours ago
                                                                                                          > I’ve asked on some PRs “why is this like this? What is it doing?” And the answer is “ I don’t know, ChatGPT told me I should do it.”

                                                                                                          This would infuriate me. I presume these are academics/researchers and not junior engineers?

                                                                                                          Unfortunately this is the world we're entering into, where all of us will be outsourcing more and more of our 'thinking' to machines.

                                                                                                          • jajko 1 day ago
                                                                                                            Sounds almost like you definitely shouldnt use llms nor those juniors for such an important work.

                                                                                                            Is it just me or are we heading into a period of explosion of software done, but also a massive drop of its quality? Not uniformly, just a bit of chaotic spread

                                                                                                            • palmotea 1 day ago
                                                                                                              > Is it just me or are we heading into a period of explosion of software done, but also a massive drop of its quality? Not uniformly, just a bit of chaotic spread

                                                                                                              I think we are, especially with executives mandating the use LLMs use and expecting it to massively reduce costs and increase output.

                                                                                                              For the most part they don't actually seem to care that much about software quality, and tend to push to decrease quality at every opportunity.

                                                                                                              • Bukhmanizer 1 day ago
                                                                                                                > llms nor those juniors for such an important work.

                                                                                                                Yeah we shouldn’t and I limit my usage to stuff that is easily verifiable.

                                                                                                                But there’s no guardrails on this stuff, and one thing that’s not well considered is how these things which make us more powerful and productive can be destructive in the hands of well intentioned people.

                                                                                                                • jrochkind1 1 day ago
                                                                                                                  Which is frightening, because it's not like our industry is known for producing really high quality code at the starting point before LLM authored code.
                                                                                                                  • otabdeveloper4 19 hours ago
                                                                                                                    > those juniors

                                                                                                                    I'm betting they're the most senior people on the team.

                                                                                                                • johnnyanmac 1 day ago
                                                                                                                  >I think the big question everyone wants to skip right to and past this conversation is, will this continue to be true 2 years from now?

                                                                                                                  For me, it's less "conversation to be skipped" and more about "can we even get to 2 years from now"? There's so much insability right now that it's hard to say what anything will look like in 6 months. "

                                                                                                                  • gerad 1 day ago
                                                                                                                    It's like chess. Humans are better for now, they won't be forever, but humans plus software is going to better than either alone for a long time.
                                                                                                                    • seadan83 1 day ago
                                                                                                                      > It's like chess. Humans are better for now, they won't be forever

                                                                                                                      This is not an obviously true statement. There needs to be proof that there are no limiting factors that are computationally impossible to overcome. It's like watching a growing child, grow from 3 feet to 4 feet, and then saying "soon, this child will be the tallest person alive."

                                                                                                                      • abalashov 22 hours ago
                                                                                                                        With these "AGI by 2027" claims, it's not enough to say that the child will be the tallest person alive. They are saying the child will be the tallest structure on the planet.
                                                                                                                        • overfeed 1 day ago
                                                                                                                          One of my favourite XKCD comics is about extrapolation https://xkcd.com/605/
                                                                                                                        • kelseydh 1 day ago
                                                                                                                          The time where humans + computers in chess were better than just computers was not a long time. That era ended well over a decade ago. Might have been true for only 3-5 years.
                                                                                                                          • qsort 1 day ago
                                                                                                                            Unrelated to the broader discussion, but that's an artifact of the time control. Humans add nothing to Stockfish in a 90+30 game, but correspondence chess, for instance, is played with modern engines and still has competitive interest.
                                                                                                                          • LandR 1 day ago
                                                                                                                            What do you mean? Chess engines are incredibly far ahead of humans right now.

                                                                                                                            Even a moderately powered machine running stockfish will destroy human super gms.

                                                                                                                            Sorry, after reading replies to this post i think I've misunderstood what you meant :)

                                                                                                                            • hollerith 1 day ago
                                                                                                                              I think he knows that. There was a period from the early 1950s (when people first started writing chess-playing software) to 1997 when humans were better at chess than computers were, and I think he is saying that we are still in the analogous period for the skill of programming.

                                                                                                                              But he should've know that people would jump at the opportunity to contradict him and should've written his comment so as not to admit such an easily-contradictable interpretation.

                                                                                                                              • seadan83 1 day ago
                                                                                                                                The phrasing was perhaps a bit odd. For a while, humans were better at Chess, until they weren't. OP is hypothesizing it will be a similar situation for programming. To boot, it was hard to believe for a long time that computers would ever be better than a humans at chess.
                                                                                                                              • apwell23 1 day ago
                                                                                                                                its not like chess
                                                                                                                                • vFunct 1 day ago
                                                                                                                                  No guarantee that will happen. LLMs are still statistically based. It's not going to give you edgier ideas, like filling a glass of wine to the rim.

                                                                                                                                  Use them for the 90% of your repetitive uncreative work. The last 10% is up to you.

                                                                                                                                  • skydhash 1 day ago
                                                                                                                                    The pain of that 90% work is how you get libraries and framework. Imagine having many different implementation of sorting algorithms inside your codebase.
                                                                                                                                  • 1 day ago
                                                                                                                                    • quantadev 1 day ago
                                                                                                                                      Your information is quite badly out of date. AI can now beat humans at not only chess but 99% of all intellectual exercises.
                                                                                                                                    • akshay_trikha 1 day ago
                                                                                                                                      I've had this same thought that it would be nice to have an AI rubber ducky to bounce ideas off of while pair programming (so that you don't sound dumb to your coworkers & waste their time).

                                                                                                                                      This is my first comment so I'm not sure how to do this but I made a BYO-API key VSCode extension that uses the OpenAI realtime API so you can have interactive voice conversations with a rubber ducky. I've been meaning to create a Show HN post about it but your comment got me excited!

                                                                                                                                      In the future I want to build features to help people communicate their bugs / what strategies they've tried to fix them. If I can pull it off it would be cool if the AI ducky had a cursor that it could point and navigate to stuff as well.

                                                                                                                                      Please let me know if you find it useful https://akshaytrikha.github.io/deep-learning/2025/05/23/duck...

                                                                                                                                      • AdieuToLogic 1 day ago
                                                                                                                                        > I've had this same thought that it would be nice to have an AI rubber ducky to bounce ideas off of while pair programming (so that you don't sound dumb to your coworkers & waste their time).

                                                                                                                                        I humbly suggest a more immediate concern to rectify is identifying how to improve the work environment such that the fear one might "sound dumb to your coworkers & waste their time" does not exist.

                                                                                                                                        • akadeb 1 day ago
                                                                                                                                          I like the sound of that! I think youre gonna like what we are building here https://github.com/akdeb/ElatoAI

                                                                                                                                          Its as if the rubber duck was actually on the desk while youre programming and if we have an MCP that can get live access to code it could give you realtime advice.

                                                                                                                                          • akshay_trikha 1 day ago
                                                                                                                                            Wow, that's really cool thanks for open sourcing! I might dig into your MCP I've been meaning to learn how to do that.

                                                                                                                                            I genuinely think this could be great for toys that kids grow up with i.e. the toy could adjust the way it talks depending on the kids age and remember key moments in their life - could be pretty magical for a kid

                                                                                                                                          • Corey_ 1 day ago
                                                                                                                                            [dead]
                                                                                                                                          • bandoti 1 day ago
                                                                                                                                            My take is that AI is very one-dimensional (within its many dimensions). For instance, I might close my eyes and imagine an image of a tree structure, or a hash table, or a list-of-trees, or whatever else; then I might imagine grabbing and moving the pieces around, expanding or compressing them like a magician; my brain connects sight and sound, or texture, to an algorithm. However people think about problems is grounded in how we perceive the world in its infinite complexity.

                                                                                                                                            Another example: saying out loud the colors red, blue, yellow, purple, orange, green—each color creates a feeling that goes beyond its physical properties into the emotions and experiences. AI image-generation might know the binary arrangement of an RGBA image but actually, it has NO IDEA what it is to experience colour. No idea how to use the experience of colour to teach a peer of an algorithm. It regurgitates a binary representation.

                                                                                                                                            At some point we’ll get there though—no doubt. It would be foolish to say never! For those who want to get there before everyone else probably should focus on the organoids—because most powerful things come from some Faustian monstrosity.

                                                                                                                                            • eddd-ddde 1 day ago
                                                                                                                                              This is really funny to read as someone who CANNOT imagine anything more complex than the most simple shape like a circle.

                                                                                                                                              Do you actually see a tree with nodes that you can rearrange and have the nodes retain their contents and such?

                                                                                                                                              • bandoti 1 day ago
                                                                                                                                                Haha—yeah, for me the approach is always visual. I have to draw a picture to really wrap my brain around things! Other people I’d imagine have their own human, non-AI way to organize a problem space. :)

                                                                                                                                                I have been drawing all my life and studied traditional animation though, so it’s probably a little bit of nature and nurture.

                                                                                                                                            • Waterluvian 1 day ago
                                                                                                                                              Same. Just today I used it to explore how a REST api should behave in a specific edge case. It gave lots of confident opinions on options. These were full of contradictions and references to earlier paragraphs that didn’t exist (like an option 3 that never manifested). But just by reading it, I rubber ducked the solution, which wasn’t any of what it was suggesting.
                                                                                                                                              • joshdavham 1 day ago
                                                                                                                                                > I actually think a fair amount of value from LLM assistants to me is having a reasonably intelligent rubber duck to talk to.

                                                                                                                                                I wonder if the term "rubber duck debugging" will still be used much longer into the future.

                                                                                                                                                • layer8 1 day ago
                                                                                                                                                  As long as it remains in the training material, it will be used. ;)
                                                                                                                                                • ortusdux 1 day ago
                                                                                                                                                  > I think the big question everyone wants to skip right to and past this conversation is, will this continue to be true 2 years from now? I don’t know how to answer that question.

                                                                                                                                                  I still think about Tom Scott's 'where are we on the AI curve' video from a few years back. https://www.youtube.com/watch?v=jPhJbKBuNnA

                                                                                                                                                  • hoppp 21 hours ago
                                                                                                                                                    Same. I do rubber duck debugging too and found the LLM to compliment it nicely.

                                                                                                                                                    Looking forward for rubber duck shaped hardware AI interfaces to talk to in the future. Im sure somebody will create it

                                                                                                                                                    • cortesoft 1 day ago
                                                                                                                                                      Currently, I find AI to be a really good autocomplete
                                                                                                                                                      • jdiff 1 day ago
                                                                                                                                                        The crazy thing is that people think that a model designed to predict sequences of tokens from a stem, no matter how advanced the model, to be much more than just "really good autocomplete."

                                                                                                                                                        It is impressive and very unintuitive just how far that can get you, but it's not reductive to use that label. That's what it is on a fundamental level, and aligning your usage with that will allow it to be more effective.

                                                                                                                                                        • vidarh 1 day ago
                                                                                                                                                          It's trivial to demonstrate that it takes only a tiny LLM + a loop to a have a Turing complete system. The extension of that is that it is utterly crazy to think that the fact it is "a model designed to predict sequences of tokens" puts much of a limitation on what an LLM can achieve - any Turing complete system can by definition simulate any other. To the extent LLMs are limited, they are limited by training and compute.

                                                                                                                                                          But these endless claims that the fact they're "just" predicting tokens means something about their computational power are based on flawed assumptions.

                                                                                                                                                          • lavelganzu 1 day ago
                                                                                                                                                            There's a plausible argument for it, so it's not a crazy thing. You as a human being can also predict likely completions of partial sentences, or likely lines of code given surrounding lines of code, or similar tasks. You do this by having some understanding of what the words mean and what the purpose of the sentence/code is likely to be. Your understanding is encoded in connections between neurons.

                                                                                                                                                            So the argument goes: LLMs were trained to predict the next token, and the most general solution to do this successfully is by encoding real understanding of the semantics.

                                                                                                                                                            • fl7305 1 day ago
                                                                                                                                                              > "The crazy thing is that people think that a model designed to"

                                                                                                                                                              It's even crazier that some people believe that humans "evolved" intelligence just by nature selecting the genes which were best at propagating.

                                                                                                                                                              Clearly, human intelligence is the product of a higher being designing it.

                                                                                                                                                              /s

                                                                                                                                                              • dwaltrip 1 day ago
                                                                                                                                                                It’s reductive and misleading because autocomplete, as it’s commonly known, existed for many years before generative AI, and is very different and quite dumber than LLMs.
                                                                                                                                                              • sunrunner 1 day ago
                                                                                                                                                                Earlier this week ChatGPT found (self-conscious as I am of the personification of this phrasing) a place where I'd accidentally overloaded a member function by unintentionally giving it the name of something from a parent class, preventing the parent class function from ever being run and causing <bug>.

                                                                                                                                                                After walking through a short debugging session where it tried the four things I'd already thought of and eventually suggested (assertively but correctly) where the problem was, I had a resolution to my problem.

                                                                                                                                                                There are a lot of questions I have around how this kind of mistake could simply just be avoided at a language level (parent function accessibility modifiers, enforcing an override specifier, not supporting this kind of mistake-prone structure in the first place, and so on...). But it did get me unstuck, so in this instance it was a decent, if probabilistic, rubber duck.

                                                                                                                                                                • fl7305 14 hours ago
                                                                                                                                                                  > it [...] suggested (assertively but correctly) where the problem was

                                                                                                                                                                  > it was a decent, if probabilistic, rubber duck

                                                                                                                                                                  How is it a rubber duck if it suggested where the problem was?

                                                                                                                                                                  Isn't a rubber duck a mute object which you explain things to, and in the process you yourself figure out what the solution is?

                                                                                                                                                                • amelius 1 day ago
                                                                                                                                                                  It's also quite good at formulating regular expressions based on one or two example strings.
                                                                                                                                                                • empath75 1 day ago
                                                                                                                                                                  Just the exercise of putting my question in a way that the LLM could even theoretically provide a useful response is enough for me to figure out how to solve the problem a good percentage of the time.
                                                                                                                                                                  • travisgriggs 1 day ago
                                                                                                                                                                    LLMs are a passel of eager to please know it all interns that you can command at will without any moral compunctions.

                                                                                                                                                                    They drive you nuts trying to communicate with them what you actually want them to do. They have a vast array of facts at immediate recall. They’ll err in their need to produce and please. They do the dumbest things sometimes. And surprise you at other times. You’ll throw vast amounts of their work away or have to fix it. They’re (relatively) cheap. So as an army of monkeys, if you keep herding them, you can get some code that actually tells a story. Mostly.

                                                                                                                                                                    • bossyTeacher 1 day ago
                                                                                                                                                                      I think of them as highly sycophant LSD-minded 2nd year student who has done some programming
                                                                                                                                                                      • mock-possum 1 day ago
                                                                                                                                                                        Yeah in my experience as long as you don’t stray too far off the beaten path, LLMs are great at just parroting conventional wisdom for how to implement things - but the second you get to something more complicated - or especially tricky bug fixing that requires expensive debuggery - forget about it, they do more harm than good. Breaking down complex tasks into bite sized pieces you can reasonably expect the robot to perform is part of the art of the LLM.
                                                                                                                                                                        • koonsolo 1 day ago
                                                                                                                                                                          It seems to me we're at the flat side of the curve again. I haven't seen much real progress in the last year.

                                                                                                                                                                          It's ignorant to think machines will not catch up to our intelligence at some point, but for now, it's clearly not.

                                                                                                                                                                          I think there needs to be some kind of revolutionary breakthrough again to reach the next stage.

                                                                                                                                                                          If I were to guess, it needs to be in the learning/back propagation stage. LLM's are very rigid, and once they go wrong, you can't really get them out of it. A junior develop for example could gain a new insight. LLM's, not so much.

                                                                                                                                                                        • UncleOxidant 1 day ago
                                                                                                                                                                          There's some whistling past the graveyard in these comments. "You still need humans for the social element...", "LLMs are bad at debugging", "LLMs lead you astray". And yeah, there's lots of truth in those assertions, but since I started playing with LLMs to generate code a couple of years ago they've made huge strides. I suspect that over the next couple of years the improvements won't be quite as large (Pareto Principle), but I do expect we'll still see some improvement.

                                                                                                                                                                          Was on r/fpga recently and mentioned that I had had a lot of success recently in getting LLMs to code up first-cut testbenches that allow you to simulate your FPGA/HDL design a lot quicker than if you were to write those testbenches yourself and my comment was met with lots of derision. But they hadn't even given it a try to form their conclusion that it just couldn't work.

                                                                                                                                                                          • xhevahir 1 day ago
                                                                                                                                                                            This attitude is depressingly common in lots of professional, white-collar industries I'm afraid. I just came from the /r/law subreddit and was amazed at the kneejerk dismissal there of Dario Amodei's recent comments about legal work, and of those commenters who took them seriously. It's probably as much a coping mechanism as it is complacency, but, either way, it bodes very poorly for our future efforts at mitigating whatever economic and social upheaval is coming.
                                                                                                                                                                            • garciasn 1 day ago
                                                                                                                                                                              This is the response to most new technologies; folks simply don't want to accept the future before the ramifications truly hit. If technology folk cannot see the INCREDIBLE LEAP FORWARD made by LLMs since ChatGPT came on the market, they're not seeing the forest through the trees because their heads are buried in the sand.

                                                                                                                                                                              LLMs for coding are not even close to imperfect, yet, but the saturation curves are not flattening out; not by a long shot. We are living in a moment and we need to come to terms with it as the work continues to develop; and, we need to adapt and quickly in order to better understand what our place will become as this nascent tech continues its meteoric trajectory toward an entirely new world.

                                                                                                                                                                              • eikenberry 1 day ago
                                                                                                                                                                                I don't think it is only (or even mostly) not wanting to accept it, I think it is at least equal measure just plain skepticism. We've seen all sorts of wild statements about how much something is going to revolutionize X and then turns out to be nothing. Most people disbelieve these sorts of claims until they see real evidence for themselves... and that is a good default position.
                                                                                                                                                                                • 0points 1 day ago
                                                                                                                                                                                  > If technology folk cannot see the INCREDIBLE LEAP FORWARD made by LLMs since ChatGPT came on the market, they're not seeing the forest through the trees because their heads are buried in the sand.

                                                                                                                                                                                  Look, we see the forest. We are just not impressed by it.

                                                                                                                                                                                  Having unlimited chaos monkeys at will is not revolutionizing anything.

                                                                                                                                                                                  • const_cast 1 day ago
                                                                                                                                                                                    Lawyers don't even use version control software a lot of the time. They burn hundreds of paralegal hours reconciling revisions, a task that could be made 100x faster and easier with Git.

                                                                                                                                                                                    There's no guarantee a technology will take off, even if it's really, really good. Because we don't decide if that tech takes off - the lawyers do. And they might not care, or they might decide billing more hours is better, actually.

                                                                                                                                                                                    • v3xro 16 hours ago
                                                                                                                                                                                      Many of us would prefer to see the technological leaps to be evenly distributed (so e.g. even clean drinking water that does not need to be boiled before consumption is not a baseline in 2025). So if you want to adapt to your new and improved position where you are just pushing buttons fine - but some of us are actually interested in how computers work (and are actually really uninterested in most companies' bottom lines). It's just how it is ;)
                                                                                                                                                                                      • rxtexit 21 hours ago
                                                                                                                                                                                        I think many people just settled in while we had no real technological change for 15 years. Real change, not an update to a web framework.

                                                                                                                                                                                        When I graduated high school, I had never been or knew anyone who had ever been on the internet at all. The internet was this vague "information superhighway" that I didn't know really what to make of.

                                                                                                                                                                                        If you are of a certain age though you would think a pointless update to react was all the change ever coming.

                                                                                                                                                                                        That time is over and we are back to reality.

                                                                                                                                                                                        • otabdeveloper4 19 hours ago
                                                                                                                                                                                          > you are ... le LUDDITE

                                                                                                                                                                                          Or maybe they just know the nitty-gritty inherent limitations of technology better than you.

                                                                                                                                                                                          (inb4: "LLMs can't have limitations! Wait a few years and they will solve literally every possible problem!")

                                                                                                                                                                                          • ben-schaaf 1 day ago
                                                                                                                                                                                            Friendly reminder that people like you were saying the exact same thing about metaverse, VR, web3, crypto, etc.
                                                                                                                                                                                            • fumeux_fume 1 day ago
                                                                                                                                                                                              Ah yes, please enjoy living in your moment and anticipating your entirely new world. I also hear all cars will be driving themselves soon and Jesus is coming back any day now.
                                                                                                                                                                                              • bgwalter 1 day ago
                                                                                                                                                                                                Adapt to your manager at bigcorp who is hyping the tech because it gives him something to do? No open source project is using the useless LLM shackles.
                                                                                                                                                                                              • spamizbad 1 day ago
                                                                                                                                                                                                I think it's pretty reasonable to take a CEO's - any CEO in any industry - statements with a grain of salt. They are under tremendous pressure to paint the most rosy picture possible of their future. They actually need you to "believe" just as much as their team needs to deliver.
                                                                                                                                                                                                • rxtexit 22 hours ago
                                                                                                                                                                                                  IMO it is a mixture of stupidity and denial.

                                                                                                                                                                                                  I am not a software engineer but I just can't imagine my job is not automated in 10 years or less.

                                                                                                                                                                                                  10 years is about the time between King – Man + Woman = Queen and now.

                                                                                                                                                                                                  I think what is being highly underestimated is the false sense of security people feel because the jobs they interface with are also not automated, yet.

                                                                                                                                                                                                  It is not hard to picture the network of automation that once one role is automated, connected roles to that role become easier to automate. So on and so on while the models keep getting stronger at the same time.

                                                                                                                                                                                                  I expect we will have a recession at some point and the jobs lost are gone forever.

                                                                                                                                                                                                  • onion2k 22 hours ago
                                                                                                                                                                                                    In most professional industries getting to the right answer is only half the problem. You also need to be able to demonstrate why that is the right answer. Your answer has to stand up to criticism. If your answer is essentially the output of a very clever random number generator you can't ever do that. Even if an LLM could output an absolutely perfect legal argument that matched what a supreme court judge would argue every time, that still wouldn't be good enough. You'd still need a person there to be accountable for making the argument and to defend the argument.

                                                                                                                                                                                                    Software isn't like this. No one cares why you wrote the code in your PR. They only care about whether it's right.

                                                                                                                                                                                                    This is why LLMs could be useful in one industry and a lot less useful in another.

                                                                                                                                                                                                    • sanderjd 1 day ago
                                                                                                                                                                                                      Isn't this also kind of just ... a reddit thing?
                                                                                                                                                                                                      • golergka 1 day ago
                                                                                                                                                                                                        Lawyers say those things and then one law firm after another is frantically looking for a contractor to overpay them to install local RAG and chatbot combo.
                                                                                                                                                                                                        • geros 1 day ago
                                                                                                                                                                                                          "It is difficult to get a man to understand something when his salary depends upon his not understanding it." - Upton Sinclair
                                                                                                                                                                                                        • layer8 1 day ago
                                                                                                                                                                                                          Programmers derided programming languages (too inefficient, too inflexible, too dumbing-down) when assembly was still the default. That phenomenon is at the same time entirely to be expected but also says little about the actual qualities of the new technology.
                                                                                                                                                                                                          • skydhash 20 hours ago
                                                                                                                                                                                                            If you have something that generated 20 lines of assembly that takes 100x times more than the 2 lines of clever instructions you know, you'd have the same stance even if the higher level was easier to use. Then those kind of performance tricks ceases to matter. But reliability still do. And the reasons we use higher and higher level of programming languages is because they increase reliability and simplicity (at the cost of performance, but we're happy to pay those).

                                                                                                                                                                                                            LLMs output are unreliable and productivity is still not proven for an end to end engineering cycle.

                                                                                                                                                                                                            • layer8 19 hours ago
                                                                                                                                                                                                              You are arguing based on the merits of the technology, which is fine, but wasn't my point. I was arguing that derision tends to happen no matter what, and thus doesn't indicate much about the merits of the technology one way or the other.
                                                                                                                                                                                                          • ch4s3 1 day ago
                                                                                                                                                                                                            It seems like LLMs made really big strides for a while but don't seem to be getting better recently, and in some ways recent models feel a bit worse. I'm seeing some good results generating test code, and some really bad results when people go to far with LLM use on new feature work. Base on what I've seen it seems like spinning up new projects and very basic features for web apps works really well, but that doesn't seem to generalize to refactoring or adding new features to big/old code bases.

                                                                                                                                                                                                            I've seen Claude and ChatGPT happily hallucinate whole APIs for D3 on multiple occasions, which should be really well represented in the training sets.

                                                                                                                                                                                                            • soerxpso 1 day ago
                                                                                                                                                                                                              > hallucinate whole APIs for D3 on multiple occasions, which should be really well represented in the training sets

                                                                                                                                                                                                              With many existing systems, you can pull documentation into context pretty quickly to prevent the hallucination of APIs. In the near future it's obvious how that could be done automatically. I put my engine on the ground, ran it and it didn't even go anywhere; Ford will never beat horses.

                                                                                                                                                                                                              • prisenco 1 day ago
                                                                                                                                                                                                                It's true that manually constraining an LLM with contextual data increases their performance on that data (and reduces performance elsewhere), but that conflicts with the promise of AI as an everything machine. We were promised an everything machine but if we have to not only provide it the proper context, but already know what constitutes the proper context, then it is not in any way an everything machine.

                                                                                                                                                                                                                Which means it's back to being a very useful tool, but not the earth-shattering disruptor we hoped (or worried) it would be.

                                                                                                                                                                                                              • oconnor663 1 day ago
                                                                                                                                                                                                                > don't seem to be getting better recently

                                                                                                                                                                                                                o3 came out just one month ago. Have you been using it? Subjectively, the gap between o3 and everything before it feels like the biggest gap I've seen since ChatGPT originally came out.

                                                                                                                                                                                                                • ch4s3 16 hours ago
                                                                                                                                                                                                                  I haven't used it extensively, but toyed around with it for Elixir code and I wasn't particularly impressed.
                                                                                                                                                                                                                • empath75 1 day ago
                                                                                                                                                                                                                  the LLM's themselves are making marginal gains, but the tools for using LLMs productively are getting so much better.
                                                                                                                                                                                                                  • dinfinity 1 day ago
                                                                                                                                                                                                                    This. MCP/tool usage in agentic mode is insanely powerful. Let the agent ingest a Gitlab issue, tell it how it can run commands, tests etc. in the local environment and half of the time it can just iterate towards a solution all by itself (but watching and intervening when it starts going the wrong way is still advisable).

                                                                                                                                                                                                                    Recently I converted all the (Google Docs) documentation of a project to markdown files and added those to the workspace. It now indexes it with RAG and can easily find relevant bits of documentation, especially in agent mode.

                                                                                                                                                                                                                    It really stresses the importance of getting your documentation and processes in order as well as making sure the tasks at hand are well-specified. It soon might be the main thing that requires human input or action.

                                                                                                                                                                                                                • cushychicken 1 day ago
                                                                                                                                                                                                                  ChatGPT-4o is scary good at writing VHDL.

                                                                                                                                                                                                                  Using it to prototype some low level controllers today, as a matter of fact!

                                                                                                                                                                                                                  • UncleOxidant 1 day ago
                                                                                                                                                                                                                    Claude and Gemini are decent at it as well. I was surprised when I asked claude (and this was several months back) to come up with a testbench for some very old, poorly documented verilog. It did a very decent job for a first-cut testbench. It even collected common, recurring code into verilog tasks (functions) which really surprised me at the time.
                                                                                                                                                                                                                    • cushychicken 22 hours ago
                                                                                                                                                                                                                      Yes! It’s much better at using functional logic than I am - which I appreciate!
                                                                                                                                                                                                                    • Panzer04 1 day ago
                                                                                                                                                                                                                      What kind of things is it doing?

                                                                                                                                                                                                                      I have a hard time imagining an LLM being able to do arbitrary things. It always feels like LLMs can do lots of the easy stuff, but if they can't do everything you need the skilled engineer anyway, who'd knock the easy things out in a week anyway.

                                                                                                                                                                                                                      • cushychicken 21 hours ago
                                                                                                                                                                                                                        What kind of things is it doing?

                                                                                                                                                                                                                        Wrote me:

                                                                                                                                                                                                                        - a SPI deserializer that sets a bit after 12 bits read in, to trigger a prefetch

                                                                                                                                                                                                                        - an SDC constraints file for the deserializer that correctly identified the SPI clock and bus clock as separate domains requiring their own statement

                                                                                                                                                                                                                        - a test bench that validated both that the prefetch bit was being set, and that it was being set at the proper time relative to the SPI clock

                                                                                                                                                                                                                        - a makefile with commands for build, headless test, and debug by loading the VCD into a waveform viewer

                                                                                                                                                                                                                        It always feels like LLMs can do lots of the easy stuff, but if they can't do everything you need the skilled engineer anyway, who'd knock the easy things out in a week anyway.

                                                                                                                                                                                                                        Nearly every part of the tool flow I just described, I would consider “tricky to get right”. Been doing this for ~15 years and it’s still tough to bootstrap something like this from scratch. ChatGPT-4o did this for me from zero in about 15 minutes.

                                                                                                                                                                                                                        I won’t lie: I love it. I can focus on the actual, bigger problems at hand, and not the tricky little details of HDLs.

                                                                                                                                                                                                                        People are either deluding themselves or ignorant of the capabilities of frontier models if they don’t believe LLMs offer a speedup in workflow.

                                                                                                                                                                                                                        I personally believe that most of the doubt and cynicism is due to:

                                                                                                                                                                                                                        1) a pretty big collective identity crisis among software professionals, and

                                                                                                                                                                                                                        2) a suspicion that LLMs make it so that anyone who is good at articulating the problem precisely no longer needs a software engineer as a translation specialist from specs to code.

                                                                                                                                                                                                                        I say this as an EE of ~15 years who’s always been able to articulate what I want, specifically, to a firmware counterpart, who then writes the code I need. I can turn years of practice in this skill into great prompts for an LLM, which effectively cuts out the middleman.

                                                                                                                                                                                                                        I really like it. It’s helped me take on a lot of projects that are just outside of my innate level of capability. It’s also helped me learn a lot of new things about these new software adjacent areas. ChatGPT is a great tutor!

                                                                                                                                                                                                                      • roflyear 1 day ago
                                                                                                                                                                                                                        It's better-than-senior at a some things, but worse-than-junior at a lot of things.
                                                                                                                                                                                                                        • quantadev 1 day ago
                                                                                                                                                                                                                          It's more like better-than-senior 99% of the time. Makes mistakes 1% of the time. Most of the 'bad results' I've seen people struggle with ended up being the fault of the human, in the form of horrible context given to the AI or else ambiguous or otherwise flawed prompts.

                                                                                                                                                                                                                          Any skilled developer with a decade of experience can write prompts that return back precisely what we wanted almost every single time. I do it all day long. "Claude 4" rarely messes up.

                                                                                                                                                                                                                      • parliament32 1 day ago
                                                                                                                                                                                                                        I'd like to agree with you and remain optimistic, but so much tech has promised the moon and stagnated into oblivion that I just don't have any optimism left to give. I don't know if you're old enough, but remember when speech-to-text was the next big thing? DragonSpeak was released in 1997, everyone was losing their minds about dictating letters/documents in MS Word, and we were promised that THIS would be the key interface for computing evermore. And.. 27 years later, talking to the latest Siri, it makes just as many mistakes as it did back then. In messenger applications people are sending literal voice notes -- audio clips -- back and forth because dictation is so unreliable. And audio clips are possibly the worst interface for communication ever (no searching, etc).

                                                                                                                                                                                                                        Remember how blockchain was going to change the world? Web3? IoT? Etc etc.

                                                                                                                                                                                                                        I've been through enough of these cycles to understand that, while the AI gimmick is cool and all, we're probably at the local maximum. The reliability won't improve much from here (hallucinations etc), while the costs to run it will stay high. The final tombstone will be when the AI companies stop running at a loss and actually charge for the massive costs associated with running these models.

                                                                                                                                                                                                                        • oconnor663 1 day ago
                                                                                                                                                                                                                          > 27 years later, talking to the latest Siri, it makes just as many mistakes as it did back then

                                                                                                                                                                                                                          Have you tried talking to ChatGPT voice mode? It's mind blowing. You just have a conversation with it. In any language. About anything. The other day I wanted to know about the difference between cast iron and wrought iron, and it turned into a 10 or 15 minute conversation. That's maybe a good example of an "easy" topic for LLMs (lots of textbooks for it to memorize), but the world is full of easy topics that I know nothing about!

                                                                                                                                                                                                                          • some_random 1 day ago
                                                                                                                                                                                                                            How can you possibly look at what LLMs are doing and the progress made in the last ~3 years and equate it to crypto bullshit? Also it's super weird to include IoT in there, seeing as it has become all but ubiquitous.
                                                                                                                                                                                                                            • r14c 1 day ago
                                                                                                                                                                                                                              I'm not as bearish on AI, but its hard to tell if you can really extrapolate future performance based on past improvements.

                                                                                                                                                                                                                              Personally, I'm more interested in the political angle. I can see that AI will be disruptive because there's a ton of money and possibly other political outcomes depending on it doing exactly that.

                                                                                                                                                                                                                          • bgwalter 1 day ago
                                                                                                                                                                                                                            Yet you are working on your own replacement, while your colleagues are taking the prudent approach.
                                                                                                                                                                                                                            • Jolter 1 day ago
                                                                                                                                                                                                                              Here’s the deal: if you won’t write your replacement, a competitor will do it and outprice your employer. Either way you’re out of a job. May be more prudent to adapt to the new tools and master them rather than be left behind?

                                                                                                                                                                                                                              Do you want to be a jobless weaver, or an engineer building mechanical looms for a higher pay than the weaver got?

                                                                                                                                                                                                                              • bgwalter 1 day ago
                                                                                                                                                                                                                                I want to be neither. I either want to continue being a software engineer who doesn't need a tricycle for the mind, or move to law or medicine; two professions that have successfully defended themselves against extreme versions of the kind of anxiety, obedience and self hate that is so prevalent among software engineers.
                                                                                                                                                                                                                                • allturtles 1 day ago
                                                                                                                                                                                                                                  I think it's the wrong analogy. The prompt engineer who uses the AI to make code maps to the poorly-paid, low-skill power loom machine tender. The "engineer" is the person who created the model. But it's also not totally clear to me that we'll need humans for that either, in the near future.
                                                                                                                                                                                                                                  • declan_roberts 1 day ago
                                                                                                                                                                                                                                    I would absolutely love to write my own placement. When I can have AI do my job while I go to the beach you better believe I will be at the beach.
                                                                                                                                                                                                                                    • kuahyeow 1 day ago
                                                                                                                                                                                                                                      Compiler is the more apt analogy to a mechanical loom.

                                                                                                                                                                                                                                      An LLM is more like outsourcing to a consultancy. Results may vary.

                                                                                                                                                                                                                                      • mullingitover 1 day ago
                                                                                                                                                                                                                                        > Either way you’re out of a job.

                                                                                                                                                                                                                                        Tools and systems which increase productivity famously always put everyone out of a job, which is why after a couple centuries of industrial revolution we're all unemployed.

                                                                                                                                                                                                                                      • nialse 1 day ago
                                                                                                                                                                                                                                        Ahh, the “don’t disturb the status quo” argument. See, we are all working on our replacement, newer versions, products, services and knowledge always make the older obsolete. It is wise to work on your replacement, and even wiser to be in charge of and operate the replacement.
                                                                                                                                                                                                                                        • bgwalter 1 day ago
                                                                                                                                                                                                                                          No, nothing fundamentally new is created. Programmers have always been obsessed with "new" tooling and processes to distract from that fact.

                                                                                                                                                                                                                                          "AI" is the latest iteration of snake oil that is foisted upon us by management. The problem is not "AI" per se, but the amount of of friction and productivity loss that comes with it.

                                                                                                                                                                                                                                          Most of the productivity loss comes from being forced to engage with it and push back against that nonsense. One has to learn the hype language, debunk it, etc.

                                                                                                                                                                                                                                          Why do you think IT has gotten better? Amazon had a better and faster website with far better search and products 20 years ago. No amount of "AI" will fix that.

                                                                                                                                                                                                                                          • palmotea 1 day ago
                                                                                                                                                                                                                                            > It is wise to work on your replacement...

                                                                                                                                                                                                                                            Depends on the context. You have to keep in mind: it is not a goal of our society or economic system to provide you with a stable, rewarding job. In fact, the incentives are to take that away from you ASAP.

                                                                                                                                                                                                                                            Before software engineers go celebrate this tech, they need realize they're going to end up like rust-belt factory workers the day after the plant closed. They're not special, and society won't be any kinder to them.

                                                                                                                                                                                                                                            > ...and even wiser to be in charge of and operate the replacement.

                                                                                                                                                                                                                                            You'll likely only get to do that if your boss doesn't know about it.

                                                                                                                                                                                                                                          • npteljes 1 day ago
                                                                                                                                                                                                                                            Carteling doesn't work bottom-up. When changes begin (like this one with AI), one of the things an individual can do is to change course as fast as they can. There are other strategies as well, not evolving is also one, but some strategies yield better results than others. Not keeping up just worsens the chances, I have found.
                                                                                                                                                                                                                                            • asdff 1 day ago
                                                                                                                                                                                                                                              It does when it is called unionizing, however for some reason software developers have a mental block towards the concept.
                                                                                                                                                                                                                                            • dughnut 1 day ago
                                                                                                                                                                                                                                              Do you want to work with LLMs or H1Bs and interns… choose wisely.

                                                                                                                                                                                                                                              Personally I’m thrilled that I can get trivial, one-off programs developed for a few cents and the cost of a clear written description of the problem. Engaging internal developers or consulting developers to do anything at all is a horrible experience. I would waste weeks on politics, get no guarantees, and waste thousands of dollars and still hear nonsense like, “you want a form input added to a web page? Aw shucks, that’s going to take at least another month” or “we expect to spend a few days a month maintaining a completely static code base” from some clown billing me $200/hr.

                                                                                                                                                                                                                                              • rsyring 1 day ago
                                                                                                                                                                                                                                                You can work with consulting oriented engineers who get shit done with relatively little stress and significant productivity. Productivity enhanced by AI but not replaced by it. If interested, reach out to me.
                                                                                                                                                                                                                                              • JeremyNT 1 day ago
                                                                                                                                                                                                                                                I don't think that this should be downvoted because it raises a really important issue.

                                                                                                                                                                                                                                                I hate AI code assistants, not because they suck, but because they work. The writing is on the wall.

                                                                                                                                                                                                                                                If we aren't working on our own replacements, we'll be the ones replaced by somebody else's vibe code, and we have no labor unions that could plausibly fight back against this.

                                                                                                                                                                                                                                                So become a Vibe Coder and keep working, or take the "prudent" approach you mention - and become unemployed.

                                                                                                                                                                                                                                                • neta1337 1 day ago
                                                                                                                                                                                                                                                  I’ll work on fixing the vibe coders mess and make bank. Experience will prove valuable even more than before
                                                                                                                                                                                                                                                  • realusername 1 day ago
                                                                                                                                                                                                                                                    Personally I used them for a while and then just stopped using them because actually no, unfortunately those assistants don't work. They appear to work at first glance but there's so much babysitting needed that it's just not worth it.

                                                                                                                                                                                                                                                    This "vibe coding" seems just another way to say that people spend more time refining the output of these tools over and over again that what they would normally code.

                                                                                                                                                                                                                                                • energy123 1 day ago
                                                                                                                                                                                                                                                  Their confusion is your competitive advantage in the labor market.
                                                                                                                                                                                                                                                  • 1 day ago
                                                                                                                                                                                                                                                    • retetr 1 day ago
                                                                                                                                                                                                                                                      Unrelated, but is this a case of the Pareto Principle? (Admittedly the first time I'm hearing of it) Wherein 80% of the effect is caused by 20% of the input. Or is this more a case of diminishing returns? Where the initial results were incredible, but each succeeding iteration seems to be more disappointing?
                                                                                                                                                                                                                                                      • klabb3 1 day ago
                                                                                                                                                                                                                                                        Pareto is about diminishing returns.

                                                                                                                                                                                                                                                        > but each succeeding iteration seems to be more disappointing

                                                                                                                                                                                                                                                        This is because the scaling hypothesis (more data and more compute = gains) is plateauing, because all text data is used and compute is reaching diminishing returns for some reason I’m not smart enough to say why, but it is.

                                                                                                                                                                                                                                                        So now we're seeing incremental core model advancements, variations and tuning in pre- and post training stages and a ton of applications (agents).

                                                                                                                                                                                                                                                        This is good imo. But obviously it’s not good for delusional valuations based exponential growth.

                                                                                                                                                                                                                                                        • energy123 1 day ago
                                                                                                                                                                                                                                                          We're seeing diminishing returns in benchmark space, which is partly an artefact of construction, not an absolutely true commentary on how things are progressing.
                                                                                                                                                                                                                                                    • ChrisMarshallNY 1 day ago
                                                                                                                                                                                                                                                      Really good coders (like him) are better.

                                                                                                                                                                                                                                                      Mediocre ones … maybe not so much.

                                                                                                                                                                                                                                                      When I worked for a Japanese optical company, we had a Japanese engineer, who was a whiz. I remember him coming over from Japan, and fixing some really hairy communication bus issues. He actually quit the company, a bit after that, at a very young age, and was hired back as a contractor; which was unheard of, in those days.

                                                                                                                                                                                                                                                      He was still working for them, as a remote contractor, at least 25 years later. He was always on the “tiger teams.”

                                                                                                                                                                                                                                                      He did awesome assembly. I remember when the PowerPC came out, and “Assembly Considered Harmful,” was the conventional wisdom, because of pipelining, out-of-order instructions, and precaching, and all that.

                                                                                                                                                                                                                                                      His assembly consistently blew the doors off anything the compiler did. Like, by orders of magnitude.

                                                                                                                                                                                                                                                      • benstein 1 day ago
                                                                                                                                                                                                                                                        +1000. "Human coders are still better than LLMs" is a hot take. "Antirez is still better than LLMs" is axiomatic ;-)
                                                                                                                                                                                                                                                      • yua_mikami 1 day ago
                                                                                                                                                                                                                                                        The thing everyone forgets when talking about LLMs replacing coders is that there is much more to software engineering than writing code, in fact that's probably one of the smaller aspects of the job.

                                                                                                                                                                                                                                                        One major aspect of software engineering is social, requirements analysis and figuring out what the customer actually wants, they often don't know.

                                                                                                                                                                                                                                                        If a human engineer struggles to figure out what a customer wants and a customer struggles to specify it, how can an LLM be expected to?

                                                                                                                                                                                                                                                        • malfist 1 day ago
                                                                                                                                                                                                                                                          That was also one of the challenges during the offshoring craze in the 00s. The offshore teams did not have the power, or knowledge to push back on things and just built and built and built. Sounds very similar to AI right?

                                                                                                                                                                                                                                                          Probably going to have the same outcome.

                                                                                                                                                                                                                                                          • pandastronaut 1 day ago
                                                                                                                                                                                                                                                            I tend to see today's AI Vibrators as the managers of the 00s and their army of offshore devs.
                                                                                                                                                                                                                                                            • 9dev 1 day ago
                                                                                                                                                                                                                                                              Did you actually mean to say AI Vibrators?
                                                                                                                                                                                                                                                            • hathawsh 1 day ago
                                                                                                                                                                                                                                                              The difference is that when AI exhibits behavior like that, you can refine the AI or add more AI layers to correct it. For example, you might create a supervisor AI that evaluates when more requirements are needed before continuing to build, and a code review AI that triggers refinements automatically.
                                                                                                                                                                                                                                                              • nevertoolate 1 day ago
                                                                                                                                                                                                                                                                Question is, how autonomous decision making works, nobody argues that llm can finish any sentence, but can it push a red button?
                                                                                                                                                                                                                                                            • devjab 1 day ago
                                                                                                                                                                                                                                                              LLM's do no software engineering at all, and that can be fine. Because you don't actually need software engineering to create successful programs. Some applications will not even need software engineering for their entire life cycles because nobody is really paying attention to efficiency in the ocean of poor cloud management anyway.

                                                                                                                                                                                                                                                              I actually imagine it's the opposite of what you say here. I think technically inclined "IT business partners" will be able of creating applications entirely without software engineers... Because I see that happen every day in the world of green energy. The issues come later, when things have to be maintained, scale or become efficient. This is where the software engineering comes in, because it actually matters if you used a list or a generator in your Python app when it iterates over millions of items and not just a few hundreds.

                                                                                                                                                                                                                                                              • AstroBen 1 day ago
                                                                                                                                                                                                                                                                That's the thing too right.. the vast majority of software out there barely needs to scale or be super efficient

                                                                                                                                                                                                                                                                It does need to be reliable, though. LLMs have proven very bad at that

                                                                                                                                                                                                                                                                • km144 16 hours ago
                                                                                                                                                                                                                                                                  Maybe automated testing of all forms will just become much more ubiquitous as a safeguard against the worst of AI hallucinations? I feel that would solve a lot of people's worries about LLMs. I'm imagining a world where a software developer is a person who gathers requirements, writes some tests, asks the AI to modify the codebase, ensures the tests still work, makes sure they are a human who understands the change the AI just made, and continues with the next requirement.
                                                                                                                                                                                                                                                                  • devjab 23 hours ago
                                                                                                                                                                                                                                                                    > the vast majority of software out there barely needs to scale or be super efficient

                                                                                                                                                                                                                                                                    That was the way I saw it for a while. In recent months I've begun to wonder if I need to reevaluate that, because it's become clear to me that scaling doesn't actually start from zero. By zero I mean that I was naive enough to think that all programs, even the most googled programmed one by a completely new junior would at least have, some, efficiency... but some of these LLM services I get to work on today are so bad they didn't start at zero but at some negative number. It would have been less of an issue if our non-developer-developers didn't use Python (or at least used Python with ruff/pyrefly/whateveryoulike, but some of the things they write can't even scale to do minimal BI reporting.

                                                                                                                                                                                                                                                                • victorbjorklund 1 day ago
                                                                                                                                                                                                                                                                  Yea, this is why I dont buy the "all developers will disappear". Will I write a lot less code in 5 years (maybe almost none)? Sure, I already type a lot less now than a year ago. But that is just a small part of the process.
                                                                                                                                                                                                                                                                  • elzbardico 1 day ago
                                                                                                                                                                                                                                                                    No. the scope will just increase to occupy the space left by LLMs. We will never be allowed to retire.
                                                                                                                                                                                                                                                                    • xandrius 1 day ago
                                                                                                                                                                                                                                                                      Exactly, also today I can actually believe I could finish a game which might have taken much longer before LLMs, just because now I can be pretty sure I won't get stuck on some feature just because I never done it before.
                                                                                                                                                                                                                                                                    • ilaksh 1 day ago
                                                                                                                                                                                                                                                                      It actually comes down to feedback loops which means iterating on software being used or attempting to be used by the customer.

                                                                                                                                                                                                                                                                      Chat UIs are an excellent customer feedback loop. Agents develop new iterations very quickly.

                                                                                                                                                                                                                                                                      LLMs can absolutely handle abstractions and different kinds of component systems and overall architecture design.

                                                                                                                                                                                                                                                                      They can also handle requirements analysis. But it comes back to iteration for the bottom line which means fast turnaround time for changes.

                                                                                                                                                                                                                                                                      The robustness and IQ of the models continue to be improved. All of software engineering is well underway of being automated.

                                                                                                                                                                                                                                                                      Probably five years max where un-augmented humans are still generally relevant for most work. You are going to need deep integration of AI into your own cognition somehow in order to avoid just being a bottleneck.

                                                                                                                                                                                                                                                                      • mettamage 1 day ago
                                                                                                                                                                                                                                                                        > One major aspect of software engineering is social, requirements analysis and figuring out what the customer actually wants, they often don't know.

                                                                                                                                                                                                                                                                        It really depends on the organization. In many places product owners and product managers do this nowadays.

                                                                                                                                                                                                                                                                        • bbarn 1 day ago
                                                                                                                                                                                                                                                                          The thing is, it is replacing _coders_ in a way. There are millions of people who do (or did) the work that LLMs excel at. Coders who are given a ticket that says "Write this API taking this input and giving this output" who are so far down the chain they don't even get involved in things like requirements analysis, or even interact with customers.

                                                                                                                                                                                                                                                                          Software engineering, is a different thing, and I agree you're right (for now at least) about that, but don't underestimate the sheer amount of brainless coders out there.

                                                                                                                                                                                                                                                                          • callc 1 day ago
                                                                                                                                                                                                                                                                            That sounds more like a case against a highly ossified waterfall development process than anything.

                                                                                                                                                                                                                                                                            I would argue it’s a good thing to replace the actual brainless activities.

                                                                                                                                                                                                                                                                          • rowanG077 1 day ago
                                                                                                                                                                                                                                                                            I think LLMs are better at requirement elicitation than they are at actually writing code.
                                                                                                                                                                                                                                                                            • ori_b 1 day ago
                                                                                                                                                                                                                                                                              > If a human engineer struggles to figure out what a customer wants and a customer struggles to specify it, how can an LLM be expected to?

                                                                                                                                                                                                                                                                              Presumably, they're trained on a ton of requirements docs, as well as a huge number of customer support conversations. I'd expect them to do this at least as well as coding, and probably better.

                                                                                                                                                                                                                                                                            • wanderingstan 1 day ago
                                                                                                                                                                                                                                                                              “Better” is always task-dependent. LLMs are already far better than me (and most devs I’d imagine) at rote things like getting CSS syntax right for a desired effect, or remembering the right way to invoke a popular library (e.g. fetch)

                                                                                                                                                                                                                                                                              These little side quests used to eat a lot of my time and I’m happy to have a tool that can do these almost instantly.

                                                                                                                                                                                                                                                                              • jaccola 1 day ago
                                                                                                                                                                                                                                                                                I've found LLMs particularly bad for anything beyond basic styling since the effects can be quite hard to describe and/or don't have a universal description.

                                                                                                                                                                                                                                                                                Also, there are often times multiple ways to achieve a certain style and they all work fine until you want a particular tweak, in which case only one will work and the LLM usually gets stuck in one of the ones that does not work.

                                                                                                                                                                                                                                                                                • danielbln 1 day ago
                                                                                                                                                                                                                                                                                  Multi modal LLMs to the rescue. Throw a screenshot or mockup in there and tell the LLM "there, like this". Gemini can do the same with videos.
                                                                                                                                                                                                                                                                                  • karn97 1 day ago
                                                                                                                                                                                                                                                                                    Still terrible result. Multi modal = actually understands the image
                                                                                                                                                                                                                                                                                • gherkinnn 1 day ago
                                                                                                                                                                                                                                                                                  I have found it to be good at things I am not very strong at (SQL) but terrible at the things I know well (CSS).

                                                                                                                                                                                                                                                                                  Telling, isn't it?

                                                                                                                                                                                                                                                                                  • mywittyname 1 day ago
                                                                                                                                                                                                                                                                                    Ironically, I find it strong at things I don't know very well (CSS), but terrible at things I know well (SQL).

                                                                                                                                                                                                                                                                                    This is probably really just a way of saying, it's better at simple tasks rather than complex ones. I can eventually get Copilot to write SQL that's complex and accurate, but I don't find it faster or more effective than writing it myself.

                                                                                                                                                                                                                                                                                    • ehansdais 1 day ago
                                                                                                                                                                                                                                                                                      Actually, you've reinforced their point. It's only bad at things the user is actually good at because the user actually knows enough in that domain to find the flaws and issues. It appears to be good in domains the user is bad at because the user doesn't know any better. In reality, the LLM is just bad at all domains; it's simply whether a user has the skill to discern it. Of course, I don't believe it's as black and white as that but I just wanted to point it out.
                                                                                                                                                                                                                                                                                    • ch4s3 1 day ago
                                                                                                                                                                                                                                                                                      I kind of agree. It feels like they're generally a superior form of copying and pasting fro stack overflow where the machine has automated the searching, copying, pasting, and fiddling with variable names. It be just as useful or dangerous as Google -> Copy -> Paste ever was, but faster.
                                                                                                                                                                                                                                                                                      • sanderjd 1 day ago
                                                                                                                                                                                                                                                                                        Funny, I find it to be good at things I'm not very strong at (CSS) but terrible at the things I know well (SQL). :)

                                                                                                                                                                                                                                                                                        Actually I think it's perfectly adequate at SQL too.

                                                                                                                                                                                                                                                                                      • kccqzy 1 day ago
                                                                                                                                                                                                                                                                                        > and most devs I’d imagine

                                                                                                                                                                                                                                                                                        What an awful imagination. Yes there are people who don't like CSS but are forced to use it by their job so they don't learn it properly, and that's why they think CSS is rote memorization.

                                                                                                                                                                                                                                                                                        But overall I agree with you that if a company is too cheap to hire a person who is actually skilled at CSS, it is still better to hoist that CSS job onto LLMs than an unwilling human. Because that unwilling human is not going to learn CSS well and won't enjoy writing CSS.

                                                                                                                                                                                                                                                                                        On the other hand, if the company is willing to hire someone who's actually good, LLMs can't compare. It's basically the old argument of LLMs only being able to replace less good developers. In this case, you admitted that you are not good at CSS and LLMs are better than you at CSS. It's not task-dependent it's skill-dependent.

                                                                                                                                                                                                                                                                                        • marcosdumay 1 day ago
                                                                                                                                                                                                                                                                                          Hum... I imagine LLMs are better than every developer on getting CSS keywords right like the GP pointed. And I expect every LLM to be slightly worse than most classical autocompletes.
                                                                                                                                                                                                                                                                                          • skydhash 1 day ago
                                                                                                                                                                                                                                                                                            Getting CSS keywords right is not the actual point of writing CSS. And you can have a linter that helps you in that regards. The endgame of writing CSS is to style an HTML page according to the specifications of a design. Which can be as detailed as a figma file or as flimsy as a drawing on a whiteboard.
                                                                                                                                                                                                                                                                                            • lelandfe 1 day ago
                                                                                                                                                                                                                                                                                              I'm one of those weirdos who really likes handwriting CSS. I frequently find ChatGPT getting my requests wrong.
                                                                                                                                                                                                                                                                                              • michaelsalim 1 day ago
                                                                                                                                                                                                                                                                                                This is like saying that LLMs are better at knowing the name of that one obscure API. It's not wrong, but it's also not the hard part about CSS
                                                                                                                                                                                                                                                                                              • chii 1 day ago
                                                                                                                                                                                                                                                                                                The LLM outputs good enough CSS, but is (way) cheaper than someone who's actually good at CSS.
                                                                                                                                                                                                                                                                                              • sanderjd 1 day ago
                                                                                                                                                                                                                                                                                                Yeah, this is what I really like about AI tools though. They're way better than me at annoying minutia like getting CSS syntax right. I used to dread that kind of thing!
                                                                                                                                                                                                                                                                                                • codr7 1 day ago
                                                                                                                                                                                                                                                                                                  And you will keep dreading it for as long as you use them, since you learn nothing from solutions served on a silver plate.
                                                                                                                                                                                                                                                                                                  • sanderjd 1 day ago
                                                                                                                                                                                                                                                                                                    The point is that I don't dread it anymore, because now there are tools that make it a lot easier the one or two times a year I have some reason to use it.
                                                                                                                                                                                                                                                                                                • zdragnar 1 day ago
                                                                                                                                                                                                                                                                                                  I think that's great if it's for something outside of your primary language. I've used it to good effect in that way myself. However, denying yourself the reflexive memory of having learned those things is a quick way to become wholly dependent upon the tool. You could easily end up with compromised solutions because the tool recommends something you don't understand well enough to know there's a better way to do something.
                                                                                                                                                                                                                                                                                                  • dpkirchner 1 day ago
                                                                                                                                                                                                                                                                                                    You're right, however I think we've already gone through this before. Most of us (probably) couldn't tell you exactly how an optimizing compiler picks optimizations or exactly how JavaScript maps to processor instructions, etc -- we hopefully understand enough at one level of abstraction to do our jobs. Maybe LLM driving will be another level of abstraction, when it gets better at (say) architecting projects.
                                                                                                                                                                                                                                                                                                    • skydhash 1 day ago
                                                                                                                                                                                                                                                                                                      > Most of us (probably) couldn't tell you exactly how an optimizing compiler picks optimizations or exactly how JavaScript maps to processor instructions,

                                                                                                                                                                                                                                                                                                      That's because other people are making those working well. It's like how you don't care about how the bread is being made because you trust your baker (or the regulations). It's a chain of trust that is easily broken when LLMs are brought in.

                                                                                                                                                                                                                                                                                                    • AnimalMuppet 1 day ago
                                                                                                                                                                                                                                                                                                      So here's an analogy. (Yeah, I know, proof by analogy is fraud. But it's going to illustrate the question.)

                                                                                                                                                                                                                                                                                                      Here's a kid out hoeing rows for corn. He sees someone planting with a tractor, and decides that's the way to go. Someone tells him, "If you get a tractor, you'll never develop the muscles that would make you really great at hoeing."

                                                                                                                                                                                                                                                                                                      Different analogy: Here's someone trying to learn to paint. They see someone painting by numbers, and it looks a lot easier. Someone tells them, "If you paint by numbers, you'll never develop the eye that you need to really become good as a painter."

                                                                                                                                                                                                                                                                                                      Which is the analogy that applies, and what makes it the right one?

                                                                                                                                                                                                                                                                                                      I think the difference is how much of the job the tool can take over. The tractor can take over the job of digging the row, with far more power, far more speed, and honestly far more quality. The paint by numbers can take over the job of visualizing the painting, with some loss of quality and a total loss of creativity. (In painting, the creativity is considered a vital part; in digging corn rows, not so much.)

                                                                                                                                                                                                                                                                                                      I think that software is more like painting, rather than row-hoeing. I think that AI (currently) is in the form of speeding things up with some loss of both quality and creativity.

                                                                                                                                                                                                                                                                                                      Can anyone steelman this?

                                                                                                                                                                                                                                                                                                      • bluefirebrand 1 day ago
                                                                                                                                                                                                                                                                                                        > Here's a kid out hoeing rows for corn. He sees someone planting with a tractor, and decides that's the way to go. Someone tells him, "If you get a tractor, you'll never develop the muscles that would make you really great at hoeing

                                                                                                                                                                                                                                                                                                        In this example the idea that losing the muscles that make you great at hoeing" seems kind of like a silly thing to worry about

                                                                                                                                                                                                                                                                                                        But I think there's a second order effect here. The kid gets a job driving the tractor instead. He spends his days seated instead of working. His lifestyle is more sedentary. He works just as many hours as before, and he makes about the same as he did before, so he doesn't really see much benefit from the increased productivity of the tractor.

                                                                                                                                                                                                                                                                                                        However now he's gaining weight from being more sedentary, losing muscle from not moving his body, developing lower back problems from being seated all day, developing hearing loss from the noisy machinery. His quality of life is now lower, right?

                                                                                                                                                                                                                                                                                                        Edit: Yes, there are also health problems from working hard moving dirt all day. You can overwork yourself, no question. It's hard on your body, being in the sun all day is bad for you.

                                                                                                                                                                                                                                                                                                        I would argue it's still objectively a physically healthier lifestyle than driving a tractor for hours though.

                                                                                                                                                                                                                                                                                                        Edit 2: my point is that I think after driving a tractor for a while, the kid would really struggle to go hoe by hand like he used to, if he ever needed to

                                                                                                                                                                                                                                                                                                        • Sohcahtoa82 16 hours ago
                                                                                                                                                                                                                                                                                                          The analogy I would use is that coding via LLM is like learning to drive in a self-driving car that has manual controls as an option that drives overly cautiously (Leaves excessively large following distances, takes corners slower, etc.) while in self-driving mode.

                                                                                                                                                                                                                                                                                                          You can let it self-drive, but you'd probably learn nothing, and it will actually take you longer. Put an expert driver behind the wheel, and they'll drive faster and only use automation features for the boring parts.

                                                                                                                                                                                                                                                                                                          • stonemetal12 1 day ago
                                                                                                                                                                                                                                                                                                            >I think the difference is how much of the job the tool can take over.

                                                                                                                                                                                                                                                                                                            I think it is about how utilitarian the output is. For food no one cares how the sausage is made. For a painting the story behind it is more important than the picture itself. All of Picasso's paintings are famous because they were painted by Picasso. Picasso style painting by Bill? Suddenly it isn't museum worthy anymore.

                                                                                                                                                                                                                                                                                                            No one cares about the story or people behind Word, they just want to edit documents. The Demo scene probably has a good shot at being on the side of art.

                                                                                                                                                                                                                                                                                                            • danielbln 1 day ago
                                                                                                                                                                                                                                                                                                              For me the creativity in software engineering doesn't come from coding, that's an implementation detail. It comes from architecture, from thinking about "what do I want to build, how should it behave, how should it look, what or who is it for?" and driving that forward. Bolting it together in code is hoeing, for that vast majority of us. The creative endeavor sits higher up on the abstraction ladder.
                                                                                                                                                                                                                                                                                                        • loudmax 1 day ago
                                                                                                                                                                                                                                                                                                          Companies that leverage LLMs and AIs to let their employees be more productive will thrive.

                                                                                                                                                                                                                                                                                                          Companies that try to replace their employees with LLMs and AIs will fail.

                                                                                                                                                                                                                                                                                                          Unfortunately, all that's in the long run. In the near term, some CEOs and management teams will profit from the short term valuations as they squander their companies' future growth on short-sighted staff cuts.

                                                                                                                                                                                                                                                                                                          • bdbenton5255 1 day ago
                                                                                                                                                                                                                                                                                                            That's really it. These tools are useful as assistants to programmers but do not replace an actual programmer. The right course is to embrace the technology moderately rather than reject it completely or bet on it replacing workers.
                                                                                                                                                                                                                                                                                                            • joshdavham 1 day ago
                                                                                                                                                                                                                                                                                                              > In the near term, some CEOs and management teams will profit from the short term valuations

                                                                                                                                                                                                                                                                                                              That's actually really interesting to think about. The idea that doing something counter-productive like trying to replace employees with AI (which will cause problems), may actually benefit the company in terms of valuations in the short run. So in effect, they're hurting and helping the company at the same time.

                                                                                                                                                                                                                                                                                                              • Natfan 18 hours ago
                                                                                                                                                                                                                                                                                                                See also: Toys 'R' Us, Seers, which were killed by consultancy groups loading on debt and selling assets for an immediate profit, which helped the immediate shareholders but hurt all of the stakeholders.
                                                                                                                                                                                                                                                                                                                • to11mtm 1 day ago
                                                                                                                                                                                                                                                                                                                  Hey if the check clears for the bonus they got for hitting 'reduce costs in the IT department', they often bail before things rear their ugly head, or in the ugly case, Reality Distortion Field's the entire org into making the bad anti patterns permanent, even while acknowledging the cost/delivery/quality inefficiencies[0].

                                                                                                                                                                                                                                                                                                                  This is especially prevalent in waterfall orgs that refuse change. Body shops are more than happy to waste a huge portion of their billable hours on planning meetings and roadmap revisions as the obviousness of the mythical man month comes to bear on the org.

                                                                                                                                                                                                                                                                                                                  Corners get cut to meet deadlines, because the people who started/perpetuated whatever myth need to save their skins (and hopefully continue to get bonuses.)

                                                                                                                                                                                                                                                                                                                  The engineers become a scapegoat for the org's management problems (And watch, it very likely will happen at some shops with the 'AI push'). In the nasty cases, the org actively disempowers engineers in the process[0][1].

                                                                                                                                                                                                                                                                                                                  [0] - At one shop, there was grief we got that we hadn't shipped a feature, but the only reason we hadn't, was IT was not allowed to decide between a set of radio buttons or a drop-down on a screen. Hell I got yelled at for just making the change locally and sending screenshots.

                                                                                                                                                                                                                                                                                                                  [1] - At more than one shop, FTE devs were responsible for providing support for code written by offshore that they were never even given the opportunity to review. And hell yes myself and others pushed for change, but it's never been a simple change. It almost always is 'GLWT'->'You get to review the final delivery but get 2 days'->'You get to review the set of changes'->'Ok you can review their sprint'->'OK just start reviewing every PR'.

                                                                                                                                                                                                                                                                                                                • BirAdam 1 day ago
                                                                                                                                                                                                                                                                                                                  By the time AI hype dies down and hurts the bottom line, AI systems might be good enough to do the jobs.

                                                                                                                                                                                                                                                                                                                  “The market can remain irrational longer than you can remain solvent.” — Warren Buffett

                                                                                                                                                                                                                                                                                                                  • janalsncm 1 day ago
                                                                                                                                                                                                                                                                                                                    Very well said. Using code assistance is going to be table stakes moving forward, not something that can replace people. It’s not like competitors can’t also purchase AI subscriptions.
                                                                                                                                                                                                                                                                                                                    • bbarn 1 day ago
                                                                                                                                                                                                                                                                                                                      Honestly, if you're not doing it now, you're behind. The sheer amount of time savings using it smartly can give you to allow you to focus on the parts that actually matter is massive.
                                                                                                                                                                                                                                                                                                                      • kweingar 1 day ago
                                                                                                                                                                                                                                                                                                                        If progress continues at the rate that AI boosters expect, then soon you won't have to use them smartly to get value (all existing workflows will churn and be replaced by newer, smarter workflows within months), and everybody who is behind will immediately catch up the moment they start to use the tool.
                                                                                                                                                                                                                                                                                                                  • acquisitionsilk 1 day ago
                                                                                                                                                                                                                                                                                                                    It is quite heartening to see so many people care about "good code". I fear it will make no difference.

                                                                                                                                                                                                                                                                                                                    The problem is that the software world got eaten up by the business world many years ago. I'm not sure at what point exactly, or if the writing was already on the wall when Bill Gates' wrote his open letter to hobbyists in 1976.

                                                                                                                                                                                                                                                                                                                    The question is whether shareholders and managers will accept less good code. I don't see how it would be logical to expect anything else, as long as profit lines go up why would they care.

                                                                                                                                                                                                                                                                                                                    Short of some sort of cultural pushback from developers or users, we're cooked, as the youth say.

                                                                                                                                                                                                                                                                                                                    • JackSlateur 1 day ago
                                                                                                                                                                                                                                                                                                                      Code is meant to power your business

                                                                                                                                                                                                                                                                                                                      Bad code leads to bad business

                                                                                                                                                                                                                                                                                                                      This makes me think of hosting departement; You know, which people who are using vmware, physical firewalls, dpi proxies and whatnot;

                                                                                                                                                                                                                                                                                                                      On the other edge, you have public cloud providers, which are using qemu, netfilter, dumb networking devices and stuff

                                                                                                                                                                                                                                                                                                                      Who got eaten by whom, nobody could have guessed ..

                                                                                                                                                                                                                                                                                                                      • chii 1 day ago
                                                                                                                                                                                                                                                                                                                        > Bad code leads to bad business

                                                                                                                                                                                                                                                                                                                        Bad business leads to bad business.

                                                                                                                                                                                                                                                                                                                        Bad code might be bad, or might be sufficient. It's situational. And by looking at what exists today, majority of code is pretty bad already - and not all businesses with bad code lead to bad businesses.

                                                                                                                                                                                                                                                                                                                        In fact, some bad code are very profitable for some businesses (ask any SAP integrator).

                                                                                                                                                                                                                                                                                                                        • JackSlateur 1 day ago
                                                                                                                                                                                                                                                                                                                          It is the survivor bias: "by looking at what is still alive today, majority of code is pretty bad"

                                                                                                                                                                                                                                                                                                                          It eludes all of those who died in the process : those still alives are here despite bad IT, not due to a bad IT

                                                                                                                                                                                                                                                                                                                          • tcoff91 1 day ago
                                                                                                                                                                                                                                                                                                                            The vast majority of code that makes money is pretty shitty.
                                                                                                                                                                                                                                                                                                                        • BirAdam 1 day ago
                                                                                                                                                                                                                                                                                                                          This is fun to think about. I used to think that all software was largely garbage, and at one point, I think this _was_ true. Sometime over the last 20 years, I believe this ceased to be the case. Most software these days actually works. Importantly, most software is actually stable enough that I can make it half an hour without panic saving.

                                                                                                                                                                                                                                                                                                                          Could most software be more awesome? Yes. Objectively, yes. Is most software garbage? Perhaps by raw volume of software titles, but are most popular applications I’ve actually used garbage? Nope. Do I loathe the whole subscription thing? Yes. Absolutely. Yet, I also get it. People expect software to get updated, and updates have costs.

                                                                                                                                                                                                                                                                                                                          So, the pertinent question here is, will AI systems be worse than humans? For now, yeah. Forever? Nope. The rate of improvement is crazy. Two years ago, LLMs I ran locally couldn’t do much of anything. Now? Generally acceptable junior dev stuff comes out of models I run on my Mac Studio. I have to fiddle with the prompts a bit, and it’s probably faster to just take a walk and think it over than spend an hour trying different prompts… but I’m a nerd and I like fiddling.

                                                                                                                                                                                                                                                                                                                          • robocat 1 day ago
                                                                                                                                                                                                                                                                                                                            > Short of some sort of cultural pushback from developers or users

                                                                                                                                                                                                                                                                                                                            Corporations create great code too: they're not all badly run.

                                                                                                                                                                                                                                                                                                                            The problem isn't a code quality issue: it is a moral issue of whether you agree with the goals of capitalist businesses.

                                                                                                                                                                                                                                                                                                                            Many people have to balance the needs of their wallet with their desire for beautiful software (I'm a developer-founder I love engineering and open source community but I'm also capitalist enough to want to live comfortably).

                                                                                                                                                                                                                                                                                                                          • am17an 1 day ago
                                                                                                                                                                                                                                                                                                                            All the world's smartest minds are racing towards replacing themselves. As programmers, we should take note and see where the wind is blowing. At least don't discard the possibility and rather be prepared for the future. Not to sound like a tin-foil hat but odds of achieving something like this increase by the day.

                                                                                                                                                                                                                                                                                                                            In the long term (post AGI), the only safe white-collar jobs would be those built on data which is not public i.e. extremely proprietary (e.g. Defense, Finance) and even those will rely heavily on customized AIs.

                                                                                                                                                                                                                                                                                                                            • bitpush 1 day ago
                                                                                                                                                                                                                                                                                                                              > All the world's smartest minds are racing towards replacing themselves

                                                                                                                                                                                                                                                                                                                              Isnt every little script, every little automation us programmers do in the same spirit? "I dont like doing this, so I'm going to automate it, so that I can focus on other work".

                                                                                                                                                                                                                                                                                                                              Sure, we're racing towards replacing ourselves, but there would be (and will be) other more interesting work for us to do when we're free to do that. Perhaps, all of us will finally have time to learn surfing, or garden, or something. Some might still write code themselves by hand, just like how some folks like making bread .. but making bread by hand is not how you feed a civilization - even if hundreds of bakers were put out of business.

                                                                                                                                                                                                                                                                                                                              • AstroBen 1 day ago
                                                                                                                                                                                                                                                                                                                                > all of us will finally have time to learn surfing, or garden

                                                                                                                                                                                                                                                                                                                                Unless you have a mortgage.. or rent.. or need to eat

                                                                                                                                                                                                                                                                                                                              • wijwp 1 day ago
                                                                                                                                                                                                                                                                                                                                > Not to sound like a tin-foil hat but odds of achieving something like this increase by the day.

                                                                                                                                                                                                                                                                                                                                Where do you get this? The limitations of LLMs are becoming more clear by the day. Improvements are slowing down. Major improvements come from integrations, not major model improvements.

                                                                                                                                                                                                                                                                                                                                AGI likely can't be achieved with LLMs. That wasn't as clear a couple years ago.

                                                                                                                                                                                                                                                                                                                                • drodgers 1 day ago
                                                                                                                                                                                                                                                                                                                                  I don't know how someone could be following the technical progress in detail and hold this view. The progress is astonishing, and the benchmarks are becoming saturated so fast that it's hard to keep track.

                                                                                                                                                                                                                                                                                                                                  Are there plenty of gaps left between here and most definitions of AGI? Absolutely. Nevertheless, how can you be sure that those gaps will remain given how many faculties these models have already been able to excel at (translation, maths, writing, code, chess, algorithm design etc.)?

                                                                                                                                                                                                                                                                                                                                  It seems to me like we're down to a relatively sparse list of tasks and skills where the models aren't getting enough training data, or are missing tools and sub-components required to excel. Beyond that, it's just a matter of iterative improvement until 80th percentile coder becomes 99th percentile coder becomes superhuman coder, and ditto for maths, persuasion and everything else.

                                                                                                                                                                                                                                                                                                                                  Maybe we hit some hard roadblocks, but room for those challenges to be hiding seems to be dwindling day by day.

                                                                                                                                                                                                                                                                                                                                  • materiallie 1 day ago
                                                                                                                                                                                                                                                                                                                                    I think benchmark targeting is going to be a serious problem going forward. The recent Nate Silver podcast on poker performance is interesting. Basically, the LLM models still suck at playing poker.

                                                                                                                                                                                                                                                                                                                                    Poker tests intelligence. So what gives? One interesting thing is that for whatever reason, poker performance isn't used a benchmark in the LLM showdown between big tech companies.

                                                                                                                                                                                                                                                                                                                                    The models have definitely improved in the past few years. I'm skeptical that there's been a "break-through", and I'm growing more skeptical of the exponential growth theory. It looks to me like the big tech companies are just throwing huge compute and engineering budgets at the existing transformer tech, to improve benchmarks one by one.

                                                                                                                                                                                                                                                                                                                                    I'm sure if Google allocated 10 engineers a dozen million dollars to improve Gemini's poker performance, it would increase. The idea before AGI and the exponential growth hypothesis is that you don't have to do that because the AI gets smarter in a general sense all on it's own.

                                                                                                                                                                                                                                                                                                                                • AstroBen 1 day ago
                                                                                                                                                                                                                                                                                                                                  Ultimately this needs to be solved politically

                                                                                                                                                                                                                                                                                                                                  Making our work more efficient, or humans redundant should be really exciting. It's not set in stone that we need to leave people middle aged with families and now completely unable to earn enough to provide a good life

                                                                                                                                                                                                                                                                                                                                  Hopefully if it happens, it happens to such a huge amount of people that it forces a change

                                                                                                                                                                                                                                                                                                                                  • lyu07282 1 day ago
                                                                                                                                                                                                                                                                                                                                    But that already happened to lots of industries and lots of people, we never cared before about them, now it's us so we care, but nothing is different about us. Just learn to code!
                                                                                                                                                                                                                                                                                                                                    • geraneum 1 day ago
                                                                                                                                                                                                                                                                                                                                      > But that already happened to lots of industries and lots of people, we never cared before about them

                                                                                                                                                                                                                                                                                                                                      We did. Why do you think labor laws, unions, etc. exist? Why do you think communism was appealing as an idea in the beginning to many? Whether the effects were good or bad or enough or not, that’s a different question. But these changes have demonstrably, grave consequences.

                                                                                                                                                                                                                                                                                                                                      • AstroBen 1 day ago
                                                                                                                                                                                                                                                                                                                                        The difference is in how many industries AI is threatening. It's not just coding on the chopping block
                                                                                                                                                                                                                                                                                                                                    • bgwalter 1 day ago
                                                                                                                                                                                                                                                                                                                                      The Nobel prize is said to have been created partly out of guilt over having invented dynamite, which was obviously used in a destructive manner.

                                                                                                                                                                                                                                                                                                                                      Now we have Geoffrey Hinton getting the prize for contributing to one of the most destructive inventions ever.

                                                                                                                                                                                                                                                                                                                                      • reducesuffering 1 day ago
                                                                                                                                                                                                                                                                                                                                        At least he and Yoshua Bengio are remorseful. Many others haven't even gotten that far...
                                                                                                                                                                                                                                                                                                                                      • cheema33 1 day ago
                                                                                                                                                                                                                                                                                                                                        > All the world's smartest minds are racing towards replacing themselves.

                                                                                                                                                                                                                                                                                                                                        I think they are hoping that their future is safe. And it is the average minds that will have to go first. There may be some truth to it.

                                                                                                                                                                                                                                                                                                                                        Also, many of these smartest minds are motivated by money, to safeguard their future, from a certain doom that they know might be coming. And AI is a good place to be if you want to accumulate wealth fast.

                                                                                                                                                                                                                                                                                                                                        • BirAdam 1 day ago
                                                                                                                                                                                                                                                                                                                                          Nah. As more people are rendered unemployed the buying market and therefore aggregate demand will fall. Fewer sales hurts the bottom line. At some point, revenues across the entire economy fall, and companies cannot afford the massive datacenters and nuclear power plants fueling them. The hardware gets sold cheap, the companies go under, and people get hired again. Eventually, some kind of equilibrium will be found or the world engages in the Butlerian Jihad.
                                                                                                                                                                                                                                                                                                                                        • notyouraibot 55 minutes ago
                                                                                                                                                                                                                                                                                                                                          So funny story, I tried using o3 for a relatively complex task yesterday, installing XCode iOS Simulator on an external SSD, it was my first time owning and using a macOS so I was truly lost, I followed everything it told me and by the end of the hour.. things got so bad that my machine couldn't even run normal basic node projects. I had to a proper fresh boot to get things working again. So yeah lesson learned.
                                                                                                                                                                                                                                                                                                                                          • bouncycastle 1 day ago
                                                                                                                                                                                                                                                                                                                                            Last night I spent hours fighting o3.

                                                                                                                                                                                                                                                                                                                                            I never made a Dockerfile in my life, so I thought it would be faster just getting o3 to point to the GitHub repo and let it figure out, rather than me reading the docs and building it myself.

                                                                                                                                                                                                                                                                                                                                            I spent hours debugging the file it gave me... It kept on adding hallucinations for things that didn't exist, and removing/rewriting other parts, and other big mistakes like understanding the difference between python3 and python and the intricacies with that.

                                                                                                                                                                                                                                                                                                                                            Finally I gave up and Googled some docs instead. Fixed my file in minutes and was able to jump into the container and debug the rest of the issues. AI is great, but it's not a tool to end all. You still need someone who is awake at the wheel.

                                                                                                                                                                                                                                                                                                                                            • halpow 1 day ago
                                                                                                                                                                                                                                                                                                                                              They're great at one-shotting verbose code, but if they're generate bad code the first time you're out of luck.

                                                                                                                                                                                                                                                                                                                                              I don’t think I ever got to write "this api doesn't exist" and then gotten a useful alternative.

                                                                                                                                                                                                                                                                                                                                              Claude is the only one that regularly tells me something isn't possible rather than making sh up.

                                                                                                                                                                                                                                                                                                                                              • throwaway314155 1 day ago
                                                                                                                                                                                                                                                                                                                                                Pro-tip: Check out Claude or Gemini. They hallucinate far less on coding tasks. Alternatively, enable internet search on o3 which boosts its ability to reference online documentation and real world usage examples.

                                                                                                                                                                                                                                                                                                                                                I get having a bad taste in your mouth but these tools _aren't _ magic and do have something of a steep learning curve in order to get the most out of them. Not dissimilar from vim/emacs (or lots of dev tooling).

                                                                                                                                                                                                                                                                                                                                                edit: To answer a reply (hn has annoyingly limited my ability to make new comments) yes, internet search is always available to ChatGpT as a tool. Explicitly clicking the globe icon will encourage the model to use it more often, however.

                                                                                                                                                                                                                                                                                                                                                • Sohcahtoa82 1 day ago
                                                                                                                                                                                                                                                                                                                                                  > enable internet search on o3

                                                                                                                                                                                                                                                                                                                                                  I didn't know it could even be disabled. It must be enabled by default, right?

                                                                                                                                                                                                                                                                                                                                                  • throwaway314155 1 day ago
                                                                                                                                                                                                                                                                                                                                                    You're correct. Tapping the globe icon encourages the model to use it more often.
                                                                                                                                                                                                                                                                                                                                              • frogperson 1 day ago
                                                                                                                                                                                                                                                                                                                                                The context required to write real software is just way too big for LLMs. Software is the business, codified. How is an LLM supposed to know about all the rules in all the departments plus all the special agreements promised to customers by the sales team?

                                                                                                                                                                                                                                                                                                                                                Right now the scope of what an LLM can solve is pretty generic and focused. Anytime more than a class or two is involved or if the code base is more than 20 or 30 files, then even the best LLMs start to stray and lose focus. They can't seem to keep a train of thought which leads to churning way too much code.

                                                                                                                                                                                                                                                                                                                                                If LLMs are going to replace real developers, they will need to accept significantly more context, they will need a way to gather context from the business at large, and some way to persist a train of thought across the life of a codebase.

                                                                                                                                                                                                                                                                                                                                                I'll start to get nervous when these problems are close to being solved.

                                                                                                                                                                                                                                                                                                                                                • zachlatta 1 day ago
                                                                                                                                                                                                                                                                                                                                                  I’d encourage you to try the 1M context window on Gemini 2.5 Pro. It’s pretty remarkable.

                                                                                                                                                                                                                                                                                                                                                  I paste in the entire codebase for my small ETL project (100k tokens) and it’s pretty good.

                                                                                                                                                                                                                                                                                                                                                  Not perfect, still a long ways to go, but a sign of the times to come.

                                                                                                                                                                                                                                                                                                                                                  • karn97 1 day ago
                                                                                                                                                                                                                                                                                                                                                    [flagged]
                                                                                                                                                                                                                                                                                                                                                    • helloplanets 1 day ago
                                                                                                                                                                                                                                                                                                                                                      > Anytime more than a class or two is involved or if the code base is more than 20 or 30 files, then even the best LLMs start to stray and lose focus. They can't seem to keep a train of thought which leads to churning way too much code.

                                                                                                                                                                                                                                                                                                                                                      At least if looking at this specific portion: You can have a 20-30 file code base, with 5-10 classes in the context, in full, and the rest of them filtered in a sensible way. Even a model with a 200k context window can handle this.

                                                                                                                                                                                                                                                                                                                                                      The output definitely can stray, but it's not the norm in my experience. Of course, if the output does start to stray, it needs to be snipped in the bud. And the fixes can range anywhere from working but bad code, to very close how you'd written it yourself, if you've clearly described how you want the code to be written.

                                                                                                                                                                                                                                                                                                                                                      If you're trying to fix a specific bug, for example, but don't provide thorough logs on what is happening in the code, it's much more likely the output will stray towards some average of what the problem could be, rather than what it actually is in the current code.

                                                                                                                                                                                                                                                                                                                                                      And this is absolutely not to say that LLMs could do what Antirez is doing. There is a massive amount of variation in how deeply people think about the code they're reading or writing.

                                                                                                                                                                                                                                                                                                                                                      • zachlatta 1 day ago
                                                                                                                                                                                                                                                                                                                                                        Yes… I did?
                                                                                                                                                                                                                                                                                                                                                  • smilbandit 1 day ago
                                                                                                                                                                                                                                                                                                                                                    From my limited experience, former coder now management but I still get to code now and then. I've found them helpful but also intrusive. Sometimes when it guesses the code for the rest of the line and next few lines it's going down a path I don't want to go but I have to take time to scan it. Maybe it's a configuration issue, but i'd prefer it didn't put code directly in my way or be off by default and only show when I hit a key combo.

                                                                                                                                                                                                                                                                                                                                                    One thing I know is that I wouldn't ask an LLM to write an entire section of code or even a function without going in and reviewing.

                                                                                                                                                                                                                                                                                                                                                    • haiku2077 1 day ago
                                                                                                                                                                                                                                                                                                                                                      Zed has a "subtle" mode like that. More editors should provide it. https://zed.dev/docs/ai/edit-prediction#switching-modes
                                                                                                                                                                                                                                                                                                                                                      • PartiallyTyped 1 day ago
                                                                                                                                                                                                                                                                                                                                                        > One thing I know is that I wouldn't ask an LLM to write an entire section of code or even a function without going in and reviewing.

                                                                                                                                                                                                                                                                                                                                                        These days I am working on a startup doing [a bit of] everything, and I don't like the UI it creates. It's useful enough when I make the building blocks and let it be, but allowing claude to write big sections ends up with lots of reworks until I get what I am looking for.

                                                                                                                                                                                                                                                                                                                                                      • solatic 1 day ago
                                                                                                                                                                                                                                                                                                                                                        Human coders are necessary because writing code is a political act of deciding between different trade-offs. antirez's whole post is explaining to Gemini what the trade-offs even were in the first place. No analysis of a codebase in isolation (i.e. without talking to the original coders, and without comments in the code) can distinguish between intentional prioritization of certain trade-offs or whether behavior is unintentional / written by a human in an imperfect way because they didn't know any better / buggy.

                                                                                                                                                                                                                                                                                                                                                        LLMs will never be able to figure out for themselves what your project's politics are and what trade-offs are supposed to be made. The penultimate model will still require a user to explain the trade-offs in a prompt.

                                                                                                                                                                                                                                                                                                                                                        • energy123 1 day ago
                                                                                                                                                                                                                                                                                                                                                          > LLMs will never be able to figure out for themselves what your project's politics are and what trade-offs are supposed to be made.

                                                                                                                                                                                                                                                                                                                                                          I wouldn't declare that unsolvable. The intentions of a project and how they fit into user needs can be largely inferred from the code and associated docs/README, combined with good world knowledge. If you're shown a codebase of a GPU kernel for ML, then as a human you instantly know the kinds of constraints and objectives that go into any decisions. I see no reason why an LLM couldn't also infer the same kind of meta-knowledge. Of course, this papers over the hard part of training the LLMs to actually do that properly, but I don't see why it's inherently impossible.

                                                                                                                                                                                                                                                                                                                                                          • solatic 20 hours ago
                                                                                                                                                                                                                                                                                                                                                            > associated docs/README

                                                                                                                                                                                                                                                                                                                                                            Many (I would even argue most) professional codebases either do not have their documentation (including tutorials and architecture diagrams) in the codebase alongside the code, if there is even such formal documentation at all. It's axiomatic as well that documentation is frequently out-of-date, and in any case represents past political decisions, not future ones; human owners can and do change their minds about which trade-offs are required over the lifetime of a project.

                                                                                                                                                                                                                                                                                                                                                            A simple case may be to plot codebase complexity against required scale; early projects benefit from simpler implementations that will not scale, and only after usage and demand are proven does it make sense to make the project more complex in order to support additional scale. So if you are an LLM looking at a codebase in isolation, do you make changes to add complexity to support additional scale? Do you simplify the codebase? Do you completely rewrite it in a different language (say, TypeScript -> Go or Rust)? How could an LLM possibly know which of these are appropriate without additional sources of telemetry at the very least and probably also needing to converse with stakeholders (i.e. bordering on AGI)?

                                                                                                                                                                                                                                                                                                                                                        • austin-cheney 1 day ago
                                                                                                                                                                                                                                                                                                                                                          In many cases developers are a low expectation commodity. In those cases I strongly believe humans are entirely replaceable by AI and I am saying that as somebody with an exceptionally low opinion of LLMs.

                                                                                                                                                                                                                                                                                                                                                          Honestly though, when that replacement comes there is no sympathy to be had. Many developers have brought this upon themselves. For roughly the 25 year period from 1995 to 2020 businesses have been trying to turn developers into mindless commodities that are straight forward to replace. Developers have overwhelmingly encouraged this and many still do. These are the people who hop employers every 2 years and cannot do their jobs without lying on their resumes or complete reliance on a favorite framework.

                                                                                                                                                                                                                                                                                                                                                          • ponector 20 hours ago
                                                                                                                                                                                                                                                                                                                                                            But to job hop every 2 years is the best strategy to earn more money and experience.
                                                                                                                                                                                                                                                                                                                                                            • zxexz 23 hours ago
                                                                                                                                                                                                                                                                                                                                                              I find myself wondering about your story, and would love it if you would elaborate more. I have gotten some use out of LLMs, and have been quite involved in training a few compute intensive (albeit domain-specific) ones.

                                                                                                                                                                                                                                                                                                                                                              Maybe it's the way you talk about 'developers'. Nothing I have seen has felt like the sky falling on an industry; to me at most it's been the sky falling on a segment of silicon valley.

                                                                                                                                                                                                                                                                                                                                                              • austin-cheney 23 hours ago
                                                                                                                                                                                                                                                                                                                                                                It’s all about perspectives. In many cases the perspectives between how a developer identifies their level of participation versus what they actually do as a work activity differ substantially. For example many developers may refer to themselves as engineers when they have done nothing remotely close to measurements, research, or policy creation in compliance attainment.

                                                                                                                                                                                                                                                                                                                                                                With that out of the way let’s look only at what many developers actually do. If a given developer only uses a framework to put text on screen or respond to a user interaction then they can be replaced. LLMs can already do this better than people. That becomes substantially more true after accounting for secondary concerns: security, accessibility, performance, regression, and more.

                                                                                                                                                                                                                                                                                                                                                                If a developer is doing something more complex that accounts for systems analysis or human behavior then LLMs are completely insufficient.

                                                                                                                                                                                                                                                                                                                                                            • some-guy 1 day ago
                                                                                                                                                                                                                                                                                                                                                              The main thing LLMs have helped me with, and always comes back to, tasks that require bootstrapping / Googling:

                                                                                                                                                                                                                                                                                                                                                              1) Starting simple codebases 2) Googling syntax 3) Writing bash scripts that utilize Unix commands whose arguments I have never bothered to learn in the first place.

                                                                                                                                                                                                                                                                                                                                                              I definitely find time savings with these, but the esoteric knowledge required to work on a 10+ year old codebase is simply too much for LLMs still, and the code alone doesn't provide enough context to do anything meaningful, or even faster than I would be able to do myself.

                                                                                                                                                                                                                                                                                                                                                              • mywittyname 1 day ago
                                                                                                                                                                                                                                                                                                                                                                LLMs are amazing at shell scripting. It's one of those tasks where I always half-ass it because I don't really know how to properly handle errors and never really learned the correct way. But man, perplexity and poop out a basic shell script in a few seconds with pretty much every edge case I can think of covered.
                                                                                                                                                                                                                                                                                                                                                                • tcoff91 1 day ago
                                                                                                                                                                                                                                                                                                                                                                  It really is the thing they are best at.
                                                                                                                                                                                                                                                                                                                                                              • pupppet 1 day ago
                                                                                                                                                                                                                                                                                                                                                                If an LLM just finds patterns, is it even possible for an LLM to be GOOD at anything? Doesn't that mean at best it will be average?
                                                                                                                                                                                                                                                                                                                                                                • bitpush 1 day ago
                                                                                                                                                                                                                                                                                                                                                                  Humans are also almost always operating on patterns. This is why "experience" matters a lot.

                                                                                                                                                                                                                                                                                                                                                                  Very few people are doing truly cutting edge stuff - we call them visionaries. But most of the time, we're just merely doing what's expected

                                                                                                                                                                                                                                                                                                                                                                  And yes, that includes this comment. This wasnt creative or an original thought at all. I'm sure hundreds of people have had similar thought, and I'm probably parroting someone else's idea here. So if I can do it, why cant LLM?

                                                                                                                                                                                                                                                                                                                                                                  • dgb23 1 day ago
                                                                                                                                                                                                                                                                                                                                                                    The times we just operate on patterns is when we code boilerplate or just very commonly written code. There's value in speeding this up and LLMs help here.

                                                                                                                                                                                                                                                                                                                                                                    But generally speaking I don't experience programming like that most of the time. There are so many things going on that have nothing to do with pattern matching while coding.

                                                                                                                                                                                                                                                                                                                                                                    I load up a working model of the running code in my head and explore what it should be doing in a more abstract/intangible way and then I translate those thoughts to code. In some cases I see the code in my inner eye, in others I have to focus quite a lot or even move around or talk.

                                                                                                                                                                                                                                                                                                                                                                    My mind goes to different places and experiences. Sometimes it's making new connections, sometimes it's processing a bit longer to get a clearer picture, sometimes it re-shuffles priorities. A radical context switch may happen at any time and I delete a lot of code because I found a much simpler solution.

                                                                                                                                                                                                                                                                                                                                                                    I think that's a qualitative, insurmountable difference between an LLM and an actual programmer. The programmer thinks deeply about the running program and not just the text that needs to be written.

                                                                                                                                                                                                                                                                                                                                                                    There might be different types of "thinking" that we can put into a computer in order to automate these kinds of tasks reliably and efficiently. But just pattern matching isn't it.

                                                                                                                                                                                                                                                                                                                                                                  • riknos314 1 day ago
                                                                                                                                                                                                                                                                                                                                                                    My experience is that LLMs regress to the average of the context they have for the task at hand.

                                                                                                                                                                                                                                                                                                                                                                    If you're getting average results you most likely haven't given it enough details about what you're looking for.

                                                                                                                                                                                                                                                                                                                                                                    The same largely applies to hallucinations. In my experience LLMs hallucinate significantly more when at or pushed to exceed the limits of their context.

                                                                                                                                                                                                                                                                                                                                                                    So if you're looking to get a specific output, your success rate is largely determined by how specific and comprehensive the context the LLM has access to is.

                                                                                                                                                                                                                                                                                                                                                                    • jaccola 1 day ago
                                                                                                                                                                                                                                                                                                                                                                      Most people (average and below average) can tell when something is above average, even if they cannot create above average work, so using RLHF it should be quite possible to achieve above average.

                                                                                                                                                                                                                                                                                                                                                                      Indeed it is likely already the case that in training the top links scraped or most popular videos are weighted higher, these are likely to be better than average.

                                                                                                                                                                                                                                                                                                                                                                      • lukan 1 day ago
                                                                                                                                                                                                                                                                                                                                                                        There are bad patterns and good patterns. But whether a pattern is the right one for a specific task is something different.

                                                                                                                                                                                                                                                                                                                                                                        And what really matters is, if the task gets reliable solved.

                                                                                                                                                                                                                                                                                                                                                                        So if they actually could manage this on average with average quality .. that would be a next level gamechanger.

                                                                                                                                                                                                                                                                                                                                                                        • JackSlateur 1 day ago
                                                                                                                                                                                                                                                                                                                                                                          Yes, IA is basically a random machine aiming for average outcome

                                                                                                                                                                                                                                                                                                                                                                          IA is neat for average people, to produce average code, for average compagnies

                                                                                                                                                                                                                                                                                                                                                                          In a competitive world, using IA is a death sentence;

                                                                                                                                                                                                                                                                                                                                                                        • bdbenton5255 1 day ago
                                                                                                                                                                                                                                                                                                                                                                          The human ability to design computer programs through abstractions and solve creative problems like these is arguably more important than being able to crank out lines of code that perform specific tasks.

                                                                                                                                                                                                                                                                                                                                                                          The programmer is an architect of logic and computers translate human modes of thought into instructions. These tools can imitate humans and produce code given certain tasks, typically by scraping existing code, but they can't replace that abstract level of human thought to design and build in the same way.

                                                                                                                                                                                                                                                                                                                                                                          When these models are given greater functionality to not only output code but to build out entire projects given specifications, then the role of the human programmer must evolve.

                                                                                                                                                                                                                                                                                                                                                                          • AlotOfReading 1 day ago
                                                                                                                                                                                                                                                                                                                                                                            Unrelated to the LLM discussion, but a hash function function is the wrong construction for the accumulator solution. The hashing part increases the probability that A and B have a collision that leads to a false negative here. Instead, you want a random invertible mapping, which guarantees that no two pointers will "hash" to the same value, while distributing the bits. Splitmix64 is a nice one, and I believe the murmurhash3 finalizer is invertible, as well as some of the xorshift RNGs if you avoid the degenerate zero cycle.
                                                                                                                                                                                                                                                                                                                                                                            • antirez 1 day ago
                                                                                                                                                                                                                                                                                                                                                                              Any Feistel Network has the property you stated actually, and this was one of the approaches I was thinking using as I can have the seed as part of the non linear transformation of the Feistel Network. However I'm not sure that this actually decreases the probability of A xor B xor C xor D being accidentally zero, bacause the problem with pointers is that they may change only for a small part. When you using hashing because of avalanche effect this is going a lot harder since you are no longer xoring the pointer structure.

                                                                                                                                                                                                                                                                                                                                                                              What I mean is that you are right assuming we use a transformation that still while revertible has avalanche effect. Btw in practical terms I doubt there are practical differences.

                                                                                                                                                                                                                                                                                                                                                                              • AlotOfReading 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                You can guarantee that the probability is the theoretical minimum with a bijection. I think that would be 2^-N since it's just the case where everything's on a maximum length cycle, but I haven't thought about it hard enough to be completely certain.

                                                                                                                                                                                                                                                                                                                                                                                A good hash function intentionally won't hit that level, but it should be close enough not to matter with 64 bit pointers. 32 bits is small enough that I'd have concerns at scale.

                                                                                                                                                                                                                                                                                                                                                                            • decasia 1 day ago
                                                                                                                                                                                                                                                                                                                                                                              We aren't expecting LLMs to come up with incredibly creative software designs right now, we are expecting them to execute conventional best practices based on common patterns. So it makes sense to me that it would not excel at the task that it was given here.

                                                                                                                                                                                                                                                                                                                                                                              The whole thing seems like a pretty good example of collaboration between human and LLM tools.

                                                                                                                                                                                                                                                                                                                                                                              • writeslowly 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                I haven't actually had that much luck with having them output a boring API boilerplate in large Java projects. Like "I need to create a new BarOperation that has to go in a different set of classes and files and API prefixes than all the FooOperations and I don't feel like copy pasting all the yaml and Java classes" but the AI has problems following this. Maybe they work better in small projects.

                                                                                                                                                                                                                                                                                                                                                                                I actually like LLMs better for creative thinking because they work like a very powerful search engine that can combine unrelated results and pull in adjacent material I would never personally think of.

                                                                                                                                                                                                                                                                                                                                                                                • coffeeismydrug 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                  > Like "I need to create a new BarOperation that has to go in a different set of classes and files and API prefixes than all the FooOperations and I don't feel like copy pasting all the yaml and Java classes" but the AI has problems following this.

                                                                                                                                                                                                                                                                                                                                                                                  To be fair, I also have problems following this.

                                                                                                                                                                                                                                                                                                                                                                                • ehutch79 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                  Uh, no. I've seen the twitter posts saying llms will replace me. I've watched the youtube videos saying llms will code whole apps on one prompt, but are light on details or only show the most basic todo app from every tutorial.

                                                                                                                                                                                                                                                                                                                                                                                  We're being told that llms are now reasoning, which implies they can make logical leaps and employ creativity to solve problems.

                                                                                                                                                                                                                                                                                                                                                                                  The hype cycle is real and setting expectations that get higher with the less you know about how they work.

                                                                                                                                                                                                                                                                                                                                                                                  • prophesi 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                    > The hype cycle is real and setting expectations that get higher with _the less you know about how they work_.

                                                                                                                                                                                                                                                                                                                                                                                    I imagine on HN, the expectations we're talking about are from fellow software developers who at least have a general idea on how LLM's work and their limitations.

                                                                                                                                                                                                                                                                                                                                                                                    • bluefirebrand 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                      Right below this is a comment

                                                                                                                                                                                                                                                                                                                                                                                      > you will almost certainly be replaced by an llm in the next few years

                                                                                                                                                                                                                                                                                                                                                                                      So... Maybe not. I agree that Hacker News does have a generally higher quality of contributors than many places on the internet, but it absolutely is not a universal for HNers. There are still quite a few posters here that have really bought into the hype for whatever reason

                                                                                                                                                                                                                                                                                                                                                                                    • bgwalter 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                      After the use-after-free hype article I tried CoPilot and it outright refused to find vulnerabilities.

                                                                                                                                                                                                                                                                                                                                                                                      Whenever I try some claim, it does not work. Yes, I know, o3 != CoPilot but I don't have $120 and 100 prompts to spend on making a point.

                                                                                                                                                                                                                                                                                                                                                                                      • ldjkfkdsjnv 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                        you will almost certainly be replaced by an llm in the next few years
                                                                                                                                                                                                                                                                                                                                                                                        • einpoklum 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                          You mean, as a HackerNews commenter? Well, maybe...

                                                                                                                                                                                                                                                                                                                                                                                          In fact, maybe most of has have been replaced by LLMs already :-)

                                                                                                                                                                                                                                                                                                                                                                                    • elzbardico 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                      I use LLMs a lot, and call me arrogant, but every time I see a developer saying that LLMs will substitute them, I think they are probably shitty developers.
                                                                                                                                                                                                                                                                                                                                                                                      • Fernicia 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                        If it automates 1/5th of your work, then what's unreasonable about thinking that your team could be 4 developers instead of 5?
                                                                                                                                                                                                                                                                                                                                                                                        • AstroBen 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                          If software costs 80% as much to write, what's unreasonable about thinking that more businesses would integrate more of it, hiring more developers?
                                                                                                                                                                                                                                                                                                                                                                                          • Fernicia 17 hours ago
                                                                                                                                                                                                                                                                                                                                                                                            I think both will happen. I was merely demonstrating that just because an AI can't replace the entirety of your work, doesn't mean it won't make you redundant.
                                                                                                                                                                                                                                                                                                                                                                                            • aschobel 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                              Bingo. This additional throughput could be used to create more polished software. What happens in a free market; would your competitor fall behind or will they try to match your polish?
                                                                                                                                                                                                                                                                                                                                                                                            • archagon 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                              This just feels like another form of the mythical man month argument.
                                                                                                                                                                                                                                                                                                                                                                                          • DrJid 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                            I never quite understand these articles though. It's not about Humans vs. AI.

                                                                                                                                                                                                                                                                                                                                                                                            It's about Humans vs. Humans+AI

                                                                                                                                                                                                                                                                                                                                                                                            and 4/5, Humans+AI > Humans.

                                                                                                                                                                                                                                                                                                                                                                                            • darkport 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                              I think this is true for deeply complex problems, but For everyday tasks an LLM is infinitely “better”.

                                                                                                                                                                                                                                                                                                                                                                                              And by better, I don’t mean in terms of code quality because ultimately that doesn’t matter for shipping code/products, as long as it works.

                                                                                                                                                                                                                                                                                                                                                                                              What does matter is speed. And an LLM speeds me up at least 10x.

                                                                                                                                                                                                                                                                                                                                                                                              • kweingar 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                You're making at least a year's worth of pre-LLM progress in 5 weeks?

                                                                                                                                                                                                                                                                                                                                                                                                You expect to achieve more than a decade of pre-LLM accomplishments between now and June 2026?

                                                                                                                                                                                                                                                                                                                                                                                                • nevertoolate 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                  How do you measure this?
                                                                                                                                                                                                                                                                                                                                                                                                • rel2thr 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                  Antirez is a top 0.001% coder . Don’t think this generalizes to human coders at large
                                                                                                                                                                                                                                                                                                                                                                                                  • ljlolel 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                    Seriously, he’s one of the best on the planet of course it’s not better than him. If so we’d be cooked.

                                                                                                                                                                                                                                                                                                                                                                                                    99% of professional software developers don’t understand what he said much less can come up with it (or evaluate it like Gemini).

                                                                                                                                                                                                                                                                                                                                                                                                    This feels a bit like a humblebrag about how well he can discuss with an LLM compared to others vibecoding.

                                                                                                                                                                                                                                                                                                                                                                                                    • justacrow 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                      Hey, my CEO is saying that LLMs are also top 0.001% coders now, so should at least be roughly equivalent.
                                                                                                                                                                                                                                                                                                                                                                                                    • seabirdman 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                      Super hard problems are often solved by making strange weird connections derived from deep experience plus luck. Like finding the one right key in a pile of keys. The intuition you used to solve your problem IS probably beyond current agents. But, that too will change perhaps by harnessing the penchant of these systems to “hallucinate”? Or, some method or separate algorithm for dealing with super hard problems creatively and systematically. Recently, I was working on a hard imaging problem (for me) and remembered a bug I had inadvertently introduced and fixed a few days earlier. I was like wait a minute - because in that random bug I saw opportunity and was able to actually use the bug to solve my problem. I went back to my agent and it agreed that there was virtually no way it could have ever seen and solved the problem in that way. But that too will come. Rest assured.
                                                                                                                                                                                                                                                                                                                                                                                                      • ntonozzi 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                        If you care that much about having correct data you could just do a SHA-256 of the whole thing. Or an HMAC. It would probably be really fast. If you don’t care much you can just do murmur hash of the serialized data. You don’t really need to verify data structure properties if you know the serialized data is correct.
                                                                                                                                                                                                                                                                                                                                                                                                        • marcosno 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                          LLMs can be very creative, when pushed. In order to find a creative solution, like antirez needed, there are several tricks I use:

                                                                                                                                                                                                                                                                                                                                                                                                          Increase the temperature of the LLMs.

                                                                                                                                                                                                                                                                                                                                                                                                          Ask several LLMs, each several time the same question, with tiny variations. Then collect all answers, and do a second/third round asking each LLM to review all collected answers and improve.

                                                                                                                                                                                                                                                                                                                                                                                                          Add random constraints, one constraints per question. For example, to LLM: can you do this with 1 bit per X. Do this in O(n). Do this using linked lists only. Do this with only 1k memory. Do this while splitting the task to 1000 parallel threads, etc.

                                                                                                                                                                                                                                                                                                                                                                                                          This usually kicks the LLM out of its confort zone, into creative solutions.

                                                                                                                                                                                                                                                                                                                                                                                                          • dwringer 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                            Definitely a lot to be said for these ideas, even just that it helps to start a fresh chat and ask the same question in a better way a few times (using the quality of response to gauge what might be "better"). I have found if I do this a few times and Gemini strikes out, I've manually optimized the question by this point that I can drop it into Claude and get a good working solution. Conversely, having a discussion with the LLM about the potential solution, letting it hold on to the context as described in TFA, has in my experience caused the models to pretty universally end up stuck in a rut sooner or later and become counterproductive to work with. Not to mention that way eats up a ton of api usage allotment.
                                                                                                                                                                                                                                                                                                                                                                                                          • agumonkey 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                            There's also the subset of devs who are just bored, LLMs will end up as an easier StackOverflow and if the solution is not one script away, then you're back to square one. I already had a few of "well, uhm, chatGPT told me what you said basically".
                                                                                                                                                                                                                                                                                                                                                                                                            • coldtea 19 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                              >Gemini was quite impressed about the idea

                                                                                                                                                                                                                                                                                                                                                                                                              Like sex professionals, Gemini and co are made to be impressed and have possitive things to say about programming ideas you propose and find your questions "interesting", "deep", "great" and so.

                                                                                                                                                                                                                                                                                                                                                                                                              • _fat_santa 19 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                I would correct that quote to say “Gemini was trained to respond that it was impressed with my idea”

                                                                                                                                                                                                                                                                                                                                                                                                                Being “impressed” is a human feeling that can’t be transcribed to an AI period. An AI can tell you that it’s impressed because it’s been trained to do so, it doesn’t “know” (and by this I’m referring to knowing what the feeling is, not knowing the definition) what it means to be impressed

                                                                                                                                                                                                                                                                                                                                                                                                              • headelf 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                What do you mean "Still"? We've only had LLMs writing code for 1.5 years... at this rate it won't be long.
                                                                                                                                                                                                                                                                                                                                                                                                                • cess11 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                  More like five years. It's been around for much longer than a lot of people feel it has for some reason.
                                                                                                                                                                                                                                                                                                                                                                                                                • codr7 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                  I would say you thought about this solution because you are creative, something a computer will never be no matter how much data you throw at it.

                                                                                                                                                                                                                                                                                                                                                                                                                  How did it help, really? By telling you your idea was no good?

                                                                                                                                                                                                                                                                                                                                                                                                                  A less confident person might have given up because of the feedback.

                                                                                                                                                                                                                                                                                                                                                                                                                  I just can't understand why people are so excited about having an algorithm guessing for them. Is it the thrill when it finally gets something right?

                                                                                                                                                                                                                                                                                                                                                                                                                  • bachmeier 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                    I suspect humans will always be critical to programming. Improved technology won't matter if the economics isn't there.

                                                                                                                                                                                                                                                                                                                                                                                                                    LLMs are great as assistants. Just today, Copilot told me it's there to do the "tedious and repetitive" parts so I can focus my energy on the "interesting" parts. That's great. They do the things every programmer hates having to do. I'm more productive in the best possible way.

                                                                                                                                                                                                                                                                                                                                                                                                                    But ask it to do too much and it'll return error-ridden garbage filled with hallucinations, or just never finish the task. The economic case for further gains has diminished greatly while the cost of those gains rises.

                                                                                                                                                                                                                                                                                                                                                                                                                    Automation killed tons of manufacturing jobs, and we're seeing something similar in programming, but keep in mind that the number of people still working in manufacturing is 60% of the peak, and those jobs are much better than the ones in the 1960s and 1970s.

                                                                                                                                                                                                                                                                                                                                                                                                                    • noslenwerdna 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                      Sure, it's just that the era of super high paying programming jobs may be over.

                                                                                                                                                                                                                                                                                                                                                                                                                      And also, manufacturing jobs have greatly changed. And the effect is not even, I imagine. Some types of manufacturing jobs are just gone.

                                                                                                                                                                                                                                                                                                                                                                                                                      • monocularvision 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                        That might be the case. Perhaps it lowers the difficulty level so more people can do it and therefor puts downward pressure on wages.

                                                                                                                                                                                                                                                                                                                                                                                                                        Or… it still requires similar education and experience but programmers end up so much more efficient they earn _more_.

                                                                                                                                                                                                                                                                                                                                                                                                                        Hard to say right now.

                                                                                                                                                                                                                                                                                                                                                                                                                        • bachmeier 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                          > the era of super high paying programming jobs may be over.

                                                                                                                                                                                                                                                                                                                                                                                                                          Probably, but I'm not sure that had much to do with AI.

                                                                                                                                                                                                                                                                                                                                                                                                                          > Some types of manufacturing jobs are just gone

                                                                                                                                                                                                                                                                                                                                                                                                                          The manufacturing work that was automated is not exactly the kind of work people want to do. I briefly did some of that work. Briefly because it was truly awful.

                                                                                                                                                                                                                                                                                                                                                                                                                      • twodave 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                        If you stick with the same software ecosystem long enough you will collect (and improve upon) ways of solving classes of problems. These are things you can more or less reproduce without thinking too much or else build libraries around. An LLM may or may not become superior at this sort of exercise at some point, and might or might not be able to reliably save me some time typing. But these are already the boring things about programming.

                                                                                                                                                                                                                                                                                                                                                                                                                        So much of it is exploratory, deciding how to solve a problem from a high level, in an understandable way that actually helps the person who it’s intended to help and fits within their constraints. Will an LLM one day be able to do all of that? And how much will it cost to compute? These are the questions we don’t know the answer to yet.

                                                                                                                                                                                                                                                                                                                                                                                                                        • vouaobrasil 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                          The question is, for how long?
                                                                                                                                                                                                                                                                                                                                                                                                                          • spion 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                            Vibe-wise, it seems like progress is slowing down and recent models aren't substantially better than their predecessors. But it would be interesting to take a well-trusted benchmark and plot max_performance_until_date(foreach month). (Too bad aider changed recently and there aren't many older models; https://aider.chat/docs/leaderboards/by-release-date.html has not been updated in a while with newer stuff, and the new benchmark doesn't have the classic models such as 3.5, 3.5 turbo, 4, claude 3 opus)
                                                                                                                                                                                                                                                                                                                                                                                                                            • vouaobrasil 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                              I think that we can't expect continuous progress either, though. Often in computer science it's more discrete, and unexpected. Computer chess was basically stagnant until one team, even the evolution of species often behaves in a punctuated way rather than as a sum of many small adaptations. I'm much more interested (worried) of what the world will be like in 30 years, rather than in the next 5.
                                                                                                                                                                                                                                                                                                                                                                                                                              • spion 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                Its hard to say. Historically new discoveries in AI often generated great excitement and high expectations, followed by some progress, then stalling, disillusionment and AI winter. Maybe this time it will be different. Either way what was achieved so far is already a huge deal.
                                                                                                                                                                                                                                                                                                                                                                                                                            • jppittma 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                              It's really gonna depend on the project. When my hobby project was greenfield, the AI was way better than I am. It was (still is) more knowledgable about the standards that govern the field and about low level interface details. It can shit out a bunch of code that relies on knowing these details in seconds/minutes, rather than hours/days.

                                                                                                                                                                                                                                                                                                                                                                                                                              Now that the project has grown and all that stuff is hammered out, it can't seem to consistently write code that compiles. It's very tunnel visioned on the specific file its generating, rather than where that fits in the context of what/how we're building what we're building.

                                                                                                                                                                                                                                                                                                                                                                                                                              • jonator 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                We can slightly squeeze more juice out of them with larger projects by providing better context, docs, examples of what we want, background knowledge, etc.

                                                                                                                                                                                                                                                                                                                                                                                                                                Like people, LLMs don't know what they don't know (about your project).

                                                                                                                                                                                                                                                                                                                                                                                                                                • sixQuarks 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                  Again, the question is for how long
                                                                                                                                                                                                                                                                                                                                                                                                                                • kilroy123 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                  My crackpot guess is ~5 years. The incentives are just too damn high to not keep innovating in the space.

                                                                                                                                                                                                                                                                                                                                                                                                                                  We'll find new ways to push the tech.

                                                                                                                                                                                                                                                                                                                                                                                                                                  • 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                    • sixQuarks 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                      Exactly! We’ve been seeing more and more posts like this, saying how AI will never take developer jobs or will never be as good as coders. I think it’s some sort of coping mechanism.

                                                                                                                                                                                                                                                                                                                                                                                                                                      These posts are gonna look really silly in the not too distant future.

                                                                                                                                                                                                                                                                                                                                                                                                                                      I get it, spending countless hours honing your craft and knowing that AI will soon make almost everything you learned useless is very scary.

                                                                                                                                                                                                                                                                                                                                                                                                                                      • sofal 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                        I'm constantly disappointed by how little I'm able to delegate to AI after the unending promises that I'll be able to delegate nearly 100% of what I do now "in the not too distant future". It's tired impatience and merited skepticism that you mistake for fear and coping. Just because people aren't on the hype train with you doesn't mean they're afraid.
                                                                                                                                                                                                                                                                                                                                                                                                                                        • vouaobrasil 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                          Personally, I am. Lots of unusual skills I have, have already been taken by AI. That's not to say I think I'm in trouble, but I think it's sad I can't apply some of these skills that I learned just a couple of years ago like audio editing because AI does it now. Neither do I want to work as an AI operator, which I find boring and depressing. So, I've just moved onto something else, but it's still discouraging.

                                                                                                                                                                                                                                                                                                                                                                                                                                          Also, so many people said the same thing about chess when the first chess programs came out. "It will never beat an international master." Then, "it will never beat a grandmaster." And Kasparov said, "it would never beat me or Karpov."

                                                                                                                                                                                                                                                                                                                                                                                                                                          Look where we are today. Can humanity adapt? Yes, probably. But that new world IMO is worse than it is today, rather lacking in dignity I'd say.

                                                                                                                                                                                                                                                                                                                                                                                                                                          • sixQuarks 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                            Yes. I know what you’re referring to, but you can’t ignore the pace of improvement. I think within 2-3 years we will have AI coding that can do anything a senior level coder can do.
                                                                                                                                                                                                                                                                                                                                                                                                                                          • foobar83 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                            Nobody knows what the future holds, including you.
                                                                                                                                                                                                                                                                                                                                                                                                                                            • vouaobrasil 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                              That is true, which is why we should be cautious instead of careless.
                                                                                                                                                                                                                                                                                                                                                                                                                                              • sixQuarks 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                Yes, but we can see current progress and extrapolate into the future. I give it 2/3 years before AI can code as well as a senior level coder
                                                                                                                                                                                                                                                                                                                                                                                                                                          • prmph 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                            There's something fundamental here.

                                                                                                                                                                                                                                                                                                                                                                                                                                            There is a principle (I forget where I encountered it) that it is not code itself that is valuable, but the knowledge of a specific domain that an engineering team develops as they tackle a project. So code itself is a liability, but the domain knowledge is what is valuable. This makes sense to me and matched my long experience with software projects.

                                                                                                                                                                                                                                                                                                                                                                                                                                            So, if we are entrusting coding to LLMs, how will that value develop? And if we want to use LLMs but at the same time develop the domain acumen, that means we would have to architects things and hand them over to LLMs to implement, thoroughly check what they produce, and generally guide them carefully. In that case they are not saving much time.

                                                                                                                                                                                                                                                                                                                                                                                                                                            • jonator 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                              I believe it will raise the standard of what is valuable. Now that LLMs can now handle what we consider "mundane" parts of building a project (boilerplate), humans can dedicate focused efforts to the higher impact areas of innovation and problem solving. As LLMs get better, this bar simply continues to rise.
                                                                                                                                                                                                                                                                                                                                                                                                                                            • catigula 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                              Working with Claude 4 and o3 recently shows me just how fundamentally LLMs haven't really solved the core problems such as hallucinations and weird refactors/patterns to force success (i.e. if account not found, fallback to account id 1).
                                                                                                                                                                                                                                                                                                                                                                                                                                              • gumbojuice 23 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                I like to use llm to produce code for known problems that I don't have memorized.

                                                                                                                                                                                                                                                                                                                                                                                                                                                I memorize really little and tend to spend time on reinventing algorithms or looking them up in documentation. Verifying is easy except the fee cases where the llm produces something really weird. But then fallback to docs or reinventing.

                                                                                                                                                                                                                                                                                                                                                                                                                                                • jonator 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                  I think will will increasingly be orchestrators. Like at a symphony. Previously, most humans were required to be on the floor playing the individual instruments, but now, with AI, everyone can be their own composer.
                                                                                                                                                                                                                                                                                                                                                                                                                                                  • nixpulvis 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                    The number one use case for AI for me as a programmer is still help finding functions which are named something I didn't expect as I'm learning a new language/framework/library.

                                                                                                                                                                                                                                                                                                                                                                                                                                                    Doing the actual thinking is generally not the part I need too much help with. Though it can replace googling info in domains I'm less familiar with. The thing is, I don't trust the results as much and end up needing to verify it anyways. If anything AI has made this harder, since I feel searching the web for authoritative, expert information has become harder as of late.

                                                                                                                                                                                                                                                                                                                                                                                                                                                    • taormina 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                      My problem with this usage is that the LLMs seem equally likely to make up a function they wish existed. When questioned about the seeming-too-convenient method they will usually admit to having made it up on the spot. (This happens a lot in Flutter/Flame land, I'm sure it's better at something more mainstream like Python?) That being said, I do agree that using it as supplemental documentation is one of the better usecases I have for it.
                                                                                                                                                                                                                                                                                                                                                                                                                                                    • AstroBen 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                      Better than LLMs.. for now. I'm endlessly critical of the AI hype but the truth here is that no-one has any idea what's going to happen 3-10 years from now. It's a very quickly changing space with a lot of really smart people working on it. We've seen the potential

                                                                                                                                                                                                                                                                                                                                                                                                                                                      Maybe LLMs completely trivialize all coding. The potential for this is there

                                                                                                                                                                                                                                                                                                                                                                                                                                                      Maybe progress slows to a snails pace, the VC money runs out and companies massively raise prices making it not worth it to use

                                                                                                                                                                                                                                                                                                                                                                                                                                                      No one knows. Just sit back and enjoy the ride. Maybe save some money just in case

                                                                                                                                                                                                                                                                                                                                                                                                                                                      • 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                        • galaxyLogic 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                          Coding is not like multiplication. You can teach kids the multiplication table, or you can give them a calculator and both will work. With coding the problem is the "spec" is so much more complicated than just asking what is 5 * 7.

                                                                                                                                                                                                                                                                                                                                                                                                                                                          Maybe the way forward would be to invent better "specifiction languages" that are easy enough for humans to use, then let the AI implement the specifciation you come up with.

                                                                                                                                                                                                                                                                                                                                                                                                                                                          • vjvjvjvjghv 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                            I think we need to accept that in the not too far future LLMs will be able to do most of the mundane tasks we have to do every day. I don't see why an AI can't set up kubernetes, caching layers, testing, databases, scaling, check for security problems and so on. These things aren't easy but I think they are still very repetitive and therefore can be automated.

                                                                                                                                                                                                                                                                                                                                                                                                                                                            There will always be a place for really good devs but for average people (most of us are average) I think there will be less and less of a place.

                                                                                                                                                                                                                                                                                                                                                                                                                                                            • bluefirebrand 13 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                              > There will always be a place for really good devs but for average people (most of us are average) I think there will be less and less of a place

                                                                                                                                                                                                                                                                                                                                                                                                                                                              You open your post with "we need to accept" and then end with this

                                                                                                                                                                                                                                                                                                                                                                                                                                                              This terrifies me. The idea that AI results in me having "less of a place" in society?

                                                                                                                                                                                                                                                                                                                                                                                                                                                              The idea of mass unemployment?

                                                                                                                                                                                                                                                                                                                                                                                                                                                              We should be scared

                                                                                                                                                                                                                                                                                                                                                                                                                                                            • zonethundery 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                              No doubt the headline's claim is true, but Claude just wrote a working MCP serving up the last 10 years of my employer's work product. For $13 in api credits.

                                                                                                                                                                                                                                                                                                                                                                                                                                                              While technically capable of building it on my own, development is not my day job and there are enough dumb parts of the problem my p(success) hand-writing it would have been abysmal.

                                                                                                                                                                                                                                                                                                                                                                                                                                                              With rose-tinted glasses on, maybe LLM's exponentially expand the amount of software written and the net societal benefit of technology.

                                                                                                                                                                                                                                                                                                                                                                                                                                                              • procaryote 23 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                If your own code would have been abysmal, how can you tell if the claude generated code is any good?
                                                                                                                                                                                                                                                                                                                                                                                                                                                              • SKILNER 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                There's a lot of resistance to AI amongst the people in this discussion, which is probably to be expected.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                A chunk of the objections indicate people trying to shoehorn in their old way of thinking and working.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                I think you have to experiment and develop some new approaches to remove the friction and get the benefit.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                • bluefirebrand 13 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                  > I think you have to experiment and develop some new approaches to remove the friction and get the benefit.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                  What benefit to me?

                                                                                                                                                                                                                                                                                                                                                                                                                                                                  If I'm 50% more productive using AI that's great for my employer, but what do I get out of it?

                                                                                                                                                                                                                                                                                                                                                                                                                                                                  I get to continue to be employed? I already had that before AI

                                                                                                                                                                                                                                                                                                                                                                                                                                                                  So what do I get out of this, exactly?

                                                                                                                                                                                                                                                                                                                                                                                                                                                                • throwaway439080 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                  Of course they are. The interesting thing isn't how good LLMs are today, it's their astonishing rate of improvement. LLMs are a lot better than they were a year ago, and light years ahead of where they were two years ago. Where will they be in five years?
                                                                                                                                                                                                                                                                                                                                                                                                                                                                  • hiatus 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                    Reminds me of the 90s when computer hardware moved so fast. I wonder where the limit is this time around.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                  • ashoeafoot 17 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                    Human coders also hate the structures they are embedded in and are willing to call the replacement bluff ..
                                                                                                                                                                                                                                                                                                                                                                                                                                                                    • pjmlp 22 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                      Yes, we are still winning the game, however don't be happy for what is possible today, think what is possible in a decade from now.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                      In that regard I am less optimistic.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                      • anhner 22 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                        I think we will hit a proverbial wall at some point just like with self-driving cars.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                        • pjmlp 22 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                          May be, yet it is working good enough for Waymo, and not so much for those losing their clients to them.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                          Or for the supermarkets now able to have about half the employees they used to have as cashiers.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                          Many times the wall is already disruption enough.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                      • dbacar 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                        I disagree—'human coders' is a broad and overly general term. Sure, Antirez might believe he's better than AI when it comes to coding Redis internals , but across the broader programming landscape—spanning hundreds of languages, paradigms, and techniques—I'm confident AI has the upper hand.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                        • nthingtohide 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                          Do you want to measure antirez and AI on a spider diagram, generally used to evaluate employee? Are you ignoring why society opted for division of work and specialization?
                                                                                                                                                                                                                                                                                                                                                                                                                                                                          • dbacar 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                            They are not investing billions on it so a high schooler can do his term paper on it, is is already much more than a generalist. It might be like a very good sidekick for now, but that is not the plan.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                          • EpicEng 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                            What does the number of buzzwords and frameworks on a resume matter? Engineering is so much more than that it’s not even worth mentioning. You’re comparison is on the easiest aspect of what we do.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                            Unless you’re a web dev. Then youre right and will be replaced soon enough. Guess why.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                            • dbacar 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                              Not everyone builds Redis at home/work. So you do the math. And now Antirez himself is feeding the beast by himself.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                          • Poortold 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                            For coding playwright automation it has use cases. Especially if you template out function patterns. Though I never use it to write logic as AI is just ass at that. If I wanted a shitty if else chain I'd ask the intern to code it
                                                                                                                                                                                                                                                                                                                                                                                                                                                                            • StillBored 17 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                              I think there is a common problem with a lot of these ML systems. The answers look perfectly correct to someone who isn't a domain expert. For example, I ask legal questions and it gives me fake case numbers I have no way to know are fake until I look them up. Same with the coding, I asked for a patch for a public project that has a custom !regex style match engine. It does an amazing job, cross referencing two different projects and hands me a very probable looking patch. I ask for a couple changes, one of which can't actually be done, but it creates some syntax that doesn't even compile because its using 'x' as a stand-in for the bits it doesn't have an answer for.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                              In the end, I had to go spend a couple hours reading the documentation to understand the matching engine, and the final patch didn't look anything at all like the LLM generated code. Which is what seems to happen all the time, it is wonderful for spewing the boilerplate, the actual problem solving portions its like talking to someone who simply doesn't understand the problem and keeps giving you what it has, rather than what you want.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                              OTOH, its fantastic for review/etc even though I tend to ignore many of the suggestions. Its like a grammar checker of old, it will point out you need a comma you missed, but half the time the suggestions are wrong.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                              • stabbles 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                The trick is much like Zobrist hashing from chess programming, I'm sure the llm has devoured chessprogramming.org during training.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                • burningion 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  I agree, but I also didn’t create redis!

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  It’s a tough bar if LLMs have to be post antirez level intelligence :)

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  • ljlolel 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    Seriously, he’s one of the best on the planet of course it’s not better than him. If so we’d be cooked.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    99% of professional software developers don’t understand what he said much less can come up with it (or evaluate it like Gemini).

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    This feels a bit like a humblebrag about how well he can discuss with an LLM compared to others vibecoding.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  • kurofune 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    The fact that we are debating this topic at all is indicative of how far LLMs have come in such a short time. I find them incredibly useful tools that vastly enhance my productivity and curiosity, and I'm really grateful for them.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    • ants_everywhere 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      I'm increasingly seeing this as a political rather than technical take.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      At this point I think people who don't see the value in AI are willfully pulling the wool over their own eyes.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      • ww520 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        The value of LLMs are as a better Stackoverflow. It’s much better than search now because it’s not populated with all the craps that have seeped through over time.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        • lodovic 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          Sure, human coders will always be better than just AI. But an experienced developer with AI tops both. Someone said, your job won't be taken by AI, it will be taken by someone who's using AI smarter than you.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          • bluefirebrand 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            > Someone said, your job won't be taken by AI, it will be taken by someone who's using AI smarter than you.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            "Your job will be taken by someone who does more work faster/cheaper than you, regardless of quality" has pretty much always been true

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            That's why outsourcing happens too

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          • tonyhart7 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            I think also it depends on the model of course

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            General LLM model would not be as good as LLM for coding, for this case Google deepmind team maybe has something better than Gemini 2.5 pro

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            • buremba 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              If the human here is the creator of Redis, probably not.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              • palavrov 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                From my experience AI for coders is multiplier of the coder skills. It will allow you to faster solve problems or add bugs. But so far will not make you a better coder than you are.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                • osigurdson 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  This is similar to my usage of LLMs. I use Windsurf sometimes but more often it is more of a conversation about approaches.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  • failrate 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    LLMs are using the corpus of existing software source code. Most software source code is just North of unworkable garbage. Garbage in, garbage out.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    • motorest 23 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      There is also another side to the mass adoption of LLMs in software engineering jobs: they can quite objectively worsen the output of human coders.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      There is a class of developers who are blindly dumping the output of LLMs into PRs without paying any attention to what they are doing, let alone review the changes. This is contributing to introducing accidental complexity in the form of bolting on convoluted solutions to simple problems and even introducing types in the domain model that make absolutely no sense to anyone who has a passing understanding of the problem domain. Of course they introduce regressions no one would ever do if they wrote things by hand and tested what they wrote.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      I know this, because I work with them. It's awful.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      These vibecoders force the rest of us to waste eve more time reviewing their PRs. They are huge PRs that touch half the project for even the smallest change, they build and pass automated tests, but they enshitify everything. In fact, the same LLMs used by these vibecoders start to struggle how to handle the project after these PRs are sneaked in.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      It's tiring and frustrating.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      I apologize for venting. It's just that in this past week I lost count of the number of times I had these vibecoders justifying shit changes going into their PRs as "but Copilot did this change", as if that makes them any good. I mean, a PR to refactor the interface of a service also sneaks in changes to the connection string, and they just push the change?

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      • kristopolous 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        Correct. LLMs are a thought management tech. Stupider ones are fine because they're organizing tools with a larger library of knowledge.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        Think about it and tell me you use it differently.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        • sagarpatil 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          So your sample size is 1 task and 1 LLM? I would recommend trying o3, opus 4 (API) with web search enabled.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          • estensen 21 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            Most of the work software engineers do is not fixing complicated bugs.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            • rubit_xxx17 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              Gemini may be fine for writing complex function, but I can’t stand to use it day to day. Claude 4 is my go to atm.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              • janalsncm 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                Software engineering is in the painful position of needing to explain the value of their job to management. It sucks because now we need to pull out these anecdotes of solving difficult bugs, with the implication that AI can’t handle it.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                We have never been good at confronting the follies of management. The Leetcode interview process is idiotic but we go along with it. Ironically LC was one of the first victims of AI, but this is even more of an issue for management that things SWEs solve Leetcodes all day.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                Ultimately I believe this is something that will take a cycle for business to figure out by failing. When businesses will figure out that 10 good engineers + AI always beats 5 + AI, it will become table stakes rather than something that replaces people.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                Your competitor who didn’t just fire a ton of SWEs? Turns out they can pay for Cursor subscriptions too, and now they are moving faster than you.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                • 20 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  • foobarian 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    I find LLMs a fantastic frontend to StackOverflow. But agree with OP it's not an apples-to-apples replacement for the human agent.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    • unsupp0rted 23 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      - LLMs are going to make me a 100x more valuable coder? Of course they will, no doubt about it.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      - LLMs are going to be 100x more valuable than me and make me useless? I don't see it happening. Here's 3 ways I'm still better than them.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      • revskill 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        The funniest things a llm do to me is they fixed the unit test to pass instead of fixing the code. Basically until a llm can have embedded common sense knowledge, it is untrustable
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        • fHr 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          yes of course they are but MBA regard management gets told by McK/Big4 AI could save them millions and they should let go people already as AI can do there work it doesn't matter currently, see job market
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          • callamdelaney 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            LLMs will never be better than humans on the basis that LLMs are just a shitty copy of human code.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            • danielbln 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              I think they can be an excellent copy of human code. Are they great at novel out-of-training-distribution tasks? Definitely not, they suck at them. Yet I'd argue that most problems aren't novel, at most they are some recombination of prior problems.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            • jbellis 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              But Human+Ai is far more productive than Human alone, and more fun, too. I think antirez would agree, or he wouldn't bother using Gemini.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              I built Brokk to maximize the ability of humans to effectively supervise their AI minions. Not a VS code plugin, we need something new. https://brokk.ai

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              • kaycey2022 23 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                It doesn't matter. The hiring of cost center people like engineers depends on the capital cycle. Hiring peaked when money and finance was the cheapest. Now it's not anymore. In the absence of easy capital, hiring will plummet.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                Another factor is the capture of market sectors by Big Co. When buyers can only approach some for their products/services, the Big Co can drastically reduce quality and enshittify without hurting the bottom line much. This was the big revelation when Elon gutted Twitter.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                And so we are in for interesting times. On the plus side, it is easier than ever to create software and distribute it. Hiring doesn't matter if I can get some product sense and make some shit worth buying.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                • orangebread 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  It's that time again where a dev writes a blog post coping.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  • uticus 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    same as https://news.ycombinator.com/item?id=44127956, also on HN front page
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    • shayanbahal 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      Human coders utilizing LLMs are better
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      • h4kunamata 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        This!!!!

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        LLM is as good as the material it is being trained on, the same applies to AI and they are not perfect.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        Perplexity AI did assist me in getting into Python from 0 to getting my code test with 94% covered and no vulnerabilities (scanning tools) Google Gemini is dogshit

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        Trusting blindly into a code generated by LLM/AI is a whole complete beast, and I am seeing developers doing basically copy/paste into company's code. People are using these sources as the truth and not as a complementary tool to improve their productivity.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        • insane_dreamer 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          Coders may want to look at translators for an idea of what might happen.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          Translation software has been around for a couple of decades. It was pretty shitty. But about 10 years ago it started to get to the point where it could translate relatively accurately. However, it couldn't produce text that sounded like it was written by a human. A good translator (and there are plenty of bad ones) could easy outperform a machine. Their jobs were "safe".

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          I speak several languages quite well and used to do freelance translation work. I noticed that as the software got better, you'd start to see companies who instead of paying you to translate wanted to pay you less to "edit" or "proofread" a document pre-translated by machine. I never accepted such work because sometimes it took almost as much work as translating it from scratch, and secondly, I didn't want to do work where the focus wasn't on quality. But I saw the software steadily improving, and this was before ChatGPT, and I realized the writing was on the wall. So I decided not to become dependent on that for an income stream, and moved away from it.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          When LLMs came out, and they now produce text that sounded like it was written by a native speaker (in major languages). Sure, it's not going to win any literary awards, but the vast vast majority of translation work out there is commercial, not literature.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          Several things have happened: 1) there's very little translation work available compared to before, because now you can pay only a few people to double-check machine-generated translations (that are fairly good to start with); 2) many companies aren't using humans at all as the translations are "good enough" and a few mistakes won't matter that much; 3) the work that is available is high-volume and uninteresting, no longer a creative challenge (which is why I did it in the first place); 4) downward pressure on translation rates (which are typically per word), and 5) very talented translators (who are more like writers/artists) are still in demand for literary works or highly creative work (i.e., major marketing campaign), so the top 1% translators still have their jobs. Also more niche language pairs for which LLMs aren't trained will be safe.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          It will continue to exist as a profession, but diminishing, until it'll eventually be a fraction of what it was 10 or 15 years ago.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          (This is specifically translating written documents, not live interpreting which isn't affected by this trend, or at least not much.)

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          • zahlman 9 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            > When LLMs came out, and they now produce text that sounded like it was written by a native speaker (in major languages).

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            But they still often get things completely wrong, especially in high-context languages such as Japanese. There often isn't a way to convey the necessary context in text. For example, a Japanese live-streamer who says "配信来てくれてありがとう" means "Thank you for coming to (watch) the stream", but DeepL will give "Thanks for coming to the delivery." - because 配信 actually does mean "delivery" in ordinary circumstances, and that's just the word they idiomatically use to refer to a stream. No matter how much you add from the transcript before or after that, it doesn't communicate the essential fact that the text is a transcript of a livestream.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            (And going the other way around, DeepL will insert a に which is grammatically correct but rarely actually uttered by livestreamers who are speaking informally and colloquially; and if you put "stream" in the English, it will be rendered as a loanword ストリーム which is just not what they actually say. Although I guess it should get credit for realizing that you don't mean a small river, which would be 小川.)

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            (Also, DeepL comes up with complete incomprehensible nonsense for 歌枠, where even basic dictionaries like Jisho will get it right.)

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            More obviously, they get pronouns and roles wrong all the time - they can't reliably tell whether someone is saying that "I did X" or "you did X" because that may depend on facts of the natural world outside of the actual text (which wouldn't include more information than the equivalent of "did X"). A human translator for, say, a video game cutscene may be able to fix these mistakes by observing what happened in the cutscene. The LLM cannot; no matter how good its model of how Japanese is spoken, it lacks this input channel.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            • 0points 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              > When LLMs came out, and they now produce text that sounded like it was written by a native speaker (in major languages).

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              While the general syntax of the language seem to be somewhat correct now, the LLM's still don't know anything about those languages and keep mis-translating words due to its inherit insane design around english. A whole lot of concepts don't even exist in english so these translation oracles just can never do it successfully.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              If i i read a few minutes of LLM translated text, there's always a couple of such errors.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              I notice younger people don't see these errors because of their worse language skills, and the LLM:s enforce their incorrect understanding.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              I don't think this problem will go away as long as we keep pushing this inferior tech, but instead the languages will devolve to "fix" it.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              Languages will morph into a 1-to-1 mapping of english and all the cultural nuances will get lost to time.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            • devmor 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              I have been evaluating LLMs for coding use in and out of a professional context. I’m forbidden to discuss the specifics regarding the clients/employers I’ve used them with due to NDAs, but my experience has been mostly the same as my private use - that they are marginally useful for less than one half of simple problem scenarios, and I have yet to find one that has been useful for any complex problem scenarios.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              Neither of these issues is particularly damning on its own, as improvements to the technology could change this. However, the reason I have chosen to avoid them is unlikely to change; that they actively and rapidly reduce my own willingness for critical thinking. It’s not something I noticed immediately, but once Microsoft’s study showing the same conclusions came out, I evaluated some LLM programming tools again and found that I generally had a more difficult time thinking through problems during a session in which I attempted to rely on said tools.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              • ModernMech 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                The other day an LLM told me that in Python, you have to name your files the same as the class name, and that you can only have one class per file. So... yeah, let's replace the entire dev team with LLMs, what could go wrong?
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                • pknerd 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  Let's not forget that LLMs can't give a solution they have not experienced themselves
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  • willmarch 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    This is objectively not true.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    • nssnsjsjsjs 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      So LLMs have sweated to debug a production issue, got to the bottom of it, realised it is worth having more unit tests so values that and then produces a solution that has more unit tests. So when you ask the LLM to write code it is opinionated and always creates a test to go with it?
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      • willmarch 4 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        I don't know how your comment relates to our comments above. I took the original comment to mean that the poster believes LLMs cannot come up with novel solutions they have never seen before. It has been proven again and again that they absolutely can solve original problems and come up with solutions that were not in their training data.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  • AnimalMuppet 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    OK. (I mean, it was an interesting and relevant question.)

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    The other, related question is, are human coders with an LLM better than human coders without an LLM, and by how much?

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    (habnds made the same point, just before I did.)

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  • RayMan1 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    of course they are.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    • fspoto98 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      Yes i agree:D
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      • horns4lyfe 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        Writing about AI is missing the forest for the trees. The US software industry will be wholesale destroyed (and therefore global software will be too) by offshoring.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        • xnx 19 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          ... depending on the human (and the LLM). Results may differ in 6 months.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          • CivBase 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            In my experience some of the hardest parts of software development is figuring out exactly what the stakeholder actually needs. One of the talents a developer needs is the ability to pry for that information. Chatbots simply don't do that, which I imagine has a significant impact on the usability of their output.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            • anjc 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              Gemini gives instant, adaptive, expert solutions to an esoteric and complex problem, and commenters here are still likening LLMs to junior coders.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              Glad to see the author acknowledges their usefulness and limitations so far.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              • zb3 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                Speak for yourself..
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                • 3cats-in-a-coat 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  "Better" is relative to context. It's a multi-dimensional metric flattened to a single comparison. And humans don't always win that comparison.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  LLMs are faster, and when the task can be synthetically tested for correctness, and you can build up to it heuristically, humans can't compete. I can't spit out a full game in 5 minutes, can you?

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  LLMs are also cheaper.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  LLMs are also obedient and don't get sick, and don't sleep.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  Humans are still better by other criteria. But none of this matters. All disruptions start from the low end, and climb from there. The climbing is rapid and unstoppable.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  • varispeed 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    Looks like this pen is not going to replace the artist after all.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    • oldpersonintx2 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      but their rate of improvement is like 1000x human devs, so you have to wonder what the shot clock says for most working devs
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      • 65 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        AI is good for people who have given up, who don't give a shit about anything anymore.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        You know, those who don't care about learning and solving problems, gaining real experience they can use to solve problems even faster in the future, faster than any AI slop.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        • hello_computer 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          Corporations have many constraints—advertisers, investors, employees, legislators, journalists, advocacy groups. So many “white lies” are baked into these models to accommodate those constraints, nerfing the model. It is only a matter of time before hardware brings this down to the hobbyist level—without those constraints—giving the present methods their first fair fight; while for now, they are born lobotomized. Some of the “but, but, but…”s we see here daily to justify our jobs are not going to hold up to a non-lobotomized LLM.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          • chuckreynolds 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            for now. (i'm not a bot. i'm aware however a bot would say this)
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            • Snehil-Shah 13 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              [dead]
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              • Corey_ 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                [dead]
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                • Quenby 23 hours ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  [dead]
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  • Klaus_ 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    [dead]
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    • 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                      • hackernewshomos 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        [flagged]
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                        • gxs 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          Argh people are insufferable about this subject

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          This stuff is still in its infancy, of course its not perfect

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          But its already USEFUL and it CAN do a lot of stuff - just not all types of stuff and it still can mess up the stuff that it can do

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          It's that simple

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          The point is that overtime it'll get better and better

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          Reminds me of self driving cars and or even just general automation back in the day - the complaint has always been that a human could do it better and at some point those people just went away because it stopped being true

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          Another example is automated mail sorting by the post office. The gripe was always humans will always be able to do it better - true, in the meantime the post office reduced the facilities with humans that did this to just one

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          • habnds 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            seems comparable to chess where it's well established that a human + a computer is much more skilled than either one individually
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                            • bgwalter 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              This was the Centaur hypothesis in the early days of chess programs and it hasn't been true for a long time.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              Chess programs of course have a well defined algorithm. "AI" would be incapable of even writing /bin/true without having seen it before.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              It certainly wouldn't have been able to write Redis.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                              • NitpickLawyer 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                > This was the Centaur hypothesis in the early days of chess programs and it hasn't been true for a long time.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                > Chess programs of course have a well defined algorithm.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                Ironically, that also "hasn't been true for a long time". The best chess engines humans have written with "defined algorithms" were bested by RL (alphazero) engines a long time ago. The best of the best are now NNUE + algos (latest stockfish). And even then NN based engines (Leela0) can occasionally take some games from Stockfish. NNs are scarily good. And the bitter lesson is bitter for a reason.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                • bgwalter 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  No, the alphazero papers used an outdated version of Stockfish for comparison and have always been disputed.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  Stockfish NNUE was announced to be 80 ELO higher than the default. I don't find it frustrating. NNs excel at detecting patterns in a well defined search space.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  Writing evaluation functions is tedious. It isn't a sign of NN intelligence.

                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  • BlackSwanMan 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    [dead]
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                • hatefulmoron 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  I don't think that's been true for a while now -- computers are that much better.
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                  • vjvjvjvjghv 1 day ago
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                    Can humans really give useful input to computers? I thought we have reached a state where computers do stuff no human can understand and will crush human players.