View all comments ➔
  • I don’t share your concerns about the profession. Even supposing for a moment that LLMs did deliver on the promise of making 1 human as productive as 5 humans were previously, that isn’t how for-profit industry has traditionally incorporated productivity gains. Instead, you’ll just have 5 humans producing 25x output. If code generation becomes less of a bottleneck (which it has been doing for decades as frameworks and tooling have matured) there will simply be more code in the world that the code wranglers will have to wrangle. Maybe if LLMs get good enough at generating usable code (still a big if for most non-trivial jobs), some people who previously focused on low-level coding concerns will be able to specialize in higher-level concerns like directing an LLM, while some people will still be writing the low-level inputs for the LLMs, sort of like how you can write applications today without needing to know the specific ins and outs of the instruction set for your CPU. I’m doubtful that that’s around the corner, but who knows. But whatever the tools we have are capable of, the output will be bounded by the abilities of the people who operate the tools, and if you have good tools that are easily replicated, as software tools are, there’s no reason not to try and maximize your output by having as many people as you can afford and cranking out as much product as you can.