For some background, I originally wanted to break into programming back when I was in college but drifted more into desktop tech support and now systems administration. SysAdmin work is draining me, though, and I want to pick back up programming and see if I can make a career out of it, but industry seems like it could be moving in a direction to rely on AI for coding. Everything I’ve heard has said AI is not there yet, but if it’s looking like it hits a point where it reaches an ability to fully automate coding, should I even bother? Am I going to be obsolete after a year? Five years?

  • VoterFrog@kbin.social
    link
    fedilink
    arrow-up
    12
    ·
    10 months ago

    One thing that is somewhat unique about software engineering is that a large part of it is dedicated to making itself more efficient and always has been. From programming languages, protocols, frameworks, and services, all of it has made programmers thousands of times more efficient than the guys who used to punch holes into cards to program the computer.

    Nothing has infinite demand, clearly, but the question is more whether or not we’re anywhere near the peak, such that more efficiency will result in an overall decrease in employment. So far, the answer has been no. The industry has only grown as it’s become more efficient.

    I still think the answer is no. There’s far more of our lives and the way people do business that can be automated as the cost of doing so is reduced. I don’t think we’re close to any kind of maximum saturation of tech.

    • AggressivelyPassive@feddit.de
      link
      fedilink
      arrow-up
      1
      arrow-down
      6
      ·
      10 months ago

      Here again, I think, is a somewhat tech-centric view on economics.

      There is only a finite amount of automation demand, simply because human labor exists.

      Inside of our tech bubble, automation simply means more “functionality” per person per time unit. What took 10 devs a year yesterday can be done by 5 people in 6 months today. That’s all five and dandy, but at some point, software clashes with the hard reality of physics. Software doesn’t produce anything, it’s often just an enabler for physical production. Lube, or grease.

      Now, that production obviously can be automated tremendously as well, but with diminishing returns. Each generation of automation is harder than the one before. And each generation has to compete with a guy in Vietnam/Kenia/Mexico. And each generation also has to compete with its own costs.

      Why do you think, chips are so incredibly expensive lately? RND costs are going through the roof, production equipment is getting harder and harder to produce, and due to the time pressure, you have to squeeze out as much money as possible out of your equipment. So prices go up. But that can’t go on forever, at Stone point the customers can’t justify/afford the expense. So there’s a kind of feedback loop.

      • VoterFrog@kbin.social
        link
        fedilink
        arrow-up
        7
        ·
        10 months ago

        Yes, what I’m saying is that lower costs for software, which AI will help with, will make software more competitive against human production labor. The standard assumption is that if software companies can reduce the cost of producing software, they’ll start firing programmers but the entire history of software engineering has shown us that that’s not true as long as the lower cost opens up new economic opportunities for software users, thus increasing demand.

        That pattern stops only when there are no economic opportunities to be unlocked. The only way I think that happens is when automation has become so prevalent that further advancement has minimal impact. I don’t think we’re there yet. Labor costs are still huge and automation is still relatively primitive.