I think this article misses the forest for the trees. The real “evil” here is capitalism, not AI. Capitalism encourages a race towards optimality with no care to what happens to workers. Just like the invention of the car put carriage makers out of business, so AI will be used to by company owners to cut costs if it serves them. It has been like this for over a 100 years, AI is just the latest technology to come along. I’m old enough to remember tons of these same doom and gloom articles about workers losing their jobs when the internet revolution hit in the late 90s. And probably many people did lose jobs, but many new jobs were created too.
This person explains all her failures: insted of adopting and using chatgpt herself, reducing price and finding more clients she did nothing.
She was writing most boring pieces of text than no one is reading (corporate blog posts and spam emails).
Refused to learn new things which would keep her in position.
Yes, some jobs disappear other appear. I believe that 90+% of today’s jobs didn’t exist even 50 years ago. Especially not without will to learn new ways of doing things. Imagine farmer with knowledge of 100 years ago. Or hotel front desk worker without computer and telephone.
For mid-level writers, which she was, using AI doesn’t work. The few remaining clients you have specifically don’t want AI to be used. So you either lie and deceive them or you stay away from AI.
And using AI to lower prices and finding new clients also doesn’t work. Writers are already competing against writers from nations with much lower cost of living who do the same work for a fraction of the cost. But the big advantage that domestic writers had was a batter grasp of the language and culture. These advantages are mostly lost if you start using AI. So if that’s your business plan you are in a race to the bottom. It’s not sustainable and you will be out of a job in maybe 3-5 years.
Thank you for good insight, I was just thinking if all here clients are satisfied with AI, then
The few remaining clients you have specifically don’t want AI to be used.
Is not completely true.
At the end of the day if an AI can do the job to an acceptable standard a human doesn’t need to be doing it.
As you say it’s happened to countless industries and will continue to happen.
Except that the ‘AI’ is fed by the work of actual humans, and as time goes on, they will be trained more and more on the imperfect output of other AIs, which will eventually result in their output being total bizarre crap. Meanwhile, humans stopped training at whatever task since they couldn’t be paid to do it anymore, so there’s no new human material.
Wow you clearly have a very good understanding of economy and of how our species has been evolving in the lady hundreds of years.
You are the same as the people who didn’t want to lose their jobs in the coal mines and in the oil rigs. BeCauSE wE wON’t HavE JOooOBs…instead of diving into the ones created by renewables.
You prefer to be in stable shity conditions then in an turbulent way to improvement
I’m really having a hard time thinking about what jobs this would create though. I get the internet thing, as people needed to create and maintain all aspects of it, so jobs are created. If some massive corporation makes the AI and all others use the AI, there’s no real infrastructure. The same IT guys maintain the systems at AI corp. What’s left to be done with it/for it by “common folk?”
There are plenty of companies out there (and growing daily) who want to do AI in house, and can’t (or don’t want) to send their data to some monolithic, blackbox company which has no transparency. The finance industry, for example, cannot send any data to some third party company like OpenAI (ChatGPT) for compliance reasons, so they are building teams to develop and maintain their own AI models in-house (SFT, RLHF, MLOps, etc).
There are lots of jobs being created in AI daily, and they’re generally high paying, but they’re also very highly skilled, so it’s difficult to retrain into them unless you already have a strong math and programming background. And the number of jobs being created is definitely a lot, lot less than the potential number of jobs lost to AI, but this may change over time.
Despite what the pseudo-intellectuals will tell you, ChatGPT is not some all powerful do everything AI. Say you want to use GPT to create your own chatbot for your company to give company specific info to people at your company, you cant just take existing chat GPT and ask it “how do I connect to the wifi” or “is the office closed on monday” you need an in-house team of people to provide properly indexed information, train and test the bot, update it, handle error reports, etc.
AI is not magic, its literally just an advanced computer script, and if your job can be replaced by an AI then it could have been replaced by a regular computer script or program, there just wasnt enough buzzwords and media hype to convince your boss to do it.
Not optimality. Maximum profit. Very different from any definition of optimal I would personally use.
Well, in business school they teach you that running a company is an exercise in maximising profits as a constrained optimisation problem, so optimality for a classical company (not one of those weird startups that doesn’t make money for 10+ years) almost always is maximum profit.
I honestly can’t tell if you’re being serious. The ‘evil’ is the same force that replaced carriages with cars? The world would be better if carriage-making was still a critical profession?