"We're nowhere near the point where an AI can do your job, but we're well past the point where your boss can be suckered into firing you and replacing you with a bot that fails at doing your job." (Pl
I couldn't have said it better myself. All of these companies firing people are doing it because they want to fire people. AI is just a convenient excuse. It's RTO all over again.
Well, there's one good thing that will come out of this: these kinds of idiotic moves will help us figure out which companies have the right kinds of management at the top, and which ones don't have any clue whatsoever.
Of course, it will come with the working class bearing the brunt of their bad decisions, but that has always been the case unfortunately. Business as usual...
My dad accidentally bought 2 chargers a few weeks ago. He tried refunding it, and what do you know, the company fired their support staff and replaced them with chat bot AIs. Anyway, the AI looked at his order and helpfully told him he had already returned the product and it had already been refunded so there was nothing left to do.
It kept doing this to him every time he tried to return the second charger, and there wasn't any other way to contact them on their site, so he ended-up leaving a 1-star review on their site complaining about the issue. Then an actual person contacted him to get it sorted-out.
If you convince a human to give you the password, that’s called social engineering. If you convince an AI send you free stuff, what kind of engineering is that?
they've already trained on stack overflow, if you want an AI that recommends a complete change of technology stack in preference to solving the problem at hand.
I don't know if it can also insult you for wanting to solve the problem?
I just hired an employee who managed things as I was on a leave of absence and things went fine without me. Getting a little pushback from MY boss now because you know, this cheaper employee just did my job.
Of course, he did it for a portion of the year after I managed to complete 3 major projects early so he didn't have to deal with them and I left a month-by-month explanation of how to do everything he had to do. And the one problem that popped up went unresolved until I returned.
That is basically the situation with AI too. You still need someone knowledgeable in the loop to describe the things it needs to do, and handle exceptions.
I had to rewrite our entire scheduling system at work to use Outlook instead of Google Calendar. The guy who wrote the Google Calendar scheduling system made it so unmaintainable that it was faster to just rewrite the entire thing from scratch (1000+ line lambda function with almost 0 abstraction).
At least 90% of what I wrote is just exception handling. There's ~15 different 4xx/5xx errors that can be returned for each endpoint, but only 1 or 2 200 responses.
"You're 100% right, you should promote me so I can train more people to be able to run things. Things falling apart whenever someone goes away is a key sign of a bad leader, not a good one. I think I've demonstrated that I've managed this department into where it can function smoothly without me needing to put full time into it and I'd do well with an opportunity to move some other things in the company forward."
"Hey, unrelated question, what's your boss's contact info?"
The issue is that one specialist can oversee how many AI job holders? How many jobs are we getting rid of that will supposedly be bolstered by the new jobs created in the fields of manufacturing and AI hosting/training?
Now how many of those jobs have or will actually materialize?
That's my issue, it'll just get placed on IT's shoulders without any additional support.
This has been my general worry: the tech is not good enough, but it looks convincing to people with no time. People don’t understand you need at least an expert to process the output, and likely a pretty smart person for the inputs. It’s “trust but verify”, like working with a really smart parrot.
AI can't replace a person just yet, but it can easily augment a persons output so only a quarter as many workers are needed. Yes, this has happened throughout history, but AI is poised to displace workers across almost every industry.
I generally agree. It’ll be interesting what happens with models, the datasets behind them (particularly copyright claims), and more localized AI models. There have been tasks where AI greatly helped and sped me up, particularly around quick python scripts to solve a rote problem, along with early / rough documentation.
However, using this output as justification to shed head count is questionable for me because of the further business impacts (succession planning, tribal knowledge, human discussion around creative efforts).
If someone is laying people off specifically to gap fill with AI, they are missing the forest for the trees. Morale impacts whether people want to work somewhere, and I’ve been fortunate enough to enjoy the company of 95% of the people I’ve worked alongside. If our company shed major head count in favor of AI, I would probably have one foot in and one foot out.
Ya what’s your point? Are you saying that the invention of computers didn’t displace a lot of jobs?
If you’re saying that AI is going to disrupt the market and displace a large number of jobs just like computers did then you’re 100% right.
Nothing is finite. AI isn’t going to be the first or last thing to shake up the world.
Eventually your skills are going to become less valued and you’ll have no choice but to retool. Either you figure out how to retool or you get left behind.
The Segway was sold at the time as a "revolution in transportation". Ditto for the Hyperloop, by the way.
Then there's the Theranos' "revolutionary" Tech as mentioned in the article.
And don't get me started on how Pets.com (which famously went bust in the .Net Crash) was also revolutionary.
There are way more situations of Tech snakeoil being peddled to the masses as "revolutionary" than there are of trully revolutionary Tech being sold as such (I can only think of 3 trully revolutionary Technologies in the last half a century: the Personal Computer, the Internet and Smartphones, and one of those was a surprise revolution whilst another was only hyped after it was actually starting to show revolutionary results, and only smartphones were hyped from the start), so the logical default position for anybody but the snakeoil salesmen trying to swindle the masses is to suspect of tall tales in Tech, the taller the tale the greater the suspicion.
(I keenly remember the early days of the Internet, and it was incredibly low-key compared to the present day hype-spectables for bullshit that never even works).
Believing such claims by default is either incredibly naive or the product of a vested financial interest in getting people to put money into it.
Well this isn’t quite true, automation and computers have replaced many jobs. They just haven’t been skilled labour.
Now AI is catching up with skilled labour, whether it’s CNNs for loss prevention, LSTM/1DCNN for anomaly detection in Time Series (e.g. biosignal, finance) or more recently llms explaining and adapting code.
In one way or another, that work, at least in part, would have been done by a person, even if it’s an intern for example.
That's where the majority of jobs are that computers and automation "took".
Large companies needed hundreds of accountants to do what a dozen can do now. Same goes for developers. Or biologists. Or architects. Or whatever else.
Don't get this mistaken with the fact that a lot of people know their job is bullshit. People like to sit there thinking 'an AI can't take my job' while at the same time thinking 'a monkey can do this job it's such a waste of time'
My job isn't bullshit, but management has no concept of the true amount of time it takes to do my job. Depending on projects I can go from 2 hours of work a week up to around 60 hours of work a week. With the majority of weeks being under 40 hours. And yet management somehow thinks that they're giving me 8 hours of work to do every day despite them regularly being the blocker to new work.
Middle management only cares that it looks like you're working (and thus their job of supervising you doing the work is necessary (apparently)), and upper management only cares that you're making them money.
No kidding huh. I'm glad we're finally having the discussion about AI and what that means for employment and things like UBI, but this is far from actual AI.
Are we actually having that discussion? All I see is people concerned about being replaced by AI asking to put constraints on it and people wanting to replace their employees with AI ignoring them. No one will get UBI or anything like it until the latter group is more concerned about a mob with pitchforks showing up at their door than they are with giving their stock price a small bump.
What really concerns me is that the modern-day version of mobs with pitchforks seems to be fascism, because fascists have learned how to create the mob and harness it for their own purposes.
This is going to be like the self checkout lanes at the store but for creative jobs.
At the end of the day, a company will be able to produce the same output with fewer people. Some stuff will be of lower quality, just like sometimes people spend time on Lemmy and then phone in some crappy work.
But all the self checkouts around me have been ripped out and replaced with cashiers again. For some reason having someone paid 30 cents over minimum wage watching a bunch of people shop on the honor system with a bunch of finicky machines didn't work.
You might just live in crime central, that's not happening everywhere. Probably on an individual cost of cashier versus lost stock basis with each location.
I sense a lot of dislike for self-checkouts and wonder why they are done so poorly where other people live. In Holland they are fine. You can self-scan with either a portable scanner or your phone while you shop, or scan the items at checkout. I've literally never had to wait for a free machine, and they work well. Some people use the registers with humans scanning for you, and they seem fine too.
There's a fairly big shoppers drugmart I went to in north Vancouver and they only have self checkout and one customer service desk. Most people still hate them from the moaning and groaning you hear while waiting in line though.
Loads of good points in that video, thanks for posting. The only argument I don't really agree with is about bias. She's implying here that a human decision maker would be less biased than the AI model. I'm not convinced by that because the training data is just a statistical record of human bias. So as long as the training data is well selected for your problem, it should be a good predictior for the likelihood of bias in your human decision maker.
I think with a human operator, we can be proactive. A person can be informed of bias, learn to recognize it, and even attempt to compensate for their own.
An AI model is working off of aggregate past data that we already know is biased. There is currently no proactive anti bias training that can be done to a AI model without massively altering the dataset, which, at some level of alteration, loses its value as true to life data.
Secondly, AI is a black box. we can’t see inner the workings of the model and determine what types of associations it is making to come to its result. So we don’t even know what part of the dataset would need to be altered to address the bias.
Lastly, the default assumption by end users will be, unless there are glaring defects, that any individual result is correct and unbiased, because “AI was made by smart people and data, and data doesn’t lie.” And because interrogating and validating the result defeats the whole purpose of using AI to cut out those steps of the process.
Everything I read was well worded and well reasoned. However, it seems like either my ADD got the better of me, or that was the article that has no end. I didn't really realize before that my attention has a word count, but I now know that it is less than this article.