US experts who work in artificial intelligence fields seem to have a much rosier outlook on AI than the rest of us.
In a survey comparing views of a nationally representative sample (5,410) of the general public to a sample of 1,013 AI experts, the Pew Research Center found that “experts are far more positive and enthusiastic about AI than the public” and “far more likely than Americans overall to believe AI will have a very or somewhat positive impact on the United States over the next 20 years” (56 percent vs. 17 percent). And perhaps most glaringly, 76 percent of experts believe these technologies will benefit them personally rather than harm them (15 percent).
The public does not share this confidence. Only about 11 percent of the public says that “they are more excited than concerned about the increased use of AI in daily life.” They’re much more likely (51 percent) to say they’re more concerned than excited, whereas only 15 percent of experts shared that pessimism. Unlike the majority of experts, just 24 percent of the public thinks AI will be good for them, whereas nearly half the public anticipates they will be personally harmed by AI.
If it was marketed and used for what it’s actually good at this wouldn’t be an issue. We shouldn’t be using it to replace artists, writers, musicians, teachers, programmers, and actors. It should be used as a tool to make those people’s jobs easier and achieve better results. I understand its uses and that it’s not a useless technology. The problem is that capitalism and greedy CEOs are ruining the technology by trying to replace everyone but themselves so they can maximize profits.
The natural outcome of making jobs easier in a profit driven business model is to either add more work or reduce the number of workers.
This is exactly the result. No matter how advanced AI gets, unless the singularity is realized, we will be no closer to some kind of 8-hour workweek utopia. These AI Silicon Valley fanatics are the same ones saying that basic social welfare programs are naive and un-implementable - so why would they suddenly change their entire perspective on life?
we will be no closer to some kind of 8-hour workweek utopia.
If you haven’t read this, it’s short and worth the time. The short work week utopia is one of two possible outcomes imagined: https://marshallbrain.com/manna1
Mayne pedantic, but:
Everyone seems to think CEOs are the problem. They are not. They report to and get broad instruction from the board. The board can fire the CEO. If you got rid of a CEO, the board will just hire a replacement.
CEOs are the figurehead, they are virtually bound by law to act sociopathically - in the interests of their shareholders over everyone else. Carl Icahn also has an interesting take on a particularly upsetting emergent property of our system of CEO selection: https://dealbreaker.com/2007/10/icahn-explains-why-are-there-so-many-idiots-running-shit
And if you get rid of the board, the shareholders will appointment a new one. If you somehow get rid of all the shareholders, like-minded people will slot themselves into those positions.
The problems are systemic, not individual.
AI has it’s place, but they need to stop trying to shoehorn it into anything and everything. It’s the new “internet of things” cramming of internet connectivity into shit that doesn’t need it.
Now your smart fridge can propose unpalatable recipes. Woo fucking hoo.
You’re saying the addition of Copilot into MS Paint is anything short of revolutionary? You heretic.
Butlerian Jihad
The first thing seen at the top of WhatsApp now is an AI query bar. Who the fuck needs anything related to AI on WhatsApp?
Android Messages and Facebook Messenger also pushed in AI as ‘something you can chat with’
I’m not here to talk to your fucking chatbot I’m here to talk to my friends and family.
It’s easier to up-sell and cross-sell if you’re talking to an AI.
Right?! It’s literally just a messenger, honestly, all I expect from it is that it’s an easy and reliable way of sending messages to my contacts. Anything else is questionable.
There are exactly 0 good reasons to use whatsapp anyways…
Yes, there are. You just have to live in one of the many many countries in the world where the overwhelming majority of the population uses whatsapp as their communication app. Like my country. Where not only friends and family, but also businesses and government entities use WhatsApp as their messaging app. I have at least a couple hundred reasons to use WhatsApp, including all my friends, all my family members, and all my clients at work. Do I like it? Not really. Do I have a choice? No. Just like I don’t have a choice on not using gmail, because that’s the email provider that the company I work for decided to go with.
SMS works fine in any country.
And you can isolate your business requirements from your personal life.
Who the fuck needs
anything related to AI onWhatsApp?
For once, most Americans are right.
I do as a software engineer. The fad will collapse. Software engineering hiring will increase but the pipeline of new engineers will is dry because no one wants to enter the career with companies hanging ai over everyone’s heads. Basic supply and demand says my skillset will become more valuable.
Someone will need to clean up the ai slop. I’ve already had similar pistons where I was brought into clean up code bases that failed being outsourced.
Ai is simply the next iteration. The problem is always the same business doesn’t know what they really want and need and have no ability to assess what has been delivered.
A complete random story but, I’m on the AI team at my company. However, I do infrastructure/application rather than the AI stuff. First off, I had to convince my company to move our data scientist to this team. They had him doing DevOps work (complete mismanagement of resources). Also, the work I was doing was SO unsatisfying with AI. We weren’t tweaking any models. We were just shoving shit to ChatGPT. Now it was be interesting if you’re doing RAG stuff maybe or other things. However, I was “crafting” my prompt and I could not give a shit less about writing a perfect prompt. I’m typically used to coding what I want but I had to find out how to write it properly: “please don’t format it like X”. Like I wasn’t using AI to write code, it was a service endpoint.
During lunch with the AI team, they keep saying things like “we only have 10 years left at most”. I was like, “but if you have AI spit out this code, if something goes wrong … don’t you need us to look into it?” they were like, “yeah but what if it can tell you exactly what the code is doing”. I’m like, “but who’s going to understand what it’s saying …?” “no, it can explain the type of problem to anyone”.
I said, I feel like I’m talking to a libertarian right now. Every response seems to be some solution that doesn’t exist.
If it walks and quacks like a speculative bubble…
I’m working in an organization that has been exploring LLMs for quite a while now, and at least on the surface, it looks like we might have some use cases where AI could prove useful. But so far, in terms of concrete results, we’ve gotten bupkis.
And most firms I’ve encountered don’t even have potential uses, they’re just doing buzzword engineering. I’d say it’s more like the “put blockchain into everything” fad than like outsourcing, which was a bad idea for entirely different reasons.
I’m not saying AI will never have uses. But as it’s currently implemented, I’ve seen no use of it that makes a compelling business case.
AI can look at a bajillion examples of code and spit out its own derivative impersonation of that code.
AI isn’t good at doing a lot of other things software engineers actually do. It isn’t very good at attending meetings, gathering requirements, managing projects, writing documentation for highly-industry-specific products and features that have never existed before, working user tickets, etc.
I work in an environment where we’re dealing with high volumes of data, but not like a few meg each for millions of users. More like a few hundred TB fed into multiple pipelines for different kinds of analysis and reduction.
There’s a shit-ton of prior art for how to scale up relatively simple web apps to support mass adoption. But there’s next to nothing about how do to what we do, because hardly anyone does. So look ma, no training set!
I too am a developer and I am sure you will agree that while the overall intelligence of models continues to rise, without a concerted focus on enhancing logic, the promise of AGI likely will remain elusive. AI cannot really develop without the logic being dramatically improved, yet logic is rather stagnant even in the latest reasoning models when it comes to coding at least.
I would argue that if we had much better logic with all other metrics being the same, we would have AGI now and developer jobs would be at risk. Given the lack of discussion about the logic gaps, I do not foresee AGI arriving anytime soon even with bigger a bigger models coming.
If we had AGI, the number of jobs that would be at risk would be enormous. But these LLMs aren’t it.
They are language models and until someone can replace that second L with Logic, no amount of layering is going to get us there.
Those layers are basically all the previous AI techniques laid over the top of an LLM but anyone that has a basic understanding of languages can tell you how illogical they are.
Agreed. I would add that not only would job loss be enormous, but many corporations are suddenly going to be competing with individuals armed with the same AI.
I mean, it hasn’t thus far.
No surprise there. We just went through how blockchain is going to drastically help our lives in some unspecified future.
All it took was for us to destroy our economy using it to figure that out!
New technologies are not the issue. The problem is billionaires will fuck it up because they can’t control their insatiable fucking greed.
exactly. we could very well work less hours with the same pay. we wouldnt be as depressed and angry as we are right now.
we just have to overthrow, what, like 2000 people in a given country?
Experts are working from their perspective, which involves being employed to know the details of how the AI works and the potential benefits. They are invested in it being successful as well, since they spent the time gaining that expertise. I would guess a number of them work in fields that are not easily visible to the public, and use AI systems in ways the public never will because they are focused on things like pattern recognition on virii or idendifying locations to excavate for archeology that always end with a human verifying the results. They use AI as a tool and see the indirect benefits.
The general public’s experience is being told AI is a magic box that will be smarter than the average person, has made some flashy images and sounds more like a person than previous automated voice things. They see it spit out a bunch of incorrect or incoherent answers, because they are using it the way it was promoted, as actually intelligent. They also see this unreliable tech being jammed into things that worked previously, and the negative outcome of the hype not meeting the promises. They reject it because how it is being pushed onto the public is not meeting their expectations based on advertising.
That is before the public is being told that AI will drive people out of their jobs, which is doubly insulting when it does a shitty job of replacing people. It is a tool, not a replacement.
The problem could be that, with all the advancements in technology just since 1970, all the medical advancements, all the added efficiencies at home and in the workplace, the immediate knowledge-availability of the internet, all the modern conveniences, and the ability to maintain distant relationships through social media, most of our lives haven’t really improved.
We are more rushed and harried than ever, life expectancy (in the US) has decreased, we’ve gone from 1 working adult in most families to 2 working adults (with more than 1 job each), income has gone down. Recreation has moved from wholesome outdoor activities to an obese population glued to various screens and gaming systems.
The “promise of the future” through technological advancement, has been a pretty big letdown. What’s AI going to bring? More loss of meaningful work? When will technology bring fewer working hours and more income - at the same time? When will technology solve hunger, famine, homelessness, mental health issues, and when will it start cleaning my freaking house and making me dinner?
When all the jobs are gone, how beneficial will our overlords be, when it comes to universal basic income? Most of the time, it seems that more bad comes from out advancements than good. It’s not that the advancements aren’t good, it’s that they’re immediately turned to wartime use considerations and profiteering for a very few.
Just about every major advance in technology like this enhanced the power of the capitalists who owned it and took power away from the workers who were displaced.
This is like asking tobacco farmers what their thoughts are on smoking.
Al Gore’s family thought that the political tide was turning against it, so they gave up tobacco farming in the late 1980s - and focused on politics.
More like asking the slaves about productivity advances in slavery. “Nothing good will come of this”.
See also; the cotton gin.
The cotton gin has been used as an argument for why slavery finally became unacceptable. Until then society “needed” slaves to do the work, but with the cotton gin and other automations the costs of slavery started becoming higher than the value.
My understanding is that the cotton gin led to more slavery as cotton production became more profitable. The machine could process cotton but not pick it, so more hands were needed for field work.
Wiki:
The invention of the cotton gin caused massive growth in the production of cotton in the United States, concentrated mostly in the South. Cotton production expanded from 750,000 bales in 1830 to 2.85 million bales in 1850. As a result, the region became even more dependent on plantations that used black slave labor, with plantation agriculture becoming the largest sector of its economy.[35] While it took a single laborer about ten hours to separate a single pound of fiber from the seeds, a team of two or three slaves using a cotton gin could produce around fifty pounds of cotton in just one day.[36] The number of slaves rose in concert with the increase in cotton production, increasing from around 700,000 in 1790 to around 3.2 million in 1850."
That is also true, the cotton gin wasn’t the total economic turning point, and the Civil War pre-dated automation’s economic turning of the corner against some economic measures of slavery’s cost, but slavery has very difficult to quantify costs, it was an entrenched lifestyle much more than a pool of day labor hanging out at Home Depot waiting for work, where both employers and employees could easily change their ways on very short notice.
After the Civil War it looks like “free person” cotton harvesting labor persisted until about 1926 - that could have changed earlier, but farm owners needed a kick in the butt to figure out how to improve:
https://www.printmag.com/creative-voices/lessons-from-cottons-slow-motion-robot-takeover/
Because it won’t. So far it’s only been used to replace people and cut costs. If it were used for what it was actually intended for then it’d be a different story.
Replacing people is a good thing. It means less people do more work. It means progress. It means products and services will get cheaper and more available. The fact that people are being replaced means that AI actually has tremendous value for our society.
Great for people getting fired or finding that now the jobs they used to have that were middle class are now lower class pay or obsolete. They will be so delighted at the progress despite their salaries and employment benefits and opportunities falling.
And it’s so nice that AI is most concentrated in the hands of billionaires who are oh so generous with improving living standards of the commoners. Wonderful.
This is collateral damage of societal progress. This is a phenomenon as old as humanity. You can’t fight it. And it has brought us to where we are now. From cavemen to space explorers.
Removed by mod
What makes you think that? You can’t just go around and insult people personally without elaborating on the reason.
Yeah, yeah, omelettes…eggs… heard it all before.
Which are separate things from people’s ability to financially support themselves.
People can have smartphones and tech the past didn’t have, but be increasingly worse off financially and unable to afford housing.
And you aren’t a space explorer.
I’m not arguing about whether innovation is cool. It is.
I however strongly disagree with your claim that people being replaced is good. That assumes society is being guided with altruism as a cornerstone of motivation to create some Star Trek future to free up people to pursue their interests, but that’s a fantasy. Innovation is simply innovation. It’s not about whether people’s lives will be improved. It doesn’t care.
World can be the most technologically advanced its ever been with space travel for the masses and still be a totalitarian dystopia. People could be poorer than ever and become corpo slaves, but it would fit under the defition of societal progress because of innovation.
Great for people getting fired or finding that now the jobs they used to have that were middle class are now lower class pay or obsolete. They will be so delighted at the progress despite their salaries and employment benefits and opportunities falling.
This shouldn’t come as a surprise. Everyone who’s suprised by that is either not educated how economy works or how societal progress works. There are always winners and losers but society makes net-positive progress as a whole.
I have no empathy for people losing their jobs. Even if I lose my job, I accept it. It’s just life. Humanity is a really big machine of many gears. Some gears replace others to make the machine run more efficient.
And it’s so nice that AI is most concentrated in the hands of billionaires who are oh so generous with improving living standards of the commoners. Wonderful.
This is just a sad excuse I’m hearing all the time. The moment society gets intense and chang is about to happen, a purpetrator needs to be found. But most people don’t realize that the people at the top change all the time when the economy changes. They die aswell. It’s a dynamic system. And there is no one purpetrator in a dynamic system. The only purpetrator is progress. And progress is like entropy. It always find its way and you cannot stop it. Those who attempt to stop it instead of adapting to it will be crushed.
I have no empathy
for people losing their jobsFTFY
I trust you’ve volunteered for it to replace you then. It being so beneficial to society, and all.
It means less people do more work.
And then those people no longer working… do what, exactly? Fewer well-paying jobs, same number of people, increasing costs. Math not working out here.
The fact that people are being replaced means that AI actually has tremendous value for our society.
Oh, it has value. Just not for society (it could that’s the sad part). For very specific people though, yeah, value. Just got to step on all the little people along the way, like we’ve always done, eh?
Yeah, rather than volunteering its more likely you lack a basic characteristic of humanity some of like to refer to as “empathy” instead. And if – giving you the benefit of the doubt – you’re just a troll… well, my statement stands.
I trust you’ve volunteered for it to replace you then. It being so beneficial to society, and all.
Yes. If I get replaced by something more efficient I accept that. I am no longer worth the position of my job. I will look for something else and try to find ways to apply some of my skillsets in other ways. I may do some further training and education, or just accept a lower paying job if that’s not possible.
And then those people no longer working… do what, exactly? Fewer well-paying jobs, same number of people, increasing costs. Math not working out here.
Can you elaborate? I don’t quiet understand what you mean by that. The people who no longer work need to find something else. There will remain only a fraction that can never find another job again. And that fraction is offset by the increased productivity of society.
Oh, it has value. Just not for society (it could that’s the sad part). For very specific people though, yeah, value. Just got to step on all the little people along the way, like we’ve always done, eh?
Can you specify “specific”? What little people? If you use very vague terminology like that you should back it up with some arguments. I personally see no reason why AI would disadvantage working people any more than the sewing machine did back in the day. Besides, when you think about it you’ll find that defining the terms you used is actually quiet difficult in a rapidly changing economy when you don’t know to whom these terms might apply to in the end.
I have a feeling you’re not actually thinking this through, or at least doing it on a very emotional level. This will not help you adapt to the changing world. The very opposite actually.