I can see some minor benefits - I use it for the odd bit of mundane writing and some of the image creation stuff is interesting, and I knew that a lot of people use it for coding etc - but mostly it seems to be about making more cash for corporations and stuffing the internet with bots and fake content. Am I missing something here? Are there any genuine benefits?
Much like automated machinery, it could in theory free the workers to do more important, valuable work and leave the menial stuff for the machine/AI. In theory this should make everyone richer as the companies can produce stuff cheaper and so more of the profits can go to worker salaries.
Unfortunately what happens is that the extra productivity doesn’t go to the workers, but just let’s the owners of the companies take more of the money with fewer expenses. Usually rather firing the human worker rather than giving them a more useful position.
So yea I’m not sure myself tbh
No no you found the actual “use” for AI as far as businesses go. They don’t care about the human cost of adopting AI and firing large swaths of workers just profits.
Which is why governments should be quickly moving to highly regulate AI and it’s uses. But governments are slow plodding things full of old people who get confused with toasters.
As always capitalism kills.
This is the part that bothers me the most, I think.
Trouble is the best way to regulate it isn’t clear. If the new tool can do the job at least as well and cheaper, just disallowing it is less beneficial to society. You can tax its use until it is only a little cheaper, but then you have to get people to approve of taxes. Et cetera
This already happened with the industrial revolution. It did make the rich awfully rich, but let’s be honest. People are way better off today too.
It’s not perfect, but it does help in the long run. Also, there’s a big difference in which country you’re in.
Capitalist-socialism will be way better off than hard core capitalism, because the mind set and systems are already in place to let it benefit the people more.
Yes, that way the government will be able to make sure it benefits the right people. And we will call it the national socialism… wait… no!
The question wasn’t "In Theory, are there any genuine benefits" it was if there are currently right now.
Most email spam detection and antimalware use ML. There are use cases in medicine with trying to predict whether someone has a condition early
It’s also being used in drug R&D to find similar compounds like antimicrobial activity, afaik.
Medical use is absolutely revolutionary. From GP’s consultations to reading tests results, radios, AI is already better than humans and will be getting better and better.
Computers are exceptionally good at storing large amount of data, and with ML they are great at taking a lot of input and inferring a result from that. This is essentially diagnosing in a nutshell.
I read that one LLM was so good at detecting TB from Xrays that they reverse engineered the “black box” code hoping for some insight doctors could use. Turns out, the AI was biased toward the age of the Xray machine that took each photo because TB is more common in developing countries that have older equipment. Womp Womp.
A large language model was used to detect TB in X-ray? Do you not just mean Machine Learning?
There are supposedly multiple Large Language Model Radiology Report Generators in development. Can’t say if any of them are actually useful at all, though.
okay, but there still needs to be a part that processes the scan images and that’s not LLM.
So you’re saying because the LLM isn’t operating the machinery and processing the data start to finish without any non-LLM software then none of it is LLM? Stay off the drugs, kid.
What’s TB?
Tuberculosis
Tuberculosis
Watch any video at random by John Green (vlogbrothers, and author of several successful books that I haven’t read) and you’ll know more than you could ever hope about TB.
That’s super interesting, TIL
I hadn’t considered this. It’s interesting stuff. My old doctor used to just Google stuff in front of me and then repeat the info as if I hadn’t been there for the last five minutes.
This sort of feels like someone using a PC for the first time in 1989 and asking what it does that they can’t do on a piece of paper with a calculator. They may not have been far off at the time, but they would be missing the point. This is a paradigm shift that allows for a single application to fulfill the role of, eventually, infinite applications. And yes it starts with mundane tasks. You know, the kind people don’t want to do themselves.
The problem is that most of the things it feels we can currently see applications for are… Kinda bad. Actually repulsive frankly. Like I don’t want those things. I don’t wanna talk to an ai to order my big mac or instead of just getting a highlighted excerpt from a webpage when I search things. I don’t want a world where artists have to compete with image generators to make a living, or where weird creepy porn that chases and satisfies ever more unrealistic expectations is the norm. I don’t want to talk to chat bots that use statistical analysis to convincingly sell me lies they don’t understand.
I just wanna talk to actual people. I wanna see art made by people, I wanna look at pictures of the bodies of actual human beings, I wanna see the animations that humans poured their soul into, I wanna see the actual text a person wrote on the subject I’m researching. I wanna do simple things, in simple ways, and the world that it feels like AI companies are offering us honestly sucks, and as soon as that door is fully opened things will just be permanently worse. Convenience is great but I don’t want a robot to feed me a weird gross regurgitation of reality or approximation of human interaction to me like a bird that chews and digests its food for its babies. I don’t wanna consume the spit-up of an overgrown algorithm. Its a gross idea of how we could engage with the world. It obfuscates the humanity of whatever it touches, and the humanity is the worthwhile part. There comes a point where the abstraction is abstracting away everything of value and leaving you with the most sanitized version.
If ai was just gonna be used to improve medicine and translate books or webpages, or as interactive accsessibiltiy tool, or do actually helpful shit maybe I wouldn’t be so opposed to it, but it feels like everything consumer or employee facing that ai is offering is awful and something I absolutely do not want. But companies don’t care, and that shitty world is gonna be the reality cause it’s profitable
Well then I guess I’d ask you to reconsider your answer but from the perspective of 1989. I’d imagine that’d be the same answer you’d give to the personal computer. AI isn’t going to make things more complicated It’s going to make things simpler. But people will create a more complicated (diverse) world in the vacuum that leaves. Just like an ox pulled plow made it easier to till farmland led to more complex agricultural societies. This type of advancement has been the story of human history since its beginning. Your perspective seems most concerned with people using this advancement against you, but our future now holds the possibility of having this AI on your side.
Using it to synopsize complicated TOS that corporations use to obfuscate what you’re agreeing to, actually answering questions instead of needing to search through ad riddled web pages, allowing more people to become artists and create their vision.
Your examples of useful ways to use AI are great. So help build or support them. If you only look at the future corporations are selling you, yeah, it’s going to look like a bleak corporate nightmare. But the truth is technology empowers the individual. So we need to do something good with that power.
TBF if a mathematician or a programmer cannot do it on paper then they’ve kind of failed and probably won’t have any notable impact. Paper math didn’t end when consumer computers came about.
Wrap it up, climate scientists, the show is over! This lad said he can do your job without the supercomputer.
You think Supercomputers are designing and building themselves, you fucking donkey? You think ChatGPT has the solution to Climate Change?
I know plenty of modern programmers who are empowered by the ease at which they can learn the trade now. Some never go deeper than front end developer, because there’s good money there. That job would look nothing like it does today if it had to be done by hand.
Ah yes, the html programmers. Top minds of our generation, them. /s
AI is a very broad topic. Unless you only want to talk about Large Language Models (like ChatGPT) or AI Image Generators (Midjourney) there are a lot of uses for AI that you seem to not be considering.
It’s great for upscaling old videos: (this would fall under image generating AI since it can be used for colorizing, improving details, and adding in additional frames) so that you end up with something like: https://www.youtube.com/watch?v=hZ1OgQL9_Cw
It’s useful for scanning an image for text and being able to copy it out (OCR).
It’s excellent if you’re deaf, or sitting in a lobby with a muted live broadcast and want to see what is being said with closed captions (Speech to Text).
Flying your own drone with object detection/avoidance.
There’s a lot more, but basically, it’s great at taking mundane tasks where you’re stuck doing the same (or similar) thing over, and over, and over again, and automating it.
Here is an alternative Piped link(s):
https://www.piped.video/watch?v=hZ1OgQL9_Cw
Piped is a privacy-respecting open-source alternative frontend to YouTube.
I’m open-source; check me out at GitHub.
Yeah that’s interesting.
I think most of those are only labelled AI to generate tech hype, though? Like, sure, machine learning and maybe even LLM can and are used for those, but it isn’t a machine given human discernable input and pretending to give human output.
“AI” is the broadest umbrella term for any of these tools. That’s why I pointed out that OP really should be a bit more specific as to what they mean with their question.
AI doesn’t have the same meaning that it had over 10 years ago when we used to use it exclusively for machines that could think for themselves.
They are the greatest gift to solo-brainstorming that I’ve ever encountered.
_ /\ _
The fruit of those brainstorming sessions are like Homer Simpson designing a new car.
You’re confusing brainstorming with content generation. LLMs are great for brainstorming: they can quickly churn out dozens of ideas for my D&D campaign, which I then look through, discard the garbage, keep the good bits of, and riff off of before incorporating into my campaign. If I just used everything it suggested blindly, yeah, nightmare fuel. For brainstorming though, it’s fantastic.
Exactly. It can generate those base-level ideas much faster and worth higher fidelity than humans can without it, and that can see us at the hobby level with DND, or up at the business level with writers rooms and such.
The important point is that you still need someone good at making the thing you want to look at and finish the thing you’re making, or you end up with paintings with too many fingers or stories full of contradictions
Any kid who uses it to craft their campaign is lazy and depriving themselves of a valuable experience, any professional who uses it to write a book, script, or study is wildly unethical, and both are creating a much much worse product than a human without reliance on them. That is the reality of a model who at 100% accuracy would be exactly as flawed as human output, and we’re nowhere near that accuracy.
But the point is that you don’t use it to make the campaign or write the book. You use it as a tool to help yourself make a campaign or write a book. Ignoring the potential of ai as a tool is silly just because it can’t do the whole job for you. That would be a bit like saying you are a fool for using a sponge when washing because it will never get everything by itself…
I get it now! You don’t use it for the thing you use it for but instead as a tool to create the thing that you’ve used it for for yourself because the magic was inside all of us but also the GPT all along. /sarcasm
“don’t feed the trolls,” they said, but did she ever listen?
No, I guess I didn’t…
I would retort that the exact opposite is true, that content generation is the only thing LLMs are good at because they often forget the context of their previous statements.
I think we’re saying the same thing there: LLMs are great at spewing out a ton of content, which makes them a great tool for brainstorming. The content they create is not necessarily trustworthy or even good, but it can be great fuel for the creative process.
My stance is that spewing out a ton of flawed unrelated content is not conducive to creative good content, and therefor LLM is not useful for writing. That hasn’t changed.
Anything that requires tons of iteration can be done way faster with AI. Finding new chemical formulas for medicine, as an example. It takes a “throw everything at the wall and see what sticks” approach, but it’s still more effective than a human.
Brute force is AI now?
brute force would be “throw at the wall one at a time until one stick”
As long as everything gets thrown it’s still brute force, but the reason they use ai for it is because it can throw a lot more a lot faster.
I think by broad definitions it can be, yes.
Think about it. AI is just throwing a ton of sample data in and filtering out the results that are least correct.
Presumably in order to determine whether the eg chemical is worth looking at in the first place
AI has some interesting use cases, but should not be trusted 100%.
Like github copilot ( or any “code copilot”):
- Good for repeating stuff but with minor changes
- Can help with common easy coding errors
- Code quality can take a big hit
- For coding beginners, it can lead to a deficit of real understanding of your code
( and because of that could lead to bugs, security backdoors… )
Like translations ( code or language ):
- Good translation of the common/big languages ( english, german…)
- Can extend a brief summary to a big wall of text ( and back )
- If wrong translated it can lead to that someone else understands it wrong and it misses the point
- It removes the “human” part. It can be most of the time depending on the context easily identified.
Like classification of text/Images for moderation:
- Help for identify bad faith text / images
- False Positives can be annoying to deal with.
But dont do anything that is IMPORTANT with AI, only use it for fun or if you know if the code/text the AI wrote is correct!
Adding to the language section, it’s also really good at guessing words if you give it a decent definition. I think this has other applications but it’s quite useful for people like me with the occasionally leaky brain.
I have sometimes the same issue!
Actually the summaries are good, but you have to know some of it anyway and then check to see if it’s just making stuff up. That’s been my experience.
An interesting point that I saw about a trail on one of the small, London Tube stations:
-
most of the features involved a human who could come and assist or review the footage. The AI being able to flag wheelchair users was good because the station doesn’t have wheelchair access with assistance.
-
when they tried to make a heuristic for automatically flagging aggressing people, they found that people with the arms up tend to be aggressive. This flagging system led to the unexpected feature that if a Transport For London (TFL) staff member needed assistance (i.e. if medical assistance was necessary, or if someone was being aggressive towards them, the TFL staff member could put their arms up to bring the attention onto them.
That last one especially seems neat. It seems like the kind of use case where AI has the most power when it’s used as a tool to augment human systems, rather than taking humans out of stuff.
“Once implemented, the system was able to identify many black men who were then immediately confronted. Confrontations with black men are now documented at 87% of aggressive confrontations in TFL locations.” /sarcasm
I don’t think designing AI to make generalizations based on physical appearances is a very good idea to start with.
While not AI. That’s my goal with my home automation. To augment my life to make certain things easier and/or more efficient.
https://www.home-assistant.io/blog/2016/01/19/perfect-home-automation/
-
AI is a revolution in learning.
deleted by creator
I don’t believe you, this seems like a bot posting.
Very true. I learned how to code surprisingly fast.
And even the mistakes the AI made was good, because it made me learn so much seeing what changes it did to fix it.
Bullshit. Reading a book on a language is just as fast and it doesn’t randomly lie or make up entire documentations as an added bonus.
It’s sped up my retouching workflows. I can automate things that a few years ago would’ve needed quite a lot of time spent with manual brush work.
Also in the creative industries, it’s a massive time saver for conceptual work. Think storyboarding and scamping, first stage visuals that kind of thing.
One of the better uses I’ve heard of is in search and rescue type situations. Using AI to find specific items, people or anomalies on a map or video feed can be helpful.
An example regarding wildfires:
Our software uses ML to detect tax fraud and since tax offices are usually understaffed they can now go after more cases. So yes?
Don’t limit your thoughts to just generative AI, which is what you are talking about. Chat bot and media generation aren’t the only uses for AI (by which I mean any trained neural network program that can do some sort of task.
Motor skills
AI can solve learn to solve the marble maze “Labyrinth” much, much faster than a human, and then speedrun it faster than any human ever has. Six hours. That’s how long it took a brand new baby AI to beat the human world record. A human that has been learning hand-eye coordination and fine motor control all of it’s life, with a brain which evolved over millions of years to do exactly that.
No special code needed. The AI didn’t need to be told how balls roll or knobs turn, or how walls block the ball. It earned all of that on the fly. The only special code it had was optical and mechanical. It knew it had “hands” in the form of two motors, and it knew how to use them. It also had eyes (a camera), and access to a neural network computer vision system. When the AI started taking illegal shortcuts, and they had to instruct it to follow the prescribed path, which is printed on the maze.
Robots could in work factories, mines, and other dangerous, dehumanizing jobs. Why do we want workers to behave like robots at a factory job? Replace them with actual robots and let them perform a human job like customer service.
Think of a robot that has actual hands and arms, feet and legs, and various “muscles”. We have it learn it’s motor control using a very accurate physics system on a computer that simulates its body. This allows the AI to learn at much faster speeds than by controlling a real robot. We can simulate thousands of robots in parallel and run the simulations much faster than real time. Train it to learn how to use it’s limbs and eyes to climb over obstacles, open doors and detain or kill people. We could replace police with them. Super agile robot cops with no racial bias or other prejudices. Arresting people and recording their crimes. Genuine benefit.
Computer Vision
AI can be trained to recognize objects, abstract shapes, people’s individual faces, emotions, people’s individual body shape, mannerisms, and gait. There are many genuine benefits to such systems. We can monitor every public location with cameras and an AI employing these tools. This would help you find lost loved ones, keep track of your kids as they navigate the city, and track criminal activity.
By recording all of this data, tagged with individual names, we can spontaneously view the public history of any person in the world for law enforcement purposes. Imagine we identify a person as a threat to public safety 10 years from now. We’d have 10 years of data showing everyone they’ve ever associated with and where they went. Then we could weed out entire networks of crime at once by finding patterns among the people they’ve associated with.
AI can even predict near future crime from an individual’s recent location history, employment history, etc. Imagine a person is fired from his job then visits a gun store then his previous place of employment. Pretty obvious what’s going on, right? But what if it happens over the period of two weeks? Difficult for a human to detect a pattern like this in all the noise of millions of people doing their everyday tasks, but easy for an AI. Genuine benefit.
Managing Production
With enough data and processing power, we can manage the entire economy without the need for capitalism. People’s needs could be calculated by an AI and production can be planned years ahead of time to optimize inputs and outputs. The economy–as it stands today–is a distributed network of human brains and various computers. AI can eliminate the need for the humans, which is good because humans are greedy and neurotic. AI can do the same job without either. Again, human’s are freed to pursue human endeavors instead of worrying about making sure each farm and factory has the resources it needs to feed and clothe everyone. Genuine benefit.
Togetherness
We will all be part of the same machine working in harmony instead of fighting over how to allocate resources. Genuine benefit!
Says the AI…
Even if it turns out to be organic, that wordwall is shill as fuck.
I was with it until it said let’s train AI robots to kill people, and then use them to track every face on the planet and use that data to “identify threats”…
OP wants a robot overseer, but also wants it to be a police state.
deleted by creator
Train it to learn how to use it’s limbs and eyes to climb over obstacles, open doors and detain or kill people. We could replace police with them. Super agile robot cops with no racial bias or other prejudices. Arresting people and recording their crimes. Genuine benefit.
I got as far as ai cops and became sceptical. Like, yeah, sure, but what you’re describing isn’t just a robot being controlled by an AI, it’s also the ai making decisions and choosing who to pursue and such, which is a known weakness right now.
And then you let them kill people.
Well, yeah. They are told to put down their weapon. They get 20 seconds to comply. If they don’t, they get killed.
Don’t discount the generative AI either!
Language generating AI like LLMs: Though we’re in early stages yet and they don’t really work for communication, these are going to be the foundation on which AI learns to talk and communicate information to people. Right now they just spit out correct-sounding responses, but eventually the trick to using that language generation to actually communicate will be resolved.
Image/video/music generating AI: How difficult it is right now, for the average person to illustrate an idea or visual concept they have! But already these image generating AI are making such illustration available to the common person. As they advance further and adjusting their output based on natural conversational language becomes more effective, this will only get better. A picture paints a thousand words…and now the inverse will also be true, as anyone will be able to create a picture with sufficient description. And the same applies to video and music.
That said I love your managing production point. It’s something I e been thinking too - centrally planned economies have always had serious issues, but if with predictive AI we can overcome the problems by accurately predicting future need, the problems with them may be solvable, and we can then take advantage of the inherent efficiency in such a planned system.
That’s funny because the whole post was sarcastically outlining a distopian nightmare.
If that kind of stuff was actually to become real, some dictator would take control of it and subjugate the entire country, or world… forever. There’d be no way to resist that level of surveillance or machine policing.
I think each one of those dystopian ideas can be done in a safe and humane way, but needless to say it is not the current trajectory.
Well written!
Nice post. A while back I read something on reddit about a theory for technological advances always being used for the worst possible nightmare scenario. But I can’t find it now. Fundamentally I’m a technological optimist but I can’t even yet fully imagine the subtle systemic issues this will cause. Except the rather obvious one:
Algorithms on social media decide what people see and that shapes their perception of the world. It is relatively easy to manipulate in subtle ways. If AI can learn to categorize the psychology of users and then “blindly anticipate” what and how they will respond to stimuli (memes / news / framing) then that will in theory allow for a total control by means of psychological manipulation. I’m not sure how close we are to this, the counter argument would be that AI or LLMs currently don’t understand at all what is going on, but public relations / advertising / propaganda works on emotional levels and doesn’t “have to make sense”. The emotional logic is much easier to categorize and generate. So even if there is no blatant evil master plan just optimizing for max profit, max engagement, could make the AI pursue (dumbly) a broad strategy that is more evil than that.
Another great one is science! Machine learning is used for physics, bio, and chem models, in things such as genetic sequencing and generation of new drugs as well as very useful in figuring out protein folding. It’s very useful in all of the iterative “grunt work” so to speak. While it may not be the best at finding effective new drugs, it can certainly arrange molecules according to the general rules of organic chemistry much faster than any human, and because of that has already led to several drug breakthroughs. AI is hugely useful! LLMs are mostly hype
I use it daily to generate basic Perl scripts that I cant be bothered to write myself. It’s fantastic.