I just got an email at work starting with: “Certainly!, here is the rephrased text:…”
People abusing AI are not even reading the slop they are sending
I get these kinds of things all the time at work. I’m a writer, and someone once sent me a document to brief me on an article I had to write. One of the topics in the briefing mentioned a concept I’d never heard of (and the article was about a subject I actually know). I Googled the term, checked official sources … nothing, it just didn’t make sense. So I asked the person who wrote the briefing what it meant, and the response was: “I don’t know, I asked ChatGPT to write it for me LOL”.
facepalm is all I can think of…lol
I am not sure what my emailer started with but what chatgpt gave it was almost unintelligible
The thing is… AI is making me smarter! I use AI as a learning tool. The absolute best thing about AI is the ability to follow up questions with additional questions and get a better understanding of a subject. I use it to ask about technical topics and flush out a better understanding that I ever got from just a text book. I have seem some instances of hallucinating in the past, but with the current generation of AI I’ve had very good results and consider it an excellent tool for learning.
For reference I’m an engineer with over 25 years of experience and I am considered an expert in my field.
Same, I use it to put me down research paths. I don’t take anything it tells me at face value, but often it will introduce me to ideas in a particular field which I can then independently research by looking up on kagi.
Instead of saying “write me some code which will generate a series of caverns in a videogame”, I ask “what are 5 common procedural level generation algorithms, and give me a brief synopsis of them”, then I can take each one of those and look them up
$100 billion and the electricity consumption of France seems a tad pricey to save a few minutes looking in a book…
I recently read that LLMs are effective for improving learning outcomes. When I read one of the meta studies, however, it seemed that many of the benefits were indirect: LLMs improved accessibility by allowing teachers to quickly tailor lessons to individual students, for example. It also seems that some students ask questions more freely and without embarrassment when chatting with an LLM, which can improve learning for those students - and this aligns with what you mention in your post. I personally have withheld follow-up questions in lectures because I didn’t want to look foolish or reveal my imperfect understanding of the topic, so I can see how an LLM could help me that way.
What the studies did not (yet) examine was whether the speed and ease of learning with LLMs were somehow detrimental to, say, retention. Sure, I can save time studying for an exam/technical interview with an LLM, but will I remember what I learned in 6 months? For some learning tasks, the long struggle is essential to a good understanding and retention (for example, writing your own code implementation of an algorithm vs. reading someone else’s). Will my reliance on AI somehow damage my ability to learn in some circumstances? I think that LLMs might be like powered exoskeletons for the mind - the operator slowly wastes away from lack of exercise.
It seems like a paradox, but learning “more, faster” might be worse in the long run.
I use it as a glorified manual. I’ll ask it about specific error codes and “how do I” requests. One problem I keep running into is I’ll tell it the exact OS version and app version I’m using and it will still give me commands that don’t work with that version. Sometimes I’ll tell it the commands don’t work and restate my parameters and it will loop around to its original response in a logic circle.
At least it doesn’t say “Never mind, I figured out the solution” like they do too often in stack exchange.
If it’s a topic that has been heavily discussed on the internet or in literature, LLMs can have good conversations about it. Take it all with a grain of salt because it will regurgitate common bad arguments as well as good ones, but if you challenge it, you can get it to argue against its own previous statements.
It doesn’t handle things that are in flux very well. Or things that require very specific consistency. It’s a probabilistic model where it looks at existing tokens and predicts what the next one is most likely to be, so questions about specific versions of something might result in a response specific to that version or it might end up weighing other tokens more than the version or maybe even start treating it all like pseudocode, where descriptive language plays a bigger role than what specifically exists.
Joke’s on you, I was already stupid to begin with.
Actually a really good article with several excellent points not having to do with AI 😊👌🏻 Worth a read
I agree. I was almost skipping it because of the title, but the article is nuanced and has some very good reflections on topics other that AI. Every technical progress is a tradeoff. The article mentions cars to get to the grocery store and how there are advantages in walking that we give up when always using a car. Are cars in general a stupid and useless technology? No, but we need to be aware of where the tradeoffs are. And eventually most of these tradeoffs are economic in nature.
By industrializing the production of carpets we might have lost some of our collective ability to produce those hand-made masterpieces of old, but we get to buy ok-looking carpets for cheap.
By reducing and industrializing the production of text content, our mastery of language is declining, but we get to read a lot of not-very-good content for free. This pre-dates AI btw, as can be seen by standardized tests in schools everywhere.
The new thing about GenAI, though is that it upends the promise that technology was going to do the grueling, boring work for us and free up time for us to do the creative things that give us joy. I feel the roles have reversed: even when I have to write an email or a piece of coding, AI does the creative piece and I’m the glorified proofreader and corrector.
I think the author was quite honest about the weak points in his thesis, by drawing comparisons with cars, and even with writing. Cars come at great cost to the environment, to social contact, and to the health of those who rely on them. And maybe writing came at great cost to our mental capabilities though we’ve largely stopped counting the cost by now. But both of these things have enabled human beings to do more, individually and collectively. What we lost was outweighed by what we gained. If AI enables us to achieve more, is it fair to say it’s making us stupid? Or are we just shifting our mental capabilities, neglecting some faculties while building others, to make best use of the new tool? It’s early days for AI, but historically, cognitive offloading has enhanced human potential enormously.
Well creating the slide was a form of cognitive offloading, but barely you still had to know how to use and what formula to use. Moving to the pocket calculator just change how you the it didn’t really increase how much thinking we off loaded.
but this is something different. With infinite content algorithms just making the next choice of what we watch amd people now blindly trusting whatever llm say. Now we are offloading not just a comolex task like sqrt of 55, but “what do i want to watch”, “how do i know this true”.
I agree that it’s on a whole other level, and it poses challenging questions as to how we might live healthily with AI, to get it to do what we don’t benefit from doing, while we continue to do what matters to us. To make matters worse, this is happening in a time of extensive dumbing down and out of control capitalism, where a lot of the forces at play are not interested in serving the best interests of humanity. As individuals it’s up to us to find the best way to live with these pressures, and engage with this technology on our own terms.
how we might live healthily with AI, to get it to do what we don’t benefit from doing,
Agree that is oir goal, but one i don’t ai with paying for training data. Also amd this the biggest. What benefits me is not what benefits the people owning the ai models
What benefits me is not what benefits the people owning the ai models
Yep, that right there is the problem
Lol, this is the 10,000 thing that makes me stupid. Get a new scare tactic.
Proof that it’s already too late ☝️
Ain’t skeerd
I mean, obviously, you need higher cognitive functioning for all that
Damn, I thought flight or fight was the most primitive function. Ah well, back to chewing on this tire.
Yeah, you know, just like my cat is scared of distant fireworks but doesn’t give a flying fuck about climate change or rise of fascism in our own country.
Oh so like when someone’s afraid of falling off the edge of the earth?
More like how some people are afraid of needles but aren’t afraid of deadly diseases. Their primitive understanding of reality allows them to draw connection between prick and pain, but not between an invisible to the naked eye organism and a gruesome death.
that picture is kinky as hell, yo
I was annoyed that it wasn’t over her mouth to implant the egg.
It implants ideas, so it goes through the eyes.
No it’s am not
No shit
(picking up phone) Hello this is Sherlock speaking
Not me tho
How are you using new AI technology?
For porn, mostly.
I did have it create a few walking tours on a vacation recently, which was pretty neat.
The maker of Deep Seek made it so it would be easier for him to do stocks, which I am doing as well. Unless you all expect us to get degree on how to manually calculate the P/E ratio, potential loss and earnings, position sizing, spread and leverage, compounding, etc., then I will keep using AI. Not everyone of us could specialise on particular areas.
You’re speedrunning Danning-Kruger with an impressive force
I use LLM’s to help with math/science/coding, and the thing it screws up the most seems to be simple math (typically units/conversion issues) so I would be weary about gleaning financial advice from a chatbot.
so I would be weary about gleaning financial advice from a chatbot.
Oh yes, I use the bots for projections, which I don’t necessarily take on the face value. Some calculations had been off but as long as I gain some actual profits, I am content enough.
You don’t need AI to do that…
A new update for ONEui on my Samsung phone has allowed me to disable Gemini from the start. I wasted no time doing so
I only ever use it to answer a question and even then I double check it’s sources. I also like making superman pics.
Add it to the list