this post was submitted on 04 Sep 2024
914 points (98.2% liked)

Technology

59495 readers
3114 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Does AI actually help students learn? A recent experiment in a high school provides a cautionary tale. 

Researchers at the University of Pennsylvania found that Turkish high school students who had access to ChatGPT while doing practice math problems did worse on a math test compared with students who didn’t have access to ChatGPT. Those with ChatGPT solved 48 percent more of the practice problems correctly, but they ultimately scored 17 percent worse on a test of the topic that the students were learning.

A third group of students had access to a revised version of ChatGPT that functioned more like a tutor. This chatbot was programmed to provide hints without directly divulging the answer. The students who used it did spectacularly better on the practice problems, solving 127 percent more of them correctly compared with students who did their practice work without any high-tech aids. But on a test afterwards, these AI-tutored students did no better. Students who just did their practice problems the old fashioned way — on their own — matched their test scores.

top 50 comments
sorted by: hot top controversial new old
[–] N0body@lemmy.dbzer0.com 205 points 2 months ago (11 children)

Traditional instruction gave the same result as a bleeding edge ChatGPT tutorial bot. Imagine what would happen if a tiny fraction of the billions spent to develop this technology went into funding improved traditional instruction.

Better paid teachers, better resources, studies geared at optimizing traditional instruction, etc.

Move fast and break things was always a stupid goal. Turbocharging it with all this money is killing the tried and true options that actually produce results, while straining the power grid and worsening global warming.

[–] TallonMetroid@lemmy.world 112 points 2 months ago (1 children)

Investing in actual education infrastructure won't get VC techbros their yachts, though.

[–] elvith@feddit.org 50 points 2 months ago

It’s the other way round: Education makes for less gullible people and for workers that demand more rights more freely and easily - and then those are coming for their yachts…

[–] Petter1@lemm.ee 15 points 2 months ago

Imagine all the money spent on war would be invested into education 🫣what a beautiful world we would live in.

load more comments (9 replies)
[–] 2ugly2live@lemmy.world 67 points 2 months ago (5 children)

I don't even know of this is ChatGPT's fault. This would be the same outcome if someone just gave them the answers to a study packet. Yes, they'll have the answers because someone (or something) gave it to them, but won't know how to get that answer without teaching them. Surprise: For kids to learn, they need to be taught. Shocker.

load more comments (5 replies)
[–] blazeknave@lemmy.world 59 points 2 months ago (4 children)

Kids who take shortcuts and don't learn suck at recalling knowledge they never had..

[–] ameancow@lemmy.world 22 points 2 months ago* (last edited 2 months ago) (1 children)

The only reason we're trying to somehow compromise and allow or even incorporate cheating software into student education is because the tech-bros and singularity cultists have been hyping this technology like it's the new, unstoppable force of nature that is going to wash over all things and bring about the new Golden Age of humanity as none of us have to work ever again.

Meanwhile, 80% of AI startups sink and something like 75% of the "new techs" like AI drive-thru orders and AI phone support go to call centers in India and Philippines. The only thing we seem to have gotten is the absolute rotting destruction of all content on the internet and children growing up thinking it's normal to consume this watered-down, plagiarized, worthless content.

load more comments (1 replies)
[–] ChickenLadyLovesLife@lemmy.world 13 points 2 months ago (2 children)

I took German in high school and cheated by inventing my own runic script. I would draw elaborate fantasy/sci-fi drawings on the covers of my notebooks with the German verb declensions and whatnot written all over monoliths or knight's armor or dueling spaceships, using my own script instead of regular characters, and then have these notebook sitting on my desk while taking the tests. I got 100% on every test and now the only German I can speak is the bullshit I remember Nightcrawler from the X-Men saying. Unglaublich!

[–] pmc@lemmy.blahaj.zone 7 points 2 months ago (2 children)

Meanwhile the teacher was thinking, "interesting tactic you've got there, admiring your art in the middle of a test"

load more comments (2 replies)
load more comments (1 replies)
load more comments (2 replies)
[–] Insig@lemmy.world 51 points 2 months ago (7 children)

At work we give a 16/17 year old, work experience over the summer. He was using chatgpt and not understanding the code that was outputing.

I his last week he asked why he doing print statement something like

print (f"message {thing} ")

load more comments (7 replies)
[–] kusivittula@sopuli.xyz 49 points 2 months ago (17 children)
load more comments (17 replies)
[–] Eiri@lemmy.ca 37 points 2 months ago (17 children)
load more comments (17 replies)
[–] maegul@lemmy.ml 33 points 2 months ago (2 children)

Yea, this highlights a fundamental tension I think: sometimes, perhaps oftentimes, the point of doing something is the doing itself, not the result.

Tech is hyper focused on removing the "doing" and reproducing the result. Now that it's trying to put itself into the "thinking" part of human work, this tension is making itself unavoidable.

I think we can all take it as a given that we don't want to hand total control to machines, simply because of accountability issues. Which means we want a human "in the loop" to ensure things stay sensible. But the ability of that human to keep things sensible requires skills, experience and insight. And all of the focus our education system now has on grades and certificates has lead us astray into thinking that the practice and experience doesn't mean that much. In a way the labour market and employers are relevant here in their insistence on experience (to the point of absurdity sometimes).

Bottom line is that we humans are doing machines, and we learn through practice and experience, in ways I suspect much closer to building intuitions. Being stuck on a problem, being confused and getting things wrong are all part of this experience. Making it easier to get the right answer is not making education better. LLMs likely have no good role to play in education and I wouldn't be surprised if banning them outright in what may become a harshly fought battle isn't too far away.

All that being said, I also think LLMs raise questions about what it is we're doing with our education and tests and whether the simple response to their existence is to conclude that anything an LLM can easily do well isn't worth assessing. Of course, as I've said above, that's likely manifestly rubbish ... building up an intelligent and capable human likely requires getting them to do things an LLM could easily do. But the question still stands I think about whether we need to also find a way to focus more on the less mechanical parts of human intelligence and education.

load more comments (2 replies)
[–] Soup@lemmy.cafe 29 points 2 months ago

Kids using an AI system trained on edgelord Reddit posts aren’t doing well on tests?

Ya don’t say.

[–] flerp@lemm.ee 28 points 2 months ago (2 children)

Like any tool, it depends how you use it. I have been learning a lot of math recently and have been chatting with AI to increase my understanding of the concepts. There are times when the textbook shows some steps that I don't understand why they're happening and I've questioned AI about it. Sometimes it takes a few tries of asking until you figure out the right question to ask to get the right answer you need, but that process of thinking helps you along the way anyways by crystallizing in your brain what exactly it is that you don't understand.

I have found it to be a very helpful tool in my educational path. However I am learning things because I want to understand them, not because I have to pass a test and that determination in me to want to understand is a big difference. Just getting hints to help you solve the problem might not really help in the long run, but it you're actually curious about what you're learning and focus on getting a deeper understanding of why and how something works rather than just getting the right answer, it can be a very useful tool.

[–] Rekorse@sh.itjust.works 16 points 2 months ago (14 children)

Why are you so confident that the things you are learning from AI are correct? Are you just using it to gather other sources to review by hand or are you trying to have conversations with the AI?

We've all seen AI get the correct answer but the show your work part is nonsense, or vice versa. How do you verify what AI outputs to you?

[–] GaMEChld@lemmy.world 8 points 2 months ago (1 children)

You check it's work. I used it to calculate efficiency in a factory game and went through and made corrections to inconsistencies I spotted. Always check it's work.

load more comments (1 replies)
load more comments (13 replies)
load more comments (1 replies)
[–] glowie@h4x0r.host 24 points 2 months ago (10 children)

Of all the students in the world, they pick ones from a "Turkish high school". Any clear indication why there of all places when conducted by a US university?

[–] catloaf@lemm.ee 19 points 2 months ago (1 children)

I'm guessing there was a previous connection with some of the study authors.

I skimmed the paper, and I didn't see it mention language. I'd be more interested to know if they were using ChatGPT in English or Turkish, and how that would affect performance, since I assume the model is trained on significantly more English language data than Turkish.

load more comments (1 replies)
[–] Phoenix3875@lemmy.world 8 points 2 months ago

The paper only says it's a collaboration. It's pretty large scale, so the opportunity might be rare. There's a chance that (the same or other) researchers will follow up and experiment in more schools.

load more comments (8 replies)
[–] Akasazh@feddit.nl 22 points 2 months ago (3 children)
load more comments (3 replies)
[–] michaelmrose@lemmy.world 22 points 2 months ago

TLDR: ChatGPT is terrible at math and most students just ask it the answer. Giving students the ability to ask something that doesn't know math the answer makes them less capable. An enhanced chatBOT which was pre-fed with questions and correct answers didn't screw up the learning process in the same fashion but also didn't help them perform any better on the test because again they just asked it to spoon feed them the answer.

references

ChatGPT’s errors also may have been a contributing factor. The chatbot only answered the math problems correctly half of the time. Its arithmetic computations were wrong 8 percent of the time, but the bigger problem was that its step-by-step approach for how to solve a problem was wrong 42 percent of the time.

The tutoring version of ChatGPT was directly fed the correct solutions and these errors were minimized.

The researchers believe the problem is that students are using the chatbot as a “crutch.” When they analyzed the questions that students typed into ChatGPT, students often simply asked for the answer.

[–] Vanth@reddthat.com 18 points 2 months ago (5 children)

I'm not entirely sold on the argument I lay out here, but this is where I would start were I to defend using chatGPT in school as they laid out in their experiment.

It's a tool. Just like a calculator. If a kid learns and does all their homework with a calculator, then suddenly it's taken away for a test, of course they will do poorly. Contrary to what we were warned about as kids though, each of us does carry a calculator around in our pocket at nearly all times.

We're not far off from having an AI assistant with us 24/7 is feasible. Why not teach kids to use the tools they will have in their pocket for the rest of their lives?

[–] filister@lemmy.world 19 points 2 months ago (10 children)

I think here you also need to teach your kid not to trust unconditionally this tool and to question the quality of the tool. As well as teaching it how to write better prompts, this is the same like with Google, if you put shitty queries you will get subpar results.

And believe me I have seen plenty of tech people asking the most lame prompts.

load more comments (10 replies)
[–] Schal330@lemmy.world 16 points 2 months ago (1 children)

As adults we are dubious of the results that AI gives us. We take the answers with a handful of salt and I feel like over the years we have built up a skillset for using search engines for answers and sifting through the results. Kids haven't got years of experience of this and so they may take what is said to be true and not question the results.

As you say, the kids should be taught to use the tool properly, and verify the answers. AI is going to be forced onto us whether we like it or not, people should be empowered to use it and not accept what it puts out as gospel.

[–] Petter1@lemm.ee 9 points 2 months ago

This is true for the whole internet, not only AI Chatbots. Kids need to get teached that there is BS around. In fact kids had to learn that even pre-internet. Every human has to learn that you can not blindly trust anything, that one has to think critically. This is nothing new. AI chatbots just show how flawed human education is these days.

load more comments (3 replies)
[–] MystikIncarnate@lemmy.ca 17 points 2 months ago

Something I've noticed with institutional education is that they're not looking for the factually correct answer, they're looking for the answer that matches whatever you were told in class. Those two things should not be different, but in my experience, they're not always the same thing.

I have no idea if this is a factor here, but it's something I've noticed. I have actually answered questions with a factually wrong answer, because that's what was taught, just to get the marks.

[–] prosp3kt@lemmy.dbzer0.com 15 points 2 months ago* (last edited 2 months ago) (2 children)

There are a part here that sounds interesting

The students who used it did spectacularly better on the practice problems, solving 127 percent more of them correctly compared with students who did their practice work without any high-tech aids. But on a test afterwards, these AI-tutored students did no better.

Do you think that these students that used ChatGPT can do the exercises "the old fashioned way"? For me it was a nightmare try to resolve a calculus problem just with the trash books that doesn't explain a damn fuck, I have to go to different resources, wolphram, youtube, but what happened when there was a problem that wasnt well explained in any resource?. I hate openAI, I want to punch Altman in the face. But this doesn't mean we have to bait this hard in the title.

load more comments (2 replies)
[–] fne8w2ah@lemmy.world 14 points 2 months ago

Taking too many shortcuts doesn't help anyone learn anything.

[–] Cornelius_Wangenheim@lemmy.world 12 points 2 months ago (6 children)

This isn't a new issue. Wolfram alpha has been around for 15 years and can easily handle high school level math problems.

[–] zarcher@lemmy.world 33 points 2 months ago

Except wolfram alpha is able to correctly explain step by step solutions. Which was an aid in my education.

load more comments (5 replies)
[–] Maggoty@lemmy.world 9 points 2 months ago

ChatGPT lies which is kind of an issue in education.

As far as seeing the answer, I learned a significant amount of math by looking at the answer for a type of question and working backwards. That's not the issue as long as you're honestly trying to understand the process.

[–] Kusimulkku@lemm.ee 8 points 2 months ago (4 children)

I've found AI helpful in asking for it to explain stuff. Why is the problem solved like this, why did you use this and not that, could you put it in simpler terms and so on. Much like you might ask a teacher.

[–] NikkiDimes@lemmy.world 7 points 2 months ago* (last edited 2 months ago) (2 children)

I think this works great if the student is interested in the subject, but if you're just trying to work through a bunch of problems so you can stop working through a bunch of problems, it ain't gonna help you.

I have personally learned so much from LLMs (although you can't really take anything at face value and have to look things up independently, but it gives you a great starting place), but it comes from a genuine interest in the questions I'm asking and things I dig at.

load more comments (2 replies)
load more comments (3 replies)
load more comments
view more: next ›