Does AI actually help students learn? A recent experiment in a high school provides a cautionary tale.

Researchers at the University of Pennsylvania found that Turkish high school students who had access to ChatGPT while doing practice math problems did worse on a math test compared with students who didn’t have access to ChatGPT. Those with ChatGPT solved 48 percent more of the practice problems correctly, but they ultimately scored 17 percent worse on a test of the topic that the students were learning.

A third group of students had access to a revised version of ChatGPT that functioned more like a tutor. This chatbot was programmed to provide hints without directly divulging the answer. The students who used it did spectacularly better on the practice problems, solving 127 percent more of them correctly compared with students who did their practice work without any high-tech aids. But on a test afterwards, these AI-tutored students did no better. Students who just did their practice problems the old fashioned way — on their own — matched their test scores.

  • trougnouf@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    5 months ago

    The title is pretty misleading. Kids who used ChatGPT to get hints/explanations rather than outright getting the answers did as well as those who had no access to ChatGPT. They probably had a much easier time studying/understanding with it so it’s a win for LLMs as a teaching tool imo.

    • Apollo42@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      5 months ago

      Is it really a win for LLMs if the study found no significant difference between those using it as a tutor and those not?

      • 8andage@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        5 months ago

        Maybe using llm assistance was less stressful or quicker than self study. The tutoring focused llm is definitely better than allowing full access to gpt itself, which is what is currently happening

      • EatATaco@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        As another poster questioned, if it saved them tine then, yes, it is absolutely a win. But if they spent the same amount of time, I would agree with you that it’s not a win.

      • trougnouf@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        Not everyone can afford a tutor or knows where to find an expert that can answer questions in any given domain. I think such a tool would have made understanding a lot of my college courses a lot easier.

  • Kusimulkku@lemm.ee
    link
    fedilink
    English
    arrow-up
    0
    ·
    5 months ago

    I’ve found AI helpful in asking for it to explain stuff. Why is the problem solved like this, why did you use this and not that, could you put it in simpler terms and so on. Much like you might ask a teacher.

    • NιƙƙιDιɱҽʂ@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      5 months ago

      I think this works great if the student is interested in the subject, but if you’re just trying to work through a bunch of problems so you can stop working through a bunch of problems, it ain’t gonna help you.

      I have personally learned so much from LLMs (although you can’t really take anything at face value and have to look things up independently, but it gives you a great starting place), but it comes from a genuine interest in the questions I’m asking and things I dig at.

      • utopiah@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        I have personally learned so much from LLMs

        No offense but that’s what the article is also highlighting, naming that students, even the good, believe they did learn. Once it’s time to pass a test designed to evaluate if they actually did, it’s not that positive.

    • JackbyDev@programming.dev
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 months ago

      To an extent, but it’s often just wrong about stuff.

      It’s been a good second step for things I have questions about that I can’t immediately find good search results for. I don’t wanna get off topic but I have major beef with Stack Overflow and posting questions there makes me anxious as hell because I’ll do so much diligence to make sure it is clear, reproducible, and not a duplicate only for my questions to still get closed. It’s a major fucking waste of my time. Why put all that effort in when it’s still going to get closed?? Anyways – ChatGPT never gets mad at me. Sure, it’s often wrong as hell but it never berates me or makes me feel stupid for asking a question. It generally gets me close enough on topics that I can search for other terms in search engines and get different results that are more helpful.

    • Homescool@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      5 months ago

      Yep. My first interaction with GPT pro lasted 36 hours and I nearly changed my religion.

      AI is the best thing to come to learning, ever. If you are a curious person, this is bigger than Gutenberg, IMO.

    • ameancow@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      5 months ago

      The only reason we’re trying to somehow compromise and allow or even incorporate cheating software into student education is because the tech-bros and singularity cultists have been hyping this technology like it’s the new, unstoppable force of nature that is going to wash over all things and bring about the new Golden Age of humanity as none of us have to work ever again.

      Meanwhile, 80% of AI startups sink and something like 75% of the “new techs” like AI drive-thru orders and AI phone support go to call centers in India and Philippines. The only thing we seem to have gotten is the absolute rotting destruction of all content on the internet and children growing up thinking it’s normal to consume this watered-down, plagiarized, worthless content.

    • ChickenLadyLovesLife@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 months ago

      I took German in high school and cheated by inventing my own runic script. I would draw elaborate fantasy/sci-fi drawings on the covers of my notebooks with the German verb declensions and whatnot written all over monoliths or knight’s armor or dueling spaceships, using my own script instead of regular characters, and then have these notebook sitting on my desk while taking the tests. I got 100% on every test and now the only German I can speak is the bullshit I remember Nightcrawler from the X-Men saying. Unglaublich!

      • pmc@lemmy.blahaj.zone
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        Meanwhile the teacher was thinking, “interesting tactic you’ve got there, admiring your art in the middle of a test”

        • ChickenLadyLovesLife@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          5 months ago

          God knows what he would have done to me if he’d caught me. He once threw an eraser at my head for speaking German with a Texas accent. In his defense, he grew up in a post-war Yugoslavian concentration camp.

      • blazeknave@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        I just wrote really small on a paper in my glasses case, or hidden data in the depths of my TI86.

        We love Nightcrawler in this house.

    • michaelmrose@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 months ago

      Actually if you read the article ChatGPT is horrible at math a modified version where chatGPT was fed the correct answers with the problem didn’t make the kids stupider but it didn’t make them any better either because they mostly just asked it for the answers.

  • terminhell@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    5 months ago

    Maybe, if the system taught more of HOW to think and not WHAT. Basically more critical thinking/deduction.

    This same kinda topic came up back when I was in middle/highschool when search engines became wide spread.

    However, LLM’s shouldn’t be trusted for factual anything, same as Joe blows blog on some random subject. Did they forget to teach cross referencing too? I’m sounding too bitter and old so I’ll stop.

    • ChickenLadyLovesLife@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 months ago

      However, LLM’s shouldn’t be trusted for factual anything, same as Joe blows blog on some random subject.

      Podcasts are 100% reliable tho

  • xelar@lemmy.ml
    link
    fedilink
    English
    arrow-up
    0
    ·
    5 months ago

    While I get that, AI could be handy for some subjects, where you wont put your future on. However using it extinsively for everything is quite an exaggeration.

        • blackbirdbiryani@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          5 months ago

          Because a huge part about learning is actually figuring out how to extract/summarise information from imperfect sources to solve related problems.

          If you use CHATGPT as a crutch because you’re too lazy to read between the lines and infer meaning from text, then you’re not exercising that particular skill.

          • billwashere@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            5 months ago

            I don’t disagree, but thats like saying using a calculator will hurt you in understanding higher order math. It’s a tool, not a crutch. I’ve used it many times to help me understand concepts just out of reach. I don’t trust anything LLMs implicitly but it can and does help me.

            • obbeel@lemmy.eco.br
              link
              fedilink
              English
              arrow-up
              0
              ·
              5 months ago

              ChatGPT hallucinations inspire me to search for real references. It teaches we cannot blindly trust on things that are said. Teachers will commonly reinforce they are correct.

            • WordBox@lemmy.world
              link
              fedilink
              English
              arrow-up
              0
              ·
              5 months ago

              Congrats but there’s a reason teachers ban calculators… And it’s not always for the pain.

              • billwashere@lemmy.world
                link
                fedilink
                English
                arrow-up
                0
                ·
                5 months ago

                Take a college physics test without a calculator if you wanna talk about pain. And I doubt you could find a single person who could calculate trig functions or logarithms long hand. At some point you move past the point to prove you can do arithmetic. It’s just not necessary.

                The real interesting thing here is whether an LLM is useful as a study aid. It looks like there is more research necessary. But an LLM is not smart. It’s a complicated next word predictor and they have been known to go off the rails for sure. And this article suggests its not as useful and you might think for new learners.

              • Skates@feddit.nl
                link
                fedilink
                English
                arrow-up
                0
                ·
                5 months ago

                There are many reasons for why some teachers do some things.

                We should not forget that one of them is “because they’re useless cunts who have no idea what they’re doing and they’re just powertripping their way through some kids’ education until the next paycheck”.

                • Zoot@reddthat.com
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  edit-2
                  5 months ago

                  Not knowing how to add 6 + 8 just because a calculator is always available isn’t okay.

                  I have friends in my DnD session who have to count the numbers together on their fingers. I feel bad for the person. Don’t blame a teacher for wanting you to be a smarter more efficient and productive person, for banning a calculator.

              • assassin_aragorn@lemmy.world
                link
                fedilink
                English
                arrow-up
                0
                ·
                5 months ago

                In some cases I’d argue, as an engineer, that having no calculator makes students better at advanced math and problem solving. It forces you to work with the variables and understand how to do the derivation. You learn a lot more manipulating the ideal gas formula as variables and then plugging in numbers at the end, versus adding numbers to start with. You start to implicitly understand the direct and inverse relationships with variables.

                Plus, learning to directly use variables is very helpful for coding. And it makes problem solving much more of a focus. I once didn’t have enough time left in an exam to come to a final numerical answer, so I instead wrote out exactly what steps I would take to get the answer – which included doing some graphical solutions on a graphing calculator. I wrote how to use all the results, and I ended up with full credit for the question.

                To me, that is the ultimate goal of math and problem solving education. The student should be able to describe how to solve the problem even without the tools to find the exact answer.

  • N0body@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    5 months ago

    Traditional instruction gave the same result as a bleeding edge ChatGPT tutorial bot. Imagine what would happen if a tiny fraction of the billions spent to develop this technology went into funding improved traditional instruction.

    Better paid teachers, better resources, studies geared at optimizing traditional instruction, etc.

    Move fast and break things was always a stupid goal. Turbocharging it with all this money is killing the tried and true options that actually produce results, while straining the power grid and worsening global warming.

    • otp@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 months ago

      Traditional instruction gave the same result as a bleeding edge ChatGPT tutorial bot.

      Interesting way of looking at it. I disagree with your conclusion about the study, though.

      It seems like the AI tool would be helpful for things like assignments rather than tests. I think it’s intellectually dishonest to ignore the gains in some environments because it doesn’t have gains in others.

      You’re also comparing a young technology to methods that have been adapted over hundreds of thousands of years. Was the first automobile entirely superior to every horse?

      I get that some people just hate AI because it’s AI. For the people interested in nuance, I think this study is interesting. I think other studies will seek to build on it.

      • Kalysta@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        The point of assignments is to help study for your test.

        Homework is forced study. If you’re just handed the answers, you will do shit on the test.

        • otp@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          0
          ·
          5 months ago

          The point of assignments is to help study for your test.

          To me, “assignment” is more of a project. Not rote practice. Applying knowledge to a bit of a longer term, multi-part project.

      • Elvith Ma'for@feddit.org
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        It’s the other way round: Education makes for less gullible people and for workers that demand more rights more freely and easily - and then those are coming for their yachts…

    • Petter1@lemm.ee
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 months ago

      Imagine all the money spent on war would be invested into education 🫣what a beautiful world we would live in.

    • littlewonder@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 months ago

      The first sentence of this comment says everything. If a technology that is still ironing out its capabilities is able to get kids almost to the level of in-person instruction, think of the potential when used in tandem with teachers, or even when it has matured into a polished version of itself.

      How many of these kids knew how to leverage a GPT while avoiding common pitfalls? Would they have performed even better if given info on creating prompts for studying?

      LLMs/GPT, and other forms of the AI boogeyman, are all just a tool we can use to augment education when it makes sense. Just like the introduction of calculators or the internet, AI isn’t going to be the easy button, nor is it going to steal all teacher’s jobs. These tools need to be studied, trained for, and applied purposely in order to be most effective.

      • petrol_sniff_king@lemmy.blahaj.zone
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        are all just a tool
        just a tool
        it’s just a tool
        a tool is a tool
        all are just tools
        it’s no more than a tool
        it’s just a tool
        it’s a tool we can use
        one of our many tools
        it’s only a tool
        these are just tools
        a tool for thee, a tool for me

        guns don’t kill people, people kill people
        the solution is simple:
        teach drunk people not to shoot their guns so much
        unless they want to
        that is the American way

        tanks don’t kill people, people kill people
        the solution is simple:
        teach drunk people not to shoot their tanks so much
        the barista who offered them soy milk
        wasn’t implying anything about their T levels
        that is the American way

        Thanks for reminding me that AI is just tools, friend.
        My memory is not so good.
        I often can’t
        remember

        • littlewonder@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          5 months ago

          Ok, I’m going to reply like you’re being serious. It is a tool and it’s out there and it’s not going anywhere. Do we allow ourselves to imagine how it can be improved to help students or do we ignore it and act like it won’t ever be something students need to learn?

    • technocrit@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 months ago

      The education system is primarily about controlling bodies and minds. So any actual education is counter-productive.

  • Vanth@reddthat.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    5 months ago

    I’m not entirely sold on the argument I lay out here, but this is where I would start were I to defend using chatGPT in school as they laid out in their experiment.

    It’s a tool. Just like a calculator. If a kid learns and does all their homework with a calculator, then suddenly it’s taken away for a test, of course they will do poorly. Contrary to what we were warned about as kids though, each of us does carry a calculator around in our pocket at nearly all times.

    We’re not far off from having an AI assistant with us 24/7 is feasible. Why not teach kids to use the tools they will have in their pocket for the rest of their lives?

    • Schal330@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 months ago

      As adults we are dubious of the results that AI gives us. We take the answers with a handful of salt and I feel like over the years we have built up a skillset for using search engines for answers and sifting through the results. Kids haven’t got years of experience of this and so they may take what is said to be true and not question the results.

      As you say, the kids should be taught to use the tool properly, and verify the answers. AI is going to be forced onto us whether we like it or not, people should be empowered to use it and not accept what it puts out as gospel.

      • Petter1@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        This is true for the whole internet, not only AI Chatbots. Kids need to get teached that there is BS around. In fact kids had to learn that even pre-internet. Every human has to learn that you can not blindly trust anything, that one has to think critically. This is nothing new. AI chatbots just show how flawed human education is these days.

    • filister@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 months ago

      I think here you also need to teach your kid not to trust unconditionally this tool and to question the quality of the tool. As well as teaching it how to write better prompts, this is the same like with Google, if you put shitty queries you will get subpar results.

      And believe me I have seen plenty of tech people asking the most lame prompts.

      • otp@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        I remember teachers telling us not to trust the calculators. What if we hit the wrong key? Lol

        Some things never change.

        • Deceptichum@quokk.au
          link
          fedilink
          English
          arrow-up
          0
          ·
          5 months ago

          I remember the teachers telling us not to trust Wikipedia, but they had utmost faith in the shitty old books that were probably never verified by another human before being published.

            • Deceptichum@quokk.au
              link
              fedilink
              English
              arrow-up
              0
              ·
              5 months ago

              Eh I find they’re usually from a more direct source. The schoolbooks are just information sourced from who knows where else.

              • qarbone@lemmy.world
                link
                fedilink
                English
                arrow-up
                0
                ·
                5 months ago

                I don’t know about your textbooks and what ages you’re referring to but I remember many of my technical textbooks had citations in the back.

                • bluewing@lemm.ee
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  5 months ago

                  Yep, students these days have no idea about the back of their books and how useful the index can be and the citations after that.

                  Even after repeatedly pointing it out, they still don’t make use of it. Despite the index being nearly a cheat code in itself.

          • otp@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            0
            ·
            5 months ago

            You’re right. The commenter who made the comparison to Wikipedia made a better point.

      • wesley@yall.theatl.social
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        Yeah it’s like if you had a calculator and 10% of the time it gave you the wrong answer. Would that be a good tool for learning? We should be careful when using these tools and understand their limitations. Gen AI may give you an answer that happens to be correct some of the time (maybe even most of the time!) but they do not have the ability to actually reason. This is why they give back answers that we understand intuitively are incorrect (like putting glue on pizza), but sometimes the mistakes can be less intuitive or subtle which is worse in my opinion.

      • Womble@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        Ask your calculator what 1-(1-1e-99) is and see if it never halucinates (confidently gives an incorrect answer) still.

  • prosp3kt@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    5 months ago

    There are a part here that sounds interesting

    The students who used it did spectacularly better on the practice problems, solving 127 percent more of them correctly compared with students who did their practice work without any high-tech aids. But on a test afterwards, these AI-tutored students did no better.

    Do you think that these students that used ChatGPT can do the exercises “the old fashioned way”? For me it was a nightmare try to resolve a calculus problem just with the trash books that doesn’t explain a damn fuck, I have to go to different resources, wolphram, youtube, but what happened when there was a problem that wasnt well explained in any resource?. I hate openAI, I want to punch Altman in the face. But this doesn’t mean we have to bait this hard in the title.

    • MBM@lemmings.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 months ago

      Maybe it’s an accident but you left out your fellow students and the teacher, in my eyes the most useful resources

  • Ilandar@aussie.zone
    link
    fedilink
    English
    arrow-up
    0
    ·
    5 months ago

    What do the results of the third group suggest? AI doesn’t appear to have hindered their ability to manage by themselves under test conditions, but it did help them significantly with their practice results. You could argue the positive reinforcement an AI tutor can provide during test preparations might help some students with their confidence and pre-exam nerves, which will allow them to perform closer to their best under exam conditions.

    • nyan@lemmy.cafe
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      5 months ago

      It suggests that the best the chatbot can do, after being carefully tailored for its job, is no better than the old methods (because the goal is for the students to be able to handle the subject matter without having to check every common operation with a third party, regardless of whether that’s a chatbot or a textbook, and the test is the best indicator of that). Therefore, spending the electricity to run an educational chatbot for highschoolers isn’t justified at this time, but it’s probably worth rechecking in a few years to see if its results have improved. It may also be worth doing extended testing to determine whether there are specific subsets of the student body that benefit more from the chatbot than others. And allowing the students to seek out an untailored chatbot on their own is strongly counterindicated.

    • cheddar@programming.dev
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 months ago

      Yep. But the post title suggest that all students who used ChatGPT did worse. Fuck this clickbait shit.

  • Allero@lemmy.today
    link
    fedilink
    English
    arrow-up
    0
    ·
    5 months ago

    Perhaps unsurprisingly. Any sort of “assistance” with answers will do that.

    Students have to learn why things work the way they do, and they won’t be able to grasp it without going ahead and doing every piece manually.