• mkwt@lemmy.world
    link
    fedilink
    arrow-up
    2
    arrow-down
    3
    ·
    3 months ago

    “gradient descent” is a jargon word for one kind of training method.

    • PM_ME_VINTAGE_30S [he/him]@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      4
      ·
      edit-2
      3 months ago

      “Gradient descent” ≈ on a “hilly” (mathematical) surface, try to find the lowest point by finding the lowest point near an initial guess. “Gradient” is basically the steepness, or rate that the thing you’re trying to optimize changes as you move through “space”. The gradient tells you mathematically which direction you need to go to reach the bottom. “Descent” means “try to find the minimum”.

      I’m glossing over a lot of details, particularly what a “surface” actually means in the high dimensional spaces that AI uses, but a lot of problems in mathematical optimization are solved like this. And one of the steps in training an AI agent is to do an optimization, which often does use a gradient descent algorithm. That being said, not every process that uses gradient descent is necessarily AI or even machine learning. I’m actually taking a course this semester where a bunch of my professor’s research is in optimization algorithms that don’t use a gradient descent!

      • mbtrhcs
        link
        fedilink
        arrow-up
        5
        ·
        3 months ago

        This is a decent explanation of gradient descent but I’m pretty sure the meme is referencing the color gradients often used to highlight when something is AI generated haha