• @btaf45@lemmy.world
    link
    fedilink
    English
    16 months ago

    Each of them is tuned by the input data, which is a long and expensive process.

    But surely the history of how this data is tuned/created is kept track of. If you want to know how a specific value is created you ideally should be able to reference the entire history of how it changed over time.

    I’m not saying this would be easy, but you could have people whose entire job is to understand this and with unlimited amounts of time to do so if it is important enough. And it seems like it would be important enough and such people would be very valuable.

    Now that AI is first taking off is exactly the time to establish the precedent that we do not let it escape the understanding and control of humans.

    • Traister101
      link
      English
      36 months ago

      Well the thing is that good AI models aren’t manually tuned. There’s not some poor intern turning a little knob and seeing if it’s slightly more accurate, it happens on its own. The more little knobs there are the better the model is. This means essentially you have no idea how any knob ultimately effects every other knob cause there’s thousands of them and any little change can completely change something else.

      Look at “simple” AI for playing like Super Mario World https://youtu.be/qv6UVOQ0F44 shits already pretty complicated and this thing is stupid. It’s only capable of playing the first level

    • @BreadstickNinja@lemmy.world
      link
      fedilink
      English
      36 months ago

      The issue is that the values of the parameters don’t correspond to traditional variables. Concepts in AI are not represented with discrete variables and quantities. A concept may be represented in a distributed way across thousands or millions of neurons. You can look at each individual neuron and say, oh, this neuron’s weight is 0.7142, and this neuron’s weight is 0.2193, etc., across all the billions of neurons in your model, but you’re not going to be able to connect a concept from the output back to the behavior of those individual parameters because they only work in aggregate.

      You can only know that an AI system knows a concept based on its behavior and output, not from individual neurons. And AI systems are quite like humans in that regard. If your professor wants to know if you understand calculus, or if the DMV wants to know if you can safely drive a car, they give you a test: can you perform the desired output behavior (a correct answer, a safe drive) when prompted? Understanding how an idea is represented across billions of parameters in an AI system is no more feasible than your professor trying to confirm you understand calculus by scanning your brain to find the exact neuronal connections that represent that knowledge.

    • @vrighter@discuss.tchncs.de
      link
      fedilink
      English
      1
      edit-2
      6 months ago

      imagine you have a simple equation:

      ax + by + cz

      The machine learning part finds values for the coefficients a, b and c.

      Even if you stepped through the code, you will see the equation be evaluated just fine, but you still won’t know why the coefficients are the way they are. Oh and there are literally billions of coefficients.