• AbouBenAdhem@lemmy.world
    link
    fedilink
    English
    arrow-up
    16
    ·
    2 days ago

    AlphaEvolve verifies, runs and scores the proposed programs using automated evaluation metrics. These metrics provide an objective, quantifiable assessment of each solution’s accuracy and quality.

    Yeah, that’s the way genetic algorithms have worked for decades. Have they figured out a way to turn those evaluation metrics directly into code improvements, or do they just keep doing a bunch of rounds of trial and error?

    • KingRandomGuy@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      9 hours ago

      The general framework for evolutionary methods/genetic algorithms is indeed old but it’s extremely broad. What matters is how you actually mutate the algorithm being run given feedback. In this case, they’re using the same framework as genetic algorithms (iteratively building up solutions by repeatedly modifying an existing attempt after receiving feedback) but they use an LLM for two things:

      1. Overall better sampling (the LLM has better heuristics for figuring out what to fix compared to handwritten techniques), meaning higher efficiency at finding a working solution.

      2. “Open set” mutations: you don’t need to pre-define what changes can be made to the solution. The LLM can generate arbitrary mutations instead. In particular, AlphaEvolve can modify entire codebases as mutations, whereas prior work only modified single functions.

      The “Related Work” (section 5) section of their whitepaper is probably what you’re looking for, see here.

  • PlantJam@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    ·
    edit-2
    1 day ago

    Can’t wait for my boss to come up with half a working solution with this and expect me to fix/finish it!