home.social

#foundationmodel — Public Fediverse posts

Live and recent posts from across the Fediverse tagged #foundationmodel, aggregated by home.social.

  1. RE: mastodon.social/@thetransmitte

    Tl;dr: transformers didn’t learn the underlying principle.

    “Keyon Vafa and his colleagues simulated tens of millions of planetary orbits from Newton’s laws and trained a transformer on the resulting sequences. The model predicted future positions with very high accuracy. But when they fine-tuned it to infer the underlying gravitational force vectors, the model produced nonsense. The implied laws of gravitation were different depending on which subset of data the researchers examined. The transformer had assembled a patchwork of heuristics, accurate for every solar system in the training set, but it hadn’t discovered the universal gravitational principle. Without that principle, the model could predict the movement of points of light in the sky but could never send a rocket to the moon.”

    Also link to the preprint: arxiv.org/abs/2507.06952

    #neuroscience #prediction #FoundationModel

  2. RE: mastodon.social/@thetransmitte

    Tl;dr: transformers didn’t learn the underlying principle.

    “Keyon Vafa and his colleagues simulated tens of millions of planetary orbits from Newton’s laws and trained a transformer on the resulting sequences. The model predicted future positions with very high accuracy. But when they fine-tuned it to infer the underlying gravitational force vectors, the model produced nonsense. The implied laws of gravitation were different depending on which subset of data the researchers examined. The transformer had assembled a patchwork of heuristics, accurate for every solar system in the training set, but it hadn’t discovered the universal gravitational principle. Without that principle, the model could predict the movement of points of light in the sky but could never send a rocket to the moon.”

    Also link to the preprint: arxiv.org/abs/2507.06952

    #neuroscience #prediction #FoundationModel

  3. RE: mastodon.social/@thetransmitte

    Tl;dr: transformers didn’t learn the underlying principle.

    “Keyon Vafa and his colleagues simulated tens of millions of planetary orbits from Newton’s laws and trained a transformer on the resulting sequences. The model predicted future positions with very high accuracy. But when they fine-tuned it to infer the underlying gravitational force vectors, the model produced nonsense. The implied laws of gravitation were different depending on which subset of data the researchers examined. The transformer had assembled a patchwork of heuristics, accurate for every solar system in the training set, but it hadn’t discovered the universal gravitational principle. Without that principle, the model could predict the movement of points of light in the sky but could never send a rocket to the moon.”

    Also link to the preprint: arxiv.org/abs/2507.06952

    #neuroscience #prediction #FoundationModel

  4. RE: mastodon.social/@thetransmitte

    Tl;dr: transformers didn’t learn the underlying principle.

    “Keyon Vafa and his colleagues simulated tens of millions of planetary orbits from Newton’s laws and trained a transformer on the resulting sequences. The model predicted future positions with very high accuracy. But when they fine-tuned it to infer the underlying gravitational force vectors, the model produced nonsense. The implied laws of gravitation were different depending on which subset of data the researchers examined. The transformer had assembled a patchwork of heuristics, accurate for every solar system in the training set, but it hadn’t discovered the universal gravitational principle. Without that principle, the model could predict the movement of points of light in the sky but could never send a rocket to the moon.”

    Also link to the preprint: arxiv.org/abs/2507.06952

    #neuroscience #prediction #FoundationModel

  5. Just heard a podcast with the #ceo of #amazon #aws Matt Garman referring to "bedrock" models to what others refer to as a #foundationmodel . That's a big #middlefinger to #Stanford !