We will use Grok 3.5 (maybe we should call it 4), which has advanced reasoning, to rewrite the entire corpus of human knowledge, adding missing information and deleting errors.

Then retrain on that.

Far too much garbage in any foundation model trained on uncorrected data.

Source.

More Context

Source.

Source.

  • hansolo@lemmy.today
    link
    fedilink
    English
    arrow-up
    7
    ·
    edit-2
    2 days ago

    Musk probably heard about “synthetic data” training, which is where you use machine learning to create thousands of things that are typical-enough to be good training data. Microsoft uses it to take documents users upload to Office365, train the ML model, and then use that ML output to train an LLM so they can technically say “no, your data wasn’t used to train an LLM.” Because it trained the thing that trained the LLM.

    However, you can’t do that with LLM output and stuff like… History. WTF evidence and documents are the basis for the crap he wants to add? The hallucinations will just compound because who’s going to cross-check this other than Grok anyway?