Get the latest tech news

Reflection 70B saga continues as training data provider releases post-mortem report


The more data the Reflection 70B creators publish about the model, the more evidence the open source AI community has to pore over.

Despite the setbacks, Chaudhary believes the “reflection tuning” approach — in which a model is given time to check its responses for accuracy before outputting them to a user — has potential and encourages further experimentation by the AI community. Yuchen Jin, co-founder and CTO of Hyperbolic Labs, a startup that offers cloud-based GPUs and other AI services on demand who initially worked hard and late to host Reflection 70B before criticizing Shumer over its discrepancies, also voiced skepticism on X toward Chaudhary’s post-mortem report, pointing out that Chaudhary’s claims on X that he “reproduced all but two of the initially reported scores,” don’t actually match with the data he provided, which show at least 4 benchmarks changing scores from before to now. Meanwhile, another Redditor, “DangerousBenefit” looked into the training data Chaudhary released today and f ound it was filled with many instances of the phrase “as an AI language model,” which indicates it could be generated primarily from OpenAI’s ChatGPT and likely wasn’t properly cleaned.

Get the Android app

Or read this on Venture Beat

Read more on:

Photo of training

training

Photo of reflection

reflection

Photo of Reflection 70B saga

Reflection 70B saga

Related news:

News photo

The Norwegian model of training to run

News photo

LinkedIn scraped user data for training before updating its terms of service

News photo

The Auto Industry Finally Has a Plan to Stop Electric Vehicle Fires | EVs can burn for days, firefighters have little or no training to deal with them, and conventional equipment isn’t effective. But new technology is being developed that should make all the difference