Get the latest tech news

A Replacement for BERT


We’re on a journey to advance and democratize artificial intelligence through open source and open science.

This blog post introduces ModernBERT, a family of state-of-the-art encoder-only models representing improvements over older generation encoders across the board, with a 8192 sequence length, better downstream performance and much faster processing. Process We stick to the original BERT’s training recipe, with some slight upgrades inspired by subsequent work: we remove the Next-Sentence Prediction objective, since then shown to add overhead for no clear gains, and increase the masking rate from 15% to 30%. To encourage this, we’re opening a call for demos until January 10th, 2025: the 5 best ones will get added to this post in a showcase section and win a $100 (or local currency equivalent) Amazon gift card, as well as a 6-month HuggingFace Pro subscription!

Get the Android app

Or read this on Hacker News

Read more on:

Photo of replacement

replacement

Photo of Bert

Bert

Related news:

News photo

Monitor replacement – using a projector for a home office setup

News photo

Nokolexbor: Drop-in replacement for Nokogiri. 5.2x faster at parsing HTML

News photo

Break your Google Pixel 9 Pro Fold main screen? A replacement costs the price of a new phone