Get the latest tech news

Using a LLM to compress text


Introduction Large language models are trained on huge datasets of text to learn the relationships and contexts of words within larger documents. These relationships are what allows the model to generate text. Recently I've read concerns about LLMs being trained on copyrighted text and reproducing it. This got me thinking:

Large language models are trained on huge datasets of text to learn the relationships and contexts of words within larger documents. I figured that, for the most part, many texts contain sections that would naturally align with the language relationships the model has learned. For the first, I decided to use the first chapter of " Alice's Adventures in Wonderland" as I assumed it would be in the model's training data.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of Text

Text

Photo of LLM

LLM

Related news:

News photo

Stardog’s Karaoke offers on-premises, zero hallucination LLM solution for enterprises

News photo

Show HN: FileKitty – Combine and label text files for LLM prompt contexts

News photo

Nebula LLM by Symbl.ai Beats Claude 3, GPT-4, Llama 3 on Emotional Intelligence