Skip to content

Latest commit

 

History

History
36 lines (20 loc) · 1.6 KB

README.md

File metadata and controls

36 lines (20 loc) · 1.6 KB

A Single Usage Is All You Need

Awards:

5 high school credit points in data science. Grade: 99%

Abstarct:

I developed and published an open-source efficient text-generation algorithm called grouped sampling to enable cheap and accessible AI text-generation services for everyone.

Causal language models are state-of-the-art text generation models that power many popular products like chat-GPT. The naive text generation algorithm requires x usages of a causal language model to generate x words, which makes it inefficient.

Grouped sampling is an alternative algorithm that manipulates the input text before passing it to the model,

forcing the model to predict the entire output at once.

Grouped sampling only requires one use of a causal language model to generate text of any length, making it much more efficient.

I compared grouped sampling and the naive algorithm in translating TED talks.

The naive algorithm required 33.049 GPU hours that costs $17.87.

Grouped sampling required 0.028 GPU hours that costs $0.015.

Grouped sampling translated more accurately by 5%-24%, measured using BERT scores.

In conclusion, grouped sampling is an accurate and efficient text-generation technique.

It is 1180 times faster and cheaper to run than the naive algorithm.