The AiEdge Newsletter is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber. Why do we keep talking about "tokens" in LLMs instead of words? It happens to be much more efficient to break the words into sub-words (tokens) for model performance!
Share this post
From Words to Tokens: The Byte-Pair Encoding…
Share this post
The AiEdge Newsletter is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber. Why do we keep talking about "tokens" in LLMs instead of words? It happens to be much more efficient to break the words into sub-words (tokens) for model performance!