Subscribe
Sign in
Home
Podcast
The AiEdge Courses
The LLM Book
Table of Content
Archive
About
Latest
Top
Discussions
Chapter 4 of The Big Book of Large Language Models is Here!
Chapter 4 of the Big Book of Large Language Models is finally here!
Mar 31
•
Damien Benveniste
18
Share this post
The AiEdge Newsletter
Chapter 4 of The Big Book of Large Language Models is Here!
Copy link
Facebook
Email
Notes
More
Reduce AI Model Operational Costs With Quantization Techniques
A deep dive into quantization and precision levels
Mar 26
•
Damien Benveniste
18
Share this post
The AiEdge Newsletter
Reduce AI Model Operational Costs With Quantization Techniques
Copy link
Facebook
Email
Notes
More
3
How To Construct Self-Attention Mechanisms For Arbitrary Long Sequences
Toward Infinite Sequence Lengths
Mar 21
•
Damien Benveniste
14
Share this post
The AiEdge Newsletter
How To Construct Self-Attention Mechanisms For Arbitrary Long Sequences
Copy link
Facebook
Email
Notes
More
How To Improve Decoding Latency With Faster Self-Attention Mechanisms
In LLMs, handling large sequences is not enough, we need to make sure the decoding process is fast.
Mar 12
•
Damien Benveniste
15
Share this post
The AiEdge Newsletter
How To Improve Decoding Latency With Faster Self-Attention Mechanisms
Copy link
Facebook
Email
Notes
More
How To Reduce The Memory Usage Of The Self-Attention
With a bit of magic, we take a very inefficient computation like the Self-Attention and make it super memory-optimized for the specific hardware we use…
Mar 6
•
Damien Benveniste
13
Share this post
The AiEdge Newsletter
How To Reduce The Memory Usage Of The Self-Attention
Copy link
Facebook
Email
Notes
More
February 2025
How To Linearize The Attention Mechanism!
Today, we talk about how to engineer attention mechanisms in O(n) complexity instead of O(n2).
Feb 26
•
Damien Benveniste
15
Share this post
The AiEdge Newsletter
How To Linearize The Attention Mechanism!
Copy link
Facebook
Email
Notes
More
Understanding The Sparse Transformers!
The First Sparse Attention: Sparse Transformers
Feb 19
•
Damien Benveniste
10
Share this post
The AiEdge Newsletter
Understanding The Sparse Transformers!
Copy link
Facebook
Email
Notes
More
Attention Is All You Need: The Original Transformer Architecture
This newsletter is the latest chapter of the Big Book of Large Language Models. You can find the preview here, and the full chapter is available in this…
Feb 12
•
Damien Benveniste
18
Share this post
The AiEdge Newsletter
Attention Is All You Need: The Original Transformer Architecture
Copy link
Facebook
Email
Notes
More
January 2025
Introducing The Big Book of Large Language Models!
For the past years, I have been creating educational content around machine learning and, specifically, large language models.
Jan 30
•
Damien Benveniste
110
Share this post
The AiEdge Newsletter
Introducing The Big Book of Large Language Models!
Copy link
Facebook
Email
Notes
More
3
Transforming Text Into Tokens: The WordPiece VS The Byte Pair Encoding Algorithm
The WordPiece tokenization algorithm
Jan 27
•
Damien Benveniste
18
Share this post
The AiEdge Newsletter
Transforming Text Into Tokens: The WordPiece VS The Byte Pair Encoding Algorithm
Copy link
Facebook
Email
Notes
More
The LLM APIs Landscape
How to make a choice?
Jan 20
•
Damien Benveniste
25
Share this post
The AiEdge Newsletter
The LLM APIs Landscape
Copy link
Facebook
Email
Notes
More
3
The Machine Learning Fundamentals Bootcamp V2: Live Sessions Starting Soon!
Get your Machine Learning Career Started!
Jan 14
•
Damien Benveniste
6
Share this post
The AiEdge Newsletter
The Machine Learning Fundamentals Bootcamp V2: Live Sessions Starting Soon!
Copy link
Facebook
Email
Notes
More
Share
Copy link
Facebook
Email
Notes
More
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts