Subscribe
Sign in
Home
Podcast
The AiEdge Courses
The LLM Book
Table of Content
Archive
About
Latest
Top
Discussions
Last Week to Register to the Build Production-Ready LLMs From Scratch Course!
From Prototype to Production: Ship Scalable LLM Systems in 6 Weeks
May 19
•
Damien Benveniste
9
Share this post
The AiEdge Newsletter
Last Week to Register to the Build Production-Ready LLMs From Scratch Course!
Copy link
Facebook
Email
Notes
More
April 2025
All About The Modern Positional Encodings In LLMs
The Positional Encoding in LLMs may appear somewhat mysterious the first time we come across the concept, and for good reasons!
Apr 28
•
Damien Benveniste
12
Share this post
The AiEdge Newsletter
All About The Modern Positional Encodings In LLMs
Copy link
Facebook
Email
Notes
More
Join us for a Free LIVE Coding Event: Build The Self-Attention in PyTorch From Scratch
Next Friday, I am inviting you to join me for an exciting live coding event. It is a completely free event where I will explain the basics of the…
Apr 25
•
Damien Benveniste
14
Share this post
The AiEdge Newsletter
Join us for a Free LIVE Coding Event: Build The Self-Attention in PyTorch From Scratch
Copy link
Facebook
Email
Notes
More
Build Production-Ready LLMs From Scratch
From Prototype to Production: Ship Scalable LLM Systems in 6 Weeks
Apr 21
•
Damien Benveniste
18
Share this post
The AiEdge Newsletter
Build Production-Ready LLMs From Scratch
Copy link
Facebook
Email
Notes
More
March 2025
Chapter 4 of The Big Book of Large Language Models is Here!
Chapter 4 of the Big Book of Large Language Models is finally here!
Mar 31
•
Damien Benveniste
21
Share this post
The AiEdge Newsletter
Chapter 4 of The Big Book of Large Language Models is Here!
Copy link
Facebook
Email
Notes
More
2
Reduce AI Model Operational Costs With Quantization Techniques
A deep dive into quantization and precision levels
Mar 26
•
Damien Benveniste
21
Share this post
The AiEdge Newsletter
Reduce AI Model Operational Costs With Quantization Techniques
Copy link
Facebook
Email
Notes
More
3
How To Construct Self-Attention Mechanisms For Arbitrary Long Sequences
Toward Infinite Sequence Lengths
Mar 21
•
Damien Benveniste
16
Share this post
The AiEdge Newsletter
How To Construct Self-Attention Mechanisms For Arbitrary Long Sequences
Copy link
Facebook
Email
Notes
More
How To Improve Decoding Latency With Faster Self-Attention Mechanisms
In LLMs, handling large sequences is not enough, we need to make sure the decoding process is fast.
Mar 12
•
Damien Benveniste
17
Share this post
The AiEdge Newsletter
How To Improve Decoding Latency With Faster Self-Attention Mechanisms
Copy link
Facebook
Email
Notes
More
How To Reduce The Memory Usage Of The Self-Attention
With a bit of magic, we take a very inefficient computation like the Self-Attention and make it super memory-optimized for the specific hardware we use…
Mar 6
•
Damien Benveniste
15
Share this post
The AiEdge Newsletter
How To Reduce The Memory Usage Of The Self-Attention
Copy link
Facebook
Email
Notes
More
February 2025
How To Linearize The Attention Mechanism!
Today, we talk about how to engineer attention mechanisms in O(n) complexity instead of O(n2).
Feb 26
•
Damien Benveniste
15
Share this post
The AiEdge Newsletter
How To Linearize The Attention Mechanism!
Copy link
Facebook
Email
Notes
More
Understanding The Sparse Transformers!
The First Sparse Attention: Sparse Transformers
Feb 19
•
Damien Benveniste
10
Share this post
The AiEdge Newsletter
Understanding The Sparse Transformers!
Copy link
Facebook
Email
Notes
More
Attention Is All You Need: The Original Transformer Architecture
This newsletter is the latest chapter of the Big Book of Large Language Models. You can find the preview here, and the full chapter is available in this…
Feb 12
•
Damien Benveniste
22
Share this post
The AiEdge Newsletter
Attention Is All You Need: The Original Transformer Architecture
Copy link
Facebook
Email
Notes
More
Share
Copy link
Facebook
Email
Notes
More
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts