Blog

#17: Red Teaming: Why Breaking LLMs Might Save Them
In earlier posts we looked at how LLMs are given direction through supervised fine-tuning / reinforcement learning from human feedback (RLHF) and how LLMs are kept safe. But just because a model can follow instructions doesn’t mean it should follow all of them.
6/22/2025
#16: Understanding How Safety is Built into LLMs
In my previous post, we explored how Large Language Models (LLMs) are trained to align with human preferences. Now, let's examine the broader landscape of LLM safety. As these models become increasingly integrated into our daily lives, ensuring their safe and ethical operation is critical.
6/2/2025
#15: Teaching LLMs to Align with Us: Supervised Fine-Tuning and RLHF
We’ve seen how large language models (LLMs) are pretrained on vast amounts of data and how they generate text one token at a time during inference.
5/25/2025
#14: How LLMs Generate Text: A Peek Inside the Inference Process
In the last few posts, we walked through how large language models (LLMs) are built — from collecting training data to breaking down words into tokens, and finally training the model on massive datasets. Today, we will discuss the inference process, which is the reason LLMs exist.
5/17/2025
#13: How LLMs Learn: A Beginner’s Guide to Training
In our last few posts, we explored what large language models (LLMs) are, how their data is collected and curated, and how that data is broken down into tokens. Now comes the question at the heart of it all: How do these models actually learn?
4/26/2025
#12 Understanding Tokenization in LLMs
In earlier posts, we explored how large language models (LLMs) work at a high level, followed by a deeper dive into how data is collected for training these models.
4/6/2025
#11 How Community Notes Work and Is it Working?
Taking a break from the LLM posts, this week I want to discuss Community Notes as pioneered by X/Twitter a few years ago.
3/23/2025
#10 Data Collection and Curation for LLM Training
In my last post, I discussed how LLMs work at a high level.
3/16/2025
#9 How LLMs Work: A Beginner’s Guide
I am sure that most of you reading this blog are aware of Large Language Models (LLMs) like ChatGPT, Claude, and DeepSeek.
3/2/2025
#8 Behind the Scenes: Amazon's OAK4 Fulfillment Center
Every second, Amazon processes over 100 orders.
2/15/2025
#7 Carl Zeiss: The Hidden Optics Powerhouse Behind Every Advanced Chip
In a previous article, I discussed ASML and its dominance in the production of Extreme Ultraviolet (EUV) technology for chip fabrication.
2/10/2025
#6 Canon Tokki: The Unsung Hero Behind our Gadgets
When you think of cutting-edge display technology, names like Samsung, LG, and Apple might come to mind.
1/25/2025
#5 ASML: The Technology Powering the Modern Digital Era
Did you know that nearly all the chips that are crucial to modern technology are printed by just a singular company?
1/12/2025
#4 LLMs and Sustainability
Did you know that the training of GPT-4 took 50–62 GWh of energy?
1/4/2025
#3 Fortune Cookies and The Future of Marketing
Recently, my family and I went to Lake Tahoe to ski.
12/26/2024
#2 Learning Spanish and The Power Law Curve
Normal distributions and Power law curves rule our lives.
12/1/2024
#1 The Future is Here - Waymo
This past weekend I got lucky to try out my first ride in a self-driving vehicle, Waymo.
11/28/2024
Welcome to Tech Unpacked
Hello and welcome to Tech Unpacked!
11/24/2024