5 min read

My reading list [wip]

Table of contents
  1. Blogs / Articles
  2. Books
  3. Articles / papers
    1. On working with LLMs
    2. Design patterns for AI agents
    3. Tokenization
    4. Memory
    5. LLM architecture

I am often asked for recommendations on reading materials for those working with AI/ML and LLMs. Here is a short list of what I read (and re-read!). This is a constantly evolving list. If you have something to recommend, let me know in the comments below!

Blogs / Articles

A small list of blogs that I follow on LLMs, applied ML, engineering and broader AI trends (in no particular order).

From the AI labs:

  • Research — Anthropic is an AI safety and research company that’s working to build reliable, interpretable, and steerable AI systems.

Business:

  • Acquired Podcast — This is what I listen to during long commutes, road trips, and when doing chores at home. Note that the episodes are seriously long-form (typically ~3h long!) and require some commitment. These have really opened my eyes to “company-making” and inspired me over the years! Some of my favorite episodes: Epic a.k.a. MyHealthOnline (2025), Ikea (2024), Costco (2023), Amazon and AWS (2022), Nvidia 1 & 2 (2022), TSMC (2021) and a follow-up with Morris Chang (2025), AirBnB (2020), Google Maps (2019), Slack (2019).

Cybersecurity:

Books

  • Artificial Intelligence: A Modern Approach (Stuart Russell and Peter Norvig) – a must-read introduction to AI and agents that is even more pertinent today, even though it was (first) written some 10+ years before LLMs and AI agents became mainstream.

Articles / papers

This list is not meant to be comprehensive. It is simply my personal bookmark list; articles I frequently share when asked about specific topics. Some of these are from blogs mentioned above. Again, these are not in any particular order.

On working with LLMs

Design patterns for AI agents

Good prompt design guides and references:

Tokenization

The importance of tokenization is often overlooked by folks working with LLMs. To me, many of the quirks of LLMs come down to tokenization strategies (and trade offs).

  • Tokenization in large language models (Sean Trott) [2024] – a gentle introduction to tokenization; interesting discussion on whether morphology really matters with tokenization (i.e. is it important for LLMs to truly understand language rules?).
  • Let’s build the GPT tokenizer (Andrej Karpathy) [2024] – a must-watch for understanding tokenization (note: 2h 13min long video). There is also a pretty cool tutorial with code, though it doesn’t capture many of the nuances and insights in Karpathy’s video.

Memory

LLM architecture