6 March, 2025
-
Strangely, Matrix Multiplications on GPUs Run Faster When Given āPredictableā Data!
Great minds discuss
flopsflops per watt; average minds discuss data; small minds discuss architecture.
-
You could have designed state of the art Positional Encoding
A complex system that works is invariably found to have evolved from a simple system that worked
-
Smuggling arbitrary data through an emoji
Is it really possible to encode arbitrary data in a single emoji? ~ YES
-
The Tragedies of Reality Are Coming for You
āif you could take any machine learning subfield, and give all their resources to a different one, what are you killing and what are you boosting?ā - Robotics
-
Being a High-Leverage Generalist
Weāre told to pick a lane early, specialise hard, and climb the ladder in our chosen field. This advice made sense in a world of stable, well-defined industries. But that world is dead.
20 February, 2025
-
New Junior Developers Canāt Actually Code
Donāt let those AI tools blind you into thinking you are a good developer, get in the dirt, lurk in stackoverflow and fight with that random guy on a random reddit post on that thing
-
Deep Research, information vs. insight, and the nature of science
To an LLM, a novel discovery is indistinguishable from an error
-
Situation Awareness - The Decade Ahead
For the pro-AGI peeps
-
Making Deep Learning Go Brrrr From First Principles
How to efficiently make your GPUs go brrrrr
7 January, 2025
-
Why Deep Learning Works Even Though It Shouldnāt
why models always get better when they are bigger and deeper, even when the amount of data they consume stays the same or gets smaller?
-
200Bn Weights of Responsibility
The mental stress of developing LLMs.
-
Things we learned about LLMs in 2024
2024 wrapped for LLM space!
-
Is AI progress slowing down?
Is model-scaling dead and inference scaling the way forward?
20 December, 2024
-
OpenAIās o3: The grand finale of AI in 2024
An initial blog on the newly released o3 by OpenAI: A step change as influential as the release of GPT-4. Reasoning language models are the current and next big thing.
-
Building effective agents
A (relative) short guide on what are Agents and how to use them in real-world scenarios.
-
Outperforming Llama 70B with Llama 3B by scaling test-time compute
Outperforming Llama3.1 70B with Llama 3.2 3B on MATH-500 using Test-Time Compute (attempt to reverse engineer o1) by Researchers at HuggingFace.
A really good blog showing that (maybe) in future small LM may out-perform Large LM with more compute time/generations given efficient search (& reward) algorithms are used in inference.
-
The Invisible OS
The ultimate evolution the invisible AI operating system.
-
An intuitive introduction to text embeddings
A good read on how text-embeddings work.
6 December, 2024
-
AI research journey and advice by Jason Wei
Some advice around doing Research work in AI
-
Goodbye, Clean Code
Let clean code guide you. Then let it go.
-
What is SwiGLU?
A good & short explaination of how SwiGLU works.. (not the why part)
27 November,2024
-
The Problem with Reasoners
Current scenario around reasoning and the secret to getting better LLM results:
bigger sizelonger inference
-
Convolutional Neural Networks (CNNs / ConvNets) - CS231n
Gentle introduction to what and hows of CNN - StandfordCS231n
-
The Unreasonable Effectiveness of Recurrent Neural Networks
A great introduction to Recurrent Neural Networkes-RNN by Andrej Karpathy
-
A Visual Guide to Quantization
The in and outs of LLM Quantization with nice visuals!
27 October, 2024
-
Software 2.0 by Andrej Karpathy
Neural Networks: The next Leap for Software?
14 July,2024
-
GPT in 60 Lines of Numpy
Implement a GPT from scratch in justĀ 60 lines ofĀ numpy