Aman Sanger
I'm Aman. I write code and am obsessed with AI.
I'm building cursor, a new way of programming using AI.
I've worked at Google, Bridgewater, you.com, and graduated from MIT.
While there, I did NLP & comp bio research. I also ran a small AI consultancy.
Thoughts
Shorter thoughts
- 4-bit weight-quantization is more expensive than 16-bit
- flash attention isn't helpful when generating tokens
- llama-1 needs multi-query attention
- instruction finetuning is underrated