tiny-llm and Practical PyTorch Learning Prerequisites
tiny-llm Exactly what I wished for. It also contains links to two existing PyTorch related courses to machine learning from Carnegie Mellon University.
tiny-llm Exactly what I wished for. It also contains links to two existing PyTorch related courses to machine learning from Carnegie Mellon University.
Arc — My new browser of choice. I love the fact that bookmarks are organized on the side panel, rather than clustered at the top of the window.
Eleventy appears to be a pretty simple static website generator that is worth exploring. A competitor to Hugo.
How I use LLMs by Andrej Karpathy — A must watch.
Panasonic S1R II — With the Sigma 28-105 f/2.8, this would be my dream camera. It is just slightly heavier than my Panasonic S5 IIx (1.57 lb vs. 1.45 lb body only).
The Ultra-Scale Playbook: Training LLMs on GPU Clusters — Amazing, and finally we have a 100-page open-source online book on how models are trained with multiple GPUs.
Crafted — What a great looking set of open-source, hand-crafted UI templates based on shadcn/ui!
Better Auth — A new authentication library that is feature-complete and easy-to-use. Compared to Lucia, which advocates a copy-and-paste approach.
Andriy Burkov’s minimalist implementation of GRPO from scratch — Rather than using a library such as Hugging Face’s TRL.
Transformer Lab — a free, open-source LLM workspace that prepares a custom dataset and fine-tunes a model using MLX on the Mac.
Lucia — Lucia, the authentication library, has adopted the design of cutting and pasting code, just like shadcn/ui, rather than implementing a library.
Unsloth.ai’s GRPO — it seems that the Unsloth implementation of GRPO uses less GPU memory, and it supports QLoRA and LoRA.
DOGE: Make AI Conferences Great Again — Zeyuan (Allen) Zhu wrote a very interesting piece on using LLMs as arbitrators in the reviewer-author discussions.
Deep Dive into LLMs like ChatGPT — Andrej Karpathy continues his top-notch hours-long education on large language models with a new episode today.
GRPO will soon be added to Apple MLX — The PR now works, using about 32 GB of memory when training Qwen2.5-0.5B.
Another simple DeepSeek R1 reproduction — This reproduction of GRPO has one distinct feature: it is exceedingly simple and quite elegant.
Fourth attempt on reproducing DeepSeek R1’s GRPO on small models — The third fourth time is the charm. I can successfully run this repo, without activating vLLM.
Lambda Labs hosts DeepSeek R1 — the dashboard is simple, nice to look at, free to use, and pretty fast when generating tokens. Overall, an excellent user experience.
How to fine-tune open LLMs in 2025 with Hugging Face — Philipp Schmid a Technical Lead at Hugging Face, posted this article on fine-tuning LLMs using Hugging Face.
On DeepSeek and Export Controls — Dario Amodei, Anthropic’s CEO, wrote a fairly long editorial on DeepSeek.
The Illustrated DeepSeek-R1 — Jay Alammar, the author of O’Reilly’s Hands-On Large Language Models, wrote a short piece on explaining DeepSeek R1 at a high level.
Qwen 2.5 7B 1M — I have just tried Qwen’s latest local model, the 7B 1M, locally in LM Studio 0.3.8 (Build 4). I loaded an entire PhD thesis into the model, and LM Studio gleefully chose inject-full-content as its content injection strategy.
Although it’s quite long, The Short Case for Nvidia Stock is a fascinating read. Also, agents are not happening yet.
Open-R1 — Hugging Face started to reproduce DeepSeek R1 in the open, and discussed the R1 technical report in a recorded YouTube video.
This website is a space for storing — and sharing, if anyone cares about these — some of the websites, code repositories, and tweets that I have read.