Instruction Finetuning GPT-2 with PyTorch
Literally coding every function from scratch.
Literally coding every function from scratch.
Literally coding every function from scratch.
A dive into OpenAI's first ever open-weight model after 5 years.
A practical workflow for adapting GPT-2 to emulate a specific influencer’s tone: continue pretraining on long-form transcripts, align with supervised fine-tuning on crafted Q&A, then polish with Direct Preference Optimization (DPO) to balance authenticity and safety.
Fundamental constraints through the history of deep learning summarized in a somewhat length post.
Supervised finetuning Llama model for better at customer servicing.
Engineering considerations when pre-training, finetuning a language model.
A sandbox showcasing the basic capabilities of an AI agent.
A lot of theory about trasnformer and its inner working.