This episode features Lewis Tunstall, machine learning engineer at Hugging Face and author of the best selling book Natural Language Processing with Transformers. He currently focuses on one of the hottest topic in NLP right now reinforcement learning from human feedback (RLHF). Lewis holds a PhD in quantum physics and his research has taken him around the world and into some of the most impactful projects including the Large Hadron Collider, the world's largest and most powerful particle accelerator. Lewis shares his unique story from Quantum Physicist to Data Scientist to Machine Learning Engineer.
Resources to learn more about Lewis Tunstall
References from the Episode
- https://www.fast.ai/
- https://jeremy.fast.ai/
- SetFit - https://arxiv.org/abs/2209.11055
- Proximal Policy Optimization
- InstructGPT
- RAFT Benchmark
- Bidirectional Language Models are Also Few-Shot Learners
- Nils Reimers - Sentence Transformers
- Jay Alammar - Illustrated Transformer
- Annotated Transformer
- Moshe Wasserblat, Intel, NLP, Research Manager
- Leandro von Werra, Co-Author of NLP with Transformers, Hugging Face Researcher
- LLMSys - https://lmsys.org/
- LoRA - Low-Rank Adaptation of Large Language Models
Resources to learn more about Learning from Machine Learning