top of page
Title:
Category:
Dive into Deep Learning
Book
URL
Authors:
Amazon Aston Zhang, Zack C. Lipton, Mu Li, Alex J. Smola
Published
1 August 2023
Review:
Dennis Kuriakose
Review Date :
15 August 2024
Summary
This book makes deep learning approachable, teaching the concepts, the context, and the code. The key ideas are distilled, minimizing the onboarding time needed to update the new practitioners. Implemented with PyTorch, NumPy/MXNet, JAX, and TensorFlow. Adopted at 500 universities from 70 countries
Review & Notes:
Excellent build up of educational content and papers for building a solid understanding of Transformer architecture and its later developments by Jay Alammar
Start with - A pictorial tutorial for the "Attention is all you need" paper where sensational "transformer architecture" in deep learning is first described (Jay Alammar)
And then proceed to
Read the Attention Is All You Need paper, the Transformer blog post (Transformer: A Novel Neural Network Architecture for Language Understanding), and the Tensor2Tensor announcement.
Watch Łukasz Kaiser’s talk walking through the model and its details
Play with the Jupyter Notebook provided as part of the Tensor2Tensor repo
Explore the Tensor2Tensor repo.
Follow-up works:
Technology Posts
bottom of page











