github d2l-ai/d2l-en v0.16.0
Release v0.16.0

latest releases: v1.0.3, v1.0.0, v1.0.0-beta0...
3 years ago

Brand-New Attention Chapter

We have added the brand-new Chapter: Attention Mechanisms:

  • Attention Cues

    • Attention Cues in Biology
    • Queries, Keys, and Values
    • Visualization of Attention
  • Attention Pooling: Nadaraya-Watson Kernel Regression

    • Generating the Dataset
    • Average Pooling
    • Nonparametric Attention Pooling
    • Parametric Attention Pooling
  • Attention Scoring Functions

    • Masked Softmax Operation
    • Additive Attention
    • Scaled Dot-Product Attention
  • Bahdanau Attention

    • Model
    • Defining the Decoder with Attention
    • Training
  • Multi-Head Attention

    • Model
    • Implementation
  • Self-Attention and Positional Encoding

    • Self-Attention
    • Comparing CNNs, RNNs, and Self-Attention
    • Positional Encoding
  • Transformer

    • Model
    • Positionwise Feed-Forward Networks
    • Residual Connection and Layer Normalization
    • Encoder
    • Decoder
    • Training

PyTorch Adaptation Completed

We have completed PyTorch implementations for Vol.1 (Chapter 1--15).

Towards v1.0

The following chapters have been significantly improved for v1.0:

  • Introduction
  • Modern Recurrent Neural Networks

Chinese Translation

The following chapters have been translated into Chinese (d2l-zh v2 Git repo, Web preview):

  • Introduction
  • Preliminaries
  • Linear Neural Networks
  • Multilayer Perceptrons
  • Deep Learning Computation
  • Convolutional Neural Networks
  • Modern Convolutional Neural Networks

Turkish Translation

The community are translating the book into Turkish (d2l-tr Git repo, Web preview). The first draft of Chapter 1--7 is complete.

Don't miss a new d2l-en release

NewReleases is sending notifications on new releases.