Skip to content

Latest commit

 

History

History
60 lines (43 loc) · 3.27 KB

README.md

File metadata and controls

60 lines (43 loc) · 3.27 KB

Chatbots

References

Recurrent Neural Networks

In order to understand the concepts of Seq2Seq models or Transformer models, the first thing to know is about RNNs,

Sequence to Sequence Model

A Sequence to Sequence Model aims to map a fixed-length input with a fixed-length output where the length of the input and output may differ.

Transformer Model (Seq2Seq with Attention)

The Transformer in NLP is a novel architecture that aims to solve sequence-to-sequence tasks while handling long-range dependencies with ease.

“The Transformer is the first transduction model relying entirely on self-attention to compute representations of its input and output without using sequence-aligned RNNs or convolution.”

Further Resources