Sitemap
Artificial Intelligence in Plain English

New AI, ML and Data Science articles every day. Follow to join our 3.5M+ monthly readers.

Follow publication

Member-only story

Working of Encoder-Decoder model

11 min readMay 25, 2023

--

LSTM and GRU models are variants of RNN. The RNN models are much easier to understand when compared to LSTM and GRU units. The beauty in understanding base concept is that, the learnings can be leveraged into LSTM and GRU units

There are many types of RNN/LSTM models and the most popular among them is the Seq2Seq model.

Types of RNN models: img src

There are two types of Seq2Seq models — the Seq2Seq model of the same length and the Seq2Seq model of different lengths.

The architecture of the model depends on the type of use case, which is to be solved. Sequence models have gained traction in the past 5 years and it’s been very active research of study, even the GPT models — on which ChatGPT was implemented using the concept of Transformers and BERT, which is based on Self-Attention model — it is based on Attention models — which uses the base of Encoder-Decoder sequence models.

The problems like Machine Translation, Question Answering, Creating Chatbots and Text Summarization could also be solved using Encoder-Decoder sequence models.

Let’s consider a problem of Machine Translation (Translating a text from one language to another) to understand the architecture of Encoder-Decoder model and how they are built. Typically Neural Networks are used…

--

--

Artificial Intelligence in Plain English
Artificial Intelligence in Plain English

Published in Artificial Intelligence in Plain English

New AI, ML and Data Science articles every day. Follow to join our 3.5M+ monthly readers.

Chamanth mvs
Chamanth mvs

Written by Chamanth mvs

Data Science and ML practitioner | I share my learnings and thoughts here

Write a response