About 145,000 results
Open links in new tab
  1. Encoder-Decoder Long Short-Term Memory Networks

    Aug 14, 2019 · The Encoder-Decoder LSTM is a recurrent neural network designed to address sequence-to-sequence problems, sometimes called seq2seq. Sequence-to-sequence …

  2. lkulowski/LSTM_encoder_decoder - GitHub

    To make sequence-to-sequence predictions using a LSTM, we use an encoder-decoder architecture. The LSTM encoder-decoder consists of two LSTMs. The first LSTM, or the …

  3. Encoder-Decoder Seq2Seq Models, Clearly Explained!! - Medium

    Mar 11, 2021 · Encoder-Decoder models were originally built to solve such Seq2Seq problems. In this post, I will be using a many-to-many type problem of Neural Machine Translation (NMT) …

  4. Seq2Seq-Encoder-Decoder-LSTM-Model | by Pradeep Dhote

    Aug 20, 2020 · The most common architecture used to build Seq2Seq models is the Encoder Decoder architecture. Encoder reads the input sequence and summarizes the information in …

  5. Encoder/Decoder LSTM model for time series forecasting

    Oct 13, 2023 · Encoder: I pass an input tensor consisting of the predictor variable and covariates at time t-2, t-1, t which will have dimensions (N, L, H_in) where L=3 (in the diagram I have …

  6. We built a multi-encoder-decoder architecture with attention mechanism between the encoders as well as between the encoder and the decoder. We achieved ROUGE-L of 12.8 and BLEU of …

  7. Building a LSTM Encoder-Decoder model using Lux.jl

    StatefulLuxLayer{true}(RNNEncoderDecoder(encoder = RNNEncoder(cell = LSTMCell(1 => 32), # 4_480 parameters, plus 1), decoder = RNNDecoder(cell = LSTMCell(1 => 32), # 4_480 …

  8. Encoder-Decoder Models •Basic premise: •Use a neural network to encode an input to an internal representation •Pass that internal representation as input to a second neural network •Use that …

  9. Exploring Encoder-Decoder Architecture with LSTMs - Medium

    Sep 5, 2024 · Sequential models employ the encoder-decoder architecture as a preeminent framework, particularly for tasks that involve mapping one sequence to another, like machine …

  10. Encoder-decoder models in sequence-to-sequence learning: A …

    Through the analysis and summary of relevant literature, this study reveals the advantages of RNN and LSTM in sequence data processing; RNN structure is simple and effective, can …

  11. GitHub - michhar/forecasting-with-lstm: LSTM encoder-decoder

    In the Forecasting with LSTM Encoder-Decoder in Keras notebook a simple encoder-decoder network is applied to a timeseries forecasting problem: predicting natural gas prices in the US …

  12. LSTM based sequence-sequence/ encoder-decoder architecture

    Jul 29, 2024 · It is a supervised machine learning problem where we send first english words sequentially to the encoder and decoder predict the word and calculate loss function to update …

  13. We propose an LSTM-based Encoder-Decoder scheme for Anomaly Detection in multi-sensor time-series (EncDec-AD). An encoder learns a vector representation of the in-put time-series …

  14. LSTM cell based encoder-decoder. “Sequence to Sequence

    Mar 24, 2024 · Encoder-Decoder Architecture: The model uses two separate LSTMs — one for the input sequence (encoder) and another for the output sequence (decoder). The encoder …

  15. Encoder Decoder Models - GeeksforGeeks

    May 2, 2025 · In an encoder-decoder model both the encoder and decoder are separate networks each one has its own specific task. These networks can be different types such as Recurrent …

  16. What's the difference between stacked LSTM and encoder-decoder LSTM

    Mar 11, 2022 · Yes, to turn this into an encoder-decoder, you need to turn off return_sequences in the first LSTM. This will create a 2D output, and so you need to use a RepeatVector …

  17. GitHub - tm4roon/pytorch-seq2seq: An Implementation of Encoder-Decoder ...

    This stacked multiple layers of an RNN with a Long Short-Term Memory (LSTM) are used for both the encoder and the decoder. Also, the global attention mechanism and input feeding …

  18. Applied LSTM: Use Cases, Types, and Challenges - G2

    May 27, 2025 · Users can use an encoder-decoder LSTM model to encode the input sequence to a context vector and share translated outputs. Speech recognition systems use LSTM models …

  19. Understanding Encoders-Decoders with Attention Based Mechanism

    Feb 1, 2021 · In the encoder-decoder model, the input sequence would be encoded as a single fixed-length context vector. We will obtain a context vector that encapsulates the hidden and …

  20. lstm_encoder_decoder.py - GitHub

    Build a LSTM encoder-decoder using PyTorch to make sequence-to-sequence prediction for time series data - lkulowski/LSTM_encoder_decoder

  21. Encoders and Decoders in Transformer Models

    May 25, 2025 · Transformer models have revolutionized natural language processing (NLP) with their powerful architecture. While the original transformer paper introduced a full encoder …

  22. LSTM_encoder_decoder/README.md~ at master - GitHub

    We use PyTorch to build the LSTM encoder-decoder in `lstm_encoder_decoder.py`. The LSTM encoder takes an input sequence and produces an encoded state (i.e., cell state and hidden …

  23. Some results have been removed