Skip to content
#

attention-model

Here are 137 public repositories matching this topic...

A recurrent attention module consisting of an LSTM cell which can query its own past cell states by the means of windowed multi-head attention. The formulas are derived from the BN-LSTM and the Transformer Network. The LARNN cell with attention can be easily used inside a loop on the cell state, just like any other RNN. (LARNN)

  • Updated Aug 20, 2018
  • Jupyter Notebook

Seq2SeqSharp is a tensor based fast & flexible encoder-decoder deep neural network framework written by .NET (C#). It has many highlighted features, such as automatic differentiation, many different types of encoders/decoders(Transformer, LSTM, BiLSTM and so on), multi-GPUs supported and so on.

  • Updated Sep 10, 2020
  • C#

Improve this page

Add a description, image, and links to the attention-model topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the attention-model topic, visit your repo's landing page and select "manage topics."

Learn more

You can’t perform that action at this time.