Skip to content
#

attention

Here are 531 public repositories matching this topic...

jingtianyilong
jingtianyilong commented Oct 22, 2020

Great work by @argusswift
I forked this repo and did some modifications to suit my case better. (You can also check my repo :-) )But the training part remains mostly the same. And I came across problem when playing with subdivision.
I assume the accumulate in the training works the same se subdivision as is in the original darknet-yolov4.
Basically you use this trick to enlarge the effec

pytorch-original-transformer

My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.

  • Updated Nov 21, 2020
  • Jupyter Notebook

Improve this page

Add a description, image, and links to the attention topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the attention topic, visit your repo's landing page and select "manage topics."

Learn more

You can’t perform that action at this time.