Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

2018-02-06 Plan #3

Closed
bmcfee opened this issue Jan 30, 2018 · 7 comments
Closed

2018-02-06 Plan #3

bmcfee opened this issue Jan 30, 2018 · 7 comments

Comments

@bmcfee
Copy link
Contributor

bmcfee commented Jan 30, 2018

Attention papers? @justinsalamon

@justinsalamon
Copy link

justinsalamon commented Jan 30, 2018

Some recommendations from @pli1988:

Good overview and references plus it's short.
http://ruder.io/deep-learning-nlp-best-practices/index.html#attention

More of a survey of attention...a bit old at this point, but good
https://arxiv.org/pdf/1507.01053.pdf

Next Level Attention
https://arxiv.org/abs/1706.03762

I think I have a couple more candidates lying around somewhere, will post later today, then would be good for those familiar with this literature to help guide the choice of which papers/blogs to focus on.

@bmcfee
Copy link
Contributor Author

bmcfee commented Jan 30, 2018

I think for a first run, either a survey or an early paper would be best. We can cover the more recent stuff later on.

@mcartwright
Copy link

I vote for Describing Multimedia Content using Attention-based Encoder–Decoder Networks

@justinsalamon
Copy link

There's also the "original" attention paper:

Neural Machine Translation by Jointly Learning to Align and Translate
https://arxiv.org/abs/1409.0473

Without having read the papers, I'm inclined to agree with @mcartwright that we should start with:
"Describing Multimedia Content using Attention-based Encoder–Decoder Networks"
https://arxiv.org/pdf/1507.01053.pdf

Unless there are any contenders, shall we go with this paper?

@justinsalamon
Copy link

justinsalamon commented Jan 31, 2018

Just saw there's a bunch of attention paper suggestions on #2. Pasting them here, @bmcfee @mcartwright @pli1988 shall we stick to this issue for attention papers?

@bmcfee:
Chiu and Raffel 17: https://arxiv.org/abs/1712.05382
Mnih 14: http://papers.nips.cc/paper/5542-recurrent-models-of-visual-attention
Gregor 15: https://arxiv.org/abs/1502.04623 (aka DRAW)
Xu 15: http://proceedings.mlr.press/v37/xuc15.html (show-attend-tell)
Arandjelovic 17: https://arxiv.org/abs/1705.08168 (look-listen-learn)

@pli1988:
I really like this survey. It starts with seq2seq first which I think is important.
https://arxiv.org/pdf/1507.01053.pdf
This blog covers a bunch of different flavors of attention and is easy to read.
http://ruder.io/deep-learning-nlp-best-practices/index.html#attention
Attention is all you need.
https://arxiv.org/pdf/1706.03762.pdf

It still seems to me like the Cho's survey paper (Describing Multimedia Content using
Attention-based Encoder–Decoder Networks) would be the best starting point - any comments?

@bmcfee
Copy link
Contributor Author

bmcfee commented Feb 1, 2018

SGTM. Do you have an alternative link to the pdf? I'm getting a 403 error.

@bmcfee
Copy link
Contributor Author

bmcfee commented Feb 2, 2018

NM, looks like it was a temporary arxiv hiccup. @justinsalamon can you send the paper out to the group?

@bmcfee bmcfee closed this as completed May 31, 2018
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants