-
Notifications
You must be signed in to change notification settings - Fork 96
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ACNN implementation has only 1 Conv layer #3
Comments
The attention mechanism (self attention) is written in Line 102 with raw torch.matmul implementations. |
Yes, but shouldn't there still be more than 1 If I run
|
Just 1 CNN layer with 1 attention layer, the implementation style before Attention Is All You Need paper. You can see in https://arxiv.org/abs/1703.03130 |
Here is a very beginning transformer1d repo https://github.com/hsd1503/transformer1d |
Thanks for the resources Hong, I will read up on that! What performance are you able to attain with this model? If you have time a table with all the different architectures and their performance on the Physionet data would be very useful! 🙏🏻 |
I think something is wrong with the ACNN implementation as the entire CNN exists out of a single Conv1d layer:
The text was updated successfully, but these errors were encountered: