8000 GitHub - kyegomez/Audio-xLSTMs: Implementation of "Audio xLSTMs: Learning Self-supervised audio representations with xLSTMs" in PyTorch
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

Implementation of "Audio xLSTMs: Learning Self-supervised audio representations with xLSTMs" in PyTorch

License

Notifications You must be signed in to change notification settings

kyegomez/Audio-xLSTMs

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
< 8000 div class="react-directory-filename-cell">
 
 
 
 
 
 
 
 

Repository files navigation

Multi-Modality

Audio xLSTMs: Learning Self-supervised audio representations with xLSTMs

Join our Discord Subscribe on YouTube Connect on LinkedIn Follow on X.com

This is a community based approach to an implementation mostly for practice. I will implement the model architecture as defined in the paper but will leave someone else to implement the training script! So please create a training script if you have the time and energy

Install

$ pip3 install -U audio-xlstm

License

MIT

Todo

  • Implement the flip module
  • Correctly leverage msltm module
  • Ensure model architecture is correct
  • Implement training script on whisper like data
  • Implement speech and audio recognition datasets

Citation

@article{xlstm,
  title={xLSTM: Extended Long Short-Term Memory},
  author={Beck, Maximilian and P{\"o}ppel, Korbinian and Spanring, Markus and Auer, Andreas and Prudnikova, Oleksandra and Kopp, Michael and Klambauer, G{\"u}nter and Brandstetter, Johannes and Hochreiter, Sepp},
  journal={arXiv preprint arXiv:2405.04517},
  year={2024}
}

About

Implementation of "Audio xLSTMs: Learning Self-supervised audio representations with xLSTMs" in PyTorch

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Sponsor this project

 

Packages

No packages published
0