trphysx

Transformers for modeling physical systems


Keywords
transformers, physics, surrogate, machine, learning, deep, deep-learning, machine-learning, physical-systems, pytorch, self-attention, transformer
License
MIT
Install
pip install trphysx==0.0.8

Documentation

Transformer PhysX

PyPI version CircleCI Documentation Status Website liscense

Transformer PhysX is a Python packaged modeled after the Hugging Face repository designed for the use of transformers for modeling physical systems. Transformers have seen recent success in both natural language processing and vision fields but have yet to fully permute other machine learning areas. Originally proposed in Transformers for Modeling Physical Systems, this projects goal is to make these deep learning advances including self-attention and Koopman embeddings more accessible for the scientific machine learning community.

Website | Documentation |Getting Started | Data

Associated Papers

Transformers for Modeling Physical Systems [ ArXiV ]

Colab Quick Start

Embedding Model Transformer
Lorenz Open In Colab Open In Colab
Cylinder Flow Open In Colab Open In Colab
Gray-Scott - -
Rossler Open In Colab Open In Colab

Road Map

This is an on going project, hence many parts are not fully developed and will be added in the near future. If you have any particular questions or features you are interested in, please make a issue request so it can be prioritized! Thanks for understanding.

  • Tutorials/ blog post to help with introduction
  • Info on each example system in docs
  • Additional Unit Testing for Better Code Coverage
  • Parallel Data training for transformer
  • Unsupervised pretraining physics

Additional Resources

Contact

Open an issue on the Github repository if you have any questions/concerns.

Citation

Find this useful or like this work? Cite us with:

@article{geneva2020transformers,
    title={Transformers for Modeling Physical Systems},
    author={Geneva, Nicholas and Zabaras, Nicholas},
    journal={arXiv preprint arXiv:2010.03957},
    year={2020}
}