https://github.com/cvjena/tensordecompositions4pinns

Code accompanying manuscript "Functional Tensor Decompositions for Physics-Informed Neural Networks"

https://github.com/cvjena/tensordecompositions4pinns

Science Score: 36.0%

This score indicates how likely this project is to be science-related based on various indicators:

  • CITATION.cff file
  • codemeta.json file
    Found codemeta.json file
  • .zenodo.json file
  • DOI references
    Found 4 DOI reference(s) in README
  • Academic publication links
    Links to: arxiv.org, acm.org
  • Academic email domains
  • Institutional organization owner
  • JOSS paper metadata
  • Scientific vocabulary similarity
    Low similarity (9.3%) to scientific vocabulary

Keywords

computer-vision pde-solver pinns
Last synced: 6 months ago · JSON representation

Repository

Code accompanying manuscript "Functional Tensor Decompositions for Physics-Informed Neural Networks"

Basic Info
  • Host: GitHub
  • Owner: cvjena
  • License: mit
  • Language: Jupyter Notebook
  • Default Branch: main
  • Homepage:
  • Size: 3.33 MB
Statistics
  • Stars: 5
  • Watchers: 1
  • Forks: 0
  • Open Issues: 0
  • Releases: 0
Topics
computer-vision pde-solver pinns
Created over 1 year ago · Last pushed about 1 year ago
Metadata Files
Readme License

README.md

TensorDecompositions4PINNs

TL;DR:

Functional Tensor Decompositions explores how tensor decomposition techniques can unlock new possibilities for variable separation in Physics-Informed Neural Networks (PINNs), overcoming the curse of dimensionality in solving high-dimensional PDEs

Abstract:

Physics-Informed Neural Networks (PINNs) have shown great promise in approximating partial differential equations (PDEs), although they remain constrained by the curse of dimensionality. In this paper, we propose a generalized PINN version of the classical variable separable method. To do this, we first show that, using the universal approxima- tion theorem, a multivariate function can be approximated by the outer product of neural networks, whose inputs are separated variables. We leverage tensor decomposition forms to separate the variables in a PINN setting. By employing Canonic Polyadic (CP), Tensor-Train (TT), and Tucker decomposition forms within the PINN framework, we create ro- bust architectures for learning multivariate functions from separate neu- ral networks connected by outer products. Our methodology significantly enhances the performance of PINNs, as evidenced by improved results on complex high-dimensional PDEs, including the 3D Helmholtz and 5D Poisson equations, among others. This research underscores the poten- tial of tensor decomposition-based variably separated PINNs to surpass the state-of-the-art, offering a compelling solution to the dimensionality challenge in PDE approximation.

Links:

For more details, refer to our paper:

Required Packages:

  • tqdm
  • jax
  • pina
  • matplotlib

Citation:

If you find this work useful, please consider citing:

```bibtex @inproceedings{vemuri2024, author = {Vemuri, Sai Karthikeya and B\"{u}chner, Tim and Niebling, Julia and Denzler, Joachim}, title = {Functional Tensor Decompositions for Physics-Informed Neural Networks}, year = {2024}, isbn = {978-3-031-78388-3}, publisher = {Springer-Verlag}, address = {Berlin, Heidelberg}, url = {https://doi.org/10.1007/978-3-031-78389-03}, doi = {10.1007/978-3-031-78389-03}, abstract = {Physics-Informed Neural Networks (PINNs) have shown continuous and increasing promise in approximating partial differential equations (PDEs), although they remain constrained by the curse of dimensionality. In this paper, we propose a generalized PINN version of the classical variable separable method. To do this, we first show that, using the universal approximation theorem, a multivariate function can be approximated by the outer product of neural networks, whose inputs are separated variables. We leverage tensor decomposition forms to separate the variables in a PINN setting. By employing Canonic Polyadic (CP), Tensor-Train (TT), and Tucker decomposition forms within the PINN framework, we create robust architectures for learning multivariate functions from separate neural networks connected by outer products. Our methodology significantly enhances the performance of PINNs, as evidenced by improved results on complex high-dimensional PDEs, including the 3d Helmholtz and 5d Poisson equations, among others. This research underscores the potential of tensor decomposition-based variably separated PINNs to surpass the state-of-the-art, offering a compelling solution to the dimensionality challenge in PDE approximation.}, booktitle = {Pattern Recognition: 27th International Conference, ICPR 2024, Kolkata, India, December 1–5, 2024, Proceedings, Part XXV}, pages = {32–46}, numpages = {15}, keywords = {Tensor Decomposition, Physics-Informed Neural Networks}, location = {Kolkata, India} }

Owner

  • Name: Computer Vision Group Jena
  • Login: cvjena
  • Kind: organization
  • Location: Jena

GitHub Events

Total
  • Release event: 1
  • Watch event: 2
  • Push event: 2
  • Create event: 2
Last Year
  • Release event: 1
  • Watch event: 2
  • Push event: 2
  • Create event: 2