Constructing Neural Network-Based Models for Simulating Dynamical Systems

Overview

Constructing Neural Network-Based Models for Simulating Dynamical Systems

Note this repo is work in progress prior to reviewing

This is a companion repo for the review paper Constructing Neural Network-Based Models for Simulating Dynamical Systems. The goal is to provide PyTorch implementations that can be used as a starting point for implementation for other applications.

If you use the work please cite it using:

{
    TODO add bibtex key
}

Installing dependencies

python3 -m pip install -r requirements.txt

Where are the models located?

The table below contains the commands necessary to train and evaluate the models described in the review paper. Each experiment can be run using default parameters by executing the script in the python interpreter as follows:

python3 experiments/
   
    .py ...

   
Name Section Command
Vanilla Direct-Solution 3.2 python3 experiments/direct_solution.py --model vanilla
Automatic Differentiation in Direct-Solution 3.3 python3 experiments/direct_solution.py --model autodiff
Physics Informed Neural Networks 3.4 python3 experiments/direct_solution.py --model pinn
Hidden Physics Networks 3.5 python3 experiments/direct_solution.py --model hnn
Direct Time-Stepper 4.2.1 python3 experiments/time_stepper.py --solver direct
Residual Time-Stepper 4.2.2 python3 experiments/time_stepper.py --solver resnet
Euler Time-Stepper 4.2.3 python3 experiments/time_stepper.py --solver euler
Neural ODEs Time-Stepper 4.2.4 python3 experiments/time_stepper.py --solver {rk4,dopri5,tsit5}
Neural State-Space Model 4.3.1 ...
Neural ODEs with input 4.3.2-3 ...
Lagrangian Time-Stepper 4.4.1 ...
Hamiltonian Time-Stepper 4.4.1 ...
Deep Potential Time-Stepper 4.4.2 ...
Deep Markov-Model 4.5.1 ...
Latent Neural ODEs 4.5.2 python3 experiments/latent_neural_odes.py
Bayesian Neural ODEs 4.5.3 ...
Neural SDEs 4.5.4 ...

Docker Image

In an effort to ensure that the code can be executed in the future, we provide a docker image. The Docker image allows the code to be run in a Linux based virtual machine on any platform supported by Docker.

To use the docker image, invoke the build command in the root of this repository:

docker build . -t python_dynamical_systems

Following this "containers" containing the code and all dependencies can be instantiated via the "run" command:

docker run -ti python_dynamical_systems bash

The command will establish an interactive connection to the container. Following this you can execute the code as if it was running on your host machine:

python3 experiments/time_stepper.py ...
You might also like...
This is a model made out of Neural Network specifically a Convolutional Neural Network model
This is a model made out of Neural Network specifically a Convolutional Neural Network model

This is a model made out of Neural Network specifically a Convolutional Neural Network model. This was done with a pre-built dataset from the tensorflow and keras packages. There are other alternative libraries that can be used for this purpose, one of which is the PyTorch library.

Pytorch modules for paralel models with same architecture. Ideal for multi agent-based systems
Pytorch modules for paralel models with same architecture. Ideal for multi agent-based systems

WideLinears Pytorch parallel Neural Networks A package of pytorch modules for fast paralellization of separate deep neural networks. Ideal for agent-b

This repository contains notebook implementations of the following Neural Process variants: Conditional Neural Processes (CNPs), Neural Processes (NPs), Attentive Neural Processes (ANPs).

The Neural Process Family This repository contains notebook implementations of the following Neural Process variants: Conditional Neural Processes (CN

Pytorch implementation of
Pytorch implementation of "Attention-Based Recurrent Neural Network Models for Joint Intent Detection and Slot Filling"

RNN-for-Joint-NLU Pytorch implementation of "Attention-Based Recurrent Neural Network Models for Joint Intent Detection and Slot Filling"

A lightweight Python-based 3D network multi-agent simulator. Uses a cell-based congestion model. Calculates risk, loudness and battery capacities of the agents. Suitable for 3D network optimization tasks.
A lightweight Python-based 3D network multi-agent simulator. Uses a cell-based congestion model. Calculates risk, loudness and battery capacities of the agents. Suitable for 3D network optimization tasks.

AMAZ3DSim AMAZ3DSim is a lightweight python-based 3D network multi-agent simulator. It uses a cell-based congestion model. It calculates risk, battery

Research on controller area network Intrusion Detection Systems

Group members information Member 1: Lixue Liang Member 2: Yuet Lee Chan Member 3: Xinruo Zhang Member 4: Yifei Han User Manual Generate Attack Packets

NVIDIA Merlin is an open source library providing end-to-end GPU-accelerated recommender systems, from feature engineering and preprocessing to training deep learning models and running inference in production.

NVIDIA Merlin NVIDIA Merlin is an open source library designed to accelerate recommender systems on NVIDIA’s GPUs. It enables data scientists, machine

Advances in Neural Information Processing Systems (NeurIPS), 2020.

What is being transferred in transfer learning? This repo contains the code for the following paper: Behnam Neyshabur*, Hanie Sedghi*, Chiyuan Zhang*.

A
A "gym" style toolkit for building lightweight Neural Architecture Search systems

A "gym" style toolkit for building lightweight Neural Architecture Search systems

Owner
Christian Møldrup Legaard
Christian Møldrup Legaard
constructing maps of intellectual influence from publication data

Influencemap Project @ ANU Influence in the academic communities has been an area of interest for researchers. This can be seen in the popularity of a

CS Metrics 13 Jun 18, 2022
Simulating Sycamore quantum circuits classically using tensor network algorithm.

Simulating the Sycamore quantum supremacy circuit This repo contains data we have obtained in simulating the Sycamore quantum supremacy circuits with

Feng Pan 46 Nov 17, 2022
Semi-supervised Video Deraining with Dynamical Rain Generator (CVPR, 2021, Pytorch)

S2VD Semi-supervised Video Deraining with Dynamical Rain Generator (CVPR, 2021) Requirements and Dependencies Ubuntu 16.04, cuda 10.0 Python 3.6.10, P

Zongsheng Yue 53 Nov 23, 2022
Dynamical Wasserstein Barycenters for Time Series Modeling

Dynamical Wasserstein Barycenters for Time Series Modeling This is the code related for the Dynamical Wasserstein Barycenter model published in Neurip

null 8 Sep 9, 2022
Dynamical movement primitives (DMPs), probabilistic movement primitives (ProMPs), spatially coupled bimanual DMPs.

Movement Primitives Movement primitives are a common group of policy representations in robotics. There are many different types and variations. This

DFKI Robotics Innovation Center 63 Jan 6, 2023
Code accompanying "Learning What To Do by Simulating the Past", ICLR 2021.

Learning What To Do by Simulating the Past This repository contains code that implements the Deep Reward Learning by Simulating the Past (Deep RSLP) a

Center for Human-Compatible AI 24 Aug 7, 2021
A python package simulating the quasi-2D pseudospin-1/2 Gross-Pitaevskii equation with NVIDIA GPU acceleration.

A python package simulating the quasi-2D pseudospin-1/2 Gross-Pitaevskii equation with NVIDIA GPU acceleration. Introduction spinor-gpe is high-level,

null 2 Sep 20, 2022
Simulating an AI playing 2048 using the Expectimax algorithm

2048-expectimax Simulating an AI playing 2048 using the Expectimax algorithm The base game engine uses code from here. The AI player is modeled as a m

Subha Ramesh 2 Jan 31, 2022
Code for Private Recommender Systems: How Can Users Build Their Own Fair Recommender Systems without Log Data? (SDM 2022)

Private Recommender Systems: How Can Users Build Their Own Fair Recommender Systems without Log Data? (SDM 2022) We consider how a user of a web servi

joisino 20 Aug 21, 2022
XtremeDistil framework for distilling/compressing massive multilingual neural network models to tiny and efficient models for AI at scale

XtremeDistilTransformers for Distilling Massive Multilingual Neural Networks ACL 2020 Microsoft Research [Paper] [Video] Releasing [XtremeDistilTransf

Microsoft 125 Jan 4, 2023