started figures
This commit is contained in:
parent
eff20b21c5
commit
02290dd8b8
4
_toc.yml
4
_toc.yml
@ -24,6 +24,10 @@
|
|||||||
- file: diffphys-code-sol.ipynb
|
- file: diffphys-code-sol.ipynb
|
||||||
- file: diffphys-dpvspinn.md
|
- file: diffphys-dpvspinn.md
|
||||||
- file: diffphys-outlook.md
|
- file: diffphys-outlook.md
|
||||||
|
- file: physgrad
|
||||||
|
sections:
|
||||||
|
- file: physgrad-comparison.ipynb
|
||||||
|
- file: physgrad-discuss.md
|
||||||
- file: old-phiflow1.md
|
- file: old-phiflow1.md
|
||||||
sections:
|
sections:
|
||||||
- file: overview-burgers-forw-v1.ipynb
|
- file: overview-burgers-forw-v1.ipynb
|
||||||
|
@ -95,8 +95,6 @@ this would cause huge memory overheads and unnecessarily slow down training.
|
|||||||
Instead, for backpropagation, we can provide faster operations that compute products
|
Instead, for backpropagation, we can provide faster operations that compute products
|
||||||
with the Jacobian transpose because we always have a scalar loss function at the end of the chain.
|
with the Jacobian transpose because we always have a scalar loss function at the end of the chain.
|
||||||
|
|
||||||
**[TODO check transpose of Jacobians in equations]**
|
|
||||||
|
|
||||||
Given the formulation above, we need to resolve the derivatives
|
Given the formulation above, we need to resolve the derivatives
|
||||||
of the chain of function compositions of the $\mathcal P_i$ at some current state $\mathbf{u}^n$ via the chain rule.
|
of the chain of function compositions of the $\mathcal P_i$ at some current state $\mathbf{u}^n$ via the chain rule.
|
||||||
E.g., for two of them
|
E.g., for two of them
|
||||||
|
1
intro.md
1
intro.md
@ -86,6 +86,7 @@ See also... Test link: {doc}`supervised`
|
|||||||
|
|
||||||
- general motivation: repeated solves in classical solvers -> potential for ML
|
- general motivation: repeated solves in classical solvers -> potential for ML
|
||||||
- PINNs: often need weighting of added loss terms for different parts
|
- PINNs: often need weighting of added loss terms for different parts
|
||||||
|
- DP intro, check transpose of Jacobians in equations
|
||||||
|
|
||||||
|
|
||||||
## TODOs , Planned content
|
## TODOs , Planned content
|
||||||
|
20
overview.md
20
overview.md
@ -36,7 +36,7 @@ natural language processing {cite}`radford2019language`,
|
|||||||
and more recently also for protein folding {cite}`alquraishi2019alphafold`.
|
and more recently also for protein folding {cite}`alquraishi2019alphafold`.
|
||||||
The field is very vibrant, and quickly developing, with the promise of vast possibilities.
|
The field is very vibrant, and quickly developing, with the promise of vast possibilities.
|
||||||
|
|
||||||
At the same time, the successes of deep learning (DL) approaches
|
On the other hand, the successes of deep learning (DL) approaches
|
||||||
has given rise to concerns that this technology has
|
has given rise to concerns that this technology has
|
||||||
the potential to replace the traditional, simulation-driven approach to
|
the potential to replace the traditional, simulation-driven approach to
|
||||||
science. Instead of relying on models that are carefully crafted
|
science. Instead of relying on models that are carefully crafted
|
||||||
@ -49,8 +49,8 @@ and _deep learning_.
|
|||||||
One central reason for the importance of this combination is
|
One central reason for the importance of this combination is
|
||||||
that DL approaches are simply not powerful enough by themselves.
|
that DL approaches are simply not powerful enough by themselves.
|
||||||
Given the current state of the art, the clear breakthroughs of DL
|
Given the current state of the art, the clear breakthroughs of DL
|
||||||
in physical applications are outstanding, the proposed techniques are novel,
|
in physical applications are outstanding.
|
||||||
sometimes difficult to apply, and
|
The proposed techniques are novel, sometimes difficult to apply, and
|
||||||
significant practical difficulties combing physics and DL persist.
|
significant practical difficulties combing physics and DL persist.
|
||||||
Also, many fundamental theoretical questions remain unaddressed, most importantly
|
Also, many fundamental theoretical questions remain unaddressed, most importantly
|
||||||
regarding data efficienty and generalization.
|
regarding data efficienty and generalization.
|
||||||
@ -78,6 +78,16 @@ at our disposal, and use them wherever we can.
|
|||||||
I.e., our goal is to _reconcile_ the data-centered
|
I.e., our goal is to _reconcile_ the data-centered
|
||||||
viewpoint and the physical simuation viewpoint.
|
viewpoint and the physical simuation viewpoint.
|
||||||
|
|
||||||
|
The resulting methods have a huge potential to improve
|
||||||
|
what can be done with numerical methods: e.g., in scenarios
|
||||||
|
where solves target cases from a certain well-defined problem
|
||||||
|
domain repeatedly, it can make a lot of sense to once invest
|
||||||
|
significant resources to train
|
||||||
|
an neural network that supports the repeated solves. Based on the
|
||||||
|
domain-specific specialization of this network, such a hybrid
|
||||||
|
could vastly outperform traditional, generic solvers. And despite
|
||||||
|
the many open questions, first publications have demonstrated
|
||||||
|
that this goal is not overly far away.
|
||||||
|
|
||||||
|
|
||||||
## Categorization
|
## Categorization
|
||||||
@ -134,10 +144,10 @@ each of the different techniques is particularly useful.
|
|||||||
|
|
||||||
To be a bit more specific, _physics_ is a huge field, and we can't cover everything...
|
To be a bit more specific, _physics_ is a huge field, and we can't cover everything...
|
||||||
|
|
||||||
```{note} The focus of this book is on...
|
```{note} The focus of this book lies on...
|
||||||
- _Field-based simulations_ (no Lagrangian methods)
|
- _Field-based simulations_ (no Lagrangian methods)
|
||||||
- Combinations with _deep learning_ (plenty of other interesting ML techniques, but not here)
|
- Combinations with _deep learning_ (plenty of other interesting ML techniques, but not here)
|
||||||
- Experiments as _outlook_ (replace synthetic data with real)
|
- Experiments as _outlook_ (replace synthetic data with real-world observations)
|
||||||
```
|
```
|
||||||
|
|
||||||
It's also worth noting that we're starting to build the methods from some very
|
It's also worth noting that we're starting to build the methods from some very
|
||||||
|
BIN
resources/pbdl-figures.key
Executable file
BIN
resources/pbdl-figures.key
Executable file
Binary file not shown.
@ -13,20 +13,20 @@ model equations exist.
|
|||||||
## Problem Setting
|
## Problem Setting
|
||||||
|
|
||||||
For supervised training, we're faced with an
|
For supervised training, we're faced with an
|
||||||
unknown function $f^*(x)=y$, collect lots of pairs of data $[x_0,y_0], ...[x_n,y_n]$ (the training data set)
|
unknown function $f^*(x)=y^*$, collect lots of pairs of data $[x_0,y^*_0], ...[x_n,y^*_n]$ (the training data set)
|
||||||
and directly train a NN to represent an approximation of $f^*$ denoted as $f$, such
|
and directly train a NN to represent an approximation of $f^*$ denoted as $f$, such
|
||||||
that $f(x)=y$.
|
that $f(x)=y \approx y^*$.
|
||||||
|
|
||||||
The $f$ we can obtain is typically not exact,
|
The $f$ we can obtain is typically not exact,
|
||||||
but instead we obtain it via a minimization problem:
|
but instead we obtain it via a minimization problem:
|
||||||
by adjusting weights $\theta$ of our representation with $f$ such that
|
by adjusting weights $\theta$ of our representation with $f$ such that
|
||||||
|
|
||||||
$\text{arg min}_{\theta} \sum_i (f(x_i ; \theta)-y_i)^2$.
|
$\text{arg min}_{\theta} \sum_i (f(x_i ; \theta)-y^*_i)^2$.
|
||||||
|
|
||||||
This will give us $\theta$ such that $f(x;\theta) \approx y$ as accurately as possible given
|
This will give us $\theta$ such that $f(x;\theta) \approx y$ as accurately as possible given
|
||||||
our choice of $f$ and the hyperparameters for training. Note that above we've assumed
|
our choice of $f$ and the hyperparameters for training. Note that above we've assumed
|
||||||
the simplest case of an $L^2$ loss. A more general version would use an error metric $e(x,y)$
|
the simplest case of an $L^2$ loss. A more general version would use an error metric $e(x,y)$
|
||||||
to be minimized via $\text{arg min}_{\theta} \sum_i e( f(x_i ; \theta) , y_i) )$. The choice
|
to be minimized via $\text{arg min}_{\theta} \sum_i e( f(x_i ; \theta) , y^*_i) )$. The choice
|
||||||
of a suitable metric is topic we will get back to later on.
|
of a suitable metric is topic we will get back to later on.
|
||||||
|
|
||||||
Irrespective of our choice of metric, this formulation
|
Irrespective of our choice of metric, this formulation
|
||||||
@ -47,12 +47,13 @@ which need to be kept small enough for a chosen application. As these topics are
|
|||||||
for classical simulations, the existing knowledge can likewise be leveraged to
|
for classical simulations, the existing knowledge can likewise be leveraged to
|
||||||
set up DL training tasks.
|
set up DL training tasks.
|
||||||
|
|
||||||
```{figure} resources/placeholder.png
|
```{figure} resources/supervised-training.jpg
|
||||||
---
|
---
|
||||||
height: 220px
|
height: 220px
|
||||||
name: supervised-training
|
name: supervised-training
|
||||||
---
|
---
|
||||||
TODO, visual overview of supervised training
|
A visual overview of supervised training. Quite simple overall, but it's good to keep this
|
||||||
|
in mind in comparison to the more complex variants we'll encounter later on.
|
||||||
```
|
```
|
||||||
|
|
||||||
## Show me some code!
|
## Show me some code!
|
||||||
|
Loading…
Reference in New Issue
Block a user