Files
pbdl-book/diffphys-discuss.md

40 lines
1.9 KiB
Markdown

Discussion
=======================
The training via differentiable physics as described so far allows us
to integrate full numerical simulations and the training of deep neural networks
interacting with these simulations. While we've only hinted at what could be
achieved via DP approaches it is nonetheless a good time to summarize the pros and cons.
## Alternatives - Noise
It is worth mentioning here that other works have proposed perturbing the inputs and
the iterations at training time with noise {cite}`sanchez2020learning` (somewhat similar to
regularizers like dropout).
This can help to prevent overfitting to the training states, and hence shares similarities
with the goals of training with DP.
However, the noise is typically undirected, and hence not as accurate as training with
the actual evolutions of simulations. Hence, this noise can be a good starting point
for training that tends to overfit, but if possible, it is preferable to incorporate the
actual solver in the training loop via a DP approach.
## Summary
To summarize the pros and cons of training NNs via differentiable physics:
✅ Pro:
- uses physical model and numerical methods for discretization
- efficiency of selected methods carries over to training
- tight coupling of physical models and NNs possible
❌ Con:
- not compatible with all simulators (need to provide gradients)
- require more heavy machinery (in terms of framework support) than previously discussed methods
Especially the last point is one that is bound to strongly improve in a fairly short time, but for now it's important to keep in mind that not every simulator is suitable for DP training out of the box. Hence, in this book we'll focus on examples using phiflow, which was designed for interfacing with deep learning frameworks.
Next we can target more some complex scenarios to showcase what can be achieved with differentiable physics.