Initial.
This commit is contained in:
28
README.md
28
README.md
@@ -33,24 +33,24 @@ dream job somewhere in a startup or a branch of the big five. You have the abili
|
||||
## What can it do for me?
|
||||
|
||||
The book’s contents is a large inventory of numerous topics relevant to DL job interviews and graduate-level exams. That places this work at the forefront of the growing trend in science to teach a core set of practical mathematical and computational skills. It is widely accepted that the training of every computer scientist must include the fundamental theorems of ML, and AI appears in the curriculum of nearly every university. This volume is designed as an excellent reference for graduates of such programs:
|
||||
● Hundreds of fully-solved problems
|
||||
● Problems from numerous areas of deep learning
|
||||
● Clear diagrams and illustrations
|
||||
● A comprehensive index
|
||||
● Step-by-step solutions to problems
|
||||
● Not just the answers given, but the work shown
|
||||
● Not just the work shown, but reasoning given where appropriate
|
||||
- Hundreds of fully-solved problems
|
||||
- Problems from numerous areas of deep learning
|
||||
- Clear diagrams and illustrations
|
||||
- A comprehensive index
|
||||
- Step-by-step solutions to problems
|
||||
- Not just the answers given, but the work shown
|
||||
- Not just the work shown, but reasoning given where appropriate
|
||||
|
||||
## Cor subject areas
|
||||
|
||||
Your curiosity will pull you through the book’s problem sets, formulas, and instructions, and as you progress, you’ll deepen your understanding of deep learning. The connections between calculus, logistic regression, entropy, and deep learning theory are intricate: work through the book, and those connections will feel intuitive. VOLUME-I of the book focuses on statistical perspectives and blends background fundamentals with core ideas and practical knowledge. There are dedicated chapters on:
|
||||
● Information Theory
|
||||
● Calculus & Algorithmic Differentiation
|
||||
● Bayesian Deep Learning & Probabilistic Programming
|
||||
● Logistic Regression
|
||||
● Ensemble Learning
|
||||
● Feature Extraction
|
||||
● Deep Learning: Expanded Chapter (100+ pages)
|
||||
- Information Theory
|
||||
- Calculus & Algorithmic Differentiation
|
||||
- Bayesian Deep Learning & Probabilistic Programming
|
||||
- Logistic Regression
|
||||
- Ensemble Learning
|
||||
- Feature Extraction
|
||||
- Deep Learning: Expanded Chapter (100+ pages)
|
||||
These chapters appear alongside numerous in-depth treatments of topics in Deep Learning with code examples in PyTorch, Python and C++.
|
||||
## Citation
|
||||
```
|
||||
|
||||
Reference in New Issue
Block a user