Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/boltzmannentropy/interviews.ai
It is my belief that you, the postgraduate students and job-seekers for whom the book is primarily meant will benefit from reading it; however, it is my hope that even the most experienced researchers will find it fascinating as well.
https://github.com/boltzmannentropy/interviews.ai
artificial-intelligence autograd bayesian-statistics convolutional-neural-networks data-science deep-learning ensemble-learning feature-extraction graduate-school information-theory interview-preparation jax jobs logistic-regression loss-functions machine-learning python pytorch pytorch-tutorial
Last synced: 25 days ago
JSON representation
It is my belief that you, the postgraduate students and job-seekers for whom the book is primarily meant will benefit from reading it; however, it is my hope that even the most experienced researchers will find it fascinating as well.
- Host: GitHub
- URL: https://github.com/boltzmannentropy/interviews.ai
- Owner: BoltzmannEntropy
- Created: 2021-09-30T17:04:46.000Z (about 3 years ago)
- Default Branch: main
- Last Pushed: 2022-01-21T20:02:37.000Z (almost 3 years ago)
- Last Synced: 2024-10-15T05:41:51.754Z (25 days ago)
- Topics: artificial-intelligence, autograd, bayesian-statistics, convolutional-neural-networks, data-science, deep-learning, ensemble-learning, feature-extraction, graduate-school, information-theory, interview-preparation, jax, jobs, logistic-regression, loss-functions, machine-learning, python, pytorch, pytorch-tutorial
- Homepage:
- Size: 21 MB
- Stars: 4,508
- Watchers: 53
- Forks: 295
- Open Issues: 7
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
Deep Learning Interviews book: Hundreds of fully solved job interview questions from a wide range of key topics in AI.
Download PDF •
About •
Errata •
# A PERSONAL NOTE:
"Keep learning, or risk becoming irrelevant."In this first volume, I purposely present a coherent, cumulative, and content-specific core curriculum of the data science field, including topics
such as information theory, Bayesian statistics, algorithmic differentiation, logistic regression, perceptrons, and convolutional neural networks.
I hope you will find this book stimulating.It is my belief that you the postgraduate students and job-seekers for whom the book is primarily meant will benefit from
reading it; however, it is my hope that even the most experienced researchers will find it fascinating as well.## **I would like to solicit corrections, criticisms, and suggestions from students and other readers. Although I have tried to eliminate errors over the multi year process of writing and revising this text, a few undoubtedly remain. In particular, some typographical infelicities will no doubt find their way into the final version.** **I hope you will forgive them.**
**Contact Amir:**
* https://www.linkedin.com/in/amirivry/
* https://scholar.google.com.mx/citations?user=rQCVwksAAAAJ&hl=iw
**Contact Shlomo:**
* https://www.linkedin.com/in/quantscientist/
* https://scholar.google.com.mx/citations?user=bM0LGgcAAAAJ&hl
This book is available for purchase through Amazon and other standard distribution channels. Please see the publisher's web page to order the book or to obtain further details on its publication. A manuscript of the book can be found below—it has been made available for personal use only and must not be sold.
* https://amazon.com/author/quantscientist
---
# Download### The PDF is available here:
https://arxiv.org/abs/2201.00650## Citation
```
@misc{kashani2021deep,
title={Deep Learning Interviews: Hundreds of fully solved job interview questions from a wide range of key topics in AI},
author={Shlomo Kashani and Amir Ivry},
year={2021},
eprint={2201.00650},
note = {ISBN 13: 978-1-9162435-4-5 },
url = {https://www.interviews.ai},
archivePrefix={arXiv},
primaryClass={cs.LG}
}
```**SELLING OR COMMERCIAL USE IS STRICTLY PROHIBITED**.
The user rights of this e-resource are specified in a licence agreement below.
You may only use this e-resource for the purposes *private study*.
Any selling/reselling of its content is strictly prohibited.This book (www.interviews.ai) was written for you: an aspiring data scientist with a quantitative background, facing down the gauntlet of the interview process in an increasingly competitive field. For most of you, the interview process is the most significant hurdle between you and a dream job.
Even though you have the ability, the background, and the motivation to excel in your target position, you might need some guidance on how to get your foot in the door.## About
The second edition of Deep Learning Interviews (The Amazon Softcover is printed in B&W) is home to hundreds of fully-solved problems, from a wide range of key topics in AI. It is designed to both rehearse interview or exam specific topics and provide machine learning M.Sc./Ph.D. students, and those awaiting an interview a well-organized overview of the field. The problems it poses are tough enough to cut your teeth on and to dramatically improve your skills-but they’re framed within thought-provoking questions and engaging stories.
That is what makes the volume so specifically valuable to students and job seekers: it provides them with the ability to speak confidently and quickly on any relevant topic, to answer technical questions clearly and correctly, and to fully understand the purpose and meaning of interview questions and answers. Those are powerful, indispensable advantages to have when walking into the interview room.
The book’s contents is a large inventory of numerous topics relevant to DL job interviews and graduate level exams. That places this work at the forefront of the growing trend in science to teach a core set of practical mathematical and computational skills. It is widely accepted that the training of every computer scientist must include the fundamental theorems of ML, and AI appears in the curriculum of nearly every university. This volume is designed as an excellent reference for graduates of such programs.
- The book spans almost 400 pages
- Hundreds of fully-solved problems
- Problems from numerous areas of deep learning
- Clear diagrams and illustrations
- A comprehensive index
- Step-by-step solutions to problems
- Not just the answers given, but the work shown
- Not just the work shown, but reasoning given where appropriateThis book was written for you: an aspiring data scientist with a quantitative background, facing down the gauntlet of the interview process in an increasingly competitive field. For most of you, the interview process is the most significant hurdle between you and a dream job. Even though you have the ability, the background, and the motivation to excel in your target position, you might need some guidance on how to get your foot in the door.
Your curiosity will pull you through the book’s problem sets, formulas, and instructions, and as you progress, you’ll deepen your understanding of deep learning. There are intricate connections between calculus, logistic regression, entropy, and deep learning theory; work through the book, and those connections will feel intuitive.## CORE SUBJECT AREAS (VOLUME-I):
VOLUME-I of the book focuses on statistical perspectives and blends background fundamentals with core ideas and practical knowledge. There are dedicated chapters on:
- Information Theory
- Calculus & Algorithmic Differentiation
- Bayesian Deep Learning & Probabilistic Programming
- Logistic Regression
- Ensemble Learning
- Feature Extraction
- Deep Learning: expanded chapter (100+ pages)These chapters appear alongside numerous in-depth treatments of topics in Deep Learning with code examples in PyTorch, Python and C++.
## Disclaimers
- "PyTorch" is a trademark of Facebook.## Licensing
- Copyright © [Shlomo Kashani, author of the book "Deep Learning Interviews"](www.interviews.ai)
Shlomo Kashani, Author of the book _Deep Learning Interviews_ www.interviews.ai: [email protected]
# Errata (May not be up to date)
## ***Minor corrections are not included.***Thank you to all the readers who pointed out these issues.
**Errata for the version** **03/12/2020** **printing and reflected in the online version:**1. Question number **PRB-267 -CH.PRB- 8.91** was removed due to lack of clarity
2. Question number **PRB-115 - CH.PRB- 5.16** was removed due to lack of clarity**Errata for the version** **05/12/2020** **printing and reflected in the online version:**
1. Page 230, Question number **PRB-178** amend “startified scross validation“ TO “stratified cross validation.“
2. Page 231, Question number **PRB-181** added a ” .“ after data-folds
3. Page 231, Question number **PRB-191** amend “an” to “a”
4. Page 234, Question number **PRB-192** “in” repeated twice
5. Page 236, Question number **PRB-194 amend** “approached” to “approaches“, “arr” to “arr001”
6. Page 247, Question number **PRB-210 amend** “an” to “a”
7. Page 258, Question number **PRB-227 amend** “A confusion metrics” to “A confusion matrix”
8. Page 271, Question number **PRB-240 amend** “MaxPool2D(4,4,)” to “MaxPool2D(4,4)”
9. Page 273, Question number **PRB-243 amend** “identity” to “identify”
10. Page 281, Question number **PRB-254 amend** “suggest” to “suggests”
11. Page 283, Question number **PRB-256** “happening” misspelled
12. Page 286, “L1, L2” amended to “Norms”
13. Page 288, Question number **SOL-184** **amend** “the full” to “is the full”
14. Page 298, Question number **SOL-208** **amend** “ou1” to “out”
15. Page 319, Question number **SOL-240** **amend** “torch.Size([1, 32, 222, 222]).” to “torch.size([1, 32, 222, 222]).“
16. Page 283, Question number **PRB-256** “happening” was misspelled**Errata for the version** **07/12/2020** **printing and reflected in the online version:**
1. Page 187, Question number **PRB-140** two missing plots (6.3, 6.4) which did not render correctly on the print version
![ball001.png](https://images.squarespace-cdn.com/content/v1/5c33c435f93fd4233f157b43/1607530038262-PGU5F0YDMFA9NON3NKZA/ball001.png?format=500w)6.3
![ball002.png](https://images.squarespace-cdn.com/content/v1/5c33c435f93fd4233f157b43/1607530124438-1U0OIE7QPO0DSMP8LKBS/ball002.png?format=500w)
6.4
**Errata for the version** **09/21/2020** **printing and reflected in the online version:**1. Page 34, Solution number **SOL-19** , 0.21886 should be 0.21305 and 0.21886 ± 1.95 × 0.21886 should be **0.21305** ± 1.95 × 0.21886
2. Page 36-7, Solution number **SOL-21** ,4.8792/0.0258 = **189.116** and not 57.3 and pi(33) = 0.01748 and not pi(33) = **0.211868**.
3. Page 49, **PRB-47** “What is the probability that the expert is a **monkey**“ should be “What is the probability that the expert is a **human**”**Errata for the version** **09/22/2020** **printing and reflected in the online version:**
1. Page 73, Solution number **SOL-56** should read ”The Hessian is generated by **differentiating**”
2. Page 57, Problem number **PRB-65** should read ”**two** neurons”**Errata for the version** **09/24/2020** **printing and reflected in the online version:**
1. Page 78, Solution number **SOL-64** , the OnOffLayer is off only if at least 150 out of 200 neurons are off. Therefore, this may be represented as a Binomial distribution and the probability for the layer to be off is :
![2020-12-24 21_08_52-E__Sync_branded_interviews.ai_amazon_21-12-2020_chap_bayes.tex - TeXstudio.png](https://images.squarespace-cdn.com/content/v1/5c33c435f93fd4233f157b43/1608836963860-U8ZD3L5L4IL5QKOAUZD8/2020-12-24+21_08_52-E__Sync_branded_interviews.ai_amazon_21-12-2020_chap_bayes.tex+-+TeXstudio.png?format=750w)