Upgrade to Pro
— share decks privately, control downloads, hide ads and more …
Speaker Deck
Features
Speaker Deck
PRO
Sign in
Sign up for free
Search
Search
Deep Learning
Search
Sponsored
·
Your Podcast. Everywhere. Effortlessly.
Share. Educate. Inspire. Entertain. You do you. We'll handle the rest.
→
Abhinav Tushar
September 10, 2015
Research
290
6
Share
Deep Learning
Introductory talk on deep learning
Abhinav Tushar
September 10, 2015
More Decks by Abhinav Tushar
See All by Abhinav Tushar
the garden of eden
lepisma
0
110
Technology
lepisma
0
96
Bio-Inspired Computing
lepisma
0
110
Maestro
lepisma
0
130
War and Economics
lepisma
0
140
Other Decks in Research
See All in Research
通時的な類似度行列に基づく単語の意味変化の分析
rudorudo11
0
250
IEEE AIxVR 2026 Keynote Talk: "Beyond Visibility: Understanding Scenes and Humans under Challenging Conditions with Diverse Sensing"
miso2024
0
150
明日から使える!研究効率化ツール入門
matsui_528
11
6.1k
討議:RACDA設立30周年記念都市交通フォーラム2026
trafficbrain
0
780
ForestCast: Forecasting Deforestation Risk at Scale with Deep Learning
satai
3
770
svc-hook: hooking system calls on ARM64 by binary rewriting
retrage
2
210
20年前に50代だった人たちの今
hysmrk
0
180
2026年1月の生成AI領域の重要リリース&トピック解説
kajikent
0
950
SREはサイバネティクスの夢をみるか? / Do SREs Dream of Cybernetics?
yuukit
3
470
オーストリア流 都市の公共交通サービス水準評価@公共交通オープンデータ最前線2026
trafficbrain
0
130
存立危機事態の再検討
jimboken
0
270
東京大学工学部計数工学科、計数工学特別講義の説明資料
kikuzo
0
210
Featured
See All Featured
It's Worth the Effort
3n
188
29k
Six Lessons from altMBA
skipperchong
29
4.2k
Building Applications with DynamoDB
mza
96
7k
StorybookのUI Testing Handbookを読んだ
zakiyama
31
6.7k
Money Talks: Using Revenue to Get Sh*t Done
nikkihalliwell
0
200
Avoiding the “Bad Training, Faster” Trap in the Age of AI
tmiket
0
120
Groundhog Day: Seeking Process in Gaming for Health
codingconduct
0
140
The B2B funnel & how to create a winning content strategy
katarinadahlin
PRO
1
330
JavaScript: Past, Present, and Future - NDC Porto 2020
reverentgeek
52
5.9k
End of SEO as We Know It (SMX Advanced Version)
ipullrank
3
4.1k
The Organizational Zoo: Understanding Human Behavior Agility Through Metaphoric Constructive Conversations (based on the works of Arthur Shelley, Ph.D)
kimpetersen
PRO
0
310
Jamie Indigo - Trashchat’s Guide to Black Boxes: Technical SEO Tactics for LLMs
techseoconnect
PRO
0
110
Transcript
D E E P L E A R N I
N G
models AE / SAE RBM / DBN CNN RNN /
LSTM Memnet / NTM agenda questions What ? Why ? How ? Next ?
what why how next What ? AI technique for learning
multiple levels of abstractions directly from raw information
what why how next Primitive rule based AI Tailored systems
Hand Crafted Program Output Input
what why how next Classical machine learning Learning from custom
features Hand Crafted Features Learning System Output Input
what why how next Deep Learning based AI Learn everything
Learned Features (Lower Level) Learned Features (Higher Level) Learning System Output Input
None
https://www.youtube.com/watch?v=Q70ulPJW3Gk PPTX PDF (link to video below)
With the capacity to represent the world in signs and
symbols, comes the capacity to change it Elizabeth Kolbert (The Sixth Extinction) “
Why The buzz ?
what why how next Google Trends Deep Learning
what why how next
Crude timeline of Neural Networks 1950 1980 1990 2000 Perceptron
Backprop & Application NN Winter
2010 Stacking RBMs Deep Learning fuss
HUGE DATA Large Synoptic Survey Telescope (2022) 30 TB/night
HUGE CAPABILITIES GPGPU ~20x speedup Powerful Clusters
HUGE SUCCESS Speech, text understanding Robotics / Computer Vision Business
/ Big Data Artificial General Intelligence (AGI)
How its done ?
what why how next Shallow Network ℎ ℎ = (,
0) = ′(ℎ, 1) = (, ) minimize
what why how next Deep Network
what why how next Deep Network More abstract features Stellar
performance Vanishing Gradient Overfitting
what why how next Autoencoder ℎ Unsupervised Feature Learning
what why how next Stacked Autoencoder Y. Bengio et. all;
Greedy Layer-Wise Training of Deep Networks
what why how next Stacked Autoencoder 1. Unsupervised, layer by
layer pretraining 2. Supervised fine tuning
what why how next Deep Belief Network 2006 breakthrough Stacking
Restricted Boltzmann Machines (RBMs) Hinton, G. E., Osindero, S. and Teh, Y.; A fast learning algorithm for deep belief nets
Rethinking Computer Vision
what why how next Traditional Image Classification pipeline Feature Extraction
(SIFT, SURF etc.) Classifier (SVM, NN etc.)
what why how next Convolutional Neural Network Images taken from
deeplearning.net
what why how next Convolutional Neural Network
what why how next Convolutional Neural Network Images taken from
deeplearning.net
what why how next Convolutional Neural Network
what why how next The Starry Night Vincent van Gogh
Leon A. Gatys, Alexander S. Ecker and Matthias Bethge; A Neural Algorithm of Artistic Style
what why how next
what why how next Scene Description CNN + RNN Oriol
Vinyals et. all; Show and Tell: A Neural Image Caption Generator
Learning Sequences
what why how next Recurrent Neural Network Simple Elman Version
ℎ ℎ = ( , ℎ−1 , 0, 1) = ′(ℎ , 2)
what why how next Long Short Term Memory (LSTM) add
memory cells learn access mechanism Sepp Hochreiter and Jürgen Schmidhuber; Long short-term memory
None
what why how next
what why how next Fooling Deep Networks Anh Nguyen, Jason
Yosinski, Jeff Clune; Deep Neural Networks are Easily Fooled
Next Cool things to try
what why how next Hyperparameter optimization bayesian Optimization methods adadelta,
rmsprop . . . Regularization dropout, dither . . .
what why how next Attention & Memory NTMs, Memory Networks,
Stack RNNs . . . NLP Translation, description
what why how next Cognitive Hardware FPGA, GPU, Neuromorphic Chips
Scalable DL map-reduce, compute clusters
what why how next Deep Reinforcement Learning deepmindish things, deep
Q learning Energy models RBMs, DBNs . . .
https://www.reddit.com/r/MachineLearning/wiki
Theano (Python) | Torch (lua) | Caffe (C++) Github is
a friend
@AbhinavTushar ?