舉報(bào)

會(huì)員
Hands-On Mathematics for Deep Learning
Mostprogrammersanddatascientistsstrugglewithmathematics,havingeitheroverlookedorforgottencoremathematicalconcepts.ThisbookusesPythonlibrariestohelpyouunderstandthemathrequiredtobuilddeeplearning(DL)models.You'llbeginbylearningaboutcoremathematicalandmoderncomputationaltechniquesusedtodesignandimplementDLalgorithms.Thisbookwillcoveressentialtopics,suchaslinearalgebra,eigenvaluesandeigenvectors,thesingularvaluedecompositionconcept,andgradientalgorithms,tohelpyouunderstandhowtotraindeepneuralnetworks.Laterchaptersfocusonimportantneuralnetworks,suchasthelinearneuralnetworkandmultilayerperceptrons,withaprimaryfocusonhelpingyoulearnhoweachmodelworks.Asyouadvance,youwilldelveintothemathusedforregularization,multi-layeredDL,forwardpropagation,optimization,andbackpropagationtechniquestounderstandwhatittakestobuildfull-fledgedDLmodels.Finally,you’llexploreCNN,recurrentneuralnetwork(RNN),andGANmodelsandtheirapplication.Bytheendofthisbook,you'llhavebuiltastrongfoundationinneuralnetworksandDLmathematicalconcepts,whichwillhelpyoutoconfidentlyresearchandbuildcustommodelsinDL.
目錄(277章)
倒序
- 封面
- 版權(quán)信息
- About Packt
- Why subscribe?
- Contributors
- About the author
- About the reviewers
- Packt is searching for authors like you
- Preface
- Who this book is for
- What this book covers
- To get the most out of this book
- Download the color images
- Conventions used
- Get in touch
- Reviews
- Section 1: Essential Mathematics for Deep Learning
- Linear Algebra
- Comparing scalars and vectors
- Linear equations
- Solving linear equations in n-dimensions
- Solving linear equations using elimination
- Matrix operations
- Adding matrices
- Multiplying matrices
- Inverse matrices
- Matrix transpose
- Permutations
- Vector spaces and subspaces
- Spaces
- Subspaces
- Linear maps
- Image and kernel
- Metric space and normed space
- Inner product space
- Matrix decompositions
- Determinant
- Eigenvalues and eigenvectors
- Trace
- Orthogonal matrices
- Diagonalization and symmetric matrices
- Singular value decomposition
- Cholesky decomposition
- Summary
- Vector Calculus
- Single variable calculus
- Derivatives
- Sum rule
- Power rule
- Trigonometric functions
- First and second derivatives
- Product rule
- Quotient rule
- Chain rule
- Antiderivative
- Integrals
- The fundamental theorem of calculus
- Substitution rule
- Areas between curves
- Integration by parts
- Multivariable calculus
- Partial derivatives
- Chain rule
- Integrals
- Vector calculus
- Derivatives
- Vector fields
- Inverse functions
- Summary
- Probability and Statistics
- Understanding the concepts in probability
- Classical probability
- Sampling with or without replacement
- Multinomial coefficient
- Stirling's formula
- Independence
- Discrete distributions
- Conditional probability
- Random variables
- Variance
- Multiple random variables
- Continuous random variables
- Joint distributions
- More probability distributions
- Normal distribution
- Multivariate normal distribution
- Bivariate normal distribution
- Gamma distribution
- Essential concepts in statistics
- Estimation
- Mean squared error
- Sufficiency
- Likelihood
- Confidence intervals
- Bayesian estimation
- Hypothesis testing
- Simple hypotheses
- Composite hypothesis
- The multivariate normal theory
- Linear models
- Hypothesis testing
- Summary
- Optimization
- Understanding optimization and it's different types
- Constrained optimization
- Unconstrained optimization
- Convex optimization
- Convex sets
- Affine sets
- Convex functions
- Optimization problems
- Non-convex optimization
- Exploring the various optimization methods
- Least squares
- Lagrange multipliers
- Newton's method
- The secant method
- The quasi-Newton method
- Game theory
- Descent methods
- Gradient descent
- Stochastic gradient descent
- Loss functions
- Gradient descent with momentum
- The Nesterov's accelerated gradient
- Adaptive gradient descent
- Simulated annealing
- Natural evolution
- Exploring population methods
- Genetic algorithms
- Particle swarm optimization
- Summary
- Graph Theory
- Understanding the basic concepts and terminology
- Adjacency matrix
- Types of graphs
- Weighted graphs
- Directed graphs
- Directed acyclic graphs
- Multilayer and dynamic graphs
- Tree graphs
- Graph Laplacian
- Summary
- Section 2: Essential Neural Networks
- Linear Neural Networks
- Linear regression
- Polynomial regression
- Logistic regression
- Summary
- Feedforward Neural Networks
- Understanding biological neural networks
- Comparing the perceptron and the McCulloch-Pitts neuron
- The MP neuron
- Perceptron
- Pros and cons of the MP neuron and perceptron
- MLPs
- Layers
- Activation functions
- Sigmoid
- Hyperbolic tangent
- Softmax
- Rectified linear unit
- Leaky ReLU
- Parametric ReLU
- Exponential linear unit
- The loss function
- Mean absolute error
- Mean squared error
- Root mean squared error
- The Huber loss
- Cross entropy
- Kullback-Leibler divergence
- Jensen-Shannon divergence
- Backpropagation
- Training neural networks
- Parameter initialization
- All zeros
- Random initialization
- Xavier initialization
- The data
- Deep neural networks
- Summary
- Regularization
- The need for regularization
- Norm penalties
- L2 regularization
- L1 regularization
- Early stopping
- Parameter tying and sharing
- Dataset augmentation
- Dropout
- Adversarial training
- Summary
- Convolutional Neural Networks
- The inspiration behind ConvNets
- Types of data used in ConvNets
- Convolutions and pooling
- Two-dimensional convolutions
- One-dimensional convolutions
- 1 × 1 convolutions
- Three-dimensional convolutions
- Separable convolutions
- Transposed convolutions
- Pooling
- Global average pooling
- Convolution and pooling size
- Working with the ConvNet architecture
- Training and optimization
- Exploring popular ConvNet architectures
- VGG-16
- Inception-v1
- Summary
- Recurrent Neural Networks
- The need for RNNs
- The types of data used in RNNs
- Understanding RNNs
- Vanilla RNNs
- Bidirectional RNNs
- Long short-term memory
- Gated recurrent units
- Deep RNNs
- Training and optimization
- Popular architecture
- Clockwork RNNs
- Summary
- Section 3: Advanced Deep Learning Concepts Simplified
- Attention Mechanisms
- Overview of attention
- Understanding neural Turing machines
- Reading
- Writing
- Addressing mechanisms
- Content-based addressing mechanism
- Location-based address mechanism
- Exploring the types of attention
- Self-attention
- Comparing hard and soft attention
- Comparing global and local attention
- Transformers
- Summary
- Generative Models
- Why we need generative models
- Autoencoders
- The denoising autoencoder
- The variational autoencoder
- Generative adversarial networks
- Wasserstein GANs
- Flow-based networks
- Normalizing flows
- Real-valued non-volume preserving
- Summary
- Transfer and Meta Learning
- Transfer learning
- Meta learning
- Approaches to meta learning
- Model-based meta learning
- Memory-augmented neural networks
- Meta Networks
- Metric-based meta learning
- Prototypical networks
- Siamese neural networks
- Optimization-based meta learning
- Long Short-Term Memory meta learners
- Model-agnostic meta learning
- Summary
- Geometric Deep Learning
- Comparing Euclidean and non-Euclidean data
- Manifolds
- Discrete manifolds
- Spectral decomposition
- Graph neural networks
- Spectral graph CNNs
- Mixture model networks
- Facial recognition in 3D
- Summary
- Other Books You May Enjoy
- Leave a review - let other readers know what you think 更新時(shí)間:2021-06-18 18:56:13
推薦閱讀
- GitHub Essentials
- ETL數(shù)據(jù)整合與處理(Kettle)
- 云計(jì)算環(huán)境下的信息資源集成與服務(wù)
- Visual Studio 2015 Cookbook(Second Edition)
- Learning JavaScriptMVC
- Live Longer with AI
- Oracle 12c云數(shù)據(jù)庫備份與恢復(fù)技術(shù)
- 深入淺出Greenplum分布式數(shù)據(jù)庫:原理、架構(gòu)和代碼分析
- 圖數(shù)據(jù)實(shí)戰(zhàn):用圖思維和圖技術(shù)解決復(fù)雜問題
- 達(dá)夢(mèng)數(shù)據(jù)庫運(yùn)維實(shí)戰(zhàn)
- 數(shù)字IC設(shè)計(jì)入門(微課視頻版)
- Internet of Things with Python
- Deep Learning with R for Beginners
- MySQL技術(shù)內(nèi)幕:InnoDB存儲(chǔ)引擎
- 離線和實(shí)時(shí)大數(shù)據(jù)開發(fā)實(shí)戰(zhàn)
- 大數(shù)據(jù)隱私保護(hù)技術(shù)與治理機(jī)制研究
- 實(shí)用預(yù)測(cè)分析
- Python數(shù)據(jù)分析入門與實(shí)戰(zhàn)
- Building Multicopter Video Drones
- Practical Data Analysis Using Jupyter Notebook
- Unity AI Programming Essentials
- 數(shù)據(jù)自助服務(wù)實(shí)踐指南:數(shù)據(jù)開放與洞察提效
- 數(shù)據(jù)庫系統(tǒng):原理、設(shè)計(jì)與編程(MOOC版)
- 數(shù)據(jù)科學(xué)技術(shù):文本分析和知識(shí)圖譜
- Box2D for Flash Games
- 數(shù)據(jù)產(chǎn)品開發(fā)與經(jīng)營(yíng):從數(shù)據(jù)資源到數(shù)據(jù)資本
- Python數(shù)據(jù)分析
- Oracle DBA手記 3:數(shù)據(jù)庫性能優(yōu)化與內(nèi)部原理解析
- 數(shù)據(jù)資源的管理與調(diào)度:云環(huán)境下數(shù)據(jù)與資源協(xié)同技術(shù)
- Access 2010數(shù)據(jù)庫技術(shù)及應(yīng)用實(shí)訓(xùn)教程