Course Outline
Part 1 – Deep Learning and DNN Concepts
Introduction AI, Machine Learning & Deep Learning
- History, basic concepts, and usual applications of artificial intelligence far from the fantasies carried by this domain.
- Collective Intelligence: aggregating knowledge shared by many virtual agents.
- Genetic algorithms: to evolve a population of virtual agents by selection.
- Usual Learning Machine: definition.
- Types of tasks: supervised learning, unsupervised learning, reinforcement learning.
- Types of actions: classification, regression, clustering, density estimation, reduction of dimensionality.
- Examples of Machine Learning algorithms: Linear regression, Naive Bayes, Random Tree.
- Machine learning VS Deep Learning: problems on which Machine Learning remains today the state of the art (Random Forests & XGBoosts).
Basic Concepts of a Neural Network (Application: multi-layer perceptron)
- Reminder of mathematical bases.
- Definition of a network of neurons: classical architecture, activation and.
- Weighting of previous activations, depth of a network.
- Definition of the learning of a network of neurons: functions of cost, back-propagation, Stochastic gradient descent, maximum likelihood.
- Modeling of a neural network: modeling input and output data according to The type of problem (regression, classification ...). Curse of dimensionality.
- Distinction between Multi-feature data and signal. Choice of a cost function according to the data.
- Approximation of a function by a network of neurons: presentation and examples.
- Approximation of a distribution by a network of neurons: presentation and examples.
- Data Augmentation: how to balance a dataset.
- Generalization of the results of a network of neurons.
- Initialization and regularization of a neural network: L1 / L2 regularization, Batch Normalization.
- Optimization and convergence algorithms.
Standard ML / DL Tools
A simple presentation with advantages, disadvantages, position in the ecosystem, and use is planned.
- Data management tools: Apache Spark, Apache Hadoop Tools.
- Machine Learning: Numpy, Scipy, Sci-kit.
- DL high level frameworks: PyTorch, Keras, Lasagne.
- Low level DL frameworks: Theano, Torch, Caffe, Tensorflow.
Convolutional Neural Networks (CNN).
- Presentation of the CNNs: fundamental principles and applications.
- Basic operation of a CNN: convolutional layer, use of a kernel,.
- Padding & stride, feature map generation, pooling layers. Extensions 1D, 2D, and 3D.
- Presentation of the different CNN architectures that brought the state of the art in classification.
- Images: LeNet, VGG Networks, Network in Network, Inception, Resnet. Presentation of Innovations brought about by each architecture and their more global applications (Convolution 1x1 or residual connections).
- Use of an attention model.
- Application to a common classification case (text or image).
- CNNs for generation: super-resolution, pixel-to-pixel segmentation. Presentation of.
- Main strategies for increasing feature maps for image generation.
Recurrent Neural Networks (RNN).
- Presentation of RNNs: fundamental principles and applications.
- Basic operation of the RNN: hidden activation, back propagation through time, Unfolded version.
- Evolutions towards the Gated Recurrent Units (GRUs) and LSTM (Long Short Term Memory).
- Presentation of the different states and the evolutions brought by these architectures.
- Convergence and vanishing gradient problems.
- Classical architectures: Prediction of a temporal series, classification ....
- RNN Encoder Decoder type architecture. Use of an attention model.
- NLP applications: word / character encoding, translation.
- Video Applications: prediction of the next generated image of a video sequence.
Generational models: Variational AutoEncoder (VAE) and Generative Adversarial Networks (GAN).
- Presentation of the generational models, link with the CNNs.
- Auto-encoder: reduction of dimensionality and limited generation.
- Variational Auto-encoder: generational model and approximation of the distribution of a given. Definition and use of latent space. Reparameterization trick. Applications and Limits observed.
- Generative Adversarial Networks: Fundamentals.
- Dual Network Architecture (Generator and discriminator) with alternate learning, cost functions available.
- Convergence of a GAN and difficulties encountered.
- Improved convergence: Wasserstein GAN, Began. Earth Moving Distance.
- Applications for the generation of images or photographs, text generation, super-resolution.
Deep Reinforcement Learning.
- Presentation of reinforcement learning: control of an agent in a defined environment.
- By a state and possible actions.
- Use of a neural network to approximate the state function.
- Deep Q Learning: experience replay, and application to the control of a video game.
- Optimization of learning policy. On-policy && off-policy. Actor critic architecture. A3C.
- Applications: control of a single video game or a digital system.
Part 2 – Theano for Deep Learning
Theano Basics
- Introduction.
- Installation and Configuration.
Theano Functions
- inputs, outputs, updates, givens.
Training and Optimization of a neural network using Theano
- Neural Network Modeling.
- Logistic Regression.
- Hidden Layers.
- Training a network.
- Computing and Classification.
- Optimization.
- Log Loss.
Testing the model
Part 3 – DNN using Tensorflow
TensorFlow Basics
- Creation, Initializing, Saving, and Restoring TensorFlow variables.
- Feeding, Reading and Preloading TensorFlow Data.
- How to use TensorFlow infrastructure to train models at scale.
- Visualizing and Evaluating models with TensorBoard.
TensorFlow Mechanics
- Prepare the Data.
- Download.
- Inputs and Placeholders.
-
Build the Graphs.
- Inference.
- Loss.
- Training.
-
Train the Model.
- The Graph.
- The Session.
- Train Loop.
-
Evaluate the Model.
- Build the Eval Graph.
- Eval Output.
The Perceptron
- Activation functions.
- The perceptron learning algorithm.
- Binary classification with the perceptron.
- Document classification with the perceptron.
- Limitations of the perceptron.
From the Perceptron to Support Vector Machines
- Kernels and the kernel trick.
- Maximum margin classification and support vectors.
Artificial Neural Networks
- Nonlinear decision boundaries.
- Feedforward and feedback artificial neural networks.
- Multilayer perceptrons.
- Minimizing the cost function.
- Forward propagation.
- Back propagation.
- Improving the way neural networks learn.
Convolutional Neural Networks
- Goals.
- Model Architecture.
- Principles.
- Code Organization.
- Launching and Training the Model.
- Evaluating a Model.
Basic Introductions to be given to the below modules (Brief Introduction to be provided based on time availability):
Tensorflow - Advanced Usage
- Threading and Queues.
- Distributed TensorFlow.
- Writing Documentation and Sharing your Model.
- Customizing Data Readers.
- Manipulating TensorFlow Model Files.
TensorFlow Serving
- Introduction.
- Basic Serving Tutorial.
- Advanced Serving Tutorial.
- Serving Inception Model Tutorial.
Requirements
Background in physics, mathematics, and programming. Involvement in image processing activities.
Delegates should have a prior understanding of machine learning concepts and should have worked with Python programming and libraries.
Testimonials (2)
The training was organized and well-planned out, and I come out of it with systematized knowledge and a good look at topics we looked at
Magdalena - Samsung Electronics Polska Sp. z o.o.
Course - Deep Learning with TensorFlow 2
I really liked the end where we took the time to play around with CHAT GPT. The room was not set up the best for this- instead of one large table a couple of small ones so we could get into small groups and brainstorm would have helped