Professional Documents
Culture Documents
AI Foundations and Applications: 8. Optimization of Learning Process
AI Foundations and Applications: 8. Optimization of Learning Process
Thien Huynh-The
HCM City Univ. Technology and Education
Jan, 2023
04/05/2023 HCMUTE 1
Course Contents
1. Introduction (1 week)
2. Intelligent Agents (1 week)
3. Machine Learning: Fundamental Concepts (3 weeks)
1. Feature Engineering
2. Linear Regression
3. Overfitting
4. Machine Learning Algorithms (3 weeks)
1. Decision Tree Note: This is the tentative contents and may
2. K-nearest neighbors be changed depending upon the teaching
3. K-means clustering progress.
5. Neural Networks (2 weeks) Note: There are no final exam of this course,
6. Deep Learning (4 weeks) therefore it has one reserved week for
marking
1. Convolutional Neural Networks
2. Applications with codes
• Homework assignment:
• Send code to teacher by email before deadline
• Over deadline is zero (no negotiate)
• Code with detailed comments (to rate how much students understand their codes)
• Run codes and record the laptop screen to demonstrate your codes (even it may be error)
• Final project:
• Send code to teacher by email before deadline
• Over deadline is zero (no negotiate)
• Code with detailed comments (to rate how much students understand their codes)
• Run codes directly on student’s laptop and show to teacher
• Answer question from teacher individually
Local minima
• The objective function of deep learning usually has many local minima
• The numerical solution obtained by the final iteration may only minimize the
objective function locally, rather than globally.
• As the gradient of the objective function's solutions approaches or becomes zero
• Instead of using only the gradient of the current step to guide the search,
momentum also accumulates the gradient of the past steps to determine the
direction to go.
• The equations of gradient descent are revised as follows.
Assignment 2 (mandatory)
Design a multilayer neural networks with input layer, 02 hidden layers (sigmoid) , output
layer (softmax). Apply the optimization methods of Momentum và Adam. Compare the
accuracy and Converging time among two methods. Assume that the MNIST dataset is
used for training and testing the neural network.
Submission by email (thienht@hcmute.edu.vn) the Python code files:
• Momentum.py
• Adam.py
• Deadline: 24/04/2023.
• Take 10% of overall subject score.