Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Slide improvements #27

Open
FunnyPhantom opened this issue Feb 5, 2023 · 30 comments
Open

Slide improvements #27

FunnyPhantom opened this issue Feb 5, 2023 · 30 comments

Comments

@FunnyPhantom
Copy link

These slides are getting used as a reference for teaching in the ML for BioInformatic course as well.
In the process of class, some points of improvement got found. This issues tries to serve as a thread for conveying these improvements.

@FunnyPhantom
Copy link
Author

In the first slide, the slide that shows the category of MLs have 2 typos that need to be fixed.
First: Recommender system (is written recommended system)
Second: Feature Elimination (is written feature elicitation)

@mahsayazdani
Copy link
Collaborator

@FunnyPhantom Thank you for your suggestion. Please provide the exact name of the slide and page numbers.

@FunnyPhantom
Copy link
Author

@mahsayazdani the file that needs improvement resides here:
Slides/Chapter_02_Classical_Models/Introduction to ML/Figs/1.jpeg

@FunnyPhantom
Copy link
Author

There is another issue in here: https://github.com/asharifiz/Introduction_to_Machine_Learning/blob/3a595142161801b224e9fd06b1e447de7dfb0749/Slides/Chapter_02_Classical_Models/Loss/Loss.tex#L269
The definition of this function should replace cosh with Logcosh.

@FunnyPhantom
Copy link
Author

@FunnyPhantom
Copy link
Author

In this slide, there are multiple improvements:
https://github.com/asharifiz/Introduction_to_Machine_Learning/blob/main/Slides/Chapter_04_Tabular_Data_Models/Chapter%204%20(ML_Models_for_Tabular_Datasets).pdf

  1. The Majority Voting slide should be after soft clustering.
  2. In the Why Majority Voting slide, the ensemble error is missing a choose(n,k).

@FunnyPhantom
Copy link
Author

  1. for the P(k) in the slide above, the condition of k > [n/2] should be removed. (but not for the ensemble error)

@FunnyPhantom
Copy link
Author

FunnyPhantom commented Apr 9, 2023

On each node, i will make my tree to use random subset of features (sqrt(n))
photo_2023-04-09_18-48-18

There is a sentence which is not there iand does not convey the sould of fthe topic.

@FunnyPhantom
Copy link
Author

Also for the xavier initialization slides, please denote whether it has been initialized with normal distribution, or uniform distribution.

@FunnyPhantom
Copy link
Author

@FunnyPhantom
Copy link
Author

Also, Early stopping slides to be put before L1/L2 regularization term.

@FunnyPhantom
Copy link
Author

FunnyPhantom commented May 2, 2023

@FunnyPhantom
Copy link
Author

@FunnyPhantom
Copy link
Author

In the same context as the comment above, the Epochs of each network should be explicitly specified.

Moreover, the epoch number should be starting from 1 and not 0 (Since zero is indicating not even once gradient decent has been ran)

@FunnyPhantom
Copy link
Author

Page 20-22 can be aggregated into one slide.

Moreover providing a famous kernel with well known function (such has horizontal or vertical edge detection) can be more helpful in the instruction process.

@FunnyPhantom
Copy link
Author

FunnyPhantom commented May 8, 2023

In page 87, the number of dimensions is written as Channel * Width * Height. The common way to denote them is Width * Height * Channel. Changing this can help reduce student's confusion.

Also there it on the bottom of the page, the number is written as n_kernels * channels * width * height. In addition to changing the suggestion above, it would be good to write the number as n_kernels x ( width * height * channel) This will reduce confusion as it is showing the first number is the number of kernels and not the tensor dimensions.

@FunnyPhantom
Copy link
Author

** VERY IMPORTANT CHANGE ** All the slides for the channel section in the CNN slides should be moved BEFORE the strides Section

@FunnyPhantom
Copy link
Author

FunnyPhantom commented May 28, 2023

Very important!

Please add the following images for the GAN slides to convey the concept better:
image ref
image ref
image

(Based on Dr. Sharifi's comments, it would be best to revise the slides for GAN rigorously and let him review the results)
(Also Dr. Sharifi was searching these search terms in the class: "CycleGan" "DiscoGan")

@FunnyPhantom
Copy link
Author

page 22/65 RNN, a simpler example which can convey the meaning better. Preferably with same dimension and with a non linear activation function

@FunnyPhantom
Copy link
Author

  • Notation in RNN slides is not consistent. It should be consistent. Ideally everything should be defined based on LSTM Notation

@FunnyPhantom
Copy link
Author

FunnyPhantom commented May 28, 2023

Page 37 RNN slides, there are two activation functions, one can be only tanh, but the other one can be tanh or sigmoid. This should be explicitely specified.

@FunnyPhantom
Copy link
Author

Page 38 RNN slides, for the Gated recurrent unit, it should have more details for its architecture. (Compare with the previous page slide which is the architecture of a simple RNN Unit)

@FunnyPhantom
Copy link
Author

In the introduction of the RNN slides, the limitation of previous model should be specified first in order to give more context about the problem RNN solves.

@FunnyPhantom
Copy link
Author

RNN limitation should be specified, before going to GRU for the same reason explained above.

@FunnyPhantom
Copy link
Author

GRU limitation should be specified before going to LSTM for the same reason above.

@FunnyPhantom
Copy link
Author

FunnyPhantom commented Jun 6, 2023

Transformer should be introduced by the limitation of LSTM. (BPTT (need to be sequential), vanishing or exploding gradient, long range dependency)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants