= Pi (i=1 to n) P(xi|+ve) * P(+ve)(Class variable)/ Sum (C=+ve to -ve) (P([x1,x2....xn],C))
= Pi (i=1 to n) P(xi|+ve) * P(+ve)(Class variable) / (( Pi (i=1 to n ) P(xi|+ve)*P(+ve))+..+ Pi (i=1 to n) P(xi|-ve)*P(-ve))
P(X,Y) = Sum (y=y1...yn) P(X,Y=y)
W2 = P(+ve|xi=1) (human approach) (Naive bayes) (naively thinking that all features are independent)
Regression : output contains real values -> (predicting the position of joints of a body given an image of a person)
Classification : output classifies to discrete set of classes
-> predicting the posture of a person (sitting,walking,standing,running) given an image of a person
(Numerical/Categorical)
Representation Learning: embedding
Deeplearning is all about hierarchical representation learning.
Metric Learning: distance( of facial features)/similarity(of fashion apparels)/relevance( of search document)
Structured Output(Models): auto-corrects dependent outputs based on output on the upper hierarchy outputs.
Types of input:
Bag of features,bag of words: ( finding whether a feature is present/not without caring where the feature occurs in the input)
eg: Using unsupervized learning to convert the input to a given set of classes(clusters) and use them as bag of features.
Spatial data(sequential data): if there local dependencies use CNN(convolutional nn) if there are near past dependencies in the data use RNN(Recurrent NN-LSTM)
eg: stock market temporal data / speech data/ image data
Non-Parametric models : k-NN(K-nearest neighbor), Decision Trees, Random Forests (independent of parameters)
-> very inaccurate because doesn't know much about the data
-> more accurate coz the knows more about the parameters from the data
Types of Learning:
supervized learning -> labeled data
unsupervized learning -> unlabeled data
exercise: *take 3s from mnist data *create a gmm model with them and *cluster them with 5/3/10 gaussians.
*take all images and cluster them to 10 gaussians.
semi-supervized learning -> combination of supervized and unsupervized models
Auto Encoder:
finding a low dimensional representation of a high dimensional data.
eg. image of 200x200 pixels create a fingerprint of image of 128 dimensions.
exercise: use the 128 dimensional data to reconstruct the 200x200 image(using inverse of the model).
Reinforcement learning:
eg:playing chess -> using the final result of the game to assign weights/score for moves that were made upto the final result. and training the model to predict based on those scores.