Skip to content

Latest commit

 

History

History
201 lines (169 loc) · 11.2 KB

README.md

File metadata and controls

201 lines (169 loc) · 11.2 KB

My work on CS50’s Introduction to AI with Python

https://cs50.harvard.edu/ai/

This course explores the concepts and algorithms at the foundation of modern artificial intelligence, diving into the ideas that give rise to technologies like game-playing engines, handwriting recognition, and machine translation. Through hands-on projects, students gain exposure to the theory behind graph search algorithms, classification, optimization, reinforcement learning, and other topics in artificial intelligence and machine learning as they incorporate them into their own Python programs. By course’s end, students emerge with experience in libraries for machine learning as well as knowledge of artificial intelligence principles that enable them to design intelligent systems of their own.

Certificate: https://courses.edx.org/certificates/2ec5ff3f06b24bb595c21e3821591538

Notes

I've taken some notes on key concepts and algorithms throughout the lectures for future reference.

Lecture 0: Search

Concepts

  • Agent: entity that perceives its environment and acts upon that environment.
  • State: a configuration of the agent and its environment.
  • Actions: choices that can be made in a state.
  • Transition model: a description of what state results from performing any applicable action in any state.
  • Path cost: numerical cost associated with a given path.
  • Evaluation function: function that estimates the expected utility of the game from a given state.

Algorithms

  • DFS (depth first search): search algorithm that always expands the deepest node in the frontier.
  • BFS (breath first search): search algorithm that always expands the shallowest node in the frontier.
  • Greedy best-first search: search algorithm that expands the node that is closest to the goal, as estimated by an heuristic function h(n).
  • A* search: search algorithm that expands node with lowest value of the "cost to reach node" plus the "estimated goal cost".
  • Minimax: adversarial search algorithm.

Projects

Lecture 1: Knowledge

Concepts

  • Sentence: an assertion about the world in a knowledge representation language.
  • Knowledge base: a set of sentences known by a knowledge-based agent.
  • Entailment: a entails b if in every model in which sentence a is true, sentence b is also true.
  • Inference: the process of deriving new sentences from old ones.
  • Conjunctive normal form: logical sentence that is a conjunction of clauses.
  • First order logic: Propositional logic.
  • Second order logic: Proposition logic with universal and existential quantification.

Algorithms

  • Model checking: enumerate all possible models and see if a proposition is true in every one of them.
  • Conversion to CNF and Inference by resolution

Projects

Lecture 2: Uncertainty

Concepts

  • Unconditional probability: degree of belief in a proposition in the absence of any other evidence.
  • Conditional probability: degree of belief in a proposition given some evidence that has already been revealed.
  • Random variable: a variable in probability theory with a domain of possible values it can take on.
  • Independence: the knowledge that one event occurs does not affect the probability of the other event.
  • Bayes' Rule: P(a) P(b|a) = P(b) P(a|b)
  • Bayesian network: data structure that represents the dependencies among random variables.
  • Markov assumption: the assumption that the current state depends on only a finite fixed number of previous states.
  • Markov chain: a sequence of random variables where the distribution of each variable follows the Markov assumption.
  • Hidden Markov Model: a Markov model for a system with hidden states that generate some observed event.

Algorithms

  • Inference by enumeration
  • Sampling
  • Likelihood weighting

Projects

Lecture 3: Optimization

Concepts

  • Optimization: choosing the best option from a set of options.

Algorithms

  • Local Search Hill climbing
    • steepest-ascent: choose the highest-valued neighbor.
    • stochastic: choose randomly from higher-valued neighbors.
    • first-choice: choose the first higher-valued neighbor.
    • random-restart: conduct hill climbing multiple times.
    • local beam search: chooses the k highest-valued neighbors.
  • Simulated annealing: early on, more likely to accept worse-valued neighbors than the current state.
  • Linear programming
    • Simplex
    • Interior-Point
  • Constraint satisfaction problems
    • Arc consistency: to make X arc-consistent with respect to Y, removing elements from X's domain until every choice for X has a possible choice for Y
    • Backtracking search

Projects

Lecture 4: Learning

Concepts

  • Supervised learning: given a data set of input-output pairs, learn a function to map inputs to outputs.
    • Classification: supervised learning task of learning a function mapping an input point to a discrete category.
    • Regression: supervised learning task of learning a function mapping and input point to a continuous value.
    • Loss function: function that express how poorly our hypothesis performs (L1, L2).
    • Overfitting: when a model fits too closely to a particular data set and therefore may fail to generalize to future data.
    • Regularization: penalizing hypotheses that are more complex to favor simpler, more general hypotheses.
    • Holdout cross-validation: splitting data into a training set and a test set, such that learning happens on the training set and is evaluated on the test set.
    • k-fold cross-validation: splitting data into k sets, and experimenting k times, using each set as a test set once, and using remaining data as training set.
  • Reinforcement learning: given a set of rewards or punishments, learn what actions to take in the future.
  • Unsupervised learning: given input data without any additional feedback, learn patterns.
  • Clustering: organizing a set of objects into groups in such a way that similar objects tend to be in the same group.

Algorithms

  • k-nearest-neighbor classification: given an input, chooses the most common class out of the k nearest data points to that input.
  • Support Vector Machines (SVM)
  • Markov decision process: model for decision-making, representing states, actions and their rewards.
  • Q-learning: method for learning a function Q(s, a), estimate of the value of performing action a in state s.
  • Greedy decision-making
  • epsilon-greedy
  • k-means clustering: clustering data based on repeatedly assigning points to clusters and updating those clusters' centers.

Projects

Lecture 5: Neural Networks

Concepts

  • Artificial neural network: mathematical model for learning inspired by biological neural networks.
  • Multilayer neural network: artificial neural network with an input layer, an output layer, and at least one hidden layer.
  • Deep neural network: neural network with multiple hidden layer.
  • Dropout: temporarily removing units - selected at random - from a neural network to prevent over-reliance on certain units.
  • Image convolution: applying a filter that adds each pixel value of an image to its neighbors, weighted according to a kernel matrix.
  • Pooling: reducing the size of an input by sampling from regions in the input.
  • Convolutional neural network: neural networks that use convolution, usually for analyzing images.
  • Recurrent neural network: neural network that generates output that feeds back into its own inputs.

Algorithms

  • Gradient descent: algorithm for minimizing loss when training neural network.
  • Backpropagation: algorithm for training neural networks with hidden layers.

Projects

Lecture 6: Language

Concepts

  • Natural language processing
  • n-gram: a continuous sequence of n items inside of a text.
  • Tokenization: the task of splitting a sequence of characters into pieces (tokens).
  • Text Categorization
    • Bag-of-words model: represent text as an unordered collection of words.
  • Information retrieval: the task of finding relevant documents in response to a user query.
    • Topic modeling: models for discovering the topics for a set of documents.
    • Term frequency: number of times a term appears in a document.
      • Function words: words that have little meaning on their own, but are used to grammatically connect other words.
      • Content words: words that carry meaning independently.
    • Inverse document frequency: measure of how common or rare a word is across documents.
  • Information extraction: the task of extracting knowledge from documents.
  • WordNet: a lexical database of semantic relations between words.
  • Word representation: looking for a way to represent the meaning of a word for further processing.
    • one-hot: representation of meaning as a vector with a single 1, and with other values as 0.
    • distribution: representation of meaning distributed across multiple values.

Algorithms

  • Markov model applied to language: generating the next word based on the previous words and a probability.
  • Naive Bayes: based on the Bayes' Rule to calculate probability of a text being in a certain category, given it contains specific words. Assuming every word is independent of each other.
    • Additive smoothing: adding a value a to each value in our distribution to smooth the data.
    • Laplace smoothing: adding 1 to each value in our distribution (pretending we've seen each value one more time than we actually have).
  • tf-idf: ranking of what words are important in a document by multiplying term frequency (TF) by inverse document frequency (IDF).
  • Automated template generation: giving AI some terms and let it look into a corpus for patterns where those terms show up together. Then it can use those templates to extract new knowledge from the corpus.
  • word2vec: model for generating word vectors.
  • skip-gram architecture: neural network architecture for predicting context words given a target word.

Projects