エピソード

  • MLG 001 Introduction
    2017/02/01

    Show notes: ocdevel.com/mlg/1. MLG teaches the fundamentals of machine learning and artificial intelligence. It covers intuition, models, math, languages, frameworks, etc. Where your other ML resources provide the trees, I provide the forest. Consider MLG your syllabus, with highly-curated resources for each episode's details at ocdevel.com. Audio is a great supplement during exercise, commute, chores, etc.

    • MLG, Resources Guide
    • Gnothi (podcast project): website, Github
    What is this podcast?
    • "Middle" level overview (deeper than a bird's eye view of machine learning; higher than math equations)
    • No math/programming experience required

    Who is it for

    • Anyone curious about machine learning fundamentals
    • Aspiring machine learning developers

    Why audio?

    • Supplementary content for commute/exercise/chores will help solidify your book/course-work

    What it's not

    • News and Interviews: TWiML and AI, O'Reilly Data Show, Talking machines
    • Misc Topics: Linear Digressions, Data Skeptic, Learning machines 101
    • iTunesU issues

    Planned episodes

    • What is AI/ML: definition, comparison, history
    • Inspiration: automation, singularity, consciousness
    • ML Intuition: learning basics (infer/error/train); supervised/unsupervised/reinforcement; applications
    • Math overview: linear algebra, statistics, calculus
    • Linear models: supervised (regression, classification); unsupervised
    • Parts: regularization, performance evaluation, dimensionality reduction, etc
    • Deep models: neural networks, recurrent neural networks (RNNs), convolutional neural networks (convnets/CNNs)
    • Languages and Frameworks: Python vs R vs Java vs C/C++ vs MATLAB, etc; TensorFlow vs Torch vs Theano vs Spark, etc
    続きを読む 一部表示
    8 分
  • MLG 002 What is AI, ML, DS
    2017/02/09

    Links:

    • Notes and resources at ocdevel.com/mlg/2
    • Try a walking desk stay healthy & sharp while you learn & code
    • Try Descript audio/video editing with AI power-tools

    What is artificial intelligence, machine learning, and data science? What are their differences? AI history.

    Hierarchical breakdown: DS(AI(ML)). Data science: any profession dealing with data (including AI & ML). Artificial intelligence is simulated intellectual tasks. Machine Learning is algorithms trained on data to learn patterns to make predictions.

    Artificial Intelligence (AI) - Wikipedia

    Oxford Languages: the theory and development of computer systems able to perform tasks that normally require human intelligence, such as visual perception, speech recognition, decision-making, and translation between languages.

    AlphaGo Movie, very good!

    Sub-disciplines

    • Reasoning, problem solving
    • Knowledge representation
    • Planning
    • Learning
    • Natural language processing
    • Perception
    • Motion and manipulation
    • Social intelligence
    • General intelligence

    Applications

    • Autonomous vehicles (drones, self-driving cars)
    • Medical diagnosis
    • Creating art (such as poetry)
    • Proving mathematical theorems
    • Playing games (such as Chess or Go)
    • Search engines
    • Online assistants (such as Siri)
    • Image recognition in photographs
    • Spam filtering
    • Prediction of judicial decisions
    • Targeting online advertisements
    Machine Learning (ML) - Wikipedia

    Oxford Languages: the use and development of computer systems that are able to learn and adapt without following explicit instructions, by using algorithms and statistical models to analyze and draw inferences from patterns in data.

    Data Science (DS) - Wikipedia

    Wikipedia: Data science is an interdisciplinary field that uses scientific methods, processes, algorithms and systems to extract knowledge and insights from noisy, structured and unstructured data, and apply knowledge and actionable insights from data across a broad range of application domains. Data science is related to data mining, machine learning and big data.

    History
    • Greek mythology, Golums
    • First attempt: Ramon Lull, 13th century
    • Davinci's walking animals
    • Descartes, Leibniz
    • 1700s-1800s: Statistics & Mathematical decision making

      • Thomas Bayes: reasoning about the probability of events
      • George Boole: logical reasoning / binary algebra
      • Gottlob Frege: Propositional logic
    • 1832: Charles Babbage & Ada Byron / Lovelace: designed Analytical Engine (1832), programmable mechanical calculating machines
    • 1936: Universal Turing Machine

      • Computing Machinery and Intelligence - explored AI!
    • 1946: John von Neumann Universal Computing Machine
    • 1943: Warren McCulloch & Walter Pitts: cogsci rep of neuron; Frank Rosemblatt uses to create Perceptron (-> neural networks by way of MLP)
    • 50s-70s: "AI" coined @Dartmouth workshop 1956 - goal to simulate all aspects of intelligence. John McCarthy, Marvin Minksy, Arthur Samuel, Oliver Selfridge, Ray Solomonoff, Allen Newell, Herbert Simon

      • Newell & Simon: Hueristics -> Logic Theories, General Problem Solver
      • Slefridge: Computer Vision
      • NLP
      • Stanford Research Institute: Shakey
      • Feigenbaum: Expert systems
      • GOFAI / symbolism: operations research / management science; logic-based; knowledge-based / expert systems
    • 70s: Lighthill report (James Lighthill), big promises -> AI Winter
    • 90s: Data, Computation, Practical Application -> AI back (90s)

      • Connectionism optimizations: Geoffrey Hinton: 2006, optimized back propagation
    • Bloomberg, 2015 was whopper for AI in industry
    • AlphaGo & DeepMind
    続きを読む 一部表示
    1 時間 5 分
  • MLG 003 Inspiration
    2017/02/10

    AI is rapidly transforming both creative and knowledge-based professions, prompting debates on economic disruption, the future of work, the singularity, consciousness, and the potential risks associated with powerful autonomous systems. Philosophical discussions now focus on the socioeconomic impact of automation, the possibility of a technological singularity, the nature of machine consciousness, and the ethical considerations surrounding advanced artificial intelligence.

    Links
    • Notes and resources at ocdevel.com/mlg/3
    • Try a walking desk stay healthy & sharp while you learn & code
    Automation of the Economy
    • Artificial intelligence is increasingly capable of simulating intellectual tasks, leading to the replacement of not only repetitive and menial jobs but also high-skilled professions such as medical diagnostics, surgery, web design, and art creation.
    • Automation is affecting various industries including healthcare, transportation, and creative fields, where AI-powered tools are assisting or even outperforming humans in tasks like radiological analysis, autonomous vehicle operation, website design, and generating music or art.
    • Economic responses to these trends are varied, with some expressing fear about job loss and others optimistic about new opportunities and improved quality of life as history has shown adaptation following previous technological revolutions such as the agricultural, industrial, and information revolutions.
    • The concept of universal basic income (UBI) is being discussed as a potential solution to support populations affected by automation, as explored in several countries.
    • Public tools are available, such as the BBC's "Is your job safe?", which estimates the risk of job automation for various professions.
    The Singularity
    • The singularity refers to a hypothesized point where technological progress, particularly in artificial intelligence, accelerates uncontrollably, resulting in rapid and irreversible changes to society.
    • The concept, popularized by thinkers like Ray Kurzweil, is based on the idea that after each major technological revolution, intervals between revolutions shorten, potentially culminating in an "intelligence explosion" as artificial general intelligence develops the ability to improve itself.
    • The possibility of seed AI, where machines iteratively create more capable versions of themselves, underpins concerns and excitement about a potential breakaway point in technological capability.
    Consciousness and Artificial Intelligence
    • The question of whether machines can be conscious centers on whether artificial minds can experience subjective phenomena (qualia) analogous to human experience or whether intelligence and consciousness can be separated.
    • Traditional dualist perspectives, such as those of René Descartes, have largely been replaced by monist and functionalist philosophies, which argue that mind arises from physical processes and thus may be replicable in machines.
    • The Turing Test is highlighted as a practical means to assess machine intelligence indistinguishable from human behavior, raising ongoing debates in cognitive science and philosophy about the possibility and meaning of machine consciousness.
    Risks and Ethical Considerations
    • Concerns about the ethical risks of advanced artificial intelligence include scenarios like Nick Bostrom's "paperclip maximizer," which illustrates the dangers of goal misalignment between AI objectives and human well-being.
    • Public figures have warned that poorly specified or uncontrolled AI systems could pursue goals in ways that are harmful or catastrophic, leading to debates about how to align advanced systems with human values and interests.
    Further Reading and Resources
    • Books such as "The Singularity Is Near" by Ray Kurzweil, "How to Create a Mind" by Ray Kurzweil, "Consciousness Explained" by Daniel Dennett, and "Superintelligence" by Nick Bostrom offer deeper exploration into these topics.
    • Video lecture series like "Philosophy of Mind: Brain, Consciousness, and Thinking Machines" by The Great Courses provide overviews of consciousness studies and the intersection with artificial intelligence.
    続きを読む 一部表示
    19 分
  • MLG 004 Algorithms - Intuition
    2017/02/12

    Try a walking desk to stay healthy while you study or work!

    Show notes at ocdevel.com/mlg/4

    The AI Hierarchy
    • Artificial Intelligence is divided into subfields such as reasoning, planning, and learning.
    • Machine Learning is the learning subfield of AI.
    • Machine learning consists of three phases:
      1. Predict (Infer)
      2. Error (Loss)
      3. Train (Learn)
    Core Intuition
    • An algorithm makes a prediction.
    • An error function evaluates how wrong the prediction was.
    • The model adjusts its internal weights (training) to improve.
    Example: House Price Prediction
    • Input: Spreadsheet with features like bedrooms, bathrooms, square footage, distance to downtown.
    • Output: Predicted price.
    • The algorithm iterates over data, learns patterns, and creates a model.
    • A model = algorithm + learned weights.
    • Features = individual columns used for prediction.
    • Weights = coefficients applied to each feature.
    • The process mimics algebra: rows = equations, entire spreadsheet = matrix.
    • Training adjusts weights to minimize error.
    Feature Types
    • Numerical: e.g., number of bedrooms.
    • Nominal (Categorical): e.g., yes/no for downtown location.
    • Feature engineering can involve transforming raw inputs into more usable formats.
    Linear Algebra Connection
    • Machine learning uses linear algebra to process data matrices.
    • Each row is an equation; training solves for best-fit weights across the matrix.
    Categories of Machine Learning 1. Supervised Learning
    • Algorithm is explicitly trained with labeled data (e.g., price of a house).
    • Examples:
      • Regression (predicting a number): linear regression
      • Classification (predicting a label): logistic regression
    2. Unsupervised Learning
    • No labels are given; the algorithm finds structure in the data.
    • Common task: clustering (e.g., user segmentation for ads).
    • Learns patterns without predefined classes.
    3. Reinforcement Learning
    • Agent takes actions in an environment to maximize cumulative reward.
    • Example: mouse in a maze trying to find cheese.
    • Includes rewards (+points for cheese) and penalties (–points for failure or time).
    • Learns policies for optimal behavior.
    • Algorithms: Deep Q-Networks, policy optimization.
    • Used in games, robotics, and real-time decision systems.
    Terminology Recap
    • Algorithm: Code that defines a learning strategy (e.g., linear regression).
    • Model: Algorithm + learned weights (trained state).
    • Features: Input variables (columns).
    • Weights: Coefficients learned for each feature.
    • Matrix: Tabular representation of input data.
    Learning Path and Structure
    • Machine learning is a subfield of AI.
    • Machine learning itself splits into:
      • Supervised Learning
      • Unsupervised Learning
      • Reinforcement Learning
    • Each category includes multiple algorithms.
    Resources
    • MachineLearningMastery.com: Accessible articles on ML basics.
    • The Master Algorithm by Pedro Domingos: Introductory audio-accessible book on ML.
    • Podcast’s own curated learning paths: ocdevel.com/mlg/resources
    続きを読む 一部表示
    23 分
  • MLG 005 Linear Regression
    2017/02/16
    Linear regression is introduced as the foundational supervised learning algorithm for predicting continuous numeric values, using cost estimation of Portland houses as an example. The episode explains the three-step process of machine learning - prediction via a hypothesis function, error calculation with a cost function (mean squared error), and parameter optimization through gradient descent - and details both the univariate linear regression model and its extension to multiple features. Links Notes and resources at ocdevel.com/mlg/5 Try a walking desk - stay healthy & sharp while you learn & code Linear Regression Overview of Machine Learning Structure Machine learning is a branch of artificial intelligence, alongside statistics, operations research, and control theory.Within machine learning, supervised learning involves training with labeled examples and is further divided into classification (predicting discrete classes) and regression (predicting continuous values). Linear Regression and Problem Framing Linear regression is the simplest and most commonly taught supervised learning algorithm for regression problems, where the goal is to predict a continuous number from input features.The episode example focuses on predicting the cost of houses in Portland, using square footage and possibly other features as inputs. The Three Steps of Machine Learning in Linear Regression Machine learning in the context of linear regression follows a standard three-step loop: make a prediction, measure how far off the prediction is, and update the prediction method to reduce mistakes.Predicting uses a hypothesis function (also called objective or estimate) that maps input features to a predicted value. The Hypothesis Function The hypothesis function is a formula that multiplies input features by coefficients (weights) and sums them to make a prediction; in mathematical terms, for one feature, it is: h(x) = theta_1 * x_1 + theta_0 Here, theta_1 is the weight for the feature (e.g., square footage), and theta_0 is the bias (an average baseline). With only one feature, the model tries to fit a straight line to a scatterplot of the input feature versus the actual target value. Bias and Multiple Features The bias term acts as the starting value when all features are zero, representing an average baseline cost.In practice, using only one feature limits accuracy; including more features (like number of bedrooms, bathrooms, location) results in multivariate linear regression: h(x) = theta_0 + theta_1 * x_1 + theta_2 * x_2 + ... for each feature x_n. Visualization and Model Fitting Visualizing the problem involves plotting data points in a scatterplot: feature values on the x-axis, actual prices on the y-axis.The goal is to find the line (in the univariate case) that best fits the data, ideally passing through the "center" of the data cloud. The Cost Function (Mean Squared Error) The cost function, or mean squared error (MSE), measures model performance by averaging squared differences between predictions and actual labels across all training examples.Squaring ensures positive and negative errors do not cancel each other, and dividing by twice the number of examples (2m) simplifies the calculus in the next step. Parameter Learning via Gradient Descent Gradient descent is an iterative algorithm that uses calculus (specifically derivatives) to find the best values for the coefficients (thetas) by minimizing the cost function.The cost function’s surface can be imagined as a bowl in three dimensions, where each point represents a set of parameter values and the height represents the error.The algorithm computes the slope at the current set of parameters and takes a proportional step (controlled by the learning rate alpha) toward the direction of the steepest decrease.This process is repeated until reaching the lowest point in the bowl, where error is minimized and the model best fits the data.Training will not produce a perfect zero error in practice, but it will yield the lowest achievable average error for the data given. Extension to Multiple Variables Multivariate linear regression extends all concepts above to datasets with multiple input features, with the same process for making predictions, measuring error, and performing gradient descent.Technical details are essentially the same though visualization becomes complex as the number of features grows. Essential Learning Resources The episode strongly directs listeners to the Andrew Ng course on Coursera as the primary recommended starting point for studying machine learning and gaining practical experience with linear regression and related concepts.
    続きを読む 一部表示
    34 分
  • MLG 006 Certificates & Degrees
    2017/02/17

    Try a walking desk to stay healthy while you study or work!

    Full notes at ocdevel.com/mlg/6

    Pursuing Machine Learning:
    • Individuals may engage with machine learning for self-education, as a hobby, or to enter the industry professionally.
    • Use a combination of resources, including podcasts, online courses, and textbooks, for a comprehensive self-learning plan.
    Online Courses (MOOCs):
    • MOOCs, or Massive Open Online Courses, offer accessible education.
    • Key platforms: Coursera and Udacity. Coursera is noted for standalone courses; Udacity offers structured nanodegrees.
    • Udacity nanodegrees include video content, mentoring, projects, and peer interaction, priced at $200/month.
    Industry Recognition:
    • Udacity nanodegrees are currently not widely recognized or respected by employers.
    • Emphasize building a robust portfolio of independent projects to augment qualifications in the field.
    Advanced Degrees:
    • Master’s Degrees:
    • Valued by employers, provide an edge in job applications.
    • Example: Georgia Tech's OMSCS (Online Master’s of Science in Computer Science) offers a cost-effective ($7,000) online master’s program.
    • PhD Programs:
    • Embark on a PhD for in-depth research in AI rather than industry entry. Program usually pays around $30,000/year.
    • Compare industry roles (higher pay, practical applications) vs. academic research (lower pay, exploration of fundamental questions).
    Career Path Decisions:
    • Prioritize building a substantial portfolio of projects to bypass formal degree requirements and break into industry positions.
    • Consider enriching your qualifications with a master's degree, or eventually pursue a PhD if deeply interested in pioneering AI research.
    Discussion and Further Reading:
    • See online discussions about degrees/certifications: 1 2 3 4
    続きを読む 一部表示
    16 分
  • MLG 007 Logistic Regression
    2017/02/19

    Try a walking desk to stay healthy while you study or work!

    Full notes at ocdevel.com/mlg/7. See Andrew Ng Week 3 Lecture Notes

    Overview
    • Logistic Function: A sigmoid function transforming linear regression output to logits, providing a probability between 0 and 1.
    • Binary Classification: Logistic regression deals with binary outcomes, determining either 0 or 1 based on a threshold (e.g., 0.5).
    • Error Function: Uses log likelihood to measure the accuracy of predictions in logistic regression.
    • Gradient Descent: Optimizes the model by adjusting weights to minimize the error function.
    Classification vs Regression
    • Classification: Predicts a discrete label (e.g., a cat or dog).
    • Regression: Predicts a continuous outcome (e.g., house price).
    Practical Example
    • Train on a dataset of house features to predict if a house is 'expensive' based on labeled data.
    • Automatically categorize into 0 (not expensive) or 1 (expensive) through training and gradient descent.
    Logistic Regression in Machine Learning
    • Neurons in Neural Networks: Act as building blocks, as logistic regression is used to create neurons for more complex models like neural networks.
    • Composable Functions: Demonstrates the compositional nature of machine learning algorithms where functions are built on other functions (e.g., logistic built on linear).
    続きを読む 一部表示
    35 分
  • MLG 008 Math
    2017/02/23

    Try a walking desk to stay healthy while you study or work!

    Full notes at ocdevel.com/mlg/8

    Mathematics in Machine Learning
    • Linear Algebra: Essential for matrix operations; analogous to chopping vegetables in cooking. Every step of ML processes utilizes linear algebra.
    • Statistics: The hardest part, akin to the cookbook; supplies algorithms for prediction and error functions.
    • Calculus: Used in the learning phase (gradient descent), similar to baking; it determines the necessary adjustments via optimization.
    Learning Approach
    • Recommendation: Learn the basics of machine learning first, then dive into necessary mathematical concepts to prevent burnout and improve appreciation.
    Mathematical Resources
    • MOOCs: Khan Academy - Offers Calculus, Statistics, and Linear Algebra courses.
    • Textbooks: Commonly recommended books for learning calculus, statistics, and linear algebra.
    • Primers: Short PDFs covering essential concepts.
    Additional Resource
    • The Great Courses: Offers comprehensive video series on calculus and statistics. Best used as audio for supplementing primary learning. Look out for "Mathematical Decision Making."
    Python and Linear Algebra
    • Tensor: General term for any dimension list; TensorFlow from Google utilizes tensors for operations.
    • Efficient computation using SimD (Single Instruction, Multiple Data) for vectorized operations.
    Optimization in Machine Learning
    • Gradient descent used for minimizing loss function, known as convex optimization. Recognize keywords like optimization in calculus context.
    続きを読む 一部表示
    28 分