Untitled
44 Questions
2 Views

Choose a study mode

Play Quiz
Study Flashcards
Spaced Repetition
Chat to Lesson

Podcast

Play an AI-generated podcast conversation about this lesson

Questions and Answers

Which of the following developments exposed a significant limitation in early neural network models?

  • The demonstration of Backpropagation by Rumelhart et al.
  • The emergence of Support Vector Machines.
  • Minsky et al.'s proof regarding the limitations of the Perceptron. (correct)
  • The creation of Arthur Samuel’s checkers program.

Deductive learning involves algorithms that learn from provided examples to map inputs to outputs, similar to supervised learning.

False (B)

In the context of machine learning, what is the primary role of 'training data'?

To train the model

The 'AI Winter' of the 1970s is closely associated with the organization known as ________.

<p>AAAI</p> Signup and view all the answers

Match each machine learning era with a key development:

<p>1950s = Arthur Samuel's Checkers Program 1980s = Backpropagation 2010s = Deep Learning 2020s = Generative AI</p> Signup and view all the answers

Which of the following statements best describes the primary goal of sentiment analysis in NLP?

<p>To determine whether data expresses positive, negative, or neutral opinions. (C)</p> Signup and view all the answers

Machine translation involves understanding the nuances of the original text by translating words directly without considering context.

<p>False (B)</p> Signup and view all the answers

How does text summarization benefit users in applications like news headlines and search engine result snippets?

<p>It provides a concise version of the original text, preserving the meaning while saving time for readers.</p> Signup and view all the answers

The field of artificial intelligence that enables computers to interpret and understand images is known as ______.

<p>computer vision</p> Signup and view all the answers

Match the following NLP techniques with their primary application:

<p>Sentiment Analysis = Understanding customer needs and brand monitoring Text Summarization = Condensing text for news headlines Machine Translation = Converting text from one language to another</p> Signup and view all the answers

Which of the following is the most complex task that computer vision can perform, building upon object detection?

<p>Estimating key points in addition to detecting and categorizing objects. (D)</p> Signup and view all the answers

Computer vision in medical imaging is primarily limited to diagnostic applications and does not extend to surgical assistance or research.

<p>False (B)</p> Signup and view all the answers

Beyond navigation, in what crucial capacity does computer vision assist self-driving cars?

<p>obstacle avoidance</p> Signup and view all the answers

In agriculture, computer vision is employed to detect product defects, sort produce, and identify areas of low ______.

<p>water</p> Signup and view all the answers

Match each application of computer vision with its primary function:

<p>Facial recognition = Identifying individuals Self-driving cars = Navigating and avoiding obstacles Robotic automation = Performing tasks based on visual input Medical imaging = Diagnostic applications</p> Signup and view all the answers

What is the foundational difference between object recognition and object detection in computer vision?

<p>Object recognition only identifies objects, while object detection also locates them. (C)</p> Signup and view all the answers

A computer vision system solely relies on pre-programmed rules without the capacity to adapt based on new visual inputs.

<p>False (B)</p> Signup and view all the answers

Besides identifying diseases, what specific diagnostic capability does computer vision bring to medical imaging?

<p>trend identification</p> Signup and view all the answers

The use of computer vision allows robots to make ______ based on visual input, expanding their functionality in automated systems.

<p>decisions</p> Signup and view all the answers

Which capability of computer vision is most directly applicable to enhancing security by monitoring property perimeters?

<p>Person detection (A)</p> Signup and view all the answers

Retinaface is primarily used for what combination of tasks?

<p>Object detection and key points estimation (C)</p> Signup and view all the answers

The website 'https://this-person-does-not-exist.com/en' generates images of real people.

<p>False (B)</p> Signup and view all the answers

What type of data is represented by categories that have a meaningful order or ranking?

<p>Ordinal Data</p> Signup and view all the answers

Data that represents categories without any inherent order or ranking is known as ______ data.

<p>Nominal</p> Signup and view all the answers

Which model is specifically mentioned for image-to-sketch generation?

<p>sketchmypic.com (B)</p> Signup and view all the answers

For what purpose is the pytorch-CycleGAN-and-pix2pix model primarily utilized?

<p>Object generation (A)</p> Signup and view all the answers

Data used for training a machine learning model cannot include image or video formats.

<p>False (B)</p> Signup and view all the answers

Match the data type with the appropriate example.

<p>Nominal Data = Colors (Red, Blue, Green) Ordinal Data = Education Level (High School, Bachelor's, Master's)</p> Signup and view all the answers

Elaine Rich defines AI as the study of techniques for solving exponentially hard problems in polynomial time. Which aspect of this definition is most critical in distinguishing AI from traditional algorithm design?

<p>The exploitation of knowledge about the specific problem domain. (C)</p> Signup and view all the answers

The Chinese Room Argument posits that a system passing the Turing Test necessarily possesses genuine understanding or consciousness.

<p>False (B)</p> Signup and view all the answers

Explain how Herbert Simon's definition of AI relates to the concept of the Turing Test.

<p>Simon's definition focuses on creating programs that exhibit behaviors considered intelligent in humans, while the Turing Test assesses a machine's ability to exhibit intelligent behavior indistinguishable from that of a human.</p> Signup and view all the answers

The ability of ________ is a key differentiator between humans and machines, as humans can learn and apply knowledge and skills.

<p>humans</p> Signup and view all the answers

Match the following concepts with their descriptions:

<p>Turing Test = A test of a machine's ability to exhibit intelligent behavior equivalent to, or indistinguishable from, that of a human. Chinese Room Argument = A thought experiment questioning whether a machine can truly 'understand' by merely manipulating symbols according to rules. Artificial Intelligence = The study of techniques for solving exponentially hard problems in polynomial time by exploiting knowledge about the problem domain. Human Intelligence = The ability to learn and apply knowledge and skills.</p> Signup and view all the answers

In the context of the Chinese Room Argument, what is the central question regarding 'understanding'?

<p>Does the agent know the semantics of the symbols it's manipulating? (B)</p> Signup and view all the answers

Which of the following scenarios best describes the application of AI as defined by Elaine Rich?

<p>Creating a chess-playing program that uses heuristics to explore only promising moves, rather than all possible moves. (D)</p> Signup and view all the answers

Contrast the objective of the Turing Test with the focus of the Chinese Room Argument in evaluating machine intelligence.

<p>The Turing Test evaluates a machine's ability to imitate human-like responses, whereas the The Chinese Room Argument questions whether this imitation constitutes genuine understanding or consciousness.</p> Signup and view all the answers

Which of the following scenarios would most likely require data scaling to ensure fair feature contribution in a predictive model?

<p>Estimating credit risk based on 'number of credit cards held' (ranging from 1 to 4) and 'annual income' (ranging from $20,000 to $200,000). (D)</p> Signup and view all the answers

Deleting an entire column with missing values is always the most appropriate strategy for handling missing data.

<p>False (B)</p> Signup and view all the answers

What is the main reason why data scaling is important when features have significantly different ranges?

<p>To prevent features with larger values from dominating the model.</p> Signup and view all the answers

__________ data can take any value within a given range, often involving decimals.

<p>Continuous</p> Signup and view all the answers

Match the data preprocessing steps with their primary purpose:

<p>Data Scaling = Transforms feature values to a similar scale. Handling Missing Data = Addresses blank observations in a dataset. Normalization = Transforms features to be on a similar scale, typically between 0 and 1. Standardization = Scales data to have a mean of 0 and a standard deviation of 1.</p> Signup and view all the answers

Which of the following is NOT a typical reason for the necessity of data preprocessing?

<p>Data preprocessing guarantees perfect model accuracy. (C)</p> Signup and view all the answers

In what scenario would imputing missing values with a central tendency measure (mean, median, or mode) be least appropriate?

<p>When the missing values are due to a systematic error and concentrated in a specific subgroup. (B)</p> Signup and view all the answers

What distinguishes normalization from standardization in feature scaling?

<p>Normalization scales values to a specific range (e.g., 0 to 1), while standardization scales data to have a mean of 0 and a standard deviation of 1. (D)</p> Signup and view all the answers

Flashcards

What is Natural Language Processing (NLP)?

A field of AI enabling computers to understand and process human language.

What is Sentiment Analysis?

Determines the emotional tone (positive, negative, or neutral) of text data.

What is Text Summarization?

Condenses a longer text into a shorter version while keeping the main meaning.

What is Machine Translation?

Automatically converting text from one language to another.

Signup and view all the flashcards

What is Computer Vision?

An AI field that enables computers to see, understand, and interpret images.

Signup and view all the flashcards

Machine Learning (ML)

A sub-field of AI where computers learn from data and experience to make predictions.

Signup and view all the flashcards

Training Data

The phase where a machine learning model learns patterns from a labeled dataset.

Signup and view all the flashcards

Testing Data

The phase where a trained model's performance is evaluated on new, unseen data.

Signup and view all the flashcards

Deductive Learning

Learning where rules are explicitly provided; applying these rules yields the answer.

Signup and view all the flashcards

Inductive Learning

Learning where the system learns a mapping from inputs to outputs based on given examples.

Signup and view all the flashcards

Retinaface

Detects faces and predicts facial landmarks.

Signup and view all the flashcards

Object Generation (People)

Generates realistic images of people.

Signup and view all the flashcards

Image to Sketch Generation

Converts images into sketches.

Signup and view all the flashcards

Data

Raw facts and figures used to train and test ML models; can be text, image, audio or video.

Signup and view all the flashcards

Nominal Data

Represents categories without inherent order.

Signup and view all the flashcards

Ordinal Data

Represents categories with a meaningful order or ranking.

Signup and view all the flashcards

Artificial Intelligence (AI)

The concept of imbuing machines with the ability to perform tasks that typically require human intelligence.

Signup and view all the flashcards

AI vs Human Intelligence

Humans can learn so AI tries to recreate this in machines.

Signup and view all the flashcards

Turing Test

A test of a machine's ability to exhibit intelligent behavior equivalent to, or indistinguishable from, that of a human.

Signup and view all the flashcards

Chinese Room Argument

A thought experiment exploring whether a computer can truly 'understand' a language by processing it based on syntactic rules alone.

Signup and view all the flashcards

Natural Language Processing (NLP)

The study of enabling computers to understand and process human language.

Signup and view all the flashcards

Computer Vision (CV)

The field of enabling computers to 'see' and interpret images.

Signup and view all the flashcards

Data Preprocessing

Preparing raw data to make it suitable for machine learning models.

Signup and view all the flashcards

Exponentially Hard Problems

Problems that are very difficult to solve, needing a long computation using brute-force methods.

Signup and view all the flashcards

Facial Recognition

Identifying individuals through analyzing visual inputs using digital images and deep learning.

Signup and view all the flashcards

Self-driving cars

Navigating and avoiding obstacles using computer vision.

Signup and view all the flashcards

Robotic automation

Performing tasks and making decisions based on visual input.

Signup and view all the flashcards

Person detection

Monitoring perimeters intelligently using person detection.

Signup and view all the flashcards

Medical Imaging

Using computer vision for various processes like diagnostics, screening, surgery, and research.

Signup and view all the flashcards

Agriculture applications

Identifying defects, sorting produce, and detecting diseases using computer vision.

Signup and view all the flashcards

Computer Vision

A field of AI that enables computers to 'see' and interpret images.

Signup and view all the flashcards

Object Recognition

Mapping an image to a category.

Signup and view all the flashcards

Object Detection

Mapping an image to a category and localizing objects.

Signup and view all the flashcards

Object detection + key points estimation

Mapping an image to a category, localizing objects, and estimating key points on the objects.

Signup and view all the flashcards

Discrete Data

Data that can only take specific, distinct values (e.g., number of cars).

Signup and view all the flashcards

Continuous Data

Data that can take any value within a range (e.g., height, temperature).

Signup and view all the flashcards

Why Preprocessing?

Raw data often contains errors, missing information, and inconsistencies.

Signup and view all the flashcards

Missing Data

Instances where some values are absent in a dataset.

Signup and view all the flashcards

Deleting Missing Values

Removing rows or columns containing missing values.

Signup and view all the flashcards

Imputing Missing Values

Replacing missing values with substitute values.

Signup and view all the flashcards

Data Scaling

Transforming feature values to a similar scale.

Signup and view all the flashcards

Normalization

Transforming features to a range between 0 and 1.

Signup and view all the flashcards

Study Notes

Overview of Machine Learning & Its Applications and Data Preprocessing

  • Artificial intelligence and machine learning history will be discussed
  • Definition and types of machine learning will be examined: Supervised, Unsupervised and Reinforcement Learning
  • Applications of machine learning, like Natural Language Processing (NLP) and Computer Vision (CV) will be presented
  • Data Preprocessing will be covered

Artificial Intelligence

  • The term "artificial intelligence" was first coined in 1956 at the Dartmouth Conference
  • Homo sapiens can control other species due to their thinking ability
  • Programs exhibiting behaviors considered intelligent when shown by humans are called intelligent - according to Herbert Simon
  • AI involves solving exponentially hard problems in polynomial time by leveraging domain knowledge - according to Elaine Rich

Human Intelligence vs Artificial Intelligence

  • Humans can learn and apply knowledge and skills, while AI recreates human intelligence in machines

Classical Problems in AI

  • Turing Test is a test of a machine's ability to exhibit intelligent behavior equivalent to, or indistinguishable from, that of a humans.
  • Alan Turing questioned if machines can think; named it "too meaningless"
  • Turing's "Imitation Game" is now known as the Turing Test (1912–1954)
  • The Loebner Prize is an annual competition where chatbots responding like humans are judged; The grand price of USD 100,000 is still open.
  • A question explored in "The Chinese Room Argument" is Can an agent locked in a room, processing Chinese questions based on syntactic rules, be said to understand Chinese?

Topics in AI

  • AI topics include: natural language understanding, speech recognition/synthesis, image processing, qualitative reasoning, computer vision, knowledge representation, neural networks, machine learning, knowledge discovery, graphics, logic, search, models, planning, problem solving, tactile sensing, adversarial reasoning and robot control

Machine Learning

  • Machine Learning is a subset of AI
  • ML enables computers to learn from data and past experiences to make predictions

Machine Learning History

  • 1950s: Arthur Samuel created the Checker's program
  • 1960s: Neural Network and Rosenblatt's Perceptron came about; limitations of Perceptron were proven by Minsky et al.
  • 1970s: AI Winter began, which was also known as AAAI (Association for the Advancement of Artificial Intelligence)
  • 1980s: Backpropagation came to be via David Rumelhart et al.
  • 1990s: Support Vector Machines were developed
  • 2010s: Deep Learning was introduced
  • 2020s: Generative AI developed

Dataset

  • Datasets include Labels, Independent and Dependent Variables, such as age(years), weight(kg), systolic BP(mm Hg) of subjects

Types of Learning Algorithms

  • Deductive Learning involves hard-coded rules being provided. An example includes multiplying numbers; you can apply this to any two numbers and find the answer
  • Inductive Learning involves examples being provided for you, with needs to learn the mapping from input to output. Supervised learning is an example.

Types of Machine Learning Algorithms

  • Supervised Learning Classification answers the question "is this A or B?"
  • Supervised Learning Regression answers the question "how much or how many?"
  • Unsupervised Learning Clustering answers the question "how is this organized?"
  • Reinforcement Learning answers the question "what should I do next.

Examples of Regression

  • Supervised Learning Algorithms can predict a continuous variable's value
  • Examples include employee salary as a function of experience, and price of a bike as a function of age
  • A Machine Learning model example includes y = f(x) also can be written as Salary = f(experience) or f = w₀ + w₁x where f is dependent output, x is independent input variable(s), w₀ is y-axis intercept, and w₁ is coefficient

Examples of Classification

  • Predict a binary (discrete) variable's value using Supervised Learning Algorithms
  • Examples include if a given image is outdoor or indoor, if a personal loan application can be approved or not, and if a given cancer cell is begin or malignant.

Examples of Clustering

  • The Un-supervised Learning Algorithm is used when there is no variable to predict; only the common pattern to group. Examples include grouping different customers based on income and spending.

Examples of Association Rule Learning

  • Un-Supervised Learning Algorithms such as Recommender Systems are used to recognize customers with similar shopping patterns, recommend products having high probability of buying

Examples of Reinforcement Learning

  • Learning is through reward and penalty when using this Algorithm

Spectrum of Supervision

  • The Spectrum of supervision increases with "semi" supervised and fully supervised data

Applications of Machine Learning: NLP

  • Natural Language Processing (NLP) is a branch of computer science/AI that allows computers to understand text and spoken words. Computer science, linguistics, and machine learning are combined to study how computers and humans communicate.
  • Sentiment analysis determines if data is positive, negative or neutral
  • Text summarization condenses text to a shorter version while keeping the content meaning. News headlines and result snippets in web searches are some uses.
  • Machine translation automatically translates text from one language to another

Sentiment Analysis

  • Sentiment analysis determines whether data is positive, negative, or neutral in business to understand customer needs and monitor brand/product sentiment

Language Translation

  • Automatically translate from one language to another, using NLP techniques to understand the structure and meaning of the original text and creating a translation conveying the meaning in the target language

Applications of Machine Learning: Computer Vision

  • Computer vision enables computers to understand and interprets the visual world. It utilizes digital images and deep learning models to identify and classify objects, and then react to what it "sees":
  • Facial Recognition: Identifying individuals through visual analysis
  • Self-driving cars use computer vision to navigate and avoid obstacles
  • Robotic Automation enables decisions based on visual input so that the robot can perform tasks.
  • Person detection is useful for intelligent perimeter monitoring
  • Medical Imaging uses computer vision for diagnostic applications, cancer treatments, surgery and research

Applications of Machine Learning: Computer Vision

  • Includes object recognition and object detection

Object Recognition

  • Object Recognition maps any given sample to one of the categories, such as indoor/kitchen and outdoor/beach.

Data Preprocessing

  • Data is a set of observations or measurements used to train and test a ML model.
  • Data can be any unprocessed value, text, speech, image, or video
  • Types of data: normal, ordinal, discrete and continuous data

Data and Types of Data

  • Nominal Data (Categorical): Represents categories without inherent order/ranking, ex: Gender & Colors.
  • Ordinal Data: Represents categories with a meaningful order/ranking, ex: Education level (High School < Bachelor's < Master's < Ph.D.)
  • Discrete Data: Numerical data that can take on specific, distinct values, ex: number of cars in a parking lot.
  • Continuous Data: Numerical data that can take any value within a given range, often involving decimals, ex: Weight, Temperature and Time.

Why is Data Preprocessing Required?

  • Raw data quality is poor
  • Real-world datasets are noisy, missing and/or inconsistent

Handling Missing Data

  • Missing data, also known as missing values, is where some of the observation in a data set are blank
  • Delete the missing values by deleting the entire row or column
  • Impute the missing values by replacing with an arbitrary value like 0, or a central tendency measure for that column

Scaling and Data Normalization

  • Data scaling is a preprocessing technique used to transform the values of features with differing scales in a dataset to a similar scale
  • The goal of data scaling is so that all features contribute equally to the mode, reducing outlier impact, and enabling a faster model convergence
  • Techniques include normalization and standardization

Importance of Feature Scaling

  • Number of bedrooms (1 to 10) and house size (500 to 5000) must be scaled for their values to contribute equally to the model when predicting house prices

Normalization

  • A goal is to transform features to be on a similar scale
  • Min-max normalization: transform feature values from their natural range (example, 100 to 900) into a standard range, which is typically [0, 1] (or sometimes [-1,1])
  • To rescale the feature values between arbitrary values [a,b]:
Xscaled = a + (X - Xmin) (b-a) /  Xmax - Xmin

where min-max normalization is suggested when the feature is uniformly distributed across a fixed range

Standardization

  • A Feature is standardized if its values have zero-mean and unit-variance
  • This is known as Z-Score normalization method where the values are centered around a zero mean with a unit standard deviation.
  • Where x1 = x - μ / σ, μ is represents the mean of the feature values and o represents the standard deviation

Normalization vs. Standardization

  • Normalization rescales value to a range between 0 and 1 and is sensitive to outlier
  • Standardization centers data around a zero mean, scales it to a standard deviation of 1 and has less sensitivity to outliers

Dealing with Outliers

  • This is another data preprocessing step in which outliers need to be detected, then handled. Training and accuracy of a ML model can decrease because of them.

Detecting and Handling Outliers

  • Ways to detect outliers is through visualization/box-plot, using standard deviation or Z-scores
  • Ways to handle outliers: remove the, use quantile-based capping and flooring, or Mean/Median imputation

Encoding Categorical Values

  • Encoding methods for Ordinal and Nominal values are required for ML Models
  • Some of the encoding methods can be used: Label, One-Hot, Effect, Binary, Base N, and Target encoding

Encoding Categorical Variables: Label Encoding

  • Each label is converted into an integer value.
  • If we apply label encoding to 'safety' feature, ['low', 'med', 'high'] will be encoded to [1, 2, 0].
  • Similarly for 'lug_boot' feature, ['small', 'med', 'big'] will be encoded to [2, 1, 0]

Encoding Categorical Variables: One-Hot Encoding

  • Each category value is mapped with a binary value of either 0 or 1.

Terminology

  • A feature vector is a numerical representation of an object
  • Class to which X belongs to is y ∈ Y, and needs to be estimated, based on the training set
  • Task is to design a classifer or decision rule of f: X → Y which decides about the class label based on X.

Features and Samples: An Example

  • Feature Vector: A vector of observations
  • Class to which X belongs: overweight, normal
  • Training Set: To need to estimate, based on training set
  • Task: to design a classifer (decision rule) in order to classify the person, classify him/her. f: X → Y
  • Sample: Each X is considered to be as a sample.

Training, Validation, and Test sets

  • Training set: examples are used to fit/learn the parameters of a model; can also be used to build a perceptron classifier
  • Validation set: It is a development set to used to tune the hyperparameters and architecture of a model.
  • Test set: This set is a learned machine learning model and its performance is assessed.

Feature Engineering

  • Raw data is transformed into features to create a predictive model using machine learning.
  • Feature Transformation transforms the features into a more suitable representation for the machine learning model.
  • Feature Selection selects a subset of relevant features used for machine-learning model training, where some info is removed
  • Example: For a dataset having a model to classify person overweight or normal; if this is the case, then the name is removed

Studying That Suits You

Use AI to generate personalized quizzes and flashcards to suit your learning preferences.

Quiz Team

More Like This

Untitled
110 questions

Untitled

ComfortingAquamarine avatar
ComfortingAquamarine
Untitled Quiz
6 questions

Untitled Quiz

AdoredHealing avatar
AdoredHealing
Untitled Quiz
18 questions

Untitled Quiz

RighteousIguana avatar
RighteousIguana
Untitled Quiz
50 questions

Untitled Quiz

JoyousSulfur avatar
JoyousSulfur
Use Quizgecko on...
Browser
Browser