Machine Learning Foundations - Unit 1
48 Questions
0 Views

Choose a study mode

Play Quiz
Study Flashcards
Spaced Repetition
Chat to Lesson

Podcast

Play an AI-generated podcast conversation about this lesson

Questions and Answers

What effect does the choice of kernel function have in kernel density estimation?

  • The variance of the data
  • The number of bins used in the histogram
  • The smoothness of the estimated density (correct)
  • The mean of the distribution

Which kernel function is commonly used in kernel density estimation?

  • Exponential kernel
  • Poisson kernel
  • Gaussian kernel (correct)
  • Binomial kernel

In k-nearest neighbor density estimation, what does the parameter 'k' represent?

  • The width of the kernel function
  • The number of bins in the histogram
  • The number of nearest neighbors considered for each point (correct)
  • The number of data points used for density estimation

What is a key advantage of k-nearest neighbor density estimation?

<p>It is simple and non-parametric (C)</p> Signup and view all the answers

Which nonparametric method is best suited for handling large datasets with unknown distribution shapes?

<p>Kernel density estimator (A)</p> Signup and view all the answers

What is the primary difference between histogram estimators and kernel estimators?

<p>Histograms use bins while kernels use points (B)</p> Signup and view all the answers

What limitation is commonly associated with kernel density estimation techniques?

<p>They require large sample sizes to be accurate (D)</p> Signup and view all the answers

What does a histogram estimator primarily rely on for its structure?

<p>The number of bins and their width (B)</p> Signup and view all the answers

Which clustering algorithm can handle clusters of varying shapes and sizes?

<p>DBSCAN (A)</p> Signup and view all the answers

Which clustering algorithm does not require the assumption of equal-sized clusters?

<p>DBSCAN (C)</p> Signup and view all the answers

Which clustering algorithm is based on the concept of nearest neighbors?

<p>K-Nearest Neighbors (C)</p> Signup and view all the answers

Which assumption does the Naïve Bayes classifier make about features?

<p>Features are independent given the class label. (D)</p> Signup and view all the answers

Which probability is calculated in the Naïve Bayes algorithm to classify a new data point?

<p>Posterior probability (C)</p> Signup and view all the answers

What is the key equation used in Bayes' Theorem?

<p>P(A|B) = P(B|A) * P(A) / P(B) (C)</p> Signup and view all the answers

In a Naïve Bayes classifier, which class is chosen as the predicted class?

<p>Class with the highest posterior probability (D)</p> Signup and view all the answers

What is the main purpose of the 'kernel trick' in SVM?

<p>To transform data into a higher-dimensional space (B)</p> Signup and view all the answers

Which kernel function is commonly used in Support Vector Machines (SVM) for non-linearly separable data?

<p>Radial Basis Function (RBF) kernel (B)</p> Signup and view all the answers

Which of the following is NOT a commonly used kernel in SVM?

<p>Logistic (B)</p> Signup and view all the answers

In Bayes' Theorem, what does the term $P(B)$ represent?

<p>Marginal probability (C)</p> Signup and view all the answers

Which statement about Naïve Bayes classifiers is accurate?

<p>It is robust to noise and irrelevant features. (A)</p> Signup and view all the answers

What is a 'support vector' in the context of SVM?

<p>A data point that is closest to the decision boundary (C)</p> Signup and view all the answers

Which activation function is most commonly used in the output layer of a binary classification neural network?

<p>Sigmoid (C)</p> Signup and view all the answers

What is the primary role of an activation function in a neural network?

<p>To introduce non-linearity into the model (D)</p> Signup and view all the answers

What is a Perceptron in the context of machine learning?

<p>The simplest form of a neural network (A)</p> Signup and view all the answers

What does the Simple Matching Coefficient measure?

<p>The proportion of matching attributes in binary data (C)</p> Signup and view all the answers

Which metric is used to calculate the correlation between two attributes?

<p>Pearson Correlation Coefficient (C)</p> Signup and view all the answers

What does the Cosine Similarity measure?

<p>The angle between two vectors (A)</p> Signup and view all the answers

Which of the following measures similarity between binary vectors?

<p>Simple Matching Coefficient (B)</p> Signup and view all the answers

What is a key advantage of using Euclidean Distance?

<p>It is easy to compute and interpret in a continuous space (B)</p> Signup and view all the answers

In what type of data is the Cosine Similarity particularly useful?

<p>Text data (C)</p> Signup and view all the answers

What does a decision tree model do?

<p>Divides data into branches to make predictions based on feature values (B)</p> Signup and view all the answers

Which algorithm is commonly used to create a decision tree?

<p>ID3 (B)</p> Signup and view all the answers

What is a primary difference between histograms and kernel density estimators?

<p>Kernels are more sensitive to bin width than histograms. (A)</p> Signup and view all the answers

How does the choice of bandwidth affect kernel density estimation?

<p>It influences the smoothness versus the bias-variance tradeoff. (D)</p> Signup and view all the answers

Which statement accurately describes nonparametric methods?

<p>They do not rely on assumptions about the distribution's form. (C)</p> Signup and view all the answers

In nonparametric density estimation, what does 'smoothing' signify?

<p>Adjusting the bandwidth to control the smoothness of the density estimate. (B)</p> Signup and view all the answers

What is commonly observed when increasing the number of bins in a histogram?

<p>It produces a more jagged density estimate. (D)</p> Signup and view all the answers

Which metric is generally the most useful when handling an imbalanced dataset?

<p>Precision (B)</p> Signup and view all the answers

What effect does kernel smoothing have compared to histograms in density estimation?

<p>It generally produces a more continuous estimate. (C)</p> Signup and view all the answers

When using histograms, what happens if the bin size is too large?

<p>Details of the data distribution are lost. (A)</p> Signup and view all the answers

How does the k-Means algorithm initialize cluster centroids?

<p>Randomly (C)</p> Signup and view all the answers

What is the role of the ‘k’ parameter in the k-Means algorithm?

<p>Number of clusters to be formed (B)</p> Signup and view all the answers

How does the k-Means algorithm update cluster centroids during each iteration?

<p>By calculating the mean of all data points in each cluster (D)</p> Signup and view all the answers

What is a major limitation of the k-Means algorithm?

<p>It is sensitive to initial centroid positions (A)</p> Signup and view all the answers

How does the k-Means algorithm determine convergence?

<p>When the centroids stop moving significantly between iterations (A)</p> Signup and view all the answers

Which distance metric is commonly used in the k-Means algorithm?

<p>Euclidean distance (A)</p> Signup and view all the answers

What is the computational complexity of the k-Means algorithm?

<p>O(n*k) (A)</p> Signup and view all the answers

Which of the following methods can help improve the performance of the k-Means algorithm?

<p>Scaling the data to have equal variance (A), Initializing centroids close to the mean of the data (B)</p> Signup and view all the answers

Flashcards

Simple Matching Coefficient

Measures the proportion of matching attributes in binary data. It focuses on the number of shared features between two data points.

Pearson Correlation Coefficient

Used to calculate the linear correlation between two attributes. It measures the strength and direction of their linear relationship.

Cosine Similarity

Measures the angle between two vectors. It indicates how similar the directions of two objects are, regardless of their magnitude.

Euclidean Distance

Calculates the straight-line distance between two points in a multi-dimensional space. It's a common way to measure dissimilarity between data points.

Signup and view all the flashcards

Decision Tree Model

Divides data into branches based on feature values to create an easily interpretable model for making predictions.

Signup and view all the flashcards

ID3 Algorithm

A common algorithm used to create decision trees by iteratively splitting the data based on the feature that offers the most information gain.

Signup and view all the flashcards

Rule-Based Classifier

Makes decisions based on a set of predefined rules that have been learned from the data. Each rule has a condition and a corresponding action.

Signup and view all the flashcards

Minkowski Distance

A generalized form of distance calculation that encompasses Euclidean and Manhattan distances. It calculates a power-based distance between points.

Signup and view all the flashcards

K-Means Initialization

The k-Means algorithm starts by placing the initial cluster centroids. There are different methods, but the standard one is randomly selecting k data points as centroids.

Signup and view all the flashcards

Role of 'k' in K-Means

'k' in k-Means represents the number of clusters you want to create. It's the predetermined number of groups the algorithm will form.

Signup and view all the flashcards

K-Means Centroid Update

During each iteration, the k-Means algorithm updates the positions of the centroids. It calculates the mean of all data points assigned to each cluster and moves the centroid to that mean.

Signup and view all the flashcards

K-Means Limitation: Initial Centroids Impact

A big issue with k-Means is its sensitivity to the initial placement of centroids. Bad starting positions can lead to suboptimal and inaccurate clusters.

Signup and view all the flashcards

K-Means Convergence

The k-Means algorithm converges when the centroids no longer shift significantly between iterations. This means the data points have stabilized in their assigned clusters.

Signup and view all the flashcards

K-Means Distance Metric

The k-Means algorithm usually employs the Euclidean distance as a metric to calculate the distance between data points and centroids.

Signup and view all the flashcards

K-Means Computational Complexity

The k-Means algorithm has a computational complexity of O(n*k). This means the runtime increases linearly with the number of data points (n) and clusters (k).

Signup and view all the flashcards

K-Means Advantage: Efficiency

One of the key strengths of k-Means is its computational efficiency. It's relatively fast compared to other clustering techniques, especially for large datasets.

Signup and view all the flashcards

Kernel Trick in SVM

A technique used in SVM to transform data into a higher-dimensional space. This allows for the separation of data that cannot be linearly separated in the original space.

Signup and view all the flashcards

Common SVM Kernels

Different kernel functions used in SVM to achieve different levels of complexity and flexibility in the decision boundary. Linear, Polynomial, and Gaussian are common examples.

Signup and view all the flashcards

Logistic Kernel

A kernel function NOT commonly used in SVMs because it's not designed for classifying data.

Signup and view all the flashcards

P(B) in Bayes' Theorem

Represents the marginal probability of event B occurring. It's the probability of observing B without considering any relationship to A.

Signup and view all the flashcards

Naive Bayes

A classification algorithm that uses Bayes' Theorem to predict the probability of a class based on features. It assumes features are independent of each other, making it simple and fast, but potentially less accurate.

Signup and view all the flashcards

Support Vector in SVM

A data point that directly impacts the decision boundary in an SVM. It's a point that is closest to the boundary line.

Signup and view all the flashcards

Activation Function

A function used in neural networks to introduce non-linearity into the model. It transforms the input data to allow for learning complex relationships.

Signup and view all the flashcards

Softmax Function

An activation function commonly used in the output layer of a multi-class classification neural network. It outputs a probability distribution over all classes, indicating the likelihood of each class being the correct prediction.

Signup and view all the flashcards

DBSCAN Clustering

A clustering algorithm that groups data points based on density. It identifies clusters as areas of high density separated by areas of low density.

Signup and view all the flashcards

K-Means Clustering

An algorithm that partitions data into k clusters by iteratively assigning data points to the closest cluster center and updating the cluster center based on the assigned data points.

Signup and view all the flashcards

Agglomerative Clustering

A hierarchical clustering method that starts with each data point as a separate cluster and then merges clusters iteratively based on their similarity until a desired number of clusters is reached.

Signup and view all the flashcards

Mean-Shift Clustering

A clustering algorithm that iteratively moves data points towards the denser regions of the data, eventually converging to cluster centers.

Signup and view all the flashcards

Naïve Bayes Classifier

A probabilistic classifier based on Bayes' Theorem that assumes independence of features given the class label. It calculates the probability of a data point belonging to each class and predicts the class with the highest probability.

Signup and view all the flashcards

Bayes' Theorem

A mathematical formula that calculates the conditional probability of an event A given that event B has occurred.

Signup and view all the flashcards

Support Vector Machine (SVM)

A supervised learning algorithm that finds a hyperplane that best separates data points of different classes. It maximizes the margin, or distance, between the hyperplane and the closest data points (support vectors).

Signup and view all the flashcards

Radial Basis Function (RBF) Kernel

A kernel function used in SVMs to handle non-linearly separable data by mapping data points into a higher-dimensional space. It calculates the similarity based on the distance between data points using a radial basis function.

Signup and view all the flashcards

Kernel Smoothing

A technique used in nonparametric density estimation, like kernel density estimation, to create a smooth and continuous approximation of the data distribution. This involves using a kernel function to weigh nearby data points, creating a smoother curve.

Signup and view all the flashcards

Bin Width (Histogram)

The size of the intervals used in histograms to group data points. A larger bin width combines more data points into each bin, leading to a smoother histogram, while a smaller bin width provides a more detailed but potentially jagged histogram.

Signup and view all the flashcards

Bandwidth (Kernel Density Estimation)

Similar to bin width, bandwidth in kernel density estimation controls the smoothness of the density estimate. A larger bandwidth creates a smoother curve with a larger influence of distant data points, while a smaller bandwidth leads to a more detailed, potentially noisy curve.

Signup and view all the flashcards

Kernel Density vs. Histogram

Kernel Density Estimation (KDE) uses kernel functions to create a smooth, continuous density estimate, while histograms use bins to group data points, leading to a stepped, discontinuous representation.

Signup and view all the flashcards

Bias-Variance Tradeoff

A central concept in machine learning. In density estimation, increasing bandwidth reduces variance (noise) but introduces bias (distortion). Conversely, decreasing bandwidth increases variance but reduces bias.

Signup and view all the flashcards

Nonparametric Methods

Statistical methods that do not make assumptions about the underlying distribution of the data. This allows them to work with more complex, real-world data distributions.

Signup and view all the flashcards

Smoothing (Density Estimation)

The process of creating a smoother, more continuous representation of the data distribution by adjusting the bandwidth or bin width in kernel density or histogram estimation.

Signup and view all the flashcards

Number of Bins and Histogram

Increasing the number of bins in a histogram generally results in a more jagged, detailed representation of the data distribution. This can reveal more subtle features within the data but potentially introduce noise.

Signup and view all the flashcards

What is the effect of bin width on histograms?

The bin width determines the number of bins in a histogram. A smaller bin width leads to more detailed visualization, but might cause gaps and instability with limited data. A larger bin width smooths out the data but might hide important details.

Signup and view all the flashcards

Why do histograms need large sample sizes?

Histograms require a sufficient amount of data to provide a reliable representation of the underlying distribution. Small sample sizes can lead to uneven and misleading bin counts.

Signup and view all the flashcards

What does the kernel function in kernel density estimation control?

In Kernel density estimation, the kernel function defines the shape of the smoothing function used to estimate the probability density. It influences the smoothness and detail level of the estimated density curve.

Signup and view all the flashcards

What is the primary purpose of kernel density estimation?

Kernel density estimation aims to create a smooth estimate of the probability density function from a given dataset. Unlike histograms, it avoids abrupt jumps and provides a more continuous representation of the distribution.

Signup and view all the flashcards

What does the parameter 'k' represent in k-nearest neighbor density estimation?

In k-nearest neighbor density estimation, 'k' determines the number of neighboring data points considered when estimating the density at a specific point. A higher 'k' value results in smoother density estimates.

Signup and view all the flashcards

What is the advantage of k-nearest neighbor density estimation?

A key advantage of k-nearest neighbor density estimation is its non-parametric nature. It doesn't require assuming a specific distribution shape. It's also simple to understand and implement.

Signup and view all the flashcards

Which nonparametric method is suitable for large datasets with unknown distributions?

Kernel density estimation is a preferred method for handling large datasets where the underlying distribution is unknown. It offers a smooth and flexible approach to density estimation.

Signup and view all the flashcards

What's the difference between histograms and kernel density estimators?

Histograms use bins to group data, creating a step-like visualization. Kernel density estimation uses smoothing functions to create a smooth curve representing the distribution. This makes kernel estimators more flexible and less prone to abrupt jumps.

Signup and view all the flashcards

Study Notes

Machine Learning Foundations - Unit 1

  • Machine learning is primarily focused on building algorithms that allow computers to learn and improve from data.
  • Examples of machine learning applications include predicting stock prices.
  • Data in machine learning is the information used to train and test models.
  • Supervised learning uses labeled data for learning.
  • Unsupervised learning aims to discover patterns or structures in unlabeled data.
  • Reinforcement Learning is a type of learning where an agent learns by interacting with an environment to maximize cumulative reward.
  • Data accuracy, computational speed and model complexity are challenges in machine learning.
  • The purpose of training in machine learning is to build a model that can predict or classify data.
  • Generalization addresses the model's ability to perform well on unseen data.

Machine Learning Foundations - Unit 1 Additional Topics (cont'd)

  • Feasibility of learning: The ability to learn effectively from available data.
  • Model complexity: The intricacy of the model; crucial to prevent overfitting.
  • Cross-validation: A measure of the discrepancy between training and testing performance.
  • Underfitting: A model that is too simplistic and fails to capture the underlying patterns in the data.
  • Overfitting: A model that is too complex and fits the noise in the training data but doesn't generalize to new data.
  • Bias-variance tradeoff: Balance between the model fitting the training data perfectly and not over-fitting.
  • Distance Metrics: Various measures for quantifying how far apart data points are, including Euclidean distance, Manhattan distance, and Minkowski distance.
  • Cosine similarity: A measure, useful for text data, of the angle between two vectors.
  • Jaccard coefficient: Measures the similarity between sets, common in text data.
  • Simple Matching Coefficient: Measures the proportion of matching attributes in binary data.
  • Pearson Correlation Coefficient: Used to calculate the correlation between two variables.
  • Distance Metrics: Euclidean distance, Manhattan distance, Minkowski distance, Cosine Similarity, Jaccard coefficient.
  • K-Nearest Neighbors (KNN): An algorithm where a new data point is classified based on the categories of the most similar nearby points.
  • KNN Challenges:: Complexity of decision boundary and high computational cost during prediction.
  • Decision Trees: Algorithm that creates a tree-like structure, effectively partitioning data based on feature values leading to predictions.
  • Decision Tree Algorithms: Apriori, ID3, and C4.5 are widely known.
  • Rule-Based Classifiers': Employ pre-defined if-then rules to classify data.
  • Polynomial Regression: Uses polynomial terms to capture non-linear relationships between variables.
  • Multicollinearity: Occurs when independent variables in a linear regression model are highly correlated.
  • Regularization Techniques: Method employed in models to prevent overfitting and improve generalization to new data like Ridge Regression and Lasso Regression.
  • Model Evaluation Metrics for Regression : MSE (Mean Squared Error), MAE (Mean Absolute Error), RMSE (Root Mean Squared Error), RMSLE (Root Mean Squared Logarithmic Error)

Clustering Analysis - Unit 2

  • k-Means Clustering: Aims to partition data into 'k' clusters based on minimizing the within-cluster variance.
  • Hierarchical Clustering: Creates clusters in a hierarchical structure (either agglomerative or divisive).
  • Agglomerative Clustering: Starts with each data point as a separate cluster and merges them based on the minimum distance.
  • Divisive Clustering: Starts with all data points in one cluster and recursively splits them into smaller clusters.
  • Ward's Method: A Hierarchical clustering method that minimizes the sum of squared differences within clusters.
  • Dendrogram: A tree-like diagram showing the hierarchical relationship between clusters.
  • Silhouette Coefficient: A measure of how similar a data point is to its own cluster compared to other clusters.
  • DBSCAN (Density-Based Spatial Clustering of Applications with Noise): Clusters data points based on density, identifying outliers as noise.
  • Core Points: Data points surrounded by a minimum number of points within a specific radius.
  • K-nearest neighbors (KNN): used in algorithms like DBSCAN
  • Feature scaling/data normalization: Improves the performance of some algorithms like KMeans.

Naïve Bayes and Support Vector Machines (SVM) - Unit 3

  • Naive Bayes: A probabilistic classifier based on Bayes' Theorem assuming features are conditionally independent given the class label.
  • Support Vector Machines (SVM): A supervised learning method that finds the optimal hyperplane to separate data points of different classes.
  • Kernel Trick: Transforms the data into a higher-dimensional space to solve non-linearly separable problems.
  • Linear Kernel: Used for linearly separable data.
  • Polynomial Kernel: Used for non-linearly separable data.
  • Gaussian Kernel: Also known as RBF (Radial Basis Function) kernel. Used for non-linearly separable data.

Studying That Suits You

Use AI to generate personalized quizzes and flashcards to suit your learning preferences.

Quiz Team

Related Documents

Description

i want same questions in pdf

More Like This

Use Quizgecko on...
Browser
Browser