🎧 New: AI-Generated Podcasts Turn your study notes into engaging audio conversations. Learn more

Mastering the Stratified Approach for Machine Learning
40 Questions
0 Views

Mastering the Stratified Approach for Machine Learning

Created by
@CozyOctopus

Podcast Beta

Play an AI-generated podcast conversation about this lesson

Questions and Answers

What is the most popular regression metric used to describe the distance between prediction and actual?

  • Mean Squared Logarithmic Error (MSLE)
  • Mean Square Error (MSE) (correct)
  • Mean Absolute Error (MAE)
  • Root Mean Square Error (RMSE)
  • What is the purpose of using a histogram/KDE model in regression?

  • To calculate Mean Directional Accuracy
  • To calculate Mean Absolute Scaled Error
  • To calculate Median Absolute Error (MedAE)
  • To visualize errors distribution (correct)
  • Why is it more difficult to make a correct assessment of a classification model?

  • Classification models do not have evaluation metrics
  • It is easier to make a correct assessment of a regression model
  • Classification models are not used in business outcomes
  • It requires more knowledge and abstract thinking (correct)
  • What is the symmetric version of Mean Absolute Percentage Error (MAPE)?

    <p>sMAPE</p> Signup and view all the answers

    What are evaluation metrics in machine learning?

    <p>Functions used to train and monitor the quality of a model</p> Signup and view all the answers

    What are the properties that cost function should meet in machine learning?

    <p>Differentiability with respect to parameters</p> Signup and view all the answers

    What is the role of evaluation metrics in assessing model accuracy?

    <p>To evaluate the performance of a model during training and testing</p> Signup and view all the answers

    Why do evaluation metrics not have to comply with restrictive mathematical properties?

    <p>They are calculated after the estimator is already created with use of different cost function</p> Signup and view all the answers

    What is the purpose of using evaluation metrics and plots dedicated to probabilities in classification tasks?

    <p>To make decisions about probability cut-off points in a responsible and aware way</p> Signup and view all the answers

    What is the Receiver Operating Characteristic Curve (ROC) used for?

    <p>To plot TPR and FPR for every probability cut-off point</p> Signup and view all the answers

    What is the difference between AUC ROC and AUC PR?

    <p>AUC ROC measures the tradeoff between true positive rate and false positive rate while AUC PR measures the tradeoff between precision and recall</p> Signup and view all the answers

    What is Log-loss or Cross entropy or Entropy used for?

    <p>To evaluate the performance of a classification model</p> Signup and view all the answers

    What is the purpose of using a stratify approach in train-validation pair?

    <p>To ensure that relative class frequencies are approximately preserved</p> Signup and view all the answers

    What is the purpose of creating several models independently on the train and validation data?

    <p>To select one best model on the testing sample</p> Signup and view all the answers

    What is cross-validation (CV)?

    <p>A technique for evaluating a machine learning model and testing its performance</p> Signup and view all the answers

    Why is cross-validation considered more robust than a single train-validation split?

    <p>Because it uses different portions of the data to validate and train the model on different iterations</p> Signup and view all the answers

    Which metric can be used as an ultimate metric to assess the quality of a model's ROC curve?

    <p>Area Under the Curve ROC (AUC ROC)</p> Signup and view all the answers

    What is the range of values that AUC ROC can take?

    <p>0.5 to 1</p> Signup and view all the answers

    Which type of classification tasks is ROC curve not well suited for?

    <p>Imbalanced classification tasks</p> Signup and view all the answers

    What is the Precision Recall curve visualization used for?

    <p>To combine precision and recall in a single visualization</p> Signup and view all the answers

    What is the interpretation of AUC ROC?

    <p>The probability that a uniformly drawn random positive has a higher score than a uniformly drawn random negative</p> Signup and view all the answers

    What is the purpose of AUC PR in highly imbalanced problems?

    <p>To get one representative number for the whole model</p> Signup and view all the answers

    What is the difference between bias and variance of a model?

    <p>Bias is the difference between the expected prediction and the correct model, and variance is the variability of the model prediction for given data points</p> Signup and view all the answers

    What is the bias/variance trade-off?

    <p>The simpler the model, the higher the bias, and the more complex the model, the higher the variance</p> Signup and view all the answers

    What is the Continuous Ranked Probability Score (CRPS)?

    <p>A metric that generalizes the mean absolute error (MAE) to the case of probabilistic forecasts</p> Signup and view all the answers

    What is the Matthews Correlation Coefficient?

    <p>A metric that measures the correlation between predicted classes and ground truth in binary classification</p> Signup and view all the answers

    What is the False Positive Rate?

    <p>The proportion of actual negative observations that are incorrectly classified as positive</p> Signup and view all the answers

    What is the F beta score?

    <p>A combination of precision and recall in one metric</p> Signup and view all the answers

    What is the True Positive Rate?

    <p>The proportion of actual positive observations that are correctly classified as positive</p> Signup and view all the answers

    What is the Positive Predictive Value?

    <p>The proportion of observations predicted as positive that are actually positive</p> Signup and view all the answers

    What is the purpose of using validation/cross validation in machine learning?

    <p>To assess the quality of our model in a quasi-objective way and to execute hyperparameter tuning safely</p> Signup and view all the answers

    Which of the following is NOT a type of cross-validation discussed in the text?

    <p>Leave-p-out</p> Signup and view all the answers

    What is a hyperparameter in machine learning?

    <p>A parameter that controls the learning process and is not estimable</p> Signup and view all the answers

    Which type of cross-validation is most commonly used for cross-sectional problems?

    <p>K-folds</p> Signup and view all the answers

    Why are there multiple types of cross-validation?

    <p>To cater to the specifics of the data, business problem, dataset size, imbalance, and computing resources</p> Signup and view all the answers

    What is a learning curve?

    <p>A plot of model learning performance over experience or time</p> Signup and view all the answers

    What is the purpose of a validation learning curve?

    <p>To give an idea of how well the model is generalizing</p> Signup and view all the answers

    What is the bias/variance trade-off?

    <p>The trade-off between overfitting and underfitting in a model</p> Signup and view all the answers

    What is the purpose of dividing a data set into training, validation, and testing sets?

    <p>To avoid overfitting the model to the training data</p> Signup and view all the answers

    What is an imbalanced dataset?

    <p>A dataset with unequal distribution of classes</p> Signup and view all the answers

    More Quizzes Like This

    Mastering Options Greeks
    24 questions
    Mastering Physics Set 4 Midterm Flashcards
    29 questions
    Mastering Biology 2 Flashcards
    35 questions
    Mastering Biology Chapter 7 Flashcards
    70 questions
    Use Quizgecko on...
    Browser
    Browser