SEA100 Explorations of AI Ethics: Fairness and Bias Quiz
12 Questions
1 Views

Choose a study mode

Play Quiz
Study Flashcards
Spaced Repetition
Chat to lesson

Podcast

Play an AI-generated podcast conversation about this lesson

Questions and Answers

What are the three main causes of bias in AI?

  • Underrepresentation, proxy measurements, and hidden protected features (correct)
  • Overrepresentation, direct measurements, and diverse decision thresholds
  • Unbalanced data, biased metrics, and subgroup thresholds
  • Group-independent predictions, equal metrics, and same model training
  • Which method ensures fairness in AI by having the same predictions when a protected feature is hidden from the model?

  • Enforcing diverse decision thresholds
  • Using biased metrics
  • Balanced data representation
  • Group-independent predictions (correct)
  • How can bias affect the results of an AI model according to the text?

  • By using complex financial models for investment recommendations
  • By discriminating based on gender due to biased training data (correct)
  • By providing audio output only for visually impaired users
  • By ensuring all subgroups are equally represented in the data
  • What is a potential risk highlighted in the text regarding data exposure in AI applications?

    <p>Storage of sensitive patient data insecurely</p> Signup and view all the answers

    In the context of AI ethics, what does it mean to use proxy measurements?

    <p>Indirectly inferring sensitive attributes to avoid bias</p> Signup and view all the answers

    Which principle of responsible AI involves ensuring the same performance metrics across different subgroups?

    <p>Equal metrics across subgroups</p> Signup and view all the answers

    According to the principles of responsible AI, who is liable for AI-driven decisions that result in harm?

    <p>The company that implemented the AI system</p> Signup and view all the answers

    Which of the following is not a principle of responsible AI as outlined by Microsoft?

    <p>Creativity</p> Signup and view all the answers

    What is one way to reduce bias in AI models?

    <p>Reduce imbalance and errors in data</p> Signup and view all the answers

    What is a common challenge associated with AI?

    <p>AI is introducing bias into models</p> Signup and view all the answers

    Which of the following is not a common AI workload?

    <p>Manual data entry</p> Signup and view all the answers

    Which of the following is not a tool provided by Microsoft Azure for responsible AI?

    <p>AI-Driven Decision Making Tool</p> Signup and view all the answers

    Study Notes

    Responsible AI

    • AI-driven decisions can lead to unintended consequences, such as an innocent person being convicted of a crime based on biased facial recognition evidence.
    • Microsoft's Responsible AI principles aim to address challenges and risks associated with AI, including bias, fairness, reliability, safety, privacy, security, inclusiveness, transparency, and accountability.

    Causes of Bias in AI

    • Three main causes of bias in AI are:
      • Data imbalance and errors
      • Using proxy measurements instead of direct measurements
      • Failing to enforce equal metrics across subgroups

    Measures of Fairness

    • Fairness in AI can be measured by:
      • Group-independent predictions
      • Same predictions when a protected feature (e.g. sex) is hidden from the model
      • Equal metrics across subgroups (e.g. accuracy and TP/FN rates for men and women)

    Methods for Fairness

    • To make AI fair, it's essential to:
      • Ensure data is balanced and all subgroups are equally represented
      • Directly measure instead of using proxy measurements
      • Enforce model training to have same metrics for subgroups
      • Use different decision thresholds for subgroups to counter bias

    Other Challenges and Risks with AI

    • Errors in AI systems can cause harm, such as an autonomous vehicle experiencing a system failure and causing a collision.
    • AI systems can expose sensitive data, such as a medical diagnostic bot trained using insecure patient data.
    • Solutions may not work for everyone, such as a predictive app providing no audio output for visually impaired users.
    • Users must trust complex AI systems, such as an AI-based financial tool making investment recommendations without clear explanations.

    Studying That Suits You

    Use AI to generate personalized quizzes and flashcards to suit your learning preferences.

    Quiz Team

    Description

    Test your knowledge on fairness, bias, and ethical considerations in artificial intelligence with this quiz based on the course SEA100 Explorations of Artificial Intelligence Ethics at Seneca College. Explore causes of bias, measures of fairness, methods for ensuring fairness, as well as other challenges and risks associated with AI.

    More Like This

    AI Ethics: Fairness and Transparency
    10 questions
    Deontological Ethics and Fairness in AI
    20 questions
    Module 2: AI Bias and Ethics
    40 questions
    Ethics of AI and Fairness
    18 questions
    Use Quizgecko on...
    Browser
    Browser