GPT-3: Language Model and Potential Implications Quiz

Choose a study mode

Play Quiz
Study Flashcards
Spaced Repetition
Chat to Lesson

Podcast

Play an AI-generated podcast conversation about this lesson

Questions and Answers

What is the training data for GPT-3?

  • Microsoft's Turing NLG
  • A filtered version of Common Crawl (correct)
  • Hundreds of billions of words
  • Review of Wikipedia

What language is GPT-3 capable of coding in?

  • CSS
  • JSX
  • Python (correct)
  • Java

What is the potential risk of GPT-3?

  • Spread of misinformation (correct)
  • Abuse of legal processes
  • Fraudulent academic essay writing
  • Social engineering pretexting

Flashcards are hidden until you start studying

Study Notes

  • GPT-3 is a third-generation language model that is 10 times larger than Microsoft's Turing NLG.

  • The training data for GPT-3 comes from a filtered version of Common Crawl.

  • GPT-3 was trained on hundreds of billions of words and is also capable of coding in CSS, JSX, and Python.

  • A 2022 review again highlighted that the training continues to include review of Wikipedia.

  • GPT-3 is "eerily good" at writing "amazingly coherent text" with only a few simple prompts.

  • GPT-3 is a machine learning algorithm that has the potential to advance both the beneficial and harmful applications of language models.

  • The potential harmful effects of GPT-3 include the spread of misinformation, spam, phishing, abuse of legal and governmental processes, fraudulent academic essay writing and social engineering pretexting.

  • The authors call for research on risk mitigation in order to avoid these dangers.

Studying That Suits You

Use AI to generate personalized quizzes and flashcards to suit your learning preferences.

Quiz Team

More Like This

Use Quizgecko on...
Browser
Browser