Podcast
Questions and Answers
What is the training data for GPT-3?
What is the training data for GPT-3?
- Microsoft's Turing NLG
- A filtered version of Common Crawl (correct)
- Hundreds of billions of words
- Review of Wikipedia
What language is GPT-3 capable of coding in?
What language is GPT-3 capable of coding in?
- CSS
- JSX
- Python (correct)
- Java
What is the potential risk of GPT-3?
What is the potential risk of GPT-3?
- Spread of misinformation (correct)
- Abuse of legal processes
- Fraudulent academic essay writing
- Social engineering pretexting
Flashcards are hidden until you start studying
Study Notes
-
GPT-3 is a third-generation language model that is 10 times larger than Microsoft's Turing NLG.
-
The training data for GPT-3 comes from a filtered version of Common Crawl.
-
GPT-3 was trained on hundreds of billions of words and is also capable of coding in CSS, JSX, and Python.
-
A 2022 review again highlighted that the training continues to include review of Wikipedia.
-
GPT-3 is "eerily good" at writing "amazingly coherent text" with only a few simple prompts.
-
GPT-3 is a machine learning algorithm that has the potential to advance both the beneficial and harmful applications of language models.
-
The potential harmful effects of GPT-3 include the spread of misinformation, spam, phishing, abuse of legal and governmental processes, fraudulent academic essay writing and social engineering pretexting.
-
The authors call for research on risk mitigation in order to avoid these dangers.
Studying That Suits You
Use AI to generate personalized quizzes and flashcards to suit your learning preferences.