Podcast
Questions and Answers
What is the marginal probability of X being rainy?
What is the marginal probability of X being rainy?
- 0
- 1/2
- 1/4
- 3/4 (correct)
What is the value of H(X | Y) computed in hartlys?
What is the value of H(X | Y) computed in hartlys?
- 0.324
- 0.075
- 0.207 (correct)
- 1.284
What is the method used to compute conditional entropy H(X | Y)?
What is the method used to compute conditional entropy H(X | Y)?
- Calculating the difference between H(X) and H(Y).
- Using joint probabilities only.
- Only considering the probabilities of X.
- Taking the sum of conditional entropies multiplied by marginal probabilities. (correct)
What is H(Y | X) computed in hartlys?
What is H(Y | X) computed in hartlys?
Which values are part of the joint probability mass function for X and Y?
Which values are part of the joint probability mass function for X and Y?
Flashcards are hidden until you start studying
Study Notes
Conditional Entropy
- Conditional Entropy H(X|Y) quantifies the uncertainty of random variable X given that another variable Y is known.
- Mathematically defined as:
- H(X|Y) = Σ p(y) H(X|Y=yi) for all y in Y
- Equivalently expressed as H(X|Y) = -Σ Σ p(x, y) log p(x|y) for all x in X and y in Y.
- Specific case for conditional entropy of X given a particular outcome of Y (yi):
- H(X|Y=yi) = -Σ p(xj|yi) log p(xj|yi).
Chain Rule
- The chain rule for joint entropy illustrates the relationship between individual entropies and conditional uncertainties:
- H(X, Y) = H(X) + H(Y|X).
- A symmetrical expression also exists:
- H(X, Y) = H(Y) + H(X|Y).
- This rule clarifies that joint uncertainty can be decomposed into the uncertainty of individual variables and their dependencies.
Example Calculations
- Two variables X (weather: sunny or rainy) and Y (temperature: above or below 70 degrees) are utilized to compute conditional entropies.
- Marginal probabilities derived from the example are:
- P(X): {3/4 rainy, 1/4 sunny}
- P(Y): {3/4 above 70, 1/4 below 70}
- Conditional entropy H(X|Y) calculated at 0.207 hartlys.
Joint Probability Mass Function
- A second example presents a joint probability mass function for variables X and Y with several discrete outcomes.
- The joint distribution is defined with probabilities summing to 1 over all possible X and Y combinations.
- Calculated entropies:
- H(X|Y) yields a result of approximately 1.284 hartlys.
- H(Y|X) results in approximately 0.324 hartlys.
Assignments
- Tasks include calculations of H(X), H(Y), and H(X,Y) in hartlys based on joint distributions and marginal probabilities provided from previous sections.
Studying That Suits You
Use AI to generate personalized quizzes and flashcards to suit your learning preferences.