Question : In probability theory, a conditional probability measures the probability of an event given that (by assumption, presumption, assertion or evidence) another event has occurred. So how do you represent P(X,Y,Z)?
Correct Answer : Get Lastest Questions and Answer : Explanation: From the definition, P(X,Y|Z) P(Y|Z) =P(X,Y,Z)/P(Z) P(Y,Z)/P(Z) =P(X,Y,Z) P(Y,Z) =P(X|Y,Z) This follows from the definition of conditional probability, applied twice: P(X,Y)=(PX|Y)P(Y) Let's look at some other problems in which we are asked to find a conditional probability.
Example 1: A jar contains black and white marbles. Two marbles are chosen without replacement. The probability of selecting a black marble and then a white marble is 0.34, and the probability of selecting a black marble on the first draw is 0.47. What is the probability of selecting a white marble on the second draw, given that the first marble drawn was black? Solution: P(White|Black) = P(Black and White)/P(Black) = 0.34/0.47 = 0.72 = 72%
Example 2: The probability that it is Friday and that a student is absent is 0.03. Since there are 5 school days in a week, the probability that it is Friday is 0.2. What is the probability that a student is absent given that today is Friday? Solution: P(Absent|Friday) = P(Friday and Absent)/P(Friday) = 0.03/0.2 = 0.15 = 15%
Example 3: At Kennedy Middle School, the probability that a student takes Technology and Spanish is 0.087. The probability that a student takes Technology is 0.68. What is the probability that a student takes Spanish given that the student is taking Technology? Solution: P(Spanish|Technology) = P(Technology and Spanish)/P(Technology) = 0.087/0.68 = 0.13 = 13%
Question : The use of log probabilities is widespread in several fields of computer science such as information theory and natural language processing because
1. Speed. Since multiplication is more expensive than addition, taking the product of a high number of probabilities is faster if they are represented in log form. (The conversion to log form is expensive, but is only incurred once.)
Explanation: 1.Speed. Since multiplication is more expensive than addition, taking the product of a high number of probabilities is faster if they are represented in log form. (The conversion to log form is expensive, but is only incurred once.) 2.Accuracy. The use of log probabilities improves numerical stability. The use of log probabilities is widespread in several fields of computer science such as information theory and natural language processing as it represents the surprisal, the minimum length of the message that specifies the outcome in an optimally efficient code.
The log of the product of probabilities is simply the sum of logs of those probabilities, which is easy to compute. The log of the sum of probabilities does not give rise to any similar shortcut, and so a transformation to log space does not help.
Because probabilities cannot exceed 1.0, getting a very large product of probabilities -- overflow -- is not a problem. But, probabilities can be very small, and their product smaller still, so underflows (very small positive values) are a real problem, especially in contexts like Bayes nets. However, the log of very small values does not have a small absolute value, and so working in log space avoids underflow.
Working in log space does not impact the time to convergence for MLE or gradient descent.
Question : Select the correct statement which applies to calculate the probabilities in log space? 1. It is more efficient to divide probabilities in log space 2. It is more efficient to substract probabilities in log space 3. Access Mostly Uused Products by 50000+ Subscribers 4. It is more efficient to sum probabilities in log space
Explanation:Speed. Since multiplication is more expensive than addition, taking the product of a high number of probabilities is faster if they are represented in log form. (The conversion to log form is expensive, but is only incurred once.) 2.Accuracy. The use of log probabilities improves numerical stability. The use of log probabilities is widespread in several fields of computer science such as information theory and natural language processing as it represents the surprisal, the minimum length of the message that specifies the outcome in an optimally efficient code.
The log of the product of probabilities is simply the sum of logs of those probabilities, which is easy to compute. The log of the sum of probabilities does not give rise to any similar shortcut, and so a transformation to log space does not help.
Because probabilities cannot exceed 1.0, getting a very large product of probabilities -- overflow -- is not a problem. But, probabilities can be very small, and their product smaller still, so underflows (very small positive values) are a real problem, especially in contexts like Bayes nets. However, the log of very small values does not have a small absolute value, and so working in log space avoids underflow.
Working in log space does not impact the time to convergence for MLE or gradient descent.
1. As long as both A > 0 and B > 0, you want very few words to be expected to never appear in spam 2. As long as both A > 0 and B > 0, you want very few words to be expected to always appear in spam 3. Access Mostly Uused Products by 50000+ Subscribers 4. All of the above.
1. Increase the probability mass of items seen zero times, and increase the probability mass of items seen at least once. 2. Decrease the probability mass of items seen zero times, and decrease the probability mass of items seen at least once. 3. Access Mostly Uused Products by 50000+ Subscribers 4. Increase the probability mass of items seen zero times, and decrease the probability mass of items seen at least once.