1) The probability density function of a Markov process is
a. p(x1,x2,x3.......xn) = p(x1)p(x2/x1)p(x3/x2).......p(xn/xn-1)
b. p(x1,x2,x3.......xn) = p(x1)p(x1/x2)p(x2/x3).......p(xn-1/xn)
c. p(x1,x2,x3......xn) = p(x1)p(x2)p(x3).......p(xn)
d. p(x1,x2,x3......xn) = p(x1)p(x2 * x1)p(x3 * x2)........p(xn * xn-1)
Answer
Explanation
|
ANSWER: p(x1,x2,x3.......xn) = p(x1)p(x2/x1)p(x3/x2).......p(xn/xn-1)
Explanation: No explanation is available for this question!
|
|
2) The capacity of Gaussian channel is
a. C = 2B(1+S/N) bits/s
b. C = B 2(1+S/N) bits/s
c. C = B(1+S/N) bits/s
d. C = B(1+S/N)2 bits/s
Answer
Explanation
|
ANSWER: C = B(1+S/N) bits/s
Explanation: No explanation is available for this question!
|
|
3) For M equally likely messages, the average amount of information H is
a. H = log 10M
b. H = log 2M
c. H = log 10M 2
d. H = 2log 10M
Answer
Explanation
|
ANSWER: H = log2M
Explanation: No explanation is available for this question!
|
|
4) The channel capacity is
a. The maximum information transmitted by one symbol over the channel
b. Information contained in a signal
c. The amplitude of the modulated signal
d. All of the above
Answer
Explanation
|
ANSWER: The maximum information transmitted by one symbol over the channel
Explanation: No explanation is available for this question!
|
|
5) The capacity of a binary symmetric channel, given H(P) is binary entropy function is
a. 1 - H(P)
b. H(P) - 1
c. 1 - H(P) 2
d. H(P) 2 - 1
Answer
Explanation
|
ANSWER: 1 - H(P)
Explanation: No explanation is available for this question!
|
|
6) According to Shannon Hartley theorem,
a. The channel capacity becomes infinite with infinite bandwidth
b. The channel capacity does not become infinite with infinite bandwidth
c. Has a tradeoff between bandwidth and Signal to noise ratio
d. Both b and c are correct
Answer
Explanation
|
ANSWER: Both b and c are correct
Explanation: No explanation is available for this question!
|
|
7) The negative statement for Shannon's theorem states that
a. If R > C, the error probability increases towards Unity
b. If R < C, the error probability is very small
c. Both a & b
d. None of the above
Answer
Explanation
|
ANSWER: If R > C, the error probability increases towards Unity
Explanation: No explanation is available for this question!
|
|
8) For M equally likely messages, M>>1, if the rate of information R ≤ C, the probability of error is
a. Arbitrarily small
b. Close to unity
c. Not predictable
d. Unknown
Answer
Explanation
|
ANSWER: Arbitrarily small
Explanation: No explanation is available for this question!
|
|
9) For M equally likely messages, M>>1, if the rate of information R > C, the probability of error is
a. Arbitrarily small
b. Close to unity
c. Not predictable
d. Unknown
Answer
Explanation
|
ANSWER: Close to unity
Explanation: No explanation is available for this question!
|
|
10) The channel capacity according to Shannon's equation is
a. Maximum error free communication
b. Defined for optimum system
c. Information transmitted
d. All of the above
Answer
Explanation
|
ANSWER: All of the above
Explanation: No explanation is available for this question!
|
|