Many times, we are required to use statistical measures to try and construct a problem. We run a program for a 10 different inputs. The times are measures in 1-second intervals and none of them took 0 secs.
a. Suppose the standard deviation of a set of times we run the program is 0. What does this tell you about the running times?
b. Suppose that the mean of the times is 1000.9 sec while the median is 1 sec. Explain what do you know about the program running times for all 10 different inputs?
c. Assume now that the mean of the times is 1000.9 sec while the median is 1 sec and the variance is 9998000. Explain what do you know about the program running times for all 10 different inputs
The question belongs to Statistics and it discusses about standard deviation of a program being run for 0 seconds.
Total Word Count 72
If you are here for the first time, you can request for a discount coupon, which can knock off upto 20% of the quoted price on any service.