i am stumped!
an electrical company repairs very large numbers of television sets and wishes to estimate the mean time taken to repair a particular fault. it is known from previous research that the standard deviation of the time taken to repair this fault is 2.5 minutes. the manager wishes to ensure that the probability that the estimate differs from the true mean by less than 30 seconds is 0.95. find how large a sample is required.
i reckon that p(|e - μ | < 0) < 0.95 where e is the estimate
is that correct? and what on earth do i do next? thanks so much for the help, i'm self-teaching and feeling very discouraged!