Question

Advertisers fear that users of DVRs (digital video recorders) will “fast forward” past commercials when they...

Advertisers fear that users of DVRs (digital video recorders) will “fast forward” past commercials when they watch a recorded program. A leading British pay television company told its advertisers that this effect might be offset because DVR users watch more TV. A sample of 15 DVR users showed a daily mean screen time of 2 hours and 26 minutes with a standard deviation of 14 minutes, compared with a daily mean of 2 hours and 7 minutes with a standard deviation of 12 minutes for a sample of 15 non-DVR users.

(a-1) Construct a 95 percent confidence interval for the difference in mean TV watching. (Round your answers to 2 decimal places.)

The 95% confidence interval is from  to  .

Homework Answers

Answer #1

1 hour = 60 minutes

Know the answer?
Your Answer:

Post as a guest

Your Name:

What's your source?

Earn Coins

Coins can be redeemed for fabulous gifts.

Not the answer you're looking for?
Ask your own homework help question
Similar Questions
Advertisers fear that users of DVRs (digital video recorders) will “fast forward” past commercials when they...
Advertisers fear that users of DVRs (digital video recorders) will “fast forward” past commercials when they watch a recorded program. A leading British pay television company told its advertisers that this effect might be offset because DVR users watch more TV. A sample of 25 DVR users showed a daily mean screen time of 2 hours and 31 minutes with a standard deviation of 13 minutes, compared with a daily mean of 2 hours and 14 minutes with a standard...
In order to determine how many hours per week freshmen college students watch television, a random...
In order to determine how many hours per week freshmen college students watch television, a random sample of 25 students was selected. It was determined that the students in the sample spent an average of 19.5 hours with a sample standard deviation of 3.9 hours watching TV per week. Please answer the following questions: (a) Provide a 95% confidence interval estimate for the average number of hours that all college freshmen spend watching TV per week. (b) Assume that a...
1. A sample size of n = 70 is drawn from a population with proportion p...
1. A sample size of n = 70 is drawn from a population with proportion p = 0.32. Let p̂ be the sample proportion. Find p̂m and p̂s . Round the standard deviation to four decimal places. Find )28.0p̂P ( > . Find )40.0p̂P ( < 2.  On a certain television channel, 20% of the commercials are local advertisers. A sample of 150 commercials is selected. Would it be unusual for more than 26% of the commercials to be local advertisers?...
A Gallup poll conducted January 17 - February 6, 2019, asked 1024 teenagers 13 years of...
A Gallup poll conducted January 17 - February 6, 2019, asked 1024 teenagers 13 years of age to 17 years of age , “Typically, how many hours per week do you spend watching TV?’’ The teenagers who participated in the survey watched, on average, 13.0 hours of television per week with a standard deviation of 2.3 hours per week. Construct a 95% confidence interval for the average number of hours that teenagers watch TV each week
A research council wants to estimate the mean length of time (in minutes) that the average...
A research council wants to estimate the mean length of time (in minutes) that the average U.S. adult spends watching television using digital video recorders (DVR’s) each day. To determine the estimate, the research council takes random samples of 35 U.S. adults and obtains the following times in minutes. 24 27 26 29 33 21 18 24 23 34 17 15 19 23 25 29 36 19 18 22 16 45 32 12 24 35 14 40 30 19 14...
A research council wants to estimate the mean length of time (in minutes) that the average...
A research council wants to estimate the mean length of time (in minutes) that the average U.S. adult spends watching television using digital video recorders (DVR’s) each day. To determine the estimate, the research council takes random samples of 35 U.S. adults and obtains the following times in minutes. 24 27 26 29 33 21 18 24 23 34 17 15 19 23 25 29 36 19 18 22 16 45 32 12 24 35 14 40 30 19 14...
A research council wants to estimate the mean length of time (in minutes) that the average...
A research council wants to estimate the mean length of time (in minutes) that the average U.S. adult spends watching television using digital video recorders (DVR’s) each day. To determine the estimate, the research council takes random samples of 35 U.S. adults and obtains the following times in minutes. 24 27 26 29 33 21 18 24 23 34 17 15 19 23 25 29 36 19 18 22 16 45 32 12 24 35 14 40 30 19 14...
A study conducted a few years ago claims that the adult males spend an average of...
A study conducted a few years ago claims that the adult males spend an average of 11 hours a week watching sports on television. A recent sample of 100 adult males showed that the mean time they spend per week watching sports on television is 9.50 hours with a standard deviation of 2.2 hours. a. Use the result to give a 95% confidence interval for the mean time they spend per week watching sports on television. a. Does the study...
An advertising executive wants to estimate the mean weekly amount of time consumers spend watching traditional...
An advertising executive wants to estimate the mean weekly amount of time consumers spend watching traditional television daily. Based on previous​ studies, the standard deviation is assumed to be 24 minutes. The executive wants to​ estimate, with​ 99% confidence, the mean weekly amount of time to within plus or minus ±66 minutes. a. What sample size is​ needed? b. If​ 95% confidence is​ desired, how many consumers need to be​ selected?
1. the data collected from the 2012 General Social Survey (GSS) in which 1,298 respondents reported...
1. the data collected from the 2012 General Social Survey (GSS) in which 1,298 respondents reported on the number of hours they watched tv per day. The results show that the average (mean) number of hours spent watching tv per day is 3.089 hours with a standard deviation of .07 hours. a. Calculate 95% confidence interval to estimate the mean number of hours of tv the average American watches per day based on these data. b. What do you notice...