Blood is taken from each of n individuals to be tested for a certain disease. Rather than test each sample separately, a pooled method is used in an attempt to reduce the number of tests needed. Part of each blood sample is taken, and these parts are combined to form a pooled sample. The pooled sample is then tested. If the result is negative, then none of the n individuals has the disease, and no further tests are needed. If the pooled sample tests positive, then each individual is tested to see which of them have the disease. a. Let X represent the number of tests that are carried out. What are the possible values of X? b. Assume that \(n=4\) individuals are to be tested, and the probability that each has the disease, independent of the others, is \(p=0.1\). Find \(\mu_X\). c. Repeat part (b) with \(n=6\) and \(p=0.2\). d. Express \(\mu_X\) as a function of n and p. e. The pooled method is more economical than performing individual tests if \(\mu_X<n\). Suppose \(n=10\). For what values of p is the pooled method more economical than performing n individual tests? Equation Transcription: Text Transcription: n=4 p=0.1 mu_X n=6 p=0.2 mu_X mu_X<n n=10
Read moreTable of Contents
Appendix B
Partial Derivatives
1
Sampling and Descriptive Statistics
1
Sampling and Descriptive Statistics
1.1
Sampling
1.1
Sampling
1.2
Summary Statistics
1.2
Summary Statistics
1.3
Graphical Summaries
1.3
Graphical Summaries
2
Probability
2
Probability
2.1
Basic Ideas
2.1
Basic Ideas
2.2
Counting Methods
2.2
Counting Methods
2.3
Conditional Probability and Independence
2.3
Conditional Probability and Independence
2.4
Random Variables
2.4
Random Variables
2.5
Linear Functions of Random Variables
2.5
Linear Functions of Random Variables
2.6
Jointly Distributed Random Variables
2.6
Jointly Distributed Random Variables
3
Propagation of Error
3
Propagation of Error
3.1
Measurement Error
3.1
Measurement Error
3.2
Linear Combinations of Measurements
3.2
Linear Combinations of Measurements
3.3
Uncertainties for Functions of One Measurement
3.3
Uncertainties for Functions of One Measurement
3.4
Uncertainties for Functions of Several Measurements
3.4
Uncertainties for Functions of Several Measurements
4
Commonly Used Distributions
4
Commonly Used Distributions
4.1
The Bernoulli Distribution
4.1
The Bernoulli Distribution
4.10
Probability Plots
4.11
The Central Limit Theorem
4.11
The Central Limit Theorem
4.12
Simulation
4.12
Simulation
4.15
4.2
The Binomial Distribution
4.2
The Binomial Distribution
4.3
The Poisson Distribution
4.3
The Poisson Distribution
4.4
Some Other Discrete Distributions
4.4
Some Other Discrete Distributions
4.5
The Normal Distribution
4.5
The Normal Distribution
4.6
The Lognormal Distribution
4.6
The Lognormal Distribution
4.7
The Exponential Distribution
4.7
The Exponential Distribution
4.8
Some Other Continuous Distributions
4.8
Some Other Continuous Distributions
4.9
Some Principles of Point Estimation
4.9
Some Principles of Point Estimation
5
Confidence Intervals
5
Confidence Intervals
5.1
Large-Sample Confidence Intervals for a Population Mean
5.1
Large-Sample Confidence Intervals for a Population Mean
5.10
Using Simulation to Construct Confidence Intervals
5.2
Confidence Intervals for Proportions
5.2
Confidence Intervals for Proportions
5.3
Small-Sample Confidence Intervals for a Population Mean
5.3
Small-Sample Confidence Intervals for a Population Mean
5.4
Confidence Intervals for the Difference Between Two Means
5.4
Confidence Intervals for the Difference Between Two Means
5.5
Confidence Intervals for the Difference Between Two Proportions
5.5
Confidence Intervals for the Difference Between Two Proportions
5.6
Small-Sample Confidence Intervals for the Difference Between Two Means
5.6
Small-Sample Confidence Intervals for the Difference Between Two Means
5.7
Confidence Intervals with Paired Data
5.7
Confidence Intervals with Paired Data
5.8
Confidence Intervals for the Variance and Standard Deviation of a Normal Population
5.8
Confidence Intervals for the Variance and Standard Deviation of a Normal Population
5.9
Prediction Intervals and Tolerance Intervals
5.9
Prediction Intervals and Tolerance Intervals
6
Hypothesis Testing
6
Hypothesis Testing
6.1
Large-Sample Tests for a Population Mean
6.1
Large-Sample Tests for a Population Mean
6.10
Tests with Categorical Data
6.11
Tests for Variances of Normal Populations
6.11
Tests for Variances of Normal Populations
6.12
Fixed-Level Testing
6.12
Fixed-Level Testing
6.13
Power
6.13
Power
6.14
Multiple Tests
6.14
Multiple Tests
6.15
Using Simulation to Perform Hypothesis Tests
6.15
Using Simulation to Perform Hypothesis Tests
6.2
Drawing Conclusions from the Results of Hypothesis Tests
6.2
Drawing Conclusions from the Results of Hypothesis Tests
6.3
Tests for a Population Proportion
6.3
Tests for a Population Proportion
6.4
Small-Sample Tests for a Population Mean
6.4
Small-Sample Tests for a Population Mean
6.5
Large-Sample Tests for the Difference Between Two Means
6.5
Large-Sample Tests for the Difference Between Two Means
6.6
Tests for the Difference Between Two Proportions
6.6
Tests for the Difference Between Two Proportions
6.7
Small-Sample Tests for the Difference Between Two Means
6.7
Small-Sample Tests for the Difference Between Two Means
6.8
Tests with Paired Data
6.8
Tests with Paired Data
6.9
Distribution-Free Tests
6.9
Distribution-Free Tests
7
Correlation and Simple Linear Regression
7
Correlation and Simple Linear Regression
7.1
Correlation
7.1
Correlation
7.2
The Least-Squares Line
7.2
The Least-Squares Line
7.3
Uncertainties in the Least-Squares Coefficients
7.3
Uncertainties in the Least-Squares Coefficients
7.4
Checking Assumptions and Transforming Data
7.4
Checking Assumptions and Transforming Data
8
Multiple Regression
8
Multiple Regression
8.1
The Multiple Regression Model
8.1
The Multiple Regression Model
8.2
Confounding and Collinearity
8.2
Confounding and Collinearity
8.3
Model Selection
8.3
Model Selection
9
Factorial Experiments
9
Factorial Experiments
9.1
One-Factor Experiments
9.1
One-Factor Experiments
9.2
Pairwise Comparisons in One-Factor Experiments
9.2
Pairwise Comparisons in One-Factor Experiments
9.3
Two-Factor Experiments
9.3
Two-Factor Experiments
9.4
Randomized Complete Block Designs
9.4
Randomized Complete Block Designs
9.5
2p Factorial Experiments
9.5
2p Factorial Experiments
10
Statistical Quality Control
10
Statistical Quality Control
10.1
Basic Ideas
10.1
Basic Ideas
10.2
Control Charts for Variables
10.2
Control Charts for Variables
10.3
Control Charts for Attributes
10.3
Control Charts for Attributes
10.4
The CUSUM Chart
10.4
The CUSUM Chart
10.5
Process Capability
10.5
Process Capability
Textbook Solutions for Statistics for Engineers and Scientists
Chapter 2 Problem 1SE
Question
A system consists of four components connected as shown.
Assume A, B, C, and D function independently. If the probabilities that A, B, C, and D fail are 0.1, 0.2, 0.05, and 0.3, respectively, what is the probability that the system functions?
Solution
Solution :
Given, a system consists four component A,B,C and D.
From the given diagram
A be the event that component function is A.
B be the event that component function is B.
C be the event that component function is C and
D be the event that component function is D.
Here we assumed independently function is A,B,C and D.
Our goal is to find
a). What is the probability that the system functions?
Step 1 of 1 :
If the probability fails
A=0.1, B=0.2, C=0.05 and C=0.3.
Then,
Subscribe to view the
full solution
full solution
Title
Statistics for Engineers and Scientists 4
Author
William Navidi
ISBN
9780073401331