The probability that a research finding is indeed true depends on:
Finding | True Relationship | ||
---|---|---|---|
Yes | No | Total | |
Yes | \[c(1-\beta)R/(R+1)\] | \[c\alpha/(R+1)\] | \[c(R+\alpha-\beta R)/(R+1)\] |
No | \[c\beta R/(R+1)\] | \[c(1-\alpha)/(R+1)\] | \[c(1-\alpha+\beta R)/(R+1)\] |
Total | \[cR/(R+1)\] | \[c/(R+1)\] | \[c\] |
A combination of various factors that tend to produce research findings when they should not be produced including:
\(1-\beta\) | R | Bias, u | Example | PPV |
---|---|---|---|---|
0.80 | 1:1 | 0.10 | Powered RCT with little bias a 1:1 pre-study odds | 0.85 |
0.95 | 2:1 | 0.30 | Confirmatory meta-analysis of good quality RCTs | 0.85 |
0.80 | 1:10 | 0.30 | Adequately powered exploratory epidemiological study | 0.20 |
0.20 | 1:1000 | 0.80 | Discovery oriented exploratory research with massive testing limited bias | 0.0015 |
Cellular and molecular neuroscience.
Given a random variable (experiment; say rolling a die) a probability measure is a population quantity that summarizes the randomness.
The Russian mathematician Andrey Nikolaevich Kolmogorov formalized these rules.
\(A = \{1\}\) and \(B = \{1, 3, 5\}\). Then
\[ P(D|+) = \frac{P(+|D)P(D)}{P(+|D)P(D) + P(+|D^c)P(D^c)} \] \[P(D^c|+) = \frac{P(+|D^c)P(D^c)}{P(+|D)P(D) + P(+|D^c)P(D^c)}\]
\[ \bar X = \sum_{i=1}^n x_i p(x_i) \] where \(p(x_i) = 1/n\)
—
—
\[Var(X) = E[X^2] - E[X]^2 = p - p^2 = p(1 - p)\] —
—
—
—
Truth | Decide | Result |
---|---|---|
\(H_0\) | \(H_0\) | Correctly accept null |
\(H_0\) | \(H_a\) | Type I error |
\(H_a\) | \(H_a\) | Correctly reject null |
\(H_a\) | \(H_0\) | Type II error |
Idea: Suppose nothing is going on - how unusual is it to see the estimate we got? Approach:
What’s the probability of getting a \(T\) statistic as large as \(2.5\)?
—
—
Term | Meaning | Common Uses |
---|---|---|
Standard deviation | The typical difference between each value and the mean value. | Describing how broadly the sample values are distributed. \[s.d.=\sqrt{\sum(X-\bar{X})^2/(N-1)}\] |
Standard error of the mean (s.e.m) | An estimate how variable the means will be if the experiment is repeated multiple times. | Inferring where the population mean is likely to lie or whether set of samples are likely to come from the sample population. \[s.e.m.=s.d./\sqrt{N}\] |
Confidence Interval (CI:95%) | with 95% confidence, the population mean will lie in this interval. | Top interfere where the population mean lies, and to compare two populations \[CI=mean\pm s.e.m. \times t_{(N-1)}\] |
Independent Data | Values from separate of the same type that are not linked | Testing hypothesis about population. |
Replicate data | Values from experiment where everything is linked as much as possible. | Serves as an internal check on performance of an experiment. |
Sampling error | Variation caused by sampling part of a population rather than measuring the whole population. | Can reveal bias in the data or problems with conduct of experiment. In binomial distributions the expected is \[\sqrt{Np(1-p)}\]; in Poisson the expected s.d. is \[\sqrt{mean}\] |
More intuitively, the statistical power can be thought of as the probability of accepting an alternative hypothesis, when the alternative hypothesis is true.