9+ Best Post Hoc Power Calculation Tests & Guide


9+ Best Post Hoc Power Calculation Tests & Guide

A statistical computation performed after a study has concluded, using the observed effect size, sample size, and alpha level to estimate the probability of detecting a true effect. For example, if a study fails to reject the null hypothesis, this calculation aims to determine if the failure was due to a lack of statistical power rather than a genuine absence of effect.

Understanding the achieved power provides context to non-significant findings. Historically, it has been used to justify underpowered studies or to claim that a non-significant result is “almost significant.” However, this application is often criticized because the computed value is directly related to the p-value and offers no additional information beyond what the p-value already conveys. Its use may lead to misinterpretations about the reliability and validity of research findings.

Given the potential for misinterpretation, subsequent discussion will delve into the appropriate interpretation of statistical power in the context of research design and the limitations associated with its retrospective computation. Furthermore, alternative approaches to interpreting non-significant results, such as confidence intervals and Bayesian methods, will be explored.

1. Retrospective assessment

Retrospective assessment forms the foundational basis of a post hoc power calculation. This calculation, by definition, occurs after data collection and analysis are complete. The process involves examining the already observed effect size, the sample size utilized, and the predetermined alpha level to estimate the statistical power achieved in the study. Without this retrospective view of the completed study, the inputs necessary for the calculation are unavailable. The temporal sequence is therefore critical; the assessment must be retrospective for the computation to be considered post hoc. For instance, if a clinical trial investigating a new drug yields a non-significant result, a retrospective power analysis aims to evaluate whether the study was adequately powered to detect a clinically meaningful effect, based on the observed differences between treatment groups.

The reliance on observed effect sizes in retrospective assessment presents both advantages and disadvantages. While it allows for a power estimate specific to the study’s actual results, it also creates a direct mathematical relationship with the p-value. Consequently, the computed power provides limited additional information. An underpowered study, identified through retrospective assessment, might prompt a larger, confirmatory trial. However, interpreting the initial non-significant result solely through the lens of power can be misleading, as it does not address potential issues with study design, measurement error, or confounding variables. A real-world example involves behavioral research; a study examining the impact of a specific intervention on student performance might yield non-significant results. A retrospective assessment would estimate the power based on the observed performance differences. If the power is low, it indicates a possible need for a larger sample or a more potent intervention in future research.

In conclusion, retrospective assessment is an indispensable element of a post hoc power calculation, enabling the computation of statistical power following data collection. While it offers insights into the study’s ability to detect a true effect, its inherent limitations, particularly the correlation with the p-value and reliance on observed effect sizes, necessitate cautious interpretation. The focus should shift towards prospective power analysis during the design phase of a study to ensure adequate power and robust results. Retrospective assessment, when used judiciously, serves as a supplementary tool for understanding the context of non-significant findings, but it should not be the primary basis for drawing definitive conclusions about the presence or absence of an effect.

2. Observed effect size

The observed effect size is a critical input into a post hoc power calculation. It represents the magnitude of the effect detected in a study, serving as a primary measure for estimating the probability of detecting a true effect if it exists. This observed value becomes the basis for determining statistical power retrospectively.

  • Magnitude Estimation

    The observed effect size provides a quantifiable estimate of the difference between groups or the strength of a relationship. Standardized measures like Cohen’s d or Pearson’s r allow for comparison across studies. For example, in a study comparing two teaching methods, the observed effect size might quantify the difference in student performance. In the context of post hoc power calculation, a larger observed effect size generally leads to a higher estimated power, assuming other factors remain constant.

  • Sampling Variability

    The observed effect size is subject to sampling variability; it is only an estimate of the true population effect. A small sample size can lead to an unstable estimate of the effect size. In a study with few participants, the observed effect size may be inflated or deflated due to random chance. Consequently, the post hoc power calculation can be misleading if based on an unreliable observed effect size.

  • Influence on Power Estimate

    The observed effect size has a direct and substantial influence on the resulting post hoc power estimate. A larger observed effect will lead to a larger power estimate. This relationship is mathematically determined and contributes to the criticism of using post hoc power, because it provides similar information as the p-value already obtained. In a study of a new drug, a large observed effect size would result in a high post hoc power, suggesting the study was likely to detect the effect if it was real.

  • Interpretation Challenges

    Relying on the observed effect size for post hoc power calculations presents interpretational challenges. Because the post hoc power is heavily influenced by the observed effect size, it adds little information beyond the p-value. It may lead researchers to overemphasize the importance of non-significant results if a moderately large effect was observed but the power was low. Appropriate interpretation necessitates careful consideration of confidence intervals and alternative statistical methods.

In conclusion, the observed effect size is an essential component of post hoc power calculation, serving as the basis for estimating statistical power after a study’s completion. Its inherent relationship with the p-value, susceptibility to sampling variability, and potential for misinterpretation underscore the need for cautious application and recognition of alternative methods for evaluating research findings.

3. Sample Size Dependent

The sample size exhibits a direct influence on the outcome of a post hoc power calculation. Power, defined as the probability of correctly rejecting a false null hypothesis, is inextricably linked to the number of observations included in a study. Studies with smaller sample sizes are inherently less likely to detect a true effect, resulting in lower post hoc power estimates. Conversely, larger sample sizes generally increase the likelihood of detecting an effect, leading to higher post hoc power. This dependency arises because a larger sample provides a more precise estimate of the population parameters, reducing the impact of random variability and increasing the accuracy of the observed effect size. For instance, a clinical trial with only 30 participants per arm may fail to detect a statistically significant difference in treatment efficacy, even if a true effect exists. A subsequent post hoc power calculation would likely reveal low power, indicating that the small sample size was insufficient to detect the effect.

The relationship between sample size and power is not linear; increasing the sample size yields diminishing returns in terms of power. Initially, increasing the sample size significantly boosts power. However, as the sample size grows, the incremental gain in power decreases. This concept is crucial for researchers when planning studies. They must carefully balance the desire for higher power with the practical constraints of time, resources, and participant availability. A study examining consumer preferences for a new product could initially show a substantial increase in power when the sample size is increased from 50 to 150 participants. However, further increasing the sample size to 300 or 500 might yield only marginal improvements in power, while significantly increasing the cost and effort of data collection. Understanding this non-linear relationship is essential for optimizing study design.

In summary, sample size is a critical determinant of post hoc power. Smaller sample sizes are prone to producing low power estimates, potentially leading to false negative conclusions. While increasing the sample size generally improves power, the gains diminish as the sample grows larger. Researchers must carefully consider the trade-offs between sample size, power, and practical limitations when designing studies. Furthermore, interpreting post hoc power calculations requires acknowledging the influence of sample size on the estimated power value, ensuring that conclusions are drawn with appropriate caution. The inherent limitations of post hoc power analysis, particularly its reliance on the observed effect size and its connection to the p-value, further emphasize the need for comprehensive study planning and careful interpretation of results.

4. P-value correlated

The p-value and the post hoc power calculation exhibit a mathematical dependency, making the latter largely redundant. The p-value represents the probability of observing data as extreme as, or more extreme than, the observed data, assuming the null hypothesis is true. The post hoc power calculation, using the observed effect size and sample size, estimates the probability of rejecting the null hypothesis. Because the observed effect size is directly derived from the data used to calculate the p-value, the resulting power estimate is intrinsically linked to the p-value itself. A smaller p-value will invariably lead to a higher post hoc power, and a larger p-value will result in lower power. This relationship renders the post hoc power calculation less informative than the p-value, as it provides no independent evidence.

The practical implication of this correlation is that the post hoc power calculation does not offer a meaningful reassessment of the statistical significance. A researcher, observing a non-significant result (e.g., p = 0.15), might attempt to justify the finding by calculating post hoc power. However, the resulting power estimate will only reflect the initial p-value. It will not provide any additional insight into whether the null hypothesis should be rejected. For example, a study investigating a new teaching method might yield a p-value of 0.20, indicating a lack of statistical significance. The post hoc power calculation, based on the observed effect size, will confirm this lack of power, but it will not alter the initial conclusion drawn from the p-value. A focus on confidence intervals or prospective power analysis would provide more valuable information.

In conclusion, the inherent correlation between the p-value and post hoc power calculations diminishes the practical utility of the latter. Because the power estimate is directly derived from the same data used to generate the p-value, the post hoc power calculation offers no new evidence or perspective. The p-value already conveys the strength of the evidence against the null hypothesis. Researchers should prioritize study design with prospective power analysis, and when interpreting results, focus on alternative methods such as confidence intervals. The redundancy and potential for misinterpretation associated with post hoc power calculations underscore the need for cautious application.

5. Interpretational Challenges

The application of a post hoc power calculation often presents interpretational challenges that stem from the calculation’s inherent limitations and potential for misuse. The primary challenge lies in the fact that the resulting power estimate is mathematically dependent on the p-value obtained from the original statistical test. This dependency creates a circularity in reasoning: a non-significant p-value will invariably lead to a low post hoc power estimate, thereby offering no additional substantive information about the study’s findings. The calculation simply restates the lack of statistical significance in terms of power, without providing new evidence regarding the presence or absence of a real effect. For instance, if a clinical trial examining the efficacy of a new drug results in a p-value of 0.25, a post hoc power calculation will likely indicate low power. The researcher, however, gains no further insight into whether the drug is truly ineffective or whether the study design was simply inadequate to detect a real effect. This redundancy undermines the value of the post hoc power calculation as a means of informing decision-making or guiding future research.

Another significant challenge arises from the temptation to use post hoc power to justify underpowered studies. A researcher, confronted with a non-significant result, might calculate post hoc power to argue that the study “almost” found a significant effect, but was simply limited by insufficient sample size. This argument can be misleading because it focuses attention on the potential for a true effect without addressing other possible explanations for the non-significant result, such as measurement error, confounding variables, or flaws in the study design. In behavioral research, for example, a study examining the impact of a specific intervention on student performance might yield a non-significant result. The researcher, after calculating post hoc power, might claim that the study was underpowered and that a larger sample size would have revealed a significant effect. However, this claim ignores the possibility that the intervention itself was ineffective, or that other factors, such as student motivation or teacher quality, were influencing the outcome. Furthermore, the reliance on the observed effect size in post hoc power calculations introduces bias, as the observed effect size is often an overestimate of the true population effect, especially in studies with small sample sizes. This can lead to inflated power estimates and unwarranted conclusions.

In summary, the interpretational challenges associated with post hoc power calculations stem from its inherent dependency on the p-value, its potential for misuse in justifying underpowered studies, and its reliance on biased estimates of the effect size. Understanding these challenges is crucial for researchers to avoid drawing misleading conclusions from post hoc power calculations. Instead, emphasis should be placed on prospective power analysis during the study design phase to ensure adequate power and robust results. When interpreting results, focus should be on alternative methods such as confidence intervals and Bayesian analysis, which provide more comprehensive and nuanced assessments of the evidence.

6. Justification Questioned

The application of post hoc power calculation as a means of justifying non-significant results in research faces increasing scrutiny. Its utility in rescuing studies with insufficient power at the design stage is significantly challenged due to inherent methodological flaws. The primary issue stems from the calculation’s dependence on the observed effect size and the obtained p-value. Since the post hoc power is derived from these values, it provides no additional information beyond what the p-value already conveys, rendering its use as a justification for non-significance logically circular. For example, a study on a novel educational intervention failing to demonstrate a statistically significant improvement in student performance might employ post hoc power to argue that the lack of significance is attributable to insufficient sample size. However, the power calculation, influenced by the non-significant p-value, merely confirms the study’s inability to detect an effect, without providing independent evidence of the intervention’s potential effectiveness or the adequacy of the research design.

Further undermining its justification, the observed effect size, a central component of post hoc power calculation, is prone to inflation, especially in studies with small sample sizes. This inflation can lead to an overestimation of the study’s power, creating a false sense of confidence in the existence of an effect that may not be real. Researchers might, therefore, mistakenly attribute the non-significant result solely to a lack of power, overlooking other critical factors such as flawed methodology, measurement error, or the presence of confounding variables. A pharmaceutical study, failing to show the efficacy of a new drug due to poor patient compliance, might calculate post hoc power to claim that the study was underpowered, obscuring the more fundamental issue of non-adherence impacting the observed effect. Consequently, the use of post hoc power to justify non-significant findings can lead to a misrepresentation of the study’s limitations and potentially misguided directions for future research.

The questioned justification of post hoc power calculations stems from their lack of independent evidentiary value and their potential to obfuscate underlying methodological issues. Prospective power analysis, conducted during the study design phase, offers a more robust approach to ensuring adequate power and minimizing the risk of false negative results. When interpreting study results, particularly non-significant findings, researchers should prioritize confidence intervals, effect size estimates, and a critical examination of the study’s design and methodology over reliance on post hoc power. This approach fosters a more transparent and rigorous assessment of research findings, avoiding the pitfalls associated with using post hoc power as a means of justifying questionable results.

7. Resource allocation

Resource allocation, encompassing the strategic deployment of financial, human, and technological assets, is inextricably linked to considerations of statistical power in research. Effective allocation decisions directly influence the feasibility of achieving sufficient power, while conversely, suboptimal allocation may necessitate post hoc power calculations, often revealing limitations in the study’s ability to detect true effects.

  • Prospective Power Analysis and Funding Justification

    Prior to initiating a study, a well-conducted power analysis informs resource allocation by estimating the sample size required to achieve a desired level of statistical power. This analysis directly influences budgetary requests for personnel, equipment, and participant recruitment. For instance, a clinical trial aimed at demonstrating the superiority of a novel treatment regimen necessitates a power analysis to determine the number of patients needed to detect a clinically meaningful difference. The outcome of this analysis dictates the scope and cost of the trial. Failure to adequately fund the study based on this initial power calculation may lead to an underpowered study, increasing the likelihood of a false negative result and necessitating post hoc power calculations to assess the study’s limitations.

  • Trade-offs Between Sample Size and Measurement Precision

    Resource constraints often necessitate trade-offs between increasing sample size and improving the precision of measurements. Allocating resources towards more accurate measurement tools or rigorously trained data collectors can reduce measurement error, thereby increasing statistical power without requiring a larger sample. Conversely, a study prioritizing a large sample size at the expense of measurement precision may suffer from reduced power due to increased noise in the data. A post hoc power calculation might reveal that the observed non-significant result is attributable not only to the sample size, but also to the high degree of measurement error stemming from suboptimal resource allocation. A study investigating the relationship between exercise and mood could allocate resources towards either recruiting more participants or employing more reliable mood assessment instruments. The decision will have significant implications for statistical power and the interpretability of the results.

  • Adaptive Designs and Interim Analyses

    Adaptive study designs, employing interim analyses to adjust sample size or treatment allocation based on accumulating data, represent a sophisticated approach to resource allocation. These designs allow for early stopping if the treatment effect is convincingly demonstrated or for increasing the sample size if the initial results are inconclusive. The decision to adjust the study design hinges on statistical power considerations at each interim analysis. Although not strictly post hoc, these interim power calculations inform ongoing resource allocation decisions. In contrast, a traditional study design lacking interim assessments may find itself underpowered at the conclusion, leading to a post hoc power calculation that reveals the missed opportunity for adaptation.

  • Impact on Generalizability

    Resource constraints can also affect the diversity and representativeness of the study sample, thereby limiting the generalizability of the findings. If resources are limited, a researcher may be tempted to recruit a more homogenous sample, reducing variability and potentially increasing statistical power. However, this comes at the cost of reduced external validity. A post hoc power calculation, even if indicating adequate power, does not address the limitation of generalizability arising from the sample composition. A survey on political attitudes conducted primarily among college students might achieve sufficient statistical power to detect certain trends, but the findings may not be representative of the broader population.

These facets illustrate the profound interplay between resource allocation and statistical power considerations. While a post hoc power calculation may offer insights into the limitations of a completed study, its value is diminished if fundamental resource allocation decisions during the planning phase compromised the study’s ability to detect true effects or produce generalizable findings. Therefore, a prospective and strategic approach to resource allocation, informed by rigorous power analysis and a clear understanding of study objectives, is paramount to conducting high-quality research.

8. Misleading inferences

The application of post hoc power calculations carries a substantial risk of generating misleading inferences regarding research findings. This arises primarily from the inherent limitations of the calculation and the potential for its misinterpretation. The dependence of post hoc power on observed effect sizes and p-values, derived from the same data, creates a circular logic that often results in inaccurate conclusions about the validity and reliability of study results.

  • Overemphasis on Non-Significant Trends

    Post hoc power may lead to an overemphasis on non-significant trends by suggesting that a larger sample size would have yielded a significant result. This interpretation often overlooks other potential explanations for the non-significance, such as flaws in the study design, measurement error, or the absence of a true effect. A study evaluating a new marketing strategy, failing to demonstrate a statistically significant increase in sales, might calculate post hoc power and conclude that the lack of significance is solely due to a small sample size. This conclusion may neglect other critical factors such as ineffective advertising or poor product quality.

  • Inflated Effect Size Estimates

    The observed effect size, a key input in the post hoc power calculation, is often an inflated estimate of the true population effect, particularly in studies with small sample sizes. This inflation can lead to an overestimation of power and a false sense of confidence in the existence of an effect. In clinical research, a preliminary study with a small patient cohort may show a large, albeit non-significant, effect of a new drug. Calculating post hoc power based on this inflated effect size could lead to the misleading inference that the drug is highly promising, even though the true effect may be much smaller or non-existent.

  • Neglect of Type II Error Considerations

    While post hoc power focuses on the probability of avoiding a Type II error (failing to reject a false null hypothesis), it often neglects the broader context of Type I error (incorrectly rejecting a true null hypothesis). Emphasizing post hoc power can lead researchers to accept a non-significant finding, arguing that the study was underpowered, without adequately considering the potential for a false positive result in a study with a larger sample size. A study evaluating the effectiveness of a new educational program might fail to demonstrate a significant improvement in student test scores but, based on post hoc power, conclude that the program is potentially effective. This conclusion disregards the possibility that a larger study could have revealed a false positive result due to confounding variables.

  • Circular Reasoning and Lack of Independent Evidence

    The most significant source of misleading inferences arises from the circular reasoning inherent in post hoc power calculations. Because the power estimate is directly derived from the p-value and observed effect size, it provides no independent evidence to support or refute the study’s findings. It merely restates the lack of statistical significance in terms of power. A study investigating the link between social media usage and mental health might find a non-significant correlation. The subsequent post hoc power calculation, based on the observed correlation and the p-value, confirms the lack of power but provides no new information regarding the true relationship between these variables.

In conclusion, the application of post hoc power calculations can readily lead to misleading inferences concerning research outcomes. The dependence of the calculation on observed effect sizes and p-values, its potential for overemphasizing non-significant trends, and its neglect of Type I error considerations all contribute to this risk. The reliance on these calculations for justifying study conclusions or guiding future research directions should be approached with caution, and greater emphasis should be placed on rigorous study design, prospective power analysis, and the careful interpretation of results within the context of broader scientific evidence.

9. Alternative approaches

The limitations inherent in post hoc power calculations necessitate consideration of alternative approaches for interpreting research findings. These alternatives mitigate the risks of drawing misleading inferences often associated with retrospective power analyses. While post hoc power calculations attempt to assess the probability of detecting an effect after a study has been completed, alternative methods offer more robust and informative strategies for evaluating the evidence. The application of these approaches influences the interpretation of results, particularly when statistical significance is not achieved. For instance, instead of relying on a post hoc power calculation to suggest that a non-significant result might be due to low power, alternative methods encourage a more comprehensive evaluation of the data and the study design. This leads to more reasoned conclusions about the presence or absence of an effect.

One prominent alternative involves focusing on confidence intervals. Confidence intervals provide a range of plausible values for the true population parameter, offering a more nuanced perspective than a simple binary assessment of statistical significance. If the confidence interval is wide and includes both clinically meaningful and null values, it indicates a lack of precision in the estimate, irrespective of the p-value or post hoc power. Another approach involves Bayesian methods, which incorporate prior knowledge or beliefs into the analysis, providing a more comprehensive assessment of the evidence. Furthermore, emphasis on effect sizes and their practical significance, rather than solely relying on statistical significance, allows for a more meaningful interpretation of research findings. For example, in a study comparing two different therapies, if the confidence interval for the difference in outcomes includes zero, the Bayesian posterior probability would be important information. The assessment shifts from whether there is any difference to whether the difference is clinically meaningful.

In conclusion, the adoption of alternative approaches, such as confidence intervals, Bayesian methods, and an emphasis on effect sizes, addresses the shortcomings of post hoc power calculations. These alternatives provide a more informative and less misleading framework for interpreting research results. By shifting the focus from retrospective power analysis to a more holistic evaluation of the evidence, researchers can draw more valid and reliable conclusions, enhancing the overall quality and impact of scientific inquiry.

Frequently Asked Questions About Post Hoc Power Calculation

This section addresses common queries and misconceptions surrounding post hoc power calculation. It provides concise and informative answers to enhance comprehension of this statistical concept.

Question 1: What precisely is a post hoc power calculation?

It is a statistical computation performed after a study has been completed. The computation uses the observed effect size, sample size, and alpha level to estimate the achieved power of the study, reflecting the probability of detecting a true effect, if one existed.

Question 2: Why is the utilization of post hoc power calculations often criticized?

The criticism stems from its mathematical dependency on the p-value. The post hoc power calculation provides limited additional information, rendering it largely redundant. Furthermore, its use can lead to misleading interpretations of non-significant results.

Question 3: Is post hoc power calculation an appropriate method for justifying non-significant findings?

No. Its application for justifying non-significant findings is generally discouraged due to its inherent limitations and potential for misinterpretation. It offers no independent evidence beyond the p-value, and can obscure other methodological issues.

Question 4: How does the observed effect size influence the post hoc power estimate?

The observed effect size directly influences the power estimate. A larger observed effect will lead to a higher power estimate. However, this effect size is subject to sampling variability and may not accurately reflect the true population effect.

Question 5: What are more reliable alternatives to post hoc power calculation?

Alternatives include focusing on confidence intervals, employing Bayesian methods, and emphasizing the practical significance of effect sizes. These approaches offer a more comprehensive and nuanced assessment of the evidence.

Question 6: How can researchers ensure adequate statistical power in their studies?

Researchers should conduct prospective power analyses during the study design phase. This ensures that the sample size is sufficient to detect a meaningful effect, if it exists. Proper planning and resource allocation are crucial.

Post hoc power calculations are viewed cautiously due to their reliance on observed data and their limited capacity to offer new insights beyond the p-value. Alternative approaches to interpreting research findings are favored for their comprehensive perspectives.

The following sections will address the practical implications of these concerns and outline best practices for statistical analysis.

Tips Regarding Post Hoc Power Calculation

The following guidelines outline prudent practices when considering or encountering post hoc power calculation in research.

Tip 1: Acknowledge Inherent Limitations: Recognize that post hoc power calculation is mathematically linked to the p-value and, therefore, provides limited additional insight. Avoid attributing excessive importance to a non-significant result solely based on the computed power.

Tip 2: Prioritize Prospective Power Analysis: Emphasize prospective power analysis during the study design phase. Determine the sample size required to achieve adequate power, mitigating the need for post hoc power calculation later.

Tip 3: Interpret Confidence Intervals: Focus on interpreting confidence intervals to assess the range of plausible values for the true population parameter. This offers a more nuanced perspective than relying solely on statistical significance or post hoc power calculation.

Tip 4: Evaluate Effect Sizes: Evaluate the magnitude and practical significance of effect sizes, irrespective of statistical significance. This allows for a more meaningful interpretation of research findings beyond the limitations of post hoc power calculation.

Tip 5: Consider Bayesian Methods: Explore the application of Bayesian methods, which incorporate prior knowledge and beliefs into the analysis. This provides a more comprehensive assessment of the evidence, offering an alternative to post hoc power calculation.

Tip 6: Critically Assess Study Design: Examine the study design for potential flaws, measurement error, or confounding variables. Avoid solely attributing non-significant results to a lack of power as indicated by post hoc power calculation.

Tip 7: Avoid Misleading Inferences: Be aware of the risk of drawing misleading inferences from post hoc power calculation. Its reliance on observed effect sizes and p-values can lead to inaccurate conclusions about the validity of study results.

Utilizing these practices enhances the rigor and transparency of research, minimizing the potential for misinterpretation associated with post hoc power calculation.

Subsequent sections will explore additional strategies for improving the quality and interpretability of research data.

Conclusion

This exploration of post hoc power calculation has revealed its inherent limitations and potential for misinterpretation. The dependence of post hoc power calculation on observed effect sizes and p-values, coupled with its tendency to offer redundant information, diminishes its utility in evaluating research findings. The application of post hoc power calculation for justifying non-significant results, or for guiding future research directions, warrants considerable caution due to the risk of drawing inaccurate conclusions.

The scientific community should prioritize rigorous study design, emphasize prospective power analyses, and embrace alternative methods for interpreting research results. Through conscientious application of statistical principles and a commitment to transparent reporting, researchers can enhance the validity and reliability of scientific inquiry.