Weekend batch
Avijeet is a Senior Research Analyst at Simplilearn. Passionate about Data Analytics, Machine Learning, and Deep Learning, Avijeet is also interested in politics, cricket, and football.
Free eBook: Top Programming Languages For A Data Scientist
Normality Test in Minitab: Minitab with Statistics
Machine Learning Career Guide: A Playbook to Becoming a Machine Learning Engineer
Writing a hypothesis is one of the essential elements of a scientific research paper. It needs to be to the point, clearly communicating what your research is trying to accomplish. A blurry, drawn-out, or complexly-structured hypothesis can confuse your readers. Or worse, the editor and peer reviewers.
A captivating hypothesis is not too intricate. This blog will take you through the process so that, by the end of it, you have a better idea of how to convey your research paper's intent in just one sentence.
The first step in your scientific endeavor, a hypothesis, is a strong, concise statement that forms the basis of your research. It is not the same as a thesis statement , which is a brief summary of your research paper .
The sole purpose of a hypothesis is to predict your paper's findings, data, and conclusion. It comes from a place of curiosity and intuition . When you write a hypothesis, you're essentially making an educated guess based on scientific prejudices and evidence, which is further proven or disproven through the scientific method.
The reason for undertaking research is to observe a specific phenomenon. A hypothesis, therefore, lays out what the said phenomenon is. And it does so through two variables, an independent and dependent variable.
The independent variable is the cause behind the observation, while the dependent variable is the effect of the cause. A good example of this is “mixing red and blue forms purple.” In this hypothesis, mixing red and blue is the independent variable as you're combining the two colors at your own will. The formation of purple is the dependent variable as, in this case, it is conditional to the independent variable.
Types of hypotheses
Some would stand by the notion that there are only two types of hypotheses: a Null hypothesis and an Alternative hypothesis. While that may have some truth to it, it would be better to fully distinguish the most common forms as these terms come up so often, which might leave you out of context.
Apart from Null and Alternative, there are Complex, Simple, Directional, Non-Directional, Statistical, and Associative and casual hypotheses. They don't necessarily have to be exclusive, as one hypothesis can tick many boxes, but knowing the distinctions between them will make it easier for you to construct your own.
A null hypothesis proposes no relationship between two variables. Denoted by H 0 , it is a negative statement like “Attending physiotherapy sessions does not affect athletes' on-field performance.” Here, the author claims physiotherapy sessions have no effect on on-field performances. Even if there is, it's only a coincidence.
Considered to be the opposite of a null hypothesis, an alternative hypothesis is donated as H1 or Ha. It explicitly states that the dependent variable affects the independent variable. A good alternative hypothesis example is “Attending physiotherapy sessions improves athletes' on-field performance.” or “Water evaporates at 100 °C. ” The alternative hypothesis further branches into directional and non-directional.
A simple hypothesis is a statement made to reflect the relation between exactly two variables. One independent and one dependent. Consider the example, “Smoking is a prominent cause of lung cancer." The dependent variable, lung cancer, is dependent on the independent variable, smoking.
In contrast to a simple hypothesis, a complex hypothesis implies the relationship between multiple independent and dependent variables. For instance, “Individuals who eat more fruits tend to have higher immunity, lesser cholesterol, and high metabolism.” The independent variable is eating more fruits, while the dependent variables are higher immunity, lesser cholesterol, and high metabolism.
Associative and casual hypotheses don't exhibit how many variables there will be. They define the relationship between the variables. In an associative hypothesis, changing any one variable, dependent or independent, affects others. In a casual hypothesis, the independent variable directly affects the dependent.
Also referred to as the working hypothesis, an empirical hypothesis claims a theory's validation via experiments and observation. This way, the statement appears justifiable and different from a wild guess.
Say, the hypothesis is “Women who take iron tablets face a lesser risk of anemia than those who take vitamin B12.” This is an example of an empirical hypothesis where the researcher the statement after assessing a group of women who take iron tablets and charting the findings.
The point of a statistical hypothesis is to test an already existing hypothesis by studying a population sample. Hypothesis like “44% of the Indian population belong in the age group of 22-27.” leverage evidence to prove or disprove a particular statement.
Writing a hypothesis is essential as it can make or break your research for you. That includes your chances of getting published in a journal. So when you're designing one, keep an eye out for these pointers:
Outside of academia, hypothesis and prediction are often used interchangeably. In research writing, this is not only confusing but also incorrect. And although a hypothesis and prediction are guesses at their core, there are many differences between them.
A hypothesis is an educated guess or even a testable prediction validated through research. It aims to analyze the gathered evidence and facts to define a relationship between variables and put forth a logical explanation behind the nature of events.
Predictions are assumptions or expected outcomes made without any backing evidence. They are more fictionally inclined regardless of where they originate from.
For this reason, a hypothesis holds much more weight than a prediction. It sticks to the scientific method rather than pure guesswork. "Planets revolve around the Sun." is an example of a hypothesis as it is previous knowledge and observed trends. Additionally, we can test it through the scientific method.
Whereas "COVID-19 will be eradicated by 2030." is a prediction. Even though it results from past trends, we can't prove or disprove it. So, the only way this gets validated is to wait and watch if COVID-19 cases end by 2030.
Quick tips on writing a hypothesis
A hypothesis should instantly address the research question or the problem statement. To do so, you need to ask a question. Understand the constraints of your undertaken research topic and then formulate a simple and topic-centric problem. Only after that can you develop a hypothesis and further test for evidence.
Once you have your research's foundation laid out, it would be best to conduct preliminary research. Go through previous theories, academic papers, data, and experiments before you start curating your research hypothesis. It will give you an idea of your hypothesis's viability or originality.
Making use of references from relevant research papers helps draft a good research hypothesis. SciSpace Discover offers a repository of over 270 million research papers to browse through and gain a deeper understanding of related studies on a particular topic. Additionally, you can use SciSpace Copilot , your AI research assistant, for reading any lengthy research paper and getting a more summarized context of it. A hypothesis can be formed after evaluating many such summarized research papers. Copilot also offers explanations for theories and equations, explains paper in simplified version, allows you to highlight any text in the paper or clip math equations and tables and provides a deeper, clear understanding of what is being said. This can improve the hypothesis by helping you identify potential research gaps.
Variables are an essential part of any reasonable hypothesis. So, identify your independent and dependent variable(s) and form a correlation between them. The ideal way to do this is to write the hypothetical assumption in the ‘if-then' form. If you use this form, make sure that you state the predefined relationship between the variables.
In another way, you can choose to present your hypothesis as a comparison between two variables. Here, you must specify the difference you expect to observe in the results.
Now that everything is in place, it's time to write your hypothesis. For starters, create the first draft. In this version, write what you expect to find from your research.
Clearly separate your independent and dependent variables and the link between them. Don't fixate on syntax at this stage. The goal is to ensure your hypothesis addresses the issue.
After preparing the first draft of your hypothesis, you need to inspect it thoroughly. It should tick all the boxes, like being concise, straightforward, relevant, and accurate. Your final hypothesis has to be well-structured as well.
Research projects are an exciting and crucial part of being a scholar. And once you have your research question, you need a great hypothesis to begin conducting research. Thus, knowing how to write a hypothesis is very important.
Now that you have a firmer grasp on what a good hypothesis constitutes, the different kinds there are, and what process to follow, you will find it much easier to write your hypothesis, which ultimately helps your research.
Now it's easier than ever to streamline your research workflow with SciSpace Discover . Its integrated, comprehensive end-to-end platform for research allows scholars to easily discover, write and publish their research and fosters collaboration.
It includes everything you need, including a repository of over 270 million research papers across disciplines, SEO-optimized summaries and public profiles to show your expertise and experience.
If you found these tips on writing a research hypothesis useful, head over to our blog on Statistical Hypothesis Testing to learn about the top researchers, papers, and institutions in this domain.
1. what is the definition of hypothesis.
According to the Oxford dictionary, a hypothesis is defined as “An idea or explanation of something that is based on a few known facts, but that has not yet been proved to be true or correct”.
The hypothesis is a statement that proposes a relationship between two or more variables. An example: "If we increase the number of new users who join our platform by 25%, then we will see an increase in revenue."
A null hypothesis is a statement that there is no relationship between two variables. The null hypothesis is written as H0. The null hypothesis states that there is no effect. For example, if you're studying whether or not a particular type of exercise increases strength, your null hypothesis will be "there is no difference in strength between people who exercise and people who don't."
• Fundamental research
• Applied research
• Qualitative research
• Quantitative research
• Mixed research
• Exploratory research
• Longitudinal research
• Cross-sectional research
• Field research
• Laboratory research
• Fixed research
• Flexible research
• Action research
• Policy research
• Classification research
• Comparative research
• Causal research
• Inductive research
• Deductive research
• Your hypothesis should be able to predict the relationship and outcome.
• Avoid wordiness by keeping it simple and brief.
• Your hypothesis should contain observable and testable outcomes.
• Your hypothesis should be relevant to the research question.
• Null hypotheses are used to test the claim that "there is no difference between two groups of data".
• Alternative hypotheses test the claim that "there is a difference between two data groups".
A research question is a broad, open-ended question you will try to answer through your research. A hypothesis is a statement based on prior research or theory that you expect to be true due to your study. Example - Research question: What are the factors that influence the adoption of the new technology? Research hypothesis: There is a positive relationship between age, education and income level with the adoption of the new technology.
The plural of hypothesis is hypotheses. Here's an example of how it would be used in a statement, "Numerous well-considered hypotheses are presented in this part, and they are supported by tables and figures that are well-illustrated."
The red queen hypothesis in evolutionary biology states that species must constantly evolve to avoid extinction because if they don't, they will be outcompeted by other species that are evolving. Leigh Van Valen first proposed it in 1973; since then, it has been tested and substantiated many times.
The father of the null hypothesis is Sir Ronald Fisher. He published a paper in 1925 that introduced the concept of null hypothesis testing, and he was also the first to use the term itself.
You need to find a significant difference between your two populations to reject the null hypothesis. You can determine that by running statistical tests such as an independent sample t-test or a dependent sample t-test. You should reject the null hypothesis if the p-value is less than 0.05.
The bottom line.
Hypothesis testing, sometimes called significance testing, is an act in statistics whereby an analyst tests an assumption regarding a population parameter. The methodology employed by the analyst depends on the nature of the data used and the reason for the analysis.
Hypothesis testing is used to assess the plausibility of a hypothesis by using sample data. Such data may come from a larger population or a data-generating process. The word "population" will be used for both of these cases in the following descriptions.
In hypothesis testing, an analyst tests a statistical sample, intending to provide evidence on the plausibility of the null hypothesis. Statistical analysts measure and examine a random sample of the population being analyzed. All analysts use a random population sample to test two different hypotheses: the null hypothesis and the alternative hypothesis.
The null hypothesis is usually a hypothesis of equality between population parameters; e.g., a null hypothesis may state that the population mean return is equal to zero. The alternative hypothesis is effectively the opposite of a null hypothesis. Thus, they are mutually exclusive , and only one can be true. However, one of the two hypotheses will always be true.
The null hypothesis is a statement about a population parameter, such as the population mean, that is assumed to be true.
If an individual wants to test that a penny has exactly a 50% chance of landing on heads, the null hypothesis would be that 50% is correct, and the alternative hypothesis would be that 50% is not correct. Mathematically, the null hypothesis is represented as Ho: P = 0.5. The alternative hypothesis is shown as "Ha" and is identical to the null hypothesis, except with the equal sign struck-through, meaning that it does not equal 50%.
A random sample of 100 coin flips is taken, and the null hypothesis is tested. If it is found that the 100 coin flips were distributed as 40 heads and 60 tails, the analyst would assume that a penny does not have a 50% chance of landing on heads and would reject the null hypothesis and accept the alternative hypothesis.
If there were 48 heads and 52 tails, then it is plausible that the coin could be fair and still produce such a result. In cases such as this where the null hypothesis is "accepted," the analyst states that the difference between the expected results (50 heads and 50 tails) and the observed results (48 heads and 52 tails) is "explainable by chance alone."
Some statisticians attribute the first hypothesis tests to satirical writer John Arbuthnot in 1710, who studied male and female births in England after observing that in nearly every year, male births exceeded female births by a slight proportion. Arbuthnot calculated that the probability of this happening by chance was small, and therefore it was due to “divine providence.”
Hypothesis testing helps assess the accuracy of new ideas or theories by testing them against data. This allows researchers to determine whether the evidence supports their hypothesis, helping to avoid false claims and conclusions. Hypothesis testing also provides a framework for decision-making based on data rather than personal opinions or biases. By relying on statistical analysis, hypothesis testing helps to reduce the effects of chance and confounding variables, providing a robust framework for making informed conclusions.
Hypothesis testing relies exclusively on data and doesn’t provide a comprehensive understanding of the subject being studied. Additionally, the accuracy of the results depends on the quality of the available data and the statistical methods used. Inaccurate data or inappropriate hypothesis formulation may lead to incorrect conclusions or failed tests. Hypothesis testing can also lead to errors, such as analysts either accepting or rejecting a null hypothesis when they shouldn’t have. These errors may result in false conclusions or missed opportunities to identify significant patterns or relationships in the data.
Hypothesis testing refers to a statistical process that helps researchers determine the reliability of a study. By using a well-formulated hypothesis and set of statistical tests, individuals or businesses can make inferences about the population that they are studying and draw conclusions based on the data presented. All hypothesis testing methods have the same four-step process, which includes stating the hypotheses, formulating an analysis plan, analyzing the sample data, and analyzing the result.
Sage. " Introduction to Hypothesis Testing ," Page 4.
Elder Research. " Who Invented the Null Hypothesis? "
Formplus. " Hypothesis Testing: Definition, Uses, Limitations and Examples ."
Harvard Business School Online's Business Insights Blog provides the career insights you need to achieve your goals and gain confidence in your business skills.
Becoming a more data-driven decision-maker can bring several benefits to your organization, enabling you to identify new opportunities to pursue and threats to abate. Rather than allowing subjective thinking to guide your business strategy, backing your decisions with data can empower your company to become more innovative and, ultimately, profitable.
If you’re new to data-driven decision-making, you might be wondering how data translates into business strategy. The answer lies in generating a hypothesis and verifying or rejecting it based on what various forms of data tell you.
Below is a look at hypothesis testing and the role it plays in helping businesses become more data-driven.
Access your free e-book today.
To understand what hypothesis testing is, it’s important first to understand what a hypothesis is.
A hypothesis or hypothesis statement seeks to explain why something has happened, or what might happen, under certain conditions. It can also be used to understand how different variables relate to each other. Hypotheses are often written as if-then statements; for example, “If this happens, then this will happen.”
Hypothesis testing , then, is a statistical means of testing an assumption stated in a hypothesis. While the specific methodology leveraged depends on the nature of the hypothesis and data available, hypothesis testing typically uses sample data to extrapolate insights about a larger population.
When it comes to data-driven decision-making, there’s a certain amount of risk that can mislead a professional. This could be due to flawed thinking or observations, incomplete or inaccurate data , or the presence of unknown variables. The danger in this is that, if major strategic decisions are made based on flawed insights, it can lead to wasted resources, missed opportunities, and catastrophic outcomes.
The real value of hypothesis testing in business is that it allows professionals to test their theories and assumptions before putting them into action. This essentially allows an organization to verify its analysis is correct before committing resources to implement a broader strategy.
As one example, consider a company that wishes to launch a new marketing campaign to revitalize sales during a slow period. Doing so could be an incredibly expensive endeavor, depending on the campaign’s size and complexity. The company, therefore, may wish to test the campaign on a smaller scale to understand how it will perform.
In this example, the hypothesis that’s being tested would fall along the lines of: “If the company launches a new marketing campaign, then it will translate into an increase in sales.” It may even be possible to quantify how much of a lift in sales the company expects to see from the effort. Pending the results of the pilot campaign, the business would then know whether it makes sense to roll it out more broadly.
Related: 9 Fundamental Data Science Skills for Business Professionals
1. alternative hypothesis and null hypothesis.
In hypothesis testing, the hypothesis that’s being tested is known as the alternative hypothesis . Often, it’s expressed as a correlation or statistical relationship between variables. The null hypothesis , on the other hand, is a statement that’s meant to show there’s no statistical relationship between the variables being tested. It’s typically the exact opposite of whatever is stated in the alternative hypothesis.
For example, consider a company’s leadership team that historically and reliably sees $12 million in monthly revenue. They want to understand if reducing the price of their services will attract more customers and, in turn, increase revenue.
In this case, the alternative hypothesis may take the form of a statement such as: “If we reduce the price of our flagship service by five percent, then we’ll see an increase in sales and realize revenues greater than $12 million in the next month.”
The null hypothesis, on the other hand, would indicate that revenues wouldn’t increase from the base of $12 million, or might even decrease.
Check out the video below about the difference between an alternative and a null hypothesis, and subscribe to our YouTube channel for more explainer content.
Statistically speaking, if you were to run the same scenario 100 times, you’d likely receive somewhat different results each time. If you were to plot these results in a distribution plot, you’d see the most likely outcome is at the tallest point in the graph, with less likely outcomes falling to the right and left of that point.
With this in mind, imagine you’ve completed your hypothesis test and have your results, which indicate there may be a correlation between the variables you were testing. To understand your results' significance, you’ll need to identify a p-value for the test, which helps note how confident you are in the test results.
In statistics, the p-value depicts the probability that, assuming the null hypothesis is correct, you might still observe results that are at least as extreme as the results of your hypothesis test. The smaller the p-value, the more likely the alternative hypothesis is correct, and the greater the significance of your results.
When it’s time to test your hypothesis, it’s important to leverage the correct testing method. The two most common hypothesis testing methods are one-sided and two-sided tests , or one-tailed and two-tailed tests, respectively.
Typically, you’d leverage a one-sided test when you have a strong conviction about the direction of change you expect to see due to your hypothesis test. You’d leverage a two-sided test when you’re less confident in the direction of change.
To perform hypothesis testing in the first place, you need to collect a sample of data to be analyzed. Depending on the question you’re seeking to answer or investigate, you might collect samples through surveys, observational studies, or experiments.
A survey involves asking a series of questions to a random population sample and recording self-reported responses.
Observational studies involve a researcher observing a sample population and collecting data as it occurs naturally, without intervention.
Finally, an experiment involves dividing a sample into multiple groups, one of which acts as the control group. For each non-control group, the variable being studied is manipulated to determine how the data collected differs from that of the control group.
Hypothesis testing is a complex process involving different moving pieces that can allow an organization to effectively leverage its data and inform strategic decisions.
If you’re interested in better understanding hypothesis testing and the role it can play within your organization, one option is to complete a course that focuses on the process. Doing so can lay the statistical and analytical foundation you need to succeed.
Do you want to learn more about hypothesis testing? Explore Business Analytics —one of our online business essentials courses —and download our Beginner’s Guide to Data & Analytics .
Hypothesis testing is the act of testing a hypothesis or a supposition in relation to a statistical parameter. Analysts implement hypothesis testing in order to test if a hypothesis is plausible or not.
In data science and statistics , hypothesis testing is an important step as it involves the verification of an assumption that could help develop a statistical parameter. For instance, a researcher establishes a hypothesis assuming that the average of all odd numbers is an even number.
In order to find the plausibility of this hypothesis, the researcher will have to test the hypothesis using hypothesis testing methods. Unlike a hypothesis that is ‘supposed’ to stand true on the basis of little or no evidence, hypothesis testing is required to have plausible evidence in order to establish that a statistical hypothesis is true.
Perhaps this is where statistics play an important role. A number of components are involved in this process. But before understanding the process involved in hypothesis testing in research methodology, we shall first understand the types of hypotheses that are involved in the process. Let us get started!
In data sampling, different types of hypothesis are involved in finding whether the tested samples test positive for a hypothesis or not. In this segment, we shall discover the different types of hypotheses and understand the role they play in hypothesis testing.
Alternative Hypothesis (H1) or the research hypothesis states that there is a relationship between two variables (where one variable affects the other). The alternative hypothesis is the main driving force for hypothesis testing.
It implies that the two variables are related to each other and the relationship that exists between them is not due to chance or coincidence.
When the process of hypothesis testing is carried out, the alternative hypothesis is the main subject of the testing process. The analyst intends to test the alternative hypothesis and verifies its plausibility.
The Null Hypothesis (H0) aims to nullify the alternative hypothesis by implying that there exists no relation between two variables in statistics. It states that the effect of one variable on the other is solely due to chance and no empirical cause lies behind it.
The null hypothesis is established alongside the alternative hypothesis and is recognized as important as the latter. In hypothesis testing, the null hypothesis has a major role to play as it influences the testing against the alternative hypothesis.
(Must read: What is ANOVA test? )
The Non-directional hypothesis states that the relation between two variables has no direction.
Simply put, it asserts that there exists a relation between two variables, but does not recognize the direction of effect, whether variable A affects variable B or vice versa.
The Directional hypothesis, on the other hand, asserts the direction of effect of the relationship that exists between two variables.
Herein, the hypothesis clearly states that variable A affects variable B, or vice versa.
A statistical hypothesis is a hypothesis that can be verified to be plausible on the basis of statistics.
By using data sampling and statistical knowledge, one can determine the plausibility of a statistical hypothesis and find out if it stands true or not.
(Related blog: z-test vs t-test )
Now that we have understood the types of hypotheses and the role they play in hypothesis testing, let us now move on to understand the process in a better manner.
In hypothesis testing, a researcher is first required to establish two hypotheses - alternative hypothesis and null hypothesis in order to begin with the procedure.
To establish these two hypotheses, one is required to study data samples, find a plausible pattern among the samples, and pen down a statistical hypothesis that they wish to test.
A random population of samples can be drawn, to begin with hypothesis testing. Among the two hypotheses, alternative and null, only one can be verified to be true. Perhaps the presence of both hypotheses is required to make the process successful.
At the end of the hypothesis testing procedure, either of the hypotheses will be rejected and the other one will be supported. Even though one of the two hypotheses turns out to be true, no hypothesis can ever be verified 100%.
(Read also: Types of data sampling techniques )
Therefore, a hypothesis can only be supported based on the statistical samples and verified data. Here is a step-by-step guide for hypothesis testing.
First things first, one is required to establish two hypotheses - alternative and null, that will set the foundation for hypothesis testing.
These hypotheses initiate the testing process that involves the researcher working on data samples in order to either support the alternative hypothesis or the null hypothesis.
Once the hypotheses have been formulated, it is now time to generate a testing plan. A testing plan or an analysis plan involves the accumulation of data samples, determining which statistic is to be considered and laying out the sample size.
All these factors are very important while one is working on hypothesis testing.
As soon as a testing plan is ready, it is time to move on to the analysis part. Analysis of data samples involves configuring statistical values of samples, drawing them together, and deriving a pattern out of these samples.
While analyzing the data samples, a researcher needs to determine a set of things -
Significance Level - The level of significance in hypothesis testing indicates if a statistical result could have significance if the null hypothesis stands to be true.
Testing Method - The testing method involves a type of sampling-distribution and a test statistic that leads to hypothesis testing. There are a number of testing methods that can assist in the analysis of data samples.
Test statistic - Test statistic is a numerical summary of a data set that can be used to perform hypothesis testing.
P-value - The P-value interpretation is the probability of finding a sample statistic to be as extreme as the test statistic, indicating the plausibility of the null hypothesis.
The analysis of data samples leads to the inference of results that establishes whether the alternative hypothesis stands true or not. When the P-value is less than the significance level, the null hypothesis is rejected and the alternative hypothesis turns out to be plausible.
As we have already looked into different aspects of hypothesis testing, we shall now look into the different methods of hypothesis testing. All in all, there are 2 most common types of hypothesis testing methods. They are as follows -
The frequentist hypothesis or the traditional approach to hypothesis testing is a hypothesis testing method that aims on making assumptions by considering current data.
The supposed truths and assumptions are based on the current data and a set of 2 hypotheses are formulated. A very popular subtype of the frequentist approach is the Null Hypothesis Significance Testing (NHST).
The NHST approach (involving the null and alternative hypothesis) has been one of the most sought-after methods of hypothesis testing in the field of statistics ever since its inception in the mid-1950s.
A much unconventional and modern method of hypothesis testing, the Bayesian Hypothesis Testing claims to test a particular hypothesis in accordance with the past data samples, known as prior probability, and current data that lead to the plausibility of a hypothesis.
The result obtained indicates the posterior probability of the hypothesis. In this method, the researcher relies on ‘prior probability and posterior probability’ to conduct hypothesis testing on hand.
On the basis of this prior probability, the Bayesian approach tests a hypothesis to be true or false. The Bayes factor, a major component of this method, indicates the likelihood ratio among the null hypothesis and the alternative hypothesis.
The Bayes factor is the indicator of the plausibility of either of the two hypotheses that are established for hypothesis testing.
(Also read - Introduction to Bayesian Statistics )
To conclude, hypothesis testing, a way to verify the plausibility of a supposed assumption can be done through different methods - the Bayesian approach or the Frequentist approach.
Although the Bayesian approach relies on the prior probability of data samples, the frequentist approach assumes without a probability. A number of elements involved in hypothesis testing are - significance level, p-level, test statistic, and method of hypothesis testing.
(Also read: Introduction to probability distributions )
A significant way to determine whether a hypothesis stands true or not is to verify the data samples and identify the plausible hypothesis among the null hypothesis and alternative hypothesis.
Be a part of our Instagram community
5 Factors Influencing Consumer Behavior
Elasticity of Demand and its Types
An Overview of Descriptive Analysis
What is PESTLE Analysis? Everything you need to know about it
What is Managerial Economics? Definition, Types, Nature, Principles, and Scope
5 Factors Affecting the Price Elasticity of Demand (PED)
6 Major Branches of Artificial Intelligence (AI)
Scope of Managerial Economics
Dijkstra’s Algorithm: The Shortest Path Algorithm
Different Types of Research Methods
A hypothesis test is exactly what it sounds like: You make a hypothesis about the parameters of a population, and the test determines whether your hypothesis is consistent with your sample data.
The p-value of a hypothesis test is the probability that your sample data would have occurred if you hypothesis were not correct. Traditionally, researchers have used a p-value of 0.05 (a 5% probability that your sample data would have occurred if your hypothesis was wrong) as the threshold for declaring that a hypothesis is true. But there is a long history of debate and controversy over p-values and significance levels.
Many of the most commonly used hypothesis tests rely on assumptions about your sample data—for instance, that it is continuous, and that its parameters follow a Normal distribution. Nonparametric hypothesis tests don't make any assumptions about the distribution of the data, and many can be used on categorical data.
An official website of the United States government
The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.
The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.
Preview improvements coming to the PMC website in October 2024. Learn More or Try it out now .
Edward barroga.
1 Department of General Education, Graduate School of Nursing Science, St. Luke’s International University, Tokyo, Japan.
2 Department of Biological Sciences, Messiah University, Mechanicsburg, PA, USA.
The development of research questions and the subsequent hypotheses are prerequisites to defining the main research purpose and specific objectives of a study. Consequently, these objectives determine the study design and research outcome. The development of research questions is a process based on knowledge of current trends, cutting-edge studies, and technological advances in the research field. Excellent research questions are focused and require a comprehensive literature search and in-depth understanding of the problem being investigated. Initially, research questions may be written as descriptive questions which could be developed into inferential questions. These questions must be specific and concise to provide a clear foundation for developing hypotheses. Hypotheses are more formal predictions about the research outcomes. These specify the possible results that may or may not be expected regarding the relationship between groups. Thus, research questions and hypotheses clarify the main purpose and specific objectives of the study, which in turn dictate the design of the study, its direction, and outcome. Studies developed from good research questions and hypotheses will have trustworthy outcomes with wide-ranging social and health implications.
Scientific research is usually initiated by posing evidenced-based research questions which are then explicitly restated as hypotheses. 1 , 2 The hypotheses provide directions to guide the study, solutions, explanations, and expected results. 3 , 4 Both research questions and hypotheses are essentially formulated based on conventional theories and real-world processes, which allow the inception of novel studies and the ethical testing of ideas. 5 , 6
It is crucial to have knowledge of both quantitative and qualitative research 2 as both types of research involve writing research questions and hypotheses. 7 However, these crucial elements of research are sometimes overlooked; if not overlooked, then framed without the forethought and meticulous attention it needs. Planning and careful consideration are needed when developing quantitative or qualitative research, particularly when conceptualizing research questions and hypotheses. 4
There is a continuing need to support researchers in the creation of innovative research questions and hypotheses, as well as for journal articles that carefully review these elements. 1 When research questions and hypotheses are not carefully thought of, unethical studies and poor outcomes usually ensue. Carefully formulated research questions and hypotheses define well-founded objectives, which in turn determine the appropriate design, course, and outcome of the study. This article then aims to discuss in detail the various aspects of crafting research questions and hypotheses, with the goal of guiding researchers as they develop their own. Examples from the authors and peer-reviewed scientific articles in the healthcare field are provided to illustrate key points.
A research question is what a study aims to answer after data analysis and interpretation. The answer is written in length in the discussion section of the paper. Thus, the research question gives a preview of the different parts and variables of the study meant to address the problem posed in the research question. 1 An excellent research question clarifies the research writing while facilitating understanding of the research topic, objective, scope, and limitations of the study. 5
On the other hand, a research hypothesis is an educated statement of an expected outcome. This statement is based on background research and current knowledge. 8 , 9 The research hypothesis makes a specific prediction about a new phenomenon 10 or a formal statement on the expected relationship between an independent variable and a dependent variable. 3 , 11 It provides a tentative answer to the research question to be tested or explored. 4
Hypotheses employ reasoning to predict a theory-based outcome. 10 These can also be developed from theories by focusing on components of theories that have not yet been observed. 10 The validity of hypotheses is often based on the testability of the prediction made in a reproducible experiment. 8
Conversely, hypotheses can also be rephrased as research questions. Several hypotheses based on existing theories and knowledge may be needed to answer a research question. Developing ethical research questions and hypotheses creates a research design that has logical relationships among variables. These relationships serve as a solid foundation for the conduct of the study. 4 , 11 Haphazardly constructed research questions can result in poorly formulated hypotheses and improper study designs, leading to unreliable results. Thus, the formulations of relevant research questions and verifiable hypotheses are crucial when beginning research. 12
Excellent research questions are specific and focused. These integrate collective data and observations to confirm or refute the subsequent hypotheses. Well-constructed hypotheses are based on previous reports and verify the research context. These are realistic, in-depth, sufficiently complex, and reproducible. More importantly, these hypotheses can be addressed and tested. 13
There are several characteristics of well-developed hypotheses. Good hypotheses are 1) empirically testable 7 , 10 , 11 , 13 ; 2) backed by preliminary evidence 9 ; 3) testable by ethical research 7 , 9 ; 4) based on original ideas 9 ; 5) have evidenced-based logical reasoning 10 ; and 6) can be predicted. 11 Good hypotheses can infer ethical and positive implications, indicating the presence of a relationship or effect relevant to the research theme. 7 , 11 These are initially developed from a general theory and branch into specific hypotheses by deductive reasoning. In the absence of a theory to base the hypotheses, inductive reasoning based on specific observations or findings form more general hypotheses. 10
Research questions and hypotheses are developed according to the type of research, which can be broadly classified into quantitative and qualitative research. We provide a summary of the types of research questions and hypotheses under quantitative and qualitative research categories in Table 1 .
Quantitative research questions | Quantitative research hypotheses |
---|---|
Descriptive research questions | Simple hypothesis |
Comparative research questions | Complex hypothesis |
Relationship research questions | Directional hypothesis |
Non-directional hypothesis | |
Associative hypothesis | |
Causal hypothesis | |
Null hypothesis | |
Alternative hypothesis | |
Working hypothesis | |
Statistical hypothesis | |
Logical hypothesis | |
Hypothesis-testing | |
Qualitative research questions | Qualitative research hypotheses |
Contextual research questions | Hypothesis-generating |
Descriptive research questions | |
Evaluation research questions | |
Explanatory research questions | |
Exploratory research questions | |
Generative research questions | |
Ideological research questions | |
Ethnographic research questions | |
Phenomenological research questions | |
Grounded theory questions | |
Qualitative case study questions |
In quantitative research, research questions inquire about the relationships among variables being investigated and are usually framed at the start of the study. These are precise and typically linked to the subject population, dependent and independent variables, and research design. 1 Research questions may also attempt to describe the behavior of a population in relation to one or more variables, or describe the characteristics of variables to be measured ( descriptive research questions ). 1 , 5 , 14 These questions may also aim to discover differences between groups within the context of an outcome variable ( comparative research questions ), 1 , 5 , 14 or elucidate trends and interactions among variables ( relationship research questions ). 1 , 5 We provide examples of descriptive, comparative, and relationship research questions in quantitative research in Table 2 .
Quantitative research questions | |
---|---|
Descriptive research question | |
- Measures responses of subjects to variables | |
- Presents variables to measure, analyze, or assess | |
What is the proportion of resident doctors in the hospital who have mastered ultrasonography (response of subjects to a variable) as a diagnostic technique in their clinical training? | |
Comparative research question | |
- Clarifies difference between one group with outcome variable and another group without outcome variable | |
Is there a difference in the reduction of lung metastasis in osteosarcoma patients who received the vitamin D adjunctive therapy (group with outcome variable) compared with osteosarcoma patients who did not receive the vitamin D adjunctive therapy (group without outcome variable)? | |
- Compares the effects of variables | |
How does the vitamin D analogue 22-Oxacalcitriol (variable 1) mimic the antiproliferative activity of 1,25-Dihydroxyvitamin D (variable 2) in osteosarcoma cells? | |
Relationship research question | |
- Defines trends, association, relationships, or interactions between dependent variable and independent variable | |
Is there a relationship between the number of medical student suicide (dependent variable) and the level of medical student stress (independent variable) in Japan during the first wave of the COVID-19 pandemic? |
In quantitative research, hypotheses predict the expected relationships among variables. 15 Relationships among variables that can be predicted include 1) between a single dependent variable and a single independent variable ( simple hypothesis ) or 2) between two or more independent and dependent variables ( complex hypothesis ). 4 , 11 Hypotheses may also specify the expected direction to be followed and imply an intellectual commitment to a particular outcome ( directional hypothesis ) 4 . On the other hand, hypotheses may not predict the exact direction and are used in the absence of a theory, or when findings contradict previous studies ( non-directional hypothesis ). 4 In addition, hypotheses can 1) define interdependency between variables ( associative hypothesis ), 4 2) propose an effect on the dependent variable from manipulation of the independent variable ( causal hypothesis ), 4 3) state a negative relationship between two variables ( null hypothesis ), 4 , 11 , 15 4) replace the working hypothesis if rejected ( alternative hypothesis ), 15 explain the relationship of phenomena to possibly generate a theory ( working hypothesis ), 11 5) involve quantifiable variables that can be tested statistically ( statistical hypothesis ), 11 6) or express a relationship whose interlinks can be verified logically ( logical hypothesis ). 11 We provide examples of simple, complex, directional, non-directional, associative, causal, null, alternative, working, statistical, and logical hypotheses in quantitative research, as well as the definition of quantitative hypothesis-testing research in Table 3 .
Quantitative research hypotheses | |
---|---|
Simple hypothesis | |
- Predicts relationship between single dependent variable and single independent variable | |
If the dose of the new medication (single independent variable) is high, blood pressure (single dependent variable) is lowered. | |
Complex hypothesis | |
- Foretells relationship between two or more independent and dependent variables | |
The higher the use of anticancer drugs, radiation therapy, and adjunctive agents (3 independent variables), the higher would be the survival rate (1 dependent variable). | |
Directional hypothesis | |
- Identifies study direction based on theory towards particular outcome to clarify relationship between variables | |
Privately funded research projects will have a larger international scope (study direction) than publicly funded research projects. | |
Non-directional hypothesis | |
- Nature of relationship between two variables or exact study direction is not identified | |
- Does not involve a theory | |
Women and men are different in terms of helpfulness. (Exact study direction is not identified) | |
Associative hypothesis | |
- Describes variable interdependency | |
- Change in one variable causes change in another variable | |
A larger number of people vaccinated against COVID-19 in the region (change in independent variable) will reduce the region’s incidence of COVID-19 infection (change in dependent variable). | |
Causal hypothesis | |
- An effect on dependent variable is predicted from manipulation of independent variable | |
A change into a high-fiber diet (independent variable) will reduce the blood sugar level (dependent variable) of the patient. | |
Null hypothesis | |
- A negative statement indicating no relationship or difference between 2 variables | |
There is no significant difference in the severity of pulmonary metastases between the new drug (variable 1) and the current drug (variable 2). | |
Alternative hypothesis | |
- Following a null hypothesis, an alternative hypothesis predicts a relationship between 2 study variables | |
The new drug (variable 1) is better on average in reducing the level of pain from pulmonary metastasis than the current drug (variable 2). | |
Working hypothesis | |
- A hypothesis that is initially accepted for further research to produce a feasible theory | |
Dairy cows fed with concentrates of different formulations will produce different amounts of milk. | |
Statistical hypothesis | |
- Assumption about the value of population parameter or relationship among several population characteristics | |
- Validity tested by a statistical experiment or analysis | |
The mean recovery rate from COVID-19 infection (value of population parameter) is not significantly different between population 1 and population 2. | |
There is a positive correlation between the level of stress at the workplace and the number of suicides (population characteristics) among working people in Japan. | |
Logical hypothesis | |
- Offers or proposes an explanation with limited or no extensive evidence | |
If healthcare workers provide more educational programs about contraception methods, the number of adolescent pregnancies will be less. | |
Hypothesis-testing (Quantitative hypothesis-testing research) | |
- Quantitative research uses deductive reasoning. | |
- This involves the formation of a hypothesis, collection of data in the investigation of the problem, analysis and use of the data from the investigation, and drawing of conclusions to validate or nullify the hypotheses. |
Unlike research questions in quantitative research, research questions in qualitative research are usually continuously reviewed and reformulated. The central question and associated subquestions are stated more than the hypotheses. 15 The central question broadly explores a complex set of factors surrounding the central phenomenon, aiming to present the varied perspectives of participants. 15
There are varied goals for which qualitative research questions are developed. These questions can function in several ways, such as to 1) identify and describe existing conditions ( contextual research question s); 2) describe a phenomenon ( descriptive research questions ); 3) assess the effectiveness of existing methods, protocols, theories, or procedures ( evaluation research questions ); 4) examine a phenomenon or analyze the reasons or relationships between subjects or phenomena ( explanatory research questions ); or 5) focus on unknown aspects of a particular topic ( exploratory research questions ). 5 In addition, some qualitative research questions provide new ideas for the development of theories and actions ( generative research questions ) or advance specific ideologies of a position ( ideological research questions ). 1 Other qualitative research questions may build on a body of existing literature and become working guidelines ( ethnographic research questions ). Research questions may also be broadly stated without specific reference to the existing literature or a typology of questions ( phenomenological research questions ), may be directed towards generating a theory of some process ( grounded theory questions ), or may address a description of the case and the emerging themes ( qualitative case study questions ). 15 We provide examples of contextual, descriptive, evaluation, explanatory, exploratory, generative, ideological, ethnographic, phenomenological, grounded theory, and qualitative case study research questions in qualitative research in Table 4 , and the definition of qualitative hypothesis-generating research in Table 5 .
Qualitative research questions | |
---|---|
Contextual research question | |
- Ask the nature of what already exists | |
- Individuals or groups function to further clarify and understand the natural context of real-world problems | |
What are the experiences of nurses working night shifts in healthcare during the COVID-19 pandemic? (natural context of real-world problems) | |
Descriptive research question | |
- Aims to describe a phenomenon | |
What are the different forms of disrespect and abuse (phenomenon) experienced by Tanzanian women when giving birth in healthcare facilities? | |
Evaluation research question | |
- Examines the effectiveness of existing practice or accepted frameworks | |
How effective are decision aids (effectiveness of existing practice) in helping decide whether to give birth at home or in a healthcare facility? | |
Explanatory research question | |
- Clarifies a previously studied phenomenon and explains why it occurs | |
Why is there an increase in teenage pregnancy (phenomenon) in Tanzania? | |
Exploratory research question | |
- Explores areas that have not been fully investigated to have a deeper understanding of the research problem | |
What factors affect the mental health of medical students (areas that have not yet been fully investigated) during the COVID-19 pandemic? | |
Generative research question | |
- Develops an in-depth understanding of people’s behavior by asking ‘how would’ or ‘what if’ to identify problems and find solutions | |
How would the extensive research experience of the behavior of new staff impact the success of the novel drug initiative? | |
Ideological research question | |
- Aims to advance specific ideas or ideologies of a position | |
Are Japanese nurses who volunteer in remote African hospitals able to promote humanized care of patients (specific ideas or ideologies) in the areas of safe patient environment, respect of patient privacy, and provision of accurate information related to health and care? | |
Ethnographic research question | |
- Clarifies peoples’ nature, activities, their interactions, and the outcomes of their actions in specific settings | |
What are the demographic characteristics, rehabilitative treatments, community interactions, and disease outcomes (nature, activities, their interactions, and the outcomes) of people in China who are suffering from pneumoconiosis? | |
Phenomenological research question | |
- Knows more about the phenomena that have impacted an individual | |
What are the lived experiences of parents who have been living with and caring for children with a diagnosis of autism? (phenomena that have impacted an individual) | |
Grounded theory question | |
- Focuses on social processes asking about what happens and how people interact, or uncovering social relationships and behaviors of groups | |
What are the problems that pregnant adolescents face in terms of social and cultural norms (social processes), and how can these be addressed? | |
Qualitative case study question | |
- Assesses a phenomenon using different sources of data to answer “why” and “how” questions | |
- Considers how the phenomenon is influenced by its contextual situation. | |
How does quitting work and assuming the role of a full-time mother (phenomenon assessed) change the lives of women in Japan? |
Qualitative research hypotheses | |
---|---|
Hypothesis-generating (Qualitative hypothesis-generating research) | |
- Qualitative research uses inductive reasoning. | |
- This involves data collection from study participants or the literature regarding a phenomenon of interest, using the collected data to develop a formal hypothesis, and using the formal hypothesis as a framework for testing the hypothesis. | |
- Qualitative exploratory studies explore areas deeper, clarifying subjective experience and allowing formulation of a formal hypothesis potentially testable in a future quantitative approach. |
Qualitative studies usually pose at least one central research question and several subquestions starting with How or What . These research questions use exploratory verbs such as explore or describe . These also focus on one central phenomenon of interest, and may mention the participants and research site. 15
Hypotheses in qualitative research are stated in the form of a clear statement concerning the problem to be investigated. Unlike in quantitative research where hypotheses are usually developed to be tested, qualitative research can lead to both hypothesis-testing and hypothesis-generating outcomes. 2 When studies require both quantitative and qualitative research questions, this suggests an integrative process between both research methods wherein a single mixed-methods research question can be developed. 1
Research questions followed by hypotheses should be developed before the start of the study. 1 , 12 , 14 It is crucial to develop feasible research questions on a topic that is interesting to both the researcher and the scientific community. This can be achieved by a meticulous review of previous and current studies to establish a novel topic. Specific areas are subsequently focused on to generate ethical research questions. The relevance of the research questions is evaluated in terms of clarity of the resulting data, specificity of the methodology, objectivity of the outcome, depth of the research, and impact of the study. 1 , 5 These aspects constitute the FINER criteria (i.e., Feasible, Interesting, Novel, Ethical, and Relevant). 1 Clarity and effectiveness are achieved if research questions meet the FINER criteria. In addition to the FINER criteria, Ratan et al. described focus, complexity, novelty, feasibility, and measurability for evaluating the effectiveness of research questions. 14
The PICOT and PEO frameworks are also used when developing research questions. 1 The following elements are addressed in these frameworks, PICOT: P-population/patients/problem, I-intervention or indicator being studied, C-comparison group, O-outcome of interest, and T-timeframe of the study; PEO: P-population being studied, E-exposure to preexisting conditions, and O-outcome of interest. 1 Research questions are also considered good if these meet the “FINERMAPS” framework: Feasible, Interesting, Novel, Ethical, Relevant, Manageable, Appropriate, Potential value/publishable, and Systematic. 14
As we indicated earlier, research questions and hypotheses that are not carefully formulated result in unethical studies or poor outcomes. To illustrate this, we provide some examples of ambiguous research question and hypotheses that result in unclear and weak research objectives in quantitative research ( Table 6 ) 16 and qualitative research ( Table 7 ) 17 , and how to transform these ambiguous research question(s) and hypothesis(es) into clear and good statements.
Variables | Unclear and weak statement (Statement 1) | Clear and good statement (Statement 2) | Points to avoid |
---|---|---|---|
Research question | Which is more effective between smoke moxibustion and smokeless moxibustion? | “Moreover, regarding smoke moxibustion versus smokeless moxibustion, it remains unclear which is more effective, safe, and acceptable to pregnant women, and whether there is any difference in the amount of heat generated.” | 1) Vague and unfocused questions |
2) Closed questions simply answerable by yes or no | |||
3) Questions requiring a simple choice | |||
Hypothesis | The smoke moxibustion group will have higher cephalic presentation. | “Hypothesis 1. The smoke moxibustion stick group (SM group) and smokeless moxibustion stick group (-SLM group) will have higher rates of cephalic presentation after treatment than the control group. | 1) Unverifiable hypotheses |
Hypothesis 2. The SM group and SLM group will have higher rates of cephalic presentation at birth than the control group. | 2) Incompletely stated groups of comparison | ||
Hypothesis 3. There will be no significant differences in the well-being of the mother and child among the three groups in terms of the following outcomes: premature birth, premature rupture of membranes (PROM) at < 37 weeks, Apgar score < 7 at 5 min, umbilical cord blood pH < 7.1, admission to neonatal intensive care unit (NICU), and intrauterine fetal death.” | 3) Insufficiently described variables or outcomes | ||
Research objective | To determine which is more effective between smoke moxibustion and smokeless moxibustion. | “The specific aims of this pilot study were (a) to compare the effects of smoke moxibustion and smokeless moxibustion treatments with the control group as a possible supplement to ECV for converting breech presentation to cephalic presentation and increasing adherence to the newly obtained cephalic position, and (b) to assess the effects of these treatments on the well-being of the mother and child.” | 1) Poor understanding of the research question and hypotheses |
2) Insufficient description of population, variables, or study outcomes |
a These statements were composed for comparison and illustrative purposes only.
b These statements are direct quotes from Higashihara and Horiuchi. 16
Variables | Unclear and weak statement (Statement 1) | Clear and good statement (Statement 2) | Points to avoid |
---|---|---|---|
Research question | Does disrespect and abuse (D&A) occur in childbirth in Tanzania? | How does disrespect and abuse (D&A) occur and what are the types of physical and psychological abuses observed in midwives’ actual care during facility-based childbirth in urban Tanzania? | 1) Ambiguous or oversimplistic questions |
2) Questions unverifiable by data collection and analysis | |||
Hypothesis | Disrespect and abuse (D&A) occur in childbirth in Tanzania. | Hypothesis 1: Several types of physical and psychological abuse by midwives in actual care occur during facility-based childbirth in urban Tanzania. | 1) Statements simply expressing facts |
Hypothesis 2: Weak nursing and midwifery management contribute to the D&A of women during facility-based childbirth in urban Tanzania. | 2) Insufficiently described concepts or variables | ||
Research objective | To describe disrespect and abuse (D&A) in childbirth in Tanzania. | “This study aimed to describe from actual observations the respectful and disrespectful care received by women from midwives during their labor period in two hospitals in urban Tanzania.” | 1) Statements unrelated to the research question and hypotheses |
2) Unattainable or unexplorable objectives |
a This statement is a direct quote from Shimoda et al. 17
The other statements were composed for comparison and illustrative purposes only.
To construct effective research questions and hypotheses, it is very important to 1) clarify the background and 2) identify the research problem at the outset of the research, within a specific timeframe. 9 Then, 3) review or conduct preliminary research to collect all available knowledge about the possible research questions by studying theories and previous studies. 18 Afterwards, 4) construct research questions to investigate the research problem. Identify variables to be accessed from the research questions 4 and make operational definitions of constructs from the research problem and questions. Thereafter, 5) construct specific deductive or inductive predictions in the form of hypotheses. 4 Finally, 6) state the study aims . This general flow for constructing effective research questions and hypotheses prior to conducting research is shown in Fig. 1 .
Research questions are used more frequently in qualitative research than objectives or hypotheses. 3 These questions seek to discover, understand, explore or describe experiences by asking “What” or “How.” The questions are open-ended to elicit a description rather than to relate variables or compare groups. The questions are continually reviewed, reformulated, and changed during the qualitative study. 3 Research questions are also used more frequently in survey projects than hypotheses in experiments in quantitative research to compare variables and their relationships.
Hypotheses are constructed based on the variables identified and as an if-then statement, following the template, ‘If a specific action is taken, then a certain outcome is expected.’ At this stage, some ideas regarding expectations from the research to be conducted must be drawn. 18 Then, the variables to be manipulated (independent) and influenced (dependent) are defined. 4 Thereafter, the hypothesis is stated and refined, and reproducible data tailored to the hypothesis are identified, collected, and analyzed. 4 The hypotheses must be testable and specific, 18 and should describe the variables and their relationships, the specific group being studied, and the predicted research outcome. 18 Hypotheses construction involves a testable proposition to be deduced from theory, and independent and dependent variables to be separated and measured separately. 3 Therefore, good hypotheses must be based on good research questions constructed at the start of a study or trial. 12
In summary, research questions are constructed after establishing the background of the study. Hypotheses are then developed based on the research questions. Thus, it is crucial to have excellent research questions to generate superior hypotheses. In turn, these would determine the research objectives and the design of the study, and ultimately, the outcome of the research. 12 Algorithms for building research questions and hypotheses are shown in Fig. 2 for quantitative research and in Fig. 3 for qualitative research.
Research questions and hypotheses are crucial components to any type of research, whether quantitative or qualitative. These questions should be developed at the very beginning of the study. Excellent research questions lead to superior hypotheses, which, like a compass, set the direction of research, and can often determine the successful conduct of the study. Many research studies have floundered because the development of research questions and subsequent hypotheses was not given the thought and meticulous attention needed. The development of research questions and hypotheses is an iterative process based on extensive knowledge of the literature and insightful grasp of the knowledge gap. Focused, concise, and specific research questions provide a strong foundation for constructing hypotheses which serve as formal predictions about the research outcomes. Research questions and hypotheses are crucial elements of research that should not be overlooked. They should be carefully thought of and constructed when planning research. This avoids unethical studies and poor outcomes by defining well-founded objectives that determine the design, course, and outcome of the study.
Disclosure: The authors have no potential conflicts of interest to disclose.
Author Contributions:
When interpreting research findings, researchers need to assess whether these findings may have occurred by chance. Hypothesis testing is a systematic procedure for deciding whether the results of a research study support a particular theory which applies to a population.
Hypothesis testing uses sample data to evaluate a hypothesis about a population . A hypothesis test assesses how unusual the result is, whether it is reasonable chance variation or whether the result is too extreme to be considered chance variation.
Effect size and statistical significance.
To carry out statistical hypothesis testing, research and null hypothesis are employed:
H A: There is a relationship between intelligence and academic results.
H A: First year university students obtain higher grades after an intensive Statistics course.
H A; Males and females differ in their levels of stress.
H o : There is no relationship between intelligence and academic results.
H o: First year university students do not obtain higher grades after an intensive Statistics course.
H o : Males and females will not differ in their levels of stress.
The purpose of hypothesis testing is to test whether the null hypothesis (there is no difference, no effect) can be rejected or approved. If the null hypothesis is rejected, then the research hypothesis can be accepted. If the null hypothesis is accepted, then the research hypothesis is rejected.
In hypothesis testing, a value is set to assess whether the null hypothesis is accepted or rejected and whether the result is statistically significant:
The probability value, or p value , is the probability of an outcome or research result given the hypothesis. Usually, the probability value is set at 0.05: the null hypothesis will be rejected if the probability value of the statistical test is less than 0.05. There are two types of errors associated to hypothesis testing:
These situations are known as Type I and Type II errors:
These errors cannot be eliminated; they can be minimised, but minimising one type of error will increase the probability of committing the other type.
The probability of making a Type I error depends on the criterion that is used to accept or reject the null hypothesis: the p value or alpha level . The alpha is set by the researcher, usually at .05, and is the chance the researcher is willing to take and still claim the significance of the statistical test.). Choosing a smaller alpha level will decrease the likelihood of committing Type I error.
For example, p<0.05 indicates that there are 5 chances in 100 that the difference observed was really due to sampling error – that 5% of the time a Type I error will occur or that there is a 5% chance that the opposite of the null hypothesis is actually true.
With a p<0.01, there will be 1 chance in 100 that the difference observed was really due to sampling error – 1% of the time a Type I error will occur.
The p level is specified before analysing the data. If the data analysis results in a probability value below the α (alpha) level, then the null hypothesis is rejected; if it is not, then the null hypothesis is not rejected.
When the null hypothesis is rejected, the effect is said to be statistically significant. However, statistical significance does not mean that the effect is important.
A result can be statistically significant, but the effect size may be small. Finding that an effect is significant does not provide information about how large or important the effect is. In fact, a small effect can be statistically significant if the sample size is large enough.
Information about the effect size, or magnitude of the result, is given by the statistical test. For example, the strength of the correlation between two variables is given by the coefficient of correlation, which varies from 0 to 1.
The hypothesis testing process can be divided into five steps:
This example illustrates how these five steps can be applied to text a hypothesis:
Step 1 : There are two populations of interest.
Population 1: People who go through the experimental procedure (drink coffee).
Population 2: People who do not go through the experimental procedure (drink water).
Step 2 : We know that the characteristics of the comparison distribution (student population) are:
Population M = 19, Population SD= 4, normally distributed. These are the mean and standard deviation of the distribution of scores on the memory test for the general student population.
Step 3 : For a two-tailed test (the direction of the effect is not specified) at the 5% level (25% at each tail), the cut off sample scores are +1.96 and -1.99.
Step 4 : Your sample score of 27 needs to be converted into a Z value. To calculate Z = (27-19)/4= 2 ( check the Converting into Z scores section if you need to review how to do this process)
Step 5 : A ‘Z’ score of 2 is more extreme than the cut off Z of +1.96 (see figure above). The result is significant and, thus, the null hypothesis is rejected.
You can find more examples here:
Correlation analysis, multiple regression.
Correlation analysis explores the association between variables . The purpose of correlational analysis is to discover whether there is a relationship between variables, which is unlikely to occur by sampling error. The null hypothesis is that there is no relationship between the two variables. Correlation analysis provides information about:
A positive correlation indicates that high scores on one variable are associated with high scores on the other variable; low scores on one variable are associated with low scores on the second variable . For instance, in the figure below, higher scores on negative affect are associated with higher scores on perceived stress
A negative correlation indicates that high scores on one variable are associated with low scores on the other variable. The graph shows that a person who scores high on perceived stress will probably score low on mastery. The slope of the graph is downwards- as it moves to the right. In the figure below, higher scores on mastery are associated with lower scores on perceived stress.
Fig 2. Negative correlation between two variables. Adapted from Pallant, J. (2013). SPSS survival manual: A step by step guide to data analysis using IBM SPSS (5th ed.). Sydney, Melbourne, Auckland, London: Allen & Unwin
2. The strength or magnitude of the relationship
The strength of a linear relationship between two variables is measured by a statistic known as the correlation coefficient , which varies from 0 to -1, and from 0 to +1. There are several correlation coefficients; the most widely used are Pearson’s r and Spearman’s rho. The strength of the relationship is interpreted as follows:
It is important to note that correlation analysis does not imply causality. Correlation is used to explore the association between variables, however, it does not indicate that one variable causes the other. The correlation between two variables could be due to the fact that a third variable is affecting the two variables.
Multiple regression is an extension of correlation analysis. Multiple regression is used to explore the relationship between one dependent variable and a number of independent variables or predictors . The purpose of a multiple regression model is to predict values of a dependent variable based on the values of the independent variables or predictors. For example, a researcher may be interested in predicting students’ academic success (e.g. grades) based on a number of predictors, for example, hours spent studying, satisfaction with studies, relationships with peers and lecturers.
A multiple regression model can be conducted using statistical software (e.g. SPSS). The software will test the significance of the model (i.e. does the model significantly predicts scores on the dependent variable using the independent variables introduced in the model?), how much of the variance in the dependent variable is explained by the model, and the individual contribution of each independent variable.
Example of multiple regression model
From Dunn et al. (2014). Influence of academic self-regulation, critical thinking, and age on online graduate students' academic help-seeking.
In this model, help-seeking is the dependent variable; there are three independent variables or predictors. The coefficients show the direction (positive or negative) and magnitude of the relationship between each predictor and the dependent variable. The model was statistically significant and predicted 13.5% of the variance in help-seeking.
t-Tests are employed to compare the mean score on some continuous variable for two groups . The null hypothesis to be tested is there are no differences between the two groups (e.g. anxiety scores for males and females are not different).
If the significance value of the t-test is equal or less than .05, there is a significant difference in the mean scores on the variable of interest for each of the two groups. If the value is above .05, there is no significant difference between the groups.
t-Tests can be employed to compare the mean scores of two different groups (independent-samples t-test ) or to compare the same group of people on two different occasions ( paired-samples t-test) .
In addition to assessing whether the difference between the two groups is statistically significant, it is important to consider the effect size or magnitude of the difference between the groups. The effect size is given by partial eta squared (proportion of variance of the dependent variable that is explained by the independent variable) and Cohen’s d (difference between groups in terms of standard deviation units).
In this example, an independent samples t-test was conducted to assess whether males and females differ in their perceived anxiety levels. The significance of the test is .004. Since this value is less than .05, we can conclude that there is a statistically significant difference between males and females in their perceived anxiety levels.
Whilst t-tests compare the mean score on one variable for two groups, analysis of variance is used to test more than two groups . Following the previous example, analysis of variance would be employed to test whether there are differences in anxiety scores for students from different disciplines.
Analysis of variance compare the variance (variability in scores) between the different groups (believed to be due to the independent variable) with the variability within each group (believed to be due to chance). An F ratio is calculated; a large F ratio indicates that there is more variability between the groups (caused by the independent variable) than there is within each group (error term). A significant F test indicates that we can reject the null hypothesis; i.e. that there is no difference between the groups.
Again, effect size statistics such as Cohen’s d and eta squared are employed to assess the magnitude of the differences between groups.
In this example, we examined differences in perceived anxiety between students from different disciplines. The results of the Anova Test show that the significance level is .005. Since this value is below .05, we can conclude that there are statistically significant differences between students from different disciplines in their perceived anxiety levels.
Chi-square test for independence is used to explore the relationship between two categorical variables. Each variable can have two or more categories.
For example, a researcher can use a Chi-square test for independence to assess the relationship between study disciplines (e.g. Psychology, Business, Education,…) and help-seeking behaviour (Yes/No). The test compares the observed frequencies of cases with the values that would be expected if there was no association between the two variables of interest. A statistically significant Chi-square test indicates that the two variables are associated (e.g. Psychology students are more likely to seek help than Business students). The effect size is assessed using effect size statistics: Phi and Cramer’s V .
In this example, a Chi-square test was conducted to assess whether males and females differ in their help-seeking behaviour (Yes/No). The crosstabulation table shows the percentage of males of females who sought/didn't seek help. The table 'Chi square tests' shows the significance of the test (Pearson Chi square asymp sig: .482). Since this value is above .05, we conclude that there is no statistically significant difference between males and females in their help-seeking behaviour.
Want to create or adapt books like this? Learn more about how Pressbooks supports open publishing practices.
The previous two chapters introduced methods for organizing and summarizing sample data, and using sample statistics to estimate population parameters. This chapter introduces the next major topic of inferential statistics: hypothesis testing.
A hypothesis is a statement or claim about a property of a population.
When conducting scientific research, typically there is some known information, perhaps from some past work or from a long accepted idea. We want to test whether this claim is believable. This is the basic idea behind a hypothesis test:
For example, past research tells us that the average life span for a hummingbird is about four years. You have been studying the hummingbirds in the southeastern United States and find a sample mean lifespan of 4.8 years. Should you reject the known or accepted information in favor of your results? How confident are you in your estimate? At what point would you say that there is enough evidence to reject the known information and support your alternative claim? How far from the known mean of four years can the sample mean be before we reject the idea that the average lifespan of a hummingbird is four years?
Hypothesis testing is a procedure, based on sample evidence and probability, used to test claims regarding a characteristic of a population.
A hypothesis is a claim or statement about a characteristic of a population of interest to us. A hypothesis test is a way for us to use our sample statistics to test a specific claim.
The population mean weight is known to be 157 lb. We want to test the claim that the mean weight has increased.
Two years ago, the proportion of infected plants was 37%. We believe that a treatment has helped, and we want to test the claim that there has been a reduction in the proportion of infected plants.
The null hypothesis is a statement about the value of a population parameter, such as the population mean (µ) or the population proportion ( p ). It contains the condition of equality and is denoted as H 0 (H-naught).
H 0 : µ = 157 or H 0 : p = 0.37
The alternative hypothesis is the claim to be tested, the opposite of the null hypothesis. It contains the value of the parameter that we consider plausible and is denoted as H 1 .
H 1 : µ > 157 or H 1 : p ≠ 0.37
The test statistic is a value computed from the sample data that is used in making a decision about the rejection of the null hypothesis. The test statistic converts the sample mean ( x̄ ) or sample proportion ( p̂ ) to a Z- or t-score under the assumption that the null hypothesis is true . It is used to decide whether the difference between the sample statistic and the hypothesized claim is significant.
The p-value is the area under the curve to the left or right of the test statistic. It is compared to the level of significance ( α ).
The critical value is the value that defines the rejection zone (the test statistic values that would lead to rejection of the null hypothesis). It is defined by the level of significance.
The level of significance ( α ) is the probability that the test statistic will fall into the critical region when the null hypothesis is true. This level is set by the researcher.
The conclusion is the final decision of the hypothesis test. The conclusion must always be clearly stated, communicating the decision based on the components of the test. It is important to realize that we never prove or accept the null hypothesis. We are merely saying that the sample evidence is not strong enough to warrant the rejection of the null hypothesis. The conclusion is made up of two parts:
1) Reject or fail to reject the null hypothesis, and 2) there is or is not enough evidence to support the alternative claim.
Option 1) Reject the null hypothesis (H 0 ). This means that you have enough statistical evidence to support the alternative claim (H 1 ).
Option 2) Fail to reject the null hypothesis (H 0 ). This means that you do NOT have enough evidence to support the alternative claim (H 1 ).
Another way to think about hypothesis testing is to compare it to the US justice system. A defendant is innocent until proven guilty (Null hypothesis—innocent). The prosecuting attorney tries to prove that the defendant is guilty (Alternative hypothesis—guilty). There are two possible conclusions that the jury can reach. First, the defendant is guilty (Reject the null hypothesis). Second, the defendant is not guilty (Fail to reject the null hypothesis). This is NOT the same thing as saying the defendant is innocent! In the first case, the prosecutor had enough evidence to reject the null hypothesis (innocent) and support the alternative claim (guilty). In the second case, the prosecutor did NOT have enough evidence to reject the null hypothesis (innocent) and support the alternative claim of guilty.
There are three different pairs of null and alternative hypotheses:
where c is some known value.
This tests whether the population parameter is equal to, versus not equal to, some specific value.
H o : μ = 12 vs. H 1 : μ ≠ 12
The critical region is divided equally into the two tails and the critical values are ± values that define the rejection zones.
A forester studying diameter growth of red pine believes that the mean diameter growth will be different if a fertilization treatment is applied to the stand.
This is a two-sided question, as the forester doesn’t state whether population mean diameter growth will increase or decrease.
This tests whether the population parameter is equal to, versus greater than, some specific value.
H o : μ = 12 vs. H 1 : μ > 12
The critical region is in the right tail and the critical value is a positive value that defines the rejection zone.
A biologist believes that there has been an increase in the mean number of lakes infected with milfoil, an invasive species, since the last study five years ago.
This is a right-sided question, as the biologist believes that there has been an increase in population mean number of infected lakes.
This tests whether the population parameter is equal to, versus less than, some specific value.
H o : μ = 12 vs. H 1 : μ < 12
The critical region is in the left tail and the critical value is a negative value that defines the rejection zone.
A scientist’s research indicates that there has been a change in the proportion of people who support certain environmental policies. He wants to test the claim that there has been a reduction in the proportion of people who support these policies.
This is a left-sided question, as the scientist believes that there has been a reduction in the true population proportion.
When the observed results (the sample statistics) are unlikely (a low probability) under the assumption that the null hypothesis is true, we say that the result is statistically significant, and we reject the null hypothesis. This result depends on the level of significance, the sample statistic, sample size, and whether it is a one- or two-sided alternative hypothesis.
When testing, we arrive at a conclusion of rejecting the null hypothesis or failing to reject the null hypothesis. Such conclusions are sometimes correct and sometimes incorrect (even when we have followed all the correct procedures). We use incomplete sample data to reach a conclusion and there is always the possibility of reaching the wrong conclusion. There are four possible conclusions to reach from hypothesis testing. Of the four possible outcomes, two are correct and two are NOT correct.
A Type I error is when we reject the null hypothesis when it is true. The symbol α (alpha) is used to represent Type I errors. This is the same alpha we use as the level of significance. By setting alpha as low as reasonably possible, we try to control the Type I error through the level of significance.
A Type II error is when we fail to reject the null hypothesis when it is false. The symbol β (beta) is used to represent Type II errors.
In general, Type I errors are considered more serious. One step in the hypothesis test procedure involves selecting the significance level ( α ), which is the probability of rejecting the null hypothesis when it is correct. So the researcher can select the level of significance that minimizes Type I errors. However, there is a mathematical relationship between α, β , and n (sample size).
The natural inclination is to select the smallest possible value for α, thinking to minimize the possibility of causing a Type I error. Unfortunately, this forces an increase in Type II errors. By making the rejection zone too small, you may fail to reject the null hypothesis, when, in fact, it is false. Typically, we select the best sample size and level of significance, automatically setting β .
A Type II error ( β ) is the probability of failing to reject a false null hypothesis. It follows that 1- β is the probability of rejecting a false null hypothesis. This probability is identified as the power of the test, and is often used to gauge the test’s effectiveness in recognizing that a null hypothesis is false.
The probability that at a fixed level α significance test will reject H 0 , when a particular alternative value of the parameter is true is called the power of the test.
Power is also directly linked to sample size. For example, suppose the null hypothesis is that the mean fish weight is 8.7 lb. Given sample data, a level of significance of 5%, and an alternative weight of 9.2 lb., we can compute the power of the test to reject μ = 8.7 lb. If we have a small sample size, the power will be low. However, increasing the sample size will increase the power of the test. Increasing the level of significance will also increase power. A 5% test of significance will have a greater chance of rejecting the null hypothesis than a 1% test because the strength of evidence required for the rejection is less. Decreasing the standard deviation has the same effect as increasing the sample size: there is more information about μ .
We are going to examine two equivalent ways to perform a hypothesis test: the classical approach and the p-value approach. The classical approach is based on standard deviations. This method compares the test statistic (Z-score) to a critical value (Z-score) from the standard normal table. If the test statistic falls in the rejection zone, you reject the null hypothesis. The p-value approach is based on area under the normal curve. This method compares the area associated with the test statistic to alpha ( α ), the level of significance (which is also area under the normal curve). If the p-value is less than alpha, you would reject the null hypothesis.
As a past student poetically said: If the p-value is a wee value, Reject Ho
Both methods must have:
There are four steps required for a hypothesis test:
A forester studying diameter growth of red pine believes that the mean diameter growth will be different from the known mean growth of 1.35 inches/year if a fertilization treatment is applied to the stand. He conducts his experiment, collects data from a sample of 32 plots, and gets a sample mean diameter growth of 1.6 in./year. The population standard deviation for this stand is known to be 0.46 in./year. Does he have enough evidence to support his claim?
Step 1) State the null and alternative hypotheses.
Step 2) State the level of significance and the critical value.
Step 3) Compute the test statistic.
Step 4) State a conclusion.
In this problem, the test statistic falls in the red rejection zone. The test statistic of 3.07 is greater than the critical value of 1.96.We will reject the null hypothesis. We have enough evidence to support the claim that the mean diameter growth is different from (not equal to) 1.35 in./year.
A researcher believes that there has been an increase in the average farm size in his state since the last study five years ago. The previous study reported a mean size of 450 acres with a population standard deviation ( σ ) of 167 acres. He samples 45 farms and gets a sample mean of 485.8 acres. Is there enough information to support his claim?
We fail to reject the null hypothesis. We do not have enough evidence to support the claim that the mean farm size has increased from 450 acres.
A researcher believes that there has been a reduction in the mean number of hours that college students spend preparing for final exams. A national study stated that students at a 4-year college spend an average of 23 hours preparing for 5 final exams each semester with a population standard deviation of 7.3 hours. The researcher sampled 227 students and found a sample mean study time of 19.6 hours. Does this indicate that the average study time for final exams has decreased? Use a 1% level of significance to test this claim.
We reject the null hypothesis. We have sufficient evidence to support the claim that the mean final exam study time has decreased below 23 hours.
The p-value is the probability of observing our sample mean given that the null hypothesis is true. It is the area under the curve to the left or right of the test statistic. If the probability of observing such a sample mean is very small (less than the level of significance), we would reject the null hypothesis. Computations for the p-value depend on whether it is a one- or two-sided test.
Steps for a hypothesis test using p-values:
Instead of comparing Z-score test statistic to Z-score critical value, as in the classical method, we compare area of the test statistic to area of the level of significance.
The Decision Rule: If the p-value is less than alpha, we reject the null hypothesis
If it is a two-sided test (the alternative claim is ≠), the p-value is equal to two times the probability of the absolute value of the test statistic. If the test is a left-sided test (the alternative claim is “<”), then the p-value is equal to the area to the left of the test statistic. If the test is a right-sided test (the alternative claim is “>”), then the p-value is equal to the area to the right of the test statistic.
Let’s look at Example 6 again.
A forester studying diameter growth of red pine believes that the mean diameter growth will be different from the known mean growth of 1.35 in./year if a fertilization treatment is applied to the stand. He conducts his experiment, collects data from a sample of 32 plots, and gets a sample mean diameter growth of 1.6 in./year. The population standard deviation for this stand is known to be 0.46 in./year. Does he have enough evidence to support his claim?
Step 2) State the level of significance.
The p-value is two times the area of the absolute value of the test statistic (because the alternative claim is “not equal”).
Step 4) Compare the p-value to alpha and state a conclusion.
Let’s look at Example 7 again.
The p-value is the area to the right of the Z-score 1.44 (the hatched area).
We fail to reject the null hypothesis. We do not have enough evidence to support the claim that the mean farm size has increased.
Let’s look at Example 8 again.
The p-value is the area to the left of the test statistic (the little black area to the left of -7.02). The Z-score of -7.02 is not on the standard normal table. The smallest probability on the table is 0.0002. We know that the area for the Z-score -7.02 is smaller than this area (probability). Therefore, the p-value is <0.0002.
We reject the null hypothesis. We have enough evidence to support the claim that the mean final exam study time has decreased below 23 hours.
Both the classical method and p-value method for testing a hypothesis will arrive at the same conclusion. In the classical method, the critical Z-score is the number on the z-axis that defines the level of significance ( α ). The test statistic converts the sample mean to units of standard deviation (a Z-score). If the test statistic falls in the rejection zone defined by the critical value, we will reject the null hypothesis. In this approach, two Z-scores, which are numbers on the z-axis, are compared. In the p-value approach, the p-value is the area associated with the test statistic. In this method, we compare α (which is also area under the curve) to the p-value. If the p-value is less than α , we reject the null hypothesis. The p-value is the probability of observing such a sample mean when the null hypothesis is true. If the probability is too small (less than the level of significance), then we believe we have enough statistical evidence to reject the null hypothesis and support the alternative claim.
(referring to Ex. 8)
Test of mu = 23 vs. < 23 |
The assumed standard deviation = 7.3 |
99% Upper | |||||
N | Mean | SE Mean | Bound | Z | P |
227 | 19.600 | 0.485 | 20.727 | -7.02 | 0.000 |
Excel does not offer 1-sample hypothesis testing.
Frequently, the population standard deviation (σ) is not known. We can estimate the population standard deviation (σ) with the sample standard deviation (s). However, the test statistic will no longer follow the standard normal distribution. We must rely on the student’s t-distribution with n-1 degrees of freedom. Because we use the sample standard deviation (s), the test statistic will change from a Z-score to a t-score.
Steps for a hypothesis test are the same that we covered in Section 2.
Just as with the hypothesis test from the previous section, the data for this test must be from a random sample and requires either that the population from which the sample was drawn be normal or that the sample size is sufficiently large (n≥30). A t-test is robust, so small departures from normality will not adversely affect the results of the test. That being said, if the sample size is smaller than 30, it is always good to verify the assumption of normality through a normal probability plot.
We will still have the same three pairs of null and alternative hypotheses and we can still use either the classical approach or the p-value approach.
Selecting the correct critical value from the student’s t-distribution table depends on three factors: the type of test (one-sided or two-sided alternative hypothesis), the sample size, and the level of significance.
For a two-sided test (“not equal” alternative hypothesis), the critical value (t α /2 ), is determined by alpha ( α ), the level of significance, divided by two, to deal with the possibility that the result could be less than OR greater than the known value.
For a one-sided test (“a less than” or “greater than” alternative hypothesis), the critical value (t α ) , is determined by alpha ( α ), the level of significance, being all in the one side.
Find the critical value you would use to test the claim that μ ≠ 112 with a sample size of 18 and a 5% level of significance.
In this case, the critical value (t α /2 ) would be 2.110. This is a two-sided question (≠) so you would divide alpha by 2 (0.05/2 = 0.025) and go down the 0.025 column to 17 degrees of freedom.
What would the critical value be if you wanted to test that μ < 112 for the same data?
In this case, the critical value would be 1.740. This is a one-sided question (<) so alpha would be divided by 1 (0.05/1 = 0.05). You would go down the 0.05 column with 17 degrees of freedom to get the correct critical value.
In 2005, the mean pH level of rain in a county in northern New York was 5.41. A biologist believes that the rain acidity has changed. He takes a random sample of 11 rain dates in 2010 and obtains the following data. Use a 1% level of significance to test his claim.
4.70, 5.63, 5.02, 5.78, 4.99, 5.91, 5.76, 5.54, 5.25, 5.18, 5.01
The sample size is small and we don’t know anything about the distribution of the population, so we examine a normal probability plot. The distribution looks normal so we will continue with our test.
The sample mean is 5.343 with a sample standard deviation of 0.397.
We will fail to reject the null hypothesis. We do not have enough evidence to support the claim that the mean rain pH has changed.
Cadmium, a heavy metal, is toxic to animals. Mushrooms, however, are able to absorb and accumulate cadmium at high concentrations. The government has set safety limits for cadmium in dry vegetables at 0.5 ppm. Biologists believe that the mean level of cadmium in mushrooms growing near strip mines is greater than the recommended limit of 0.5 ppm, negatively impacting the animals that live in this ecosystem. A random sample of 51 mushrooms gave a sample mean of 0.59 ppm with a sample standard deviation of 0.29 ppm. Use a 5% level of significance to test the claim that the mean cadmium level is greater than the acceptable limit of 0.5 ppm.
The sample size is greater than 30 so we are assured of a normal distribution of the means.
Step 4) State a Conclusion.
The test statistic falls in the rejection zone. We will reject the null hypothesis. We have enough evidence to support the claim that the mean cadmium level is greater than the acceptable safe limit.
BUT, what happens if the significance level changes to 1%?
The critical value is now found by going down the 0.01 column with 50 degrees of freedom. The critical value is 2.403. The test statistic is now LESS THAN the critical value. The test statistic does not fall in the rejection zone. The conclusion will change. We do NOT have enough evidence to support the claim that the mean cadmium level is greater than the acceptable safe limit of 0.5 ppm.
The level of significance is the probability that you, as the researcher, set to decide if there is enough statistical evidence to support the alternative claim. It should be set before the experiment begins.
We can also use the p-value approach for a hypothesis test about the mean when the population standard deviation ( σ ) is unknown. However, when using a student’s t-table, we can only estimate the range of the p-value, not a specific value as when using the standard normal table. The student’s t-table has area (probability) across the top row in the table, with t-scores in the body of the table.
Estimating P-value from a Student’s T-table
If your test statistic is 3.789 with 3 degrees of freedom, you would go across the 3 df row. The value 3.789 falls between the values 3.482 and 4.541 in that row. Therefore, the p-value is between 0.02 and 0.01. The p-value will be greater than 0.01 but less than 0.02 (0.01<p<0.02).
If your level of significance is 5%, you would reject the null hypothesis as the p-value (0.01-0.02) is less than alpha ( α ) of 0.05.
If your level of significance is 1%, you would fail to reject the null hypothesis as the p-value (0.01-0.02) is greater than alpha ( α ) of 0.01.
Software packages typically output p-values. It is easy to use the Decision Rule to answer your research question by the p-value method.
(referring to Ex. 12)
Test of mu = 0.5 vs. > 0.5
95% Lower | ||||||
N | Mean | StDev | SE Mean | Bound | T | P |
51 | 0.5900 | 0.2900 | 0.0406 | 0.5219 | 2.22 | 0.016 |
Additional example: www.youtube.com/watch?v=WwdSjO4VUsg .
Frequently, the parameter we are testing is the population proportion.
Recall that the best point estimate of p , the population proportion, is given by
when np (1 – p )≥10. We can use both the classical approach and the p-value approach for testing.
The steps for a hypothesis test are the same that we covered in Section 2.
The test statistic follows the standard normal distribution. Notice that the standard error (the denominator) uses p instead of p̂ , which was used when constructing a confidence interval about the population proportion. In a hypothesis test, the null hypothesis is assumed to be true, so the known proportion is used.
A botanist has produced a new variety of hybrid soy plant that is better able to withstand drought than other varieties. The botanist knows the seed germination for the parent plants is 75%, but does not know the seed germination for the new hybrid. He tests the claim that it is different from the parent plants. To test this claim, 450 seeds from the hybrid plant are tested and 321 have germinated. Use a 5% level of significance to test this claim that the germination rate is different from 75%.
This is a two-sided question so alpha is divided by 2.
The test statistic does not fall in the rejection zone. We fail to reject the null hypothesis. We do not have enough evidence to support the claim that the germination rate of the hybrid plant is different from the parent plants.
Let’s answer this question using the p-value approach. Remember, for a two-sided alternative hypothesis (“not equal”), the p-value is two times the area of the test statistic. The test statistic is -1.81 and we want to find the area to the left of -1.81 from the standard normal table.
Now compare the p-value to alpha. The Decision Rule states that if the p-value is less than alpha, reject the H 0 . In this case, the p-value (0.0702) is greater than alpha (0.05) so we will fail to reject H 0 . We do not have enough evidence to support the claim that the germination rate of the hybrid plant is different from the parent plants.
You are a biologist studying the wildlife habitat in the Monongahela National Forest. Cavities in older trees provide excellent habitat for a variety of birds and small mammals. A study five years ago stated that 32% of the trees in this forest had suitable cavities for this type of wildlife. You believe that the proportion of cavity trees has increased. You sample 196 trees and find that 79 trees have cavities. Does this evidence support your claim that there has been an increase in the proportion of cavity trees?
Use a 10% level of significance to test this claim.
This is a one-sided question so alpha is divided by 1.
The test statistic is larger than the critical value (it falls in the rejection zone). We will reject the null hypothesis. We have enough evidence to support the claim that there has been an increase in the proportion of cavity trees.
Now use the p-value approach to answer the question. This is a right-sided question (“greater than”), so the p-value is equal to the area to the right of the test statistic. Go to the positive side of the standard normal table and find the area associated with the Z-score of 2.49. The area is 0.9936. Remember that this table is cumulative from the left. To find the area to the right of 2.49, we subtract from one.
p-value = (1 – 0.9936) = 0.0064
The p-value is less than the level of significance (0.10), so we reject the null hypothesis. We have enough evidence to support the claim that the proportion of cavity trees has increased.
(referring to Ex. 15)
Test of p = 0.32 vs. p > 0.32
90% Lower | ||||||
Sample | X | N | Sample p | Bound | Z-Value | p-Value |
1 | 79 | 196 | 0.403061 | 0.358160 | 2.49 | 0.006 |
Using the normal approximation. |
When people think of statistical inference, they usually think of inferences involving population means or proportions. However, the particular population parameter needed to answer an experimenter’s practical questions varies from one situation to another, and sometimes a population’s variability is more important than its mean. Thus, product quality is often defined in terms of low variability.
Sample variance S 2 can be used for inferences concerning a population variance σ 2 . For a random sample of n measurements drawn from a normal population with mean μ and variance σ 2 , the value S 2 provides a point estimate for σ 2 . In addition, the quantity ( n – 1) S 2 / σ 2 follows a Chi-square ( χ 2 ) distribution, with df = n – 1.
The properties of Chi-square ( χ 2 ) distribution are:
Alternative hypothesis:
where the χ 2 critical value in the rejection region is based on degrees of freedom df = n – 1 and a specified significance level of α .
As with previous sections, if the test statistic falls in the rejection zone set by the critical value, you will reject the null hypothesis.
A forester wants to control a dense understory of striped maple that is interfering with desirable hardwood regeneration using a mist blower to apply an herbicide treatment. She wants to make sure that treatment has a consistent application rate, in other words, low variability not exceeding 0.25 gal./acre (0.06 gal. 2 ). She collects sample data (n = 11) on this type of mist blower and gets a sample variance of 0.064 gal. 2 Using a 5% level of significance, test the claim that the variance is significantly greater than 0.06 gal. 2
H 0 : σ 2 = 0.06
H 1 : σ 2 >0.06
The critical value is 18.307. Any test statistic greater than this value will cause you to reject the null hypothesis.
The test statistic is
We fail to reject the null hypothesis. The forester does NOT have enough evidence to support the claim that the variance is greater than 0.06 gal. 2 You can also estimate the p-value using the same method as for the student t-table. Go across the row for degrees of freedom until you find the two values that your test statistic falls between. In this case going across the row 10, the two table values are 4.865 and 15.987. Now go up those two columns to the top row to estimate the p-value (0.1-0.9). The p-value is greater than 0.1 and less than 0.9. Both are greater than the level of significance (0.05) causing us to fail to reject the null hypothesis.
(referring to Ex. 16)
Test and CI for One Variance
Method | ||
Null hypothesis | Sigma-squared | = 0.06 |
Alternative hypothesis | Sigma-squared | > 0.06 |
The chi-square method is only for the normal distribution.
Test | |||
Method | Statistic | DF | P-Value |
Chi-Square | 10.67 | 10 | 0.384 |
Excel does not offer 1-sample χ 2 testing.
To test a claim about μ when σ is known.
Natural Resources Biometrics Copyright © 2014 by Diane Kiernan is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License , except where otherwise noted.
Statistics By Jim
Making statistics intuitive
By Jim Frost 2 Comments
The Kruskal Wallis test is a nonparametric hypothesis test that compares three or more independent groups. Statisticians also refer to it as one-way ANOVA on ranks. This analysis extends the Mann Whitney U nonparametric test that can compare only two groups.
If you analyze data, chances are you’re familiar with one-way ANOVA that compares the means of at least three groups. The Kruskal Wallis test is the nonparametric version of it. Because it is nonparametric, the analysis makes fewer assumptions about your data than its parametric equivalent.
Many analysts use the Kruskal Wallis test to determine whether the medians of at least three groups are unequal. However, it’s important to note that it only assesses the medians in particular circumstances. Interpreting the analysis results can be thorny. More on this later!
If you need a nonparametric test for paired groups or a single sample , consider the Wilcoxon signed rank test .
Learn more about Parametric vs. Nonparametric Tests and Hypothesis Testing Overview .
At its core, the Kruskal Wallis test evaluates data ranks. The procedure ranks all the sample data from low to high. Then it averages the ranks for all groups. If the results are statistically significant, the average group ranks are not all equal. Consequently, the analysis indicates whether any groups have values that rank differently. For instance, one group might have values that tend to rank higher than the other groups.
The Kruskal Wallis test doesn’t involve medians or other distributional properties—just the ranks. In fact, by evaluating ranks, it rolls up both the location and shape parameters into a single evaluation of each group’s average rank.
When their average ranks are unequal, you know a group’s distribution tends to produce higher or lower values than the others. However, you don’t know enough to draw conclusions specifically about the distributions’ locations (e.g., the medians).
However, when you hold the distribution shapes constant, the Kruskal Wallis test does tell us about the median. That’s not a property of the procedure itself but logic. If several distributions have the same shape, but the average ranks are shifted higher and lower, their medians must differ. But we can only draw that conclusion about the medians when the distributions have the same shapes.
These three distributions have the same shape, but the red and green are shifted right to higher values. Wherever the median falls on the blue distribution, it’ll be in the corresponding position in the red and blue distributions. In this case, the analysis can assess the medians.
But, if the shapes aren’t similar, we don’t know whether the location, shape, or a combination of the two produced the statistically significant Kruskal Wallis test.
Like all statistical analyses, the Kruskal Wallis test has assumptions. Ensuring that your data meet these assumptions is crucial.
Violating these assumptions can lead to incorrect conclusions.
Consider using the Kruskal Wallis test in the following cases:
Learn more about the Normal Distribution .
If you have 3 – 9 groups and more than 15 observations per group or 10 – 12 groups and more than 20 observations per group, you might want to use one-way ANOVA even when you have nonnormal data. The central limit theorem causes the sampling distributions to converge on normality, making ANOVA a suitable choice.
One-way ANOVA has several advantages over the Kruskal Wallis test, including the following:
In short, use this nonparametric method when you’re specifically interested in the medians, have ordinal data, or can’t use one-way ANOVA because you have a small, nonnormal sample.
Like one-way ANOVA, the Kruskal Wallis test is an “omnibus” test. Omnibus tests can tell you that not all your groups are equal, but it doesn’t specify which pairs of groups are different.
Specifically, the Kruskal Wallis test evaluates the following hypotheses:
Again, if the distributions have similar shapes, you can replace “average ranks” with “medians.”
Imagine you’re studying five different diets and their impact on weight loss. The Kruskal Wallis test can confirm that at least two diets have different results. However, it won’t tell you exactly which pairs of diets have statistically significant differences.
So, how do we solve this problem? Enter post hoc tests. Perform these analyses after (i.e., post) an omnibus analysis to identify specific pairs of groups with statistically significant differences. A standard option includes Dunn’s multiple comparisons procedure. Other options include performing a series of pairwise Mann-Whitney U tests with a Bonferroni correction or the lesser-known but potent Conover-Iman method.
Learn about Post Hoc Tests for ANOVA .
Imagine you’re a healthcare administrator analyzing the median number of unoccupied beds in three hospitals. Download the CSV dataset: KruskalWallisTest .
For this Kruskal Wallis test, the p-value is 0.029, which is less than the typical significance level of 0.05. Consequently, we can reject the null hypothesis that all groups have the same average rank. At least one group has a different average rank than the others.
Furthermore, if the three hospital distributions have the same shape, we can conclude that the medians differ.
At this point, we might decide to use a post hoc test to compare pairs of hospitals.
May 20, 2024 at 2:07 pm
Sir kruskal walllis test is Two tailed or one tailed test??
May 20, 2024 at 3:55 pm
It’s a one-tailed test in the same sense that the F-test for one-way ANOVA is one-tailed.
Table of Contents
The Z.TEST function in Excel is a powerful tool that allows users to perform statistical hypothesis tests with ease. This function calculates the probability of a sample mean being equal to a specified population mean, using the standard normal distribution. It is commonly used in research and data analysis to determine the significance of a sample mean compared to a known population mean. To use the Z.TEST function, the user must input the data range for the sample and the known population mean. The function will then return a p-value that can be compared to a chosen significance level to determine if the null hypothesis should be rejected. This allows for efficient and accurate hypothesis testing, making it a valuable tool for decision-making and drawing conclusions from data.
This article describes the formula syntax and usage of the Z.TEST function in Microsoft Excel.
Returns the one-tailed P-value of a z-test.
For a given hypothesized population mean, x, Z.TEST returns the probability that the sample mean would be greater than the average of observations in the data set (array) — that is, the observed sample mean.
To see how Z.TEST can be used in a formula to compute a two-tailed probability value, see the Remarks section below.
Z.TEST(array,x,[sigma])
The Z.TEST function syntax has the following arguments:
Array Required. The array or range of data against which to test x.
x Required. The value to test.
Sigma Optional. The population (known) standard deviation. If omitted, the sample standard deviation is used.
If array is empty, Z.TEST returns the #N/A error value.
Z.TEST is calculated as follows when sigma is not omitted:
Z.TEST( array,x,sigma ) = 1- Norm .S.Dist ((Average(array)- x) / (sigma/√n),TRUE)
or when sigma is omitted:
Z.TEST( array,x ) = 1- Norm .S.Dist ((Average(array)- x) / (STDEV(array)/√n),TRUE)
where x is the sample mean AVERAGE(array), and n is COUNT(array).
Z.TEST represents the probability that the sample mean would be greater than the observed value AVERAGE(array), when the underlying population mean is μ0. From the symmetry of the Normal distribution, if AVERAGE(array) < x, Z.TEST will return a value greater than 0.5.
The following Excel formula can be used to calculate the two-tailed probability that the sample mean would be further from x (in either direction) than AVERAGE(array), when the underlying population mean is x:
=2 * MIN(Z.TEST(array,x,sigma), 1 – Z.TEST(array,x,sigma)).
Copy the example data in the following table, and paste it in cell A1 of a new Excel worksheet. For formulas to show results, select them, press F2, and then press Enter. If you need to, you can adjust the column widths to see all the data.
| ||
3 | ||
6 | ||
7 | ||
8 | ||
6 | ||
5 | ||
4 | ||
2 | ||
1 | ||
9 | ||
|
|
|
=Z.TEST(A2:A11,4) | One-tailed probability-value of a z-test for the data set above, at the hypothesized population mean of 4 (0.090574) | 0.090574 |
=2 * MIN(Z.TEST(A2:A11,4), 1 – Z.TEST(A2:A11,4)) | Two-tailed probability-value of a z-test for the data set above, at the hypothesized population mean of 4 (0.181148) | 0.181148 |
=Z.TEST(A2:A11,6) | One-tailed probability-value of a z-test for the data set above, at the hypothesized population mean of 6 (0.863043) | 0.863043 |
=2 * MIN(Z.TEST(A2:A11,6), 1 – Z.TEST(A2:A11,6)) | Two-tailed probability-value of a z-test for the data set above, at the hypothesized population mean of 6 (0.273913) | 0.273913 |
The blood microbiome is probably not real.
Up until recently, if bacteria were detected in your blood you would be in a world of trouble. Blood was long considered to be sterile, meaning free of viable microorganisms like bacteria. When disease-causing bacteria spread to the blood, they can cause a life-threatening septic shock.
But the use of DNA sequencing technology has allowed researchers to more easily detect something that had been reported as early as the late 1960s: bacteria can be found in the blood and not cause disease.
As we begin to map out and understand the complex microbial ecosystem that lives in our gut and elsewhere in the body, we contemplate an important question: is there such a thing as a blood microbiome?
Our large intestine is not sterile; it is teeming with bacteria. But there are parts of the body that were long thought to be devoid of microorganisms. The brain. Bones. A variety of internal fluids, like our synovial fluid and peritoneal fluid. And, importantly, the blood.
Blood is made up of a liquid called plasma filled with red blood cells, whose main function is to carry oxygen to our cells. It also transports white blood cells, important to monitor for and fight off infections, as well as platelets, involved in clotting.
In the 1960s, a team of Italian researchers published multiple papers describing “mycoplasm-like forms”—meaning shapes that look like a particular type of bacteria that often contaminate cells cultured in the lab—in the blood of healthy people. This finding was confirmed in 1977 by a different team, which reported that four out of the 60 blood samples they had drawn from healthy volunteers showed bacteria growing in them. These types of tests, however, were rudimentary compared to what we have access to now. In the 2000s, they were mostly supplanted by DNA testing.
While we can sequence the entire DNA of any bacteria found in the blood, the technique most often used is 16S rRNA gene sequencing. I have always admired physicists’ penchant for quirky names: gluons, neutrinos, and charm quarks. Molecular biologists, by comparison, tend to be more sober. Yes, we have genes like Sonic hedgehog and proteins called scramblases; usually, though, we have to contend with the dryness of “16S rRNA.” You see, RNA is a molecule with many uses. Messenger RNA (or mRNA) acts as a disposable copy of a gene, a template for the production of a specific protein. Transfer RNA (or tRNA) actually brings the building blocks of a protein to where they are being assembled. And ribosomal RNA (or rRNA) is the main component of the giant protein factories in our cells known as ribosomes. One of its subunits is made up of, among others, a particular string of RNA known as the 16S rRNA.
The cool thing about the gene that codes for this 16S rRNA molecule is that it is very old and it mutates at a slow rate. By reading its precise sequence, scientists can tell which species it belongs to. Most of the studies of the putative blood microbiome use this technique to tell which species of bacteria are present in the blood being tested. The limitation of this test, however, is that dead bacteria have DNA too. The fact that DNA from the 16S rRNA gene of a precise bacterial species was detected in someone’s blood does not mean these bacteria were alive. For there to be a microbiome in the blood, these microorganisms need to live.
Which brings us to another important point of discussion. In order for scientists to agree that a blood microbiome exists, they first need to decide on the definition of a microbiome, and this is still a point of contention. In 2020, while companies were more than happy to sell hyped-up services testing your gut microbiome and claiming to interpret what it meant for your health, actual experts in the field met to agree on just what the word meant. “We are lacking,” they wrote , “a clear commonly agreed definition of the term ‘microbiome’.” For example, do viruses qualify? A microbiome implies life but viruses live on the edge, pun intended: they have the genetic blueprint for life yet they cannot reproduce on their own.
These experts proposed that the word “microbiome” should refer to the sum of two things: the microbiota, meaning the living microorganisms themselves, and their theatre of activity. It’s like saying that the Earth is not simply the life forms it houses, but also all of their individual components, and the traces they leave behind, and the environmental conditions in which they thrive or die. The microbiome is made up of bacteria and other microorganisms, yes, but also their proteins, lipids, sugars, and DNA and RNA molecules, as well as the signalling molecules and toxins that get exchanged within their theatre. (This is where viruses were sorted, by the way: not as part of the living microbiota but as belonging to the theatre of activity of the microbiome.)
The microbiome is a community, and this community has a distinct habitat.
So, what does the evidence say? Is our blood truly host to a thriving community of microorganisms or is something else going on?
Initial studies of the alleged blood microbiome were small . The amounts of bacteria that were being reported based on DNA sequencing were tiny. If this microbiome existed, it seemed sparse, more “asteroid field in real life” than “asteroid field in the movies.”
An issue looming over this early research is that of contamination. If bacteria are detected in a blood sample, were they really in the blood… or did they contaminate supplies along the way? When blood is drawn, the skin, which has its own microbiome, is punctured. The area is usually swabbed with alcohol to kill bacteria, and the supplies used should be sterile, but suffice to say that from the blood draw to the DNA extraction to the DNA amplification to the sequencing of this DNA, bacteria can be introduced into the system. In fact, it is such common knowledge that certain bacteria are found inside of the laboratory kits used by scientists that this ecosystem has its own name: the kitome. One way to rule out these contaminants is to simultaneously run negative controls alongside samples every step of the way, to make sure that these negative controls are indeed free of bacteria. But early papers rarely reported when controls were used.
Last year, results from what purports to be the largest study ever into the question of whether the blood microbiome exists were published in Nature Microbiology . A total of 9,770 healthy individuals were tested. The conclusion? Yes, some bacteria could be found in their blood, but the evidence contradicted the claim of an ecosystem. In 84% of the samples tested, no bacteria were detected. In most of the other samples, only one species was found. In an ecosystem, you would expect to see species appearing together repeatedly, but this was not the case here. And the species they found most often in their samples were known to contaminate these types of laboratory experiments.
So, what were the few bacteria found in the blood and not recognized as contaminants doing there in the first place if they were not part of a healthy microbiome? The authors lean toward an alternative explanation that had been floated for many years: these bacteria are transient. They end up in the blood from other parts of the body, either because of some minor leak or through their active transportation into the blood by agents such as dendritic cells. Like pedestrians wandering off onto the highway, these bacteria do not normally live in the blood but they can be seen there when we look at the right moment.
This blood microbiome story could end here and simply be an interesting example of scientific research homing in on a curious finding, testing a hypothesis, and ultimately refuting it (or at the very least providing strong evidence against it). But given the incentives of modern research and the social-media spotlight cast on the academic literature, there are two slightly worrying angles here that merit discussion.
Scientists are more and more incentivized to find practical applications for their research. It’s not enough, for example, to study bacteria that survive at incredibly high temperatures; we must be assured that the DNA replication enzyme these bacteria possess will one day be used in laboratories all over the world to conduct research, identify criminals, and test samples for the presence of a pandemic-causing coronavirus.
In researching this topic, I came across many papers claiming the existence of “blood microbiome signatures” for certain diseases that are not known to be infectious. We are thus not talking about infections leaking in the blood and causing sepsis. I saw reports of signatures for cardiovascular disease , liver disease , heart attacks , even for gastrointestinal disease in dogs . The idea is that these signatures could soon be turned into (profitable) diagnostic tests. The problem, of course, is that these studies are based on the hypothesis that a blood microbiome is real; that its equilibrium can be affected by disease; and that these changes can be reliably detected and interpreted.
But if the blood microbiome is imaginary, we are just chasing ghosts. This is not unlike the time that scientists were publishing signatures of microRNAs in the blood for every possible cancer. When I looked at the published literature in grad school, I realized that the multiple signatures reported for a single cancer barely overlapped . They were just chance findings. Compare enough variables in a small sample set and you will find what appears to be an association.
My second concern is that the transitory leakage of bacteria into the blood, as evidenced by the recent Nature Microbiology paper, will be used as confirmation of a pseudoscientific entity: leaky gut syndrome. At the end of their paper, the researchers hypothesize that these bacteria end up in the blood because the integrity of certain barriers in the body are compromised during disease or during periods of stress. The “net” in our gut gets a bit porous, and some of our colon’s bacteria end up in circulation, though they are not causing disease as far as we can tell. A form of leaky gut is known to exist in certain intestinal diseases , likely to be a consequence and not a cause. But leaky gut syndrome, favoured by non-evidence-based practitioners, does not appear to be real, yet many websites portray it as the one true cause of all diseases, a real epidemic. Nuanced scientific findings have a history of being stolen, distorted, and toyed with by fake doctors to give credence to their pet theories. Though I have yet to see examples of it, I suspect work done on this hypothesized blood microbiome will similarly get weaponized.
You have been warned.
Take-home message: - Our blood was long considered to be sterile, meaning free of viable microbes, unless a dangerous infection leaked into it, causing sepsis - Studies have provided evidence for the presence of bacteria in the blood of some healthy humans, leading to the hypothesis that, much like in our gut, our blood is host to a microbiome - The largest study ever done on the topic provided strong evidence against this hypothesis. It seems that when non-disease-causing bacteria find themselves in our blood, it is temporary and occasional
@CrackedScience
The story linking nutrition and health has unexpected twists 28 jun 2024.
Office for science and society.
Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.
Scientific Reports volume 14 , Article number: 15201 ( 2024 ) Cite this article
Metrics details
With the rapid advancement of educational technology, the flipped classroom approach has garnered considerable attention owing to its potential for enhancing students’ learning capabilities. This research delves into the flipped classroom teaching methodology, employing the Unified Theory of Acceptance and Use of Technology (UTAUT), learning engagement theory, and the 4C skills (comprising communication, collaboration, creativity, and critical thinking) to investigate its effects on learning capabilities. The research surveyed 413 students from three universities in Jiangxi Province, employing stratified random sampling. SPSS 24.0 and Amos were used for structural equation modeling and hypothesis testing analysis. The findings indicate that: (1) Performance expectancy, effort expectancy, and peer influence significantly enhance students’ learning engagement in the flipped classroom. (2) Students’ learning engagement in the flipped classroom notably promotes their learning capabilities. (3) Performance expectancy, effort expectancy, and peer influence can significantly boost learning capabilities by increasing learning engagement. (4) Personality traits significantly moderate the effect of peer influence on learning engagement, highlighting the crucial role of individual differences in learning. (5) The level of students’ learning engagement is differentially influenced by performance expectancy and peer influence across various academic disciplines. Ultimately, this research provides valuable insights for educational policymakers and guides improvements in teaching practices, collectively advancing educational quality and equity.
With the rapid development of technology and continuous innovation in educational philosophy, the flipped classroom, as a cutting-edge teaching method, has garnered widespread attention globally. This teaching approach was first proposed by two high school teachers, Jonathan Bergmann and Aaron Sams, in 2007, and has quickly gained widespread application at all levels of education 1 . The core concept of the flipped classroom lies in upending the traditional teaching model, extending students’ learning activities from the classroom to outside, thereby transforming the classroom into a hub for deep learning and practical activities. As such, exploring the impact of flipped classroom teaching on students’ learning abilities constitutes an important area of research. Currently, scholars both domestically and internationally focus their research on flipped classroom teaching primarily in the following two areas:
One aspect concerns studies on the implementation effects of the flipped classroom model. Numerous studies have confirmed its significant role in promoting students’ active participation, enhancing learning motivation, and improving academic performance. In biology instruction, Flores-González and Flores-González 2 found that the flipped classroom fosters self-regulated learning and active engagement among students. Within an online educational environment, Cuetos 3 research indicates that the flipped classroom model elevates university students’ learning motivation and academic achievement. In the field of English teaching, Dewi et al. 4 study also revealed that students hold a positive attitude towards the new teaching mode of the flipped classroom. The flipped classroom provides students with more personalized and autonomous learning opportunities, allowing learning to align more closely with individual learning styles and paces 5 , 6 . Additionally, the impact of the flipped classroom on students’ academic success has garnered considerable attention. Semab and Naureen 7 research demonstrates the positive effect of the flipped classroom model in enhancing students’ academic achievements. Studies by Roehl et al. 8 and Tucker 9 emphasize that the flipped classroom promotes the cultivation of students’ practical application abilities by reallocating classroom time for in-depth discussions, problem-solving, and hands-on activities.
Another area of research pertains to effective flipped classroom teaching design and the challenges it faces. Effective teaching design serves as the cornerstone of a successful flipped classroom. Nicholas 10 underscores multiple teaching factors to consider when implementing a flipped classroom in graduate education. Meanwhile, Yang et al. 11 designed a flipped classroom teaching model based on blended learning and experimentally verified the model’s effectiveness in enhancing students’ academic performance and level of learning engagement. For STEM education, Shofiyah et al. 12 proposed a flipped classroom teaching template grounded in the 5E model, the validity and reliability of its content and structure have also been verified.
The research findings from the two aforementioned aspects provide rich references and insights for this research, yet there are two gaps in existing research. On the one hand, although the flipped classroom teaching method has been widely studied, few studies have explored the impact of the flipped classroom on enhancing 4C skills (communication, collaboration, creativity, and critical thinking), using the 4C skills as a measure of learning capability. On the other hand, despite the general agreement in previous studies on the positive effects of the flipped classroom on learning outcomes, few have investigated the differing impacts of personality and subject differences in flipped classroom teaching.
Therefore, this research seeks to answer the following questions:
Question 1: What are the key factors that affect students’ learning engagement in the flipped classroom teaching method?
Question 2: Does learning engagement further influence students’ 4C learning capabilities?
Question 3: Do individual differences (such as personality) and subject differences have a moderating effect between the relevant variables in this research?
To answer these questions, this research combines the UTAUT model, learning engagement theory, and the 4C skills analysis framework to construct a comprehensive research model. By collecting data from 413 students from three universities in Jiangxi Province, this research utilizes structural equation modeling for data analysis to unveil the complex relationships between performance expectancy, effort expectancy, peer influence, learning engagement, and learning capability in the flipped classroom.
The significance of this research lies not only in illustrating the specific impact of the flipped classroom on students’ learning capabilities through empirical analysis, providing a scientific basis for the further promotion and application of this teaching model, but also in deepening the understanding of the essence of the flipped classroom teaching model by integrating multiple theoretical frameworks. Furthermore, the research findings will provide strong support for educational policy formulation and teaching practice improvement, collectively advancing educational quality and equity.
Compared to existing research, the innovation of this research is primarily reflected in model construction and variable setting. Firstly, while preserving the core variables of the UTAUT model, this research innovatively integrates learning engagement theory and the 4C skills framework, offering a new perspective for understanding the interactions of various factors in the flipped classroom and their impact on learning capabilities. Secondly, in terms of variable setting, this research includes personality traits and subject differences as moderating variables in the analysis and utilizes 4C skills as a measure of learning capabilities, which is a novel approach in the field of flipped classroom teaching research.
In the following chapters, this research will conduct a literature review, derive research hypotheses, describe the survey process, perform model inspection, and finally discuss and present research implications.
Theory and application of the utaut.
The UTAUT model, proposed by Venkatesh et al. 13 , integrates key variables from multiple theoretical models, including the Theory of Reasoned Action (TRA), Theory of Planned Behavior (TPB), Technology Acceptance Model (TAM), Motivational Model (MM), Combined TAM and TPB (C-TAM), and Innovation Diffusion Theory (IDT). Its aim is to provide a more comprehensive and accurate framework to explain and predict users’ acceptance behavior towards new technologies.
The dimensions of the UTAUT model have demonstrated strong explanatory power in technology adoption research across multiple domains, particularly in education. The model has been applied to studies on the acceptance behavior of technologies such as Massive Open Online Courses (MOOCs) and AI chatbots. Li and Zhao 14 combined the UTAUT model with Social Presence Theory to analyze factors influencing students’ continued use of MOOCs, finding that the UTAUT model positively affects students’ satisfaction and intention to continue using MOOCs. Meanwhile, Tian et al. 15 utilized both UTAUT and ECM models to explore Chinese graduate students’ acceptance and utilization of AI chatbot technology, discovering that “confirmation” and “satisfaction” from the ECM model have a greater impact on user behavior than the UTAUT model.
The flipped classroom, as an innovative teaching model in higher education, has garnered significant attention regarding its application effects and influencing factors. In this domain, both Alyoussef 16 , Agyei and Razi 17 have conducted in-depth explorations utilizing the UTAUT model. Alyoussef 16 revealed the central mediating role of perceived usefulness and perceived ease of use in students’ acceptance of the flipped classroom, using a sample of students from a university in Saudi Arabia. This finding not only indicates students’ positive attitude towards the flipped classroom but also further confirms the positive role of this teaching model in enhancing learning outcomes. Agyei and Razi 17 enriched the UTAUT model by introducing variables such as experience expectancy, parent-school involvement, perceived behavioral control, and self-efficacy to deeply analyze high school students’ acceptance of using online resources for flipped classroom learning. Their results showed that performance expectancy, effort expectancy, parent-school involvement, students’ self-efficacy, and experience expectancy all significantly positively impact students’ willingness to learn. However, perceived behavioral control did not show a significant effect in their research.
In this research, we focus on the college student population whose social connections are primarily reflected in peer relationships. Therefore, we choose to represent the social influence element in the UTAUT model with the peer influence, aiming to be more aligned with the actual situation of this specific group. Meanwhile, considering the widespread popularity of technology in modern higher education, we do not consider facilitating conditions as a core factor in our research, although this does not imply that this element can be ignored in all environments.
Learning engagement theory, initially proposed by Fredricks et al. 18 in the field of educational psychology, aims to delve into students’ cognitive, emotional, and behavioral engagement exhibited during the learning process. This theory focuses on assessing the enthusiasm and depth of students’ participation in learning activities, encompassing three key dimensions: cognitive engagement, emotional engagement, and behavioral engagement.
In the process of deeply exploring the theory of learning engagement, researchers have conducted extensive studies targeting different educational environments and learning formats. The issue of learning engagement is particularly prominent in the field of MOOCs and online learning. Although MOOCs have attracted a large number of learners due to their openness and flexibility, the low completion rate has always been one of the challenges they face 19 . Studies have shown that students’ intrinsic motivations (such as interest) and extrinsic motivations (such as perceived knowledge value) have a significant impact on their learning engagement in MOOCs. From the perspective of self-determination theory, Lan and Hew 20 research adopted a mixed-method approach to investigate learning engagement in MOOCs. The study found that perceived ability and emotional engagement have a significant impact on students who complete MOOCs, and different dimensions of learning engagement can predict learners’ perceived learning effectiveness. Meanwhile, in the online learning environment, significant changes have occurred in the way students interact with teachers and peers, which requires educators to pay more attention to cultivating students’ autonomous learning abilities, computer and network skills, and online communication abilities to promote their learning engagement 21 .
Voogt et al. 22 emphasized the importance of 4C skills in 21st-century education in their study. These skills include Communication, Collaboration, Creativity, and Critical Thinking, which focus on cultivating students’ comprehensive literacy to adapt to the complex needs of modern society. Especially when measuring learning capability, the 4C skills provide a comprehensive and appropriate framework.
In the context of exploring the improvement of learning abilities, integrating these four skills—critical thinking, communication, collaboration, and creativity—into the learning process is particularly critical. Specifically, critical thinking skills enable learners to identify true and false information and adapt to environmental changes 23 . Communication skills, including effective speaking, listening, and writing, are key to improving interpersonal efficiency 24 . Collaboration emphasizes working together in a team environment, facilitating knowledge sharing and problem-solving 25 . Innovation ability is the key to gaining an advantage in modern social competition, requiring continuous learning, challenging traditional concepts, and maintaining sensitivity to new technologies 26 . By integrating these abilities, the learning process can be more comprehensive, improving flexibility and effectiveness in responding to various challenges.
Although previous studies have explored the role of the flipped classroom model in promoting student collaboration, criticism, and innovation abilities 9 , 27 , there is still a lack of in-depth exploration of the specific impact and mechanism of communication ability in this model. Therefore, this research aims to comprehensively integrate 4C abilities and deeply explore the overall impact of the flipped classroom teaching model on these abilities.
This research combines the UTAUT model, learning engagement theory, and the 4C theory, selecting Performance expectancy (PE), Effort expectancy (EEX), Peer influence (PI), Learning engagement (ENGA), and Learning capability (SKIL) as research constructs to explore the impact mechanism of flipped classroom teaching on college students’ learning capability. The specific definitions of the variables are shown in Table 1 :
The relationship between performance expectancy, effort expectancy, peer influence, and learning engagement.
Singh 28 , Riddle and Gier 29 and Clark 30 and other scholars have found that flipped classrooms can improve students’ test scores and course engagement by replacing traditional lectures with micro-lectures and activity-based learning strategies, thus demonstrating that performance expectancy have a significant positive impact on students’ engagement in learning. The fully online flipped classroom model was more effective than the online flipped model in supporting student behavioral engagement, suggesting the importance of effort expectancy in an online environment 31 . Ruiz 32 study demonstrated that integrating interactive technology and peer instruction into a flipped classroom can positively affect student engagement, i.e., the importance of peer influence in enhancing student engagement and learning. Based on the above, the following hypotheses are proposed in this research:
Performance expectancy has a significant positive impact on students’ learning engagement.
Effort expectancy has a significant positive impact on students’ learning engagement.
Peer influence has a significant positive impact on students’ learning engagement.
Learning Engagement Theory emphasizes the cognitive, emotional, and behavioral engagement of students in the learning process and these factors are believed to positively influence the enhancement of learning capabilities. Zimmerman 33 states that when students believe they can succeed in a learning task, they are more likely to invest more energy and effort, which promotes learning capabilities. Pekrun et al. 34 explored the relationship between students’ engagement in learning and learning competence from an emotional perspective, which further supports the positive link between engagement in learning and increased learning competence. Martin and Bolliger 35 noted that student engagement in learning directly impacts online learning outcomes, improves student performance in online programs and is considered an important factor in measuring teaching quality. Taken together, these studies suggest that learning engagement can have a significant positive impact on students’ learning capabilities by influencing their self-efficacy, motivation, learning strategies, and emotional engagement. Based on the above, this research proposes the following hypotheses:
Students’ learning engagement has a significant positive impact on learning capability.
According to the aforementioned literature, it is evident that performance expectancy, effort expectancy, and peer influence have a positive effect on the enhancement of learning capabilities through learning engagement. This is supported by studies from Chen and Wu 36 , Buabeng-Andoh 37 , Kuo et al. 38 , Zimmerman 33 , Pekrun et al. 34 , and Martin and Bolliger 35 . Additionally, research by Jamaludin and Osman 39 , Wang 40 , and Nerantzi 41 also corroborate the notion that performance expectancy, effort expectancy, and peer influence impact learning capabilities through learning engagement. Based on this information, the current study proposes the following hypotheses:
Performance expectancy has a significant positive impact on learning capability through students’ learning engagement.
Effort expectancy has a significant positive impact on learning capability through students’ learning engagement.
Peer influence has a significant positive impact on learning capability through students’ learning engagement.
Personality.
Personality is a relatively stable individual difference in behavior, emotion, and cognition exhibited by an individual that encompasses traits, habits, attitudes, and values. Eysenck and Eysenck and Eysenck 42 Proposed the introversion–extraversion theory to explain and describe individual personality differences, which divides human personality into two types: introverted and extroverted. Chuang et al. 43 , Kim et al. 44 and other scholars explored the differences in classroom performance of students with different personality traits in a flipped classroom, with Wang et al. 45 noted that students with moderate openness performed best in flipped classrooms, while students with high openness performed best in online learning situations. However, there are relatively few studies on personality as a moderating variable in the UTAUT model because UTAUT focuses primarily on technology acceptance and use behaviors, while individual differences, including personality, are usually more prominent in other models. Combined with the traits of the subjects under study, introverted students may be more inclined to show higher learning engagement in independent learning environments, while extroverted students may be more adept at collaborating with peers, therefore, the variable of personality is added as a moderating variable in this study. Based on the above, the following hypotheses are proposed in this research:
Personality has a moderating effect between performance expectancy and learning engagement.
Personality has a moderating effect between effort expectancy and learning engagement.
Personality has a moderating effect between peer influence and learning engagement.
Personality has a moderating effect between learning engagement and learning capability.
Based on the perspective of individual differences, subject differences may have different impacts on students’ engagement and learning capabilities in the flipped classroom environment. Liu et al. 46 explored the impact of a flipped classroom integrating subjects on student learning capabilities in different health professional fields. Meanwhile, Fan 47 explored the application of social influences, school motivation, and gender differences in educational psychology and found that disciplinary background may influence student motivation and engagement. Students may exhibit different learning preferences and strategies in humanities and social sciences and science and technology disciplinary contexts. Meanwhile, subject differences in the UTAUT model have been investigated by focusing on the similarities and differences in individuals’ acceptance of technology in different subject areas. Therefore, this research proposes the following hypotheses about the moderating effects of disciplinary differences:
Discipline has a moderating effect between performance expectancy and learning engagement.
Subject has a moderating effect between effort expectancy and learning engagement.
Subject has a moderating effect between peer influence and learning engagement.
Subject has a moderating effect between learning engagement and learning capabilities.
Taking the above into account, a theoretical model can be constructed about the enhancement of students’ learning capabilities by flipped classroom teaching, which is shown in Fig. 1 .
Theoretical model of flipped classroom teaching to enhance students’ learning capability.
Collection method.
This research employed a questionnaire survey for data collection. The survey was conducted from November to December 2023, targeting students who had participated in flipped classroom learning in three universities in Jiangxi Province. To ensure the representativeness and breadth of the sample, the research team adopted a stratified random sampling method. A total of 450 eligible students were selected as survey respondents. This sampling strategy aimed to ensure that the sample comprehensively reflected the characteristics and opinions of students at different levels, thereby enhancing the reliability and validity of the research.
During the survey process, a detailed questionnaire was distributed to each participating student, and they were offered a cash incentive to encourage honest and thoughtful responses. Through this approach, we hoped to gather high-quality data that truly reflected students’ flipped classroom experiences, providing a solid foundation for subsequent analysis and research.
A total of 450 questionnaires were distributed, with 418 returned (93% response rate) and 413 valid questionnaires ultimately obtained (99% validity rate). Throughout the questionnaire distribution process, we strictly adhered to ethical principles in academic research, particularly regarding obtaining informed consent from participants. All students participating in this research were clearly informed about the purpose, methodology, potential risks, and their rights related to this research.
The questionnaire employed a 7-point Likert scale, with the numbers 1, 2, 3, 4, 5, 6, and 7 representing “strongly disagree,” “disagree,” “somewhat disagree,” “neutral,” “somewhat agree,” “agree,” and “strongly agree,” respectively. “Strongly disagree” indicates that the situation described in the item completely contradicts reality, “strongly agree” indicates complete agreement with reality, and “neutral” indicates a middle ground. The research tool used in this research consists of six parts. The first part is a questionnaire on student background variables (including subject, personal personality, etc.). The second part is a performance expectancy questionnaire, adopting questionnaire content designed by Devisakti and Ramayah 48 and others for performance expectancy. It focuses on specific measurement items for university students’ performance expectancy of flipped classrooms and consists of 4 questions. The third part is an effort expectancy questionnaire, utilizing content designed by Zou et al. 49 and others for effort expectancy. It aims to measure specific aspects of university students’ effort expectancy in flipped classrooms and includes 3 questions. The fourth part is a peer influence questionnaire, based on content designed by Zhonggen and Xiaozhi 50 , Khlaisang et al. 51 , and others. It measures specific aspects of peer influence in flipped classrooms for university students and comprises 3 questions. The fifth part is a learning engagement questionnaire, using content designed by Qureshi et al. 52 . This section includes behavioral, emotional, and cognitive engagement as sub-dimensions and aims to measure specific aspects of university students’ learning engagement in flipped classrooms with 12 questions. The sixth part is a personal capability enhancement questionnaire, employing content designed by Arshad and Akram 53 , C.-H. S. Liu 54 , Baruch and Lin 55 and others. It includes sub-dimensions such as communication skills, cooperation skills, innovation skills, and critical thinking skills, aiming to measure specific aspects of personal capability enhancement in university students after learning in flipped classrooms, with 16 questions.
In this research, data collection resulted in the recovery of 418 questionnaires. After discarding 5 invalid samples, 413 valid samples remained. The basic characteristics are shown in Table 2 . The primary data collected in this research includes subjects and personal personality traits. Among the 418 surveyed students, 199 were from science and engineering, accounting for 48.18% of the total; 214 were from humanities and social sciences, accounting for 51.82%. In terms of personality traits, 201 students were introverted, accounting for 48.67% of the total, while 212 students were extroverted, representing 51.33% of the total.
In this research, SPSS 24.0 and AMOS software were used to perform structural equation modeling analysis on the data to explore the impact of flipped classroom teaching on students’ learning abilities. The analysis primarily consists of two parts: the measurement model (including reliability testing, convergent validity testing, and discriminant validity testing) and the structural model (including model fit analysis, path analysis, mediation effect analysis, and moderation effect analysis).
Questionnaire reliability test.
Table 3 presents the internal consistency of the questionnaire dimensions. The internal consistency (Cronbach’s α) of all dimensions is higher than 0.7, indicating good reliability of the questionnaire sample data. All items will be retained for subsequent analysis.
Table 4 presents the standardized factor loadings of each measurement item, as well as the composite reliability and average variance extracted (AVE) for each dimension. The standardized factor loadings range from 0.648 to 0.851, the composite reliability falls between 0.806 and 0.913, and the AVE is between 0.58 and 0.683. These values meet the criteria established by Fornell and Larcker 56 , indicating good convergence validity of the research.
This research employed the Average Variance Extracted approach to evaluate the discriminant validity. In order to have sufficient discriminant validity, each construct’s square root of the AVE should be greater than the correlation coefficients between the constructs, according to Fornell and Larcker 56 . Table 5 displays the data demonstrating that the square roots of the AVEs for each of the components are higher than the associated correlation coefficients. The good discriminant validity of the model is confirmed by this finding. Make sure every research concept is unique and not just a reflection of other variables in the model by using discriminant validity.
The values of the model fit indices for the structural equation model fall within an acceptable range, as Table 6 shows. Values fewer than 3 or 5 (depending on the criterion employed) are generally considered suggestive of a good fit, therefore the χ 2 /DF value of 1.002 is desirable. A excellent match is shown by the RMSEA value of 0.002, which is significantly less than the typical threshold of 0.08. A good model fit is indicated by the SRMR value of 0.038, which is less than the suggested maximum of 0.08. Both the CFI and the TLI values of 0.999 and 0.999 are near to 1, indicating a very good fit to the data. A good model fit is further confirmed by the values of the GFI and AGFI, which are 0.938 and 0.933, respectively, above the generally recognized criterion of 0.90. Together, these fit indices imply that the structural equation model provides a good match to the data, indicating that the model accurately captures the observed data.
According to the path coefficient analysis results presented in Table 7 and Fig. 2 , Performance Expectancy (PE) (b = 0.244, p < 0.001), Effort Expectancy (EEX) (b = 0.284, p < 0.001), and Peer Influence (PI) (b = 0.242, p < 0.001) all have a significant positive impact on Learning Engagement (ENGA). Furthermore, Learning Engagement (ENGA) also significantly and positively affects the Learning Capability (SKIL) (b = 0.838, p < 0.001). Therefore, hypotheses H1, H2, H3, and H4 are supported.
Path analysis results. Note: * p < 0.05;** p < 0.01;*** p < 0.001, All present standardized regression coefficients.
Furthermore, the findings show that whereas learning engagement accounts for 72.5% of the variance in the improvement of learning skills, performance expectancy, effort expectancy, and peer influence collectively account for 76.6% of the variance in learning engagement. These results underline the significance of Performance Expectancy, Effort Expectancy, and Peer Influence in determining Learning Engagement and validate the research hypothesis. Additionally, they attest to the importance of learning engagement in improving learning capability. This emphasizes how important these factors are to the flipped classroom learning environment and how they all work together to enhance student learning capabilities.
Based on the indirect effects analysis in the mediation model shown in Table 8 , it is noticed that the p -values are significant and the confidence intervals do not include 0 in all three mediated hypothesis paths (PI → ENGA → SKIL, EEX → ENGA → SKIL, and PE → ENGA → SKIL). This indicates that the mediation effects are valid in all cases. Specifically:
PE has a significant indirect effect on the enhancement of SKIL through ENGA, supporting Hypothesis H7. EEX has a significant indirect effect on the enhancement of SKIL through ENGA, supporting Hypothesis H6. PE has a significant indirect effect on the enhancement of SKIL through ENGA, supporting Hypothesis H5.
These findings demonstrate the critical role of ENGA as a mediator in the relationship between PE, EEX, PI, and the enhancement of SKIL. This supports the theoretical framework proposed in the research, highlighting the importance of these constructs in the context of flipped classroom learning environments.
When considering individual personality as a moderating variable, among the 413 respondents, there were 201 introverted students and 212 extroverted students. Table 9 presents the regression coefficient values for the two groups, showing the comparison of slope differences between them. Table 10 displays the moderation effect test of the model. Among the four cross-group comparisons of slopes, the path of PI → ENGA reaches a significant level, indicating that the moderation effect is partially established, thus confirming Hypothesis H8C. From the values in Table 9 , it can be observed that the regression coefficient of PI → ENGA for introverted students is significantly higher than that of extroverted students.
When considering the subject as a moderating variable, among the 413 respondents, there were 199 science and engineering students and 214 humanities and social science students. Table 9 shows the regression coefficient values for the two groups, representing the comparison of slope differences between them. Table 10 presents the moderation effect test of the model. Among the eight cross-group comparisons of slopes, the paths of PE → ENGA and PI → ENGA reach significant levels, indicating that the moderation effect is partially established, thus confirming Hypotheses H9A and H9C. From the values in Table 9 , it can be seen that the regression coefficient of PE → ENGA for science and engineering students is significantly higher than that of humanities and social science students, while the regression coefficient of PI → ENGA for humanities and social science students is significantly higher than that of science and engineering students.
Main research findings and conclusions.
Based on Table 11 , the main findings of this research are as follows:
The four main effect hypotheses, H1, H2, H3, and H4, are validated. Analyses confirming H1, H2, and H3 demonstrate that Performance Expectancy, Effort Expectancy, and Peer Influence significantly positively influence Learning Engagement, aligning with research by Singh 28 , Riddle and Gier 29 , Clark 30 , Jia et al. 31 , and Ruiz 32 . The validation of H4 shows that Learning Engagement significantly positively impacts Learning Capability, consistent with Pekrun et al. 34 , Zimmerman 33 , and Martin and Bolliger 35 .
The three mediation effect hypotheses, H5, H6, and H7, are supported. Analyses for H5, H6, and H7 indicate that Performance Expectancy, Effort Expectancy, and Peer Influence significantly and positively influence Learning Capabilities through Learning Engagement, aligning with findings by Jamaludin and Osman 39 , Wang 40 , Kobayashi 57 , and Nerantzi 41 .
The two moderating effect hypotheses, H8C, H9A, and H9C, are validated. Analysis for H8C shows a significant moderating effect of personality between Peer Influence and Learning Engagement, consistent with Wang et al. 45 . Analyses for H9A and H9C reveal significant moderating effects of the subject field on the paths from Performance Expectancy to Learning Engagement and Peer Influence to Learning Engagement, echoing Fan 47 .
Performance expectancy, effort expectancy, and peer influence notably enhance students’ learning engagement in the flipped classroom teaching. Performance expectancy encourages students to set specific learning goals, work towards achieving them, and believe in their potential to excel academically, thereby increasing their focus and dedication to studies. At the same time, a clear understanding of the required effort makes students appreciate every moment in the flipped classroom, fully investing themselves in the learning experience. Moreover, peer influence is essential as students collaborate and interact, fostering a supportive learning environment that further fuels their motivation and boosts learning engagement.
Students’ learning engagement in the flipped classroom teaching significantly enhances their learning capabilities. Highly engaged students exhibit more active participation in class discussions, thereby refining their oral expression and improving their communication skills for smoother and more efficient interactions. Additionally, the flipped classroom frequently necessitates group work, fostering a sense of teamwork and collaboration among students. Through rigorous reflection and problem-solving exercises, students cultivate critical thinking abilities, allowing them to objectively and comprehensively analyze issues. Moreover, the flipped classroom underscores independent learning and exploration, thus motivating students to tackle problems from diverse perspectives and stimulating their innovative thinking and creativity. As a result, students who demonstrate high engagement in the flipped classroom achieve not only academic excellence but also substantial improvements in their communication, collaboration, critical thinking, and creativity skills.
Performance expectancy, effort expectancy, and peer influence can significantly enhance learning capability by increasing learning engagement. Clear performance expectancy fuels students’ motivation, keeping them laser-focused on learning tasks. To achieve better outcomes, students become proactive in communicating with peers and teachers, thus honing their communication skills. Additionally, they collaborate more with classmates to solve problems, strengthening their teamwork abilities. Effort expectancy helps students understand that to reach their learning goals, consistent effort is key. This realization sharpens their critical thinking and drives them to continuously refine their learning methods. In this journey, students also experiment with novel learning strategies, fostering creativity. Moreover, peer influence plays a pivotal role in shaping the learning environment. Mutual encouragement and imitation among peers create a positive atmosphere, prompting deeper engagement and significantly boosting students’ communication, collaboration, critical thinking, and creativity skills.
Personality plays a significant moderating role in the impact of peer influence on learning engagement, emphasizing the role of individual differences in learning. Our personality traits guide us in choosing our peers, often leading similar-minded students to form study groups. This, in turn, shapes the way peers influence each other and how engaged they are in learning through unique interaction styles. Additionally, our personalities determine our preferred learning methods. Extroverted students, for example, might enjoy learning through group discussions and hands-on activities, while introverted students might prefer solo research and quiet reading. These personality-driven choices further shape our attitudes, motivation, and how we handle learning challenges.
In various subject areas, students’ learning engagement is influenced to differing extents by performance expectancy and peer influence. The level of difficulty of the subjects and the students’ personal interests directly shape their performance expectancy. For instance, the abstract logic in science and engineering subjects may pose a challenge to students, influencing their expectations, while the memorization and comprehension in humanities and social sciences may be relatively easier, leading to more optimistic Performance Expectancy. Additionally, subject characteristics shape Peer Influence, as problem-solving in science and engineering subjects and text interpretation in humanities and social sciences subjects guide different peer interaction patterns. These differences create distinct competitive and collaborative atmospheres among peers, ultimately impacting students’ Learning Engagement.
One of the significant contributions of this research is the substantial enhancement and expansion of the Unified Theory of Acceptance and Use of Technology (UTAUT) model’s application in the flipped classroom environment, providing crucial theoretical and practical insights to the field of education. By integrating variables such as performance expectancy, effort expectancy, and peer influence, and incorporating learning engagement and learning capability into the model, this research innovatively constructs an improved UTAUT model. This not only verifies the direct impact of these variables on learning capability but also reveals the mediating role of learning engagement. The research underscores the importance of performance expectancy, effort expectancy, and peer influence in promoting deep learning, active participation, and enhancing learning capabilities, especially in autonomous learning environments like flipped classrooms. These findings offer empirical evidence for understanding and improving flipped classroom design and provide educators with critical information for designing and implementing more effective strategies.
The second contribution of this research is demonstrating that learning engagement has a significant positive impact on learning capability, highlighting its central role in students’ learning processes. This finding has important implications for educational practice, providing a basis for improving flipped classroom design and emphasizing the necessity of deep learning engagement.
The third contribution lies in exploring how personality and disciplinary backgrounds moderate the effects of peer influence and learning engagement, as well as the impact of performance expectancy on learning engagement. This offers a unique perspective for understanding the influence of individual and disciplinary differences on learning. Furthermore, the model’s innovation lies in its focus on not only the impact of peer influence and performance expectancy on learning engagement but also the moderating effects of personality traits and disciplinary backgrounds. Thus, this research theoretically enhances the empirical foundation of educational psychology and behavioral science, opening new avenues for future educational practice and research. It stimulates research on optimizing learning environments by considering individual and disciplinary characteristics and provides guidance for educators on how to design and improve courses, helping students with different personalities and disciplinary backgrounds better adapt to flipped classrooms.
Strategies for implementing flipped classrooms in practice.
Successfully translating flipped classroom research findings into teaching practice is a gradual process. This process requires teachers to comprehensively and deeply understand the teaching philosophy and methods of the flipped classroom, fully recognizing its significant advantages in enhancing students’ learning initiative and deep engagement. Subsequently, based on specific course content and students’ actual needs, teachers should carefully plan and create preview videos aimed at stimulating students’ curiosity and effectively imparting core knowledge. In the classroom environment, teachers should create diversified interactive learning activities, such as group discussions, role-playing, or experimental operations, aimed at promoting students’ internalization and application of the learned knowledge. Simultaneously, establishing an efficient feedback system is crucial to enable teachers to grasp students’ learning progress and encountered problems in real-time, providing precise guidance and support. This series of processes smoothly transitions the research results of the flipped classroom into practical teaching strategies, thereby significantly improving teaching quality and optimizing students’ learning outcomes.
The research findings of flipped classrooms have profoundly impacted instructional design, emphasizing student-centeredness and focusing on students’ active learning and collaborative inquiry. In different educational backgrounds, the teaching strategies of flipped classrooms need to be adjusted accordingly to adapt to specific teaching environments and student needs. In the basic education stage, where students’ autonomous learning ability is relatively weak, teachers can design more guiding and interesting preview videos. Meanwhile, teacher-student interaction should be strengthened in the classroom to help students better understand and master knowledge. In contrast, in the higher education stage, where students possess stronger autonomous learning and inquiry abilities, teachers can set more challenging and research-oriented preview tasks and classroom activities to stimulate students’ innovative thinking and critical reflection. Through such adjustment strategies, the teaching model of the flipped classroom can be optimized to maximize its teaching effectiveness in different educational backgrounds.
The effective implementation of flipped classrooms relies on advanced educational technology platforms, which directly provides evidence for government investment in educational informatization. Based on this, the government can more targetedly increase support for educational technology innovation, thereby promoting the growth and progress of related industries. Additionally, the emphasis on students’ autonomous learning and collaborative inquiry in flipped classrooms aligns with the concept of quality education advocated in current educational reforms. Therefore, the government can formulate corresponding policies based on this, actively encouraging and guiding schools to explore and practice innovative teaching methods such as flipped classrooms. Moreover, educational equity is also a crucial aspect that cannot be ignored. The government should strive to ensure that all students have equal access to high-quality educational resources and technical support. Through this series of comprehensive and logical policy formulation and implementation, the application of educational technologies such as flipped classrooms in teaching will be more widely promoted and developed in-depth.
Research limitations.
Sample selection: this research randomly selected 450 students as samples from three universities in Jiangxi Province: Jiangxi University of Finance and Economics, Nanchang University, and Jiangxi Normal University. Although this sample size has a certain degree of representativeness, it is still relatively limited and may not fully reflect the actual situation of all college students. Additionally, the sample only comes from universities in Jiangxi Province, and geographical restrictions may affect the universality of the results.
Potential self-selection bias: students willing to participate in the questionnaire survey may have stronger concerns and interests in issues related to learning engagement and learning capabilities. This may cause the sample to deviate from the overall distribution to some extent.
Limitations of result interpretation: this research mainly draws conclusions based on questionnaire surveys and structural equation modeling analysis. However, questionnaire surveys inherently rely on respondents’ self-reports, which may be subject to subjective bias or memory reconstruction.
Expanding sample scope and diversity: future research can consider selecting more diverse universities nationwide as samples to increase the representativeness and universality of the research. Simultaneously, other student groups besides college students, such as middle school students or graduate students, can also be included.
Controlling self-selection bias: to more accurately assess the relationship between learning engagement and learning capabilities, future research can adopt more rigorous sampling methods, such as multi-stage sampling, to further reduce self-selection bias.
Deeply exploring influencing factors: this research initially explores the impact of personality and disciplinary differences on learning engagement. Future research can further delve into other potential influencing factors, such as family background, learning environment, teacher support, etc., to more comprehensively reveal the complex relationship between learning engagement and learning capabilities.
All methods were carried out in accordance with relevant guidelines and regulations. All experimental protocols were approved by the Academic Committee of the School of International Economics and Trade, Jiangxi University of Finance and Economics. Informed consent was obtained from all subjects and/or their legal guardian(s).
The data that support the findings of this research are available from the corresponding author upon reasonable request.
Sohrabi, B. & Iraj, H. Implementing flipped classroom using digital media: A comparison of two demographically different groups perceptions. Comput. Hum. Behav. 60 , 514–524 (2016).
Article Google Scholar
Flores-González, E. & Flores-González, N. The flipped classroom as a tool for learning at High School. J. Crit. Pedagog./Rev. de Pedagog. Crít. 6 (16), 10 (2022).
Google Scholar
Cuetos, M. J. Application of the flipped classroom model to stimulate University students’ learning with online education. J. Technol. Sci. Educ. 13 (1), 368–380 (2023).
Dewi, N. S. S., Padmadewi, N. & Santosa, M. H. The implementation of flipped classroom model in teaching English to Sapta Andika junior high school students in academic year 2019/2020. J. Educ. Res. Eval. 5 (1), 125–135 (2021).
Doung-In, S. Flip your classroom: Reach every student in every class every day. Walailak J. Learn. Innov. 3 (2), 71–78 (2017).
Strayer, J. F. How learning in an inverted classroom influences cooperation, innovation and task orientation. Learn. Environ. Res. 15 , 171–193 (2012).
Semab, T. & Naureen, S. Effects of flipped classroom model on academic achievement of students at elementary level. Pakistan J. Soc. Res. 4 (1), 393–401 (2022).
Roehl, A., Reddy, S. L. & Shannon, G. J. The flipped classroom: An opportunity to engage millennial students through active learning. J. Fam. Consum. Sci. 105 (2), 44 (2013).
Tucker, B. The flipped classroom. Educ. Next 12 (1), 82–83 (2012).
Nicholas, C. Teaching considerations for implementing a flipped classroom approach in postgraduate studies: The case of MBA. In Blended Learning in Practice . 67–80 (Spring, 2021).
Yang, H., Shao, Y., Bai, X., Ma, M., Liu, Y., & Liu, S. The design and application of flipped classroom teaching model based on blended learning: A case study of junior high school information technology course. In Proceedings of the 2023 14th International Conference on E-Education, E-Business, E-Management and E-Learning (2023).
Shofiyah, N., Wulandari, F. E., Mauliana, M. I. & Maghfiroh, L. 5E-based flipped classroom teaching model templates for STEM education. Indones. J. Cult. Community Dev. https://doi.org/10.21070/ijccd.v14i2.917 (2023).
Venkatesh, V., Morris, M. G., Davis, G. B. & Davis, F. D. User acceptance of information technology: Toward a unified view. MIS Q. 27 , 425–478 (2003).
Li, Y. & Zhao, M. The study on the influence factors of intention to continue using MOOCs: integrating UTAUT model and social presence. Interact. Learn. Environ. https://doi.org/10.1080/10494820.2024.2318562 (2024).
Tian, W., Ge, J., Zhao, Y. & Zheng, X. AI Chatbots in Chinese higher education: Adoption, perception, and influence among graduate students—an integrated analysis utilizing UTAUT and ECM models. Front. Psychol. 15 , 1268549 (2024).
Article PubMed PubMed Central Google Scholar
Alyoussef, I. Y. Acceptance of a flipped classroom to improve university students’ learning: An empirical study on the TAM model and the unified theory of acceptance and use of technology (UTAUT). Heliyon 8 (12), e12529 (2022).
Agyei, C. & Razi, Ö. The effect of extended UTAUT model on EFLs’ adaptation to flipped classroom. Educ. Inf. Technol. 27 (2), 1865–1882. https://doi.org/10.1007/s10639-021-10657-2 (2022).
Fredricks, J. A., Blumenfeld, P. C. & Paris, A. H. School engagement: Potential of the concept, state of the evidence. Rev. Educ. Res. 74 (1), 59–109 (2004).
Liu, Y., Zhang, M., Qi, D. & Zhang, Y. Understanding the role of learner engagement in determining MOOCs satisfaction: A self-determination theory perspective. Interact. Learn. Environ. 31 (9), 6084–6098 (2023).
Lan, M. & Hew, K. F. Examining learning engagement in MOOCs: A self-determination theoretical perspective using mixed method. Int. J. Educ. Technol. High. Educ. 17 (1), 7 (2020).
Ojo, A. O., Ravichander, S., Tan, C.N.-L., Anthonysamy, L. & Arasanmi, C. N. Investigating student’s motivation and online learning engagement through the lens of self-determination theory. J. Appl. Res. High. Educ. https://doi.org/10.1108/JARHE-09-2023-0445 (2024).
Voogt, J., Fisser, P., Pareja Roblin, N., Tondeur, J. & van Braak, J. Technological pedagogical content knowledge—A review of the literature. J. Comput. Assist. Learn. 29 (2), 109–121 (2013).
Solichah, H., Jailani, J. & Handayani, M. T. PEMBELAJARAN TIPE NHT UNTUK MENDUKUNG KETRAMPILAN COMMUNICATION, COLLABORATION, CRITICAL THINKING, DAN PROBLEM SOLVING. AKSIOMA J. Program Studi Pendidik. Mat. 11 (2), 381–1390 (2022).
Sari, D. M. M. & Wardhani, A. K. Critical thinking as learning and innovation skill in the 21st century. J. English Lang. Pedagog. 3 (2), 27–34 (2020).
Pratama, S., Haenilah, E. Y. & Adha, M. M. Is there a need for an e-module focused on contextual teaching and learning to improve student critical thinking? A preliminary examination into needs assessment. Int. J. Educ. Stud. Soc. Sci. 2 , 108 (2022).
Nuraini, U., Restuningdiah, N., Sidharta, E. A., & Utami, H. Based learning to improve students’ critical thinking skills in studying business ethics. In International Conference on Strategic Issues of Economics, Business and, Education (ICoSIEBE 2020) (2021)
Zainuddin, Z. & Halili, S. H. Flipped classroom research and trends from different fields of study. Int. Rev. Res. Open Distrib. Learn. 17 (3), 313–340 (2016).
Singh, N. “A Little Flip Goes a Long Way”—The impact of a flipped classroom design on student performance and engagement in a first-year undergraduate economics classroom. Educ. Sci. 10 (11), 319 (2020).
Riddle, E. & Gier, E. Flipped classroom improves student engagement, student performance, and sense of community in a nutritional sciences course (P07-007-19). Curr. Dev. Nutr. 3 (1), nzz032.P07-007-19 (2019).
Article PubMed Central Google Scholar
Clark, K. R. The effects of the flipped model of instruction on student engagement and performance in the secondary mathematics classroom. J. Educ. Online 12 (1), 91–115 (2015).
Jia, C., Hew, K. F., Jiahui, D. & Liuyufeng, L. Towards a fully online flipped classroom model to support student learning outcomes and engagement: A 2-year design-based study. Internet High. Educ. 56 , 100878 (2023).
Ruiz, C. G. The effect of integrating Kahoot! and peer instruction in the Spanish flipped classroom: The student perspective. J. Span. Lang. Teach. 8 (1), 63–78 (2021).
Zimmerman, B. J. Self-efficacy: An essential motive to learn. Contemp. Educ. Psychol. 25 (1), 82–91 (2000).
Article CAS PubMed Google Scholar
Pekrun, R., Goetz, T., Titz, W. & Perry, R. P. Academic emotions in students’ self-regulated learning and achievement: A program of qualitative and quantitative research. Educ. Psychol. 37 (2), 91–105 (2002).
Martin, F. & Bolliger, D. U. Engagement matters: Student perceptions on the importance of engagement strategies in the online learning environment. Online Learn. 22 (1), 205–222 (2018).
Chen, C.-M. & Wu, C.-H. Effects of different video lecture types on sustained attention, emotion, cognitive load, and learning performance. Comput. Educ. 80 , 108–121 (2015).
Buabeng-Andoh, C. Exploring University students’ intention to use mobile learning: A research model approach. Educ. Inf. Technol. 26 (1), 241–256 (2021).
Kuo, Y.-C., Walker, A. E., Belland, B. R., Schroder, K. E. & Kuo, Y.-T. A case study of integrating Interwise: Interaction, internet self-efficacy, and satisfaction in synchronous online learning environments. Int. Rev. Res. Open Distrib. Learn. 15 (1), 161–181 (2014).
Jamaludin, R. & Osman, S. Z. M. The use of a flipped classroom to enhance engagement and promote active learning. J. Educ. Pract. 5 (2), 124–131 (2014).
Wang, F. H. An exploration of online behaviour engagement and achievement in flipped classroom supported by learning management system. Comput. Educ. 114 , 79–91 (2017).
Nerantzi, C. The use of peer instruction and flipped learning to support flexible blended learning during and after the COVID-19 Pandemic. Int. J. Manag. Appl. Res. 7 (2), 184–195 (2020).
Eysenck, H. J. & Eysenck, S. B. The biological basis of personality. In Personality structure and measurement (Psychology Revivals) 49–62 (Routledge, 2013).
Chuang, H. H., Weng, C. Y. & Chen, C. H. Which students benefit most from a flipped classroom approach to language learning?. Br. J. Edu. Technol. 49 (1), 56–68 (2018).
Kim, M., Roh, S. & Ihm, J. The relationship between non-cognitive student attributes and academic achievements in a flipped learning classroom of a pre-dental science course. Korean J. Med. Educ. 30 (4), 339 (2018).
Wang, L., Tian, Y., Lei, Y., & Zhou, Z. The influence of different personality traits on learning achievement in three learning situations. Blended Learning. In New Challenges and Innovative Practices: 10th International Conference, ICBL 2017, (Hong Kong, China, June 27–29, 2017).
Liu, Q. et al. The effectiveness of blended learning in health professions: Systematic review and meta-analysis. J. Med. Internet Res. 18 (1), e2 (2016).
Fan, W. Social influences, school motivation and gender differences: An application of the expectancy-value theory. Educ. Psychol. 31 (2), 157–175 (2011).
Devisakti, A. & Ramayah, T. Sense of belonging and grit in e-learning portal usage in higher education. Interact. Learn. Environ. 31 , 1–15 (2021).
Zou, C., Li, P. & Jin, L. Integrating smartphones in EFL classrooms: Students’ satisfaction and perceived learning performance. Educ. Inf. Technol. 27 (9), 12667–12688 (2022).
Zhonggen, Y. & Xiaozhi, Y. An extended technology acceptance model of a mobile learning technology. Comput. Appl. Eng. Educ. 27 (3), 721–732 (2019).
Khlaisang, J., Songkram, N., Huang, F. & Teo, T. Teachers’ perception of the use of mobile technologies with smart applications to enhance students’ thinking skills: A study among primary school teachers in Thailand. Interact. Learn. Environ. 31 (8), 5037–5058 (2023).
Qureshi, A., Wall, H., Humphries, J. & Balani, A. B. Can personality traits modulate student engagement with learning and their attitude to employability?. Learn. Individ. Differ. 51 , 349–358 (2016).
Arshad, M. & Akram, M. S. Social media adoption by the academic community: Theoretical insights and empirical evidence from developing countries. Int. Rev. Res. Open Distrib. Learn. https://doi.org/10.19173/irrodl.v19i3.3500 (2018).
Liu, C.-H.S. Remodelling progress in tourism and hospitality students’ creativity through social capital and transformational leadership. J. Hosp. Leis. Sport Tour. Educ. 21 , 69–82 (2017).
CAS Google Scholar
Baruch, Y. & Lin, C.-P. All for one, one for all: Coopetition and virtual team performance. Technol. Forecast. Soc. Chang. 79 (6), 1155–1168 (2012).
Fornell, C. & Larcker, D. F. Evaluating structural equation models with unobservable variables and measurement error. J. Mark. Res. 18 (1), 39–50 (1981).
Kobayashi, K. D. Using flipped classroom and virtual field trips to engage students. HortTechnology 27 (4), 458–460 (2017).
Plageras, A., Xenakis, A., Kalovrektis, K. & Vavouyios, D. An application study of the UTAUT methodology for the flipped classroom model adoption by applied sciences and technology teachers. Int. J. Emerg. Technol. Learn. (Online) 18 (2), 190 (2023).
Kissi, P. S., Nat, M. & Armah, R. B. The effects of learning–family conflict, perceived control over time and task-fit technology factors on urban–rural high school students’ acceptance of video-based instruction in flipped learning approach. Educ. Technol. Res. Dev. 66 (6), 1547–1569 (2018).
Tang, Y. & Hew, K. F. Effects of using mobile instant messaging on student behavioral, emotional, and cognitive engagement: A quasi-experimental study. Int. J. Educ. Technol. High. Educ. 19 , 1–22 (2022).
Article CAS Google Scholar
Download references
Project-driven teaching method makes great innovation and practice in the brand planning course (JXJG—17—4—12); To guide college students to participate in the social practice of colleges and universities serving the local economic and social development Explore ——Take Brand Planning as an example(JG2022017).
Authors and affiliations.
School of International Trade and Economics, Jiangxi University of Finance and Economics, Nanchang, 330013, China
Yufan Pan & Wang He
You can also search for this author in PubMed Google Scholar
Conceptualization: Y.P. and W.H.; Data curation: Y.P. and W.H.; Formal analysis: Y.P. and W.H.; Investigation: Y.P. and W.H.; Methodology: Y.P. and W.H.; Project administration: Y.P. and W.H.; Sofware: Y.P. and W.H., Writing—original draf: Y.P. and W.H. ; Writing—review & editing: Y.P. and W.H.
Correspondence to Wang He .
Competing interests.
The authors declare no competing interests.
Publisher's note.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ .
Reprints and permissions
Cite this article.
Pan, Y., He, W. Research on the influencing factors of promoting flipped classroom teaching based on the integrated UTAUT model and learning engagement theory. Sci Rep 14 , 15201 (2024). https://doi.org/10.1038/s41598-024-66214-7
Download citation
Received : 30 January 2024
Accepted : 28 June 2024
Published : 02 July 2024
DOI : https://doi.org/10.1038/s41598-024-66214-7
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative
By submitting a comment you agree to abide by our Terms and Community Guidelines . If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.
Sign up for the Nature Briefing: AI and Robotics newsletter — what matters in AI and robotics research, free to your inbox weekly.
IMAGES
VIDEO
COMMENTS
Step 5: Present your findings. The results of hypothesis testing will be presented in the results and discussion sections of your research paper, dissertation or thesis.. In the results section you should give a brief summary of the data and a summary of the results of your statistical test (for example, the estimated difference between group means and associated p-value).
Medical providers often rely on evidence-based medicine to guide decision-making in practice. Often a research hypothesis is tested with results provided, typically with p values, confidence intervals, or both. Additionally, statistical or research significance is estimated or determined by the investigators. Unfortunately, healthcare providers may have different comfort levels in interpreting ...
HYPOTHESIS TESTING. A clinical trial begins with an assumption or belief, and then proceeds to either prove or disprove this assumption. In statistical terms, this belief or assumption is known as a hypothesis. Counterintuitively, what the researcher believes in (or is trying to prove) is called the "alternate" hypothesis, and the opposite ...
Formulate the Hypotheses: Write your research hypotheses as a null hypothesis (H 0) and an alternative hypothesis (H A).; Data Collection: Gather data specifically aimed at testing the hypothesis.; Conduct A Test: Use a suitable statistical test to analyze your data.; Make a Decision: Based on the statistical test results, decide whether to reject the null hypothesis or fail to reject it.
Hypothesis testing is a scientific method used for making a decision and drawing conclusions by using a statistical approach. It is used to suggest new ideas by testing theories to know whether or not the sample data supports research. A research hypothesis is a predictive statement that has to be tested using scientific methods that join an ...
A hypothesis test is a procedure used in statistics to assess whether a particular viewpoint is likely to be true. They follow a strict protocol, and they generate a 'p-value', on the basis of which a decision is made about the truth of the hypothesis under investigation.All of the routine statistical 'tests' used in research—t-tests, χ 2 tests, Mann-Whitney tests, etc.—are all ...
Hypothesis testing is a crucial procedure to perform when you want to make inferences about a population using a random sample. These inferences include estimating population properties such as the mean, differences between means, proportions, and the relationships between variables. This post provides an overview of statistical hypothesis testing.
However, in order to use hypothesis testing, you need to re-state your research hypothesis as a null and alternative hypothesis. Before you can do this, it is best to consider the process/structure involved in hypothesis testing and what you are measuring. This structure is presented on the next page. Understand the structure of hypothesis ...
6. Write a null hypothesis. If your research involves statistical hypothesis testing, you will also have to write a null hypothesis. The null hypothesis is the default position that there is no association between the variables. The null hypothesis is written as H 0, while the alternative hypothesis is H 1 or H a.
Hypothesis testing involves various statistical tests, each suited to different types of data and research questions. Understanding these tests and knowing when to use them is crucial for accurate ...
Hypothesis testing is as old as the scientific method and is at the heart of the research process. Research exists to validate or disprove assumptions about various phenomena. The process of validation involves testing and it is in this context that we will explore hypothesis testing.
The research methods you use depend on the type of data you need to answer your research question. If you want to measure something or test a hypothesis, use quantitative methods. If you want to explore ideas, thoughts and meanings, use qualitative methods. If you want to analyze a large amount of readily-available data, use secondary data.
A falsifiable hypothesis is a statement, or hypothesis, that can be contradicted with evidence. In empirical (data-driven) research, this evidence will always be obtained through the data. In statistical hypothesis testing, the hypothesis that we formally test is called the null hypothesis.
Hypothesis testing is a statistical method used to determine if there is enough evidence in a sample data to draw conclusions about a population. It involves formulating two competing hypotheses, the null hypothesis (H0) and the alternative hypothesis (Ha), and then collecting data to assess the evidence.
A research hypothesis is an assumption or a tentative explanation for a specific process observed during research. Unlike a guess, research hypothesis is a calculated, educated guess proven or disproven through research methods. ... He published a paper in 1925 that introduced the concept of null hypothesis testing, and he was also the first to ...
The first step in testing hypotheses is the transformation of the research question into a null hypothesis, H 0, and an alternative hypothesis, H A. 6 The null and alternative hypotheses are concise statements, usually in mathematical form, of 2 possible versions of "truth" about the relationship between the predictor of interest and the outcome in the population.
Hypothesis testing is an act in statistics whereby an analyst tests an assumption regarding a population parameter. The methodology employed by the analyst depends on the nature of the data used ...
3. One-Sided vs. Two-Sided Testing. When it's time to test your hypothesis, it's important to leverage the correct testing method. The two most common hypothesis testing methods are one-sided and two-sided tests, or one-tailed and two-tailed tests, respectively. Typically, you'd leverage a one-sided test when you have a strong conviction ...
Hypothesis Testing is a statistical concept to verify the plausibility of a hypothesis that is based on data samples derived from a given population, using two competing hypotheses. ... Alternative Hypothesis (H1) or the research hypothesis states that there is a relationship between two variables (where one variable affects the other). ...
P-Values. The p-value of a hypothesis test is the probability that your sample data would have occurred if you hypothesis were not correct. Traditionally, researchers have used a p-value of 0.05 (a 5% probability that your sample data would have occurred if your hypothesis was wrong) as the threshold for declaring that a hypothesis is true.
Hypothesis-testing (Quantitative hypothesis-testing research) - Quantitative research uses deductive reasoning. - This involves the formation of a hypothesis, collection of data in the investigation of the problem, analysis and use of the data from the investigation, and drawing of conclusions to validate or nullify the hypotheses.
Probability value and types of errors. The probability value, or p value, is the probability of an outcome or research result given the hypothesis.Usually, the probability value is set at 0.05: the null hypothesis will be rejected if the probability value of the statistical test is less than 0.05.
Components of a Formal Hypothesis Test. The null hypothesis is a statement about the value of a population parameter, such as the population mean (µ) or the population proportion (p).It contains the condition of equality and is denoted as H 0 (H-naught).. H 0: µ = 157 or H 0: p = 0.37. The alternative hypothesis is the claim to be tested, the opposite of the null hypothesis.
Consequently, we can reject the null hypothesis that all groups have the same average rank. At least one group has a different average rank than the others. Furthermore, if the three hospital distributions have the same shape, we can conclude that the medians differ. At this point, we might decide to use a post hoc test to compare pairs of ...
It is commonly used in research and data analysis to determine the significance of a sample mean compared to a known population mean. To use the Z.TEST function, the user must input the data range for the sample and the known population mean. ... This allows for efficient and accurate hypothesis testing, making it a valuable tool for decision ...
Putting the diagnostic cart before the horse This blood microbiome story could end here and simply be an interesting example of scientific research homing in on a curious finding, testing a hypothesis, and ultimately refuting it (or at the very least providing strong evidence against it).
This research delves into the flipped classroom teaching methodology, employing the Unified Theory of Acceptance and Use of Technology (UTAUT), learning engagement theory, and the 4C skills ...