What would be examples of numbers 1 through 4? I'm trying to read those but they don't make much sense to me.
Good question! So, here are the four highest quality categories of evidence that I listed: 1) meta-analyses (a report on a group of randomized controlled trial studies that both summarizes overall findings and which critiques the quality of the studies analyzed in the meta-analysis), 2) individual randomized controlled trial studies published in peer reviewed journals, 3) non-randomized but otherwise well-designed studies published in peer reviewed journals, and 4) well designed non-experimental studies published in peer reviewed journals.
Let's start with defining an experimental study. In an experimental study, research basically begins with a hypothesis related to dependent and independent variables. What happens with the dependent variable(s) depends on what happens with the independent variable(s). For example, someone could hypothesize that, all other factors being controlled for, a new method (independent variable) of executing a particular Taekwondo technique produces more power (dependent variable) than the currently used method of executing that technique. Or a new method (independent variable) of training Taekwondo athletes produces a greater number (dependent variable) of Olympic gold medalists than the current method. Well, it's all very well and good to have a hypothesis, but until it's actually tested, it's just a guess. If that hypothesis is stated as fact, it still isn't fact. It's just an opinion. In a well-designed experimental study, extraneous factors are controlled for so they don't confound the results. For example, you wouldn't make your experimental group (the one using the new method for the technique or the one undergoing the new method of training) be the top 30 Taekwondo competitors in a nation and have your control group (the one using the current method of executing the technique or the one undergoing the current method of training) be a group of 30 white belts. You would make sure that the *only* difference, on average, between the experimental group and the control group is the experimental intervention.
Non-experimental studies don't have variables that the researchers are manipulating (e.g., method of technique execution or method of training). They could be, for example, descriptive studies (e.g., a study that uses a survey to assess how many Taekwondo practitioners study in a Kukkiwon school, an ITF school, or an independent school; a study that uses a survey to assess how many Taekwondo practitioners are explicitly interested in self-defense). Such studies generally collect demographic data (e.g., gender, age, years of practice, style of Taekwondo practiced, socio-economic status, race/ethnicity/cultural heritage/nationality, level of education, history of being the victim of assault) so that patterns can be analyzed and described (hence, the term "descriptive study"). For example, maybe a researcher is interested in whether there is a trend related to gender or age (e.g., are more men than women interested in self defense or is it the reverse or are they interested in equal numbers?).
Non-experimental studies can also be qualitative (e.g., they're not about numbers). Qualitative research includes anthropological, ethnographic, and naturalistic research. For example, carefully designed, detailed, and systematic interviews (with open-ended questions that yield direct quotations recorded verbatim) could be conducted with the remaining leaders of the Kwans that worked to unify and develop Taekwondo to explore their original intentions which are often debated today (e.g., the purpose of poomsae, the purpose of sparring, and so on). The resulting quotations would be analyzed word by word using any number of qualitatitve research analysis techniques to discover concepts and patterns related to the researchers’ interest. This kind of qualitative research is extremely labor intensive and time consuming and yields results that are completely different (i.e., of a significantly higher reliability and validity) than a casual or journalistic interview.
So, within the category of experimental studies we have subcategories of randomized and non-randomized studies. In randomized studies, first, the largest pool of participants appropriate to the hypothesis to be tested is identified (e.g., all Kukkiwon-certified dojang owners in the US [to test, say, a hypothesis of whether post-certification annual revenue is higher than pre-certification annual revenue among dojang owners in the US] or all child students in all schools that are part of a particular system of Taekwondo unique to an area in Germany [to study whether a new curriculum used by that system of Taekwondo results in higher rates of retention of child students than a prior curriculum]). Then, participants are selected at random from within that pool.
Because of the challenges (e.g., labor, cost) that true random selection poses, convenience samples are more often used (e.g., Kukkiwon-certified dojang owners in the US who happen to participate in Martial Talk and are willing to participate in the research because they know the researcher who also participates in Martial Talk; the child students in the “future black belts club” of a school that is part of a particular system of Taekwondo unique to an area in Germany who are willing to participate). Non-random convenience samples open up the very real possibility that the research results may not be fully generalizable to the full population of interest; they could apply only to the non-random sample used because it has unique features that the larger population of interest does not consistently have.
Meta-analyses are analyses of analyses. So, in a meta-analysis of a group of randomized controlled trial studies, researchers would first conduct a literature review to identify all studies that have been conducted related to the hypothesis of interest based on a set of clearly defined inclusion criteria. Studies that were of poor quality or otherwise did not adequately match the inclusion criteria would be excluded. Studies that were of high quality and fully matched the inclusion criteria would be included. So, let’s say, there were many studies related to whether a new method (independent variable) of executing a particular Taekwondo technique produces more power (dependent variable) than the currently used method of executing that technique. The researchers conducting a meta-analysis would find and examine all of those studies, determine the flaws of each, exclude individual studies that failed to meet the clearly defined inclusion criteria that were created based on the hypothesis and the principles of scientific inquiry. Then, they would use specialized research techniques to look for and analyze any trends that consistently appear across studies. The results of a meta-analysis can range from finding strong trends that clearly support or refute the hypothesis to finding no consistent trends at all (some of the studies support the hypothesis while some of the studies refute the hypothesis) to determining that enough or enough appropriately designed research has not yet been conducted to determine whether the hypothesis is supported or not.
All of this said, many researchers (including myself) would argue that well-conducted qualitative research should rank higher on the hierarchy of evidence and that the only reason it doesn’t is because of a Western bias for quantification. Some issues simply cannot be researched quantitatively, although they can be researched qualitatively. There is no reason (other than bias) to consider qualititative research on issues that cannot be quantified as inherently inferior.
Also, people can seek to learn from external authority (e.g., the opinions of peers and colleagues, the opinions of leaders or institutions, the results of research of various levels of quality), from internal authority (personal experiential learning), or a synthesis of both. Which routes of learning we seek most tends to vary with where we are in our development and what sources are most valued in our culture.
Taekwondo is ripe for research of all kinds. But research requires money--often lots of money. Societies fund what they value. I hope that anyone in a position to fund and/or conduct research in the field of Taekwondo will do so. There is so much of value waiting to be discovered.
Cynthia