This is a guest post from Melissa Kardish - a PhD student at UC Davis - writing about a recent paper from work she did at her prior position. The citation for the paper she is writing about is below:
Kardish MR, Mueller UG, Amador-Vargas S, Dietrich EI, Ma R, Barrett B and Fang C-C (2015) Blind trust in unblinded observation in Ecology, Evolution, and Behavior. Front. Ecol. Evol. 3:51. doi: 10.3389/fevo.2015.00051
Here is her post.
Blind trust in unblinded observation in Ecology, Evolution and Behavior
We recently published our study in Frontiers
in Ecology and Evolution where we found that a remarkable number of studies
that could be affected by observer bias didn’t indicate whether or not they
blinded their research. In fact only 13.3% of studies reported this:
We tried to make this a very transparent study. In addition to
journal level data in the main article, we include in our supplemental material
a table with the score for every article we read for this study (a summary of
these scores per journal can be found in Figure S2 included here). If anything,
our results under-represent the amount of studies that could have been scored
blind (the real underreporting/underuse of blind observation is probably less
than the 13.3% we report). For instance, we did not assess that there was
potential for bias in the scoring of microsatellite markers (scored as unlikely
to have observer bias). However, we did identify one study which was based on
data from microsatellites which did blindly score their markers and report this
scoring in their methods (and was therefore scored as “blind” in our
study). We also considered a study
blind in its entirety for the purposes of our scoring if only one aspect is
reported even if other experiments could also have been influenced by observer
bias (Check out our supplemental methods for more ways we
conservatively scored in our study).
We recognize that not all EEB studies can be blinded due to a
variety of logistical or hypothesis driven reasons; however, we encourage such
studies to accurately report this rationale and consider and attempt to
minimize observer bias when designing experiments.
Thus far we have had a great response from the surveyed
journals. Many of them have notified their editors about the lack of blind
observation that we found reported in their journal. One journal has even
notified us of plans already in place to address this issue at their next editorial
board meeting.
We’re excited to have this work out there and hope this will
inspire people to blind their studies and accurately report the science they
are doing. We’re also excited to have the study published in an open-access
format where we hope the encouragement for blind observation can reach all
levels of science. Finally, as reporting of science in our fields improves in
the coming years, we hope this study can serve as a template to address other
potential concerns in experimental design and reporting.
For our article check out:
http://journal.frontiersin.org/article/10.3389/fevo.2015.00051/full
For our supplemental text and methods:
http://journal.frontiersin.org/file/downloadfile/10238/octet-stream/Data%20Sheet%201.PDF/711/2/142945
For our data table with each of the articles we surveyed:
http://journal.frontiersin.org/file/downloadfile/10235/octet-stream/Table%201.XLSX/711/1/142945
No comments:
Post a Comment