Вы находитесь на странице: 1из 10

See discussions, stats, and author profiles for this publication at: https://www.researchgate.

net/publication/236107029

CONSTRUCTION AND VALIDATION OF AN INSTRUMENT TO EVALUATE


CRITICAL READING OF RESEARCH PAPERS

Article · November 2011

CITATIONS READS

3 12,995

5 authors, including:

Héctor Cobos-Aguilar Patricia Pérez Cortés


Universidad de Monterrey Universidad de Monterrey
48 PUBLICATIONS   62 CITATIONS    23 PUBLICATIONS   29 CITATIONS   

SEE PROFILE SEE PROFILE

Arturo Rafael Vázquez-Guerrero

30 PUBLICATIONS   19 CITATIONS   

SEE PROFILE

Some of the authors of this publication are also working on these related projects:

Health and Law Research View project

Clinical research View project

All content following this page was uploaded by Héctor Cobos-Aguilar on 21 May 2014.

The user has requested enhancement of the downloaded file.


CONSTRUCTION AND VALIDATION OF AN INSTRUMENT TO
EVALUATE CRITICAL READING OF RESEARCH PAPERS
Héctor Cobos Aguilar1, Patricia Pérez Cortés2, Arturo Vázquez Guerrero3,
3 3
Cynthia D. Cobos Herrera Jasel A. Tapia Orozco
1
Universidad de Monterrey, Internship Program Coordinator (MEXICO)
2
Universidad de Monterrey, Clinical Sciences Department (MEXICO)
3
Universidad de Monterrey, Surgical and Medical Residents.
hector.cobos@udem.edu.mx, patricia.perez@udem.edu.mx

Abstract
Introduction: In order to evaluate complex skills as research, valid and reliable instruments must be
constructed. Purpose: To report the process of construction, validity and reliability of an instrument to
assess the critical reading of medical papers. Methodology: An electronic research of published
medical papers was performed. After several revisions six research designs were finally selected:
validation of instruments, surveys, cases and controls, diagnostic tests, clinical trials, and cohorts. An
abstract from each paper was made in Spanish emphasizing methodological issues (designs,
instruments used, sample size, blind evaluations, bias, the selection of statistical methods, and
others). From the abstract several items were constructed which were grouped in headings related to
the methodological issues and the indicators of critical reading used: Interpretation (hidden aspects),
judgments (evaluating the best methods), and proposals (suggesting better methods) were explored.
Each item had a true o false answer. An initial 157 item was constructed. Validation: Five experts were
invited to evaluate, in two independent rounds, if the instrument had theory, construction or content
validity. The experts suggested changes that were made in these two rounds. At the same time they
answered the final items as true or false with an agreement 4/5 or 5/5 to be acceptable. A pilot
application was made to students for further adjustments. The final version of the instrument had 108
items, 36 for each indicator and 18 for each design, the half of the correct answers were true and the
other half were false. Application: The instrument was resolved by three different experienced groups
in critical reading G1 (n: 7) professors of medical specialties G2 (n: 23) medical interns in an active
research course and G3 (n: 24) medical students without any research course. The answers were
determined as True, False or, Don’t Know. Each correct answer (True or False) added one point and
each incorrect answer subtracted one point. Questions answered with Do not know did not add or
subtract any points. This is how the final grade, which was expressed as group medians, was
determined. The grading was performed through an electronic system especially created to minimize
data capturing errors and it was carried out blind by staff unrelated to the research. The data were
analyzed with version 15 of the SPSS program. A random level was determined in the
groups.Statistics: The intra e inter agreement raters were obtained. 20-KR was used to calculate the
reliability. Kruskal-Wallis test were used for comparing the groups . Spearman test relating school
average and global medians were calculated. The random level was determined as well. Results. The
raters agreed in validity of instrument. The whole reliability was 0.75. The inter agreement value was
0.82 and the intra agreement was 0.80. The global median for G1 was 62, for G2: 28 and for G3:11.
The random level calculated was 17. All the results were significant for G1 <=0.01 for global, the three
indicators and the six designs. The level of answer random was 0% per G1, 13% for G2 and 83% for
G3. Comments: The validity and reliability of the instrument is good to evaluate the critical reading.
The contrast results among different experienced groups and the rand values sustained its reliability.
The instruments should be valid and reliable to assess this complex skill.
Keywords: Validity, reliability instrument, critical reading.

1 INTRODUCTION
It is currently pressing to use methods that will guarantee consistency in observations when searching
for variables by means of instruments - scales, questionnaires etc. – and that frequently depend on
the designer’s experience. Although usually referring to qualitative indicators [1], their construction
depends on reviewers mastering the elements to be evaluated in their professional practice. Due to
the need to elaborate observation or measurement instruments such as psychological or psychometric

Proceedings of ICERI2011 Conference. ISBN: 978-84-615-3324-4


14th-16th November 2011, Madrid, Spain. 005286
scales, methodologically rigorous criteria were established early on in order to guarantee
observational consistency [2,3]. Aspects that can be explored with instruments are as broad as the
fields requiring investigation. Their use is more extensive in disciplines exploring more qualitative
factors. Among them, are those related to education and especially the evaluation of learning.
However, the processes used for their elaboration in the research realm, are not referred to in our
country’s different publications [4].
Advances in methodology and its contextual demands have led to the search of quality criteria for the
elaboration of instruments, including their intrinsic and extrinsic values and their consistency [5,6]. As
these qualities were refined, the authors or whoever constructed the instruments became increasingly
more demanding in these aspects [7,8].
Intrinsic validity is mainly dependent on theoretical validity that in turn, supports the ideological
substrate upon which the instrument’s quest is based; the construct’s validity refers to the grammatical
and semantic structure of the instrument’s components and content validity represents a balanced
form of the criteria or domains that conform the theoretical framework of the attribute under
investigation.
Extrinsic validity refers to concurrent validity that in turn, alludes to similar results obtained with a
different instrument measuring the same attributes under the assumption that both are valid and
consistent. Predictive validity refers to the measurements obtained by the respondents at a later date
and that change in direct or inverse relation when the measured attribute mutates [9]. Simultaneously,
different procedures that will guarantee internal validity are progressively introduced. One is the
Delphos technique [10] that has proven adequate in assuring intrinsic validity [11].
According to the number of answers per item in a scale, consistency is determined by variance
analysis using Kuder-Richardson’s formula for two answers and Cronbach’s alpha for three or more
answers; these are easy to obtain in view of currently used informational technology that even allow
for specifications including confidence intervals [12]. The elaboration of instruments in such varied
areas requires the assistance of ‘experts” that may, at least partly guarantee their quality.
As of now and in our environment (social sciences and even education in order to evaluate skills such
as learning), it is still difficult to elaborate an instrument for initial evaluation with the qualitative
characteristics previously mentioned and to find validators (reviewers) considered experts in the area
under study in order to obtain the best results. There are no universally accepted criteria for their
selection and even those identified are occasionally controversial.
Selection criteria are qualitative and include experience in the corresponding professional realm, a
solid academic training, past elaboration or validation of research instruments, publication of
observations etc.; however, no consensus has been reached. There are few reports in Mexico
evaluating the consistency of validators as experts and this underscores the difficulty –fortunate- in
obtaining unanimous agreement on the correct answers exploring certain attributes once an
instrument’s intrinsic qualities have been ascertained.
Therefore, attempts have been made to temper these contradictory aspects inherent to instrument
elaboration and that will guarantee an independent and reliable evaluation.
In our milieu, Viniegra has long proposed an educational perspective focusing on participation and the
autonomous elaboration of knowledge by the student [15,16,17]. He proposes an epistemological
viewpoint that among other aspects, recommends the construction and application of valid and reliable
instruments in order to obtain rigorous observations with guaranteed reliability. From this perspective,
every professor should elaborate his research instruments based on his professional and docent
experience, thus instilling his unique personal vision.
Based on this educational stance, numerous studies on instrument elaboration in the area of
education have yielded several procedures attempting to assure their validity; these include exploring
clinical aptitudes, critical reading of factual or educational research papers [18] at different complexity
levels, different scenarios and specialties as well as personal positions towards research and even
medicine per se [19, 20, 21, 22, 23, 24].
In the context of this proposal, an attempt is made for the student to elaborate his own knowledge in
an autonomous manner; it emphasizes the development of various methodological aptitudes
dependent on critical abilities and that include critical reading of factual or educational research
papers. Several reports have shown adequate results when students are spurred to develop their own
knowledge with the use of reading guidelines specifically constructed to critique the most relevant

005287
methodological aspects of the different research study designs presented in published research
papers [25].
In researching this methodological aptitude, every report is accompanied by a reference on the
construction of the instrument pretending to measure the development of critical reading among
students. Instruments are reported as validated and consistency reports are also adapted.
Nevertheless, they do not meticulously refer to the validation process and hence, do not contribute
previous experience to the construction process and little is known on the variability of the answers
provided by experts to the items being explored [26]. In other research areas with more empirical
viewpoints, this exercise is common [27].
This study is aimed at identifying inter-observer and intra-observer variability among experts in the
construction of instruments designed to measure critical reading of research papers.

1.1 Objectives
Describe the validation process and consistency of an instrument evaluating critical reading of
research papers. Determine intra and inter-observer agreement among experts designing an
instrument to evaluate critical reading development.

2 METHODS

2.1 Design
Validation and consistency of a measuring instrument.

2.2 Variables
Independent: experience in the validation of instruments evaluating critical reading of factual research
and critical reading of research. Dependent: intra and inter-observer consistency values. The following
hypotheses were proposed: a) inter and intra-observer response consensus among experts evaluating
critical reading of research is above 0.60. Independent variable: experience in the validation of
instruments evaluating critical reading of factual research and critical reading of research. Dependent
variable: results obtained by expert consensus.

2.3 Population
Five experts in critical reading of research and that collaborated in the instrument’s validation.

2.4 Sample selection


For the experts’ convenience: individuals with experience in critical reading of research and that
fulfilled at least, the following first four conditions: a) coursed, as students, a critical reading of
research course lasting over 40 hours b) conducted more than two courses on critical reading of
research as professors or adjunct professors c) elaborated one or more instruments evaluating critical
reading of factual research d) presented more than two research studies in factual or educational
forums at the local or national level e) published, at least, one research paper f) completed a master’s
degree in science or education. A written invitation was forwarded to the experts, suggesting they
collaborate in our study and detailing their involvement; they all accepted.

2.5 Instrument elaboration.


The author reviewed several papers focusing on their methodological qualities and finally selected six
that were representative of the most common study designs used by physicians; each paper was
summarized and included the most relevant methodology aspects. The selected study designs
included: case-control, survey, instrument elaboration, diagnostic tests, clinical trials and cohort
studies. For each of the six summaries, a series of postulates relating to the theoretical framework of
critical reading proposed by Viniegra were elaborated. Viniegra’s postulates include: Interpretation: To
decipher what is implicit in the text such as underlying ideas and theories; the problem under scrutiny;
hypothesis; type of study and its design; measuring instrument(s); type of data collection. Judgment:
To differentiate strong from weak points, that which is valid from the spurious, the relevant from the
trivial and the pertinent from the non-pertinent in the research problem; the conclusions, the type of

005288
study and its design, measuring instruments, the conditions of data application and collection and the
analysis of the information. Proposal: To think of components, elements or theoretical or
methodological procedures that would further strengthen and validate or emphasize novelty, relevance
or pertinence of the critiqued research paper. Based on these three indicators, several items exploring
the papers’ methodology were generated. Grammar rules were strictly followed in order to avoid
confusion when reading the summaries. Items were to be answered as “true” or “false” according to
the expert’s experience. They also had to evaluate the theoretical validity, construction and content of
each summary, assertion and corresponding items, as well as the overall instrument.
From the beginning, we decided to balance the entire instrument with: 108 items, 18 for each study
design and 36 for each indicator. The items considered “false” or “true” were also balanced. In order to
guarantee that number, the instrument initially included 157 items.

2.6 Validation by experts.


In order to guarantee the instrument’s intrinsic value according to Delphos’ technique, two validation
rounds were conducted by the selected experts according to the previously mentioned requirements.
Due to transportation problems, the instrument was first individually e-mailed to the experts so none of
the evaluators knew the other reviewers. After this initial review, issues relating to the instrument’s
wording, syntax, the construct’s and the content’s validity, were raised; experts were also asked to
answer the items they considered sufficiently appropriate to be part of the final instrument, as “correct”
or “incorrect”. The author separated the items that the five evaluators were in complete agreement
upon, in terms of construction and answers (“correct” or “incorrect”).
After completing the suggested modifications to the summaries’ and items’ wording and content,
experts were contacted a second time in order to conduct an independent second evaluation. On this
occasion, the items that had not been previously agreed upon in terms of construction or answers
were presented anew. There were no further observations regarding the summaries’ construction and
only a few regarding the presented items. After this round, the items that were not agreed upon or that
were not considered adequate were discarded.
Finally, the author elaborated the instrument with the 108 items that had been previously balanced
among the three indicators: 36 for interpretation, judgment and proposal, 18 for the previously
mentioned study designs and 54 to be answered as “true” or “false”.

2.7 Pilot test


Once completed, the instrument was presented as a pilot test to a group of recently admitted hospital
interns and was not modified.

2.8 Inter and intra-observer agreement.


To evaluate inter-evaluator agreement, the independent answers of all five experts to the 108 items of
the final instrument were compared. Answers were requested in the form of “correct” or “incorrect”.
The expert agreement index was used.
To determine intra-evaluator agreement, the instrument was independently applied in person to all five
experts a second time, 3 weeks after the initial application. The experts were unaware of this second
application and were only informed once they were gathered. It was applied in person in order to avoid
remembering the answers by consulting electronic records. The degree of answer agreement on both
occasions was determined [28]. The correct or incorrect answers were graded as 5/5 or 4/5 according
to their responses during the first and second rounds.

2.9 Consistency
The instrument’s internal consistency was determined with Kuder-Richardson’s formula [20]. The
results of three groups with different experience in critical reading of factual research were compared.
These groups were selected and conformed as follows: a) a natural group of medical students
beginning the fifth semester of their training and with no previous exposure to critical reading of
research papers (n=24), b) a group of sixth year students (social service) partially exposed to
educational strategies developing critical reading of research papers (n=23) and c) a group of full-time
professors with a medical specialty and with previous experience in critical reading (n=7).

005289
The instrument’s application was programmed independently for all three groups. Each of the
instrument’s items were answered as “true”, “false” or “I don’t know” and they were graded by adding
one point for each correct answer and subtracting one point for each incorrect answer. An “I don’t
know” answer neither subtracted nor added points. An individual extraneous to the study graded the
instrument after the first and second application as well as after application to the three groups; it was
subsequently submitted for analysis with a capture program especially designed to minimize errors.
The three groups were compared with medians and ranges of the interpretation, judgment and
proposal indicators as well as of the included study designs. Kruskal-Wallis’ test was used to compare
the three groups’ medians. The answers randomness factor in the three groups was determined to
classify the results into various categories (random, very low, low, intermediate, high, very high) [29].

3 RESULTS
Of the 157 items initially proposed, 68% agreement was obtained in 106 according to the answers
(4/5, 5/5). The 51 items that had not been agreed upon were sent for a second validation and
agreement was reached for 29 items. The instrument finally included the items that fulfilled the
previous criterion in both rounds and comprised 132 of the original total (84%). The rest of the items
were discarded.
Of the 132, another 24 were rejected. The elimination criterion was determined according to the
experts’ comments on the item’s ambiguity or lack of clarity or because the answers were deemed too
easy.
The final instrument included 108 balanced items, 18 of which focused on study design and 36 on
interpretation, judgment and proposals; half of the answers were “true” and half were ‘false”.
According to Kuder-Richardson’s formula, consistency was 0.75. Inter-evaluator agreement was 0.82
for the five questions in both rounds (0.38 for 5/5 and 0.44 for 4/5). Intra-observer agreement ranged
from 0.70 in evaluator five to 0.9 in evaluator three. The rest scored above 0.8.
Fig. 1 shows the results of the comparison between the three groups and the significant difference
obtained in the medians.

Table 1. Results of the three contrast groups in the three subcomponents and six study designs.
Overall Int Jud Pro CC Coh Dxt Surv RCT Instr

Grp. 108 18 18   18   18   18   18   18   18   18  

1 (n=7) 62 16 24 25 10 10 10 12 12 12

2 (n=23) 28 8 14 10 6 5 7 4 2 7

3 (n=24) 11 2 5 7 3 0 2 0 1 4

0.01 0.01   0.01   0.01   0.01   0.01   0.01   0.01   0.01   0.01  

* Kruskal-Wallis. Mann-Witney U <0.05 favoring grp. 1 in all cases. NS χ 2 when comparing all three indicators.
Fig. 2 shows the asymmetry favoring group 1 when the three student groups’ results were translated
into levels.

005290
Fig. 2. Levels reached by students after excluding randomness factor.    
Percentages are in parentheses.
Groups
Level Range 1 (n=7) 2(n=23) 3(n=24)
Very High >86
High 69-85 2 (0.30)
Intermediate 52-68 5 (0.70) 1 (0.04)
Low 35-51 8 (0.35)
Very Low 18-34 11 (0.48) 4 (0.17)
Random <17 3 (0.13) 20 (0.83)

4 COMMENTS
We believe that this study presents the key aspects of methodological rigor required for the
construction, validation and consistency determination of an instrument evaluating critical reading of
factual research papers.
The authors are professors actively involved in educational strategies condoning participation when
developing critical reading of research papers; they have been continuously involved in the elaboration
of these instruments for the past 15 years and they use new reading guidelines developed for each
course.
The first elaborated instruments evaluating the development of critical reading only included four study
designs and we now retrospectively believe that their content validity has increased since students
face research problems requiring different methodological design approaches.
This instrument is balanced for the indicators of the subcomponents of critical reading as well as in the
number of items for each summary. We added facets on the construction and validation of instruments
and surveys since we believe they are necessary to evaluate the students’ progress. “True” and “false”
answers were also balanced in order to prevent an answer bias. This could provide further reliability to
the obtained results.
The validation process conducted by the evaluators also contributed to the instrument’s theoretic,
construct and content validity. The first, supported by Viniegra’s contributions relate to the critical
analysis of experience and the development of complex skills, including critical reading particularly of
factual research papers. Construct and content validity were assured with Delphi’s technique as
required in these cases.
The results obtained in consistency permit a more rigorous evaluation of this complex skill.
One of the problems we faced was the selection of so-called experts that we have named evaluators.
There are no references relating to this point; in general, when publications refer to evaluators, they
only require that they have experience in the area of interest although this does not assure the
expert’s quality. From an educational viewpoint, the complex skill of critical experience is not obtained
through academic degrees but rather through the continuous, decanted, reflexive and progressively
challenging exercise of research critique, factual in this case, and closely related to docent activities
and the development of educational strategies leading to the students’ autonomous acquisition of
knowledge through the elaboration of reading guidelines that foster critical thinking.
The criteria for tutor selection in this article were not validated per se nor their consistency determined,
but the agreement results are significantly above the hypothesis’ expectations and confirm that they
are appropriate indicators.
Another indirect fact favoring evaluator selection was the high rate of agreement in the first and
second rounds, leading to an instrument with a 5/5 and 4/5 agreement ratio for all correct or incorrect
answers (100%).
However, intra-evaluator agreement –this article’s main objective- further confirms this point. We
believe that the fact that evaluators were unaware of the second evaluation of the correct and
incorrect answers as well as the time elapsed between evaluations, increased the study’s value. The
actual distance between evaluators may hinder collaboration in instrument evaluation; this can be

005291
improved using electronic media but evaluators should be blinded to the previously reported results
when intra-evaluator agreement is under scrutiny.
Agreement values exceeded those presented in the hypothesis (0.60) so selection criteria do confirm
individual experience (0.72 to 0.90) that we consider adequate. As mentioned in the introduction,
variability among expert evaluators is not frequently studied and we consider it to be indispensable,
particularly when the criteria defining an expert are not well established.
On the other hand, we also determined the kappa value with confidence intervals and they were not
congruent with the results obtained in our intra-observer agreements. This apparent contradiction has
a theoretical basis. The kappa statistic is not the most appropriate in this case because it
predominantly focuses on the prevalence of a research aspect unrelated to education and learning;
thus, we chose agreement as the measurement evaluating this point.
Table 1emphasizes this point, since the contrast between the three groups connote the students’ lack
of experience in this transcendental skill required for their future medical practice; it is most developed
in the group of professors.
The instrument also reveals the authors’ progress and suggests that the elaborated instruments will be
permanently complex.
Table 2 confirms the distance between the three groups when the randomness factor is excluded. The
results support the consistency values determined with Kuder-Richardson’s formula. The different
expression of results in three strata clearly relates to experience in research and supports our
validation results since the instrument does bring to light evidence reflecting the development of this
complex ability.
More precise studies investigating evaluator consistency will also increase their skills and foster the
availability of highly trained personnel that can play a role in these activities that are not limited to
evaluation instruments, but may also encompass research protocols in different institutional settings.

REFERENCES
[1] Pope C, Ziebland S, Mays N. Qualitative research in health care. Analyzing qualitative data.
BMJ. 2000;320:114-16.

[2] Bland JM, Altman DG. Validating scales and indexes. BMJ 2002;324:306-7.

[3] Prieto G, Muñiz J. Un modelo para evaluar la calidad de los test en España. Papeles del
psicólogo 2000;67:65-77.

[4] Boynton PM, Greenhalg T. Selecting, designing, and developing your questionnaire. BMJ
2004;328;1312-5.

[5] Merino SC, Lautenschlager GJ. Comparación Estadística de la Confiabilidad Alfa de Cronbach:
Aplicaciones en la Medición Educacional y Psicológica. Revista de Psicología de la Universidad
de Chile. 2003; XII (2):127-136.

[6] Briones G. Validez y confiabilidad des pruebas de medición. En: Métodos y técnicas de
investigación para las ciencias sociales. Ed. Trillás.México, D.F. 2003.

[7] Boynton PM, Greenhalgh T. Hands-on guide to questionnaire research. Selecting, designing,
and developing your questionnaire. BMJ 2004;328:1312–5

[8] Mays N, Pope C. Qualitative research in health care. Assessing quality in qualitative research.
BMJ. 2000; 320:50-2.

[9] Bisquerra R. Métodos de investigación educativa Guía Práctica. Ed. Ceac. Barcelona, España.
1989.

[10] Jones J, Hunter D. Consensus methods for medical and health services research. BMJ
1995;311:376-380.

005292
[11] Landeta, Jon. El método Delphi. Una Técnica de previsión para la incertidumbre. Ariel.
Barcelona y Godet, Michel. Manuel de Prospective Strategique. Dunod. Paris (1999).

[12] Ledesma R. Alpha CI: un programa de cálculo de intervalos de confianza para el coeficiente
alfa de Cronbach. Psico-USF, 2004;9(1):31-37.

[13] Pless IB, When reviers disagree. Inj Prev 2006;12:211.

[14] Van Rooyen S, Godlee F, Evans S, Black N, Smith R. Effect of open review on quality of
reviews and on reviewers’ recommendations: a randomized trial. BMJ 1999;318:23-27 .

[15] Viniegra VL. El desafío de la educación en el IMSS: cómo constituirse en la avanzada de la


superación institucional. Rev Med IMSS 2005;43(4):305-321.

[16] Viniegra VL. Educación y crítica. El proceso de elaboración de conocimiento. México: Paidós;
2002.

[17] Viniegra-Velázquez L. La experiencia reflexiva y la educación. Rev.Inv.Clin. 2008; 60(2):133-


156.

[18] Cobos AH, Insfrán SMD, Pérez CP, Elizaldi LNE, Hernández DE, Barrera MJ. Lectura crítica de
investigación en el internado de pregrado en hospitales generales. Rev Med IMSS 2005; 43 (2):
117-124.

[19] Soler HE, Sabido SC, Sainz VL, Mendoza SH, Gil AI, González SR.
Confiabilidad de un instrumento para evaluar la aptitud clínica en residentes de medicina
familiar. Arch Med Fam 2005; 7 (1): 14-17.

[20] Pantoja PM, Barrera M J, Insfrán SMD. Instrumento para evaluar aptitud clínica en
anestesiología. Rev Med IMSS 2003; 41 (1): 15-22.

[21] Rivera CJM, Leyva GFA, Leyva SCA. Desarrollo de la aptitud clínica de médicos internos de
pregrado en anemias carenciales mediante una estrategia educativa promotora de la
participación. Rev Invest Clin 2005; 57 (6): 784-793.

[22] Soler HE, Sabido SC, Sainz VL, Mendoza SH, Gil AI, González SR. Confiabilidad de un
instrumento para evaluar la aptitud clínica en residentes de medicina familiar. Arch Med Fam
2005; 7 (1): 14-17.

[23] Pantoja PM, Barrera M J, Insfrán SMD. Instrumento para evaluar aptitud clínica en
anestesiología. Rev Med IMSS 2003; 41 (1): 15-22.

[24] Rivera CJM, Leyva GFA, Leyva SCA. Desarrollo de la aptitud clínica de médicos internos de
pregrado en anemias carenciales mediante una estrategia educativa promotora de la
participación. Rev Invest Clin 2005; 57 (6): 784-793.

[25] Matus MR, Leyva GF, Viniegra VL, Lectura crítica en estudiantes de enfermería: efectos de una
estrategia educativa. Rev. Enferm. IMSS. 2002; 10 (2): 67- 72.

[26] Robles P A, Viniegra V L. Capacidad de lectura crítica de investigación clínica en grupos de


residentes médicos. Rev Invest Clin 1997; 49: 117-22.

[27] López CJM, Araiza AR, Rodríguez MR, Murguía MC. Construcción y validación inicial de un
instrumento para medir el estilo de vida en pacientes con diabetes mellitus tipo 2.. SPM
2003;45(4)259-268.

[28] Siegel S, Castellan JS. Capítulo 8. Medidas de asociación y sus pruebas de significación. En:
Estadística no paramétrica aplicada a las ciencias de la conducta. 4ª. Ed. Ed. Trillás. México,
D.F. 2005. p: 325-333.

[29] Pérez-Padilla J.R., Viniegra V.L. Método para calcular la distribución de las calificaciones
esperadas por azar en un examen del tipo falso, verdadero, no sé. Rev.Inv.Clin. 1989; 41:375-9

005293
ADDENDUM 1 (FRAGMENT)
Role of the active and passive smoker in lung cancer
Lung cancer may have multiple etiologies. The aim of the study is to identify its relation with passive
and active smoking among other risk factors. 385 patients (cases) with a pathology-confirmed
diagnosis of lung cancer were studied in tertiary care hospitals. 898 healthy controls recruited in the
community were also studied with multi-staged surveys, in a randomized manner and stratified by age.
Cases and controls (1:2 ratio) completed a questionnaire applied by two trained interviewers that were
unaware of the study’s purpose. It included questions on active or passive tobacco exposure, number
of smoked cigarettes per day and age at which smoking began (if he smoked). Data were captured
twice and analyzed with SPSS. Bivariate odds ratios (OR) and 95% confidence intervals (CI) were
used for statistical analysis.

Table 1. Results of exposure.


Factor OR CI(0.95)
Smoker No 1.0
Yes 4.0 2.9-5.5
Previous smoker 3.2 2.2-4.5
Number of cigarrettes
Smokes 2-4 cigs./day 1.4 0.9-2.6
Smokes 5-10 cigs./day 4.4 2.9-6.6
Passive smoker home 1.8 1.3-2.4
Began at age/years
20 to 65 3.4 2.4-5.0
15 to 19 4.0 2.7-5.8
5 to 14 5.3 3.5-7.9

IN TERMS OF THE STUDY’S DESIGN


1. ___ It seeks causality.
2. ___ It is longitudinal.
THE FOLLOWING PROCEDURES INCREASE THE STUDY’S VALIDITY

3. ___ The case-control ratio.


4. ___ The interviewers’ lack of awareness of the study’s purpose.
5. ___ Capturing data twice.
6. ___ Using confidence intervals (CI).
ACCORDING TO THE RESULTS
7. ___ Smoking 2 to 4 cigarettes per day suggests the early development of lung cancer.
8. ___ Passive smokers have as much risk of developing cancer than those who smoke 2 to 4
cigarettes per day.
THE FOLLOWING ARE PROPOSALS BY THE AUTHOR
9. ___ Determine the sample size.
10. ___ Use relative risk instead of OR.
11. ___ Conduct a multivariate analysis.
12. ___ Avoid concentration bias.

005294

View publication stats

Вам также может понравиться