CMST 4N03 – PRODUCING AND VIEWING THE NEWS
DR. ALEXANDRE SEVIGNY
LECTURE 6
Measurement Techniques
- Defining Measurement
- Stevens (1951): “Measurmeent is the assignment of numerals to objects or revents according to rules.”
- In CA: our objects and events are messages units.
- Measurement theory (classic test theory) assumes that there is a “true” value for each variable on each unit.
- We attempt, but are often unable to discover the exact true value due to a variety of sources of error:
m = t + e
A measured score is the result of a true score and an error score. We attempt to minimize the e .
There is a true number of acts of violence in a hockey game, as defined by our codebook. However, rarely will we be able to capture them all in our study because of various reason: coder misinterpretation, coer error, coder fatigue, and recording errors.
These errors can be:
- Random Errors: unsystematics
- Nonrandom Errors: also called bias, it is a threat to accuracy. It involves systematic bias to measuring procedure.
- Variability, Reliability, Accuracy and Precision
- Reliability: the extent to which a measuring procedure yields the same results on repeated trials (checking each other’s coding)
- Validity: the extent to which a measuring procedure represents the intended, and only the intended, concept. “Are we masuring what we want to measure?”
- Accuracy: the extent to which a measuring procedure is free of bias.
- Precision: the fineness of distinction made between categories or levels of a measure.
- Eg.: measuring a character’s age in years is more precise than measuring in decades.
- Extreme precision can be counter-productive: measuring her age in days is likely more precise, but too tedious and error prone to be useful.
- How the standards interrelate:
- See diagram on Pg 114.
- Target A – what we always hope to achieve – a valid measure, one that is reliable, accurate and precise.
- Target B – good reliability and precision but poor accuracy and suspect validity
- Target C – good precision but low reliability and mixed accuracy, threatening overall validity
- Target D – both reliability an accuracy are low, an INVALID attempt at measurement
- Target E – one shot attempt (no reliability assessed) in which the measure is very imprecise. Large caliber bullet hits the target but takes out most of the rest of the target. The validity here is poor.
- Types of Validity Assessment
- External validity: can the results be extrapolated to other times, settings, etc.
- we can assess the representativeness of the sample
- whether the measurement process is true to life (also called ecological validity). This is also useful for replicability.
- Internal validity: the match up of a concpeual definition and an operational definition (measurement).
- Face validity – this assesses whether the measure taps the desired concept. You take a step back and examine the measures freshly and as objectively as possible.
- Internal Validity: the match between conceptual definition and operationalization.
- Criterion Validity: this is the extent to whicha measure taps an established standard or important behaviour that is external to the measure (Carmines & Zeller, 1979).
- Predictive : the standard or behaviour occurs after the measure
- Concurrent: the standard or behaviour exists at the same time as the measure
- Eg.: Pershad & Verma, 1995 used as their standard the clinical diagnoses of schizophrenia when individuals’s open-ended responses to inkblots werwe content analysed.
- Gottschalk & Gleser 1969 : used four different types of criterion measuer to validate their measure of psychological constructs through the CA of verabl samples:
- Psychological
- Physiological
- Pharmacological
- Biochemical
To validate their anxiety scale, they looked at the relationship between individuals’ scores n theat scale and:
(a)their diagnoses by linical psychologists
(b)blood pressure and skin temperatures at the times their speech samples were collected
The researchers conducted experiments
(c)administering tranquilizers, validating lower anxiety scores for the treatment groups.
They also found their content analysis scale to be validated by:
(d) the presence of greater lasma-free fatty acids (blood tests) for those with higher anxiety scores.
- Content Validity
- Reflects the extent to which the measure reflects the full domain of the concept being measured.
- Smith (1999) treid to tap a wide variety of aspects of female sex-role stereotyping in film characters. She measured 27 characteristics, traits and behaviours that had been identified in past research as associated primarily with women in the hope of being exhaustive.
- Construct Validity: the extent to which a measure is related to other measures (constructs) in a way consistent with hpotheses derived from theory.
- It is hard to get this right.
- Some interesting attempts at validation:
- Bales’s (1950) interactin process analysis system
- Stiles’s (1980) taxonomy of verbal respones modes
- Fisher’s (1970) decision proposal coding sstem
- Poole & Folger (1981) : tested the validity assumpton that a coding scheme should relatee to the meanings of the utterances as judged by the interactants …. They called this representational validity.
- Operationalization
- This is process of developing measures… the construction of actual, concrete measurement techniques (Babbie 1995).
- For content analsis, this means the construction of a Coding Scheme
- This means either a set of dictionaries or a set of measure in a codebook.
- Categories or Levels that are exhaustive
- This means an appropriate code for each and every unit coded. The category “other” is necessary.
- Look at table on page 118
- Categories or Levels that are Mutually Exclusive
- There should only be ONE appropriate code for each and every unit coded.
- If there is the possibility for mulitple codes, then these should be broken down into separate measurs. Look at table on page 119
- The categories are not mutually exclusive here. It is better to take the checklist approach of the second figure on 119.
- An Appropriate Level of Measurement
- Each variable should be measured with categories that are at te highest level of measurement possible, given the goals of the measure.
- Stevens has his Four Levels of Measurement:
- Nominal – the lowest and least sophisticated level.
- A nominal scale consistes of a st of categories that are distinct from one another.
- The order of categories is arbitrary, reording the categories make not difference in the meaning of the scale.
- In Box 6.1, there are a number of examples of nominal measures (gender, marital status, etc.)
- Ordinal – a set of categories that are rank ordered on some continuum.
- The use of numbers is for maintaining the propoer ordering, bu the numbers do not signify equal intervals between the groups. We can’t add the numbers together, they are just there to indicate placement in a list.
- An interval scale: consists of categories or levels represented by numbers that are quantitative or numeric in the ordinary sense. It differs from ordinary numbers in that the zero point is arbitrary. It is unusual to use this measure in content analysis – there is little advantage to this interval measure compared to ratio measure.
- A ratio – the most sophisticated or highest level of measurement.
- A ratio scale consists of categories or levels represented by numbers that are quantitative or numeric in the ordinary sense, including a true or meaningful zero point.
- Chronological age is constructed to be a ratio measurement
- Variables vs. Measures of Variables: A common mistake is to think that level of measurement is attached to the variable rather than to a particular measure of the a variable.
- A variable can be measured at different levels. Notice that in the smaple codebook, character age is measured in two ways – one at the ratio level (estimate the charcter’s chronological age) and one at the ordinal level (1 = child, 2 = adolescent, etc.)
- Computer Coding
- We will not be doing this sort of coding in this class.
- Human Coding
- All measures of human content analysis have to be fully explicated in a codebook.
- The codebook corresponds to a coding form.
- You have to make the codebook as COMPLETE and UNAMBIGUOUS as possible.
- Even the smallest, most mundane details must be spelled out.
- You can choose where to put the most detail – I suggest putting more detail into the codebook, not the form.
- Coder Training –
- It is IMPERATIVE to train your coders well!
- Pilot coding – this is practice coding that will enable you to assess overall viability and reliatbility of protocol.
- Sometimes you will change you codebook based on the pilot coding exercise.
- Final coding – this is done by each member INDEPENDENTLY (no chatting or comparing notes). This way you can test the reliability of your code book.
- Blind Coding – here the coders do not know the purpose of the study.
- Demand characteristic: the tendency of participants to give the experimenter the anwers he/she wants.
- Judge Indepedence – the freedom of coders to make judgements without input from the researcher.
- The Proceses
- See diagram on page 134
- Medium Modality and coding
- Coding in the same modality makes sense (if you are listening to music, code for just music).
- Researchers need to know the limitations of their medium.
- Tips for Coding
- Text
- – people work better with HARDCOPY, this way they can easily mark up the pages.
- Using computers can bring in more confusion than benefit.
- Static Images
- Audio Messages
- Moving Images
- Multimedia