Content Analysis in the Study of Crime, Media, and Popular Culture
Summary and Keywords
Content analysis is considered both a quantitative and a qualitative research method. The overarching goal of much of the research using this method is to demonstrate and understand how crime, deviance, and social control are represented in the media and popular culture. Unlike surveys of public opinions about crime issues, which seek to know what people think or feel about crime, content analysis of media and popular culture aims to reveal a culture’s story about crime. Unlike research that examines how individuals’ patterns of media consumption shape their attitudes about crime and control, content analysis appraises the meaning and messages within the media sources themselves. Media and popular culture sources are viewed as repositories of cultural knowledge, which capture past and present ideas about crime, while creating and reinforcing a culture’s shared understanding about crime.
In content analysis, media and popular culture portrayals of crime issues are the primary sources of data. These portrayals include a range of sources, such as newspapers, movies, television programs, advertisements, comic books, novels, video games, and Internet content. Depending on their research questions, researchers draw samples from their selected sources, usually with additional selection boundaries, such as timeframe, genre, and topic (e.g., movies about gangs released from 1960 to 1990).
There are two primary approaches to conducting content analysis. In quantitative forms of content analysis, researchers code and count the occurrence of elements designated by the researcher prior to the study (e.g., the number of times a violent act occurs). In qualitative forms of content analysis, the researchers focus on the narrative, using an open-ended protocol to record information. The approaches are complementary, as each reveals unique yet overlapping concepts crucial to understanding how the media and popular culture produce and reproduce ideas about crime.
Defining Content Analysis
As a research method, content analysis exists somewhere between purely quantitative and purely qualitative. In the study of crime in the media, research ranges from studies that count or otherwise quantify texts for the purpose of statistical analysis to studies that explore presentation and representation of crime-related issues. Even in those quantitatively oriented studies, results are given qualitative consideration. Increasingly, in the criminological study of media and popular culture, content analysis is typically viewed as a qualitative methodology.
Content analysis is more than watching TV or movies, or reading newspapers or comics, and then reporting what is presented in the medium. How the story is told and how characters are portrayed are often more telling than are specific plot points. Content analysis requires systematically watching or reading with an analytical and critical eye, going beyond what is presented and looking for deeper meanings and messages to which media consumers are exposed. This exposure contributes to the social construction of crime and deviance, that is, to people’s beliefs about what is deviant, who is criminal, and how to control crime. The media captures and frames the broader cultural story about criminal justice. The primary purpose of content analysis in the study of crime and justice has evolved from identifying the prevalence of the topic or terms under study into revealing the cultural frames. The results from content analysis, then, offer evidence that allows for a more critical appraisal of how crime and justice are socially constructed.
The past 40 years have seen substantial growth in the application of content analysis to a range of issues, including crime and violence (Neuendorf, 2002), moving beyond text on the page to “text” in visual and moving images. Indeed, one of the earliest studies to employ the method, the Payne Fund Studies, coded for violence and other content in films in the 1930s. As the century progressed and attention shifted to violence in television, content analysis became a core methodology of the Cultural Indicators Project (Gerbner & Gross, 1976; Morgan & Shanahan, 2010). This project has influenced media research since the 1970s, including the National Television Violence Study of the later 1990s (Smith et al., 1998).
As research has progressed, however, scholars called for greater attention to the context in which the content is presented, arguing that an act or an incident could not be fully understood without referring to the circumstances of its presentation in media or the broader socio-cultural context. Such work, coupled with expanding opportunities for consumers to encounter crime-related content across a variety of media sources, also stimulated analyses that placed more emphasis on the latent content itself. That is, some research looks beyond the action to the less obvious, but still critical, message and meaning being produced and reproduced in the media and popular culture. The advent of academic journals such as Crime, Media, Culture, Journal of Criminal Justice and Popular Culture, and the recent Journal of Qualitative Criminal Justice and Criminology also speak to the emergence of content analysis and other qualitative techniques in the study of crime and social control.
Sampling the Media Universe
For those studying crime issues presented in media and popular culture, there is a wide array of text-based sources, including novels, textbooks, newspapers, magazines, and comic books and graphic novels. There is also a wide array of audio-visual sources, including movies, television, and video games, each with a myriad of genres and formats. Music, in lyrics, video, and performance, is yet another source. Finally, the rise of electronic and social media further broadens the range of sources, from traditional news sources to Twitter conversations to YouTube content to user-generated forums like Reddit. The type and genre of media to be studied are often identified as research questions are developed.
As with most social research, it is often not feasible to examine an entire population of media texts or sources. For example, even if one could access every copy of comic books featuring Batman, it is likely impractical, due to constraints on time or resources, to read and code hundreds, if not thousands, of books. On the other hand, one could watch and code every cinematic release featuring Batman (Bosch, 2016). Thus, the decision between reviewing an entire population or a sample of the population is driven by both research questions and practical considerations.
With research questions and practical considerations in mind, sampling entails additional decision points. For example, suppose one is interested in news presentations of crime in editorials or commentaries. First, one needs to decide among newspapers, news magazines, televisions news, or Internet news. If U.S. newspapers are selected, then one needs to determine the sampling frame, including the time period of interest and which papers to select. Will papers be randomly selected from the universe of U.S. papers or from papers with a certain circulation level? Will there be a degree of stratification, such as random samples from designated geographic units? Or, will newspapers be selectively chosen based on other research interests, such as tracking a specific event in a specific location? If, however, one wants to compare across media types, then similar decisions need to be made that can be applied to each type. Within each source, there may be several stories, editorials, or commentaries, so researchers need to decide whether to review all of them, or more practically, decide how to sample among them, necessitating another round of sampling decisions.
In sum, as this brief example illustrates, sampling for the purpose of content analysis entails a good deal of complexity. Regardless of design, samples should be selected so that they reasonably represent the population and yield sufficient numbers for analysis. Researchers should also take care to record all decision points, so that the sample can be replicated by others.
Quantitative Content Analysis
Originally developed for use with written texts, quantitative content analysis (QCA) aims to distill the many words presented in a text into meaningful categories. These categories can then be treated as variables, allowing for a descriptive interpretation of the texts, or functioning as variables in statistical analyses. QCA has expanded beyond the written word to many other types of media, but the basic principal of classifying larger content into smaller categories remains at the core of the method. Through analysis of how these categories inter-relate with each other and intersect with the broader cultural context, the goal is to discover how materials communicate meaning and what meaning is communicated.
Basic Methodology of QCA
A coding schema is central to the method. As with sampling, the development of a coding schema is driven by precedents in the research or theoretical literature, and by the teams’ research questions. Weber (1990) outlines several basic steps in the coding process. The first step is to define recording units, that is, whether coders should attend to certain words, phrases, images, or overarching themes of a passage or piece. Depending on research questions, recording units often are some combination of these or other units. For example, in a television show, researchers may want to know what words were in the dialogue and also the overarching theme of the conversation between characters.
The second step in developing a coding schema is to define categories. One may think of these as boxes to be marked on a rubric, even if computer software is aiding in the coding process. The categories may be defined narrowly or broadly. For example, the presence of “violence” between two characters may be narrowly defined by physical contact or broadly defined to include abusive language. The categories may be mutually exclusive, or an incident may be coded into multiple categories. This decision may be influenced by the analytic intent; basic statistical assumptions are violated when categories are not exclusive.
Once a preliminary rubric is established, the third step is a pre-test, in which a subsample is coded. The pre-test process should reveal where categories may need further refinement or where coding rules require additional clarity. The pre-test also produces information regarding the accuracy or correctness with which human coders or computerized coding programs are classifying the text. If accuracy is low, the coding rules should be revised. Step four, then, is revising the coding schema. The fifth step is to pre-test again. This process should continue until the coding process, whether human- or computer-coded, yields an adequate level of accuracy.
The sixth step is to code the full sample using the established schema. Following data collection, the final step of the coding process is checking the achieved accuracy of the human coders or the computer program. Individual human coders may fatigue over time, thus making more mistakes, or their interpretation of categories may shift slightly over time, resulting in misclassification. Computerized output should be reviewed to confirm whether code rules were applied correctly. During the process, for example, the program may encounter text combinations not anticipated by the programmers or not present in the pre-test, resulting in misclassification.
As with other forms of measurement, issues of reliability and validity may emerge in QCA (Neuendorf, 2002). Particularly with several human coders, a primary issue is inter-rater reliability. In brief, inter-rater reliability is the extent to which different people code the same text in the same way. Differences, for example, may occur when coding rules or categories are not clear, or when there are cognitive differences across coders. The pre-test process and adequate coder training may reduce these differences, but inter-rater reliability should also be assessed at the end of coding. Various statistical tests exist to assess inter-rater reliability.
Validity can broadly be divided into internal and external. Internal validity refers to the match between concepts and their operationalized definitions in variables. There is no parallel statistical metric to assess internal validity, but there are several dimensions that researchers may consider. External validity, in contrast, refers to the generalizability and replicability of the results generated by a measure. Breadth and representativeness of the sample improve generalizability, while a full accounting of the procedures of the coding and variable creation improve replicability.
Analysis and Interpretation in QCA
Analyzing data generated by the coding process can take many forms. Again, the analysis of the data is driven by existing theory and the established research questions. Once data are collected, however, researchers fully quantify the data by creating variables from the coded data that are most meaningful for the hypotheses they want to explore and the analyses they want to conduct. These analyses may range from completely descriptive in nature to mean-difference or correlational tests to multivariate regression models.
In interpreting the analytical results, researchers bring the accumulated evidence to bear on the research questions, determining what story their results tell about the texts and their content. Regardless of the analytical technique used in QCA, any interpretation of quantified content must be corroborated by reference to the original texts. That is, researchers should compare their interpretation of the data to a subsample of their source documents. If the story of the data analysis reasonably represents the story within the documents, the interpretations of analytical results are not just products of classification schemes or statistical techniques. If the stories do not match, then reconsideration of the analysis or interpretation is necessary. In short, although QCA aims to quantify what could be considered qualitative information, it nonetheless retains a portion of qualitative art in the final interpretation of the data.
Example of QCA
One example of QCA is Britto and Dabney’s (2010) analysis of justice issues on political talk shows, which illustrates the coding and analysis process. Britto and Dabney were interested in crime content on these programs and, in particular, how the content was politicized. They selected the central primetime talk show across each of the three major cable news networks. Over the course of six months, they randomly selected one day per week to record the shows. They chose this approach for two reasons: to avoid one particular news story dominating conversation (e.g., a high-profile crime) and to ensure they were watching a “typical” example of the show. Coding was performed at two levels of analysis. At the program level, the schema included numeric coding categories for the number of segments in an episode, speaking time given to guests, and racial/ethnic and gender characteristics of offenders and victims in crime stories. At the individual level of analysis, the schema included codes for guest characteristics and guests’ interactions with hosts. These categories became variables in the analyses. Coders received four hours of training, which included a discussion of how concepts were operationalized, a practice coding session, and the follow-up discussion. During data collection, coders were instructed to watch their assigned episodes at least twice in order to code at each analytic level separately.
Britto and Dabney’s statistical analysis of their enumerated data began with a presentation of guest profiles, comparing population numbers to the demographic characteristics of guests, and then comparing general show guests to guests in justice-related segments. The analysis continued with a description of the amount of justice-related content on each show, then compared shows to each other. To address their major research questions, the analysis examined the interactions of hosts with their guests, using chi-square tests to reveal differences across guests’ political persuasions. The analysis also presented ratios of offender and victim characteristics, comparing these ratios across shows and to official United States crime data. Britto and Dabney used these statistical presentations as evidence in answering their research questions; however, they also corroborated their results by presenting qualitative descriptions of each of the source programs.
Other examples of QCA include Welch, Weber, and Edwards’ (2000) analysis of corrections debates in the New York Times; Cavender, Bond-Maupin, and Jurik’s (1999) analysis of gender in reality crime television; and Chiricos and Eschholz’s (2002) analysis of crime and race/ethnicity in local television news.
Ethnographic Content Analysis
The quantitative approach to content analysis, while useful, may result in removing the coded content from the context surrounding it (Muzzatti, 2006). For example, a “violent action” in a television program may be counted, but the scene or setting in which it occurred, the offender’s motive for the violent act, the victim’s reaction, and other visual or auditory detail are lost. To address the shortcomings of QCA, Altheide (1987) proposed for ethnographic content analysis (ECA).
Like QCA, the goal of ECA is to discover how materials communicate meaning; however, in ECA, meaning is assumed to be present in various modalities, such as text, format, visual and auditory style, and in the positioning of one piece of information among others. Although this approach may involve some degree of enumeration, the emphasis is on descriptive and conceptual data, similar to what one might record in an observational study. The procedures for data collection, analysis, and interpretation are designed to be reflexive, with the researcher constantly engaged in the material, comparing observations as the process unfolds, and attentive to conceptual and theoretical nuances as they arise.
Basic Methodology in ECA
The basic methodology of ECA is outlined by Altheide and Schneider (2013). A review of the literature should guide the selection of the specific problem or issue to examine, as well as the medium to examine. Once researchers decide on a medium, they should learn about its production process and context. For example, if researchers are interested in studying comic books, they should become familiar with how comics are created, developed, and disseminated. Next, researchers should become familiar with several key examples to understand elements of formatting or general patterns in how the text and imagery are presented. Together, an understanding of the production process and the patterning within the source material provide essential background for creating a data collection protocol.
Constructing an ECA protocol is similar to developing a coding schema in QCA in the sense that the protocol is developed, tested, and re-tested. The format of an ECA protocol, however, is oriented conceptually rather than categorically. A protocol is designed as a means by which to query the information source—to ask questions, capture definitions, understand meanings, and reveal processes. Theoretically derived concepts, drawn from the research literature and from evaluation of key examples, guide the creation of a preliminary protocol. The protocol should have appropriate preset coding categories, (e.g., citation, date, length of text), but most of the conceptual categories should be open-ended for the researcher to input pertinent information, whether it be a count, a quotation, or a narrative description. During protocol development, new conceptual classifications are expected to emerge; as a result, original categories may be modified, and new categories may be incorporated into the protocol.
Data collection using the established protocol entails populating the categories with a wealth of descriptive examples. Throughout data collection, the researcher remains attuned to the process, such as including additional notes about emergent themes or observations about how pieces in the sample connect to each other. Indeed, Altheide and Schneider (2013) recommend a midpoint analysis of the gathered data in order to detect emergent themes or interaction among concepts, which may lead to refinement of the data collection process. Previously collected data may need recoded in light of these refinements; newly collected data can proceed under a revised protocol.
Issues of reliability in ECA are approached differently than QCA. The open-ended format of an ECA protocol precludes the use of standard metrics to assess inter-rater reliability, but efforts can be taken to ensure a level of agreement among coders. For example, as part of the training process, coders should also become familiar with the production process of the medium under study, as well as with the patterns in key examples. Familiarity with the sources facilitates more detailed coding. Research teams employing ECA can promote consistency in coding their observations by coding the same sources independently, then meeting to discuss meanings of concepts and how they are interpreted, as well as recurrent and emergent themes. Coder agreement may be achieved by working together to recode a source. Persistent disagreement may be indicative of another dimension that needs to be in a revised protocol. In general, in ECA, individual consistency in coding is more critical than inter-rater reliability, given the priority placed on reflexivity during data collection.
Analysis and Interpretation in ECA
Data analysis in ECA, like data collection, is a reflexive process. Notes are carefully read, re-read, compared to other notes, sorted, and read again. The early stages of analysis allow the researcher to explore, describe, explain, and perhaps theoretically link elements of the data. In addition, researchers attend to differences within the various conceptual categories. As analysis continues, researchers take notes on their notes, writing small summaries of overarching themes, concepts, and divergent ideas. Summaries include supporting details, such as descriptions of or quotes from the data sources. These smaller summaries are then combined, identifying typical cases and exceptions to those cases, and documenting unexpected elements that push the data in intriguing directions.
Analytical interpretation extends from this process. The small summaries are reviewed in light of the data collection protocol, by which the researcher can return to the original questions of interest. Referring to the protocol allows the researcher to sort the summaries into more distinctive categories to answer those questions, as well as to determine what does not fit and why it does not fit. There may be variation in the nature of the original documents, the concept may have been an unanticipated but important part of the story, or it may suggest a direction for future research. Like QCA, the results of evidence accumulated in ECA are brought to bear on the research questions. ECA, however, is designed not only to tell the story presented in the data but also to reveal, discuss, and contextualize the manifest and latent meaning of that story by grounding it in the social world and the broader social processes by which meaning is produced.
Example of ECA
Welsh, Fleming, and Dowler (2011) conducted an ECA of crime movies to investigate how crime and victimization were constructed in film, and to discover the messages presented about justice modalities. Given their intention to uncover these themes and messages, they outlined four questions, suggested by the research literature. To select a sample from the population of American films made from 1930–2009, they performed crime and justice-related keyword searches using the movie reference site imdb.com. From this first set of results, they then employed a theoretical sampling approach, selecting films based on two criteria, derived from their research questions, regarding plot narratives of the films. To collect data, each of the 30 films in the final sample was watched twice. The research questions served as the basis for an open-ended protocol. Coding entailed taking detailed notes about characters’ interactions and about the films’ narratives by observing both dialogue and visual imagery.
Welsh and his coauthors analyzed their data using a constructivist grounded theory approach, which exemplifies the reflexive nature of the ECA process: details from the protocol were read repeatedly and comparatively, with analysis drawing on existing theory. Their analysis revealed three primary themes across the films. In interpreting these three themes, the authors identified several sub-themes, as well as points of contrast. Supporting evidence for each theme is presented as a mix of narrative summary and discussion by the authors and as strategically placed quotations from the films. Where appropriate, connections were made to existing theory, either as points of comparison or as extensions to earlier work. Thus, Welch and his coauthors offered answers to their research questions, supporting their interpretations with descriptive evidence from their sources and with reference to prior literature.
Other examples of ECA include: Altheide and Michalowski’s (1999) analysis of fear discourse in newspapers; Kort-Butler’s (2013) study of crime and justice representations in superhero cartoons; and Myers’ (2012) analysis of televised reports on juvenile detention using program transcripts.
Content Analysis, Crime, and Control
Content analyses of crime and justice issues have covered many genres of media and popular culture. In addition to those examples listed in the sections “Example of QCA” and “Example of ECA,” other genres include, but are not limited to, television crime dramas (Cavender & Deutsh, 2007); televised documentaries and realty television (Cecil, 2010); television commercials (Maguire, Sandage, & Weatherby, 2000); comic books (Phillips & Strobl, 2006); music (Hunnicutt & Andrews, 2009); criminal justice textbooks (Burns & Katovich, 2006); Internet news (Sjøvaag & Stavelin, 2012); and movie reviews posted online by imdb.com users (Gosselt, van Hoof, Gent, & Fox, 2015).
Across these genres, content analyses in the past few decades have challenged the accuracy of media presentations of crime and justice issues, generally finding that the media often exaggerate the reality of crime by focusing on unusual or rare crimes. Media presentations also tend to misrepresent offenders and victims in ways that do not represent their actual distribution in the population by race, gender, or age, such that young male minorities are seen as the default “typical” criminal. Thus, the nature of crime, as portrayed in the media and popular culture, is violent, the image of the criminal is the cold yet rational predator “other,” and the image of the victim is the innocent prey (Cavender, 2004; Kappeler & Potter, 2005; Surette, 2014).
Content analyses have also revealed the ways in which the media frame stories about crime to correspond to and reinforce these images. Mainstream media depictions of crime and justice generally present messages that conform to and promote the dominant ideology about “the crime problem” and how to manage it (Altheide, 1997), namely through established channels of reactive policing and punitive punishments. As that story changes in response to larger cultural shifts in public and political attitudes about crime and control, the methods offered by content analysis are being deployed to understand how the mediated images and meanings about crime and justice will change in turn.
Key reference books for those interested in conducting content analysis include Altheide and Schneider (2013) and Neuendorf (2002). These texts describe the techniques of content analysis in further detail, provide prodigious examples for constructing rubrics and analyzing data, and list various software that may assist researchers in their analyses. Methods offered by cultural criminologists provide additional insight into the use of ethnography and visual analysis in studying media and popular culture (Ferrell, Hayward, & Young, 2008; Kane, 2004).
Several edited volumes contain chapters of content analytic research, including Media, Process, and the Social Construction of Crime, edited by Barak (1994); Entertaining Crime, edited by Fishman and Cavender (1998); Making Trouble, edited by Ferrell and Websdale (1999); Constructing Crime, edited by Potter and Kappeler (2006); Framing Crime, edited by Hayward and Presdee (2010); and Volume 14 of Sociology of Crime, Law, and Deviance, edited by Deflem (2010), which focuses on popular culture, crime, and control.
Altheide, D. L. (1987). Ethnographic content analysis. Qualitative Sociology, 10, 65–77.Find this resource:
Altheide, D. L. (1997). The news media, the problem frame, and the production of fear. Sociological Quarterly, 38, 647–668.Find this resource:
Altheide, D. L., & Michalowski, R. S. (1999). Fear in the news: A discourse in control. Sociological Quarterly, 40, 475–503.Find this resource:
Altheide, D. L., & Schneider, C. J. (2013). Qualitative media analysis (2d ed.). Los Angeles: SAGE.Find this resource:
Barak, G. (Ed.). (1994). Media, process, and the social construction of crime: Studies in newsmaking criminology. New York: Garland.Find this resource:
Bosch, B. (2016). “Why So Serious?” Threat, Authoritarianism, and Depictions of Crime, Law, and Order in Batman Films. Criminology, Criminal Justice, Law, & Society, 17, 37–54.Find this resource:
Britto, S., & Dabney, D. A. (2010). “Fair and balanced?” Justice issues on political talk shows. American Journal of Criminal Justice, 35, 198–218.Find this resource:
Burns, R. G., & Katovich, M. A. (2006). Melodramatic and consentient images in introductory criminal justice textbooks. Journal of Criminal Justice, 34, 101–114.Find this resource:
Cavender, G. (2004). Media and crime policy: A reconsideration of David Garland’s The Culture of Control. Punishment and Society, 6, 335–348.Find this resource:
Cavender, G., Bond-Maupin, L., & Jurik, N. C. (1999). The construction of gender in reality crime TV. Gender and Society, 13, 643–663.Find this resource:
Cavender, G., & Deutch, S. K. (2007). CSI and moral authority: The police and science. Crime, Media, Culture, 3, 67–81.Find this resource:
Cecil, D. K. (2010). Televised images of jail: Lessons in controlling the unruly. Sociology of Crime, Law, and Deviance, 14, 67–88.Find this resource:
Chiricos, T., & Eschholz, S. (2002). The racial and ethnic typification of crime and the criminal typification of race and ethnicity in local television news. Journal of Research in Crime and Delinquency, 39, 400–420.Find this resource:
Deflem, M. (Ed.). (2010). Popular culture, crime, and social control. Sociology of Crime, Law, and Deviance, 14. Bingley, U.K.: Emerald Group.Find this resource:
Ferrell, J., Hayward, K., & Young, J. (2008). Cultural criminology: An invitation. London: SAGE.Find this resource:
Ferrell, J., & Websdale, N. (Eds.). (1999). Making trouble: Cultural constructions of crime, deviance, and control. New Brunswick, NJ: Transaction.Find this resource:
Fishman, M., & Cavender, G. (Eds.). (1998). Entertaining crime: Televised reality programs. New York: Aldine de Gruyter.Find this resource:
Gerbner, G., & Gross, L. (1976). Living with television: The violence profile. Journal of Communication, 26, 173–199.Find this resource:
Gosselt, J., Van Hoof, J., Gent, B., & Fox, J. P. (2015). Violent frames: Analyzing Internet movie database reviewers’ text descriptions of media violence and gender differences from 39 years of U.S. action, thriller, crime, and adventure movies. International Journal of Communication, 9, 547–567.Find this resource:
Hayward, K., & Presdee, M. (Eds.). (2010). Framing crime: Cultural criminology and the image. New York: Routledge.Find this resource:
Hunnicutt, G., & Andrews, K. H. (2009). Tragic narratives in popular culture: Depictions of homicide in rap music. Sociological Forum, 24, 611–630.Find this resource:
Kane, S. C. (2004). The unconventional methods of cultural criminology. Theoretical Criminology, 8, 303–321.Find this resource:
Kappeler, V. E., & Potter, G. W. (2005). The mythology of crime and criminal justice (4th ed.). Long Grove, IL: Waveland Press.Find this resource:
Kort-Butler, L. A. (2013). Justice league? Depictions of justice in children’s superhero cartoons. Criminal Justice Review, 38, 50–70.Find this resource:
Maguire, B., Sandage, D., & Weatherby, G. A. (2000). Violence, morality, and television commercials. Sociological Spectrum, 20, 121–143.Find this resource:
Morgan, M., & Shanahan, J. (2010). The state of cultivation. Journal of Broadcasting & Electronic Media, 54, 337–355.Find this resource:
Muzzatti, S. L. (2006). Cultural criminology: A decade and counting of criminological chaos. In W. S. DeKeserdy & B. Perry (Eds.), Advancing critical criminology: Theory and application (pp. 63–82). Lanham, MD: Lexington.Find this resource:
Myers, R. R. (2012). “Society must be protected from the child”: The construction of U.S. juvenile detention as necessary and normal. Critical Criminology, 20, 395–407.Find this resource:
Neuendorf, K. A. (2002). The content analysis guidebook. Thousand Oaks, CA: SAGE.Find this resource:
Phillips, N. D., & Strobl, S. (2006). Cultural criminology and kryptonite: Apocalyptic and retributive constructions of crime and justice in comic books. Crime, Media, Culture, 2, 304–331.Find this resource:
Potter, G. W., & Kappeler, V. E. (Eds.). (2006). Constructing crime: Perspectives on making news and social problems (2d ed.). Long Grove, IL: Waveland Press.Find this resource:
Sjøvaag, H., & Stavelin, E. (2012). Web media and the quantitative content analysis: Methodological challenges in measuring online news content. Convergence: The International Journal of Research into New Media Technologies, 18, 215–229.Find this resource:
Smith, S. L., Wilson, B. J., Kunkel, D., Linz, D., Potter, W. J., Colvin, C. M., & Donnerstein, E. (1998). Violence in television programming overall: University of California Santa Barbara study. In Center for Communication and Social Policy. (Ed.), National Television Violence Study (Vol. 3, pp. 5–220). Thousand Oaks, CA: SAGE.Find this resource:
Surette, R. (2014). Media, crime, and criminal justice: Images, realities, and policies (5th ed.). Stamford, CT: Cengage Learning.Find this resource:
Weber, R. P. (1990). Basic content analysis (2d ed.) Newbury Park, CA: SAGE.Find this resource:
Welch, M., Weber, L., & Edwards, W. (2000). “All the news that’s fit to print”: a content analysis of the correctional debate in the New York Times. The Prison Journal, 80, 245–264.Find this resource:
Welsh, A., Fleming, T., & Dowler, K. (2011). Constructing crime and justice on film: Meaning and message in cinema. Contemporary Justice Review, 14, 457–476.Find this resource: