Hostname: page-component-76fb5796d-x4r87 Total loading time: 0 Render date: 2024-04-26T23:58:16.242Z Has data issue: false hasContentIssue false

Measure for Measure: An Experimental Test of Online Political Media Exposure

Published online by Cambridge University Press:  04 January 2017

Andrew M. Guess*
Affiliation:
Department of Political Science, Columbia University, New York, NY, USA, e-mail: amg2232@columbia.edu
Rights & Permissions [Opens in a new window]

Abstract

Core share and HTML view are not available for this content. However, as you have access to this content, a full PDF is available via the ‘Save PDF’ action button.

Self-reported measures of media exposure are plagued with error and questions about validity. Since they are essential to studying media effects, a substantial literature has explored the shortcomings of these measures, tested proxies, and proposed refinements. But lacking an objective baseline, such investigations can only make relative comparisons. By focusing specifically on recent Internet activity stored by Web browsers, this article's methodology captures individuals' actual consumption of political media. Using experiments embedded within an online survey, I test three different measures of media exposure and compare them to the actual exposure. I find that open-ended survey prompts reduce overreporting and generate an accurate picture of the overall audience for online news. I also show that they predict news recall at least as well as general knowledge. Together, these results demonstrate that some ways of asking questions about media use are better than others. I conclude with a discussion of survey-based exposure measures for online political information and the applicability of this article's direct method of exposure measurement for future studies.

Type
Research Article
Copyright
Copyright © The Author 2014. Published by Oxford University Press on behalf of the Society for Political Methodology 

References

Althaus, S. L., and Tewksbury, D. H. 2007. Toward a new generation of media use measures for the ANES. Report to the Board of Overseers of the ANES.Google Scholar
Bartels, L. M. 1993. Messages received: The political impact of media exposure. American Political Science Review 87(2): 267–85.Google Scholar
Belli, R. F., Traugott, M. W., and Beckmann, M. N. 2001. What leads to voting overreports? Contrasts of overreporters to validated voters and admitted nonvoters in the American National Election Studies. Journal of Official Statistics 17(4): 479–98.Google Scholar
Berinsky, A., Huber, G., and Lenz, G. 2012. Evaluating online labor markets for experimental research: Amazon.com's Mechanical Turk. Political Analysis 20(3): 351–68.Google Scholar
Berinsky, A. J., Margolis, M. F., and Sances, M. W. 2013. Separating the shirkers from the workers? Making sure respondents pay attention on self-administered surveys. American Journal of Political Science.Google Scholar
Burton, S., and Blair, E. 1991. Task conditions, response formulation processes, and response accuracy for behavioral frequency questions in surveys. Public Opinion Quarterly 55(1): 5079.Google Scholar
Dilliplane, S., Goldman, S. K., and Mutz, D. C. 2013. Televised exposure to politics: New measures for a fragmented media environment. American Journal of Political Science 57(1): 236–48.Google Scholar
Gentzkow, M., and Shapiro, J. M. 2011. Ideological segregation online and offline. Quarterly Journal of Economics 126):1799–839.Google Scholar
Gerber, A. S., Gimpel, J. G., Green, D. P., and Shaw, D. R. 2011. How large and long-lasting are the persuasive effects of televised campaign ads? Results from a randomized field experiment. American Political Science Review 105(1): 135–50.CrossRefGoogle Scholar
Graber, D. A. 1997. Mass media and American politics. Washington, DC: CQ Press.Google Scholar
Hindman, M. 2008. The myth of digital democracy. Princeton, NJ: Princeton University Press.Google Scholar
Humphreys, M., Sanchez de la Sierra, R., and van der Windt, P. 2013. Fishing, commitment, and communication: A proposal for comprehensive nonbinding research registration. Political Analysis 21(1): 120.Google Scholar
Iyengar, S. 2011. Experimental designs for political communication research: Using new technology and online participant pools to overcome the problem of generalizability. In Sourcebook for political communication research: Methods, measures, and analytical techniques, eds. Bucy, E. P. and Holbert, R. L. New York: Routledge.Google Scholar
Iyengar, S., and Kinder, D. R. 1987. News that matters: Television and American opinion. Chicago: University of Chicago Press.Google Scholar
Krosnick, J. A. 1991. Response strategies for coping with the cognitive demands of attitude measures in surveys. Applied Cognitive Psychology 5(3): 213–36.Google Scholar
LaCour, M. J. 2013. A balanced news diet, not selective exposure: Results from Erie to Arbitron. Paper presented at the Annual Meeting of the American Political Science Association Annual, Chicago.Google Scholar
Lazarsfeld, P. F., Berelson, B., and Gaudet, H. 1944. The people's choice: How the voter makes up his mind in a presidential campaign. New York: Columbia University Press.Google Scholar
Leeper, T. 2013. Crowdsourcing with R and the MTurk API. Political Methodologist 20(2): 27.Google Scholar
Mullainathan, S., and Shleifer, A. 2005. The market for news. American Economic Review 95(4): 1031–53.Google Scholar
Munson, S., Lee, S. Y., and Resnick, P. 2013. Encouraging reading of diverse political viewpoints with a browser widget. International Conference on Weblogs and Social Media.Google Scholar
Mutz, D., and Young, L. 2011. Communication and public opinion. Public Opinion Quarterly 75(5): 1018–44.Google Scholar
Pew Research Center for the People & the Press. 2012. In Changing News Landscape, Even Television is Vulnerable. http://www.people-press.org/2012/09/27/in-changing-news-landscape-even-television-is-vulnerable/(accessed November 24. 2013).Google Scholar
Price, V., and Zaller, J. 1993. Who gets the news? Alternative measures of news reception and their implications for research. Public Opinion Quarterly 57(2): 133–64.Google Scholar
Prior, M. 2009a. The immensely inflated news audience: Assessing bias in self-reported news exposure. Public Opinion Quarterly 73(1): 130–43.Google Scholar
Prior, M. 2009b. Improving media effects research through better measurement of news exposure. Journal of Politics 71(3): 893908.Google Scholar
Prior, M., and Lupia, A. 2008. Money, time, and political knowledge: Distinguishing quick recall and political learning skills. American Journal of Political Science 52(1): 169–83.Google Scholar
Romantan, A., Hornik, R., Price, V., Cappella, J., and Viswanath, K. 2008. A comparative analysis of the performance of alternative measures of exposure. Communication Methods and Measures 2(1–2): 8099.Google Scholar
Smyth, J., Dillman, D., Christian, L., and Stern, M. 2006. Comparing check-all and forced-choice question formats in Web surveys. Public Opinion Quarterly 70(1): 6677.Google Scholar
Sudman, S., and Bradburn, N. M. 1982. Asking questions: A practical guide to questionnaire design. San Francisco: Jossey-Bass.Google Scholar
Sudman, S., Bradburn, N. M., and Schwarz, N. 1996. Thinking about answers: The application of cognitive processes to survey methodology. San Francisco: Jossey-Bass.Google Scholar
Tourangeau, R., Rips, L. J., and Rasinski, K. 2000. The psychology of survey response. Cambridge, UK: Cambridge University Press.Google Scholar
Supplementary material: PDF

Guess supplementary material

Appendix

Download Guess supplementary material(PDF)
PDF 2.1 MB