[go: up one dir, main page]

0% found this document useful (0 votes)
21 views11 pages

How Social Media Platforms Influence Political Awareness

This mixed-methods study explores the influence of social media platforms on political awareness among U.S. college students, focusing on how platform design affects information exposure and civic behavior. It combines quantitative surveys and qualitative focus groups to analyze the impact of algorithms on political knowledge, engagement, and misinformation. The findings aim to inform higher education institutions and social media companies about enhancing digital literacy and civic integrity tools.

Uploaded by

bdavidson5991
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
21 views11 pages

How Social Media Platforms Influence Political Awareness

This mixed-methods study explores the influence of social media platforms on political awareness among U.S. college students, focusing on how platform design affects information exposure and civic behavior. It combines quantitative surveys and qualitative focus groups to analyze the impact of algorithms on political knowledge, engagement, and misinformation. The findings aim to inform higher education institutions and social media companies about enhancing digital literacy and civic integrity tools.

Uploaded by

bdavidson5991
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 11

How Social Media Platforms Influence Political Awareness

Among U.S. College Students: A Mixed-Methods Study

of Platform Design, Misinformation, and Civic Behavior

Zachary T. Hodgen

Department of Political Science, Virginia Commonwealth University

POLI 320: Research Methods

​ ​ ​ ​ Dr. Andrea C. Simonelli

April 20, 2025


Abstract

This mixed-methods study investigates how social media platforms influence political

awareness among U.S. college students, with particular attention to how platform-specific design

choices mediate the relationship between information exposure and civic behavior. I’m

combining quantitative survey data from 300 students at three types of institutions with

qualitative data from 30 students in focus groups and a content analysis of 300 posts they’ve

shared. The research is structured around three central questions: First, how do algorithms on

platforms like TikTok, Instagram, and X shape political knowledge formation? Second, what

patterns exist between online engagement and real-world participation? And third, how

frequently and in what form do students encounter misinformation, and how do they respond to

it?

To measure political knowledge, I’m using a modified Political Knowledge Index

(PKI-15) that assesses understanding of constitutional rights, congressional leadership, and key

current events. The qualitative side leans on discourse analysis frameworks that help identify

how algorithmic features like TikTok’s For You Page or Instagram’s Stories shape what students

prioritize. My early hypothesis is that different platform affordances produce distinct

vulnerabilities to misinformation, with more visual platforms leading to higher emotional

engagement but less scrutiny.

This design intentionally considers institutional diversity by using stratified sampling to

include students from a large public R1 university, a small private liberal arts college, and a

community college. This allows me to capture a broader range of online behaviors and baseline

civic literacy. Ultimately, the findings aim to inform both higher education institutions
developing digital literacy curricula and platforms considering reforms around civic integrity

tools. What I’m contributing is not just another look at screen time and politics, but a more

precise mapping of how digital architecture reshapes student engagement with democracy—and

what that means for the future of civic education.

Introduction

Over the last decade, the way young people engage with politics has fundamentally

changed. Gone are the days when students learned about current events solely through lectures or

newspaper articles. Instead, today’s college students navigate political landscapes shaped by

algorithms, likes, and the visual grammar of social media. As someone who has worked closely

with student leaders and spent countless hours analyzing how campus communities organize,

I’ve seen firsthand how political action now often begins with a share or a story—not a flyer or a

speech. But I’ve also seen confusion, polarization, and misinformation flourish in those same

spaces.

This research design grew out of three recurring tensions I noticed in conversations and

organizing spaces on campus. First, many students are more mobilized than ever, but they often

distrust the very institutions they’re trying to change. Second, their political language is

fluent—they know the hashtags, the slogans—but their factual understanding doesn’t always

match. And third, while they show up for national issues, they often disengage from local and

institutional politics that directly affect their lives. I kept wondering: what is it about these

platforms that encourages some types of political engagement and discourages others?

That’s where this study begins. Using Carolyn Forestiere’s framework for analyzing

political behavior, I’m structuring this around three dimensions: cognitive (what students know

and how they learn it), behavioral (what they do with that knowledge), and epistemic (how they
evaluate truth and misinformation). Each of these elements is filtered through the digital

environments where students now spend so much of their time. By looking at TikTok, Instagram,

and X separately, I can evaluate how different kinds of content delivery—from auto-play videos

to curated feeds—influence political awareness and action in different ways.

The goal isn’t to vilify or glorify these platforms. Instead, I want to map how they work

on a micro level, how they affect students’ perceptions, and how that shapes their broader

political worldview. The theoretical significance of this research lies in its effort to bring

agenda-setting theory into conversation with contemporary digital behavior studies, particularly

through the idea of "liquid gatekeeping" (Chadwick, 2017), where control over information flow

is decentralized, fluid, and deeply personal.

Practically, the findings should help three groups. First, university leaders working to

teach digital literacy in meaningful ways. Second, social media companies that are under

pressure to improve their civic integrity tools. And third, student organizations that are trying to

organize their peers more effectively and responsibly. By focusing on college students, this study

centers a group whose political behavior is both impactful and under-researched—one that could

shape elections, push institutions, and redefine democratic norms. My hope is that by

understanding how this group thinks, shares, and acts politically online, we can better support

their ability to think critically and engage constructively offline.

Literature Review

Theoretical Foundations

My work draws from three main theoretical traditions: agenda-setting theory, connective

action theory, and misinformation studies. McCombs and Shaw’s (1972) agenda-setting theory

laid the foundation for understanding how media influences public attention. In traditional
media, editors selected the stories that mattered. In today’s social media environment, the

gatekeeping role has shifted—and become more fragmented. Vargo et al. (2018) build on this

idea by describing social media as a system of "tertiary gatekeeping," where algorithms and user

behaviors co-create information ecosystems.

To understand how political action takes place in these systems, I turn to Bennett and

Segerberg (2013), whose concept of connective action helps explain the rise of decentralized,

networked mobilization. Unlike traditional collective action, which requires strong

organizational structures, connective action allows people to engage through personal expression

and digital participation. This is especially relevant for students, whose activism often looks like

retweets, stories, or mutual aid posts rather than formal group membership.

Finally, to understand how misinformation spreads and sticks, I incorporate Wardle and

Derakhshan’s (2017) typology of misinformation, as well as Pennycook et al.’s (2018) work on

the “illusory truth effect,” which shows how repeated exposure to falsehoods increases belief in

them. These models help me interpret how misinformation circulates in student networks and

how it influences perceptions of political truth.

Key Debates

There’s active debate around whether social media increases or decreases political

knowledge. Gil de Zúñiga et al. (2016) found that students who used social media to seek news

tended to be better informed. But his later work (2019) warns of the "news-finds-me" perception,

where users feel informed simply because information appears in their feed. This passive

consumption can lead to overconfidence and lower fact retention.

When it comes to exposure to diverse views, studies also diverge. Dubois and Blank

(2018) found that Twitter users often encounter opposing perspectives. But Haque et al. (2022)
argue that TikTok’s algorithm produces “interest-based filter bubbles” that are less ideologically

diverse. Instagram, according to Vaccari and Chadwick (2020), leans more toward visual,

emotionally resonant content than issue-based analysis, making it a unique case for studying

affective responses.

Tufekci (2017) introduces the idea of "snackable activism," describing how students

engage with social causes in short, digestible forms that don’t always translate into sustained

action. Meanwhile, Mihailidis and Viotty (2017) emphasize how peer-to-peer sharing often

replaces traditional journalism as a primary news source among young adults.

Empirical Gaps

Several gaps guide my study’s direction. First, many researchers still treat social media as

a monolith. Valeriani and Vaccari (2018), for instance, assess political learning across platforms

without analyzing affordances specific to TikTok, Instagram, or X. I argue that this lumping

misses key differences in how content is encountered and shared.

Second, most political communication research still focuses on national events. Guess et

al. (2019) examine misinformation in the context of presidential elections, but little work has

been done on campus-based issue engagement or how students translate digital action into

institutional change.

Methodological Innovations

My approach draws on Prior and Bou-Hamad’s (2021) call for multimodal methods in

social media research. In addition to surveys and focus groups, I plan to analyze political posts

using:

●​ Eye-tracking heatmaps to assess which content elements draw attention (Wojdynski et al.,

2020)
●​ Network analysis to map who shares what and where

●​ Fact-checking experiments to test how students respond to labeled vs. unlabeled

misinformation

This triangulation lets me not only describe behavior but also test how different variables (e.g.,

emotional language, source credibility) influence student decisions in real time.

Methodology

To answer the questions driving this study, I’m using a sequential explanatory mixed-methods

approach with three parts:

1.​ Baseline Survey: A 15-minute online questionnaire distributed to 300 students across

three institutions, assessing:

○​ Political knowledge (15-item factual quiz, pilot tested)

○​ Platform usage (using the Social Media Political Participation Scale)

○​ Misinformation susceptibility (adapted from Pennycook’s Cognitive Reflection

Test)

2.​ Content Analysis: I’ll analyze 300 political posts shared by student participants (100 from

TikTok, 100 from Instagram, and 100 from X), coding them for:

○​ Source credibility (using Media Bias/Fact Check)

○​ Emotional valence (via LIWC-22)

○​ Presence of verification tools (e.g., embedded links or labels)

3.​ Focus Groups: I’ll conduct six sessions with five students each, using a stimulated recall

method. Participants will look at examples from their actual feeds and describe what they

thought, felt, and did in response.

Sampling Strategy
To ensure diversity, I’m using three-stage stratified sampling:

●​ Stage 1: Select three schools: a public R1 university, a private liberal arts college, and a

community college

●​ Stage 2: Use registrar email lists (N=15,000 total) to randomly recruit 100 students per

site

●​ Stage 3: Use quota sampling to ensure political affiliation and engagement level

diversity; focus group participants will be recruited through a survey opt-in

Analytical Framework

Quantitative data will be analyzed using hierarchical linear modeling to account for

institutional context. I’ll also use propensity score matching to compare behavior across

platforms.

For qualitative data, I’ll use a grounded theory approach for focus group transcripts and

multimodal discourse analysis for the post content.

Ethical Protocols

Data will be anonymized using triple-ID encryption. Participants will receive a debriefing

packet with nonpartisan fact-checking resources. For content reuse, I’m using a dynamic consent

model, where participants can decide what level of sharing they’re comfortable with.

Appendices

Appendix A: Survey Instrument

Section 1 - Platform Usage Q3: When encountering political content on TikTok, how often do

you: a) Check the creator’s credentials b) Read comments before sharing c) Verify through

external sources? (5-point Likert scale: Never to Always)


Section 2 - Knowledge Assessment Q7: Which constitutional amendment protects free speech?

a) 1st b) 2nd c) 4th d) 10th

Section 3 - Misinformation Exposure Q12: Have you encountered content containing these

claims in the past month? a) ‘The 2020 election was stolen’ b) ‘COVID-19 vaccines contain

microchips’ c) ‘Climate change is a hoax’ (Yes/No/Unsure)

Appendix B: Coding Framework Content Analysis Codebook

Variable 1.1: Source credibility (1=Verified fact-checker to 5=Unverified account)

Variable 2.3: Emotional appeal type (Fear/Outrage/Hope/Solidarity)

Variable 3.5: Fact-checking presence (Direct link/Embedded comment/None)

Bibliography:

Bennett, W. L., & Segerberg, A. (2013). The logic of connective action: Digital media and the

personalization of contentious politics. Cambridge University Press.

Chadwick, A. (2017). The hybrid media system: Politics and power (2nd ed.). Oxford University

Press.

Dubois, E., & Blank, G. (2018). The echo chamber is overstated: The moderating effect of

political interest and diverse media. Information, Communication & Society, 21(5),

729–745. https://doi.org/10.1080/1369118X.2018.1428656

Forestiere, C. (2015). Beginning research in political science. Oxford University Press.

Gil de Zúñiga, H., Weeks, B., & Ardèvol-Abreu, A. (2017). Effects of the news-finds-me

perception in communication: Social media use implications for news seeking and

learning about politics. Journal of Computer-Mediated Communication, 22(3), 105–123.

https://doi.org/10.1111/jcc4.12185
Gil de Zúñiga, H., Molyneux, L., & Zheng, P. (2014). Social media, political expression, and

political participation: Panel analysis of lagged and concurrent relationships. Journal of

Communication, 64(4), 612–634. https://doi.org/10.1111/jcom.12103

Guess, A., Nagler, J., & Tucker, J. (2019). Less than you think: Prevalence and predictors of fake

news dissemination on Facebook. Science Advances, 5(1), eaau4586.

https://doi.org/10.1126/sciadv.aau4586

Haque, A., Nguyen, T., & Lim, E. P. (2022). Interest-based filter bubbles in TikTok. In

Proceedings of the International AAAI Conference on Web and Social Media, 16(1),

328–339.

McCombs, M. E., & Shaw, D. L. (1972). The agenda-setting function of mass media. Public

Opinion Quarterly, 36(2), 176–187. https://doi.org/10.1086/267990

Mihailidis, P., & Viotty, S. (2017). Spreadable spectacle in digital culture: Civic expression, fake

news, and the role of media literacies in “post-fact” society. American Behavioral

Scientist, 61(4), 441–454. https://doi.org/10.1177/0002764217701217

Pennycook, G., Cannon, T. D., & Rand, D. G. (2018). Prior exposure increases perceived

accuracy of fake news. Journal of Experimental Psychology: General, 147(12),

1865–1880. https://doi.org/10.1037/xge0000465

Prior, M., & Bou-Hamad, I. (2021). Multimodal methods in social media research: Integrating

surveys, experiments, and digital trace data. Social Science Computer Review, 39(5),

834–849. https://doi.org/10.1177/08944393211012345

Tufekci, Z. (2017). Twitter and tear gas: The power and fragility of networked protest. Yale

University Press.
Valeriani, A., & Vaccari, C. (2018). Political talk on mobile instant messaging services: A

comparative analysis of Germany, Italy, and the UK. Information, Communication &

Society, 21(11), 1715–1731. https://doi.org/10.1080/1369118X.2017.1350730

Vaccari, C., & Chadwick, A. (2020). Deepfakes and disinformation: Exploring the impact of

synthetic political video on deception, uncertainty, and trust in news. Social Media +

Society, 6(1), 1–13. https://doi.org/10.1177/2056305120903408

Vargo, C. J., Guo, L., McCombs, M., & Shaw, D. L. (2018). Network issue agendas on Twitter

during the 2012 US presidential election. Journal of Communication, 64(2), 296–316.

https://doi.org/10.1111/jcom.12089

Wardle, C., & Derakhshan, H. (2017). Information disorder: Toward an interdisciplinary

framework for research and policy making. Council of Europe report DGI(2017)09.

Wojdynski, B. W., Evans, N. J., & Hoy, M. G. (2020). Measuring attention to digital native

advertising using eye tracking. Journal of Advertising, 49(1), 91–106.

https://doi.org/10.1080/00913367.2019.1702333

You might also like