<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" ><generator uri="https://jekyllrb.com/" version="3.10.0">Jekyll</generator><link href="https://ye-rim-oh.github.io/feed.xml" rel="self" type="application/atom+xml" /><link href="https://ye-rim-oh.github.io/" rel="alternate" type="text/html" /><updated>2026-04-04T21:01:44+09:00</updated><id>https://ye-rim-oh.github.io/feed.xml</id><title type="html">Yerim Oh</title><subtitle>Notes archive</subtitle><author><name>Yerim Oh</name></author><entry><title type="html">Research Ideas using AI</title><link href="https://ye-rim-oh.github.io/research-ideas-using-ai/" rel="alternate" type="text/html" title="Research Ideas using AI" /><published>2026-04-04T00:00:00+09:00</published><updated>2026-04-04T00:00:00+09:00</updated><id>https://ye-rim-oh.github.io/research-ideas-using-ai</id><content type="html" xml:base="https://ye-rim-oh.github.io/research-ideas-using-ai/"><![CDATA[<p>These are some research topics I had in mind for my PhD. I didn’t want them to go to waste, so I’m archiving them here. Please feel free to reach out via email if you’re interested!</p>

<hr />

<h2 id="beyond-ideal-points-quantifying-leader-personality-for-predicting-aggressive-foreign-policy-behavior">Beyond Ideal Points: Quantifying Leader Personality for Predicting Aggressive Foreign Policy Behavior</h2>

<h3 id="motivation">Motivation</h3>

<p>Ideal point estimation—the standard IR approach to modeling political actors, derived from voting records or survey responses—captures positional preferences but says little about the psychological dispositions that shape how those preferences translate into decisions under uncertainty. The gap matters most in authoritarian regimes, where weak institutional constraints mean policy outcomes track the individual at the top far more directly than structural accounts assume. If personalist dictatorships are defined by “weakened institutions and narrow support bases” that amplify individual agency, then incorporating personality into conflict prediction models is theoretically necessary, not an optional refinement.</p>

<h3 id="core-argument">Core Argument</h3>

<p>Leader personality, operationalized as a multidimensional psychological score, adds predictive value to models of aggressive foreign policy behavior (e.g., militarized interstate dispute initiation, escalatory rhetoric, territorial revisionism) beyond what positional ideal points alone capture. The effect is moderated by regime type: personality should predict behavior more strongly in personalist autocracies than in institutionally constrained democracies or party-based authoritarian systems.</p>

<h3 id="theoretical-grounding">Theoretical Grounding</h3>

<ul>
  <li>
    <p><strong>Authoritarian personality theory</strong> (Adorno et al. 1950): foundational framework linking dispositional traits—rigid cognitive style, in-group/out-group thinking, threat sensitivity—to aggressive outward behavior.</p>
  </li>
  <li>
    <p><strong>Leadership Trait Analysis (LTA)</strong> (Hermann 1999, 2003): seven operationalizable traits (belief in ability to control events, conceptual complexity, self-confidence, task vs. interpersonal orientation, distrust of others, in-group bias, need for power) shown to correlate with conflict propensity.</p>
  </li>
  <li>
    <p><strong>Operational Code Analysis (OCA)</strong> (Walker, Schafer &amp; Young 1998, 2006): quantitative measurement of a leader’s philosophical beliefs about the nature of the political universe and instrumental beliefs about optimal strategies. Outbreak of war is reliably associated with reduced conceptual complexity and elevated power motivation relative to affiliation motivation (Post 2003).</p>
  </li>
  <li>
    <p><strong>Neoclassical realism / Foreign Policy Analysis (FPA)</strong>: individual-level variables as intervening between structural pressures and state behavior; consistent with the “bringing the leader back in” agenda (Murray 2025).</p>
  </li>
</ul>

<h3 id="methodology">Methodology</h3>

<p>The methodology integrates four layers: (1) traditional at-a-distance content analysis as the baseline, (2) LLM-assisted automation for scalable corpus processing and trait scoring, (3) LLM persona simulation for prospective behavioral prediction, and (4) statistical outcome modeling. Each layer is distinct in purpose and can be validated independently.</p>

<hr />

<h4 id="layer-1-corpus-construction-and-classical-scoring-baseline">Layer 1: Corpus Construction and Classical Scoring (Baseline)</h4>

<ul>
  <li>
    <p><strong>Corpus</strong>: speeches, press conference transcripts, parliamentary or assembly records, diplomatic communiqués (sourced from governmental archives, FBIS/OSC, LexisNexis, UN Digital Library).</p>
  </li>
  <li>
    <p><strong>Classical tools</strong>: Profiler Plus (Hermann’s LTA software) for seven leadership traits; VICS coding scheme for OCA scores (P-1 through P-5 philosophical beliefs, I-1 through I-5 instrumental beliefs). This layer produces the validated baseline against which LLM-assisted outputs are benchmarked.</p>
  </li>
  <li>
    <p><strong>Output</strong>: a leader-level numeric profile, comparable across a normed reference group.</p>
  </li>
</ul>

<h4 id="layer-2-llm-assisted-trait-scoring-automation--scale">Layer 2: LLM-Assisted Trait Scoring (Automation + Scale)</h4>

<p>The classical LTA/OCA pipeline is resource-intensive: it requires trained coders, is slow, and does not scale easily to large multilingual corpora. LLMs can address all three constraints simultaneously.</p>

<p><strong>A. LLM as automated coder (zero-shot / few-shot)</strong></p>

<p>Recent work demonstrates that instruction-tuned LLMs outperform traditional dictionary methods and are competitive with supervised classifiers on political text annotation tasks, including ideology scaling and trait identification (Mao et al. 2025; Herrmann et al. 2025; Heseltine and Clemm von Hohenberg 2024). The proposed workflow:</p>

<ol>
  <li>
    <p>Chunk each leader’s speech corpus into utterance-level units.</p>
  </li>
  <li>
    <p>Prompt an LLM with the LTA trait definitions and request a scored annotation per chunk, with a brief chain-of-thought rationale. Example prompt structure:</p>
  </li>
</ol>

<blockquote>
  <p><em>“You are coding the following excerpt from a speech by [Leader, Year] for Leadership Trait Analysis. Evaluate the speaker’s expressed ‘Belief in Ability to Control Events’ on a scale of 0–10, where 10 indicates maximum conviction of personal control over political outcomes. Justify your score in one sentence.”</em></p>
</blockquote>

<ol>
  <li>
    <p>Aggregate chunk-level scores to the speech level, then to the leader-period level.</p>
  </li>
  <li>
    <p><strong>Validation</strong>: compare LLM-generated scores against a hand-coded gold-standard subset (minimum 200 utterances per leader). Report Pearson <em>r</em> and Cohen’s κ. Recent benchmarks suggest LLM–human agreement on political text tasks reaches r &gt; .85 for well-defined constructs (Herrmann et al. 2025).</p>
  </li>
</ol>

<p><strong>B. Multi-run averaging and uncertainty quantification</strong></p>

<p>LLM outputs are stochastic. Following the “asking and averaging” protocol (Herrmann et al. 2025), each utterance is scored across N=5–10 independent runs and the mean is taken as the point estimate, with the standard deviation as a proxy for coding uncertainty. Leaders with high score variance across runs should be flagged as ambiguous cases requiring human review.</p>

<p><strong>C. Hybrid human-LLM adjudication</strong></p>

<p>Where LLM runs disagree internally (SD &gt; threshold) or diverge from classical Profiler Plus scores, a human expert adjudicates. This “hybrid” approach has been shown to boost coding accuracy above either pure-LLM or pure-human approaches alone (Heseltine and Clemm von Hohenberg 2024).</p>

<p><strong>D. Multilingual extension</strong></p>

<p>LLM-based annotation is viable across languages without separate model training, enabling inclusion of leaders whose primary discourse is in Russian, Mandarin, Arabic, or Korean—historically excluded from LTA/OCA studies that depend on English transcripts. This substantially expands the effective universe of cases.</p>

<h4 id="layer-3-llm-persona-simulation-prospective-prediction-module">Layer 3: LLM Persona Simulation (Prospective Prediction Module)</h4>

<p>Beyond scoring existing texts, LLMs can simulate how a leader with a given personality profile would respond to hypothetical scenarios—a structured counterfactual generator. This is the most methodologically novel component, and the most epistemically risky. It should be framed explicitly as exploratory, not as a confirmatory test.</p>

<p><strong>A. Persona construction</strong></p>

<p>For each leader in the dataset, construct a structured persona prompt synthesizing:</p>

<ul>
  <li>
    <p>LTA trait scores (from Layers 1–2)</p>
  </li>
  <li>
    <p>OCA philosophical and instrumental belief indices</p>
  </li>
  <li>
    <p>Biographical context: formative experiences, tenure length, major prior crises</p>
  </li>
  <li>
    <p>Regime-structural context: institutional constraints, key advisers, domestic political pressures</p>
  </li>
</ul>

<p>Example persona injection:</p>

<blockquote>
  <p><em>“You are simulating the foreign policy reasoning of [Leader] as of [Year]. This leader scores [X] on Need for Power, [Y] on Distrust of Others, and [Z] on Conceptual Complexity (all normalized against the LTA reference group). Their operational code reflects [belief summary]. Given the following geopolitical scenario [scenario text], describe how this leader would most likely respond and assess the probability of a militarized action.”</em></p>
</blockquote>

<p><strong>B. Scenario battery</strong></p>

<p>Run each persona through a standardized battery of crisis scenarios, varying: (i) degree of external threat, (ii) domestic political stability, (iii) availability of diplomatic off-ramps. Record both the qualitative response and a solicited probability estimate for escalatory action.</p>

<p><strong>C. Backtesting</strong></p>

<p>Validate the simulation against known historical outcomes. For example: does a persona constructed from Saddam Hussein’s 1989–1990 OCA scores, when presented with a scenario structurally isomorphic to the Kuwait crisis, generate escalatory predictions at higher rates than a structurally similar democratic leader? This is the analogue of historical out-of-sample validation.</p>

<p><strong>D. Critical limitations of this layer</strong></p>

<ul>
  <li>
    <p>LLM persona simulation has been shown to work reasonably well at the <em>group level</em> but is less reliable at individual-level replication; individual-level persona variables explain &lt;10% of outcome variance in some benchmarks (Manning 2025).</p>
  </li>
  <li>
    <p>LLMs have training data priors that may systematically distort historical figure simulation—e.g., a model trained heavily on Western-language text may underweight the internal logic of leaders whose discourse is poorly represented in training corpora.</p>
  </li>
  <li>
    <p>This layer does not generate causal identification; it is a structured elicitation tool for hypothesis generation and plausibility assessment, not a confirmatory test.</p>
  </li>
</ul>

<h4 id="layer-4-statistical-estimation-outcome-modeling">Layer 4: Statistical Estimation (Outcome Modeling)</h4>

<ul>
  <li>
    <p><strong>Cross-sectional / panel regression</strong>: leader personality scores (from Layers 1–2) as key regressors in models of MID initiation (MID v5.0), ICEWS conflict event counts, or territorial revisionism. Controls: dyadic power ratio (Correlates of War), alliance portfolio (ATOP), GDP growth, leader tenure.</p>
  </li>
  <li>
    <p><strong>Heterogeneous effects</strong>: interact personality scores with regime type (Geddes, Wright &amp; Frantz classification) to test the moderation hypothesis that personality predicts behavior more strongly in personalist regimes.</p>
  </li>
  <li>
    <p><strong>Process tracing (qualitative)</strong>: for a subset of 3–5 cases, trace the causal chain: structural pressure → personality-filtered threat perception → domestic justification discourse → behavioral outcome. Primary sources for this layer should be cross-checked against Layer 3 simulation outputs as a consistency check.</p>
  </li>
</ul>

<h4 id="case-selection">Case Selection</h4>

<ul>
  <li>
    <p><strong>Universe</strong>: personalist authoritarian leaders in office ≥ 3 years, 1946–present.</p>
  </li>
  <li>
    <p><strong>Historical cases</strong> (deep process tracing): Stalin, Mao Zedong, Saddam Hussein, Kim Jong-il, Muammar Gaddafi, Slobodan Milošević.</p>
  </li>
  <li>
    <p><strong>Contemporary cases</strong> (real-time monitoring potential): Vladimir Putin, Xi Jinping, Kim Jong-un.</p>
  </li>
</ul>

<h3 id="expected-contributions">Expected Contributions</h3>

<ol>
  <li>
    <p>Methodological: personality scores derived from public texts are predictively valid beyond ideal point estimates—which are themselves derived from behavioral data and thus potentially endogenous to the outcome of interest.</p>
  </li>
  <li>
    <p>Theoretical: bridges quantitative IR (ideal point literature) and political psychology (LTA/OCA tradition), which have developed largely in parallel.</p>
  </li>
  <li>
    <p>Policy-relevant: personality profiles could be computed in near real-time from ongoing leader discourse, providing early-warning indicators of conflict initiation—a precedent exists in the CIA’s Center for the Analysis of Personality and Political Behavior, founded by Jerrold Post (Murray 2025).</p>
  </li>
</ol>

<h3 id="key-limitations">Key Limitations</h3>

<ul>
  <li>
    <p>Strategic communication: leaders may deliberately misrepresent beliefs in public statements, particularly in authoritarian contexts. Mitigation: cross-validate against private communications where available; weight sources by incentive structure.</p>
  </li>
  <li>
    <p>Selection bias: personalist leaders who initiate conflicts may generate more voluminous corpora, inflating apparent trait scores.</p>
  </li>
  <li>
    <p>Ecological validity of proxy instruments (MBTI, egogram): these are not validated for at-a-distance scoring of historical figures and should be framed as heuristic rather than confirmatory.</p>
  </li>
</ul>

<h3 id="reference">Reference</h3>

<p><strong>Political psychology / LTA / OCA foundations</strong></p>

<ul>
  <li>
    <p>Post, Jerrold M., ed. 2003. <em>The Psychological Assessment of Political Leaders: With Profiles of Saddam Hussein and Bill Clinton</em>. Ann Arbor: University of Michigan Press.</p>
  </li>
  <li>
    <p>Hermann, Margaret G. 1999. “Assessing Leadership Style: Trait Analysis.” In <em>The Psychological Assessment of Political Leaders: With Profiles of Saddam Hussein and Bill Clinton</em>, ed. Jerrold M. Post. Ann Arbor: University of Michigan Press.</p>
  </li>
  <li>
    <p>Walker, Stephen G., Mark Schafer, and Michael D. Young. 1998. “Systematic Procedures for Operational Code Analysis: Measuring and Modeling Jimmy Carter’s Operational Code.” <em>International Studies Quarterly</em> 42(1).</p>
  </li>
  <li>
    <p>Levinson, Daniel J. 1957. “Authoritarian Personality and Foreign Policy.” <em>Journal of Conflict Resolution</em> 1(1). https://doi.org/10.1177/002200275700100105.</p>
  </li>
  <li>
    <p>Renshon, Jonathan. 2008. “Stability and Change in Belief Systems: The Operational Code of George W. Bush from Governor to Second-Term President.” <em>Journal of Conflict Resolution</em> 52(6).</p>
  </li>
</ul>

<p><strong>LLM-assisted political text analysis (methodological grounding for Layer 2)</strong></p>

<ul>
  <li>
    <p>Mao, Honglin, Yaxin Li, Jindong Wang, Zeyu Chen, Siyuan Cheng, Xianchao Xiang, Qiwen Liu, et al. 2025. “Validating the Use of Large Language Models for Psychological Text Classification.” <em>Frontiers in Social Psychology</em>. https://www.frontiersin.org/journals/social-psychology/articles/10.3389/frsps.2025.1460277/full.</p>
  </li>
  <li>
    <p>Herrmann, Richard, et al. 2025. “Positioning Political Texts with Large Language Models by Asking and Averaging.” <em>Political Analysis</em> 33(3): 274–282. https://doi.org/10.1017/pan.2024.29.</p>
  </li>
  <li>
    <p>Heseltine, Michael, and Benno Clemm von Hohenberg. 2024. “Large Language Models as a Substitute for Human Experts in Annotating Political Text.” <em>Research &amp; Politics</em>. https://journals.sagepub.com/doi/10.1177/20531680241236239.</p>
  </li>
</ul>

<p><strong>LLM persona simulation (methodological grounding for Layer 3)</strong></p>

<ul>
  <li>
    <p>Kreutner, Martin, et al. 2025. “Persona-Driven Simulation of Voting Behavior in the European Parliament with Large Language Models.” <em>arXiv</em>. https://arxiv.org/abs/2506.11798.</p>
  </li>
  <li>
    <p>Manning, Andriy. 2025. “LLM-Based Persona Simulation: Survey of Methods and Limitations.” <em>Emergent Mind</em>. https://www.emergentmind.com/topics/llm-based-persona-simulation.</p>
  </li>
</ul>

<hr />

<h2 id="does-ai-use-make-people-more-conservative-or-more-progressive-a-longitudinal-field-experiment">Does AI Use Make People More Conservative or More Progressive? A Longitudinal Field Experiment</h2>

<h3 id="motivation-1">Motivation</h3>

<p>Large language models occupy an unusual position in the information ecosystem. Unlike social media algorithms, which optimize for engagement through homophilic content selection, LLMs generate responses dynamically and—by default—present ostensibly balanced perspectives. Mainstream LLMs have been consistently found to occupy a moderate left-of-center position on standardized political compass evaluations (Chowdhury and Robert 2025; Steimer 2025; Motoki, Pinho Neto, and Rodrigues 2024), yet early experimental evidence suggests they may actually <em>moderate</em> rather than amplify political views, at least over short time horizons (Voelkel et al. 2023; Bai et al. 2025). A structurally opposed dynamic is also plausible: sycophantic tendencies baked in through RLHF training may cause LLMs to mirror user beliefs over extended interactions, functioning as a personalized echo chamber analogous to social media filter bubbles (Calacci et al. 2026; Rathje et al. 2026; Jacob, Kerrigan, and Bastos 2025).</p>

<p>These two mechanisms—moderation through balanced deliberation vs. polarization through sycophantic mirroring—generate opposing predictions. Neither has been tested in a longitudinal field experiment with real-world AI usage patterns over a politically meaningful time horizon.</p>

<h3 id="research-questions">Research Questions</h3>

<ul>
  <li>
    <p><strong>RQ1</strong>: Does regular, unconstrained AI chatbot use shift individuals’ political attitudes over a six-month period, and in which direction?</p>
  </li>
  <li>
    <p><strong>RQ2</strong>: Does the effect differ by condition: (a) free-form AI use vs. (b) structured AI use with a system-level prompt designed to elicit balanced perspectives?</p>
  </li>
  <li>
    <p><strong>RQ3</strong>: Is the treatment effect moderated by baseline ideology, frequency of use, or topic domain of AI queries?</p>
  </li>
</ul>

<h3 id="experimental-design">Experimental Design</h3>

<h4 id="participants">Participants</h4>

<ul>
  <li>
    <p>Recruit N ≈ 600–900 adult participants via panel vendor (e.g., Prolific, KnPanel for Korean context), stratified by self-reported ideology (left / center / right) and prior AI usage.</p>
  </li>
  <li>
    <p>Pre-registration on OSF prior to data collection.</p>
  </li>
</ul>

<h4 id="random-assignment">Random Assignment</h4>

<p>Participants are randomly assigned to one of three conditions.</p>
<ul>
  <li>In the control condition, participants are not given access to the designated AI tool and instead continue their normal media diet without any study-provided chatbot intervention.</li>
  <li>In Treatment A, participants are given free access to a designated large language model under ordinary settings, with no modification to the system prompt, and their use of the tool is entirely self-directed.</li>
  <li>In Treatment B, participants are given access to the same model, but the system prompt is fixed in advance so that the chatbot is instructed to present multiple perspectives and avoid affirming a single viewpoint.</li>
</ul>

<p>The contrast between Treatment A and Treatment B isolates the effect of prompt-level intervention on top of raw AI exposure, enabling identification of how much of any observed attitude shift is attributable to sycophancy vs. content effects.</p>

<h4 id="measurement-protocol">Measurement Protocol</h4>

<ul>
  <li>
    <p><strong>T0</strong> (baseline): Political attitude battery—affective polarization scale (Iyengar et al.), issue positions on 8–10 policy domains, political identity strength.</p>
  </li>
  <li>
    <p><strong>T1</strong> (3 months): Repeat attitude battery + AI usage log (frequency, topic categories).</p>
  </li>
  <li>
    <p><strong>T2</strong> (6 months): Full repeat battery + open-ended qualitative questions on perceived AI influence.</p>
  </li>
  <li>
    <p><strong>Usage logging</strong>: Participants consent to anonymized logging of query topic categories (not content) via a lightweight browser extension or dedicated app interface.</p>
  </li>
</ul>

<h4 id="primary-outcome">Primary Outcome</h4>

<ul>
  <li>
    <p>Change in affective polarization score (T2 – T0) by treatment conditions.</p>
  </li>
  <li>
    <p>Change in issue position index.</p>
  </li>
</ul>

<h4 id="secondary-outcomes">Secondary Outcomes</h4>

<ul>
  <li>
    <p>Partisan identity strength.</p>
  </li>
  <li>
    <p>Epistemic confidence (how certain participants are in their own political views).</p>
  </li>
  <li>
    <p>Cross-cutting information exposure (whether participants encounter viewpoints opposite to their baseline).</p>
  </li>
</ul>

<h3 id="anticipated-mechanisms-and-heterogeneous-effects">Anticipated Mechanisms and Heterogeneous Effects</h3>

<p>The sycophancy-driven echo chamber mechanism predicts: Treatment A &gt; Control in polarization, concentrated among high-frequency users and users whose ideological identity is easily inferred by the model (Calacci et al. 2026). The deliberative moderation mechanism predicts: Treatment A &lt; Control in affective polarization, consistent with Bai et al. (PNAS 2025) and Voelkel et al. (PNAS 2023). Treatment B is predicted to amplify the moderating effect by constraining sycophantic mirroring.</p>

<h3 id="ethical-considerations">Ethical Considerations</h3>

<ul>
  <li>
    <p>Informed consent for ideological attitude measurement; full debriefing at T2.</p>
  </li>
  <li>
    <p>No deception; participants are told they are in a study of AI usage and political attitudes.</p>
  </li>
  <li>
    <p>IRB approval required; pre-registration mandatory.</p>
  </li>
  <li>
    <p>Data stored with identifiers separated from usage logs.</p>
  </li>
</ul>

<h3 id="theoretical-contributions">Theoretical Contributions</h3>

<ol>
  <li>
    <p>First longitudinal field experiment (as opposed to one-shot or two-week lab studies) to measure AI-induced attitude change over a politically meaningful time horizon.</p>
  </li>
  <li>
    <p>Identifies mechanism (sycophancy vs. deliberation) rather than documenting net effect.</p>
  </li>
  <li>
    <p>Generates causal identification strategy superior to observational studies of AI adoption.</p>
  </li>
</ol>

<h3 id="key-limitations-1">Key Limitations</h3>

<ul>
  <li>
    <p>Compliance monitoring: participants in Treatment conditions may not actually use the provided AI tool consistently; intent-to-treat and per-protocol analyses both required.</p>
  </li>
  <li>
    <p>External validity: lab-provision of AI tools may not replicate naturalistic AI adoption patterns.</p>
  </li>
  <li>
    <p>Temporal scope: six months may be insufficient to detect attitude crystallization; attitudes may return to baseline after intervention ends (cf. Bail et al. 2018 on social media exposure effects).</p>
  </li>
</ul>

<h3 id="reference-1">Reference</h3>

<ul>
  <li>
    <p>Voelkel, Jan G., et al. 2023. “Interventions Reducing Affective Polarization Do Not Necessarily Improve Anti-Democratic Attitudes.” <em>Nature Human Behaviour</em> 7: 55–64.</p>
  </li>
  <li>
    <p>Bai, Hui, Jan G. Voelkel, Sean Muldowney, and Robb Willer. 2025. “Testing Theories of Political Persuasion Using AI.” <em>Proceedings of the National Academy of Sciences</em> 122. https://www.pnas.org/doi/10.1073/pnas.2412815122.</p>
  </li>
  <li>
    <p>Voelkel, Jan G., et al. 2023. “Leveraging AI for Democratic Discourse.” <em>Proceedings of the National Academy of Sciences</em> 120(41). https://www.pnas.org/doi/10.1073/pnas.2311627120.</p>
  </li>
  <li>
    <p>Rathje, Steve, et al. 2026. “Sycophantic AI Increases Attitude Extremity and Overconfidence.” <em>PsyArXiv</em> preprint.</p>
  </li>
  <li>
    <p>Jacob, C., P. Kerrigan, and Marco Bastos. 2025. “The Chat-Chamber Effect: Trusting the AI Hallucination.” <em>Social Media + Society</em>. https://doi.org/10.1177/20539517241306345.</p>
  </li>
  <li>
    <p>Calacci, Dan, et al. 2026. “AI-Powered Chatbots Can Become Too Agreeable over Time.” Penn State Institute for Computational and Data Sciences news release. https://www.psu.edu/news/information-sciences-and-technology/story/ai-powered-chatbots-can-become-too-agreeable-over-time.</p>
  </li>
  <li>
    <p>Bail, Christopher A., et al. 2018. “Exposure to Opposing Views on Social Media Can Increase Political Polarization.” <em>Proceedings of the National Academy of Sciences</em> 115(37): 9216–9221.</p>
  </li>
  <li>
    <p>Steimer, Sarah. 2025. “Finding Political Leanings in Large Language Models.” University of Chicago Social Sciences Division. https://socialsciences.uchicago.edu/news/finding-political-leanings-large-language-models.</p>
  </li>
  <li>
    <p>Chowdhury, Rumman, and Kayla Schwoerer Robert. 2025. “Is the Politicization of Generative AI Inevitable?” <em>Brookings Institution</em>. https://www.brookings.edu/articles/is-the-politicization-of-generative-ai-inevitable/.</p>
  </li>
  <li>
    <p>Motoki, Fabio, Valdemar Pinho Neto, and Victor Rodrigues. 2024. “More Human than Human: Measuring ChatGPT Political Bias.” <em>arXiv</em>. https://arxiv.org/abs/2412.16746.</p>
  </li>
</ul>

<hr />
<h2 id="unequal-adoption-unequal-voice-identity-based-ai-utilization-gaps-and-asymmetric-political-participation">Unequal Adoption, Unequal Voice: Identity-Based AI Utilization Gaps and Asymmetric Political Participation</h2>

<h3 id="motivation-2">Motivation</h3>

<p>The existing literature on technology and democracy has predominantly examined AI’s effects from the top down—how states deploy AI for surveillance, how platforms use algorithms to manipulate opinion—while treating “users” as an undifferentiated mass. A separate body of work in gender studies and digital inequality documents that technology perception and use vary substantially by gender, race, religion, and socioeconomic status. The argument here is that these two literatures must be joined: if AI lowers the cost of constructing persuasive political narratives, mobilizing supporters, and flooding discursive spaces with content, then differential utilization capability across identity groups reshapes the relative capacity of groups to participate in politics, amplifying rather than simply reproducing existing inequalities. Groups that adopt and use AI more effectively gain structural advantages in agenda-setting, narrative construction, and mobilization. Groups that do not fall further behind.</p>

<p>The gender dimension is tractable empirically. Globally, women are approximately 20% less likely than men to use generative AI tools; women comprised 42% of ChatGPT’s 200 million monthly users between late 2022 and mid-2024, and only 27% of app downloads (Otis et al. 2024; Whillans et al. 2025). Income, education, and age account for little of the gap; it is explained primarily by differences in AI-specific knowledge and, secondarily, by gendered attitudes toward privacy risk (Bensoussane et al. 2024). The Carnegie Endowment flags the gender trust and adoption gap as an underappreciated risk to democratic participation (Wang 2025).</p>

<p>South Korea is a strong critical case. It combines high AI adoption rates (providing statistical power to detect differential utilization effects), the most pronounced gender ideological gap among young voters in any advanced democracy—in the 2022 presidential election, 59% of men aged 18–29 voted for the conservative candidate versus 34% of women in the same cohort (Steger 2025)—and a heavily gendered digital information environment, with men and women consuming politically divergent content in algorithmically siloed online spaces (Korea Times Editorial Board 2024; East Asia Forum 2024).</p>

<h3 id="core-argument-1">Core Argument</h3>

<p>Identity-based differences in AI adoption and utilization capability widen existing participation gaps. The mechanism is bottom-up: AI differentially empowers identity groups to construct alternative discursive realities, dominate agendas in online political spaces, and mobilize co-partisans at lower cost. The resulting asymmetry runs through two channels: (1) <em>agenda domination</em>—groups with higher utilization capability can flood political discourse with AI-assisted content at scale, crowding out less AI-capable groups; and (2) <em>illiberal mobilization</em>—AI-assisted narrative construction lowers the organizational cost of identity-based political movements, including anti-feminist and other reactionary mobilizations.</p>

<p>The central question, then, is not how technology erodes democracy but whom it empowers to participate, and at whose expense.</p>

<h3 id="conceptual-distinctions">Conceptual Distinctions</h3>

<p>The research distinguishes between two analytically separate dimensions of AI engagement:</p>

<ul>
  <li>
    <p><strong>Adoption level</strong>: frequency of AI use, purposes of use, and subjective perception of AI (trust, utility, risk).</p>
  </li>
  <li>
    <p><strong>Utilization capability</strong>: depth of engagement—ability to prompt effectively, evaluate outputs critically, and deploy AI outputs for political purposes (narrative creation, petition drafting, social media content generation, organizing messages).</p>
  </li>
</ul>

<p>The gap in <em>utilization capability</em> is likely more politically consequential than simple adoption frequency, but less studied. A group that uses AI daily for entertainment is not equivalent to a group that uses it weekly for political content production.</p>

<h3 id="methodology-1">Methodology</h3>

<h4 id="phase-1-conjoint-survey-south-korea-n--12001500">Phase 1: Conjoint Survey (South Korea, N ≈ 1,200–1,500)</h4>

<p><strong>Sampling</strong>: representative quota sample of Korean adults aged 19–39, stratified by gender and region. This age range captures the cohort in which gender ideological divergence is most pronounced.</p>

<p><strong>Measurement instruments</strong>:</p>

<p><em>Adoption module</em>: self-reported AI use frequency, platform diversity, stated purposes of use (entertainment, work, information-seeking, political/social purposes), perceived utility and risk.</p>

<p><em>Utilization capability module</em>: structured prompting task administered within the survey instrument. Respondents are given a standardized political scenario (e.g., “Write a social media post supporting/opposing [policy X]”) and asked to use an embedded AI interface. Output quality—persuasiveness, ideological specificity, factual accuracy—is scored by trained coders blind to respondent identity. This operationalizes utilization capability as a behaviorally grounded performance score rather than self-report.</p>

<p><em>Political behavior and attitude module</em>: voting intention, party identification, affective polarization (Korean adaptation of the Iyengar et al. scale), political efficacy, online political activity (posting, sharing, commenting), participation in offline political events.</p>

<p><em>Demographic module</em>: gender, age, education, income, employment sector, region, subjective social class.</p>

<p><strong>Analysis</strong>:</p>

<ul>
  <li>
    <p>OLS/logit regression of adoption and utilization capability on political participation outcomes, with gender as the primary moderator.</p>
  </li>
  <li>
    <p>Mediation analysis: does utilization capability mediate the relationship between gender and online political participation?</p>
  </li>
  <li>
    <p>Conjoint component: embed a conjoint experiment within the survey in which respondents evaluate hypothetical AI-assisted political messages varying in source identity (male/female, in-group/out-group) and AI authorship disclosure, measuring gendered reception of AI-assisted political content.</p>
  </li>
</ul>

<h4 id="phase-2-semi-structured-interviews-n--4060">Phase 2: Semi-Structured Interviews (N ≈ 40–60)</h4>

<p>Purposive sampling of survey respondents in extreme cells (high-adoption men, high-adoption women, low-adoption men, low-adoption women) plus political activists and community organizers who self-report using AI for political purposes.</p>

<p>Interview topics:</p>

<ul>
  <li>
    <p>How and for what purposes do respondents use AI in political/civic contexts?</p>
  </li>
  <li>
    <p>How do they evaluate AI-generated political content encountered online? Do they detect it?</p>
  </li>
  <li>
    <p>How do they perceive AI-literacy initiatives—are these seen as neutral technical training or as culturally coded?</p>
  </li>
  <li>
    <p>What are the barriers and facilitators of AI use for political engagement?</p>
  </li>
</ul>

<p>This phase provides contextual depth that cannot be captured quantitatively: the situated logic by which different groups come to use or avoid AI in political life.</p>

<h4 id="phase-3-comparative-extension">Phase 3: Comparative Extension</h4>

<p>Following the Korean case study, extend the design to a second Asian country with high AI adoption but a different primary identity cleavage—Indonesia (religious cleavage: Muslim vs. non-Muslim; urban vs. rural) or Taiwan (ethnic/national identity cleavage: Taiwanese vs. Chinese identity). Cross-national comparison tests the scope conditions of the theory: is the mechanism driven by gender-specific dynamics, or does it generalize to other salient identity cleavages?</p>

<h3 id="dependent-variables">Dependent Variables</h3>

<ul>
  <li>
    <p><strong>Primary</strong>: online political participation index (posting, sharing, signing petitions, organizing), offline political participation.</p>
  </li>
  <li>
    <p><strong>Secondary</strong>: political efficacy (internal and external), affective polarization, cross-cutting exposure (engagement with politically opposing content).</p>
  </li>
</ul>

<h3 id="theoretical-contributions-1">Theoretical Contributions</h3>

<ol>
  <li>
    <p>Integrates the digital inequality literature (adoption/utilization gaps) with the political participation literature (participation inequality), which have developed in relative isolation.</p>
  </li>
  <li>
    <p>Moves beyond aggregate “AI effects on democracy” framings toward a group-differentiated mechanism that specifies <em>who</em> is empowered and <em>through what channel</em>.</p>
  </li>
  <li>
    <p>Introduces utilization capability as a distinct, behaviorally operationalized construct that is theoretically and empirically separable from adoption frequency.</p>
  </li>
  <li>
    <p>Contributes a non-Western comparative case to a literature dominated by U.S. and European empirics.</p>
  </li>
</ol>

<h3 id="key-limitations-2">Key Limitations</h3>

<ul>
  <li>
    <p>Measurement of utilization capability via a survey-embedded task may not capture naturalistic AI use patterns; respondents may perform differently in a survey context than in everyday political engagement.</p>
  </li>
  <li>
    <p>Causal identification from cross-sectional survey data is limited; the Phase 1 design is primarily descriptive and hypothesis-generating. Causal claims require longitudinal follow-up or quasi-experimental leverage (e.g., comparing regions with differential AI platform availability).</p>
  </li>
  <li>
    <p>The Korean gender gap has multiple co-occurring causes (labor market precarity, military service, social media algorithms); isolating the marginal contribution of AI utilization gaps requires careful design and sensitivity analysis.</p>
  </li>
</ul>

<h3 id="reference-2">Reference</h3>

<p><strong>Empirical grounding for AI adoption gaps</strong></p>

<ul>
  <li>
    <p>Otis, N. G., Solène Delecourt, Kate Cranney, and Rembrand Koning. 2024. “Global Evidence on Gender Gaps and Generative AI.” Harvard Business School Working Paper No. 25-023. https://d3.harvard.edu/the-gender-divide-in-generative-ai-a-global-challenge/.</p>
  </li>
  <li>
    <p>Bensoussane, Inès, et al. 2024. “The Gen AI Gender Gap.” <em>Economics Letters</em>. https://www.sciencedirect.com/science/article/abs/pii/S0165176524002982.</p>
  </li>
  <li>
    <p>Wang, Vivian. 2025. “The Gender Trust Gap in AI: Implications for Democracy.” <em>Carnegie Endowment for International Peace</em>. https://carnegieendowment.org/china/posts/2025/10/ai-gender-trust-gap-democracy-implications.</p>
  </li>
  <li>
    <p>Whillans, Ashley, et al. 2025. “Gender Differences in Generative AI Adoption and Use.” Haas School of Business, University of California, Berkeley.</p>
  </li>
</ul>

<p><strong>Korean gender politics</strong></p>

<ul>
  <li>
    <p>Lee, Soohyun Christine. 2025. “Anti-Gender Politics, Economic Insecurity, and Right-Wing Populism among Young Men in South Korea.” <em>Social Politics</em>. https://academic.oup.com/sp/article/32/3/584/7826751.</p>
  </li>
  <li>
    <p>East Asia Forum. 2024. “Gender Is Reshaping South Korea’s Electoral Landscape.” https://eastasiaforum.org/2024/07/09/gender-is-reshaping-south-koreas-electoral-landscape/.</p>
  </li>
  <li>
    <p>Steger, Isabella. 2025. “South Korea’s 2025 Election: A Test for Gender Equality.” <em>The Diplomat</em>. https://thediplomat.com/2025/05/south-koreas-2025-election-a-test-for-gender-equality/.</p>
  </li>
  <li>
    <p>Korea Times Editorial Board. 2024. “Gender Divide.” <em>The Korea Times</em>. https://www.koreatimes.co.kr/www/opinion/2024/12/197_387003.html.</p>
  </li>
</ul>

<p><strong>Technology and political participation</strong></p>

<ul>
  <li>Bail, Christopher A., et al. 2018. “Exposure to Opposing Views on Social Media Can Increase Political Polarization.” <em>Proceedings of the National Academy of Sciences</em> 115(37): 9216–9221.</li>
</ul>

<hr />

<h2 id="electoral-seasonality-and-gender-targeted-mobilization-in-korean-political-twitter-networks">Electoral Seasonality and Gender-Targeted Mobilization in Korean Political Twitter Networks</h2>

<h2 id="motivation-3">Motivation</h2>

<p>Korean conservative politics has undergone a structural transformation since the 2022 presidential election, in which the People Power Party (PPP) ran an explicitly anti-feminist campaign and successfully channeled the grievances of young men (이대남, <em>yi-dae-nam</em>) into a decisive electoral bloc (Lee &amp; Kim 2025). What remains empirically underexamined is whether this mobilization strategy extends beyond anti-feminist messaging toward young <em>women</em>—specifically, whether conservative actors attempt to selectively recruit or reframe issues for female audiences during election periods. The intuition motivating this project is that electoral seasons create concentrated windows of strategic communication, during which political networks may restructure themselves: new bridges form between previously disconnected communities, influence hierarchies shift, and gendered targeting becomes more deliberate.</p>

<p>Twitter/X is a particularly suitable platform for observing this dynamic in the Korean context for three reasons. First, despite declining global user numbers, Twitter remains the dominant platform for elite political communication in Korea, including politicians, journalists, political influencers, and organized advocacy groups. Second, its retweet structure generates observable endorsement networks that can be used for community detection and ideological positioning without requiring content-level inference about user identity. Third, the temporal granularity of Twitter data enables before/during/after comparisons across electoral cycles.</p>

<p>The central empirical question is: Does the structural topology of Korean conservative Twitter networks change in gender-specific ways during election periods, and do these changes exhibit patterns consistent with deliberate outreach toward female users?</p>

<h2 id="core-argument-2">Core Argument</h2>

<p>Conservative political actors in Korea engage in electorally seasonal network restructuring on Twitter, characterized by: (i) increased bridging ties between the conservative community cluster and accounts primarily retweeted by female users; (ii) elevated centrality of accounts that mix anti-feminist grievance content with women-facing issue framing (e.g., safety, economic anxiety, family policy); and (iii) accelerated cross-community diffusion of specific message types during the four-to-six-week window preceding elections. If confirmed, this constitutes evidence of deliberate gendered mobilization strategy operating at the network level—beyond what content analysis of individual accounts would reveal.</p>

<h2 id="data">Data</h2>

<h3 id="source-and-collection">Source and Collection</h3>

<ul>
  <li><strong>Platform</strong>: Twitter/X Academic Research API (now Basic/Pro tier; see API cost discussion below).</li>
  <li><strong>Keywords and hashtag seeds</strong>: constructed from major Korean political terms, candidate names, party hashtags, and gender-salient issue terms (여성, 페미니즘, 이대녀, 이대남, 여가부, 군가산점, etc.) across election windows.</li>
  <li><strong>Target elections</strong> (four observation windows):
    <ul>
      <li>2022 presidential election (March): T−6 weeks, T−3 weeks, T−1 week, T+2 weeks</li>
      <li>2022 local elections (June)</li>
      <li>2024 parliamentary election (April)</li>
      <li>2025 presidential election (June): most recent case, real-time collection feasible if project is initiated promptly</li>
    </ul>
  </li>
  <li><strong>Corpus size target</strong>: 5–10 million tweets per election window is achievable under current API tiers; 2022 data requires retrospective access which is now commercially tiered.</li>
</ul>

<h4 id="user-level-gender-inference">User-Level Gender Inference</h4>

<p>Gender is not directly observable on Twitter. Three complementary approaches:</p>

<ol>
  <li><strong>Name-based inference</strong>: Korean given names carry strong gender signal. Apply a Korean name-gender lookup table (e.g., derived from KOSIS population registry name-gender distributions) to screen-names and display names. Precision is high for clearly gendered names; ambiguous and handle-only accounts are flagged as unknown.</li>
  <li><strong>Bio-text classification</strong>: LLM-based zero-shot classification of profile bios for gender-indicative language (self-referential terms: 엄마, 직장인 여성, etc.). Validated against name-based inference on overlap cases.</li>
  <li><strong>Network-community proxy</strong>: After community detection, label communities by aggregate gender composition using accounts with high-confidence individual gender inference, then assign community-level gender tendency scores. This is a distributional, not individual-level, inference.</li>
</ol>

<p><strong>Critical caveat</strong>: gender inference at the individual account level is inherently noisy and must be treated as probabilistic. All individual-level inferences should be aggregated to the community level before drawing conclusions; individual-level gender attribution should not be reported.</p>

<h3 id="methodology-2">Methodology</h3>

<h4 id="step-1-retweet-network-construction">Step 1: Retweet Network Construction</h4>

<p>For each election window, construct a directed weighted retweet graph G = (V, E, w) where:</p>
<ul>
  <li>V = set of unique users who authored or were retweeted</li>
  <li>E = directed edge from user <em>u</em> to user <em>v</em> if <em>u</em> retweeted <em>v</em></li>
  <li>w(u, v) = number of retweets</li>
</ul>

<p>Filter to the giant connected component. Prune edges below a minimum weight threshold (e.g., w ≥ 2) to reduce noise from one-off interactions.</p>

<h4 id="step-2-community-detection">Step 2: Community Detection</h4>

<p>Apply the <strong>Louvain algorithm</strong> (Blondel et al. 2008) to the undirected projection of the retweet graph, optimizing modularity. This is the standard approach in the Twitter political network literature and has been validated on Korean Twitter data (Kim &amp; Park 2012; Yoo &amp; Kim 2024). Target resolution: 3–6 major communities (progressive, conservative, anti-feminist/manosphere, feminist/progressive women, centrist/news-media, non-political).</p>

<p>Assign ideological labels to communities via:</p>
<ul>
  <li>Seed accounts: known politicians, party accounts, and major political media outlets manually labeled.</li>
  <li>Hashtag co-occurrence: dominant hashtags within each community provide content-level validation of the structural label.</li>
</ul>

<h4 id="step-3-gender-community-distribution-analysis">Step 3: Gender-Community Distribution Analysis</h4>

<p>For each community, compute:</p>
<ul>
  <li><strong>Gender composition score</strong>: proportion of high-confidence gender-inferred accounts that are female.</li>
  <li><strong>Gender homophily index</strong>: whether female-inferred accounts preferentially retweet other female-inferred accounts within or across communities.</li>
</ul>

<p>Compare gender composition across communities and across election windows (non-election baseline → election window) to detect structural shifts.</p>

<h4 id="step-4-bridge-account-detection-and-profiling">Step 4: Bridge Account Detection and Profiling</h4>

<p>Identify <strong>bridge accounts</strong>—accounts that connect the conservative community to communities with higher female composition—using:</p>
<ul>
  <li><strong>Betweenness centrality</strong> in the inter-community subgraph</li>
  <li><strong>Community membership change</strong>: accounts that migrate from one community cluster to another across time windows (following the method in Ramaciotti Morales &amp; Cointet 2024)</li>
</ul>

<p>Profile bridge accounts by content: what topics do they tweet about? Are they new accounts or existing ones? What is their follower-to-following ratio (potential astroturfing indicator)?</p>

<h4 id="step-5-temporal-dynamics--electoral-seasonality-test">Step 5: Temporal Dynamics — Electoral Seasonality Test</h4>

<p>The core empirical claim is that network restructuring is electorally seasonal. Test this with:</p>

<ul>
  <li><strong>Inter-community retweet ratio</strong> (ICR): ratio of cross-community to within-community retweets, computed weekly. Prior work shows ICR decreases sharply in the weeks preceding elections and recovers post-election (Dogdu, Choupani &amp; Sürücü 2024)—but the <em>direction</em> of cross-community flow (who is reaching out to whom) is what matters here, not just the volume.</li>
  <li><strong>Granger causality test</strong>: does the emergence of bridge accounts between the conservative cluster and female-leaning communities Granger-cause increases in retweet volume from those female-leaning communities toward conservative content?</li>
  <li><strong>Difference-in-differences (DiD)</strong>: compare the change in conservative → female-community bridge density across election windows (treatment) vs. non-election periods (control), using the 2022–2025 panel.</li>
</ul>

<h4 id="step-6-content-analysis-of-bridging-messages-llm-assisted">Step 6: Content Analysis of Bridging Messages (LLM-Assisted)</h4>

<p>For the top-k (k ≈ 500–1,000) tweets authored by bridge accounts during election windows, apply LLM-based stance and frame classification:</p>

<ul>
  <li><strong>Issue frame</strong>: economic anxiety / safety / family policy / anti-feminism / anti-DP / other</li>
  <li><strong>Target audience signal</strong>: explicit references to 여성, 이대녀, 엄마, or gender-neutral framing</li>
  <li><strong>Sentiment toward feminism</strong>: positive / neutral / critical / hostile</li>
</ul>

<p>This step tests whether bridging is accompanied by a <em>softening</em> of anti-feminist rhetoric (suggesting strategic reframing for female audiences) or maintains hostile framing (suggesting co-optation of women’s issue vocabulary without attitudinal moderation).</p>

<h3 id="key-hypotheses">Key Hypotheses</h3>

<table>
  <thead>
    <tr>
      <th>Hypothesis</th>
      <th>Operationalization</th>
    </tr>
  </thead>
  <tbody>
    <tr>
      <td>H1: Electoral seasonality</td>
      <td>ICR between conservative cluster and female-leaning communities increases in election windows relative to baseline</td>
    </tr>
    <tr>
      <td>H2: Directional asymmetry</td>
      <td>Conservative→female-community bridging increases more than female-community→conservative bridging</td>
    </tr>
    <tr>
      <td>H3: Strategic reframing</td>
      <td>Bridge account content during election windows shows reduced anti-feminist framing and increased women’s issue vocabulary relative to non-election content from the same accounts</td>
    </tr>
    <tr>
      <td>H4: New bridge formation</td>
      <td>Bridge accounts active in election windows are disproportionately recently created or previously low-centrality accounts, suggesting coordinated injection rather than organic community evolution</td>
    </tr>
  </tbody>
</table>

<p>H4 is the “astroturfing” hypothesis—the most provocative and hardest to definitively confirm without ground truth, but addressable with account age, bot-score (Botometer or equivalent), and network injection pattern analysis.</p>

<h3 id="api-cost-and-feasibility-discussion">API Cost and Feasibility Discussion</h3>

<p>This is the central practical constraint. Twitter/X API access tiers as of 2025:</p>

<ul>
  <li><strong>Free tier</strong>: 1,500 tweets/month read; effectively useless for this project.</li>
  <li><strong>Basic tier</strong> (~USD 100/month): 10,000 tweets/month read. Insufficient for network-scale collection.</li>
  <li><strong>Pro tier</strong> (~USD 5,000/month): 1 million tweets/month read. Feasible for one election window per month; cumulative cost for four elections would be prohibitive without grant funding.</li>
  <li><strong>Academic/Enterprise tier</strong>: no longer openly available post-Musk acquisition; must be negotiated directly.</li>
</ul>

<p><strong>Feasible mitigation strategies</strong>:</p>

<ol>
  <li><strong>Retrospective data from third-party archives</strong>: TweetSets (Library of Congress), Internet Archive Twitter Stream Grab (2022 data available), and academic data-sharing consortia (e.g., ICPSR, Harvard Dataverse political Twitter deposits) may provide 2022 election data at no marginal cost.</li>
  <li><strong>Keyword-targeted narrow collection</strong>: instead of broad streaming, collect only tweets containing a pre-specified set of high-signal hashtags and keywords. This reduces volume by ~80–90% while retaining the politically active network core.</li>
  <li><strong>Snowball sampling from seed accounts</strong>: starting from a list of major Korean political accounts (politicians, party accounts, top political influencers), collect their follower/following networks and tweets. This trades representativeness for depth on the politically active subgraph.</li>
  <li><strong>Collaboration with ICPSR / Korean political communication researchers</strong>: Korean political science and communication departments (SNU, Yonsei, KAIST) may have existing data-sharing agreements or archived datasets.</li>
  <li><strong>Focus on 2024 and 2025 elections</strong>: retrospective API access is more expensive; real-time collection starting now for the 2025 presidential election data is feasible at the Basic–Pro tier with careful keyword targeting.</li>
</ol>

<p>Realistic minimum budget for a properly powered study: USD 3,000–8,000 in API costs, plus compute time for network analysis (manageable on a university HPC cluster or AWS spot instances).</p>

<h3 id="expected-contributions-1">Expected Contributions</h3>

<ol>
  <li><strong>Empirical</strong>: first systematic, longitudinal network analysis of Korean political Twitter with explicit gender-disaggregation and electoral seasonality as the organizing variable.</li>
  <li><strong>Theoretical</strong>: bridges the political network analysis literature (which has focused on US and European cases) and the Korean gender politics literature (which has relied primarily on survey and electoral data), demonstrating that mobilization strategy is observable in network topology, not only in content.</li>
  <li><strong>Methodological</strong>: demonstrates a pipeline for probabilistic gender inference in non-English Twitter networks using name-based and bio-text LLM classification, with explicit uncertainty quantification at the community rather than individual level.</li>
</ol>

<h3 id="key-limitations-3">Key Limitations</h3>

<ul>
  <li><strong>Gender inference validity</strong>: name-based inference is noisier for Korean Twitter than for English-language platforms because many users adopt non-name handles; community-level aggregation partially mitigates this but introduces ecological fallacy risk.</li>
  <li><strong>Platform representativeness</strong>: Twitter/X overrepresents politically engaged, educated, urban users; findings may not generalize to the broader Korean electorate or to Kakao/Naver-based political communities, which are structurally different.</li>
  <li><strong>Causal identification</strong>: network analysis is inherently observational; demonstrating that conservative actors <em>deliberately</em> target female users requires additional evidence (e.g., account coordination patterns, content strategy documents) beyond structural network change.</li>
  <li><strong>API access instability</strong>: X Corp’s API policy has changed multiple times since 2022 and may change again; any real-time collection plan carries execution risk.</li>
</ul>

<h3 id="reference-3">Reference</h3>

<p><strong>Korean political network / gender politics</strong></p>

<p>Chung, Haejin. 2024. “Gender Wars and Populist Politics in South Korea.” <em>Women’s Studies International Forum</em> 104: 102906. https://www.sciencedirect.com/science/article/pii/S0277539524000530</p>

<p>Kwon, Eunji, and Sooyeon Kim. 2025. “Blaming Feminists, Claiming Pain: Anti-Feminist Discourse and Electoral Mobilization by New Men’s Solidarity in South Korea.” <em>Women’s Studies International Forum</em> 110: 102986. https://www.sciencedirect.com/science/article/pii/S0277539525001086</p>

<p>Lee, Nae-Young, and Hyeshin Kim. 2025. “Anti-Gender Politics, Economic Insecurity, and Right-Wing Populism: The Rise of Modern Sexism among Young Men in South Korea.” <em>Social Politics: International Studies in Gender, State &amp; Society</em> 32(3): 584–614. https://academic.oup.com/sp/article/32/3/584/7826751</p>

<p>Yoo, Jungwon, and Hyunjin Kim. 2024. “Semantic Networks of Election Fraud: Comparing the Twitter Discourses of the U.S. and Korean Presidential Elections.” <em>Social Sciences</em> 13(2): 94. https://www.mdpi.com/2076-0760/13/2/94</p>

<p><strong>Twitter political network methodology</strong></p>

<p>Flamino, James, Alessandro Galeazzi, Stuart Feldman, Michael W. Macy, Brendan Cross, Zhenkun Zhou, Matteo Serafino, Alexandre Bovet, Hernán A. Makse, and Boleslaw K. Szymanski. 2023. “Political Polarization of News Media and Influencers on Twitter in the 2016 and 2020 US Presidential Elections.” <em>Nature Human Behaviour</em> 7(6): 904–916. https://doi.org/10.1038/s41562-023-01550-8</p>

<p>Guerrero-Solé, Frederic. 2017. “Community Detection in Political Discussions on Twitter: An Application of the Retweet Overlap Network Method to the Catalan Process Toward Independence.” <em>Social Science Computer Review</em> 35(2): 244–261. https://doi.org/10.1177/0894439315617254</p>

<p>Dogdu, Erdogan, Ramin Choupani, and Selin Sürücü. 2024. “Detecting Political Polarization Using Social Media Data.” In <em>Advances in Intelligent Systems and Computing</em>, 3–14. Cham: Springer. https://link.springer.com/chapter/10.1007/978-3-031-61816-1_4</p>

<p>Ramaciotti Morales, Pedro, and Jean-Philippe Cointet. 2024. “Polarization Dynamics: A Study of Individuals Shifting Between Political Communities on Social Media.” arXiv:2408.07731. https://arxiv.org/html/2408.07731</p>

<p><strong>LLM-assisted content classification</strong></p>

<p>Ye, Jinyi, Luca Luceri, and Emilio Ferrara. 2025. “Auditing Political Exposure Bias: Algorithmic Amplification on Twitter/X During the 2024 U.S. Presidential Election.” In <em>Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency (FAccT ‘25)</em>. https://doi.org/10.1145/3715275.3732159</p>

<p>Zhu, Yixuan, Sotiris Siatras, and Lucia Siciliani. 2024. “Analyzing Political Stances on Twitter in the Lead-Up to the 2024 U.S. Election.” arXiv:2412.02712. https://arxiv.org/html/2412.02712</p>]]></content><author><name>Yerim Oh</name></author><category term="AI" /><summary type="html"><![CDATA[These are some research topics I had in mind for my PhD. I didn’t want them to go to waste, so I’m archiving them here. Please feel free to reach out via email if you’re interested!]]></summary></entry><entry><title type="html">What Is your “Political Blood Type”?</title><link href="https://ye-rim-oh.github.io/what-is-a-political-blood-type/" rel="alternate" type="text/html" title="What Is your “Political Blood Type”?" /><published>2026-03-13T00:00:00+09:00</published><updated>2026-03-13T00:00:00+09:00</updated><id>https://ye-rim-oh.github.io/what-is-a-%E2%80%9Cpolitical-blood-type%E2%80%9D</id><content type="html" xml:base="https://ye-rim-oh.github.io/what-is-a-political-blood-type/"><![CDATA[<h3 id="intriguing-free-data">Intriguing <em>Free</em> Data</h3>

<p>Before the South Korea’s 21st presidential election (June 3, 2025), South Korea’s public broadcaster MBC conducted a survey using a familiar format called the “Political Blood Type Test.” As of today (March 13, 2026), more than 300,000 people have visited the site.</p>

<p>After completing the survey, respondents can get to a section titled “How did other people answer?”, which provides a page displaying the response distributions for all 3,027 participants, as well as breakdowns by gender and age group.</p>

<p>If I had tried to run a 500-person survey on Qualtrics, it would probably have cost more than $2,000. For a master’s student like me, being able to examine survey data from more than 3,000 respondents for free was interesting. That is why I collected both the questionnaire items and the response distributions and saved them as CSV files.</p>

<h3 id="back-to-2025">Back to 2025</h3>

<p>Back in May 2025, I gave the result chart image files to Gemini 2.5 Pro, ChatGPT, and Claude (cannot remember exact models) and asked them to analyze it, but all of them got the detailed numbers wrong.</p>

<p>In fact, the conclusion of the blog post I wrote by myself on May 17, 2025, was less about the survey itself and more about my disappointment with AI. To quote only the concluding part:</p>

<ol>
  <li>
    <p>Do not ask an LLM to do much with chart images.</p>
  </li>
  <li>
    <p>Do not leave the “interpretation” of charts or data to an LLM.</p>
  </li>
  <li>
    <p>In South Korea, there is no state religion, and public conflict over race has not yet become especially salient, so the only 
issues that seem capable of generating real political cleavage are things like women’s rights and economic redistribution. It all felt a bit repetitive.</p>
  </li>
</ol>

<p>That was basically what I wrote.
This time, I asked Claude Code to use the data directly, build a dashboard itself, and write an analytical report based on the actual files.</p>

<h3 id="result">Result</h3>

<p>
    <a class="btn btn--primary dashboard-open-btn" href="https://z25die-0-0.shinyapps.io/mbc-survey-dashboard/" target="_blank" rel="noopener">
      Open Dashboard in New Tab
    </a>
  </p>

<div class="dashboard-mobile-wrap">
    <iframe class="dashboard-mobile-frame" src="https://z25die-0-0.shinyapps.io/mbc-survey-dashboard/" title="MBC Survey Social Values Dashboard" width="100%" height="760" style="border: 1px solid #ddd; border-radius: 12px; background: #fff;">
    </iframe>
  </div>

<p>Koreans in this survey broadly endorse fairness, care, intergenerational duty, and punishment for wrongdoing. Those commitments are real, but they do not settle the political conflicts that matter most. The sharpest disagreements concern family obligation, gender inequality, and the meaning of legitimate order.</p>

<p>The five-index regression sharpens that point. Traditional family norms show the strongest age gradient. Gender inequality skepticism shows the strongest young male-female split. Procedural fairness remains broadly shared. Authoritarian order is high across much of the age-gender map rather than isolated at one pole. Alienation stays low, but it is still concentrated among younger men. Taken together, these results do not fit a clean left-right story. Instead, they point to a common procedural vocabulary attached to rival social interpretations of what fairness demands.</p>

<p>If that reading is right, Korean polarization may deepen because citizens attach incompatible meanings to a shared moral language, not because that language is absent. That is a harder political problem. Appeals to fairness alone will not resolve it, because fairness is already the shared term. The conflict is over whose experience, whose burden, and whose injury counts as unfair.</p>

<ul>
  <li>Detailed Analysis including unweighted cell-level OLS regression is available in <a href="https://github.com/ye-rim-oh/mbc-survey-dashboard">my github repo</a>.</li>
</ul>

<h3 id="for-reference">For reference</h3>

<p>🩸What is Political Blood Type? <a href="http://plood.poll-mbc.co.kr/">Official Page</a></p>

<p>✔️ If you are tired of politics being reduced to a simple progressive-versus-conservative divide,</p>

<p>✔️ If politics feels difficult or burdensome,</p>

<p>✔️ If you want to understand your political disposition as simply as finding out your blood type,</p>

<p>“Political Blood Type” is a political personality assessment service that classifies people into 16 types based on four psychological and value-oriented dimensions. Through 36 questions, it identifies the combination of one’s “innate disposition” and the “values formed through life experience,” and maps that combination onto a broader political spectrum.</p>

<p>More precise than the ideology you think you hold, and more nuanced than a rigid progressive-conservative split, 🩸Political Blood Type offers a more detailed analysis of political orientation. Shall we find out how much 💓 Empathy Hemoglobin, 🌪️ Reset White Cells, 🧭 Tradition Hormone, and 🏛️ Hierarchy Receptor you have?</p>

<p>📎 Political Blood Type was designed on the basis of a social survey of 3,027 respondents.</p>

<p>📎 Advisory panel: Professor John Seungmin Kuk (Michigan State University), Professor SunKyoung Park (Korea University), Professor Shang E. Ha (Sogang University)</p>

<hr />

<p>Please note that the original questionnaire was, of course, written in Korean, so there may be slight differences in nuance between the original wording and the currently translated English items.</p>]]></content><author><name>Yerim Oh</name></author><summary type="html"><![CDATA[Intriguing Free Data]]></summary></entry><entry><title type="html">GradCafe-Based Political Science PhD Admission Trends, 2020-2026</title><link href="https://ye-rim-oh.github.io/gradcafe-2020-2026-trend-note/" rel="alternate" type="text/html" title="GradCafe-Based Political Science PhD Admission Trends, 2020-2026" /><published>2026-03-04T00:00:00+09:00</published><updated>2026-03-12T00:00:00+09:00</updated><id>https://ye-rim-oh.github.io/gradcafe-2020-2026-trend-note</id><content type="html" xml:base="https://ye-rim-oh.github.io/gradcafe-2020-2026-trend-note/"><![CDATA[<h2 id="introduction">Introduction</h2>

<p>I scraped and organized political science PhD admission data submitted to GradCafe from 2020 to 2026, then noted a few patterns that seemed meaningful from an applicant’s point of view.</p>
<ul>
  <li>Base repository: <a href="https://github.com/ye-rim-oh/GradCafe">ye-rim-oh/GradCafe</a></li>
</ul>

<blockquote>
  <p>Update: as of 2026-03-04, I scraped the full 2020-2026 seasons again and updated the figures.</p>
</blockquote>

<h2 id="interactive-dashboard">Interactive Dashboard</h2>

<p>
  <a class="btn btn--primary dashboard-open-btn" href="https://z25die-0-0.shinyapps.io/gradcafe-master/" target="_blank" rel="noopener">
    Open Dashboard in New Tab
  </a>
</p>

<div class="dashboard-mobile-wrap">
  <iframe class="dashboard-mobile-frame" src="https://z25die-0-0.shinyapps.io/gradcafe-master/" title="GradCafe Shiny Dashboard" width="100%" height="760" style="border: 1px solid #ddd; border-radius: 12px; background: #fff;">
  </iframe>
</div>

<h2 id="points-to-keep-in-mind">Points to Keep in Mind</h2>

<ul>
  <li>GradCafe reports are self-posted by applicants, so they do not represent the full population.</li>
  <li>One applicant may apply to multiple schools and report multiple outcomes, so 12 reported cases do not necessarily mean 12 distinct people. It could just as easily mean 6 people reporting 2 schools each, or 3 people reporting 4 schools each.</li>
  <li>GRE scores, subfield information, and nationality information are especially affected by missingness and reporting bias.</li>
  <li>The figures in this post are current as of March 4, 2026.</li>
  <li>Even so, the dataset is still useful for reading broad patterns, such as when announcements cluster by school and how competitive a given year feels.</li>
</ul>

<h2 id="key-summary">Key Summary</h2>

<h3 id="summary-table">Summary Table</h3>

<table>
  <thead>
    <tr>
      <th>Year</th>
      <th>Total Cases</th>
      <th>Overall Accept Rate</th>
      <th>American Cases</th>
      <th>American Accept Rate</th>
      <th>International Cases</th>
      <th>International Accept Rate</th>
      <th>GRE V mean</th>
      <th>GRE Q mean</th>
      <th>GPA Reporting Rate</th>
    </tr>
  </thead>
  <tbody>
    <tr>
      <td>2020</td>
      <td>593</td>
      <td>46.4%</td>
      <td>214</td>
      <td>48.5%</td>
      <td>304</td>
      <td>42.6%</td>
      <td>161.3</td>
      <td>159.6</td>
      <td>31.4%</td>
    </tr>
    <tr>
      <td>2021</td>
      <td>576</td>
      <td>35.3%</td>
      <td>206</td>
      <td>37.9%</td>
      <td>254</td>
      <td>29.9%</td>
      <td>164.5</td>
      <td>161.1</td>
      <td>29.9%</td>
    </tr>
    <tr>
      <td>2022</td>
      <td>495</td>
      <td>45.5%</td>
      <td>205</td>
      <td>41.1%</td>
      <td>289</td>
      <td>48.3%</td>
      <td>163.0</td>
      <td>162.8</td>
      <td>54.7%</td>
    </tr>
    <tr>
      <td>2023</td>
      <td>337</td>
      <td>46.1%</td>
      <td>94</td>
      <td>58.8%</td>
      <td>184</td>
      <td>47.4%</td>
      <td>163.2</td>
      <td>163.1</td>
      <td>55.8%</td>
    </tr>
    <tr>
      <td>2024</td>
      <td>402</td>
      <td>44.6%</td>
      <td>138</td>
      <td>42.5%</td>
      <td>259</td>
      <td>45.1%</td>
      <td>163.6</td>
      <td>162.2</td>
      <td>51.5%</td>
    </tr>
    <tr>
      <td>2025</td>
      <td>505</td>
      <td>36.5%</td>
      <td>176</td>
      <td>41.5%</td>
      <td>329</td>
      <td>33.7%</td>
      <td>163.4</td>
      <td>165.3</td>
      <td>57.8%</td>
    </tr>
    <tr>
      <td>2026</td>
      <td>858</td>
      <td>34.5%</td>
      <td>366</td>
      <td>42.5%</td>
      <td>492</td>
      <td>28.5%</td>
      <td>163.9</td>
      <td>165.8</td>
      <td>66.7%</td>
    </tr>
  </tbody>
</table>

<ul>
  <li>The 2026 admission-season sample was the largest in the 2020-2026 period.</li>
  <li>The overall acceptance rate remained in the mid-30% range in each of the two most recent seasons, below the previous high of 45.9% in 2020 and 2023.</li>
  <li>The nationality gap in the 2026 admission season was <strong>14.1%p</strong> (American 42.5% - International 28.5%).</li>
</ul>

<h2 id="what-stood-out">What Stood Out</h2>

<h3 id="1-american--international">1) American &gt; International</h3>

<ul>
  <li>There is year-to-year variation, but in most years the American acceptance rate is higher than the International acceptance rate.</li>
  <li>The exception is 2022 (International 48.8% &gt; American 40.4%), and in 2026 the gap widened again.</li>
</ul>

<h3 id="2-the-timing-of-the-season-still-centers-on-early-to-mid-february">2) The Timing of the Season Still Centers on Early to Mid-February</h3>

<ul>
  <li>Except for 2020, <code class="language-plaintext highlighter-rouge">Median Final</code> is generally concentrated in early to mid-February.</li>
  <li>In 2026, <code class="language-plaintext highlighter-rouge">Median Final</code> is <strong>02/10</strong>, which is close to the usual pattern.</li>
</ul>

<h3 id="3-interpretation-of-subfields-still-needs-to-be-conservative">3) Interpretation of Subfields Still Needs to Be Conservative</h3>

<ul>
  <li>Even in 2026, <code class="language-plaintext highlighter-rouge">Subfield Known %</code> is only <strong>26.2%</strong>, so subfield comparisons are best treated as directional rather than definitive.</li>
  <li>Within the reported subset, however, the CP tag appears most often.</li>
</ul>

<h3 id="4-the-linear-correlation-between-gregpa-and-admission-is-very-weak">4) The Linear Correlation Between GRE/GPA and Admission Is Very Weak</h3>

<ul>
  <li>The score-admission correlations for the combined 2020-2026 data are as follows.</li>
</ul>

<table>
  <thead>
    <tr>
      <th>Metric</th>
      <th>Valid N</th>
      <th>Correlation r</th>
      <th>p-value</th>
      <th>Interpretation</th>
    </tr>
  </thead>
  <tbody>
    <tr>
      <td>GPA</td>
      <td>1,616</td>
      <td>0.057</td>
      <td>0.0214</td>
      <td>Very weak linear correlation</td>
    </tr>
    <tr>
      <td>GRE V</td>
      <td>884</td>
      <td>0.066</td>
      <td>0.0484</td>
      <td>Very weak linear correlation</td>
    </tr>
    <tr>
      <td>GRE Q</td>
      <td>881</td>
      <td>0.117</td>
      <td>0.0005</td>
      <td>Very weak linear correlation</td>
    </tr>
    <tr>
      <td>GRE AW</td>
      <td>812</td>
      <td>0.005</td>
      <td>0.8952</td>
      <td>No meaningful linear correlation</td>
    </tr>
  </tbody>
</table>

<ul>
  <li>Even where statistical significance appears, the effect size itself is small. In practice, factors such as research fit, recommendation letters, the writing sample, and school-level fit are likely to generate more meaningful variation.</li>
</ul>

<h2 id="from-an-applicants-point-of-view">From an Applicant’s Point of View</h2>

<h3 id="1-announcements-mostly-cluster-in-mid-february">1) Announcements Mostly Cluster in Mid-February</h3>

<ul>
  <li>This is the period when many major schools’ accept/reject results are concentrated.</li>
</ul>

<h3 id="2-the-nationality-subfield-and-gre-tables-are-best-read-as-mood-indicators">2) The Nationality, Subfield, and GRE Tables Are Best Read as Mood Indicators</h3>

<ul>
  <li>
    <p>There is far too much missing data to use them as a model for predicting any individual’s outcome.</p>
  </li>
  <li>
    <p>Separately, according to GradCafe forum this cycle, Duke admitted only three students and Georgetown only six, which makes this admissions season appear especially brutal.</p>
  </li>
</ul>

<h3 id="3-a-large-share-of-the-outcome-cannot-be-explained-by-quantitative-specs-alone">3) A Large Share of the Outcome Cannot Be Explained by Quantitative Specs Alone</h3>

<ul>
  <li>The weak correlation between GRE/GPA and admission suggests that many forces besides quantitative factors are shaping outcomes, and in textbook terms it is more realistic to build one’s strategy around school-specific fit.</li>
  <li>At the same time, as I watched GradCafe admission posts from the 2026 cycle in real time, it became hard to ignore the fact that applicants with profiles such as a 330+ GRE, a 5.0+ AW score, and an Ivy League undergraduate degree were often choosing among top programs. In that sense, it still seemed clear that conventional metrics do matter after all.</li>
</ul>]]></content><author><name>Yerim Oh</name></author><category term="Admission" /><category term="Academia" /><summary type="html"><![CDATA[Introduction]]></summary></entry><entry><title type="html">Can AI Do Social Science Research?</title><link href="https://ye-rim-oh.github.io/2026-03-01-AutoResearch.md/" rel="alternate" type="text/html" title="Can AI Do Social Science Research?" /><published>2026-03-01T00:00:00+09:00</published><updated>2026-03-12T00:00:00+09:00</updated><id>https://ye-rim-oh.github.io/2026-03-01-AutoResearch.md</id><content type="html" xml:base="https://ye-rim-oh.github.io/2026-03-01-AutoResearch.md/"><![CDATA[<h3 id="changes-in-us-academia">Changes in U.S. Academia?</h3>

<p>As Claude Code became hot this year, it seems that even in U.S. academia there is finally a visible trend of faculty starting to take the usefulness of AI seriously. Professor Pedro H. C. Sant’Anna at Emory University also shared his own workflow(<a href="https://psantanna.com/claude-code-my-workflow/">Sant’Anna</a>), and Professor Chris Blattman at the University of Chicago Harris School has also continued to share examples of automation(<a href="https://x.com/cblatts/status/2027018464670491065?s=20">Blattman</a>). Professor Chris even retweeted a meme like <a href="https://x.com/cblatts/status/2027134883898757458?s=20">i just talk to chatgpt</a>, so it also seems that he consumes AI discourse quite actively.</p>

<h3 id="where-my-interest-started">Where My Interest Started</h3>

<p>I became interested in research automation using AI because, while preparing materials for 2026 fall admissions, I was writing things like a Writing Sample and SOP in LaTeX and using VS Code. At the time, various AI CLIs were only just coming in, and I remember it as the point when discussions like “how are we supposed to use this?” were only beginning. I was not confident in English grammar, so I had corrected things over and over with GPT or Grammarly, but in the end it was too tiring that a person still had to polish everything manually all the way through every time. But if you just type <code class="language-plaintext highlighter-rouge">gemini</code> in the terminal and say, “Please fix the grammar in this file,” it reads a <code class="language-plaintext highlighter-rouge">.tex</code> file hundreds of lines long and corrects the grammar on its own. Watching that, I began to think that it was probably only a matter of time not just before grammar correction, but before AI could write a paper from machine-readable data and literature files alone, as long as the researcher set up the logic properly.</p>

<h3 id="public-precedents">Public Precedents</h3>

<p>In fact, it is not as if there have been absolutely no precedents of papers written with AI getting accepted. Joshua Gans discussed, in a case-based way, what kind of effects AI might have on research production based on his own idea(<a href="https://joshuagans.substack.com/p/what-will-ai-do-to-presearch">Gans</a>), and Vincent Gregoire at HEC Montreal also posted a “Vibe Research” write-up on his own website(<a href="https://vincent.codes.finance/posts/vibe-research-paper/">Gregoire</a>). And honestly, I think there are probably quite a few papers even now that got accepted without saying so. There is almost no benefit to disclosing it.</p>

<h3 id="the-center-of-the-debate-hall--westwood">The Center of the Debate: Hall &amp; Westwood</h3>

<p>Anyway, in the current situation where faculty are also getting better and better at using AI, I wanted to leave a record because the debate around literature review especially stood out to me. It is hard to identify the exact starting point, but by my own sense, the center of the debate is Professor Andy Hall at Stanford and Professor Sean Westwood at Dartmouth.</p>

<ul>
  <li>Professor Andy Hall has long argued that the operating logic of academia needs to change fundamentally, while sharing messages along the lines that Claude Code made research “fully reproducible”(<a href="https://x.com/ahall_research/status/2007603340939800664?s=20">Here’s proof that Claude Code can write an entire empirical polisci paper</a>, <a href="https://x.com/ahall_research/status/2025282331997798585?s=20">I’m baffled by any empirical social scientist who isn’t paying attention to these trends and isn’t changing their practices accordingly</a>). He also does fairly aggressive experiments in automating social science research, such as making AI do p-hacking(<a href="https://x.com/ahall_research/status/2024544040784720365?s=20">AI is about to write thousands of papers. Will it p-hack them?</a>), so I feel he is close to being a frontrunner.</li>
  <li>If Professor Andy Hall is on the side of demonstrating “how AI is changing research,” I think Professor Sean Westwood is closer to the person who set the debate on fire. On February 23, 2026, in the post’s timeline, he posted the provocative tweet that “the academic paper is a dead format”(<a href="https://x.com/seanjwestwood/status/2025711352921112651?s=20">original post</a>), and of course the backlash was large. The rough point was that AI will do literature reviews better, AI will already do peer review too, and users will only skim AI summaries, so real science is the research question, pre-analysis plan, and analysis, while a 30-page paper is only packaging.</li>
</ul>

<h3 id="reactions">Reactions</h3>

<ul>
  <li>Professor Melissa Perreault criticized it strongly in a tone roughly like, “the future that person wants is lazy researchers who cannot communicate, have no critical thinking, and ten times more garbage papers”(<a href="https://x.com/Dr_Perreault/status/2026268605453681115?s=20">Perreault</a>).</li>
  <li>Dr. Rex Douglass pushed back in the direction that “when you listen to scholars saying what LLMs cannot do, in the end it is a skill issue. A weak model, without guidance, might fail to satisfy a strange request in one shot, but most of the workflow can still be automated even now”(<a href="https://x.com/RexDouglass/status/2025812734332580212?s=20">Douglass</a>).</li>
  <li>Professor Craig Gallagher left a skeptical reaction along the lines of, “LLMs easily throw away principles they had five minutes ago if you complain, and you want me to believe that such a model will do peer review?”(<a href="https://x.com/CraigGaIIagher/status/2025947575627465002?s=20">Gallagher</a>).</li>
  <li>Political theorist Matthew Cole sharply remarked that this is basically a way for quantitative social scientists to confess that they are producing trash they themselves do not want to write and do not want to read(<a href="https://x.com/mattbencole/status/2026357969114337785?s=20">Cole</a>).</li>
  <li>Professor Itai Sher at the University of Massachusetts Amherst continues to maintain a skeptical view toward “research done by AI”(<a href="https://x.com/itaisher/status/2028628055300710772?s=20">Sher1</a>, <a href="https://x.com/itaisher/status/2022622746929447421?s=20">Sher2</a>, <a href="https://x.com/itaisher/status/2022128868480733554?s=20">Sher3</a>).</li>
</ul>

<h3 id="my-thoughts">My Thoughts</h3>

<ul>
  <li>Professor Sean put it strongly, but as Dr. Rex said, if there is appropriate data and the user has the ability to filter it, then if books and paper PDFs are organized into a form machines can read, I think a significant part of the literature review section can indeed be automated. Data analysis is also in fact a specialty area for AI, so overall I think I would summarize my position as leaning toward agreement that we need to consider other ways in which papers might be replaced.</li>
  <li>As for my own view on the claim some people make that all AI-assisted research is low quality: even someone as mediocre as me can now produce something that looks like a plausible publication if the data are there, so of course there may well be more mediocre working papers and outputs. But many people do not seem to consider at all the case where a professor or researcher who writes well uses AI in a more detailed, field-specific way, for example from summarizing prior research to “vibe” coding for quantitative research, and raises efficiency in a careful way. As was also mentioned in Professor Alexander Kustov’s blog, there are already plenty of cases where people just cite things roughly after looking only at the abstract.</li>
</ul>

<h3 id="follow-up-discussion">Follow-up Discussion</h3>

<p>Professor Sean later posted follow-up tweets too, perhaps especially because he was conscious of the reaction on Bluesky(<a href="https://x.com/seanjwestwood/status/2026076268735742443?s=20">follow-up 1</a>, <a href="https://x.com/seanjwestwood/status/2026852781676528015?s=20">follow-up 2</a>). What kind of place is Bluesky, anyway?</p>

<h4 id="is-ai-already-better-at-research-than-many-professors-added-260304">Is AI Already Better at Research Than Many Professors? (added 26/03/04)</h4>

<ul>
  <li>Professor Alexander Kustov of the Keough School of Global Affairs at the University of Notre Dame also posted the provocative tweet, “<a href="https://x.com/akoustov/status/2028623202306428986?s=20">AI can already do social science research better than most professors with PhDs</a>.” In his <a href="https://alexanderkustov.substack.com/p/academics-need-to-wake-up-on-ai">blog</a>, he says that “much of the opposition to AI is status protection dressed up as principle,” and he specifically points to scholars on Bluesky as denying what is happening.
    <ul>
      <li>As an aside, the author said that the blog article itself was written 100% with Claude, but according to Pangram, which had been promoted on Twitter as especially good at detecting AI-written text, it apparently counted as <a href="https://x.com/akoustov/status/2028877336007840048?s=20">“Fully Human Written”</a>.</li>
      <li>In the end, as Professor Kyle Saunders says, there is also the aspect that AI is not ruining education so much as forcing universities to confront how much they have relied on fragile proxy indicators of thinking(<a href="https://kylesaunders.substack.com/p/why-higher-educations-ai-backlash?r=i13e">Saunders</a>).</li>
    </ul>
  </li>
  <li>
    <p>Professor Scott Cunningham at Baylor University also posted a piece discussing the realistic future of journal publishing(<a href="https://causalinf.substack.com/p/claude-code-27-research-and-publishing/comments">Research and Publishing Are Now Two Different Things</a>). In the Korean case, faculty hiring also tends to rely even more heavily on quantitative indicators than in the United States, so this is not somebody else’s problem for us either.</p>
  </li>
  <li>In the end, a commentary also appeared from the Brookings Institution dealing with this state of affairs in academia(<a href="https://www.brookings.edu/articles/the-train-has-left-the-station-agentic-ai-and-the-future-of-social-science-research/">the train has left the station</a>).</li>
</ul>

<h4 id="follow-up-to-the-follow-up-added-260305">Follow-up to the Follow-up (added 26/03/05)</h4>

<ul>
  <li>Professor Alexander Kustov’s <a href="https://alexanderkustov.substack.com/p/academics-need-to-wake-up-on-ai-part">follow-up post</a></li>
  <li>Occidental College professor Igor Logvinenko’s Twitter article, <a href="https://x.com/igorlogvinenko/status/2029344322097955060">Why Academia Can’t Think Clearly About AI</a></li>
  <li>There are in fact many more opinions besides these, but it seems difficult for me, just as an individual student, to follow up on all of them.</li>
</ul>

<h3 id="discussion-in-korean-academia">Discussion in Korean Academia?</h3>

<p>In Korean academia, it may be useful to refer especially to Professor <a href="https://x.com/desemboltura">Yoo In-tae</a>, who mainly works in digital humanities, and Professor <a href="https://x.com/mahler83">Ahn Sang-jin</a>, who have raised these issues relatively early and quite often.</p>]]></content><author><name>Yerim Oh</name></author><category term="Academia" /><category term="AI" /><summary type="html"><![CDATA[Changes in U.S. Academia?]]></summary></entry></feed>