Where AI meets human systems
Katherine Elkins' research bridges computational methods with humanistic and social-scientific inquiry, investigating how information flows shape human behavior, social outcomes, and institutional decision-making across political, cultural, and technological systems.
AI Safety & LLM Evaluation
Elkins investigates fundamental vulnerabilities in how large language models process language — particularly negation, prohibition, and persuasion. Her research with Jon Chun has revealed that many open-source models endorse prohibited actions 77% of the time when instructions contain syntactic negation, a finding with significant implications for AI safety in high-stakes deployment contexts. This work contributes to evaluation frameworks as part of the NIST AI Safety Institute Consortium, where Elkins and Chun serve as Principal Investigators representing the 25,000-member Modern Language Association.
Related work includes multi-agent behavioral simulation of high-stakes decisions — benchmarking over 90 model/reasoning combinations for judicial recidivism prediction through the Notre Dame-IBM Tech Ethics Lab — and research on how emotional framing can manipulate LLM decision-making, effectively "hacking" model outputs through affective rather than logical channels. This line of research demonstrates why humanities expertise in language, rhetoric, and narrative is essential to AI safety evaluation, not merely supplementary.
Computational Social Science
A central thread of Elkins' research applies computational methods to social-scientific questions — how people make decisions, how institutions allocate resources, and how information systems shape political and economic outcomes. With Jon Chun, she has developed multi-agent behavioral simulations that benchmark how well generative AI can predict human behavior in high-stakes contexts, including judicial recidivism prediction (testing 90+ model/reasoning combinations through the Notre Dame-IBM Tech Ethics Lab) and economic decision-making under uncertainty. This work sits at the intersection of behavioral economics, political science, and AI, using language models as instruments for studying human judgment rather than replacing it.
Elkins and Chun have mentored over 300 student research projects since 2016 that apply computational methods to social science questions across virtually every department at Kenyon College — political discourse analysis, economic modeling, public health messaging, criminal justice outcomes, social media dynamics, cultural analytics, and behavioral prediction. This body of student work, downloaded from institutions in 198 countries, demonstrates that computational social science is most productive when researchers bring domain expertise from the social sciences and humanities rather than treating computation as a discipline-free tool. Their course "Programming Humanity" trains students to build and critically evaluate AI systems that model social phenomena, producing original research that has been cited in peer-reviewed journals.
Computational Humanities & SentimentArcs
Katherine Elkins developed SentimentArcs, the first computational methodology for analyzing emotional arc across full-length literary narratives. Published in The Shapes of Stories: Sentiment Analysis for Narrative (Cambridge University Press, 2022), the method uses a large ensemble of NLP models — from simple lexicon-based approaches to state-of-the-art transformers — with human-in-the-loop validation to surface narrative structure in works ranging from Woolf to Kafka to Proust. Novel self-supervised metrics jointly optimize across every possible corpus-model combination, solving the generalization problem that limits traditional supervised sentiment analysis.
The SentimentArcs methodology has been adopted globally. Student and faculty research applying SentimentArcs and related computational tools has been downloaded over 95,000 times from more than 4,000 institutions across 198 countries via the Digital Kenyon repository. Student research projects using SentimentArcs have been replicated across hundreds of applications spanning literature, social media analysis, political discourse, financial narratives, medical narratives, legal documents, and more. The open-source framework is available on GitHub, and collaborative extensions include MultiSentimentArcs for multimodal (dialogue and image) analysis of film and a greybox XAI ensemble integrating GPT models with interpretable white-box classifiers.
Archival Intelligence
Elkins is Principal Investigator for a Schmidt Sciences Humanities and AI Virtual Institute (HAVI) grant — one of 23 teams selected worldwide from a highly competitive international pool — building AI tools to rescue endangered cultural archives in New Orleans. The $330,000, 18-month project, titled "Archival Intelligence," uses machine learning to process, transcribe, and make accessible historical documents from communities whose records are at risk of permanent loss due to environmental degradation, institutional neglect, and resource constraints.
Central to the project's methodology is community-governed data sovereignty: the communities whose heritage is being preserved maintain control over how their archives are accessed, represented, and used. This approach reflects Elkins' broader conviction that AI systems for cultural heritage must be designed in partnership with the people whose histories they preserve, not imposed from outside. The project is conducted in collaboration with New Orleans-based cultural institutions and represents a model for ethical AI application in the humanities. Coverage has appeared on NPR/WOSU (February 2026).
AI Governance & Comparative Regulation
Elkins conducts comparative analysis of AI regulation across the EU, China, and the United States, examining risk frameworks, enforcement approaches, and the tension between innovation and safety. Her co-authored policy paper with International Public AI, "If Open Source Is to Win, It Must Go Public," analyzes near- to mid-term risks and opportunities in open-source generative AI, arguing that meaningful AI governance requires public participation and transparency in AI development.
Her governance research extends to behavioral auditing of LLM decision-making systems, where she and Chun developed one of the earliest ethics-based audit methodologies for probing normative values in commercial and open-source language models including GPT-4. Findings revealed troubling authoritarian tendencies and cultural biases in normative frameworks across multiple models — work that informed subsequent approaches to AI alignment evaluation. The code for this research is publicly available at github.com/KatherineElkins.
Translation, Narrative & Affective AI
Elkins' research on literary translation uses large language models to assess how well translations preserve the reading experience from one language to another — a fundamentally humanistic question that computational methods can illuminate at scale. A current project focuses on multiple translations of Proust's A la recherche du temps perdu, using emotional equivalence metrics to evaluate what is gained and lost when prose moves between French and English. This work builds on her expertise as editor of Philosophical Approaches to Proust's In Search of Lost Time (Oxford University Press, 2022) and her longstanding scholarly interest in consciousness, memory, and embodied aesthetic experience.
Related research examines hyperpersuasion in language models — evaluating how persuasive AI-generated text can be relative to human-authored prose — and the emotional architecture of narrative across media, from literature and film to political speech and social media. Her Audible.com lecture series on "The Giants of French Literature" and "The Modern Novel" has reached an international audience, reflecting her commitment to making humanistic scholarship accessible beyond the academy.
Human-Centered AI Education
In 2016, Elkins and Jon Chun co-created what they describe as the world's first human-centered AI curriculum at Kenyon College, integrating computational methods with ethics, governance, and humanistic inquiry within the Integrated Program in Humane Studies (IPHS). The curriculum is distinctive for attracting over 90% non-STEM majors, with 61% women, 13% Black, and 11% Latinx students — demographics that stand in stark contrast to traditional computer science programs. Students conduct original research projects applying AI to questions drawn from their own disciplines, producing work that has been downloaded from institutions in 198 countries.
Elkins and Chun have mentored over 300 student research projects since 2016 across virtually every department at Kenyon College, spanning computational social science, digital humanities, AI safety, political discourse analysis, economic modeling, cultural analytics, and behavioral prediction. The curriculum is described in their paper "The Crisis of Artificial Intelligence: A New Digital Humanities Curriculum" and is now celebrating its roots in IPHS's 50th anniversary at aristotletoai.com. Elkins has presented on this educational model at OpenAI's Higher Education Forum, UNESCO, and dozens of universities.