Deploying and developing cutting-edge methodological techniques at the intersection of field research and computational social science to address complex problems of scientific and policy importance. Located @ Rice University.
You can download the latest version of our CMAP* Visualization Toolkit here or access an online version on google collab.
* CMAP = Cultural Mapping and Pattern Analysis
⚠️ Repository Status: This GitHub repository is currently under development. Documentation, code, and features are actively being updated/added and may be incomplete or subject to change.
The Computational Ethnography Lab (CEL) is an interdisciplinary group developing and applying computational tools to analyze qualitative data. While affiliated scholars pursue diverse topics, our work concentrates on social inequality, health, institutions, networks, and culture. Our core commitments include building community, practicing methodological transparency, and advancing social science knowledge for broader public benefit.
Computational ethnography is a methodological approach that combines in-depth analysis of text data produced in research—such as ethnographic field notes, interview transcripts, historical documents, or web posts—with tools for pattern analysis and visualization. This approach maintains a focus on situating meaning and behavior in real-world contexts, while integrating tools from computational social science (CSS). Natural language processing (NLP), machine learning (ML), and various types of data visualization help represent patterns, test emerging typologies, facilitate comparison, and analyze individual and team-based datasets alongside in-depth human reading.
A core principle is connecting levels of analysis—much like alternating between digital maps and street views—to gain a broader vantage that may be obscured by either the scope (quantitative) or granularity (qualitative) of a single mode of inquiry. In this sense, computational ethnography extends the classical sociological strategy of using multiple methods, data types, and forms of analysis to better understand topics of importance (Du Bois 1899, 1900; Small 2011).
Computational ethnography extends rather than replaces human reasoning by linking aggregate patterns with contextual details, and by enabling the analysis of larger volumes of text than would be feasible without computation. The goal is not to remove, but to add to, the contextual depth that distinguishes ethnographic work, while enhancing the capacity to answer new questions or revisit old ones in new ways (c.f. Abramson et al. 2018, 2025). This goal overlaps with emergent scholarship at the intersection of ‘scaling up’ field methods and connecting “big data” to in-depth content to better understand complex issues in fields like health, culture, stratification, and social networks (Abramson and Dohan 2015; Breiger 2015; Hanson and Theis 2024; Nelson 2020; DiMaggio 2015).
Corey M. Abramson, Ph.D. Associate Professor of Sociology, Rice University
Co-Director, Center for Computational Insights on Inequality and Society (CIISR)
Sociology: stratification and inequality, health and medicine, culture, aging, networks, methodology (qualitative, computational and mixed methods), theory
Health policy: aging, alzheimer’s disease and related dementias, social determinants of health, decision making, end of life, methodology (mixed-methods, big “qual” data)
"Sunday": A photo from Professor Abramson’s fieldwork.
Books: The End Game: How Inequality Shapes Our Final Years (Harvard University Press)
Beyond the Case (Oxford University Press, with Neil Gong)
Research , Books , Bio/CV , Personal Webpage
You can read our most recent review piece here here.
A heatmap visualizing patterned variation in interviews
A semantic network illustrating the relationships between concepts in field notes and interviews.
A t-SNE plot.
A word-cloud, with user defined concept groups represented as color coding.
Data translation tool
Papers
Blogs and Interactive Notebooks:
Video Tutorials:
See above:
No, not in our opinion. Because:
Ethnography has long encompassed varied research paradigms—including conventional science-oriented approaches to field observation, case-based alternatives that aim for complementarity, individual and team-based research, and critical as well as humanist traditions. We even published a book on how varied research paradigms– such as the extended case method, behavioralism, phenomenologies, postmodernism, and grounded theory– approach the ethnographic craft, including the use of computation: Beyond the Case (Oxford, 2020).
Fields including sociology have long combined multiple research methods, often with a goal of triangulation or connecting different levels of analysis. Long before modern computing or the term mixed-methods gained popularity, scholars including Du Bois used fieldwork, history, data visualization, and surveys in parallel to generate insights. In the modern era, Anthropologist Roy D’Andrade introduced heat maps to visualize linguistic domains and cognitive schema in the 1970s, ethnographers have employed social network plots to chart cliques, and scholars of culture (e.g. Bearman, Cerulo, Mohr) have been combining rich historical information with formal and computational analyses since shortly after the entry of modern computing.
The use of computation was already widespread when I began my career as an undergraduate at Berkeley’s Center for Ethnographic Research. This is seen not only in computer-assisted qualitative data analysis software (CAQDAS- which I taught for decades), statistical computing, and the current proliferation of AI, but in myriad digital tools used in methods assumed to be analog (like ethnography). The contours of algorithmic/ordinal society shape our work in ways both overt and subtle. Grammar checkers and auto-complete suggestions nudge our use of language, indices of books and articles in libraries shape what we read (and how works get cited), sorting algorithms order our web searches, targeted marketing and news feeds shape our perceptions and mood during downtime, even before chatbots, automations and deepfakes. This makes thoughtful use of qualitative computing, and perhaps social scientific inquiry using in-depth methods to understand life in situ, more important than ever. The challenge of navigating technology is not new however, and has some inspiring parallels in how thinkers repurposed technologies like filing cabinets to facilitate rigorous and imaginative social inquiry in previous eras.
Yes. Absolutely.
The tools were assembled based on interest and feedback over the past five years, and build on prior work in teaching and developing qualitative methodology. Many of the visualizations (heatmaps, semantic networks, wordclouds, dimensionality reduction) have been used across diverse research paradigms and use cases—academic research and community organizing, quantitative social science and computational text analysis, studies of temporality and health policy, and scientific as well as humanistic applications.
Feel free to use and adapt. If you develop something new and useful, please consider sharing. If you have suggestions, that’s great too. Regardless, we hope you find something of value.
License information is located in the toolkit repositories. This is free software, intended to encourage collaboration and broader non-commercial use with attribution—ideally to address shared challenges rather than create barriers.
For any inquiries or errors, please email me at corey.abramson@rice.edu
You can read a recent review piece about underlying principles of combining computation and qualitative analysis here and see more examples (over time). Check out this page over time for updates.
You can also send an email if you have thoughts, have an example to share, or want to contribute/collaborate/comment.