<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (NISO Z39.96-2019) Journal Publishing DTD v1.2 20190208//EN" "https://jats.nlm.nih.gov/publishing/1.2/JATS-journalpublishing1-mathml3.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="1.2" xml:lang="en" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
  <front>
    <journal-meta>
  <journal-id journal-id-type="publisher-id">CCR</journal-id>
  <journal-title-group>
    <journal-title>Computational Communication Research</journal-title>
  </journal-title-group>
  <issn pub-type="ppub" />
  <issn pub-type="epub">2665-9085</issn>
  <publisher>
    <publisher-name>Amsterdam University Press</publisher-name>
    <publisher-loc>Amsterdam</publisher-loc>
  </publisher>
</journal-meta><article-meta>
      <article-id pub-id-type="publisher-id">CCR2025.1.13.LIND</article-id><article-id pub-id-type="doi">10.5117/CCR2025.1.13.LIND</article-id><article-categories><subj-group subj-group-type="heading"><subject>Article</subject></subj-group></article-categories><title-group>
        <article-title>Grounding the Comparative Turn in Communications: A Framework for Validating Multilingual Computational Text Analysis</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <name>
            <surname>Lind</surname>
            <given-names>Fabienne</given-names>
          </name>
          <aff>Department of Communication Science, University of Vienna, Austria</aff>
        </contrib>
        <contrib contrib-type="author">
          <name>
            <surname>Schoonvelde</surname>
            <given-names>Martijn</given-names>
          </name>
          <aff>Faculty of Arts, University of Groningen, The Netherlands</aff>
        </contrib>
        <contrib contrib-type="author">
          <name>
            <surname>Baden</surname>
            <given-names>Christian</given-names>
          </name>
          <aff>Department of Communication and Journalism, The Hebrew University of Jerusalem, Israel</aff>
        </contrib>
        <contrib contrib-type="author">
          <name>
            <surname>Dolinsky</surname>
            <given-names>Alona O.</given-names>
          </name>
          <aff>Department of Communication Science, <italic>Vrije Universiteit</italic> Amsterdam, The Netherlands</aff>
        </contrib>
        <contrib contrib-type="author">
          <name>
            <surname>Pipal</surname>
            <given-names>Christian</given-names>
          </name>
          <aff>Department of Communication and Media Research, University of Zurich, Switzerland</aff>
        </contrib>
        <contrib contrib-type="author">
          <name>
            <surname>Velden</surname>
            <given-names>Mariken A.C.G. van der</given-names>
          </name>
          <aff>Department of Communication Science, <italic>Vrije Universiteit</italic> Amsterdam, The Netherlands</aff>
        </contrib>
      </contrib-group>
      <pub-date pub-type="epub"><year>2025</year></pub-date><volume>7</volume><issue>1</issue><fpage>1</fpage><permissions><copyright-statement>© The authors</copyright-statement><copyright-year>2025</copyright-year><copyright-holder>The authors</copyright-holder><license license-type="open-access"><license-p>This is an open access article distributed under the CC BY 4.0 license <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link></license-p></license></permissions><abstract>
    <title>Abstract</title><p>Following the progressing internationalisation of social science research and the computational turn in the field, researchers are increasingly adopting computational text analysis (CTA) methods to compare textual data across multiple cases and languages. In these settings, it is not only the mapping between construct and measures that requires validation, but also the equivalence of this mapping across languages and cases. However, although the validation requirements in multilingual analyses exceed those in monolingual studies, current research shows that validation is often insufficiently and inconsistently addressed in comparative multilingual CTA. To support more robust comparative research, this article presents a framework for validating findings obtained from multilingual textual data. The framework outlines validation strategies for four key stages of a typical multilingual CTA workflow: corpus, input data, process, and output. It directly tackles the challenge of approaching equivalence across contexts and languages in these stages and moves beyond the common practice of identifying problems only at the final stage of research.</p>
  </abstract>
      <kwd-group>
        <title>Keywords:</title><kwd>cross-lingual</kwd>
        <kwd>comparative research</kwd>
        <kwd>text as data</kwd>
        <kwd>computational text analysis</kwd>
        <kwd>validation framework</kwd>
        <kwd>internationalisation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="S1">
      
      
      
      
    <title>Introduction</title><p id="S1.p1">Over the past decade, the internationalisation of social science research has finally been picking up speed (e.g., Henriksen, <xref rid="bib.bibx20" ref-type="bibr">2016</xref>; Scharkow &amp; Trepte, <xref rid="bib.bibx48" ref-type="bibr">2023</xref>). As scholars engage more globally, there is growing recognition of the need to confront ongoing power imbalances in the field (Demeter, <xref rid="bib.bibx14" ref-type="bibr">2019</xref>), prompting efforts to broaden the scope of research beyond the dominance of WEIRD (Western, Educated, Industrialised, Rich, and Democratic) countries (Henrich et al., <xref rid="bib.bibx19" ref-type="bibr">2010</xref>). While comparative survey research has embraced this global shift, textual research has lagged, partly due to the cultural and linguistic complexities embedded in texts, which makes comparisons across cases<xref rid="id1" ref-type="fn" specific-use="fn"><sup>1</sup></xref> more complex (e.g., Rössler, <xref rid="bib.bibx46" ref-type="bibr">2012</xref>). Only in recent years have rapid advances in computational methods increasingly allowed researchers to analyse textual data not only at scale, but also in a multilingual and comparative manner.</p><p id="S1.p2">Multilingual computational text analysis (CTA) is an emerging and fast-evolving field. Scholars apply it to study climate change coverage across Germany, India, South Africa, and the United States (Wozniak et al., <xref rid="bib.bibx60" ref-type="bibr">2021</xref>), foreign policy reporting in over 100 countries (Baum &amp; Zhukov, <xref rid="bib.bibx6" ref-type="bibr">2019</xref>), or the salience of the EU in national legislatures (Rauh &amp; Wilde, <xref rid="bib.bibx42" ref-type="bibr">2018</xref>). Methodological innovations have made such studies increasingly feasible. These range from multilingual dictionaries (e.g., Baden et al., <xref rid="bib.bibx5" ref-type="bibr">2018</xref>; Proksch et al., <xref rid="bib.bibx40" ref-type="bibr">2019</xref>), multilingual supervised machine learning (e.g., Lind et al., <xref rid="bib.bibx30" ref-type="bibr">2021</xref>), multilingual topic modeling (e.g., Chan et al., <xref rid="bib.bibx10" ref-type="bibr">2020</xref>; Lind et al., <xref rid="bib.bibx29" ref-type="bibr">2022</xref>; Lucas et al., <xref rid="bib.bibx31" ref-type="bibr">2015</xref>; Reber, <xref rid="bib.bibx43" ref-type="bibr">2019</xref>), multilingual representation models based on transformers (e.g., Vries, <xref rid="bib.bibx13" ref-type="bibr">2021</xref>; Licht, <xref rid="bib.bibx27" ref-type="bibr">2023</xref>; Laurer, <xref rid="bib.bibx26" ref-type="bibr">2023</xref>), and, most recently, to prompt-based multilingual generative models (e.g., Rathje et al., <xref rid="bib.bibx41" ref-type="bibr">2024</xref>).</p><p id="S1.p3">Nevertheless, despite these developments, a persistent challenge remains: there is little consensus on how to validate multilingual CTA for comparative research. Most validation discussions remain rooted in single-language contexts and provide limited guidance for cross-lingual applications (e.g., Atteveldt &amp; Peng, <xref rid="bib.bibx57" ref-type="bibr">2018</xref>; Birkenmaier et al., <xref rid="bib.bibx9" ref-type="bibr">2024</xref>; Grimmer et al., <xref rid="bib.bibx17" ref-type="bibr">2022</xref>; Song et al., <xref rid="bib.bibx52" ref-type="bibr">2020</xref>). Even in single-language contexts, recent reviews find that validation is inconsistently applied and often underreported; for instance, only 54% of CTA articles in top communication journals include any form of validation (Stecker et al., <xref rid="bib.bibx54" ref-type="bibr">2024</xref>). As Birkenmaier et al. (<xref rid="bib.bibx9" ref-type="bibr">2024</xref>) note, the lack of coherent standards and field-specific guidance for CTA has led to fragmented and ad hoc validation practices, threatening the credibility and comparability of findings.</p><p id="S1.p4">In multilingual settings, the need for rigorous validation is especially pressing. Recent research shows strong demand for improved validation practices in these contexts. A survey of authors of published CTA studies (Dolinsky et al., <xref rid="bib.bibx15" ref-type="bibr">2024</xref>) found that those working with multiple languages expressed significantly more concern about the validity of their findings than those focusing on English or a single other language. However, this demand has not translated into widespread implementation of validation strategies. A review of published social science literature (Baden et al., <xref rid="bib.bibx4" ref-type="bibr">2022</xref>)<xref rid="id2" ref-type="fn" specific-use="fn"><sup>2</sup></xref> shows that validation in multilingual CTA is neither sufficiently nor consistently addressed.</p><p id="S1.p5">In this article, we address this gap by proposing a framework for validating multilingual measurement in CTA for comparative research. We argue that to ensure comparability of textual measures across languages and cases, equivalence must be validated at multiple stages of the research process. Our framework identifies four key stages for validation: corpus, input data, process, and output. These stages reflect a typical CTA workflow, offering a structured and actionable approach to validation. The chronological structure helps guide researchers through validation from data collection to final measurement.</p><p id="S1.p6">Crucially, each stage raises distinct challenges for establishing comparability. For example, without a corpus appropriate to the research objective, later validation steps may lose their meaning. Similarly, problems uncovered during output validation often stem from earlier stages, such as biased input selection or inadequate processing. A key motivation for our four-stage framework is that the need for validation far exceeds demonstrating performance relative to a given benchmark of the final measures, and that corpus, input data, and process constitute key considerations with far-reaching implications for measurement validity.</p><p id="S1.p7">This perspective also highlights how multilingual comparative text analysis introduces challenges that monolingual, single-context studies largely avoid. Whereas a researcher working in one language ‘only’ needs to ensure that the measures obtained meaningfully map onto their construct of interest, researchers working in multiple languages and with numerous cases need to provide evidence that this link between the conceptual and empirical realms is equivalent across languages and cases. Only when such equivalence is established can substantial comparisons of cases be deemed meaningful.</p><p id="S1.p8">This framework is especially timely, as researchers increasingly rely on pre-trained large language models (LLMs), which often lack transparency regarding their training data or language coverage, and tend to perform unevenly across languages due to resource imbalances and structural biases (Bender et al., <xref rid="bib.bibx7" ref-type="bibr">2021</xref>; Mate et al., <xref rid="bib.bibx33" ref-type="bibr">2023</xref>). Without systematic validation, such tools risk introducing hidden biases or misinterpreting cross-lingual patterns as substantive findings.</p><p id="S1.p9">To illustrate how the framework can be applied in practice, we use a running example drawn from the computational detection of incivility. Specifically, we build on the premise of replicating Muddiman and Stroud’s (<xref rid="bib.bibx35" ref-type="bibr">2017</xref>) classic study on incivility in online news comment sections in a multilingual, internationally comparative setting. This example is particularly well-suited for our purposes because incivility is expressed in highly variable linguistic and cultural forms. By drawing on this example throughout, we aim to concretely illustrate how language- and context-sensitive validation can be addressed through specific considerations and evaluations at each stage of a typical multilingual CTA pipeline.</p><p id="S1.p10">While specific validation strategies will necessarily depend on the setting and purpose of each study, the framework is sufficiently general to apply to a wide range of analytical approaches, including dictionary-based models, topic modeling, and transformer-based architectures. It offers guidance to researchers seeking to draw meaningful measurements from multilingual text data in comparative research projects across diverse methodological settings. Developing widely accepted standards for validating the use of CTA methods is crucial for building trust in their application to multilingual comparative research. It represents a necessary prerequisite for their broader adoption in an increasingly global research community.</p></sec>
    <sec id="S2">
      
      
      
      
    <title>Scope conditions, key terms, and goals</title><p id="S2.p1">Our validation framework is designed for research that aims for measures that can be validly compared across languages and cases, and thus for comparative research designs (Volk, <xref rid="bib.bibx58" ref-type="bibr">2022</xref>). As such, it supports the rapidly expanding community of researchers who study text beyond the monolingual English default (Baden et al., <xref rid="bib.bibx3" ref-type="bibr">2022</xref>). To ensure comparability, the framework follows an etic approach at the level of construct definition, starting from a universal definition of the concept, assumed to involve the same theoretical dimensions across contexts (Esser &amp; Vliegenthart, <xref rid="bib.bibx16" ref-type="bibr">2017</xref>). While it is possible to compare constructs defined emically (i.e., in context-specific ways), such comparisons are typically qualitative or interpretive. By focusing on etic definitions, our framework ensures that cross-national and cross-linguistic comparisons can be made using shared, conceptually equivalent criteria (Wirth &amp; Kolb, <xref rid="bib.bibx59" ref-type="bibr">2004</xref>). For example, a cross-national extension of Muddiman and Stroud’s (<xref rid="bib.bibx35" ref-type="bibr">2017</xref>) study might adopt their modular definition of incivility, comprising name-calling, profanity, racism, and threats, as a shared baseline. This reflects an etic strategy: maintaining conceptual equivalence across settings. The next step would be to determine how to measure this construct in each of the cases and languages under study.</p><p id="S2.p2">We take a broad view of measurement validity, referring to the extent to which a given CTA method accurately captures the intended meanings in a text (Grimmer et al., <xref rid="bib.bibx17" ref-type="bibr">2022</xref>). In multi-case, multi-language research, achieving validity also requires careful attention to equivalence and comparability. When analysing an etic concept across different cases and languages, it is essential that the measures consistently reflect the same underlying construct. This alignment, referred to as “equivalent mapping”, is a central goal in comparative research. Following (Volk, <xref rid="bib.bibx58" ref-type="bibr">2022</xref>, p. 80), we define equivalence as the condition in which “two objects, structures, or categories have an equal value or the same function”.</p><p id="S2.p3">Ensuring equivalence does not mean enforcing a uniform measurement strategy across all contexts. For example, in a comparative study of uncivil discourse, researchers must account for how certain expressions function differently across contexts. A term like “gay”, which may be neutral or reclaimed in some settings, can still carry a strongly derogatory connotation in others. Similarly, culturally specific slurs such as the German “grünlinksversifft” (“green-left-filthy”) have no direct equivalent in the U.S. political discourse, illustrating how seemingly analogous categories can diverge sharply in language and meaning. The key is that measures must reflect meaningful case differences relevant to the research question. At the same time, they must avoid distortions caused by unnoticed inconsistencies in measurement. A robust validation process helps distinguish true cross-case differences from those arising due to methodological artifacts (He &amp; Vijver, <xref rid="bib.bibx18" ref-type="bibr">2012</xref>). Both validation goals—equivalent mapping and measure comparability—are summarised in Figure 1.</p><fig id="S2.F1"><label>Figure 1:</label><caption><title>Measurement and validity objectives for comparative research designs. Note. This simplified illustration assumes only two cases, each of which is linked to exactly one language; actual cases may be considerably more complex.</title></caption>
        
        
        
        
      <graphic xlink:href="Figure 1.png" /></fig><p id="S2.p4">Two linguistic considerations are crucial for establishing equivalence in multilingual comparative designs: semantic comparability, which concerns the literal meaning of language and the relationship between signs and their referents, and pragmatic comparability, which relates to the social meaning of language and how it is interpreted within specific contexts (Hovy &amp; Prabhumoye, <xref rid="bib.bibx23" ref-type="bibr">2021</xref>; Licht &amp; Lind, <xref rid="bib.bibx28" ref-type="bibr">2023</xref>).</p><p id="S2.p5">A key challenge for semantic comparability concerns the extraction of comparable meanings from text data despite significant variation in how they are encoded in each language, through dimensions like script (e.g., Latin, Cyrillic, Hebrew), morphology (word formation), and syntax (see Shababo et al., <xref rid="bib.bibx50" ref-type="bibr">2023</xref>, for sentence structure). In essence, expressions in different languages do not align, impeding direct numerical comparison. What is more, comparable meanings are often expressed differently across languages, and the associated connotations or conceptual boundaries can vary as well. For example, a single word in one language may correspond to multiple distinct concepts in another (e.g., consider Hebrew: 
    An error in the conversion from LaTeX to XML has occurred here.
  ’wyb h‘m, German: “Volksfeind”, but English: “enemy of the people”). Moreover, many words lack straightforward translations altogether (Sigismondi, <xref rid="bib.bibx51" ref-type="bibr">2018</xref>). In addition, it is possible that some information that is consequential for an intended analysis, such as information on the grammatical case, gender, or tense of expressed contents, is available in some languages, but not in others. As a result, equivalent measurement strategies may need to operate with very different depths of information, potentially distorting results.</p><p id="S2.p6">The second consideration concerns pragmatic comparability. From a social science perspective, language can rarely be studied without taking into account the context in which language is produced (see also Hovy &amp; Prabhumoye, <xref rid="bib.bibx23" ref-type="bibr">2021</xref>). In comparative research designs, it is important to consider the divergent meanings and connotations of semantically similar expressions, owing to the influence of historical, political, and social traditions on language use. For example, the label “socialism” may evoke a legacy of authoritarianism in parts of Eastern Europe. At the same time, in conservative U.S. political discourse, it often functions as a generalized slur against progressive policies. For a valid comparative analysis, it is thus crucial to ensure that measured contents are valid not only in terms of their semantic meaning, but also in terms of how equivalent meanings are commonly expressed in different contexts.</p></sec>
    <sec id="S3">
      
      
      
      
    <title>A framework for validating multilingual textual analysis</title><p id="S3.p1">Having clarified the challenges of semantic and pragmatic equivalence, we now turn to the framework itself. It provides a structured approach for validating multilingual CTA, with a particular focus on approaching equivalence across languages and contexts. The framework is designed to guide researchers through key decision points in typical CTA workflows, offering validation strategies.</p><p id="S3.p2">Validation strategies refer to the methods researchers use to ensure that recorded scores accurately reflect the intended concepts (Adcock &amp; Collier, <xref rid="bib.bibx1" ref-type="bibr">2001</xref>). These strategies may involve reflecting on methodological choices and their alignment with theoretical constructs, as well as empirically testing how well the chosen procedures perform.</p><p id="S3.p3">The framework (summarised in Figure 2) organizes these strategies across four stages that align with a typical CTA workflow: (1) corpus, (2) input data, (3) process, and (4) output.</p><fig id="S3.F2"><label>Figure 2:</label><caption><title>A validation framework for multilingual comparative research designs.</title></caption>
        
        
        
        
      <p><graphic xlink:href="Figure_2.png" /></p></fig><p id="S3.p4">The corpus stage concerns the valid selection of equivalent document sources and samples across multiple cases and languages. The input data stage addresses how these documents should be unitized and quantitatively represented for subsequent text analysis procedures. The process stage involves the application of models, while the output stage focuses on evaluating the resulting measures. These stages follow a largely sequential structure, where each step builds upon the outcomes of the previous one. Problems not identified early can compound in later stages, so each stage should be approached carefully and deliberately to ensure that only validated results progress.</p><p id="S3.p5">The corpus stage holds a foundational role in the validation framework because errors introduced here are difficult and costly to correct downstream. Document sources may not be easily recollected or reproduced, and platform APIs or repositories often change over time. That said, thorough validation at any stage can never mitigate or replace attention at another stage, as fatal validity and comparability issues may arise independently at every stage. Hence, there is no inherent hierarchy. Each stage is essential, though issues at earlier stages are typically more difficult to remedy later. Problems should be addressed or identified where they happen.</p><p id="S3.p6">Alternative organizing principles for validation exist, such as psychometric categories (e.g., face, content, criterion, discriminant validity; see Adcock &amp; Collier, <xref rid="bib.bibx1" ref-type="bibr">2001</xref>; Cronbach &amp; Meehl, <xref rid="bib.bibx12" ref-type="bibr">1955</xref>) or linguistic levels (e.g., morphological, syntactic, semantic, pragmatic; see  O’Grady et al., <xref rid="bib.bibx37" ref-type="bibr">2017</xref>). Our workflow-based structure integrates these perspectives where they most directly affect research decisions: for example, criterion and face validity during output validation; content validity during input data and process stages; and linguistic challenges as well as pragmatic challenges across the pipeline stages. This chronological structure guides researchers through validation from data collection to final measurement, highlighting key equivalence challenges at each step.</p><p id="S3.p7">The framework is informed by a systematic review of 854 quantitative text analysis studies published between January 2016 and September 2020 in the top 20 Web of Science journals in communication, political science, sociology, and psychology (Baden et al., <xref rid="bib.bibx4" ref-type="bibr">2022</xref>). We manually coded these studies for reported validation strategies. Only about 45% justified the comparability of document samples across languages and cases. While preprocessing steps were frequently documented, only 5% discuss or show their suitability for achieving comparable multilingual representations. Just 10% reported validation of modeling choices, and only 30% compared outputs with convergent or divergent measures. These findings underscore the lack of standardized validation practices and motivate our contribution.</p><p id="S3.p8">In what follows, we present the four stages of the framework in greater detail, highlighting key objectives, considerations, and evaluation techniques for each.</p><sec id="S3.SS1">
        
        
        
        
      <title>Corpus Validation</title><sec id="S3.SS1.SSS1">
          
          
          
          
        <title>Objective</title><p id="S3.SS1.SSS1.p1">In all comparative research, the validity of findings depends on the equivalence of the materials being compared (Palicki et al., <xref rid="bib.bibx38" ref-type="bibr">2023</xref>; Rinke et al., <xref rid="bib.bibx44" ref-type="bibr">2022</xref>). In a text as data project, this means that a researcher must first select equivalent document sources. While not feasible for all research projects, given that some studies rely on “found data” (Salganik, <xref rid="bib.bibx47" ref-type="bibr">2017</xref>, pp. 82-83) with unknown generative processes, materials need to be sufficiently equivalent in relation to the research question at hand.<xref rid="id3" ref-type="fn" specific-use="fn"><sup>3</sup></xref> This guarantees that detected differences are not mere byproducts of subtle differences in the selection of documents. During the process of selecting document sources it is vital to ensure the equivalence of the source sampling procedure across cases, and make note of any remaining discrepancies (e.g., when specific corpora, for which no suitable equivalent exists in other cases, are included for substantive reasons) so that they can be addressed later in the analysis.</p><p id="S3.SS1.SSS1.p2">When equivalent document sources have been identified, in many cases, corpus creation furthermore requires the selection of equivalent document samples. This can involve narrowing inclusion criteria to specific relevant document populations (usually, using keywords) or creating stratified samples that capture key dimensions of variability in the population under study. Whenever the cases under study also involve different languages, researchers need to additionally ensure that their retrieval strategies yield equivalent document samples (e.g., translated keywords often find or miss different subsets of relevant debates, owing to divergent connotations and language-specific expressions).<xref rid="id4" ref-type="fn" specific-use="fn"><sup>4</sup></xref> To achieve this goal, researchers need to reflect on potential sources of sampling bias and empirically evaluate whether sample equivalence has been achieved.</p></sec><sec id="S3.SS1.SSS2">
          
          
          
          
        <title>Considerations</title><p id="S3.SS1.SSS2.p1">Suppose we wanted to replicate Muddiman and Stroud’s (<xref rid="bib.bibx35" ref-type="bibr">2017</xref>) study of news user commentary in the New York Times (NYT) across multiple languages. In such a scenario, a first consideration concerns resource bias and retrieval bias, as recommended by (Grimmer et al., <xref rid="bib.bibx17" ref-type="bibr">2022</xref>).</p><p id="S3.SS1.SSS2.p2">Resource bias arises when the availability of documents differs systematically across languages and cases. For their study, Muddiman &amp; Stroud (<xref rid="bib.bibx35" ref-type="bibr">2017</xref>) were able to access all comments posted in response to NYT coverage since the inauguration of its commentary feature, including also comments that were subsequently moderated or deleted. Besides the obvious challenges of determining which media outlet maps well onto the NYT, other media outlets may have opened commentary sections later, selectively, or failed to keep long-term records, and very few, if any, will grant comparable access to moderated content. In some countries, media primarily rely on social media platforms to enable user commentary, subjecting the posting, moderation, and archiving of content to external platform policies and norms. In a similar vein, analyses of X (formerly Twitter) content need to account for the platform’s very different roles and adoption rates across contexts, considering whether comparable political discourse might occur on various social networks instead. Media outlets, platforms, and other public forums can differ between countries concerning their policies for moderating or surpressing uncivil content, adding systematic distortions.</p><p id="S3.SS1.SSS2.p3">Retrieval bias occurs when the sampling strategy for identifying relevant documents works differently between cases in ways that introduce bias into subsequent measurements. For example, a search string for identifying swearwords may have high precision and recall in one language, but perform considerably worse when translated into another language (e.g., Russian, profanity is often rooted in sexual language; in German, it tends to use fecal metaphors; and in the U.S., profanity is frequently abbreviated [“bs”] or blanked [“f***”]).<xref rid="id5" ref-type="fn" specific-use="fn"><sup>5</sup></xref> In most cases, it cannot be assumed that identical sampling strategies, applied across cases and languages, will reliably generate equivalent samples.</p></sec><sec id="S3.SS1.SSS3">
          
          
          
          
        <title>Evaluation</title><p id="S3.SS1.SSS3.p1">When evaluating equivalence among language-specific documents, reliance on domain expertise or involvement of case experts becomes essential (Palicki et al., <xref rid="bib.bibx38" ref-type="bibr">2023</xref>; Rinke et al., <xref rid="bib.bibx44" ref-type="bibr">2022</xref>). Often, collecting equivalent samples across languages requires comparative reasoning. For instance, in cases where text sources are media outlets, a comparison of market shares within each case can guide equivalent source selection (Rössler, <xref rid="bib.bibx46" ref-type="bibr">2012</xref>). Suppose document samples are retrieved with search strings in different languages for different cases. In that case, these retrieved documents should be compared with human-annotated data to assess recall and precision, and misclassifications should be inspected (Mahl et al., <xref rid="bib.bibx32" ref-type="bibr">2022</xref>). A good example of such careful practice is provided by Rauh &amp; Wilde (<xref rid="bib.bibx42" ref-type="bibr">2018</xref>), who offer a transparent account of how they evaluated and ensured cross-lingual equivalence in their selection of EU legislative debates.</p></sec></sec><sec id="S3.SS2">
        
        
        
        
      <title>Input Data Validation</title><sec id="S3.SS2.SSS1">
          
          
          
          
        <title>Objective</title><p id="S3.SS2.SSS1.p1">After selecting the corpus for each language, researchers need to ensure that the produced units and textual representations are equivalent across cases. In most comparative research, this may require actions such as harmonising unitization or stripping uninformative, idiosyncratic content across cases. When dealing with multilingual text, linguistic perspectives also come into play. Factors like how languages segment sentences into words, compound word formation, function word affixation, and morphological variations demand attention. Researchers are encouraged to reflect on and empirically test the equivalence of the data inputs across languages and cases. Importantly, the input data should be selected with the intended analytical model in mind, as different approaches (e.g., dictionary methods, topic models, or transformers) may place different demands on the structure and quality of input data.</p></sec><sec id="S3.SS2.SSS2">
          
          
          
          
        <title>Considerations</title><p id="S3.SS2.SSS2.p1">Data input validation first concerns unitization, i.e., the determination of what level of text relevant meanings are expressed and to what units of text they pertain. For multilingual analyses, this is especially important as differences in the meaning contained within each unit of text may confound equivalent measurement. Cultural, stylistic, and linguistic differences affect unitization, such as in the case of languages that prefer long (e.g., German) or short (e.g., English) sentences, or in Chinese speech, which does not segment words in writing, resulting in substantively different meanings of paragraphs. To ensure comparability, researchers may need to adapt the unit of analysis per language—for instance, combining several English sentences to match a dense German one, or using word segmentation in Chinese. Justifying these choices with linguistic and contextual insight is key to valid cross-language analysis.</p><p id="S3.SS2.SSS2.p2">In CTA, unitization does not end here, however; rather, texts need to be further broken down into so-called “tokens” - typically derived from words - whose distribution over different documents presents the primary source of information for many computational algorithms. However, tokenization is far from language-neutral. When deciding on the appropriate set of tokenisation steps, researchers need to consider in what form relevant meanings are encoded in the text. To ensure that equivalent information is retained in multiple languages and cases, it may be required to use different preprocessing strategies for different languages and cases so that the obtained textual representations are comparable and adequate for the intended analysis. For instance, many languages create numerous distinct tokens for the same word, depending on its case, gender, etc., which appear as independent from one another to a computer unless harmonised. In English, this may be addressed by stemming (as used by Muddiman &amp; Stroud, <xref rid="bib.bibx35" ref-type="bibr">2017</xref>). However, in languages that do not inflect words solely at the end, more sophisticated forms of lemmatization are required. For example, Hebrew has pre-, in-, and suffixes: “
    An error in the conversion from LaTeX to XML has occurred here.
  lSn’w”, “
    An error in the conversion from LaTeX to XML has occurred here.
  Sn’w”, “
    An error in the conversion from LaTeX to XML has occurred here.
  Swn’”, “
    An error in the conversion from LaTeX to XML has occurred here.
  tSn’” all reduce to the same root meaning “hate,” requiring more advanced processing to treat them as equivalent.</p><p id="S3.SS2.SSS2.p3">Similarly, information expressed by conjunctions, prepositions, and other tokens typically discarded as “stopwords” in English (e.g., Muddiman &amp; Stroud, <xref rid="bib.bibx35" ref-type="bibr">2017</xref>) is often conveyed through morphological variation, word order, or other mechanisms in other languages. For example, the English phrase “son of a bitch” risks losing key relational meaning if “of” is removed as a stopword. In contrast, the equivalent Arabic expression “
    An error in the conversion from LaTeX to XML has occurred here.
  ibn al-‘AhiraT¿” retains its structure and meaning through morphological encoding, illustrating how English-centric preprocessing can inadvertently distort multilingual semantic comparisons. These linguistic properties may confound a comparative study of uncivil language in unforeseen ways.</p><p id="S3.SS2.SSS2.p4">Additionally, in certain languages, incivility might manifest through the use of explicit swear words, while in others, it might be expressed through specific multi-word phrases. In Dutch, the intensifying use of disease-related terms such as <italic>kankerlijer</italic> (cancer sufferer) or <italic>tyfushond</italic> (typhus dog) creates compound insults that carry strong emotional weight, a pattern less common in English. Furthermore, the target of uncivil language is more easily observable in languages that hardwire grammatical function in particular word inflections (e.g., Turkish encodes the accusative case). Similarly, when one is interested in the extent to which uncivil language is gendered, it helps to consider how each of the languages under study distinguishes genders (if at all).</p><p id="S3.SS2.SSS2.p5">To determine input data validity, some extent of linguistic familiarity – and to the extent that pragmatic equivalence is required, cultural familiarity – is necessary. Researchers are generally advised to consult with native speakers and case experts when preprocessing unfamiliar languages and carefully validate outputs from standard pre-processing tools (such as those in NLTK, Bird et al., <xref rid="bib.bibx8" ref-type="bibr">2009</xref>), which may not generalize across languages. Tasks such as token reduction<xref rid="id6" ref-type="fn" specific-use="fn"><sup>6</sup></xref>. and word sense disambiguation<xref rid="id7" ref-type="fn" specific-use="fn"><sup>7</sup></xref> must be adapted with care, and multilingual preprocessing should be approached not as a standardized pipeline, but as a language-sensitive design challenge.</p></sec><sec id="S3.SS2.SSS3">
          
          
          
          
        <title>Evaluation</title><p id="S3.SS2.SSS3.p1">Ideally, researchers can offer evidence that input data alignment is established. If an analysis relies on preprocessing, a way to test the validity of the input data is to check if the meaning that is studied can still be easily recognised from the preprocessed text. At the same time, all variations that are lost should be immaterial to the pursued research question. One strategy that should help with detecting validity issues and biases in multilingual analysis at the level of preprocessing is to translate a few texts before preprocessing them and subject them to the respective preprocessing stages. The key test of cross-lingual validity is that the representations obtained after preprocessing map well onto one another, such that how relevant information is expressed (e.g., by single tokens, by token collocation patterns, or by sequentially arranged tokens) is similar between languages and cases. In studies where translation is part of the preprocessing stage, such as multilingual CTA projects that bring corpora into a shared representational space (e.g., English language text), back-translation is an effective validation technique. By iteratively applying machine translation tools, this method assesses the semantic stability of translated texts by testing whether repeated cycles of translation yield consistent and interpretable results (Chew et al., <xref rid="bib.bibx11" ref-type="bibr">2025</xref>). As one example of a study that presents extensive input validation steps, Segev (<xref rid="bib.bibx49" ref-type="bibr">2019</xref>) discusses and argues for the selection of certain word types and the exclusion of others to increase the cross-linguistic and cross-national validity of the measurement instrument.</p></sec></sec><sec id="S3.SS3">
        
        
        
        
      <title>Process Validation</title><sec id="S3.SS3.SSS1">
          
          
          
          
        <title>Objective</title><p id="S3.SS3.SSS1.p1">In addition to striving for equivalent textual representations, the researcher must also ensure the equivalence of the intended modeling strategy across languages and cases. In any use of CTA, the primary consideration governing validity is whether the algorithm indeed recognises the meaning intended to be measured. In multilingual projects, researchers must confirm that the algorithms are equally effective in all languages and cases. In practical terms, validation efforts concern model development, model selection, and model calibration.</p></sec><sec id="S3.SS3.SSS2">
          
          
          
          
        <title>Considerations</title><p id="S3.SS3.SSS2.p1">For their recognition of incivility, Muddiman &amp; Stroud (<xref rid="bib.bibx35" ref-type="bibr">2017</xref>) relied on a carefully validated dictionary of one-word tokens. Transferring the same design to different languages, however, numerous sources of bias arise from linguistic and context-related factors that may compromise the effectiveness of the same modeling strategy. A deeper understanding of the algorithmic design is a prerequisite for understanding how modeling assumptions and linguistic features interact across languages.</p><p id="S3.SS3.SSS2.p2">Most rule-based algorithms (e.g., dictionaries or search strings) offer considerable researcher control and thus generally facilitate reflection upon modeling choices across languages. Enabled to recognize that terms included in Muddiman and Stroud’s (<xref rid="bib.bibx35" ref-type="bibr">2017</xref>) incivility dictionary are polysemic, context-dependent, or require multi-word expressions in other languages, researchers can relatively easily add disambiguation criteria or shift toward a multi-word enabled strategy, while adjusting keywords for case-specific contexts as needed. For example, in Quebecois French, liturgical terms like <italic>tabarnak</italic> and <italic>câlice</italic> function as strong profanities and would likely be included in a dictionary of incivility for that context. In contrast, the same terms would appear semantically benign in continental French and thus be excluded. This contrast underscores how regional cultural variation affects which expressions register as uncivil and must be accounted for in multilingual modeling.</p><p id="S3.SS3.SSS2.p3">Matters are more complicated for unsupervised, supervised algorithms, or more recent textual representations such as embeddings, where association measures are typically hard-coded into the algorithm. Researchers must assess whether document-level word co-occurrences (common in classic machine learning) suffice as representations of the expressed meaning, or whether more restrictive measures need to be obtained (e.g., by designing specific features that focus estimation on particular patterns in the data). Moreover, unsupervised algorithms tend to make strict assumptions about the generative structure of a text. For instance, many clustering algorithms assume that every token belongs to exactly one cluster, while factor analytic or topic modeling algorithms permit that the same token may contribute to more than one cluster. Especially in languages with many homonyms (i.e., words that have the same spelling) or words that appear as part of different multi-word expressions, some clustering solutions are inappropriate for certain analyses, as the same token cannot be constrained to one role exactly <xref rid="id8" ref-type="fn" specific-use="fn"><sup>8</sup></xref>. Languages that form compound words (that is, two or more words joined together to create a single, distinct word with a unified meaning) may pose issues as key terms in the analysis are too infrequent to constitute recognised patterns. For instance, if we were to search for patterns in the topical structure of uncivil comments, the algorithm has much better chances to recognize <italic>welfare</italic> and <italic>cheat</italic> (both words common also in other contexts) as a pattern than <italic>Sozialbetrüger</italic> in German, a relatively rare term unique to the topic). Especially for supervised algorithms, providing substantive justifications for the choice of specific algorithms or hyperparameters often presents a major challenge, as very little is known how exactly modelling language in a hierarchical fashion (e.g., decision tree algorithms), in a vector space (e.g., SVM), or a neural network might affect classification. While it is evident (and plausible) that such choices matter, virtually no theory is available for justifying specific decisions.</p><p id="S3.SS3.SSS2.p4">Such considerations are not only relevant for models that rely on word co-occurrences (such as bag-of-words models), but also for models that represent words as embeddings. For example, word embeddings represent each word as a vector of real numbers that is based on the premise that words that appear a lot near each other in a set of documents should have similar vector representations (for an overview, see  Rodriguez &amp; Spirling, <xref rid="bib.bibx45" ref-type="bibr">2022</xref>). While early embeddings were primarily developed for English and faced challenges in other languages due to data scarcity and linguistic differences, recent advances have produced embeddings that cover dozens of languages (e.g., Polyglot) as well as multilingual embedding systems (Licht, <xref rid="bib.bibx27" ref-type="bibr">2023</xref>; Vries, <xref rid="bib.bibx13" ref-type="bibr">2021</xref>), with promising cross-lingual alignment (Chew et al., <xref rid="bib.bibx11" ref-type="bibr">2025</xref>). Nevertheless, theoretical concerns persist regarding how grammatical word order and syntactic structures—such as the Subject-Verb-Object order in English versus Subject-Object-Verb or other variations in other languages—may influence the semantic quality and comparability of embeddings. Understanding these effects remains an important open question, especially when embeddings are used for measurement across languages.</p><p id="S3.SS3.SSS2.p5">Returning to our ongoing example of detecting uncivil language, different languages may have different swearing cultures, which will have implications for either approach. For instance, in some languages, name-calling relies on well-known slurs or offensive adjectives that frequently co-occur and are easy to detect. In others, incivility may take more subtle or culturally specific forms. In Dutch, for instance, calling someone a <italic>pannenkoek</italic> (pancake) is mildly offensive. It is roughly akin to calling someone an idiot, even though the literal meaning is benign. In some online subcultures, such as incel communities, incivility involves the creative coining of novel terms (<italic>neologisms</italic>) or the reappropriation of innocuous words (e.g., <italic>foid</italic>, <italic>roastie</italic>, or <italic>skype</italic> used derogatorily), which can be difficult to detect without cultural or contextual knowledge. In East Asian languages like Korean, supervised machine learning may, to a significant extent, simply rely on honorific forms to decide that no incivility can be present. In contrast, in most Western languages, such a form is unavailable or degraded, and thus unavailable for classification. When measuring uncivil language, such cultural differences should be explicitly acknowledged in the codebook to be used for annotating training documents for multilingual supervised classification or in the interpretation of topics in a multilingual topic model.</p></sec><sec id="S3.SS3.SSS3">
          
          
          
          
        <title>Evaluation</title><p id="S3.SS3.SSS3.p1">The easiest way to ascertain process validity is to subject a few texts to the algorithm’s transformation and determine whether the represented information (e.g., associations) reflects meaningful features of the processed text, in the case of multilingual analysis, whether it reflects the same meaningful features in equivalent ways. In general, for multilingual analysis, it is desirable to determine that equivalent modeling choices indeed emphasise and discount equivalent information. If researchers create their own set of manually annotated training documents that inform the model, testing intercoder-reliability across languages and cases for these training materials serves as another process validation strategy. Calculating metrics like Machine Translation Accuracy (MTA) can serve as an effective validation strategy for comparing the multilingual outputs of various models. For an implementation that assesses the consistency of top words per topic across languages via MTA, refer to Lind et al. (<xref rid="bib.bibx29" ref-type="bibr">2022</xref>). Validation strategies encompass further activities like visualising how inputs relate to outputs and identifying features that are most influential in the model’s decision-making process (see, for example, the LIME technique as implemented in Ho &amp; Chan (<xref rid="bib.bibx21" ref-type="bibr">2023</xref>)). When relying on generative LLMs like ChatGPT, researchers can, for example, interrogate these models to explain their reasoning for why they annotate text in one way or another, and they can do so in multiple languages (see e.g., Kuzman et al., <xref rid="bib.bibx25" ref-type="bibr">2023</xref>). As a positive example, Maurer &amp; Diehl (<xref rid="bib.bibx34" ref-type="bibr">2020</xref>) offer important details on how they validated their French and English dictionaries against crowd-sourced benchmarks and by further employing a qualitative pre-analysis of individual keywords.</p></sec></sec><sec id="S3.SS4">
        
        
        
        
      <title>Output Validation</title><sec id="S3.SS4.SSS1">
          
          
          
          
        <title>Objective</title><p id="S3.SS4.SSS1.p1">The fourth validation stage concerns the quality of the obtained measures and their equivalence across languages and across cases concerning criterion validity (the obtained measures correlate with a trusted third variable) and face validity (the results appear plausible for each case) (Krippendorff, <xref rid="bib.bibx24" ref-type="bibr">2004</xref>). For comparative multilingual applications, output validation needs to be considered for each included language and case.</p></sec><sec id="S3.SS4.SSS2">
          
          
          
          
        <title>Considerations</title><p id="S3.SS4.SSS2.p1">Our discussion on output validation focuses on the type and quality of a third variable that can serve as a trusted benchmark for case and language comparisons. Researchers are advised to assess available third variables for either the same construct (for testing convergent validity) or distinctly different constructs (for testing divergent validity). When selecting or developing a human-coded benchmark as a third variable, as was the case in Muddiman and Stroud’s (<xref rid="bib.bibx35" ref-type="bibr">2017</xref>) study, its quality can be assessed by inspecting the quality of the codebooks, human coder training, and intercoder reliability tests. Codebook definitions should apply to all languages and cases (i.e., etic constructs), while rules and examples may need to be adjusted to different contexts and settings (e.g., racist incivility tends to focus on different ethnic groups in different countries). As with monolingual projects, it is advantageous to train all involved coders in joint sessions and to clarify issues or adjust the codebook collaboratively (Rössler, <xref rid="bib.bibx46" ref-type="bibr">2012</xref>). Intercoder reliability tests should cover reliability across languages/cases as well as within languages/cases (Peter &amp; Lauf, <xref rid="bib.bibx39" ref-type="bibr">2002</xref>). If all coders are skilled in one language, they can code the same material to establish intercoder reliability (Hopmann et al., <xref rid="bib.bibx22" ref-type="bibr">2016</xref>). Care should be taken to ensure that the manually coded material is representative of all languages and cases. To achieve this, a random subset per language and case can be translated into a common language. To further assess intercoder reliability within language/case, at least two coders code the original language material per case. If a measured meaning is far more dominant, or largely absent, in one compared case, careful interpretation of performance scores is crucial; high accuracy does not necessarily imply valid discrimination (e.g., if incivility appears in only 3% of cases, not coding it achieves 97% accuracy). Employing chance- and category frequency-corrected metrics might be necessary for informative measurement validity. For example, when examining uncivil language, do we see a higher recall or precision in some languages than others? And what features are predictive of observed differences?</p></sec><sec id="S3.SS4.SSS3">
          
          
          
          
        <title>Evaluation</title><p id="S3.SS4.SSS3.p1">Applying the convergent validation strategies involves the selection of a high-quality gold standard as described above and the calculation of recall and precision scores for each language and case in the dataset. For example, Lind et al. (<xref rid="bib.bibx30" ref-type="bibr">2021</xref>) compare the results from a supervised classification approach with a benchmark coded by native speakers. Temporão et al. (<xref rid="bib.bibx56" ref-type="bibr">2018</xref>) assess the convergent and predictive validity of their computationally obtained ideological scaling measures of social media users. Another key strategy for output validation, then, is to examine misclassifications carefully (Ho &amp; Chan, <xref rid="bib.bibx21" ref-type="bibr">2023</xref>). Systematic issues in precision are detected by inspecting positively classified but irrelevant documents. Recall issues can be examined by drawing a random sample of texts for manual classification and examining which texts are missed by the model. When a proper gold standard exists, examining false positives and negatives can uncover possible biases and direct enhanced performance of the employed tools. Last but not least, it is of course advisable to critically check the results manually with case experts for their plausibility.</p></sec></sec></sec>
    <sec id="S4">
      
      
      
      
    <title>Conclusion</title><p id="S4.p1">As the community of social scientists who study various languages at scale continues to globalize and the comparative turn in computational text analysis requires theoretical grounding, we have provided a framework for validating social science measurements across languages and cases.</p><p id="S4.p2">Following a definition of validity in the context of such projects, the framework includes practical recommendations for assessing the validity of etic constructs measured across diverse languages and cases. It covers techniques for validating equivalent data sources and sample selection, preparation of equivalent input features, selection of equivalence data processing methods, attainment of equivalent measurements, and ultimately achieving equivalent mapping and thus comparable results. Despite advancements in CTA, collaboration with experts in relevant languages and regions, as well as knowledge about training procedures, remains crucial. Issues around resource and retrieval bias have not disappeared in the age of LLMs, which, for many social science applications, need to be fine-tuned to identify a particular construct in smaller language-specific samples of data.</p><p id="S4.p3">One immediate question that arises is how to deal with measures when validation techniques raise validity concerns that cannot be resolved? A first and crucial mitigation action is to report and reflect on the detected problems, as it enables future research to build on better information on related measurements. This way, the field can accumulate knowledge, build theory, and develop further recommendable strategies. The second is to make a considered decision about the extent to which the measurements can be used to make substantive comparative statements about the cases. The measurements may not be suitable for comparative statements, but they may be suitable for statements per case or for comparisons among a subset of the cases. Furthermore, researchers can explore error correction methods to account for misclassifications (i.e., Bachl &amp; Scharkow, <xref rid="bib.bibx2" ref-type="bibr">2017</xref>; TeBlunthuis et al., <xref rid="bib.bibx55" ref-type="bibr">2023</xref>).</p><p id="S4.p4">In light of this need to attend systematically to issues of validation in CTA, questions about the practical implementation of such validation immediately arise. Executing the validation requirements outlined in the framework requires significant resources. However, since the quality of the analysis is crucially dependent on the quality of the data and measurements, these are resources well-spent. Many of these steps can be supported by better research infrastructures (e.g., the Masakhane project, which works on African languages, Nekoto et al., <xref rid="bib.bibx36" ref-type="bibr">2020</xref>, or OPTED) and open science initiatives. For instance, curated inventories for data sources and tools that support the input data and process stages can save research teams valuable time and resources. These services can also facilitate collaboration between research partners and provide access to language and case experts from diverse backgrounds, which is essential for all validation stages (see Spirling, <xref rid="bib.bibx53" ref-type="bibr">2023</xref>, for a similar argument in the context of LLMs).</p><p id="S4.p5">The framework we propose outlines an ideal standard for validation in comparative text analysis, especially for projects that aim for conceptual and measurement equivalence across languages and cases. While it sets out what best practice looks like, we acknowledge that fully adhering to this standard may not always be feasible in large-scale studies. Even if resources were to improve significantly in the coming years and a project had a very large budget, the kind of methodological rigor and emphasis on human evaluation by native speakers and case experts that are central to our framework mean that projects covering 50+ countries and languages are hardly recommendable. For projects that do aim at such scale, we recommend prioritizing validation for a strategically selected subset of cases (e.g., high-variance contexts or linguistic clusters), using these as benchmarks for broader interpretation. While such strategies cannot fully replace comprehensive validation, they can help identify systematic issues and improve the interpretability of large-scale comparisons.</p><p id="S4.p6">A limitation of the framework is that it does not provide a detailed breakdown of the input, process, and output validation stages for specific analytical methods, such as distinct preprocessing steps or particular approaches like dictionary methods, topic models, or transformer-based architectures. As the framework is intended to be broadly applicable across methods, such detail was beyond the scope of this article. Nonetheless, we hope it offers a useful structure to support more specific extensions in future work.</p><p id="S4.p7">In conclusion, we encourage future research to define constructs comprehensively; to document and justify methodological choices; to report any validation efforts taken; and to prudently consider the detected validity concerns in the subsequent use of measurements. Not only should authors dedicate some space to enabling readers to follow taken validation steps and thus build confidence in the measurement, but reviewers need to demand and scrutinise presented validation efforts, and editors need to set aside the requisite space for the purpose. Transparency not only facilitates an informed assessment of measurement validity in CTA but also lays the groundwork for systematic theoretical and methodological discussions on best practices in validation.</p></sec>
    <sec id="S5">
      
      
      
      
    <title>Funding Note</title><p id="S5.p1">This work was supported by funding from the European Union’s Horizon 2020 research and innovation programme under Grant Agreement No. 951832.</p></sec>
    <sec id="S6">
      
      
      
      
    <title>Acknowledgement</title><p id="S6.p1">We sincerely thank the reviewers for their valuable feedback and thoughtful suggestions.</p></sec>
    </body>
  <back>
    <fn-group><title>Notes</title><fn id="id1" symbol="1"><p id="footnote1">
            
            
            
            
          Cases encompass a spectrum of macro-level units, ranging from countries and regions to markets and beyond.</p></fn><fn id="id2" symbol="2"><p id="footnote2">
            
            
            
            
          It builds on a review of practises reported in the quantitative text-based social science literature that was developed as part of the text analysis infrastructure project OPTED (https://opted.eu/) spanning multiple countries and languages</p></fn><fn id="id3" symbol="3"><p id="footnote3">
                
                
                
                
              Random samples and population-level analyses are unaffected by this step.</p></fn><fn id="id4" symbol="4"><p id="footnote4">
                
                
                
                
              When researchers select documents using CTA (i.e., keyword-based methods), this is basically a validation task in itself. It is considered best practice to apply the input, process, and output stages of the validation framework to the document selection step as well.</p></fn><fn id="id5" symbol="5"><p id="footnote5">
                
                
                
                
              Biases may even arise with seemingly trivial sampling criteria - for instance, the keyword “Trump” may work reasonably well in non-US media for identifying coverage of the former US president, but
may generate precision problems in US media (which may also cover other members of the family to some extent), and it will completely miss relevant transcripts of “The Late Show”, since its host Colbert systematically replaces that name with slanderous nicknames.
</p></fn><fn id="id6" symbol="6"><p id="footnote6">
                
                
                
                
              For example, synthetic languages (that express grammar by means of affixes or morphology) contain significantly fewer highly common words than English, such that discounting the most common terms mostly eliminates articles, prepositions, conjunctions etc. in English, but reaches well into common nouns used for construct formation in other languages (e.g., 
    An error in the conversion from LaTeX to XML has occurred here.
  byt [house] in Hebrew, which appears as part of very many construct state nouns: 
    An error in the conversion from LaTeX to XML has occurred here.
  byt spr [school], 
    An error in the conversion from LaTeX to XML has occurred here.
  byt mSp.t [court], 
    An error in the conversion from LaTeX to XML has occurred here.
  byt knst [synagogue])</p></fn><fn id="id7" symbol="7"><p id="footnote7">
                
                
                
                
              For example, disambiguation may be worth considering for important homonyms, where the same spelling is used to express different meanings (e.g., ‘rock’, which can be a mass of stone, a musical genre, a movement, and more; some languages contain many more homonyms than others, so especially for Semitic languages, Part Of Speech tagging may be necessary to efficiently distinguish the many possible meanings of identical character sequences).</p></fn><fn id="id8" symbol="8"><p id="footnote8">
                
                
                
                
              For example, the English phrase <italic>“parliamentary inquiry committee chairman”</italic> consists of four tokens, each of which also co-occurs in many other contexts. By contrast, the equivalent German word <italic>“Untersuchungsausschussvorsitzender”</italic> is a single, very rare token that does not co-occur with others in the same way.</p></fn></fn-group><ref-list><title>References</title>
      <ref id="bib.bibx1"><mixed-citation publication-type="journal"><string-name><surname>Adcock</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name><surname>Collier</surname>, <given-names>D.</given-names></string-name> (<year>2001</year>). <article-title>Measurement validity: A shared standard for qualitative and quantitative research</article-title>. <source><italic>American Political Science Review</italic></source>, <volume>95</volume>(<issue>3</issue>), <fpage>529</fpage>–<lpage>546</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1017/S0003055401003100">https://doi.org/10.1017/S0003055401003100</ext-link></mixed-citation></ref>
      <ref id="bib.bibx2"><mixed-citation publication-type="journal"><string-name><surname>Bachl</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Scharkow</surname>, <given-names>M.</given-names></string-name> (<year>2017</year>). <article-title>Correcting measurement error in content analysis</article-title>. <source><italic>Communication Methods and Measures</italic></source>, <volume>11</volume>(<issue>2</issue>), <fpage>87</fpage>–<lpage>104</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/19312458.2017.1305137">https://doi.org/10.1080/19312458.2017.1305137</ext-link></mixed-citation></ref>
      <ref id="bib.bibx3"><mixed-citation publication-type="journal"><string-name><surname>Baden</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Pipal</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Schoonvelde</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>van der Velden</surname>, <given-names>M. A. C. G.</given-names></string-name> (<year>2022</year>). <article-title>Three gaps in computational text analysis methods for social sciences: A research agenda</article-title>. <source><italic>Communication Methods and Measures</italic></source>, <volume>16</volume>(<issue>1</issue>), <fpage>1</fpage>–<lpage>18</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/19312458.2021.1916027">https://doi.org/10.1080/19312458.2021.1916027</ext-link></mixed-citation></ref>
      <ref id="bib.bibx4"><mixed-citation publication-type="journal"><string-name><surname>Baden</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Dolinsky</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Lind</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Pipal</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Schoonvelde</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Shababo</surname>, <given-names>G.</given-names></string-name>, &amp; <string-name><surname>van der Velden</surname>, <given-names>M. A.</given-names></string-name> (<year>2022</year>). <article-title>Integrated standards and context-sensitive recommendations for the validation of multilingual computational text analysis</article-title>. <source><italic>Horizon Project OPTED: Deliverable 6.2</italic></source>, </mixed-citation></ref>
      <ref id="bib.bibx5"><mixed-citation publication-type="book"><string-name><surname>Baden</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Jungblut</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Micevski</surname>, <given-names>I.</given-names></string-name>, <string-name><surname>Stalpouskaya</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Tenenboim-Weinblatt</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Berganza Conde</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name><surname>Dimitrakopoulou</surname>, <given-names>D.</given-names></string-name> (<year>2018</year>). <source><italic>The INFOCORE Dictionary: A multilingual dictionary for automatically analyzing conflict-related discourse</italic></source>. </mixed-citation></ref>
      <ref id="bib.bibx6"><mixed-citation publication-type="journal"><string-name><surname>Baum</surname>, <given-names>M. A.</given-names></string-name>, &amp; <string-name><surname>Zhukov</surname>, <given-names>Y. M.</given-names></string-name> (<year>2019</year>). <article-title>Media ownership and news coverage of international conflict</article-title>. <source><italic>Political Communication</italic></source>, <volume>36</volume>(<issue>1</issue>), <fpage>36</fpage>–<lpage>63</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/10584609.2018.1493002">https://doi.org/10.1080/10584609.2018.1493002</ext-link></mixed-citation></ref>
      <ref id="bib.bibx7"><mixed-citation publication-type="journal"><string-name><surname>Bender</surname>, <given-names>E. M.</given-names></string-name>, <string-name><surname>Gebru</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>McMillan-Major</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name><surname>Shmitchell</surname>, <given-names>S.</given-names></string-name> (<year>2021</year>). <article-title>On the dangers of stochastic parrots: Can language models be too big?</article-title>. <source><italic>Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency</italic></source>, <fpage>610</fpage>–<lpage>623</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1145/3442188.3445922">https://doi.org/10.1145/3442188.3445922</ext-link></mixed-citation></ref>
      <ref id="bib.bibx8"><mixed-citation publication-type="book"><string-name><surname>Bird</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Loper</surname>, <given-names>E.</given-names></string-name>, &amp; <string-name><surname>Klein</surname>, <given-names>E.</given-names></string-name> (<year>2009</year>). <source><italic>Natural Language Processing with Python</italic></source>. <publisher-name>O'Reilly Media Inc.</publisher-name>, <publisher-loc>Sebastopol, CA</publisher-loc>.</mixed-citation></ref>
      <ref id="bib.bibx9"><mixed-citation publication-type="journal"><string-name><surname>Birkenmaier</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Lechner</surname>, <given-names>C. M.</given-names></string-name>, &amp; <string-name><surname>Wagner</surname>, <given-names>C.</given-names></string-name> (<year>2024</year>). <article-title>The Search for Solid Ground in Text as Data: A Systematic Review of Validation Practices and Practical Recommendations for Validation</article-title>. <source><italic>Communication Methods and Measures</italic></source>, <volume>18</volume>(<issue>3</issue>),  <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/19312458.2023.2285765">https://doi.org/10.1080/19312458.2023.2285765</ext-link></mixed-citation></ref>
      <ref id="bib.bibx10"><mixed-citation publication-type="journal"><string-name><surname>Chan</surname>, <given-names>C. H.</given-names></string-name>, <string-name><surname>Zeng</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Wessler</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Jungblut</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Welbers</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Bajjalieh</surname>, <given-names>J. W.</given-names></string-name>, &amp; <string-name><surname>Althaus</surname>, <given-names>S. L.</given-names></string-name> (<year>2020</year>). <article-title>Reproducible extraction of cross-lingual topics (rectr)</article-title>. <source><italic>Communication Methods and Measures</italic></source>, <volume>14</volume>(<issue>4</issue>), <fpage>285</fpage>–<lpage>305</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/19312458.2020.1812559">https://doi.org/10.1080/19312458.2020.1812559</ext-link></mixed-citation></ref>
      <ref id="bib.bibx11"><mixed-citation publication-type="journal"><string-name><surname>Chew</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Chakraborti</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Weisman</surname>, <given-names>W.</given-names></string-name>, &amp; <string-name><surname>Frey</surname>, <given-names>S.</given-names></string-name> (<year>2025</year>). <article-title>Evaluating machine translation solutions for accessible multi-language text analysis: A back-translation based approach</article-title>. <source><italic>Computational Communication Research</italic></source>, <volume>7</volume>(<issue>1</issue>), <fpage>1</fpage>–<lpage>24</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.5117/CCR2025.1.001.CHEW">https://doi.org/10.5117/CCR2025.1.001.CHEW</ext-link></mixed-citation></ref>
      <ref id="bib.bibx12"><mixed-citation publication-type="journal"><string-name><surname>Cronbach</surname>, <given-names>L. J.</given-names></string-name>, &amp; <string-name><surname>Meehl</surname>, <given-names>P. E.</given-names></string-name> (<year>1955</year>). <article-title>Construct validity in psychological tests</article-title>. <source><italic>Psychological Bulletin</italic></source>, <volume>52</volume>(<issue>4</issue>), <fpage>281</fpage>–<lpage>302</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1037/h0040957">https://doi.org/10.1037/h0040957</ext-link></mixed-citation></ref>
      <ref id="bib.bibx13"><mixed-citation publication-type="journal"><string-name><surname>De Vries</surname>, <given-names>E.</given-names></string-name> (<year>2021</year>). <article-title>The sentiment is in the details: A language-agnostic approach to dictionary expansion and sentence-level sentiment analysis in news media</article-title>. <source><italic>Computational Communication Research</italic></source>, <volume>4</volume>(<issue>2</issue>), <fpage>424</fpage>–<lpage>462</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.5117/CCR2021.2.006.DEVR">https://doi.org/10.5117/CCR2021.2.006.DEVR</ext-link></mixed-citation></ref>
      <ref id="bib.bibx14"><mixed-citation publication-type="journal"><string-name><surname>Demeter</surname>, <given-names>M.</given-names></string-name> (<year>2019</year>). <article-title>The winner takes it all: International inequality in communication and media studies today</article-title>. <source><italic>Journalism &amp; Mass Communication Quarterly</italic></source>, <volume>96</volume>(<issue>1</issue>), <fpage>37</fpage>–<lpage>59</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1177/1077699018804506">https://doi.org/10.1177/1077699018804506</ext-link></mixed-citation></ref>
      <ref id="bib.bibx15"><mixed-citation publication-type="book"><string-name><surname>Dolinsky</surname>, <given-names>A. O.</given-names></string-name>, <string-name><surname>Schoonvelde</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Pipal</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Baden</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Lind</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Shababo</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Van der Velden</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Zalik</surname>, <given-names>A.</given-names></string-name> (<year>2024</year>). <source><italic>Challenges for multilingual computational text analysis researchers: evidence from a survey of social scientists</italic></source>.  <ext-link ext-link-type="uri" xlink:href="https://osf.io/9mybf_v2">https://osf.io/9mybf_v2</ext-link></mixed-citation></ref>
      <ref id="bib.bibx16"><mixed-citation publication-type="book"><string-name><surname>Esser</surname>, <given-names>F.</given-names></string-name>, &amp; <string-name><surname>Vliegenthart</surname>, <given-names>R.</given-names></string-name> (<year>2017</year>). <chapter-title>Comparative research methods</chapter-title>. In <source><italic>The International Encyclopedia of Communication Research Methods</italic></source> (pp. <fpage>1</fpage>–<lpage>22</lpage>). <publisher-name>Wiley</publisher-name>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1002/9781118901731.iecrm0035">https://doi.org/10.1002/9781118901731.iecrm0035</ext-link></mixed-citation></ref>
      <ref id="bib.bibx17"><mixed-citation publication-type="book"><string-name><surname>Grimmer</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Roberts</surname>, <given-names>M. E.</given-names></string-name>, &amp; <string-name><surname>Stewart</surname>, <given-names>B. M.</given-names></string-name> (<year>2022</year>). <source><italic>Text as data: A new framework for machine learning and the social sciences</italic></source>. <publisher-name>Princeton University Press</publisher-name>, <publisher-loc>Princeton, NJ</publisher-loc>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.23943/princeton/9780691207551.001.0001">https://doi.org/10.23943/princeton/9780691207551.001.0001</ext-link></mixed-citation></ref>
      <ref id="bib.bibx18"><mixed-citation publication-type="journal"><string-name><surname>He</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>van de Vijver</surname>, <given-names>F.</given-names></string-name> (<year>2012</year>). <article-title>Bias and equivalence in cross-cultural research</article-title>. <source><italic>Online Readings in Psychology and Culture</italic></source>, <volume>2</volume>(<issue>2</issue>), <fpage>2307</fpage>–<lpage>0919</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.9707/2307-0919.1111">https://doi.org/10.9707/2307-0919.1111</ext-link></mixed-citation></ref>
      <ref id="bib.bibx19"><mixed-citation publication-type="journal"><string-name><surname>Henrich</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Heine</surname>, <given-names>S. J.</given-names></string-name>, &amp; <string-name><surname>Norenzayan</surname>, <given-names>A.</given-names></string-name> (<year>2010</year>). <article-title>Most people are not WEIRD</article-title>. <source><italic>Nature</italic></source>, <volume>466</volume>(<issue>7302</issue>), <fpage>29</fpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1038/466029a">https://doi.org/10.1038/466029a</ext-link></mixed-citation></ref>
      <ref id="bib.bibx20"><mixed-citation publication-type="journal"><string-name><surname>Henriksen</surname>, <given-names>D.</given-names></string-name> (<year>2016</year>). <article-title>The rise in co-authorship in the social sciences (1980--2013)</article-title>. <source><italic>Scientometrics</italic></source>, <volume>107</volume>(<issue>2</issue>), <fpage>455</fpage>–<lpage>476</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1007/s11192-016-1849-x">https://doi.org/10.1007/s11192-016-1849-x</ext-link></mixed-citation></ref>
      <ref id="bib.bibx21"><mixed-citation publication-type="journal"><string-name><surname>Ho</surname>, <given-names>J. C. T.</given-names></string-name>, &amp; <string-name><surname>Chan</surname>, <given-names>C. H.</given-names></string-name> (<year>2023</year>). <article-title>Evaluating transferability in multilingual text analyses</article-title>. <source><italic>Computational Communication Research</italic></source>, <volume>5</volume>(<issue>2</issue>),  <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.5117/CCR2023.2.003.HO">https://doi.org/10.5117/CCR2023.2.003.HO</ext-link></mixed-citation></ref>
      <ref id="bib.bibx22"><mixed-citation publication-type="book"><string-name><surname>Hopmann</surname>, <given-names>D. N.</given-names></string-name>, <string-name><surname>Esser</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>de Vreese</surname>, <given-names>C. H.</given-names></string-name>, <string-name><surname>Aalberg</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>van Aelst</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Berganza</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name><surname>Strömbäck</surname>, <given-names>J.</given-names></string-name> (<year>2016</year>). <chapter-title>How we did it: Approach and methods</chapter-title>. In <source><italic>Comparing Political Journalism</italic></source> (pp. <fpage>10</fpage>–<lpage>21</lpage>). <publisher-name>Routledge</publisher-name>, <publisher-loc>London</publisher-loc>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.4324/9781315622286-2">https://doi.org/10.4324/9781315622286-2</ext-link></mixed-citation></ref>
      <ref id="bib.bibx23"><mixed-citation publication-type="journal"><string-name><surname>Hovy</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name><surname>Prabhumoye</surname>, <given-names>S.</given-names></string-name> (<year>2021</year>). <article-title>Five sources of bias in natural language processing</article-title>. <source><italic>Language and Linguistics Compass</italic></source>, <volume>15</volume>(<issue>8</issue>), <fpage>e12432</fpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/lnc3.12432">https://doi.org/10.1111/lnc3.12432</ext-link></mixed-citation></ref>
      <ref id="bib.bibx24"><mixed-citation publication-type="book"><string-name><surname>Krippendorff</surname>, <given-names>K.</given-names></string-name> (<year>2004</year>). <source><italic>Content Analysis: An Introduction to Its Methodology</italic></source>. <publisher-name>Sage Publications</publisher-name>, <publisher-loc>Thousand Oaks, CA</publisher-loc>.</mixed-citation></ref>
      <ref id="bib.bibx25"><mixed-citation publication-type="book"><string-name><surname>Kuzman</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Mozetič</surname>, <given-names>I.</given-names></string-name>, &amp; <string-name><surname>Ljubešić</surname>, <given-names>N.</given-names></string-name> (<year>2023</year>). <source><italic>ChatGPT: Beginning of an end of manual linguistic data annotation? Use case of automatic genre identification</italic></source>. </mixed-citation></ref>
      <ref id="bib.bibx26"><mixed-citation publication-type="journal"><string-name><surname>Laurer</surname>, <given-names>M.</given-names></string-name> (<year>2023</year>). <article-title>Lowering the language knowledge barrier -- investigating deep transfer learning and machine translation for multilingual analyses of political texts</article-title>. <source><italic>Computational Communication Research</italic></source>, <volume>5</volume>(<issue>2</issue>), <fpage>1</fpage>–<lpage>28</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.5117/CCR2023.2.001.LAUR">https://doi.org/10.5117/CCR2023.2.001.LAUR</ext-link></mixed-citation></ref>
      <ref id="bib.bibx27"><mixed-citation publication-type="journal"><string-name><surname>Licht</surname>, <given-names>H.</given-names></string-name> (<year>2023</year>). <article-title>Cross-lingual classification of political texts using multilingual sentence embeddings</article-title>. <source><italic>Political Analysis</italic></source>, <volume>31</volume>(<issue>3</issue>), <fpage>366</fpage>–<lpage>379</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1017/pan.2022.29">https://doi.org/10.1017/pan.2022.29</ext-link></mixed-citation></ref>
      <ref id="bib.bibx28"><mixed-citation publication-type="journal"><string-name><surname>Licht</surname>, <given-names>H.</given-names></string-name>, &amp; <string-name><surname>Lind</surname>, <given-names>F.</given-names></string-name> (<year>2023</year>). <article-title>Going cross-lingual: A guide to multilingual text analysis</article-title>. <source><italic>Computational Communication Research</italic></source>, <volume>5</volume>(<issue>2</issue>), <fpage>1</fpage>–<lpage>31</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/doi.org/10.5117/CCR2023.2.2.LICH">https://doi.org/doi.org/10.5117/CCR2023.2.2.LICH</ext-link></mixed-citation></ref>
      <ref id="bib.bibx29"><mixed-citation publication-type="journal"><string-name><surname>Lind</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Eberl</surname>, <given-names>JM.</given-names></string-name>, <string-name><surname>Eisele</surname>, <given-names>O.</given-names></string-name>, <string-name><surname>Heidenreich</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Galyga</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Boomgaarden</surname>, <given-names>H. G.</given-names></string-name> (<year>2022</year>). <article-title>Building the bridge: Topic modeling for comparative research</article-title>. <source><italic>Communication Methods and Measures</italic></source>, <volume>16</volume>(<issue>2</issue>), <fpage>93</fpage>–<lpage>118</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/19312458.2021.2015572">https://doi.org/10.1080/19312458.2021.2015572</ext-link></mixed-citation></ref>
      <ref id="bib.bibx30"><mixed-citation publication-type="journal"><string-name><surname>Lind</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Heidenreich</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Kralj</surname>, <given-names>C.</given-names></string-name>, &amp; <string-name><surname>Boomgaarden</surname>, <given-names>H. G.</given-names></string-name> (<year>2021</year>). <article-title>Greasing the wheels for comparative communication research: Supervised text classification for multilingual corpora</article-title>. <source><italic>Computational Communication Research</italic></source>, <volume>3</volume>(<issue>3</issue>), <fpage>1</fpage>–<lpage>30</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.5117/CCR2021.3.001.LIND">https://doi.org/10.5117/CCR2021.3.001.LIND</ext-link></mixed-citation></ref>
      <ref id="bib.bibx31"><mixed-citation publication-type="journal"><string-name><surname>Lucas</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Nielsen</surname>, <given-names>R. A.</given-names></string-name>, <string-name><surname>Roberts</surname>, <given-names>M. E.</given-names></string-name>, <string-name><surname>Stewart</surname>, <given-names>B. M.</given-names></string-name>, <string-name><surname>Storer</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name><surname>Tingley</surname>, <given-names>D.</given-names></string-name> (<year>2015</year>). <article-title>Computer-assisted text analysis for comparative politics</article-title>. <source><italic>Political Analysis</italic></source>, <volume>23</volume>(<issue>2</issue>), <fpage>254</fpage>–<lpage>277</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1093/pan/mpu019">https://doi.org/10.1093/pan/mpu019</ext-link></mixed-citation></ref>
      <ref id="bib.bibx32"><mixed-citation publication-type="journal"><string-name><surname>Mahl</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>von Nordheim</surname>, <given-names>G.</given-names></string-name>, &amp; <string-name><surname>Guenther</surname>, <given-names>L.</given-names></string-name> (<year>2022</year>). <article-title>Noise pollution: A multi-step approach to assessing the consequences of (not) validating search terms on automated content analyses</article-title>. <source><italic>Digital Journalism</italic></source>, <volume>11</volume>(<issue>2</issue>), <fpage>298</fpage>–<lpage>320</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/21670811.2021.2023931">https://doi.org/10.1080/21670811.2021.2023931</ext-link></mixed-citation></ref>
      <ref id="bib.bibx33"><mixed-citation publication-type="journal"><string-name><surname>Mate</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Sebök</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Wordliczek</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Stolicki</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name><surname>Feldmann</surname>, <given-names>Á.</given-names></string-name> (<year>2023</year>). <article-title>Machine translation as an underrated ingredient? Solving classification tasks with large language models for comparative research</article-title>. <source><italic>Computational Communication Research</italic></source>, <volume>5</volume>(<issue>2</issue>), <fpage>1</fpage>–<lpage>34</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.5117/CCR2023.2.004.MATE">https://doi.org/10.5117/CCR2023.2.004.MATE</ext-link></mixed-citation></ref>
      <ref id="bib.bibx34"><mixed-citation publication-type="journal"><string-name><surname>Maurer</surname>, <given-names>P.</given-names></string-name>, &amp; <string-name><surname>Diehl</surname>, <given-names>T.</given-names></string-name> (<year>2020</year>). <article-title>What kind of populism? Tone and targets in the Twitter discourse of French and American presidential candidates</article-title>. <source><italic>European Journal of Communication</italic></source>, <volume>35</volume>(<issue>5</issue>), <fpage>453</fpage>–<lpage>468</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1177/0267323120940903">https://doi.org/10.1177/0267323120940903</ext-link></mixed-citation></ref>
      <ref id="bib.bibx35"><mixed-citation publication-type="journal"><string-name><surname>Muddiman</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name><surname>Stroud</surname>, <given-names>N. J.</given-names></string-name> (<year>2017</year>). <article-title>News values, cognitive biases, and partisan incivility in comment sections</article-title>. <source><italic>Journal of Communication</italic></source>, <volume>67</volume>(<issue>4</issue>), <fpage>586</fpage>–<lpage>609</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/jcom.12313">https://doi.org/10.1111/jcom.12313</ext-link></mixed-citation></ref>
      <ref id="bib.bibx36"><mixed-citation publication-type="journal"><string-name><surname>Nekoto</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Marivate</surname>, <given-names>V.</given-names></string-name>, <string-name><surname>Matsila</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Fasubaa</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Fagbohungbe</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Akinola</surname>, <given-names>S. O.</given-names></string-name>, &amp; <string-name><surname>Bashir</surname>, <given-names>A.</given-names></string-name> (<year>2020</year>). <article-title>Participatory research for low-resourced machine translation: A case study in African languages</article-title>. <source><italic>Findings of the Association for Computational Linguistics: EMNLP 2020</italic></source>, <fpage>2144</fpage>–<lpage>2160</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.18653/v1/2020.findings-emnlp.195">https://doi.org/10.18653/v1/2020.findings-emnlp.195</ext-link></mixed-citation></ref>
      <ref id="bib.bibx37"><mixed-citation>
        
          
          
          
        
        
        
          <year>2017</year>
          <person-group person-group-type="author"><name><surname>al.</surname><given-names>O’Grady et </given-names></name></person-group>
          
          
        
        W. O’Grady, J. Archibald, M. Aronoff and J. Rees-Miller
        “Contemporary Linguistics: An Introduction”
        Boston, MA: Bedford/St. Martin’s, 2017
      </mixed-citation></ref>
      <ref id="bib.bibx38"><mixed-citation publication-type="journal"><string-name><surname>Palicki</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Walter</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>van Atteveldt</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Beazer</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name><surname>Bravo</surname>, <given-names>I.</given-names></string-name> (<year>2023</year>). <article-title>Selecting relevant documents for multilingual content analysis: An evaluation of keyword and semantic similarity search approaches</article-title>. <source><italic>Computational Communication Research</italic></source>, <volume>5</volume>(<issue>2</issue>), <fpage>1</fpage>–<lpage>54</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.5117/CCR2023.2.PALI">https://doi.org/10.5117/CCR2023.2.PALI</ext-link></mixed-citation></ref>
      <ref id="bib.bibx39"><mixed-citation publication-type="journal"><string-name><surname>Peter</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Lauf</surname>, <given-names>E.</given-names></string-name> (<year>2002</year>). <article-title>Reliability in cross-national content analysis</article-title>. <source><italic>Journalism and Mass Communication Quarterly</italic></source>, <volume>79</volume>(<issue>4</issue>), <fpage>815</fpage>–<lpage>832</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1177/107769900207900403">https://doi.org/10.1177/107769900207900403</ext-link></mixed-citation></ref>
      <ref id="bib.bibx40"><mixed-citation publication-type="journal"><string-name><surname>Proksch</surname>, <given-names>SO.</given-names></string-name>, <string-name><surname>Lowe</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Wäckerle</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Soroka</surname>, <given-names>S.</given-names></string-name> (<year>2019</year>). <article-title>Multilingual sentiment analysis: A new approach to measuring conflict in legislative speeches</article-title>. <source><italic>Legislative Studies Quarterly</italic></source>, <volume>44</volume>(<issue>1</issue>), <fpage>97</fpage>–<lpage>131</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/lsq.12212">https://doi.org/10.1111/lsq.12212</ext-link></mixed-citation></ref>
      <ref id="bib.bibx41"><mixed-citation publication-type="journal"><string-name><surname>Rathje</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Mirea</surname>, <given-names>D. M.</given-names></string-name>, <string-name><surname>Sucholutsky</surname>, <given-names>I.</given-names></string-name>, <string-name><surname>Marjieh</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Robertson</surname>, <given-names>C. E.</given-names></string-name>, &amp; <string-name><surname>Van Bavel</surname>, <given-names>J. J.</given-names></string-name> (<year>2024</year>). <article-title>GPT is an effective tool for multilingual psychological text analysis</article-title>. <source><italic>Proceedings of the National Academy of Sciences</italic></source>, <volume>121</volume>(<issue>34</issue>), <fpage>e2308950121</fpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1073/pnas.2308950121">https://doi.org/10.1073/pnas.2308950121</ext-link></mixed-citation></ref>
      <ref id="bib.bibx42"><mixed-citation publication-type="journal"><string-name><surname>Rauh</surname>, <given-names>C.</given-names></string-name>, &amp; <string-name><surname>De Wilde</surname>, <given-names>P.</given-names></string-name> (<year>2018</year>). <article-title>The opposition deficit in EU accountability: Evidence from over 20 years of plenary debate in four member states</article-title>. <source><italic>European Journal of Political Research</italic></source>, <volume>57</volume>(<issue>1</issue>), <fpage>194</fpage>–<lpage>216</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/1475-6765.12203">https://doi.org/10.1111/1475-6765.12203</ext-link></mixed-citation></ref>
      <ref id="bib.bibx43"><mixed-citation publication-type="journal"><string-name><surname>Reber</surname>, <given-names>U.</given-names></string-name> (<year>2019</year>). <article-title>Overcoming language barriers: Assessing the potential of machine translation and topic modeling for the comparative analysis of multilingual text corpora</article-title>. <source><italic>Communication Methods and Measures</italic></source>, <volume>13</volume>(<issue>2</issue>), <fpage>102</fpage>–<lpage>125</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/19312458.2019.1610977">https://doi.org/10.1080/19312458.2019.1610977</ext-link></mixed-citation></ref>
      <ref id="bib.bibx44"><mixed-citation publication-type="journal"><string-name><surname>Rinke</surname>, <given-names>E. M.</given-names></string-name>, <string-name><surname>Dobbrick</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Löb</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Zirn</surname>, <given-names>C.</given-names></string-name>, &amp; <string-name><surname>Wessler</surname>, <given-names>H.</given-names></string-name> (<year>2022</year>). <article-title>Expert-informed topic models for document set discovery</article-title>. <source><italic>Communication Methods and Measures</italic></source>, <volume>16</volume>(<issue>1</issue>), <fpage>39</fpage>–<lpage>58</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/19312458.2021.2014861">https://doi.org/10.1080/19312458.2021.2014861</ext-link></mixed-citation></ref>
      <ref id="bib.bibx45"><mixed-citation publication-type="journal"><string-name><surname>Rodriguez</surname>, <given-names>P. L.</given-names></string-name>, &amp; <string-name><surname>Spirling</surname>, <given-names>A.</given-names></string-name> (<year>2022</year>). <article-title>Word embeddings: What works, what doesn’t, and how to tell the difference for applied research</article-title>. <source><italic>The Journal of Politics</italic></source>, <volume>84</volume>(<issue>1</issue>), <fpage>101</fpage>–<lpage>115</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1086/714944">https://doi.org/10.1086/714944</ext-link></mixed-citation></ref>
      <ref id="bib.bibx46"><mixed-citation publication-type="book"><string-name><surname>Rössler</surname>, <given-names>P.</given-names></string-name> (<year>2012</year>). <chapter-title>Comparative content analysis</chapter-title>. In <source><italic>The Handbook of Comparative Communication Research</italic></source> (pp. <fpage>481</fpage>–<lpage>490</lpage>). <publisher-name>Routledge</publisher-name>.</mixed-citation></ref>
      <ref id="bib.bibx47"><mixed-citation publication-type="book"><string-name><surname>Salganik</surname>, <given-names>M. J.</given-names></string-name> (<year>2017</year>). <source><italic>Bit by Bit: Social Research in the Digital Age</italic></source>. <publisher-name>Princeton University Press</publisher-name>.</mixed-citation></ref>
      <ref id="bib.bibx48"><mixed-citation publication-type="journal"><string-name><surname>Scharkow</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Trepte</surname>, <given-names>S.</given-names></string-name> (<year>2023</year>). <article-title>National diversity at conferences of the International Communication Association</article-title>. <source><italic>Annals of the International Communication Association</italic></source>,  <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/23808985.2023.2175968">https://doi.org/10.1080/23808985.2023.2175968</ext-link></mixed-citation></ref>
      <ref id="bib.bibx49"><mixed-citation publication-type="journal"><string-name><surname>Segev</surname>, <given-names>E.</given-names></string-name> (<year>2019</year>). <article-title>From where does the world look flatter? A comparative analysis of foreign coverage in world news</article-title>. <source><italic>Journalism</italic></source>, <volume>20</volume>(<issue>7</issue>), <fpage>924</fpage>–<lpage>942</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1177/1464884917709344">https://doi.org/10.1177/1464884917709344</ext-link></mixed-citation></ref>
      <ref id="bib.bibx50"><mixed-citation publication-type="book"><string-name><surname>Shababo</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Baden</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Dolinsky</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Lind</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Pipal</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Schoonvelde</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Smoliarova</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>van der Velden</surname>, <given-names>M. A. C. G.</given-names></string-name>, &amp; <string-name><surname>Zalik</surname>, <given-names>A.</given-names></string-name> (<year>2023</year>). <source><italic>How do linguistic differences impact computational text analysis methods? A road map for future validation and integrated strategy development</italic></source>.  <ext-link ext-link-type="uri" xlink:href="https://opted.eu/fileadmin/user_upload/k_opted/OPTED_Deliverable_D6.3.pdf">https://opted.eu/fileadmin/user_upload/k_opted/OPTED_Deliverable_D6.3.pdf</ext-link></mixed-citation></ref>
      <ref id="bib.bibx51"><mixed-citation publication-type="journal"><string-name><surname>Sigismondi</surname>, <given-names>P.</given-names></string-name> (<year>2018</year>). <article-title>Exploring translation gaps: The untranslatability and global diffusion of 'cool'</article-title>. <source><italic>Communication Theory</italic></source>, <volume>28</volume>(<issue>3</issue>), <fpage>292</fpage>–<lpage>310</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1111/comt.12123">https://doi.org/10.1111/comt.12123</ext-link></mixed-citation></ref>
      <ref id="bib.bibx52"><mixed-citation publication-type="journal"><string-name><surname>Song</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Tolochko</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Eberl</surname>, <given-names>JM.</given-names></string-name>, <string-name><surname>Eisele</surname>, <given-names>O.</given-names></string-name>, <string-name><surname>Greussing</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Heidenreich</surname>, <given-names>T.</given-names></string-name>, &amp; <string-name><surname>Boomgaarden</surname>, <given-names>H. G.</given-names></string-name> (<year>2020</year>). <article-title>In validations we trust? The impact of imperfect human annotations as a gold standard on the quality of validation of automated content analysis</article-title>. <source><italic>Political Communication</italic></source>, <volume>37</volume>(<issue>4</issue>), <fpage>550</fpage>–<lpage>572</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/10584609.2020.1811890">https://doi.org/10.1080/10584609.2020.1811890</ext-link></mixed-citation></ref>
      <ref id="bib.bibx53"><mixed-citation publication-type="journal"><string-name><surname>Spirling</surname>, <given-names>A.</given-names></string-name> (<year>2023</year>). <article-title>Why open-source generative AI models are an ethical way forward for science</article-title>. <source><italic>Nature</italic></source>, <volume>616</volume>(<issue>7957</issue>), <fpage>413</fpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1038/d41586-023-02395-7">https://doi.org/10.1038/d41586-023-02395-7</ext-link></mixed-citation></ref>
      <ref id="bib.bibx54"><mixed-citation publication-type="journal"><string-name><surname>Stecker</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Balluff</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Lind</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Dinhopl</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Waldherr</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name><surname>Boomgaarden</surname>, <given-names>H. G.</given-names></string-name> (<year>2024</year>). <article-title>Tools of the trade–when are software tools mentioned in computational text analysis research?</article-title>. <source><italic>Computational Communication Research</italic></source>, <volume>6</volume>(<issue>1</issue>), <fpage>1</fpage>–<lpage>21</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.55563/ccc/abcd12">https://doi.org/10.55563/ccc/abcd12</ext-link></mixed-citation></ref>
      <ref id="bib.bibx55"><mixed-citation publication-type="book"><string-name><surname>TeBlunthuis</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Hase</surname>, <given-names>V.</given-names></string-name>, &amp; <string-name><surname>Chan</surname>, <given-names>CH.</given-names></string-name> (<year>2023</year>). <source><italic>Misclassification in automated content analysis causes bias in regression. Can we fix it? Yes, we can!</italic></source>.  <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2307.06483">https://arxiv.org/abs/2307.06483</ext-link></mixed-citation></ref>
      <ref id="bib.bibx56"><mixed-citation publication-type="journal"><string-name><surname>Temporão</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Kerckhove</surname>, <given-names>C. V.</given-names></string-name>, <string-name><surname>van der Linden</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Dufresne</surname>, <given-names>Y.</given-names></string-name>, &amp; <string-name><surname>Hendrickx</surname>, <given-names>J. M.</given-names></string-name> (<year>2018</year>). <article-title>Ideological scaling of social media users: A dynamic lexicon approach</article-title>. <source><italic>Political Analysis</italic></source>, <volume>26</volume>(<issue>4</issue>), <fpage>457</fpage>–<lpage>473</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1017/pan.2018.19">https://doi.org/10.1017/pan.2018.19</ext-link></mixed-citation></ref>
      <ref id="bib.bibx57"><mixed-citation publication-type="journal"><string-name><surname>van Atteveldt</surname>, <given-names>W.</given-names></string-name>, &amp; <string-name><surname>Peng</surname>, <given-names>TQ.</given-names></string-name> (<year>2018</year>). <article-title>When communication meets computation: Opportunities, challenges, and pitfalls in computational communication science</article-title>. <source><italic>Communication Methods and Measures</italic></source>, <volume>12</volume>(<issue>2-3</issue>), <fpage>81</fpage>–<lpage>92</lpage>. <ext-link ext-link-type="doi" xlink:href="https://doi.org/10.1080/19312458.2018.1458084">https://doi.org/10.1080/19312458.2018.1458084</ext-link></mixed-citation></ref>
      <ref id="bib.bibx58"><mixed-citation publication-type="book"><string-name><surname>Volk</surname>, <given-names>S. C.</given-names></string-name> (<year>2022</year>). <source><italic>Comparative communication research</italic></source>. <publisher-name>Springer Fachmedien Wiesbaden</publisher-name>.</mixed-citation></ref>
      <ref id="bib.bibx59"><mixed-citation publication-type="book"><string-name><surname>Wirth</surname>, <given-names>W.</given-names></string-name>, &amp; <string-name><surname>Kolb</surname>, <given-names>S.</given-names></string-name> (<year>2004</year>). <chapter-title>Designs and methods of comparative political communication research</chapter-title>. In <source><italic>Comparing political communication: Theories, cases, and challenges</italic></source> (pp. <fpage>87</fpage>–<lpage>111</lpage>). <publisher-name>Cambridge University Press</publisher-name>.</mixed-citation></ref>
      <ref id="bib.bibx60"><mixed-citation publication-type="journal"><string-name><surname>Wozniak</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Wessler</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Chan</surname>, <given-names>C. H.</given-names></string-name>, &amp; <string-name><surname>Lueck</surname>, <given-names>J.</given-names></string-name> (<year>2021</year>). <article-title>The event-centered nature of global public spheres: The UN climate change conferences, Fridays for future, and the (limited) transnationalization of media debates</article-title>. <source><italic>International Journal of Communication</italic></source>, <volume>15</volume>, <fpage>27</fpage>.</mixed-citation></ref>
    </ref-list>
    </back>
</article>