<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" dtd-version="1.3" article-type="brief-report">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Water</journal-id>
<journal-title-group>
<journal-title>Frontiers in Water</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Water</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2624-9375</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/frwa.2026.1749745</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Perspective</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>A perspective on using large language models for human data in human-water research: Why we should be cautious?</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Sch&#x000FC;ck</surname> <given-names>Fredrik</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<xref ref-type="author-notes" rid="fn001"><sup>&#x02020;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<uri xlink:href="https://loop.frontiersin.org/people/3277897"/>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Department of Sustainable Development, Environmental Sciences and Engineering, KTH Royal Institute of Technology</institution>, <city>Stockholm</city>, <country country="se">Sweden</country></aff>
<aff id="aff2"><label>2</label><institution>Hydrology Research, Swedish Meteorological and Hydrological Institute (SMHI)</institution>, <city>Norrk&#x000F6;ping</city>, <country country="se">Sweden</country></aff>
<author-notes>
<corresp id="c001"><label>&#x0002A;</label>Correspondence: Fredrik Sch&#x000FC;ck, <email xlink:href="mailto:fredrik.schuck@smhi.se">fredrik.schuck@smhi.se</email></corresp>
<fn fn-type="other" id="fn001"><label>&#x02020;</label><p>ORCID: Fredrik Sch&#x000FC;ck <uri xlink:href="https://orcid.org/0009-0001-4107-8440">orcid.org/0009-0001-4107-8440</uri></p></fn></author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-26">
<day>26</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>8</volume>
<elocation-id>1749745</elocation-id>
<history>
<date date-type="received">
<day>19</day>
<month>11</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>10</day>
<month>02</month>
<year>2026</year>
</date>
<date date-type="accepted">
<day>10</day>
<month>02</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2026 Sch&#x000FC;ck.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Sch&#x000FC;ck</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-26">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>The rise of human-water research in hydrology has increased the need for data on human behavior and decision-making. To address this demand, hydrologists have embraced methods from the social sciences, such as surveys, interviews, and agent-based modeling (ABM). However, collecting human data is expensive and time-consuming. Therefore, some social scientists have started evaluating whether Large Language Models (LLMs) could serve as a tool for generating human-like data for surveys and social simulations. This approach have the potential to transform human-water research by making access to human data faster and more affordable. Yet, human-water researchers should be cautious about adopting this method. The method has faced criticism in the social sciences, and similar caveats would apply to human-water research. While LLMs can provide responses based on different demographic personas, they fail to replicate human complexity and diversity. They also pose challenges to scientific rigor due to limited transparency and the risks of hallucinations. Most importantly, LLM-generated data fails to accurately represent marginalized groups, and undermines efforts to make human&#x02013;water research more participatory, inclusive, and transformative.</p></abstract>
<kwd-group>
<kwd>coupled human-water systems</kwd>
<kwd>generative AI</kwd>
<kwd>human data</kwd>
<kwd>large language models</kwd>
<kwd>sociohydrology</kwd>
<kwd>water management</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was not received for this work and/or its publication.</funding-statement>
</funding-group>
<counts>
<fig-count count="0"/>
<table-count count="0"/>
<equation-count count="0"/>
<ref-count count="80"/>
<page-count count="7"/>
<word-count count="6502"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Water and Human Systems</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<label>1</label>
<title>Introduction</title>
<p>Human-water system research has emerged as an important interdisciplinary lens to understand how humans interact with water systems in the Anthropocene. Examples of these interdisciplinary research efforts include sociohydrology, social-ecological systems (SES), and the International Association of Hydrological Sciences (IAHS) scientific decades&#x00027; Panta Rhei and HELPING (<xref ref-type="bibr" rid="B21">Gain et al., 2021</xref>; <xref ref-type="bibr" rid="B37">Kreibich et al., 2025</xref>; <xref ref-type="bibr" rid="B59">Sivapalan et al., 2012</xref>; <xref ref-type="bibr" rid="B6">Arheimer et al., 2024</xref>). To explore these new perspectives, researchers need to collect human data, such as data on human behaviors, attitudes, and decisions, in addition to environmental data (<xref ref-type="bibr" rid="B11">Blair and Buytaert, 2016</xref>). However, this data collection is challenging as it requires considerable time, resources, and knowledge of appropriate data-collection methods (<xref ref-type="bibr" rid="B46">Pande and Sivapalan, 2017</xref>; <xref ref-type="bibr" rid="B54">Salganik, 2019</xref>; <xref ref-type="bibr" rid="B58">Schwitter et al., 2022</xref>).</p>
<p>Human-water research has previously adopted methods from the social sciences, including surveys, text mining, and agent-based modeling (ABM) (<xref ref-type="bibr" rid="B14">Bubeck and Thieken, 2018</xref>; <xref ref-type="bibr" rid="B24">Haer et al., 2020</xref>; <xref ref-type="bibr" rid="B40">Madruga de Brito et al., 2020</xref>; <xref ref-type="bibr" rid="B18">Di Baldassarre et al., 2015</xref>) to address this new data need. While these methods have enhanced our understanding of human-water interactions (<xref ref-type="bibr" rid="B37">Kreibich et al., 2025</xref>), the collection is challenging, especially in the last decades, due to a decline in response rate for surveys (<xref ref-type="bibr" rid="B30">Jabkowski and Cichocki, 2025</xref>) or a decreased ability to collect social media data (<xref ref-type="bibr" rid="B44">Murthy, 2024</xref>; <xref ref-type="bibr" rid="B66">Trezza, 2023</xref>). To address these data collection challenges, Large Language Models (LLMs) have recently received attention in social sciences as an alternative tool, capable of producing synthetic human data to generate new datasets or fill gaps in existing ones (<xref ref-type="bibr" rid="B36">Kozlowski and Evans, 2025</xref>; <xref ref-type="bibr" rid="B23">Grossmann et al., 2023</xref>). Since OpenAI released ChatGPT in 2022, LLMs have become more accessible and have changed how people, including researchers, view the capabilities of AI (<xref ref-type="bibr" rid="B77">Zhang et al., 2025</xref>).</p>
<p>In hydrology and human-water research, the adoption of LLMs remains limited but growing (<xref ref-type="bibr" rid="B39">Ma et al., 2025</xref>). Novel approaches include a hydrology-informed chatbot, WaterGPT (<xref ref-type="bibr" rid="B51">Ren et al., 2024</xref>), or the extraction of hydrological data from a vast collection of research studies (<xref ref-type="bibr" rid="B42">Miao et al., 2024</xref>). However, only three studies have applied LLMs in human-water research (<xref ref-type="bibr" rid="B31">Kadiyala et al., 2025</xref>; <xref ref-type="bibr" rid="B9">Batista et al., 2025</xref>; <xref ref-type="bibr" rid="B79">Zhu et al., 2024</xref>), highlighting a research topic in its infancy. Even so, we should be cautious about adopting LLMs in human-water research, as the application of LLM-generated human data is currently debated in the social sciences due to challenges with scientific rigor and ethics, including misinformation, replicability, and hidden biases (<xref ref-type="bibr" rid="B53">Rossi et al., 2024</xref>; <xref ref-type="bibr" rid="B71">Weidinger et al., 2022</xref>). This critique is also relevant for hydrology, especially since hydrologists generally lack formal training in studying human behavior and attitudes.</p>
<p>In this article, I argue that although LLMs offer potential for addressing data collection challenges in human-water research, we should be careful to incorporate them. In the following sections, I will introduce LLMs, review their emerging incorporation in social sciences and hydrology, and discuss their risks and possible paths forward.</p></sec>
<sec id="s2">
<label>2</label>
<title>Large language models in science</title>
<sec>
<label>2.1</label>
<title>Brief introduction to large language models</title>
<p>Large language models are deep neural networks trained on vast amounts of data to interpret and generate text. The training data consists of text that is converted into tokens, e.g., words and phrases, which are then transformed into numerical presentations, called word embeddings. Word embedding is a transformation of a word to a vector so that related words are close in the vector space. The tokenized text is used to train an unsupervised transformer model that consists of either an encoder, a decoder, or both (<xref ref-type="bibr" rid="B67">Vaswani et al., 2017</xref>).</p>
<p>Earlier language models often used the encoder BERT for natural language processing (NLP) tasks while more recent models, including Gemma and GPT-4, use a decoder model structure, commonly a generative pre-trained transformer model (GPT) (<xref ref-type="bibr" rid="B17">Devlin et al., 2019</xref>; <xref ref-type="bibr" rid="B70">Wang et al., 2024</xref>). Both approaches utilize an attention mechanism that determines the importance of each word in its textual context (<xref ref-type="bibr" rid="B67">Vaswani et al., 2017</xref>; <xref ref-type="bibr" rid="B7">Bahdanau et al., 2016</xref>). Hence, LLMs combine linguistic and computational methods to produce human-like text by predicting the most probable sequence of words in response to a given prompt (<xref ref-type="bibr" rid="B53">Rossi et al., 2024</xref>; <xref ref-type="bibr" rid="B50">Raiaan et al., 2024</xref>).</p>
<p>The vast scale of training data is central to LLMs&#x00027; ability to generate human-like text. To achieve this, LLM companies obtain text from the internet, including books, and scientific and code databases (<xref ref-type="bibr" rid="B50">Raiaan et al., 2024</xref>). However, the training data and its sources are rarely disclosed, and concerns have been raised that LLM companies disregard copyrights and licenses (<xref ref-type="bibr" rid="B8">Bail, 2024</xref>; <xref ref-type="bibr" rid="B62">Stokel-Walker and Van Noorden, 2023</xref>). The data are used to train models containing billions of parameters, a process that requires extensive computational resources and time (<xref ref-type="bibr" rid="B73">Widder et al., 2024</xref>). Some models, such as GPT-4, are fine-tuned to reduce undesirable behavior, like hate speech, and promote more polite outputs (<xref ref-type="bibr" rid="B8">Bail, 2024</xref>; <xref ref-type="bibr" rid="B65">T&#x000F6;rnberg et al., 2023</xref>). As with the data, the fine-tuning process is often undisclosed, and thus most LLMs are black-box text predictors (<xref ref-type="bibr" rid="B73">Widder et al., 2024</xref>).</p>
</sec>
<sec>
<label>2.2</label>
<title>Large language models in social sciences</title>
<p>LLMs have gained traction for their potential to transform sociological methods and research (<xref ref-type="bibr" rid="B23">Grossmann et al., 2023</xref>; <xref ref-type="bibr" rid="B8">Bail, 2024</xref>; <xref ref-type="bibr" rid="B80">Ziems et al., 2024</xref>). (<xref ref-type="bibr" rid="B23">Grossmann et al. 2023</xref>) argue that the LLMs&#x00027; ability to mimic human behavior could redefine social science, reducing the need for using other methods such as crowdsourcing. Instead, they envision that LLMs can generate human-like data without involving participants and thus provide what (<xref ref-type="bibr" rid="B27">Horton 2023</xref>) describes as <italic>homo silicus</italic>, a computational model of humans. The researcher in favor of this argues that it is a scalable, fast, and cost-effective method that also protects human participants (<xref ref-type="bibr" rid="B2">Agnew et al., 2024</xref>).</p>
<p>To explore LLMs as stand-ins for humans, sociologists have investigated how well LLMs can replicate human behavior. For instance, (<xref ref-type="bibr" rid="B19">Dillion et al. 2023</xref>) assessed GPT-3.5 on moral scenarios and found its judgments aligned closely with human responses. Horton found that LLMs generally reproduced behavior similar to humans in four socio-economic experiments, and (<xref ref-type="bibr" rid="B41">Mei et al. 2024</xref>) explored LLMs&#x00027; performance on the &#x0201C;Big Five&#x0201D; personality traits by employing six psychological games, including prisoner&#x00027;s dilemma. Similarly, (<xref ref-type="bibr" rid="B3">Aher et al. 2023</xref>) introduced the method &#x0201C;<italic>Turing Experiments</italic>&#x0201D; to evaluate whether GPT models could behave indistinguishably from humans in four physiological experiments, including Milgram&#x00027;s (<xref ref-type="bibr" rid="B43">1963</xref>) shock experiment. However, because GPT is trained on scientific data, the model may already be familiar with these experiments. To address this, Aher et al. altered the experiments&#x00027; description, but it is uncertain if they succeeded in making the LLM unbiased to past experiment results.</p>
<p>In many studies, LLMs are conditioned with personalities and demographic data to alter the outputs and explore if they can take the role of a human. The traits can be gender (<xref ref-type="bibr" rid="B41">Mei et al., 2024</xref>; <xref ref-type="bibr" rid="B3">Aher et al., 2023</xref>), surnames representing different ethnic groups (<xref ref-type="bibr" rid="B3">Aher et al., 2023</xref>), or political affiliation (<xref ref-type="bibr" rid="B27">Horton, 2023</xref>). Such conditioning could be useful in applications in human-water research. For example, an LLM could model flood adaptation strategies for different demographic groups or explore the safe-development paradox by simulating risk perception. However, LLMs are trained on text and not behavior, and although behavior might be implicit, research shows LLMs&#x00027; limitations in understanding basic human reasoning (<xref ref-type="bibr" rid="B16">Cui et al., 2023</xref>; <xref ref-type="bibr" rid="B61">Stevenson et al., 2025</xref>). This makes it difficult to interpret how well <italic>homo silicus</italic> mimic real humans.</p>
<p>Other studies have scaled the human-conditioning approach by incorporating LLMs in survey studies to generate synthetic responses (<xref ref-type="bibr" rid="B5">Argyle et al., 2023</xref>; <xref ref-type="bibr" rid="B10">Bisbee et al., 2023</xref>; <xref ref-type="bibr" rid="B26">H&#x000E4;m&#x000E4;l&#x000E4;inen et al., 2023</xref>; <xref ref-type="bibr" rid="B34">Kim and Lee, 2024</xref>). For instance, Argyle examined LLMs&#x00027; ability to generate opinions about political parties and voting behavior in the US when conditioning the model with demographic backstories from the American National Election Studies (ANES). They showed that LLMs&#x00027; answers were similar to human respondents and argue that LLMs, &#x0201C;<italic>when properly conditioned</italic>&#x0201D; can produce human-like opinions along fine-grained demographic axes.</p>
<p>As surveys are an important tool in human-water research (<xref ref-type="bibr" rid="B37">Kreibich et al., 2025</xref>), for example, to study sociohydrological concepts such as the near-miss effect (<xref ref-type="bibr" rid="B13">Bogani et al., 2023</xref>) or flood recovery (<xref ref-type="bibr" rid="B14">Bubeck and Thieken, 2018</xref>), leveraging LLMs could help with data collection. However, researchers have shown that while LLMs can reproduce average answers from specific demographic groups, their ability to capture the variance and complexity of human opinions is limited (<xref ref-type="bibr" rid="B10">Bisbee et al., 2023</xref>; <xref ref-type="bibr" rid="B12">Boelaert et al., 2025</xref>), and some researchers report misrepresentation of representation of demographic groups (<xref ref-type="bibr" rid="B69">Wang et al., 2025</xref>; <xref ref-type="bibr" rid="B78">Zhou et al., 2025</xref>). This bias is unevenly distributed between demographic groups and could be worse for marginalized groups. (<xref ref-type="bibr" rid="B69">Wang et al. 2025</xref>) found that GPT, to a larger extent, misrepresented marginalized groups, such as Black and non-binary people. They hypothesized that out-of-group populations in the training data are influenced by political discourse. LLMs also perform worse at generating non-US synthetic opinions (<xref ref-type="bibr" rid="B1">Abeliuk et al., 2025</xref>), probably due to the overrepresentation of English and US-centric text in the training data. These limitations pose a risk in human-water research, as individuals subjected to floods and droughts can be marginalized and may have limited influence on the data generated about them (<xref ref-type="bibr" rid="B71">Weidinger et al., 2022</xref>; <xref ref-type="bibr" rid="B56">Savelli et al., 2021</xref>).</p>
<p>Another important tool for social science and coupled human-water system modeling is ABMs, where humans are simulated as computational agents with their behavior governed by decision rules (<xref ref-type="bibr" rid="B75">Wooldridge, 2009</xref>). In earlier ABMs, the agents were often controlled by simplistic rules, but efforts have been made to create more elaborate agents to mimic the complexity of society, for example, reinforcement learning (<xref ref-type="bibr" rid="B28">Hung et al., 2022</xref>) and bounded rationality (<xref ref-type="bibr" rid="B64">Taberna et al., 2023</xref>; <xref ref-type="bibr" rid="B72">Wens et al., 2019</xref>). Despite this, the increased complexity of ABMs has not always contributed to explanatory power, as modeling human behavior is challenging (<xref ref-type="bibr" rid="B63">Sun et al., 2016</xref>), and data scarcity on multi-agent interactions limits modeling of the human-water system (<xref ref-type="bibr" rid="B57">Sch&#x000FC;ck et al., 2025</xref>).</p>
<p>In light of this, researchers have noted that incorporating LLMs into ABMs could be a way forward, as it would be able to generate &#x0201C;<italic>social interactions of individuals with specific characteristics and beliefs</italic>&#x0201D; (<xref ref-type="bibr" rid="B23">Grossmann et al., 2023</xref>). So far, LLMs in ABMs have been used to explore echo chambers (<xref ref-type="bibr" rid="B65">T&#x000F6;rnberg et al., 2023</xref>), human-like interactions (<xref ref-type="bibr" rid="B33">Kaiya et al., 2023</xref>; <xref ref-type="bibr" rid="B47">Park et al., 2023</xref>), and people&#x00027;s attitudes to gender discrimination and nuclear energy (<xref ref-type="bibr" rid="B22">Gao et al., 2023</xref>). (<xref ref-type="bibr" rid="B65">T&#x000F6;rnberg et al. 2023</xref>) used ANES data to create LLM agents with demographic personas and implemented them in an ABM that represented a social media platform where agents could share, comment, and like news articles. The authors examined how political affiliations and social media bubbles fostered a toxic environment. The integrated LLM-ABM approach allowed for controlling the platform&#x00027;s feed algorithms without requiring collaboration with social media companies, and it avoided exposing humans to toxic environments. However, the setup couldn&#x00027;t be validated, and the influence of LLMs&#x00027; politeness bias was not evaluated.</p></sec>
<sec>
<label>2.3</label>
<title>Large language models in hydrology and human-water research</title>
<p>In hydrology, LLMs have had a limited impact so far. A few studies have discussed the implications of LLMs in hydrological research and have mainly focused on the coding and writing aid abilities of LLMs (<xref ref-type="bibr" rid="B20">Foroumandi et al., 2023</xref>; <xref ref-type="bibr" rid="B25">Halloran et al., 2023</xref>; <xref ref-type="bibr" rid="B29">Irvine et al., 2023</xref>) and their ability to describe hydrological concepts (<xref ref-type="bibr" rid="B39">Ma et al., 2025</xref>; <xref ref-type="bibr" rid="B25">Halloran et al., 2023</xref>; <xref ref-type="bibr" rid="B32">Kadiyala et al., 2024</xref>). Both (<xref ref-type="bibr" rid="B51">Ren et al. 2024</xref>) and (<xref ref-type="bibr" rid="B76">Xu et al. 2025</xref>) expanded this by respectively developing a hydrological-informed LLM, called WaterGPT, by finetuning an existing GPT model with hydrological data. Other hydrological studies used LLMs to extract basin areas from abstracts in their review (<xref ref-type="bibr" rid="B42">Miao et al., 2024</xref>) or integrated ChatGPT as a user interface for an AI-driven flood detection system (<xref ref-type="bibr" rid="B38">Kumbam and Vejre, 2024</xref>).</p>
<p>LLMs have been adopted to a more limited extent in human-water research, yet NLP has been used before, as several studies have used it to extract data from newspapers and social media (<xref ref-type="bibr" rid="B40">Madruga de Brito et al., 2020</xref>; <xref ref-type="bibr" rid="B68">Veigel et al., 2025</xref>). However, in the last 2 years, three studies have been published that incorporate LLMs in human-water research, where one study utilized LLMs to generate human-like data (<xref ref-type="bibr" rid="B31">Kadiyala et al., 2025</xref>). The other two papers either used LLM to perform sentiment analysis on meeting notes in a drought-affected community (<xref ref-type="bibr" rid="B9">Batista et al., 2025</xref>) or developed an integrated GIS-LLM chatbot to provide personalized flood-related information to increase people&#x00027;s flood risk perception (<xref ref-type="bibr" rid="B79">Zhu et al., 2024</xref>). (<xref ref-type="bibr" rid="B31">Kadiyala et al. 2025</xref>) developed an ABM where LLMs were used to create agents with demographic personas from traits such as age, health, and cultural background. These LLM agent groups were then tasked with allocating a budget for different flood mitigation projects. With the LLM responses, the authors&#x00027; could explore how the constellation of personas influenced the allocation. Similarly to the sociology articles, the authors noted that LLMs failed to provided variance within and between different demographic personas but note that LLM simulations</p>
<disp-quote><p>&#x0201C;<italic>can guide urban planning, infrastructure development, and disaster risk reduction by providing data-driven insights tailored to specific community requirements</italic>&#x0201D;.</p></disp-quote>
<p>Despite the limited implementation of LLMs thus far, hydrologist researchers envision that LLMs will radically change hydrological modeling with automation and knowledge integration that was previously impossible (<xref ref-type="bibr" rid="B39">Ma et al., 2025</xref>). Human-water research might follow this trend to incorporate LLMs similarly to how it&#x00027;s happening in the social sciences.</p>
</sec>
</sec>
<sec sec-type="discussion" id="s3">
<label>3</label>
<title>Discussion</title>
<p>This emerging use of LLMs could enable simple implementations of human data in hydrology, potentially reducing expensive, time-consuming, and complicated data collection, preparation, and integration. However, I believe that it poses significant risks of producing believable yet flawed and unethical research. Specifically, the application of LLMs in human-water research runs the risk of increasing the gap between research and vulnerable communities as public engagement decreases. In the following discussion, I will outline both the caveats and potential opportunities to use LLMs with caution.</p>
<p>One issue is that LLMs can hallucinate and provide incorrect information. This issue has been met with the argument that all models are wrong, but some are useful and that model output from any model has to be validated (<xref ref-type="bibr" rid="B53">Rossi et al., 2024</xref>; <xref ref-type="bibr" rid="B27">Horton, 2023</xref>). However, understanding when the model is incorrect can be difficult, as there is no indication of uncertainty in the output. (<xref ref-type="bibr" rid="B60">Spitale et al. 2023</xref>) called this combination of certainty and incorrectness <italic>compelling disinformation</italic> and showed that GPT-3 is better at creating believable disinformation than humans. For instance, ChatGPT can provide sources to research papers that do not exist (<xref ref-type="bibr" rid="B20">Foroumandi et al., 2023</xref>). Conversely, LLMs can be overly correct for the purpose of modeling humans, as they can give exact answers to numerical questions (<xref ref-type="bibr" rid="B3">Aher et al., 2023</xref>). This probably stems from that LLMs are not predicting human behavior, rather they predict the most likely groups of words (<xref ref-type="bibr" rid="B19">Dillion et al., 2023</xref>; <xref ref-type="bibr" rid="B4">Amirizaniani et al., 2024</xref>).</p>
<p>Another issue is that the LLMs exhibit biased behavior when representing humans. Some researchers see this as a possibility to steer the LLM by prompting it to mimic certain groups, genders, etc. (<xref ref-type="bibr" rid="B31">Kadiyala et al., 2025</xref>; <xref ref-type="bibr" rid="B65">T&#x000F6;rnberg et al., 2023</xref>; <xref ref-type="bibr" rid="B3">Aher et al., 2023</xref>). However, models are generally politically left-leaning, altruistic, and pro-climate action and show insignificant difference when prompted to represent other groups (<xref ref-type="bibr" rid="B41">Mei et al., 2024</xref>; <xref ref-type="bibr" rid="B55">Santurkar et al., 2023</xref>). Since it also has a limited ability to represent marginalized and non-US populations (<xref ref-type="bibr" rid="B41">Mei et al., 2024</xref>; <xref ref-type="bibr" rid="B69">Wang et al., 2025</xref>; <xref ref-type="bibr" rid="B1">Abeliuk et al., 2025</xref>), the appropriateness of generated data from demographic personas should be questioned.</p>
<p>In addition to the limitation of low variance and misrepresentation of non-Americans, LLMs have a risk of generating stereotypes; for instance, LLMs have propagated gender and racial stereotypes (<xref ref-type="bibr" rid="B71">Weidinger et al., 2022</xref>; <xref ref-type="bibr" rid="B35">Kotek et al., 2023</xref>). This reproduction of disinformation and stereotypes complicates validation&#x0201E; as outputs might align with our own prejudices and thus reinforce confirmation biases. This is problematic as researchers have suggested that LLMs can be used to scale research on underrepresented populations (<xref ref-type="bibr" rid="B53">Rossi et al., 2024</xref>). Additionally, LLMs are prone to value-lock-in; thus, the LLM represents values present in the training data that are no longer relevant, for instance, missing the change of the meaning of the word queer (<xref ref-type="bibr" rid="B71">Weidinger et al., 2022</xref>).</p>
<p>It is argued by some social scientists that LLM methods can be more ethical than traditional methods since LLMs can generate data where experiments would either be unethical to carry out on humans or for vulnerable groups where there are data gaps (<xref ref-type="bibr" rid="B27">Horton, 2023</xref>; <xref ref-type="bibr" rid="B3">Aher et al., 2023</xref>). However, this approach risks that communities that have historically been excluded will continue to be neglected and misrepresented as researchers take control of their story (<xref ref-type="bibr" rid="B2">Agnew et al., 2024</xref>). Thus, LLM-generated data of marginalized communities can reproduce patterns of historical colonialism (<xref ref-type="bibr" rid="B53">Rossi et al., 2024</xref>; <xref ref-type="bibr" rid="B15">Couldry and Mejias, 2019</xref>). I argue that this exclusion and misrepresentation of local communities is the opposite of the aim of human-water research, where interdisciplinary and participatory efforts are important (<xref ref-type="bibr" rid="B45">Nakamura et al., 2025</xref>). For instance, a goal of the IAHS scientific decade HELPING is to embrace local and indigenous knowledge (<xref ref-type="bibr" rid="B6">Arheimer et al., 2024</xref>). The use of LLMs to represent people, especially non-Western vulnerable communities, would be a step backwards for the human-water aim of participatory efforts and engaging local communities. Additionally, understudied human-water interactions might remain hidden when there is less interaction between researchers and communities due to higher reliance on LLM-data.</p>
<p>The lack of replicability and transparency in LLM-generated data poses a critical challenge for rigorous research. For instance, (<xref ref-type="bibr" rid="B10">Bisbee et al. 2023</xref>) found that the generated personas were not reproducible between different versions of ChatGPT, even if identical prompts were used. The behavior of the model can also change in the same model version (<xref ref-type="bibr" rid="B53">Rossi et al., 2024</xref>). This becomes problematic as studies most often can only include model versions and prompts in their methods, as training data and model parameters are undisclosed by the LLM companies. As the training data is unknown, researchers also have difficulties assessing LLMs&#x00027; biases (<xref ref-type="bibr" rid="B8">Bail, 2024</xref>). Thus, following the FAIR data standard (<xref ref-type="bibr" rid="B74">Wilkinson et al., 2016</xref>) and assessing the external validity will be challenging. On top of LLMs being potentially harmful to scientific rigor, the use raises ethical concerns of infringing copyright (<xref ref-type="bibr" rid="B49">Rahman and Santacana, 2023</xref>), environmental impact (<xref ref-type="bibr" rid="B52">Rillig et al., 2023</xref>), and exploitation of workers (<xref ref-type="bibr" rid="B48">Perrigo, 2023</xref>).</p>
<p>Another question is how thorough technical understanding of LLMs&#x00027; we need. I believe that since most LLMs are opaque in terms of their biases, training data, and limitations, researchers must have sufficient knowledge to use them responsibly. This also includes ethical concerns, such as environmental impact. While a complete technical expertise may be unnecessary, we should be cautious about adopting technologies simply because they are easy to use.</p>
<p>How should we then use LLMs in human-water research? Arguably, LMs are powerful coding and writing aids. LLMs also have a strong capability in analyzing and summarizing vast amounts of text, though we should carefully check the outputs. Additionally, they have been suggested for prototyping research ideas (<xref ref-type="bibr" rid="B19">Dillion et al., 2023</xref>; <xref ref-type="bibr" rid="B47">Park et al., 2023</xref>). For instance, researchers can test concepts before employing expensive data collection campaigns. However, it is uncertain how helpful it will be, as the results should always be validated with real data due to problems with hallucination and bias (<xref ref-type="bibr" rid="B53">Rossi et al., 2024</xref>).</p>
<p>Another possibility is open-source LLM projects built for research purposes, where the training data, model training, and fine-tuning processes are shared, like EleutherAI (<xref ref-type="bibr" rid="B73">Widder et al., 2024</xref>). Open-source models could thus provide more transparency and replicability, and new results show that the performance of open-source models is getting closer to that of proprietary LLMs (<xref ref-type="bibr" rid="B8">Bail, 2024</xref>). Another benefit is that research would become more independent of large tech companies. For instance, Meta&#x00027;s LLM model was open source but gradually became more opaque (<xref ref-type="bibr" rid="B8">Bail, 2024</xref>). Despite this progress and the prospect of open-source LLMs, there are still challenges to be addressed, such as the representation of groups, ethical considerations of collecting data, and environmental impact, as well as whether generating human data is epistemically sound (<xref ref-type="bibr" rid="B53">Rossi et al., 2024</xref>).</p></sec>
<sec id="s4">
<label>4</label>
<title>Conclusions</title>
<p>The use of LLMs could fundamentally transform how human-water research is conducted. They have the potential to simplify complex data collection and make it more accessible, potentially leading to new insights about human-water interactions. However, as illustrated in this article, it remains unclear how well these insights will translate into real-world applications and to what degree of certainty we can use the model outputs. From my perspective, LLMs pose more risks than benefits, particularly concerning ethics related to inclusion and representation. The HELPING decade and human-water research are just as much about understanding human-water systems as they are about engaging communities affected by climate change and hydrological hazards. Therefore, LLMs might steer us in the wrong direction by excluding those with knowledge and those who might need our help.</p></sec>
</body>
<back>
<sec sec-type="data-availability" id="s5">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="author-contributions" id="s6">
<title>Author contributions</title>
<p>FS: Conceptualization, Methodology, Writing &#x02013; original draft, Writing &#x02013; review &#x00026; editing, Formal analysis, Investigation.</p>
</sec>
<ack><title>Acknowledgments</title><p>I am grateful for the valuable comments and suggestions from the two reviewers. I am also thankful to Derya Akbaba for her valuable feedback on the draft of this paper.</p>
</ack>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s8">
<title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p></sec>
<sec sec-type="disclaimer" id="s9">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Abeliuk</surname> <given-names>A.</given-names></name> <name><surname>Gaete</surname> <given-names>V.</given-names></name> <name><surname>Bro</surname> <given-names>N.</given-names></name></person-group> (<year>2025</year>). <article-title>Fairness in LLM-generated surveys [internet]</article-title>. <source>arXiv</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2501.15351">http://arxiv.org/abs/2501.15351</ext-link> (Accessed November 1, 2025).</mixed-citation>
</ref>
<ref id="B2">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Agnew</surname> <given-names>W.</given-names></name> <name><surname>Bergman</surname> <given-names>A. S.</given-names></name> <name><surname>Chien</surname> <given-names>J.</given-names></name> <name><surname>D&#x000ED;az</surname> <given-names>M.</given-names></name> <name><surname>El-Sayed</surname> <given-names>S.</given-names></name> <name><surname>Pittman</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>&#x0201C;The illusion of artificial inclusion,&#x0201D;</article-title> in <source>Proceedings of the CHI Conference on Human Factors in Computing Systems [Internet]</source>, <fpage>1</fpage>&#x02013;<lpage>12</lpage>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2401.08572">http://arxiv.org/abs/2401.08572</ext-link> (Accessed November 7, 2025). doi: <pub-id pub-id-type="doi">10.1145/3613904.3642703</pub-id></mixed-citation>
</ref>
<ref id="B3">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Aher</surname> <given-names>G. V.</given-names></name> <name><surname>Arriaga</surname> <given-names>R. I.</given-names></name> <name><surname>Kalai</surname> <given-names>A. T.</given-names></name></person-group> (<year>2023</year>). <article-title>&#x0201C;Using large language models to simulate multiple humans and replicate human subject studies,&#x0201D;</article-title> in <source>Proceedings of the 40th International Conference on Machine Learning [Internet]</source>, eds. <person-group person-group-type="editor"><name><surname>Krause</surname> <given-names>A.</given-names></name> <name><surname>Brunskill</surname> <given-names>E.</given-names></name> <name><surname>Cho</surname> <given-names>K.</given-names></name> <name><surname>Engelhardt</surname> <given-names>B.</given-names></name> <name><surname>Sabato</surname> <given-names>S.</given-names></name> <name><surname>Scarlett</surname> <given-names>J.</given-names></name></person-group>, <fpage>337</fpage>&#x02013;<lpage>371</lpage>. <source>PMLR (Proceedings of Machine Learning Research</source>). Available online at: <ext-link ext-link-type="uri" xlink:href="https://proceedings.mlr.press/v202/aher23a.html">https://proceedings.mlr.press/v202/aher23a.html</ext-link> (Accessed October 24, 2025).</mixed-citation>
</ref>
<ref id="B4">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Amirizaniani</surname> <given-names>M.</given-names></name> <name><surname>Martin</surname> <given-names>E.</given-names></name> <name><surname>Sivachenko</surname> <given-names>M.</given-names></name> <name><surname>Mashhadi</surname> <given-names>A.</given-names></name> <name><surname>Shah</surname> <given-names>C.</given-names></name></person-group> (<year>2024</year>). <article-title>Do LLMs exhibit human-like reasoning? Evaluating theory of mind in LLMs for open-ended responses [internet]</article-title>. <source>arXiv</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2406.05659">http://arxiv.org/abs/2406.05659</ext-link> (Accessed November 7, 2025).</mixed-citation>
</ref>
<ref id="B5">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Argyle</surname> <given-names>L. P.</given-names></name> <name><surname>Busby</surname> <given-names>E. C.</given-names></name> <name><surname>Fulda</surname> <given-names>N.</given-names></name> <name><surname>Gubler</surname> <given-names>J. R.</given-names></name> <name><surname>Rytting</surname> <given-names>C.</given-names></name> <name><surname>Wingate</surname> <given-names>D.</given-names></name></person-group> (<year>2023</year>). <article-title>Out of one, many: using language models to simulate human samples</article-title>. <source>Polit. Anal</source>. <volume>31</volume>, <fpage>337</fpage>&#x02013;<lpage>351</lpage>. doi: <pub-id pub-id-type="doi">10.1017/pan.2023.2</pub-id></mixed-citation>
</ref>
<ref id="B6">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Arheimer</surname> <given-names>B.</given-names></name> <name><surname>Cudennec</surname> <given-names>C.</given-names></name> <name><surname>Castellarin</surname> <given-names>A.</given-names></name> <name><surname>Grimaldi</surname> <given-names>S.</given-names></name> <name><surname>Heal</surname> <given-names>K. V.</given-names></name> <name><surname>Lupton</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>The IAHS science for solutions decade, with Hydrology Engaging Local People IN one Global world (HELPING)</article-title>. <source>Hydrol. Sci. J</source>. <volume>69</volume>, <fpage>1417</fpage>&#x02013;<lpage>1435</lpage>. doi: <pub-id pub-id-type="doi">10.1080/02626667.2024.2355202</pub-id></mixed-citation>
</ref>
<ref id="B7">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Bahdanau</surname> <given-names>D.</given-names></name> <name><surname>Cho</surname> <given-names>K.</given-names></name> <name><surname>Bengio</surname> <given-names>Y.</given-names></name></person-group> (<year>2016</year>). <article-title>Neural machine translation by jointly learning to align and translate [internet]</article-title>. <source>arXiv</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/1409.0473">http://arxiv.org/abs/1409.0473</ext-link> (Accessed October 24, 2025).</mixed-citation>
</ref>
<ref id="B8">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bail</surname> <given-names>C. A.</given-names></name></person-group> (<year>2024</year>). <article-title>Can generative AI improve social science?</article-title> <source>Proc. Natl. Acad. Sci</source>. <volume>121</volume>:<fpage>e2314021121</fpage>. doi: <pub-id pub-id-type="doi">10.1073/pnas.2314021121</pub-id></mixed-citation>
</ref>
<ref id="B9">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Batista</surname> <given-names>T. L.</given-names></name> <name><surname>de Carvalho Studart</surname> <given-names>T. M.</given-names></name> <name><surname>Duarte</surname> <given-names>M. G.</given-names></name> <name><surname>Souza Filho</surname> <given-names>F. D. A. D.</given-names></name></person-group> (<year>2025</year>). <article-title>Modeling sentiment&#x02013;hydrology interaction using LLM: insights for adaptive governance in Cear&#x000E1;&#x00027;s water management</article-title>. <source>Water</source>. <volume>17</volume>:<fpage>2615</fpage>. doi: <pub-id pub-id-type="doi">10.3390/w17172615</pub-id></mixed-citation>
</ref>
<ref id="B10">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Bisbee</surname> <given-names>J.</given-names></name> <name><surname>Clinton</surname> <given-names>J.</given-names></name> <name><surname>Dorff</surname> <given-names>C.</given-names></name> <name><surname>Kenkel</surname> <given-names>B.</given-names></name> <name><surname>Larson</surname> <given-names>J.</given-names></name></person-group> (<year>2023</year>). <article-title>Synthetic replacements for human survey data? The perils of large language models [internet]</article-title>. <source>SocArXiv.</source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://osf.io/5ecfa">https://osf.io/5ecfa</ext-link> (Accessed October 24, 2025).</mixed-citation>
</ref>
<ref id="B11">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Blair</surname> <given-names>P.</given-names></name> <name><surname>Buytaert</surname> <given-names>W.</given-names></name></person-group> (<year>2016</year>). Socio-hydrological modelling: a review asking &#x02018;why, what and how?&#x00027; Hydrol Earth Syst <italic>Sci</italic>. <volume>20</volume>, <fpage>443</fpage>&#x02013;<lpage>478</lpage>. doi: <pub-id pub-id-type="doi">10.5194/hess-20-443-2016</pub-id></mixed-citation>
</ref>
<ref id="B12">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Boelaert</surname> <given-names>J.</given-names></name> <name><surname>Coavoux</surname> <given-names>S.</given-names></name> <name><surname>Ollion</surname> <given-names>&#x000C9;.</given-names></name> <name><surname>Petev</surname> <given-names>I.</given-names></name> <name><surname>Pr&#x000E4;g</surname> <given-names>P.</given-names></name></person-group> (<year>2025</year>). <article-title>Machine bias. How do generative language models answer opinion polls?</article-title> <source>Sociol. Methods Res</source>. <volume>54</volume>, <fpage>1156</fpage>&#x02013;<lpage>1196</lpage>. doi: <pub-id pub-id-type="doi">10.1177/00491241251330582</pub-id></mixed-citation>
</ref>
<ref id="B13">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bogani</surname> <given-names>A.</given-names></name> <name><surname>Faccenda</surname> <given-names>G.</given-names></name> <name><surname>Riva</surname> <given-names>P.</given-names></name> <name><surname>Richetin</surname> <given-names>J.</given-names></name> <name><surname>Pancani</surname> <given-names>L.</given-names></name> <name><surname>Sacchi</surname> <given-names>S.</given-names></name></person-group> (<year>2023</year>). <article-title>The near-miss effect in flood risk estimation: a survey-based approach to model private mitigation intentions into agent-based models</article-title>. <source>Int. J. Disaster Risk Reduct. [Internet]</source> <volume>89</volume>:<fpage>103629</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ijdrr.2023.103629</pub-id></mixed-citation>
</ref>
<ref id="B14">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bubeck</surname> <given-names>P.</given-names></name> <name><surname>Thieken</surname> <given-names>A. H.</given-names></name></person-group> (<year>2018</year>). <article-title>What helps people recover from floods? Insights from a survey among flood-affected residents in Germany</article-title>. <source>Reg. Environ. Change</source> <volume>18</volume>, <fpage>287</fpage>&#x02013;<lpage>296</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s10113-017-1200-y</pub-id></mixed-citation>
</ref>
<ref id="B15">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Couldry</surname> <given-names>N.</given-names></name> <name><surname>Mejias</surname> <given-names>U. A.</given-names></name></person-group> (<year>2019</year>). <article-title>Data colonialism: rethinking big data&#x00027;s relation to the contemporary subject</article-title>. <source>Telev. New Media</source>. <volume>20</volume>, <fpage>336</fpage>&#x02013;<lpage>349</lpage>. doi: <pub-id pub-id-type="doi">10.1177/1527476418796632</pub-id></mixed-citation>
</ref>
<ref id="B16">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Cui</surname> <given-names>R.</given-names></name> <name><surname>Lee</surname> <given-names>S.</given-names></name> <name><surname>Hershcovich</surname> <given-names>D.</given-names></name> <name><surname>S&#x000F8;gaard</surname> <given-names>A.</given-names></name></person-group> (<year>2023</year>). <article-title>What does the failure to reason with &#x02018;respectively&#x00027; in zero/few-shot settings tell us about language models? [Internet]</article-title>. <source>arXiv.</source> Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2305.19597">http://arxiv.org/abs/2305.19597</ext-link> (Accessed October 31, 2025).</mixed-citation>
</ref>
<ref id="B17">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Devlin</surname> <given-names>J.</given-names></name> <name><surname>Chang</surname> <given-names>M. W.</given-names></name> <name><surname>Lee</surname> <given-names>K.</given-names></name> <name><surname>Toutanova</surname> <given-names>K.</given-names></name></person-group> (<year>2019</year>). <article-title>BERT: pre-training of deep bidirectional transformers for language understanding [internet]</article-title>. <source>arXiv</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/1810.04805">http://arxiv.org/abs/1810.04805</ext-link> (Accessed October 24, 2025).</mixed-citation>
</ref>
<ref id="B18">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Di Baldassarre</surname> <given-names>G.</given-names></name> <name><surname>Viglione</surname> <given-names>A.</given-names></name> <name><surname>Carr</surname> <given-names>G.</given-names></name> <name><surname>Kuil</surname> <given-names>L.</given-names></name> <name><surname>Yan</surname> <given-names>K.</given-names></name> <name><surname>Brandimarte</surname> <given-names>L.</given-names></name> <etal/></person-group>. (<year>2015</year>). <article-title>Debates&#x02014;perspectives on socio-hydrology: capturing feedbacks between physical and social processes</article-title>. <source>Water Resour. Res</source>. <volume>51</volume>, <fpage>4770</fpage>&#x02013;<lpage>4781</lpage>. doi: <pub-id pub-id-type="doi">10.1002/2014WR016416</pub-id></mixed-citation>
</ref>
<ref id="B19">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dillion</surname> <given-names>D.</given-names></name> <name><surname>Tandon</surname> <given-names>N.</given-names></name> <name><surname>Gu</surname> <given-names>Y.</given-names></name> <name><surname>Gray</surname> <given-names>K.</given-names></name></person-group> (<year>2023</year>). <article-title>Can AI language models replace human participants?</article-title> <source>Trends Cogn. Sci</source>. <volume>27</volume>, <fpage>597</fpage>&#x02013;<lpage>600</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.tics.2023.04.008</pub-id></mixed-citation>
</ref>
<ref id="B20">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Foroumandi</surname> <given-names>E.</given-names></name> <name><surname>Moradkhani</surname> <given-names>H.</given-names></name> <name><surname>Sanchez-Vila</surname> <given-names>X.</given-names></name> <name><surname>Singha</surname> <given-names>K.</given-names></name> <name><surname>Castelletti</surname> <given-names>A.</given-names></name> <name><surname>Destouni</surname> <given-names>G.</given-names></name></person-group> (<year>2023</year>). <article-title>ChatGPT in hydrology and earth sciences: opportunities, prospects, and concerns</article-title>. <source>Water Resour. Res</source>. <volume>59</volume>:e2023WR036288. doi: <pub-id pub-id-type="doi">10.1029/2023WR036288</pub-id></mixed-citation>
</ref>
<ref id="B21">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gain</surname> <given-names>A. K.</given-names></name> <name><surname>Hossain</surname> <given-names>M. S.</given-names></name> <name><surname>Benson</surname> <given-names>D.</given-names></name> <name><surname>Di Baldassarre</surname> <given-names>G.</given-names></name> <name><surname>Giupponi</surname> <given-names>C.</given-names></name> <name><surname>Huq</surname> <given-names>N.</given-names></name></person-group> (<year>2021</year>). <article-title>Social-ecological system approaches for water resources management</article-title>. <source>Int. J. Sustain. Dev. World Ecol</source>. <volume>28</volume>, <fpage>109</fpage>&#x02013;<lpage>124</lpage>. doi: <pub-id pub-id-type="doi">10.1080/13504509.2020.1780647</pub-id></mixed-citation>
</ref>
<ref id="B22">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Gao</surname> <given-names>C.</given-names></name> <name><surname>Lan</surname> <given-names>X.</given-names></name> <name><surname>Lu</surname> <given-names>Z.</given-names></name> <name><surname>Mao</surname> <given-names>J.</given-names></name> <name><surname>Piao</surname> <given-names>J.</given-names></name> <name><surname>Wang</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>S3: social-network simulation system with large language model-empowered agents [internet]</article-title>. <source>arXiv.</source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2307.14984">https://arxiv.org/abs/2307.14984</ext-link> (Accessed October 24, 2025). doi: <pub-id pub-id-type="doi">10.2139/ssrn.5604712</pub-id></mixed-citation>
</ref>
<ref id="B23">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Grossmann</surname> <given-names>I.</given-names></name> <name><surname>Feinberg</surname> <given-names>M.</given-names></name> <name><surname>Parker</surname> <given-names>D. C.</given-names></name> <name><surname>Christakis</surname> <given-names>N. A.</given-names></name> <name><surname>Tetlock</surname> <given-names>P. E.</given-names></name> <name><surname>Cunningham</surname> <given-names>W. A.</given-names></name></person-group> (<year>2023</year>). <article-title>AI and the transformation of social science research</article-title>. <source>Science [Internet]</source> <volume>380</volume>:<fpage>1108</fpage>. doi: <pub-id pub-id-type="doi">10.1126/science.adi1778</pub-id></mixed-citation>
</ref>
<ref id="B24">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Haer</surname> <given-names>T.</given-names></name> <name><surname>Husby</surname> <given-names>T. G.</given-names></name> <name><surname>Botzen</surname> <given-names>W. J. W.</given-names></name> <name><surname>Aerts</surname> <given-names>J. C. J. H.</given-names></name></person-group> (<year>2020</year>). <article-title>The safe development paradox: an agent-based model for flood risk under climate change in the European Union</article-title>. <source>Glob. Environ. Change Hum. Policy Dimens</source>. <volume>60</volume>:<fpage>102009</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.gloenvcha.2019.102009</pub-id></mixed-citation>
</ref>
<ref id="B25">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Halloran</surname> <given-names>L. J. S.</given-names></name> <name><surname>Mhanna</surname> <given-names>S.</given-names></name> <name><surname>Brunner</surname> <given-names>P.</given-names></name></person-group> (<year>2023</year>). <article-title>AI tools such as ChatGPT will disrupt hydrology, too</article-title>. <source>Hydrol. Process</source>. <volume>37</volume>:<fpage>e14843</fpage>. doi: <pub-id pub-id-type="doi">10.1002/hyp.14843</pub-id></mixed-citation>
</ref>
<ref id="B26">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>H&#x000E4;m&#x000E4;l&#x000E4;inen</surname> <given-names>P.</given-names></name> <name><surname>Tavast</surname> <given-names>M.</given-names></name> <name><surname>Kunnari</surname> <given-names>A.</given-names></name></person-group> (<year>2023</year>). <article-title>&#x0201C;Evaluating large language models in generating synthetic HCI research data: a case study,&#x0201D;</article-title> in <source>Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems [Internet]</source> (<publisher-loc>Hamburg Germany</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>19</lpage>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://dl.acm.org/doi/10.1145/3544548.3580688">https://dl.acm.org/doi/10.1145/3544548.3580688</ext-link> (Accessed November 1, 2025). doi: <pub-id pub-id-type="doi">10.1145/3544548.3580688</pub-id></mixed-citation>
</ref>
<ref id="B27">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Horton</surname> <given-names>J.</given-names></name></person-group> (<year>2023</year>). <source>Large Language Models as Simulated Economic Agents: What Can We Learn from Homo Silicus? [Internet]</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>National Bureau of Economic Research, w31122</publisher-name>. Report No.: w31122. Available online at: <ext-link ext-link-type="uri" xlink:href="http://www.nber.org/papers/w31122.pdf">http://www.nber.org/papers/w31122.pdf</ext-link> (Accessed October 24, 2025). doi: <pub-id pub-id-type="doi">10.3386/w31122</pub-id></mixed-citation>
</ref>
<ref id="B28">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hung</surname> <given-names>F.</given-names></name> <name><surname>Son</surname> <given-names>K.</given-names></name> <name><surname>Yang</surname> <given-names>Y. C. E.</given-names></name></person-group> (<year>2022</year>). <article-title>Investigating uncertainties in human adaptation and their impacts on water scarcity in the Colorado river Basin, United States</article-title>. <source>J. Hydrol. [Internet]</source> <volume>612</volume>:<fpage>128015</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.jhydrol.2022.128015</pub-id></mixed-citation>
</ref>
<ref id="B29">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Irvine</surname> <given-names>D. J.</given-names></name> <name><surname>Halloran</surname> <given-names>L. J. S.</given-names></name> <name><surname>Brunner</surname> <given-names>P.</given-names></name></person-group> (<year>2023</year>). <article-title>Opportunities and limitations of the ChatGPT advanced data analysis plugin for hydrological analyses</article-title>. <source>Hydrol. Process</source>. <volume>37</volume>:<fpage>e15015</fpage>. doi: <pub-id pub-id-type="doi">10.1002/hyp.15015</pub-id></mixed-citation>
</ref>
<ref id="B30">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jabkowski</surname> <given-names>P.</given-names></name> <name><surname>Cichocki</surname> <given-names>P.</given-names></name></person-group> (<year>2025</year>). <article-title>Survey response rates in European comparative surveys: a 20-year decline irrespective of sampling frames or survey modes</article-title>. <source>Qual. Quant</source>. <volume>59</volume>, <fpage>635</fpage>&#x02013;<lpage>655</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11135-024-01993-9</pub-id></mixed-citation>
</ref>
<ref id="B31">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kadiyala</surname> <given-names>L.</given-names></name> <name><surname>Sajja</surname> <given-names>R.</given-names></name> <name><surname>Sermet</surname> <given-names>Y.</given-names></name> <name><surname>Muste</surname> <given-names>M.</given-names></name> <name><surname>Demir</surname> <given-names>I.</given-names></name></person-group> (<year>2025</year>). <article-title>AI-driven decision-making for water resource planning and hazard mitigation using automated multi-agents</article-title>. <source>J. Hydroinform.</source> <volume>27</volume>, <fpage>1217</fpage>&#x02013;<lpage>1237</lpage>. doi: <pub-id pub-id-type="doi">10.2166/hydro.2025.042</pub-id></mixed-citation>
</ref>
<ref id="B32">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kadiyala</surname> <given-names>L. A.</given-names></name> <name><surname>Mermer</surname> <given-names>O.</given-names></name> <name><surname>Samuel</surname> <given-names>D. J.</given-names></name> <name><surname>Sermet</surname> <given-names>Y.</given-names></name> <name><surname>Demir</surname> <given-names>I.</given-names></name></person-group> (<year>2024</year>). <article-title>The implementation of multimodal large language models for hydrological applications: a comparative study of GPT-4 vision, Gemini, LLaVa, and multimodal-GPT</article-title>. <source>Hydrology</source><volume>11</volume>:<fpage>148</fpage>. doi: <pub-id pub-id-type="doi">10.3390/hydrology11090148</pub-id></mixed-citation>
</ref>
<ref id="B33">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Kaiya</surname> <given-names>Z.</given-names></name> <name><surname>Naim</surname> <given-names>M.</given-names></name> <name><surname>Kondic</surname> <given-names>J.</given-names></name> <name><surname>Cortes</surname> <given-names>M.</given-names></name> <name><surname>Ge</surname> <given-names>J.</given-names></name> <name><surname>Luo</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Lyfe agents: generative agents for low-cost real-time social interactions [internet]</article-title>. <source>arXiv</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2310.02172">http://arxiv.org/abs/2310.02172</ext-link> (Accessed November 5, 2025).</mixed-citation>
</ref>
<ref id="B34">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Kim</surname> <given-names>J.</given-names></name> <name><surname>Lee</surname> <given-names>B.</given-names></name></person-group> (<year>2024</year>). <article-title>AI-augmented surveys: leveraging large language models and surveys for opinion prediction [internet]</article-title>. <source>arXiv</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2305.09620">http://arxiv.org/abs/2305.09620</ext-link> (Accessed October 24, 2025).</mixed-citation>
</ref>
<ref id="B35">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Kotek</surname> <given-names>H.</given-names></name> <name><surname>Dockum</surname> <given-names>R.</given-names></name> <name><surname>Sun</surname> <given-names>D.</given-names></name></person-group> (<year>2023</year>). <article-title>&#x0201C;Gender bias and stereotypes in large language models,&#x0201D;</article-title> in <source>Proceedings of The ACM Collective Intelligence Conference [Internet]</source> (<publisher-loc>Delft Netherlands</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>12</fpage>&#x02013;<lpage>24</lpage>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://dl.acm.org/doi/10.1145/3582269.3615599">https://dl.acm.org/doi/10.1145/3582269.3615599</ext-link>. doi: <pub-id pub-id-type="doi">10.1145/3582269.3615599</pub-id> (Accessed November 7, 2025).</mixed-citation>
</ref>
<ref id="B36">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kozlowski</surname> <given-names>A. C.</given-names></name> <name><surname>Evans</surname> <given-names>J.</given-names></name></person-group> (<year>2025</year>). <article-title>Simulating subjects: the promise and peril of artificial intelligence stand-ins for social agents and interactions</article-title>. <source>Sociol. Methods Res</source>. <volume>54</volume>, <fpage>1017</fpage>&#x02013;<lpage>1073</lpage>. doi: <pub-id pub-id-type="doi">10.1177/00491241251337316</pub-id></mixed-citation>
</ref>
<ref id="B37">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kreibich</surname> <given-names>H.</given-names></name> <name><surname>Sivapalan</surname> <given-names>M.</given-names></name> <name><surname>AghaKouchak A</surname> <given-names>Addor N, Aksoy H, Arheimer B</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Panta Rhei: a decade of progress in research on change in hydrology and society</article-title>. <source>Hydrol. Sci. J</source>. <volume>70</volume>, <fpage>1210</fpage>&#x02013;<lpage>1236</lpage>. doi: <pub-id pub-id-type="doi">10.1080/02626667.2025.2469762</pub-id></mixed-citation>
</ref>
<ref id="B38">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Kumbam</surname> <given-names>P. R.</given-names></name> <name><surname>Vejre</surname> <given-names>K. M.</given-names></name></person-group> (<year>2024</year>). <article-title>FloodLense: a framework for ChatGPT-based real-time flood detection [internet]</article-title>. <source>arXiv</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2401.15501">http://arxiv.org/abs/2401.15501</ext-link> (Accessed November 5, 2025).</mixed-citation>
</ref>
<ref id="B39">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ma</surname> <given-names>Z.</given-names></name> <name><surname>Peng</surname> <given-names>B.</given-names></name> <name><surname>Yue</surname> <given-names>Z.</given-names></name> <name><surname>Zeng</surname> <given-names>H.</given-names></name> <name><surname>Pan</surname> <given-names>M.</given-names></name> <name><surname>Wu</surname> <given-names>X.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Embracing large language model (LLM) technologies in hydrology research</article-title>. <source>Environ. Res. Water</source> <volume>1</volume>:<fpage>022001</fpage>. doi: <pub-id pub-id-type="doi">10.1088/3033-4942/addd43</pub-id></mixed-citation>
</ref>
<ref id="B40">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Madruga de Brito</surname> <given-names>M.</given-names></name> <name><surname>Kuhlicke</surname> <given-names>C.</given-names></name> <name><surname>Marx</surname> <given-names>A.</given-names></name></person-group> (<year>2020</year>). <article-title>Near-real-time drought impact assessment: a text mining approach on the 2018/19 drought in Germany</article-title>. <source>Environ. Res. Lett</source>. <volume>15</volume>:<fpage>1040a</fpage>9. doi: <pub-id pub-id-type="doi">10.1088/1748-9326/aba4ca</pub-id></mixed-citation>
</ref>
<ref id="B41">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mei</surname> <given-names>Q.</given-names></name> <name><surname>Xie</surname> <given-names>Y.</given-names></name> <name><surname>Yuan</surname> <given-names>W.</given-names></name> <name><surname>Jackson</surname> <given-names>M. O.</given-names></name></person-group> (<year>2024</year>). <article-title>A turing test of whether AI chatbots are behaviorally similar to humans</article-title>. <source>Proc. Natl. Acad. Sci</source>. <volume>121</volume>:<fpage>e2313925121</fpage>. doi: <pub-id pub-id-type="doi">10.1073/pnas.2313925121</pub-id></mixed-citation>
</ref>
<ref id="B42">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Miao</surname> <given-names>C.</given-names></name> <name><surname>Hu</surname> <given-names>J.</given-names></name> <name><surname>Moradkhani</surname> <given-names>H.</given-names></name> <name><surname>Destouni</surname> <given-names>G.</given-names></name></person-group> (<year>2024</year>). <article-title>Hydrological research evolution: a large language model-based analysis of 310,000 studies published globally between 1980 and 2023</article-title>. <source>Water Resour. Res</source>. <volume>60</volume>:<fpage>e2024W</fpage>R038077. doi: <pub-id pub-id-type="doi">10.1029/2024WR038077</pub-id></mixed-citation>
</ref>
<ref id="B43">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Milgram</surname> <given-names>S.</given-names></name></person-group> (<year>1963</year>). <article-title>Behavioral study of obedience</article-title>. <source>J. Abnorm. Soc. Psychol</source>. <volume>67</volume>, <fpage>371</fpage>&#x02013;<lpage>378</lpage>. doi: <pub-id pub-id-type="doi">10.1037/h0040525</pub-id></mixed-citation>
</ref>
<ref id="B44">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Murthy</surname> <given-names>D.</given-names></name></person-group> (<year>2024</year>). <article-title>Sociology of Twitter/X: trends, challenges, and future research directions</article-title>. <source>Annu. Rev. Sociol</source>. <volume>50</volume>, <fpage>169</fpage>&#x02013;<lpage>190</lpage>. doi: <pub-id pub-id-type="doi">10.1146/annurev-soc-031021-035658</pub-id></mixed-citation>
</ref>
<ref id="B45">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nakamura</surname> <given-names>S.</given-names></name> <name><surname>Kreibich</surname> <given-names>H.</given-names></name> <name><surname>Haeffner</surname> <given-names>M.</given-names></name> <name><surname>Mukherjee</surname> <given-names>J.</given-names></name> <name><surname>Di Baldassarre</surname> <given-names>G.</given-names></name> <name><surname>Sakamoto</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Expanding and mainstreaming sociohydrology toward transdisciplinary praxis</article-title>. <source>Front. Water [Internet]</source> <volume>7</volume>:<fpage>1629905</fpage>. doi: <pub-id pub-id-type="doi">10.3389/frwa.2025.1629905</pub-id></mixed-citation>
</ref>
<ref id="B46">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pande</surname> <given-names>S.</given-names></name> <name><surname>Sivapalan</surname> <given-names>M.</given-names></name></person-group> (<year>2017</year>). <article-title>Progress in socio-hydrology: a meta-analysis of challenges and opportunities</article-title>. <source>WIREs Water</source> <volume>4</volume>:<fpage>e1193</fpage>. doi: <pub-id pub-id-type="doi">10.1002/wat2.1193</pub-id></mixed-citation>
</ref>
<ref id="B47">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Park</surname> <given-names>J. S.</given-names></name> <name><surname>O&#x00027;Brien</surname> <given-names>J.</given-names></name> <name><surname>Cai</surname> <given-names>C. J.</given-names></name> <name><surname>Morris</surname> <given-names>M. R.</given-names></name> <name><surname>Liang</surname> <given-names>P.</given-names></name> <name><surname>Bernstein</surname> <given-names>M. S.</given-names></name></person-group> (<year>2023</year>). <article-title>&#x0201C;Generative agents: interactive simulacra of human behavior,&#x0201D;</article-title> in <source>Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology [Internet]</source> (<publisher-loc>UIST&#x00027;23</publisher-loc>) (New York, NY, USA: Association for Computing Machinery). doi: <pub-id pub-id-type="doi">10.1145/3586183.3606763</pub-id></mixed-citation>
</ref>
<ref id="B48">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Perrigo</surname> <given-names>B.</given-names></name></person-group> (<year>2023</year>). <article-title>Exclusive: the $2 per hour workers who made ChatGPT safer</article-title>. <source>TIME [Internet]</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://time.com/6247678/openai-chatgpt-kenya-workers/">https://time.com/6247678/openai-chatgpt-kenya-workers/</ext-link> (Accessed November 9, 2025).</mixed-citation>
</ref>
<ref id="B49">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rahman</surname> <given-names>N.</given-names></name> <name><surname>Santacana</surname> <given-names>E.</given-names></name></person-group> (<year>2023</year>). <article-title>&#x0201C;Beyond fair use: legal risk evaluation for training LLMs on copyrighted text,&#x0201D;</article-title> in <source>ICML Workshop on Generative AI and Law</source> (Accessed November 9, 2025).</mixed-citation>
</ref>
<ref id="B50">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Raiaan</surname> <given-names>M. A. K.</given-names></name> <name><surname>Mukta</surname> <given-names>M. S. H.</given-names></name> <name><surname>Fatema</surname> <given-names>K.</given-names></name> <name><surname>Fahad</surname> <given-names>N. M.</given-names></name> <name><surname>Sakib</surname> <given-names>S.</given-names></name> <name><surname>Mim</surname> <given-names>M. M. J.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>A review on large language models: architectures, applications, taxonomies, open issues and challenges</article-title>. <source>IEEE Access</source>. <volume>12</volume>, <fpage>26839</fpage>&#x02013;<lpage>26874</lpage>. doi: <pub-id pub-id-type="doi">10.1109/ACCESS.2024.3365742</pub-id></mixed-citation>
</ref>
<ref id="B51">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ren</surname> <given-names>Y.</given-names></name> <name><surname>Zhang</surname> <given-names>T.</given-names></name> <name><surname>Dong</surname> <given-names>X.</given-names></name> <name><surname>Li</surname> <given-names>W.</given-names></name> <name><surname>Wang</surname> <given-names>Z.</given-names></name> <name><surname>He</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>WaterGPT: training a large language model to become a hydrology expert</article-title>. <source>Water</source> <volume>16</volume>:<fpage>3075</fpage>. doi: <pub-id pub-id-type="doi">10.3390/w16213075</pub-id></mixed-citation>
</ref>
<ref id="B52">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rillig</surname> <given-names>M. C.</given-names></name> <name><surname>&#x000C5;gerstrand</surname> <given-names>M.</given-names></name> <name><surname>Bi</surname> <given-names>M.</given-names></name> <name><surname>Gould</surname> <given-names>K. A.</given-names></name> <name><surname>Sauerland</surname> <given-names>U.</given-names></name></person-group> (<year>2023</year>). <article-title>Risks and benefits of large language models for the environment</article-title>. <source>Environ. Sci. Technol</source>. <volume>57</volume>, <fpage>3464</fpage>&#x02013;<lpage>3466</lpage>. doi: <pub-id pub-id-type="doi">10.1021/acs.est.3c01106</pub-id></mixed-citation>
</ref>
<ref id="B53">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rossi</surname> <given-names>L.</given-names></name> <name><surname>Harrison</surname> <given-names>K.</given-names></name> <name><surname>Shklovski</surname> <given-names>I.</given-names></name></person-group> (<year>2024</year>). <article-title>The problems of LLM-generated data in social science research</article-title>. <source>Sociologica</source> <volume>18</volume>, <fpage>145</fpage>&#x02013;<lpage>168</lpage>. doi: <pub-id pub-id-type="doi">10.6092/issn.1971-8853/19576</pub-id></mixed-citation>
</ref>
<ref id="B54">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Salganik</surname> <given-names>M. J.</given-names></name></person-group> (<year>2019</year>). <source>Bit by Bit: Social Research in the Digital Age. First Paperback Printing</source>. <publisher-loc>Princeton, Oxford</publisher-loc>: <publisher-name>Princeton University Press, 423</publisher-name>.</mixed-citation>
</ref>
<ref id="B55">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Santurkar</surname> <given-names>S.</given-names></name> <name><surname>Durmus</surname> <given-names>E.</given-names></name> <name><surname>Ladhak</surname> <given-names>F.</given-names></name> <name><surname>Lee</surname> <given-names>C.</given-names></name> <name><surname>Liang</surname> <given-names>P.</given-names></name> <name><surname>Hashimoto</surname> <given-names>T.</given-names></name></person-group> (<year>2023</year>). <article-title>Whose opinions do language models reflect? [Internet]</article-title>. <source>arXiv</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2303.17548">http://arxiv.org/abs/2303.17548</ext-link> (Accessed November 7, 2025).</mixed-citation>
</ref>
<ref id="B56">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Savelli</surname> <given-names>E.</given-names></name> <name><surname>Rusca</surname> <given-names>M.</given-names></name> <name><surname>Cloke</surname> <given-names>H.</given-names></name> <name><surname>Di Baldassarre</surname> <given-names>G.</given-names></name></person-group> (<year>2021</year>). <article-title>Don&#x00027;t blame the rain: social power and the 2015&#x02013;2017 drought in Cape Town</article-title>. <source>J. Hydrol</source>. <volume>594</volume>:<fpage>125953</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.jhydrol.2020.125953</pub-id></mixed-citation>
</ref>
<ref id="B57">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sch&#x000FC;ck</surname> <given-names>F.</given-names></name> <name><surname>Arheimer</surname> <given-names>B.</given-names></name> <name><surname>Mazzoleni</surname> <given-names>M.</given-names></name> <name><surname>Brandimarte</surname> <given-names>L.</given-names></name></person-group> (<year>2025</year>). <article-title>A systematic mapping review of hydrological hazard management in agent-based systems</article-title>. <source>Environ. Res. Lett</source>. <volume>20</volume>:<fpage>113003</fpage>. doi: <pub-id pub-id-type="doi">10.1088/1748-9326/ae0fb2</pub-id></mixed-citation>
</ref>
<ref id="B58">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Schwitter</surname> <given-names>N.</given-names></name> <name><surname>Pretari</surname> <given-names>A.</given-names></name> <name><surname>Marwa</surname> <given-names>W.</given-names></name> <name><surname>Lombardini</surname> <given-names>S.</given-names></name> <name><surname>Liebe</surname> <given-names>U.</given-names></name></person-group> (<year>2022</year>). <article-title>Big data and development sociology: an overview and application on governance and accountability through digitalization in Tanzania</article-title>. <source>Front. Sociol. [Internet]</source> <volume>7</volume>:<fpage>909458</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fsoc.2022.909458</pub-id></mixed-citation>
</ref>
<ref id="B59">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sivapalan</surname> <given-names>M.</given-names></name> <name><surname>Savenije</surname> <given-names>H. H. G.</given-names></name> <name><surname>Bl&#x000F6;schl</surname> <given-names>G.</given-names></name></person-group> (<year>2012</year>). <article-title>Socio-hydrology: a new science of people and water</article-title>. <source>Hydrol. Process</source>. <volume>26</volume>, <fpage>1270</fpage>&#x02013;<lpage>1276</lpage>. doi: <pub-id pub-id-type="doi">10.1002/hyp.8426</pub-id></mixed-citation>
</ref>
<ref id="B60">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Spitale</surname> <given-names>G.</given-names></name> <name><surname>Biller-Andorno</surname> <given-names>N.</given-names></name> <name><surname>Germani</surname> <given-names>F.</given-names></name></person-group> (<year>2023</year>). <article-title>AI model GPT-3 (dis)informs us better than humans</article-title>. <source>Sci. Adv.</source> <volume>9</volume>:<fpage>eadh1850</fpage>. doi: <pub-id pub-id-type="doi">10.1126/sciadv.adh1850</pub-id></mixed-citation>
</ref>
<ref id="B61">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Stevenson</surname> <given-names>C. E.</given-names></name> <name><surname>Pafford</surname> <given-names>A.</given-names></name> <name><surname>van der Maas</surname> <given-names>H. L. J.</given-names></name> <name><surname>Mitchell</surname> <given-names>M.</given-names></name></person-group> (<year>2025</year>). <article-title>Can large language models generalize analogy solving like children can? [Internet]</article-title>. <source>arXiv</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2411.02348">http://arxiv.org/abs/2411.02348</ext-link> (Accessed October 31, 2025).</mixed-citation>
</ref>
<ref id="B62">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Stokel-Walker</surname> <given-names>C.</given-names></name> <name><surname>Van Noorden</surname> <given-names>R.</given-names></name></person-group> (<year>2023</year>). <article-title>What ChatGPT and generative AI mean for science</article-title>. <source>Nature</source> <volume>614</volume>, <fpage>214</fpage>&#x02013;<lpage>216</lpage>. doi: <pub-id pub-id-type="doi">10.1038/d41586-023-00340-6</pub-id></mixed-citation>
</ref>
<ref id="B63">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sun</surname> <given-names>Z.</given-names></name> <name><surname>Lorscheid</surname> <given-names>I.</given-names></name> <name><surname>Millington</surname> <given-names>J. D.</given-names></name> <name><surname>Lauf</surname> <given-names>S.</given-names></name> <name><surname>Magliocca</surname> <given-names>N. R.</given-names></name> <name><surname>Groeneveld</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2016</year>).Simple or complicated agent-based models? A complicated issue. <source>Environ. Model Softw</source>. <volume>86</volume>, <fpage>56</fpage>&#x02013;<lpage>67</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.envsoft.2016.09.006</pub-id></mixed-citation>
</ref>
<ref id="B64">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Taberna</surname> <given-names>A.</given-names></name> <name><surname>Filatova</surname> <given-names>T.</given-names></name> <name><surname>Hadjimichael</surname> <given-names>A.</given-names></name> <name><surname>Noll</surname> <given-names>B.</given-names></name></person-group> (<year>2023</year>). <article-title>Uncertainty in boundedly rational household adaptation to environmental shocks</article-title>. <source>Proc. Natl. Acad. Sci. U. S. A. [Internet]</source> <volume>120</volume>:<fpage>2215675120</fpage>. doi: <pub-id pub-id-type="doi">10.1073/pnas.2215675120</pub-id></mixed-citation>
</ref>
<ref id="B65">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>T&#x000F6;rnberg</surname> <given-names>P.</given-names></name> <name><surname>Valeeva</surname> <given-names>D.</given-names></name> <name><surname>Uitermark</surname> <given-names>J.</given-names></name> <name><surname>Bail</surname> <given-names>C.</given-names></name></person-group> (<year>2023</year>). <article-title>Simulating social media using large language models to evaluate alternative news feed algorithms [internet]</article-title>. <source>arXiv</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2310.05984">http://arxiv.org/abs/2310.05984</ext-link> (Accessed Septembet 29, 2025).</mixed-citation>
</ref>
<ref id="B66">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Trezza</surname> <given-names>D.</given-names></name></person-group> (<year>2023</year>). <article-title>To scrape or not to scrape, this is dilemma. The post-API scenario and implications on digital research</article-title>. <source>Front. Sociol. [Internet]</source> <volume>8</volume>:<fpage>1145038</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fsoc.2023.1145038</pub-id></mixed-citation>
</ref>
<ref id="B67">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Vaswani</surname> <given-names>A.</given-names></name> <name><surname>Shazeer</surname> <given-names>N.</given-names></name> <name><surname>Parmar</surname> <given-names>N.</given-names></name> <name><surname>Uszkoreit</surname> <given-names>J.</given-names></name> <name><surname>Jones</surname> <given-names>L.</given-names></name> <name><surname>Gomez</surname> <given-names>A. N.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>Attention is all you need [internet]</article-title>. <source>arXiv</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/1706.03762">http://arxiv.org/abs/1706.03762</ext-link> (Accessed October 24, 2025). doi: <pub-id pub-id-type="doi">10.65215/ysbyhc05</pub-id></mixed-citation>
</ref>
<ref id="B68">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Veigel</surname> <given-names>N.</given-names></name> <name><surname>Kreibich</surname> <given-names>H.</given-names></name> <name><surname>de Bruijn</surname> <given-names>J. A.</given-names></name> <name><surname>Aerts</surname> <given-names>J. C. J. H.</given-names></name> <name><surname>Cominola</surname> <given-names>A.</given-names></name></person-group> (<year>2025</year>). <article-title>Content analysis of multi-annual time series of flood-related Twitter (X) data</article-title>. <source>Nat. Hazards Earth Syst. Sci</source>. <volume>25</volume>, <fpage>879</fpage>&#x02013;<lpage>891</lpage>. doi: <pub-id pub-id-type="doi">10.5194/nhess-25-879-2025</pub-id></mixed-citation>
</ref>
<ref id="B69">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>A.</given-names></name> <name><surname>Morgenstern</surname> <given-names>J.</given-names></name> <name><surname>Dickerson</surname> <given-names>J. P.</given-names></name></person-group> (<year>2025</year>). <article-title>Large language models that replace human participants can harmfully misportray and flatten identity groups</article-title>. <source>Nat. Mach. Intell</source>. <volume>7</volume>, <fpage>400</fpage>&#x02013;<lpage>411</lpage>. doi: <pub-id pub-id-type="doi">10.1038/s42256-025-00986-z</pub-id></mixed-citation>
</ref>
<ref id="B70">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Z.</given-names></name> <name><surname>Chu</surname> <given-names>Z.</given-names></name> <name><surname>Doan</surname> <given-names>T. V.</given-names></name> <name><surname>Ni</surname> <given-names>S.</given-names></name> <name><surname>Yang</surname> <given-names>M.</given-names></name> <name><surname>Zhang</surname> <given-names>W.</given-names></name></person-group> (<year>2024</year>). <article-title>History, development, and principles of large language models-an introductory survey [internet]</article-title>. <source>arXiv</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2402.06853">http://arxiv.org/abs/2402.06853</ext-link> (Accessed October 24, 2025).</mixed-citation>
</ref>
<ref id="B71">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Weidinger</surname> <given-names>L.</given-names></name> <name><surname>Uesato</surname> <given-names>J.</given-names></name> <name><surname>Rauh</surname> <given-names>M.</given-names></name> <name><surname>Griffin</surname> <given-names>C.</given-names></name> <name><surname>Huang</surname> <given-names>P. S.</given-names></name> <name><surname>Mellor</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>&#x0201C;Taxonomy of risks posed by language models,&#x0201D;</article-title> in <source>2022 ACM Conference on Fairness Accountability and Transparency [Internet]</source> (<publisher-loc>Seoul Republic of Korea</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>214</fpage>&#x02013;<lpage>229</lpage>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://dl.acm.org/doi/10.1145/3531146.3533088">https://dl.acm.org/doi/10.1145/3531146.3533088</ext-link> (Accessed November 9, 2025). doi: <pub-id pub-id-type="doi">10.1145/3531146.3533088</pub-id></mixed-citation>
</ref>
<ref id="B72">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wens</surname> <given-names>M.</given-names></name> <name><surname>Johnson</surname> <given-names>J.</given-names></name> <name><surname>Zagaria</surname> <given-names>C.</given-names></name> <name><surname>Veldkamp</surname> <given-names>T.</given-names></name></person-group> (<year>2019</year>). <article-title>Integrating human behavior dynamics into drought risk assessment-a sociohydrologic, agent-based approach</article-title>. <source>WILEY Interdiscip. Rev. WATER</source> <volume>6</volume>:<fpage>e1345</fpage>. doi: <pub-id pub-id-type="doi">10.1002/wat2.1345</pub-id></mixed-citation>
</ref>
<ref id="B73">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Widder</surname> <given-names>D. G.</given-names></name> <name><surname>Whittaker</surname> <given-names>M.</given-names></name> <name><surname>West</surname> <given-names>S. M.</given-names></name></person-group> (<year>2024</year>). <article-title>Why &#x02018;open&#x00027; AI systems are actually closed, and why this matters</article-title>. <source>Nature</source> <volume>635</volume>, <fpage>827</fpage>&#x02013;<lpage>833</lpage>. doi: <pub-id pub-id-type="doi">10.1038/s41586-024-08141-1</pub-id></mixed-citation>
</ref>
<ref id="B74">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wilkinson</surname> <given-names>M. D.</given-names></name> <name><surname>Dumontier</surname> <given-names>M.</given-names></name> <name><surname>Aalbersberg</surname> <given-names>I. J.</given-names></name> <name><surname>Appleton</surname> <given-names>G.</given-names></name> <name><surname>Axton</surname> <given-names>M.</given-names></name> <name><surname>Baak</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>The FAIR guiding principles for scientific data management and stewardship</article-title>. <source>Sci. Data</source>. <volume>3</volume>:<fpage>160018</fpage>. doi: <pub-id pub-id-type="doi">10.1038/sdata.2016.18</pub-id></mixed-citation>
</ref>
<ref id="B75">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Wooldridge</surname> <given-names>M.</given-names></name></person-group> (<year>2009</year>). <source>An Introduction to Multiagent Systems</source>. <publisher-loc>Chichester</publisher-loc>: <publisher-name>John Wiley Sons</publisher-name>.</mixed-citation>
</ref>
<ref id="B76">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>B.</given-names></name> <name><surname>Li</surname> <given-names>Z.</given-names></name> <name><surname>Yang</surname> <given-names>Y.</given-names></name> <name><surname>Wu</surname> <given-names>G.</given-names></name> <name><surname>Wang</surname> <given-names>C.</given-names></name> <name><surname>Tang</surname> <given-names>X.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Evaluating and advancing large language models for water knowledge tasks in engineering and research</article-title>. <source>Environ. Sci. Technol. Lett</source>. <volume>12</volume>, <fpage>289</fpage>&#x02013;<lpage>296</lpage>. doi: <pub-id pub-id-type="doi">10.1021/acs.estlett.5c00038</pub-id></mixed-citation>
</ref>
<ref id="B77">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>Khan</surname> <given-names>S. A.</given-names></name> <name><surname>Mahmud</surname> <given-names>A.</given-names></name> <name><surname>Yang</surname> <given-names>H.</given-names></name> <name><surname>Lavin</surname> <given-names>A.</given-names></name> <name><surname>Levin</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Exploring the role of large language models in the scientific method: from hypothesis to discovery</article-title>. <source>NPJ Artif. Intell</source>. <volume>1</volume>:<fpage>14</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s44387-025-00019-5</pub-id></mixed-citation>
</ref>
<ref id="B78">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>M.</given-names></name> <name><surname>Yu</surname> <given-names>L.</given-names></name> <name><surname>Geng</surname> <given-names>X.</given-names></name> <name><surname>Luo</surname> <given-names>L.</given-names></name></person-group> (<year>2025</year>). <article-title>ChatGPT vs social surveys: probing objective and subjective silicon population [internet]</article-title>. <source>arXiv</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2409.02601">http://arxiv.org/abs/2409.02601</ext-link> (Accessed November 1, 2025).</mixed-citation>
</ref>
<ref id="B79">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhu</surname> <given-names>J.</given-names></name> <name><surname>Dang</surname> <given-names>P.</given-names></name> <name><surname>Cao</surname> <given-names>Y.</given-names></name> <name><surname>Lai</surname> <given-names>J.</given-names></name> <name><surname>Guo</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>A flood knowledge-constrained large language model interactable with GIS: enhancing public risk perception of floods</article-title>. <source>Int. J. Geogr. Inf. Sci</source>. <volume>38</volume>, <fpage>603</fpage>&#x02013;<lpage>625</lpage>. doi: <pub-id pub-id-type="doi">10.1080/13658816.2024.2306167</pub-id></mixed-citation>
</ref>
<ref id="B80">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ziems</surname> <given-names>C.</given-names></name> <name><surname>Held</surname> <given-names>W.</given-names></name> <name><surname>Shaikh</surname> <given-names>O.</given-names></name> <name><surname>Chen</surname> <given-names>J.</given-names></name> <name><surname>Zhang</surname> <given-names>Z.</given-names></name> <name><surname>Yang</surname> <given-names>D.</given-names></name></person-group> (<year>2024</year>). <article-title>Can large language models transform computational social science?</article-title> <source>Comput. Linguist</source>. <volume>50</volume>, <fpage>237</fpage>&#x02013;<lpage>291</lpage>. doi: <pub-id pub-id-type="doi">10.1162/coli_a_00502</pub-id></mixed-citation>
</ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1919153/overview">Mariana Madruga de Brito</ext-link>, Helmholtz Association of German Research Centres (HZ), Germany</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0002">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/720372/overview">Adeyemi Oludapo Olusola</ext-link>, York University, Canada</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1710921/overview">Bruce Lankford</ext-link>, University of East Anglia, United Kingdom</p>
</fn>
</fn-group>
</back>
</article>