http://dx.doi.org/10.24016/2026.v12.497

EDITORIAL

 

 

Ethical and Regulatory Gaps in Using Generative AI for Mental Health Support in Low- and Middle-Income Countries

 

Leonardo Rojas-Mezarina1 *, David Villarreal-Zegarra2

1 Facultad de Medicina, Universidad Nacional Mayor de San Marcos, Lima, Peru.

2 Digital Health Research Center, Instituto Peruano de Orientación Psicológica, Lima, Peru.

 

* Correspondence: leonardo.rojas@unmsm.edu.pe.

 

Received: December 13, 2025 | Reviewed: December 27, 2025 | Accepted: January 02, 2026 | Published Online: January 02, 2026.

 

CITE IT AS:

Rojas-Mezarina, L., Villarreal-Zegarra, D. (2026). Ethical and Regulatory Gaps in Using Generative AI for Mental Health Support in Low- and Middle-Income Countries. Interacciones, 12, e497. http://dx.doi.org/10.24016/2026.v12.497

 

 

The accelerated adoption of generative artificial intelligence (AI) models, such as ChatGPT and Gemini, as well as other conversational agents, has transformed how people worldwide seek mental health information and support (Thirunavukarasu et al., 2023). These large language models (LLMs) are being used at massive scale; for example, ChatGPT alone has been reported to have hundreds of millions of weekly active users (Chatterji et al., 2025). Users interact with these systems to receive guidance related to anxiety, depression, or crisis situations, marking an unprecedented shift in the digital health ecosystem (Ayers et al., 2023). However, while generative AI promises to expand access to physical and mental health resources, it also introduces ethical and regulatory risks that remain insufficiently addressed (Meskó & Topol, 2023), particularly in low- and middle-income regions such as Latin America. In these settings, AI models developed in high-income countries are widely deployed without necessarily assessing the potential risks of bias that this entails (Hussain et al., 2025).

In low- and middle-income countries (LMICs), the governance architecture for health-related generative artificial intelligence, encompassing standards, accountability, transparency, and enforceable data protection, lags behind its real-world implementation. We observe that AI-based systems are increasingly integrated into daily life without adequate standards for safety, transparency, or data protection (Morley et al., 2020). The risks arising from their therapeutic or quasi-therapeutic use in mental health therefore warrant urgent examination.

First, using LLMs for mental health support entails processing intimate and highly sensitive information, including symptoms, trauma narratives, medication histories, and crisis-related disclosures (Mandal et al., 2025; Wang et al., 2025). These interactions can also generate sensitive inferences (e.g., suicide risk, substance use, or exposure to abuse) even when users do not explicitly disclose them, increasing the potential for privacy harms if data are mishandled.

Second, the technological infrastructure that enables these services is commonly located outside the jurisdictions of LMICs, under privacy policies that permit the use, storage, and training on personal user data (Vollmer et al., 2020). This extraterritoriality complicates enforcement and redress mechanisms and weakens cross-border accountability, particularly where local regulatory agencies have limited technical capacity or unclear legal authority over foreign providers.

Third, foundational model development and data management remain opaque, including uncertainty regarding the provenance of training corpora, data governance practices, and safeguards to meet expectations of medical confidentiality (Bommasani et al., 2023). The “black box” nature of these systems also complicates auditability and post hoc investigation when harmful outputs occur, limiting effective oversight (Ethical AI governance group, 2023).

In the Peruvian case, the Law on Personal Data Protection (Law No. 29733) is insufficient to address emerging generative AI risks because it does not encompass critical aspects such as sensitive inferences, algorithmic reuse, or re-identification risks (Smart & Montori, 2025). This gap is particularly salient because, while Peru’s data protection framework shares broad intent with comprehensive regimes such as the EU’s GDPR, it is not directly comparable to sector-specific U.S. frameworks such as HIPAA and does not yet address AI-specific risks (e.g., sensitive inferences, algorithmic reuse, and re-identification). As a result, users may be exposed to privacy violations with emotional, clinical, and societal consequences.

Generative AI models are predominantly trained on data in English and within Western sociocultural contexts, and may generate erroneous or inaccurate responses when used in non-English or racially diverse populations (Omiye et al., 2023). Furthermore, several language models are optimized for English tokenization, which may result in lower performance when interacting with languages such as Spanish or Portuguese. This has direct implications for populations in LMICs, where cultural, linguistic, and socioeconomic factors deeply influence the experience and expression of mental health issues. Previous studies show that AI models can produce biased, culturally inappropriate, or clinically incorrect responses, reinforcing existing inequities in marginalized groups, such as Afro-descendant populations or those in poverty (Cross et al., 2024; Omiye et al., 2023). For example, an AI-generated response might misinterpret local idioms associated with emotional suffering or provide recommendations that overlook the structural realities of unequal access to healthcare services. This is especially relevant in LMICs, where the social and cultural determinants of mental health are complex. These biases can amplify disparities and affect the quality of support received (Ahluwalia et al., 2025). This relates to the principle of health equity, reminding us that no technology is inherently neutral or universal.

The expansion of generative AI use in mental health also raises questions regarding legal liability. Who is responsible for harm derived from a potentially dangerous, incomplete, or erroneous recommendation? This debate has intensified following recent lawsuits against AI providers for adverse outcomes associated with their responses. For example, documented harms include responses that promote discrimination, hate speech, or exclusion; harms arising from misinformation or malicious uses; harms related to human-computer interaction; and environmental or socioeconomic harms (Weidinger et al., 2022).

At the clinical level, available evidence is limited. Although some exploratory studies suggest that AI-based chatbots can support psychological interventions (Baek et al., 2025; Li et al., 2023; Vaidyam et al., 2019; Villarreal-Zegarra et al., 2024), few of these generative systems have been evaluated through large, rigorous clinical trials that support their efficacy or safety, as most sample sizes are small and follow-up times are short (Li et al., 2023). Furthermore, the structural "black box" problem of generative AI models hinders a clear understanding of how responses are generated, making it challenging to assess risks, validate recommendations, and ensure therapeutic coherence (Ethical AI governance group, 2023).

Without robust data derived from clinical studies and without clear legislative mechanisms for clinical and regulatory oversight, the integration of generative AI into mental health practices may pose more risk than benefit. This is particularly relevant because many LMICs lack governance frameworks adapted to generative AI (Smart & Montori, 2025; Stanford Center for Digital Health, 2025). Specifically in Latin America, the absence of specific legislation on AI in healthcare, combined with the lack of local clinical trials or studies that include Latin American populations, creates a context where the risks outweigh the potential benefits.

This situation demands coordinated actions among multiple actors. In this letter to the editor, we call upon researchers, AI technology developers, and policymakers. First, research teams must prioritize validation studies in real-world contexts, using heterogeneous samples that represent underrepresented populations, including clinical trials, evaluations of cultural bias, and analyses of unintended effects. This will ensure that the performance of AI-based applications in healthcare is sufficiently robust and minimizes biases that can generate inequities. Second, developers must incorporate ethical principles into the design of these applications, guaranteeing transparency, traceability of algorithmic decisions, and safety safeguards in crisis scenarios. Third, regulators and public health authorities must develop specific guidelines for generative AI technologies that address privacy, equity, civil liability, and clinical validity. This includes adapting frameworks, such as those of the WHO on digital governance, to the realities of low- and middle-income countries.

We believe that low- and middle-income countries have the opportunity to anticipate risks and establish an ethical and regulatory framework that protects users, particularly those seeking emotional and mental health support in vulnerable situations. Without these actions, the promise of generative AI in health services could become a new form of digital and health inequity.

 

ORCID

Leonardo Rojas-Mezarina: https://orcid.org/0000-0003-0293-7107

David Villarreal-Zegarra: https://orcid.org/0000-0002-2222-4764

 

AUTHORS’ CONTRIBUTION

Leonardo Rojas-Mezarina: Conceptualization, investigation, writing - original draft, and approval of the final version.

David Villarreal-Zegarra: Review, investigation, writing - original draft, and approval of the final version.

 

FUNDING

This paper has been self-financed.

 

CONFLICT OF INTEREST

The authors declare that there were no conflicts of interest in the collection, analysis, or writing of the manuscript.

 

ACKNOWLEDGMENTS

Not applicable.

 

REVIEW PROCESS

This study has been reviewed by external peers in double-blind mode. The editor in charge was Renzo Rivera. The review process is included as supplementary material 1.

 

DATA AVAILABILITY STATEMENT

Not applicable.

 

DECLARATION OF THE USE OF GENERATIVE ARTIFICIAL INTELLIGENCE

We used DeepL to translate specific sections of the manuscript to Spanish and Grammarly to improve the wording of certain sections. The final version of the manuscript was reviewed and approved by all authors.

 

DISCLAIMER

The authors are responsible for all statements made in this article.

 

REFERENCES

Ahluwalia, M., Sehgal, S., Lee, G., Agu, E., & Kpodonu, J. (2025). Disparities in artificial intelligence-based tools among diverse minority populations: Biases, barriers, and solutions. JACC: Advances, 4(5), 101742. https://doi.org/10.1016/j.jacadv.2025.101742

Ayers, J. W., Poliak, A., Dredze, M., Leas, E. C., Zhu, Z., Kelley, J. B., Faix, D. J., Goodman, A. M., Longhurst, C. A., Hogarth, M., & Smith, D. M. (2023). Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum. JAMA Internal Medicine, 183(6), 589–596. https://doi.org/10.1001/jamainternmed.2023.1838

Baek, G., Cha, C., & Han, J.-H. (2025). AI chatbots for psychological health for health professionals: Scoping review. JMIR Human Factors, 12, e67682. https://doi.org/10.2196/67682

Bommasani, R., Klyman, K., Longpre, S., Kapoor, S., Maslej, N., Xiong, B., Zhang, D., & Liang, P. (2023). The foundation model transparency index (arXiv:2310.12941). arXiv. https://doi.org/10.48550/arXiv.2310.12941

Chatterji, A., Cunningham, T., Deming, D. J., Hitzig, Z., Ong, C., Shan, C. Y., & Wadman, K. (2025). How people use ChatGPT (Working Paper No. 34255). National Bureau of Economic Research. https://doi.org/10.3386/w34255

Cross, J. L., Choma, M. A., & Onofrey, J. A. (2024). Bias in medical AI: Implications for clinical decision-making. PLOS Digital Health, 3(11), e0000651. https://doi.org/10.1371/journal.pdig.0000651

Ethical AI Governance Group. (2023). Beyond the black box: Shaping a responsible AI landscape. KPMG.

Hussain, S. A., Bresnahan, M., & Zhuang, J. (2025). Can artificial intelligence revolutionize healthcare in the Global South? A scoping review of opportunities and challenges. Digital Health, 11, 20552076251348024. https://doi.org/10.1177/20552076251348024

Li, H., Zhang, R., Lee, Y.-C., Kraut, R. E., & Mohr, D. C. (2023). Systematic review and meta-analysis of AI-based conversational agents for promoting mental health and well-being. NPJ Digital Medicine, 6(1), 236. https://doi.org/10.1038/s41746-023-00979-5

Mandal, A., Chakraborty, T., & Gurevych, I. (2025). Towards privacy-aware mental health AI models: Advances, challenges, and opportunities (arXiv:2502.00451). arXiv. https://doi.org/10.48550/arXiv.2502.00451

Meskó, B., & Topol, E. J. (2023). The imperative for regulatory oversight of large language models (or generative AI) in healthcare. NPJ Digital Medicine, 6(1), 120. https://doi.org/10.1038/s41746-023-00873-0

Morley, J., Floridi, L., Kinsey, L., & Elhalal, A. (2020). From what to how: An initial review of publicly available AI ethics tools, methods and research to translate principles into practices. Science and Engineering Ethics, 26(4), 2141–2168. https://doi.org/10.1007/s11948-019-00165-5

Omiye, J. A., Lester, J. C., Spichak, S., Rotemberg, V., & Daneshjou, R. (2023). Large language models propagate race-based medicine. NPJ Digital Medicine, 6(1), 195. https://doi.org/10.1038/s41746-023-00939-z

Smart, S., & Montori, V. M. (2025, April 23). Peru’s AI regulatory boom: Quantity without depth? https://www.hks.harvard.edu/centers/carr-ryan/our-work/carr-ryan-commentary/perus-ai-regulatory-boom-quantity-without-depth

Stanford Center for Digital Health. (2025). AI for health in low- and middle-income countries. CDH.

Thirunavukarasu, A. J., Ting, D. S. J., Elangovan, K., Gutierrez, L., Tan, T. F., & Ting, D. S. W. (2023). Large language models in medicine. Nature Medicine, 29(8), 1930–1940. https://doi.org/10.1038/s41591-023-02448-8

Vaidyam, A. N., Wisniewski, H., Halamka, J. D., Kashavan, M. S., & Torous, J. B. (2019). Chatbots and conversational agents in mental health: A review of the psychiatric landscape. Canadian Journal of Psychiatry, 64(7), 456–464. https://doi.org/10.1177/0706743719828977

Villarreal-Zegarra, D., Reategui-Rivera, C. M., García-Serna, J., Quispe-Callo, G., Lázaro-Cruz, G., Centeno-Terrazas, G., Galvez-Arevalo, R., Escobar-Agreda, S., Dominguez-Rodriguez, A., & Finkelstein, J. (2024). Self-administered interventions based on natural language processing models for reducing depressive and anxiety symptoms: Systematic review and meta-analysis. JMIR Mental Health, 11, e59560. https://doi.org/10.2196/59560

Vollmer, S., Mateen, B. A., Bohner, G., Király, F. J., Ghani, R., Jonsson, P., Cumbers, S., Jonas, A., McAllister, K. S. L., Myles, P., Granger, D., Birse, M., Branson, R., Moons, K. G. M., Collins, G. S., Ioannidis, J. P. A., Holmes, C., & Hemingway, H. (2020). Machine learning and artificial intelligence research for patient benefit: 20 critical questions on transparency, replicability, ethics, and effectiveness. BMJ, 368, l6927. https://doi.org/10.1136/bmj.l6927

Wang, Y.-F., Li, M.-D., Wang, S.-H., Fang, Y., Sun, J., Lu, L., & Yan, W. (2025). Large language models in clinical psychiatry: Applications and optimization strategies. World Journal of Psychiatry, 15(11), 108199. https://doi.org/10.5498/wjp.v15.i11.108199

Weidinger, L., Uesato, J., Rauh, M., Griffin, C., Huang, P.-S., Mellor, J., Glaese, A., Cheng, M., Balle, B., Kasirzadeh, A., Biles, C., Brown, S., Kenton, Z., Hawkins, W., Stepleton, T., Birhane, A., Hendricks, L. A., Rimell, L., Isaac, W., … Gabriel, I. (2022). Taxonomy of risks posed by language models.