Referências

ABADJI, J. et al. Towards a Cleaner Document-Oriented Multilingual Crawled Corpus., a2022. Disponível em: <https://arxiv.org/abs/2201.06642>
ABADJI, J. et al. Towards a Cleaner Document-Oriented Multilingual Crawled Corpus. Proceedings of the Thirteenth Language Resources and Evaluation Conference. Anais...Marseille, France: European Language Resources Association, jun. b2022. Disponível em: <https://aclanthology.org/2022.lrec-1.463>
ABDALLA, M. et al. The Elephant in the Room: Analyzing the Presence of Big Tech in Natural Language Processing Research. Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Anais...Toronto, Canada: Association for Computational Linguistics, 2023. Disponível em: <https://aclanthology.org/2023.acl-long.734>
ABERCROMBIE, G. et al. Mirages. On Anthropomorphism in Dialogue Systems. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. Anais...Singapore: Association for Computational Linguistics, 2023. Disponível em: <https://aclanthology.org/2023.emnlp-main.290>
ABID, A.; FAROOQI, M.; ZOU, J. Persistent Anti-Muslim Bias in Large Language Models. Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society. Anais...Virtual Event USA: ACM, jul. 2021. Disponível em: <https://dl.acm.org/doi/10.1145/3461702.3462624>
AGHAJANYAN, A.; GUPTA, S.; ZETTLEMOYER, L. Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tuning. (C. Zong et al., Eds.)Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers), Virtual Event, August 1-6, 2021. Anais...Association for Computational Linguistics, 2021. Disponível em: <https://doi.org/10.18653/v1/2021.acl-long.568>
AGIRRE, E. Cross-Lingual Word Embeddings. Computational Linguistics, v. 46, n. 1, p. 245–248, mar. 2020.
AISERA. LLM Evaluation: Key Metrics and Frameworks. https://aisera.com/blog/llm-evaluation/, 2024.
ALBALAK, A. et al. A Survey on Data Selection for Language Models., 2024. Disponível em: <https://arxiv.org/abs/2402.16827>
ALKHAMISSI, B. et al. Investigating Cultural Alignment of Large Language Models. (L.-W. Ku, A. Martins, V. Srikumar, Eds.)Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Anais...Bangkok, Thailand: Association for Computational Linguistics, ago. 2024. Disponível em: <https://aclanthology.org/2024.acl-long.671/>
AMARAL, V. DO et al. A RAG-Powered Academic Chatbot with Ontology-Driven Factual Verification. Anais Estendidos do XL Simpósio Brasileiro de Bancos de Dados. Anais...Porto Alegre, RS, Brasil: SBC, 2025. Disponível em: <https://sol.sbc.org.br/index.php/sbbd_estendido/article/view/37615>
AMODEI, D. et al. Concrete Problems in AI Safety. arXiv preprint arXiv:1606.06565. Anais...2016.
ANISUZZAMAN, D. et al. Fine-tuning large language models for specialized use cases. Mayo Clinic Proceedings: Digital Health, v. 3, n. 1, p. 100184, 2025.
ARVAN, M.; PINA, L.; PARDE, N. Reproducibility in Computational Linguistics: Is Source Code Enough? Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing. Anais...Abu Dhabi, United Arab Emirates: Association for Computational Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.emnlp-main.150>
ASAI, A. et al. Self-rag: Learning to retrieve, generate, and critique through self-reflection. 2024.
ASSUNÇÃO, I. V.; JANSON, S. F. Afinal, o que é privacidade? Um panorama histórico do direito à privacidade no ordenamento constitucional brasileiro. Internet & Sociedade, v. 5, n. 1, 2024.
BAHDANAU, D.; CHO, K.; BENGIO, Y. Neural Machine Translation by Jointly Learning to Align and Translate. (Y. Bengio, Y. LeCun, Eds.)3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings. Anais...San Diego, California.: 2015. Disponível em: <http://arxiv.org/abs/1409.0473>
BAI, J. et al. Qwen technical report. arXiv preprint arXiv:2309.16609, 2023.
BAI, X. et al. Explicitly unbiased large language models still form biased associations. Proceedings of the National Academy of Sciences, v. 122, n. 8, p. e2416228122, fev. 2025.
BANERJEE, S.; LAVIE, A. METEOR: An Automatic Metric for MT Evaluation with Improved Correlation with Human Judgments. (J. Goldstein et al., Eds.)Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization. Anais...Ann Arbor, Michigan: Association for Computational Linguistics, jun. 2005. Disponível em: <https://aclanthology.org/W05-0909>
BARRETT, A. M. et al. Benchmark Early and Red Team Often: A framework for assessing and managing dual-hazards of AI foundational models. UC Berkeley Center for Long-Term Cybersecurity, 2024. Disponível em: <https://cltc.berkeley.edu/wp-content/uploads/2024/05/Dual-Use-Benchmark-Early-Red-Team-Often.pdf>
BARRY LYNN, M. VON T.; MONTOYA, K. AI in the Public Interest: Confronting the Monopoly Threat. Open Markets Institute, 2023. Disponível em: <https://www.openmarketsinstitute.org/publications/report-ai-in-the-public-interest-confronting-the-monopoly-threat>
BELINKOV, Y.; GEHRMANN, S.; PAVLICK, E. Interpretability and Analysis in Neural NLP. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: Tutorial Abstracts. Anais...Online: Association for Computational Linguistics, 2020. Disponível em: <https://www.aclweb.org/anthology/2020.acl-tutorials.1>
BELINKOV, Y.; GLASS, J. Analysis Methods in Neural Language Processing: A Survey. Transactions of the Association for Computational Linguistics, v. 7, p. 49–72, 2019.
BELTAGY, I.; PETERS, M. E.; COHAN, A. Longformer: The Long-Document Transformer. CoRR, v. abs/2004.05150, 2020.
BELZ, A. et al. A Systematic Review of Reproducibility Research in Natural Language Processing. Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume. Anais...Online: Association for Computational Linguistics, abr. 2021. Disponível em: <https://aclanthology.org/2021.eacl-main.29>
BENCKE, L. et al. Can we trust LLMs as relevance judges? Anais do XXXIX Simpósio Brasileiro de Bancos de Dados. Anais...Porto Alegre, RS, Brasil: SBC, 2024. Disponível em: <https://sol.sbc.org.br/index.php/sbbd/article/view/30724>
BENDER, E. M. On achieving and evaluating language-independence in NLP. Linguistic Issues in Language Technology, v. 6, 2011.
BENDER, E. M. et al. On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜. Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency. Anais...: FAccT ’21.New York, NY, USA: Association for Computing Machinery, 2021. Disponível em: <https://doi.org/10.1145/3442188.3445922>
BENDER, E. M. Resisting Dehumanization in the Age of AI. Current Directions in Psychological Science, v. 33, n. 2, p. 114–120, abr. 2024.
BENDER, E. M.; FRIEDMAN, B. Data Statements for Natural Language Processing: Toward Mitigating System Bias and Enabling Better Science. Transactions of the Association for Computational Linguistics, v. 6, p. 587–604, 2018.
BENDER, E. M.; HANNA, A. The AI Con: How to fight big tech’s hype and create the future we want. [s.l.] Random House, 2025.
BENDER, E. M.; KOLLER, A. Climbing towards NLU: On Meaning, Form, and Understanding in the Age of Data. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Anais...Online: Association for Computational Linguistics, jul. 2020. Disponível em: <https://aclanthology.org/2020.acl-main.463>
BENGIO, Y. et al. A Neural Probabilistic Language Model. J. Mach. Learn. Res., v. 3, n. null, p. 1137–1155, mar. 2003.
BENOTTI, L.; BLACKBURN, P. Ethics consideration sections in natural language processing papers. Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing. Anais...Abu Dhabi, United Arab Emirates: Association for Computational Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.emnlp-main.299>
BERTSCH, A. et al. Unlimiformer: Long-Range Transformers with Unlimited Length Input. CoRR, v. abs/2305.01625, 2023.
BERWICK, R. C.; CHOMSKY, N. Por que apenas nós? Linguagem e evolução. [s.l.] SciELO-Editora UNESP, 2017.
BIANCHI, F.; HOVY, D. On the Gap between Adoption and Understanding in NLP. Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021. Anais...Online: Association for Computational Linguistics, ago. 2021. Disponível em: <https://aclanthology.org/2021.findings-acl.340>
BIBAL, A. et al. Is Attention Explanation? An Introduction to the Debate. Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Anais...Dublin, Ireland: Association for Computational Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.acl-long.269>
BIBER, D. Variation across Speech and Writing. Cambridge: Cambridge University Press, 1988.
BIBER, D.; CONRAD, S. Register, Genre, and Style. Cambridge: Cambridge University Press, 2009.
BIBER, D.; FINEGAN, E. Sociolinguistic Perspectives On Register. [s.l.] Oxford University Press, 1994.
BINZ, M. et al. How should the advancement of large language models affect the practice of science? Proceedings of the National Academy of Sciences, v. 122, n. 5, p. e2401227121, fev. 2025.
BIRD, S. Decolonising Speech and Language Technology. Proceedings of the 28th International Conference on Computational Linguistics. Anais...Barcelona, Spain (Online): International Committee on Computational Linguistics, dez. 2020. Disponível em: <https://aclanthology.org/2020.coling-main.313>
BIRD, S. Local Languages, Third Spaces, and other High-Resource Scenarios. Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Anais...Dublin, Ireland: Association for Computational Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.acl-long.539>
BLODGETT, S. L. et al. Language (Technology) is Power: A Critical Survey of Bias in NLP. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Anais...Online: Association for Computational Linguistics, a2020. Disponível em: <https://www.aclweb.org/anthology/2020.acl-main.485>
BLODGETT, S. L. et al. Language (Technology) is Power: A Critical Survey of “Bias” in NLP. (D. Jurafsky et al., Eds.)Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Anais...Online: Association for Computational Linguistics, jul. b2020. Disponível em: <https://aclanthology.org/2020.acl-main.485/>
BLOM, J. D. A dictionary of hallucinations. [s.l.] Springer, 2010.
BOSTROM, N. Superintelligence: Paths, Dangers, Strategies. Oxford, UK: Oxford University Press, 2014.
BOWMAN, S. R.; DAHL, G. What Will it Take to Fix Benchmarking in Natural Language Understanding? (K. Toutanova et al., Eds.)Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Anais...Online: Association for Computational Linguistics, jun. 2021. Disponível em: <https://aclanthology.org/2021.naacl-main.385/>
BRANDES, N. et al. ProteinBERT: a universal deep-learning model of protein sequence and function. Bioinform., v. 38, n. 8, p. 2102–2110, 2022.
BRENNAN, K.; KAK, A.; WEST, S. M. The AGI Mythology: The Argument to End All Arguments. Em: Artificial Power: 2025 Landscape Report. [s.l.] AI Now Institute, 2025.
BRENNEIS, A. Assessing dual use risks in AI research: necessity, challenges and mitigation strategies. Research Ethics, v. 21, n. 2, p. 302–330, 2025.
BROUSSARD, M. More Than a Glitch: Confronting Race, Gender, and Ability Bias in Tech. 1. ed. Cambridge, Massachusetts: MIT Press, 2023.
BROWN, T. B. et al. Language Models are Few-Shot Learners. (H. Larochelle et al., Eds.)Advances in Neural Information Processing Systems. Anais...Curran Associates, Inc., 2020. Disponível em: <https://proceedings.neurips.cc/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html>
BRUM, H.; NUNES, M. DAS G. V. Building a Sentiment Corpus of Tweets in Brazilian Portuguese. (N. C. (Conference chair) et al., Eds.)Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018). Anais...Miyazaki, Japan: European Language Resources Association (ELRA), mar. 2018.
BURCHELL, L. et al. An Expanded Massive Multilingual Dataset for High-Performance Language Technologies (HPLT). (W. Che et al., Eds.)Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Anais...Vienna, Austria: Association for Computational Linguistics, jul. 2025. Disponível em: <https://aclanthology.org/2025.acl-long.854/>
CARLSMITH, J. Is Power-Seeking AI an Existential Risk? [s.l.] Open Philanthropy, 2022.
CARMO, D. et al. PTT5: Pretraining and validating the T5 model on Brazilian Portuguese data. CoRR, v. abs/2008.09144, 2020.
CASTILHO, S. et al. DELA Corpus - A Document-Level Corpus Annotated with Context-Related Issues. Proceedings of the Sixth Conference on Machine Translation. Anais...Online: Association for Computational Linguistics, nov. 2021. Disponível em: <https://aclanthology.org/2021.wmt-1.63>
CENTRE FOR INFORMATION POLICY LEADERSHIP, C. The Limitations of Consent as a Legal Basis for Data Processing in the Digital Society. Washington DC, London, BrusselsCentre for Information Policy Leadership & Hunton Andrews Kurth LLP; Bae Kim & Lee, 2024. Disponível em: <https://www.informationpolicycentre.com/uploads/5/7/1/0/57104281/cipl_bkl_limitations_of_consent_legal_basis_data_processing_dec24.pdf>
CERVANTES, M. DE. D. Quixote de La Mancha — Primeira Parte. [s.l.] eBooksBrasil, 1605.
CHALAMALASETTI, K. et al. clembench: Using Game Play to Evaluate Chat-Optimized Language Models as Conversational Agents. (H. Bouamor, J. Pino, K. Bali, Eds.)Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. Anais...Singapore: Association for Computational Linguistics, dez. 2023. Disponível em: <https://aclanthology.org/2023.emnlp-main.689>
CHANDRA, M. et al. From Lived Experience to Insight: Unpacking the Psychological Risks of Using AI Conversational Agents. Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency. Anais...Athens Greece: ACM, jun. 2025. Disponível em: <https://dl.acm.org/doi/10.1145/3715275.3732063>
CHEN, G. H. et al. Humans or LLMs as the Judge? A Study on Judgement Bias. Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, a2024.
CHEN, K. et al. How Susceptible are Large Language Models to Ideological Manipulation? Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing. Anais...Miami, Florida, USA: Association for Computational Linguistics, b2024. Disponível em: <https://aclanthology.org/2024.emnlp-main.952>
CHEN, T. et al. CopyBench: Measuring Literal and Non-Literal Reproduction of Copyright-Protected Text in Language Model Generation. (Y. Al-Onaizan, M. Bansal, Y.-N. Chen, Eds.)Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing. Anais...Miami, Florida, USA: Association for Computational Linguistics, nov. c2024. Disponível em: <https://aclanthology.org/2024.emnlp-main.844/>
CHENG, J. et al. Dated Data: Tracing Knowledge Cutoffs in Large Language Models., 2024. Disponível em: <https://arxiv.org/abs/2403.12958>
CHILD, R. et al. Generating Long Sequences with Sparse Transformers. CoRR, v. abs/1904.10509, 2019.
CHO, K. et al. Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation. (A. Moschitti, B. Pang, W. Daelemans, Eds.)Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP 2014, October 25-29, 2014, Doha, Qatar, A meeting of SIGDAT, a Special Interest Group of the ACL. Anais...ACL, 2014. Disponível em: <https://doi.org/10.3115/v1/d14-1179>
CHOMSKY, N. Aspects of the Theory of Syntax. Cambridge, MA: MIT Press, 1965.
CHOWDHERY, A. et al. PaLM: Scaling Language Modeling with Pathways. CoRR, v. abs/2204.02311, 2022.
CHRISTIANO, P. F. et al. Deep Reinforcement Learning from Human Preferences. (I. Guyon et al., Eds.)Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA. Anais...2017. Disponível em: <https://proceedings.neurips.cc/paper/2017/hash/d5e2c0adad503c91f91df240d0cd4e49-Abstract.html>
CLARK, K. et al. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators. 8th International Conference on Learning Representations, ICLR 2020. Anais...Addis Ababa, Ethiopia: OpenReview.net, abr. 2020. Disponível em: <https://openreview.net/forum?id=r1xMH1BtvB>
COHEN, K. B. et al. Three Dimensions of Reproducibility in Natural Language Processing. Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018). Anais...Miyazaki, Japan: European Language Resources Association (ELRA), 2018. Disponível em: <https://aclanthology.org/L18-1025>
COLLOBERT, R.; WESTON, J. A unified architecture for natural language processing: deep neural networks with multitask learning. (W. W. Cohen, A. McCallum, S. T. Roweis, Eds.)Machine Learning, Proceedings of the Twenty-Fifth International Conference (ICML 2008), Helsinki, Finland, June 5-9, 2008. Anais...: ACM International Conference Proceeding Series.ACM, 2008. Disponível em: <https://doi.org/10.1145/1390156.1390177>
COMMON CRAWL. Common Crawl – Open Repository of Web Crawl Data. https://commoncrawl.org, [s.d.].
CONNEAU, A.; LAMPLE, G. Cross-Lingual Language Model Pretraining. Em: Proceedings of the 33rd International Conference on Neural Information Processing Systems. Red Hook, NY, USA: Curran Associates Inc., 2019.
CORMACK, G. V.; CLARKE, C. L.; BUETTCHER, S. Reciprocal rank fusion outperforms condorcet and individual rank learning methods. Proceedings of the 32nd international ACM SIGIR conference on Research and development in information retrieval. Anais...2009.
CORMEN, T. et al. Introduction to Algorithms. Em: 2. ed. [s.l.] MIT Press; McGraw-Hill, 2001.
CORRÊA, N. K. et al. Tucano: Advancing Neural Text Generation for Portuguese. Patterns, 2025.
CORTES, C.; VAPNIK, V. Support-Vector Networks. Machine Learning, v. 20, n. 3, p. 273–297, set. 1995.
CORTIZ, D. et al. A Weakly Supervised Dataset of Fine-Grained Emotions in Portuguese. Anais do XIII Simpósio Brasileiro de Tecnologia da Informação e da Linguagem Humana. Anais...Porto Alegre, RS, Brasil: SBC, 2021. Disponível em: <https://sol.sbc.org.br/index.php/stil/article/view/17786>
COULDRY, N.; MEJIAS, U. A. The Costs of Connection: How Data Is Colonizing Human Life and Appropriating It for Capitalism. [s.l.] Stanford University Press, 2019.
CRESPO, M. C. R. M. et al. Carolina: a General Corpus of Contemporary Brazilian Portuguese with Provenance, Typology and Versioning Information., 2023. Disponível em: <https://arxiv.org/abs/2303.16098>
CRONBACH, L. J. Studies of acquiescence as a factor in the true–false test. Journal of Educational Psychology, v. 33, p. 401–415, 1942.
CUEVA, R. et al. Inteligência Artificial e Desafios Regulatórios. [s.l.] Forense, 2026.
CURRY, A.; CERCAS CURRY, A. Computer says “No”: The Case Against Empathetic Conversational AI. (A. Rogers, J. Boyd-Graber, N. Okazaki, Eds.)Findings of the Association for Computational Linguistics: ACL 2023. Anais...Toronto, Canada: Association for Computational Linguistics, jul. 2023. Disponível em: <https://aclanthology.org/2023.findings-acl.515/>
D’ARCY, A.; BENDER, E. M. Ethics in Linguistics. Annual Review of Linguistics, v. 9, n. 1, p. 49–69, jan. 2023.
DAI, Z. et al. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context. (A. Korhonen, D. R. Traum, L. Màrquez, Eds.)Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, July 28- August 2, 2019, Volume 1: Long Papers. Anais...Association for Computational Linguistics, 2019. Disponível em: <https://doi.org/10.18653/v1/p19-1285>
DARCIO, L. et al. LattesRex: Building ChatBots for Semi-Structured Documents. Proceedings of the 14th Brazilian Symposium in Information and Human Language Technology (STIL 2025). Anais...Brazil: 2025. Disponível em: <https://github.com/Lucas-Darcio/LattesRex>
DAVIES, M.; FERREIRA, M. J. Corpus do Português: Web/Dialects. https://www.corpusdoportugues.org/web-dial/, 2016.
DAVIES, M.; FERREIRA, M. J. Corpus do Português: NOW. https://www.corpusdoportugues.org/now/, 2018.
DETTMERS, T. et al. QLoRA: Efficient Finetuning of Quantized LLMs. arXiv preprint arXiv:2305.14314, 2023.
DEVLIN, J. et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. (J. Burstein, C. Doran, T. Solorio, Eds.)Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019. Anais...Minneapolis, MN, USA: Association for Computational Linguistics, 2019. Disponível em: <https://doi.org/10.18653/v1/n19-1423>
DIERK, C.; HEALEY, J.; DOGAN, M. D. Evaluating LLMs in Experiential Context: Insights from a Survey of Recent CHI Publications. Human-centered Evaluation and Auditing of Language Models Workshop (HEAL), CHI ’25. Anais...Yokohama, Japan: ACM, 2025. Disponível em: <https://heal-workshop.github.io/chi2025_papers/43_Evaluating_LLMs_in_Experien.pdf>
DING, N. et al. Parameter-efficient fine-tuning of large-scale pre-trained language models. Nature machine intelligence, v. 5, n. 3, p. 220–235, 2023.
DONG, Q. et al. A Survey for In-context Learning. CoRR, v. abs/2301.00234, 2023.
DOU, G. et al. Avoiding Copyright Infringement via Large Language Model Unlearning. (L. Chiruzzo, A. Ritter, L. Wang, Eds.)Findings of the Association for Computational Linguistics: NAACL 2025. Anais...Albuquerque, New Mexico: Association for Computational Linguistics, abr. 2025. Disponível em: <https://aclanthology.org/2025.findings-naacl.288/>
EDITORIAL. Emotional risks of AI companions demand attention. Nature Machine Intelligence, v. 7, n. 7, p. 981–982, jul. 2025.
ES, S. et al. Ragas: Automated evaluation of retrieval augmented generation. Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics: System Demonstrations. Anais...2024.
ETHAYARAJH, K.; JURAFSKY, D. Utility is in the Eye of the User: A Critique of NLP Leaderboards. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). Anais...Online: Association for Computational Linguistics, 2020. Disponível em: <https://www.aclweb.org/anthology/2020.emnlp-main.393>
FAN, A.; LEWIS, M.; DAUPHIN, Y. Hierarchical Neural Story Generation. Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Anais...Melbourne, Australia: Association for Computational Linguistics, jul. 2018. Disponível em: <https://aclanthology.org/P18-1082>
FÄRBER, F. et al. IaraMed: A Women’s Healthcare Chatbot for Portuguese Speakers. Anais do XXV Simpósio Brasileiro de Computação Aplicada à Saúde. Anais...Porto Alegre, RS, Brasil: SBC, 2025. Disponível em: <https://sol.sbc.org.br/index.php/sbcas/article/view/35558>
FEIJÓ, D. DE V.; MOREIRA, V. P. Mono vs Multilingual Transformer-based Models: a Comparison across Several Language Tasks. CoRR, v. abs/2007.09757, 2020.
FINGER, M. et al. Carolina: General Corpus of Contemporary Brazilian Portuguese with Provenance and Typology Information., 2020. Disponível em: <https://c4ai.inova.usp.br/portal/carolina>
FIRTH, J. R. A synopsis of linguistic theory 1930–1955. [s.l.] Blackwell, 1957. p. 1–32
FISHER, J. et al. Biased LLMs can Influence Political Decision-Making. (W. Che et al., Eds.)Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Anais...Vienna, Austria: Association for Computational Linguistics, jul. 2025. Disponível em: <https://aclanthology.org/2025.acl-long.328/>
FLORIDI, L. Translating Principles into Practices of Digital Ethics: Five Risks of Being Unethical. Philosophy & Technology, v. 32, n. 2, p. 185–193, jun. 2019.
FOKKENS, A. et al. Offspring from Reproduction Problems: What Replication Failure Teaches Us. Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Anais...Sofia, Bulgaria: Association for Computational Linguistics, ago. 2013. Disponível em: <https://aclanthology.org/P13-1166>
FORT, K.; ADDA, G.; COHEN, K. B. Amazon Mechanical Turk: Gold Mine or Coal Mine? Computational Linguistics, v. 37, n. 2, p. 413–420, jun. 2011.
FRANCISCO, P. A. P.; HUREL, L. M.; RIELLI, M. M. Regulação do Reconhecimento Facial no Setor Público: Avaliação de Experiências Internacionais. Instituto Igarapé - DataPrivacyBR, 2020. Disponível em: <https://www.dataprivacybr.org/wp-content/uploads/2021/11/regulacao-do-reconhecimento-facial-no-setor-publico.pdf>
FYFE, S. et al. Apophenia, theory of mind and schizotypy: perceiving meaning and intentionality in randomness. Cortex, v. 44, n. 10, p. 1316–1325, 2008.
GABRIEL, I. Artificial Intelligence, Values and Alignment. Minds and Machines, v. 30, n. 3, p. 411–437, 2020.
GAJULAMANDYAM, D. K. et al. Domain Specific Finetuning of LLMs Using PEFT Techniques. 2025 IEEE 15th Annual Computing and Communication Workshop and Conference (CCWC). Anais...2025.
GALILEO AI. Generative AI and LLM Insights. https://galileo.ai/blog/, 2024.
GAO, M. et al. LLM-based NLG Evaluation: Current Status and Challenges. ArXiv, 2024. Disponível em: <https://arxiv.org/abs/2402.01383>
GAO, T.; YAO, X.; CHEN, D. SimCSE: Simple Contrastive Learning of Sentence Embeddings. (M.-F. Moens et al., Eds.)Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021. Anais...Association for Computational Linguistics, 2021. Disponível em: <https://doi.org/10.18653/v1/2021.emnlp-main.552>
GARCIA, E. A. S. et al. RoBERTaLexPT: A Legal RoBERTa Model pretrained with deduplication for Portuguese. (P. Gamallo et al., Eds.)Proceedings of the 16th International Conference on Computational Processing of Portuguese - Vol. 1. Anais...Santiago de Compostela, Galicia/Spain: Association for Computational Lingustics, mar. 2024. Disponível em: <https://aclanthology.org/2024.propor-1.38>
GATT, A.; KRAHMER, E. Survey of the State of the Art in Natural Language Generation: Core tasks, applications and evaluation. Journal of Artificial Intelligence Research, v. 61, n. 1, p. 65–170, 2018.
GEBRU, T. et al. Datasheets for datasets. Communications of the ACM, v. 64, n. 12, p. 86–92, dez. 2021.
GEBRU, T.; BENDER, E. M.; MCMILLAN-MAJOR, A. Statement from the listed authors of Stochastic Parrots on the AI pause” letter., 2023. Disponível em: <https://www.dair-institute.org/blog/letter-statement-March2023/>
GEBRU, T.; TORRES, E. P. The TESCREAL bundle: Eugenics and the promise of utopia through artificial general intelligence. First Monday, abr. 2024.
GEHMAN, S. et al. RealToxicityPrompts: Evaluating Neural Toxic Degeneration in Language Models. (T. Cohn, Y. He, Y. Liu, Eds.)Findings of the Association for Computational Linguistics: EMNLP 2020. Anais...Online: Association for Computational Linguistics, nov. 2020. Disponível em: <https://aclanthology.org/2020.findings-emnlp.301/>
GEVA, M.; GUPTA, A.; BERANT, J. Injecting Numerical Reasoning Skills into Language Models. (D. Jurafsky et al., Eds.)Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, July 5-10, 2020. Anais...Association for Computational Linguistics, 2020. Disponível em: <https://doi.org/10.18653/v1/2020.acl-main.89>
GOLUMBIA, D. Cyberlibertarianism: The Right-Wing Politics of Digital Technology. [s.l.] University of Minnesota Press, 2024.
GONG, Z. et al. Continual Pre-training of Language Models for Math Problem Understanding with Syntax-Aware Memory Network. (S. Muresan, P. Nakov, A. Villavicencio, Eds.)Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, Dublin, Ireland, May 22-27, 2022. Anais...Association for Computational Linguistics, 2022. Disponível em: <https://doi.org/10.18653/v1/2022.acl-long.408>
GONZALEZ-CABELLO, M. et al. Fairness in crowdwork: Making the human AI supply chain more humane. Business Horizons, v. 68, n. 5, p. 645–657, 2025.
GOODFELLOW, I.; BENGIO, Y.; COURVILLE, A. Deep Learning. [s.l.] MIT Press, 2016. v. 1
GOOGLE CLOUD. Best Practices with Large Language Models. https://cloud.google.com/vertex-ai/generative-ai/docs/learn/prompt-best-practices?hl=en, 2025.
GRAY, M. L.; SURI, S. Ghost work: How to stop Silicon Valley from building a new global underclass. [s.l.] Harper Business, 2019.
GRAY WIDDER, D.; WEST, S.; WHITTAKER, M. Open (For Business): Big Tech, Concentrated Power, and the Political Economy of Open AI. SSRN Electronic Journal, 2023.
GU, Y. et al. Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing. ACM Trans. Comput. Healthcare, v. 3, n. 1, out. 2021.
GURURANGAN, S. et al. Don’t Stop Pretraining: Adapt Language Models to Domains and Tasks. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Anais...Online: Association for Computational Linguistics, jul. 2020. Disponível em: <https://aclanthology.org/2020.acl-main.740>
HADDAD, I. Artificial Intelligence and Data in Open Source. Linux Foundation, 2022. Disponível em: <https://www.linuxfoundation.org/hubfs/LF%20Research/Artificial%20Intelligence%20and%20Data%20in%20Open%20Source%20-%20Report.pdf?hsLang=en>
HANNIGAN, T. R.; MCCARTHY, I. P.; SPICER, A. Beware of botshit: How to manage the epistemic risks of generative chatbots. Business Horizons, v. 67, n. 5, p. 471–486, 2024.
HAVENS, L. et al. Situated Data, Situated Systems: A Methodology to Engage with Power Relations in Natural Language Processing Research. (M. R. Costa-jussà et al., Eds.)Proceedings of the Second Workshop on Gender Bias in Natural Language Processing. Anais...Barcelona, Spain (Online): Association for Computational Linguistics, dez. 2020. Disponível em: <https://aclanthology.org/2020.gebnlp-1.10/>
HE, K. et al. Deep Residual Learning for Image Recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016. Anais...IEEE Computer Society, 2016. Disponível em: <https://doi.org/10.1109/CVPR.2016.90>
HE, P. et al. Deberta: decoding-Enhanced Bert with Disentangled Attention. 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. Anais...OpenReview.net, 2021. Disponível em: <https://openreview.net/forum?id=XPZIaotutsD>
HEDDERICH, M. A.; OULASVIRTA, A. Explaining crowdworker behaviour through computational rationality. Behaviour & Information Technology, v. 44, n. 3, p. 552–573, 2025.
HELLER, J. Legal AI benchmarking: CoCounsel – from code to courtroom: The meticulous testing of CoCounsel’s professional-grade AI., 23 out. 2024. Disponível em: <https://www.thomsonreuters.com/en-us/posts/innovation/legal-ai-benchmarking-cocounsel/>. Acesso em: 20 ago. 2025
HENDREN, S. All technology is assistive: Six design rules on disability. Em: SAYERS, J. (Ed.). Making Things and Drawing Boundaries: Experiments in the Digital Humanities. [s.l.] University of Minnesota Press, Minneapolis, MN, 2014.
HENDRYCKS, D. et al. Unsolved Problems in ML Safety. arXiv preprint arXiv:2109.13916, 2021.
HERSHCOVICH, D. et al. Towards Climate Awareness in NLP Research. Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing. Anais...Abu Dhabi, United Arab Emirates: Association for Computational Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.emnlp-main.159>
HICKS, M. T.; HUMPHRIES, J.; SLATER, J. ChatGPT is bullshit. Ethics and Information Technology, v. 26, n. 2, p. 38, jun. 2024.
HOCHREITER, S. Untersuchungen zu dynamischen neuronalen Netzen. Diploma, Technische Universität München, v. 91, n. 1, p. 31, 1991.
HOCHREITER, S.; SCHMIDHUBER, J. Long Short-Term Memory. Neural Computation, v. 9, n. 8, p. 1735–1780, nov. 1997.
HOFFMANN, J. et al. Training Compute-Optimal Large Language Models. CoRR, v. abs/2203.15556, 2022.
HOFMANN, V. et al. AI generates covertly racist decisions about people based on their dialect. Nature, v. 633, n. 8028, p. 147–154, set. 2024.
HOLTZMAN, A. et al. The Curious Case of Neural Text Degeneration. ICLR. Anais...OpenReview.net, 2020. Disponível em: <http://dblp.uni-trier.de/db/conf/iclr/iclr2020.html#HoltzmanBDFC20>
HOOKER, S. Moving beyond “algorithmic bias is a data problem”. Patterns, v. 2, n. 4, p. 100241, abr. 2021.
HORNIK, K.; STINCHCOMBE, M. B.; WHITE, H. Multilayer feedforward networks are universal approximators. Neural Networks, v. 2, n. 5, p. 359–366, 1989.
HOUAMEGNI, L. R. P.; GEDIKLI, F. Evaluating the Effectiveness of Large Language Models in Automated News Article Summarization., 2025. Disponível em: <https://arxiv.org/abs/2502.17136>
HOULSBY, N. et al. Parameter-Efficient Transfer Learning for NLP. (K. Chaudhuri, R. Salakhutdinov, Eds.)Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA. Anais...: Proceedings of Machine Learning Research.PMLR, 2019. Disponível em: <http://proceedings.mlr.press/v97/houlsby19a.html>
HOVY, D.; SPRUIT, S. L. The Social Impact of Natural Language Processing. Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). Anais...Berlin, Germany: Association for Computational Linguistics, 2016. Disponível em: <http://aclweb.org/anthology/P16-2096>
HOWARD, J.; RUDER, S. Universal Language Model Fine-tuning for Text Classification. Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Anais...Melbourne, Australia: Association for Computational Linguistics, jul. a2018. Disponível em: <^5^>
HOWARD, J.; RUDER, S. Universal Language Model Fine-tuning for Text Classification. (I. Gurevych, Y. Miyao, Eds.)Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Anais...Melbourne, Australia: Association for Computational Linguistics, jul. b2018. Disponível em: <https://aclanthology.org/P18-1031/>. Acesso em: 25 jul. 2025
HU, E. J. et al. LoRA: Low-Rank Adaptation of Large Language Models., 2021. Disponível em: <https://arxiv.org/abs/2106.09685>
HU, E. J. et al. LoRA: Low-Rank Adaptation of Large Language Models. The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. Anais...OpenReview.net, 2022. Disponível em: <https://openreview.net/forum?id=nZeVKeeFYf9>
HUANG, J.; YANG, D.; POTTS, C. Demystifying Verbatim Memorization in Large Language Models. (Y. Al-Onaizan, M. Bansal, Y.-N. Chen, Eds.)Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing. Anais...Miami, Florida, USA: Association for Computational Linguistics, nov. a2024. Disponível em: <https://aclanthology.org/2024.emnlp-main.598/>
HUANG, L. et al. A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions. ACM Transactions on Information Systems, v. 43, n. 2, p. 1–58, b2024.
II, S. M. W. GSM8K Benchmark. Klu; https://klu.ai/glossary/GSM8K-eval, 2025.
IP, J.; VONGTHONGSRI, K. deepeval., ago. 2025. Disponível em: <https://github.com/confident-ai/deepeval>
IVGI, M.; SHAHAM, U.; BERANT, J. Efficient Long-Text Understanding with Short-Text Models. Transactions of the Association for Computational Linguistics, v. 11, p. 284–299, 2023.
JACKSON, P.; MOULINIER, I. Natural Language Processing for Online Applications – Text retrieval, extraction and categorization. [s.l.] John Benjamins, 2002.
JAIN, S.; WALLACE, B. C. Attention is not Explanation. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). Anais...Minneapolis, Minnesota: Association for Computational Linguistics, 2019. Disponível em: <https://aclanthology.org/N19-1357>
JAKUBČEK, M. et al. The TenTen corpus family. Proceedings of the 7th International Corpus Linguistics Conference (CL). Anais...2013.
JAN, E. et al. Multitask-Bench: Unveiling and Mitigating Safety Gaps in LLMs Fine-tuning. (O. Rambow et al., Eds.)Proceedings of the 31st International Conference on Computational Linguistics. Anais...Abu Dhabi, UAE: Association for Computational Linguistics, jan. 2025. Disponível em: <https://aclanthology.org/2025.coling-main.606/>
JAPKOWICZ, N.; BOUKOUVALAS, Z. Machine Learning Evaluation: Towards Reliable and Responsible AI. [s.l.] Cambridge University Press, 2024.
JENSEN, K. N.; PLANK, B. Fine-tuning vs From Scratch: Do Vision & Language Models Have Similar Capabilities on Out-of-Distribution Visual Question Answering? (N. Calzolari et al., Eds.)Proceedings of the Thirteenth Language Resources and Evaluation Conference. Anais...Marseille, France: European Language Resources Association, jun. 2022. Disponível em: <https://aclanthology.org/2022.lrec-1.161/>
JEONG, S. et al. Adaptive-rag: Learning to adapt retrieval-augmented large language models through question complexity. arXiv preprint arXiv:2403.14403, 2024.
JI, Z. et al. Survey of Hallucination in Natural Language Generation. ACM Comput. Surv., v. 55, n. 12, mar. 2023.
JIANG, L.; WAGNER, C. How Low is Low? Crowdworker Perceptions of Microtask Payments in Work versus Leisure Situations. Proceedings of the 2024 CHI Conference on Human Factors in Computing Systems. Anais...: CHI ’24.New York, NY, USA: Association for Computing Machinery, 2024. Disponível em: <https://doi.org/10.1145/3613904.3642601>
JIAO, J. et al. Navigating LLM Ethics: Advancements, Challenges, and Future Directions. ArXiv, v. abs/2406.18841, 2024.
JIN, X. et al. Lifelong Pretraining: Continually Adapting Language Models to Emerging Corpora. Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Anais...Seattle, United States: Association for Computational Linguistics, jul. 2022. Disponível em: <https://aclanthology.org/2022.naacl-main.351>
JIN, Z. et al. How Good Is NLP? A Sober Look at NLP Tasks through the Lens of Social Impact. Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021. Anais...Online: Association for Computational Linguistics, 2021. Disponível em: <https://aclanthology.org/2021.findings-acl.273>
JOHNSTON, S. F. Alvin Weinberg and the Promotion of the Technological Fix. Technology and Culture, v. 59, n. 3, p. 620–651, 2018.
JOSHI, M. et al. TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension. (R. Barzilay, M.-Y. Kan, Eds.)Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, ACL 2017, Vancouver, Canada, July 30 - August 4, Volume 1: Long Papers. Anais...Association for Computational Linguistics, 2017. Disponível em: <https://doi.org/10.18653/v1/P17-1147>
JOSHI, M. et al. SpanBERT: Improving Pre-training by Representing and Predicting Spans. Transactions of the Association for Computational Linguistics, v. 8, p. 64–77, 2020.
JOSHI, N.; TALY, A.; MUPPALLA, D. LLM-Cite: Cheap Fact Verification with Attribution via URL Generation., 2025. Disponível em: <https://openreview.net/forum?id=qb2QRoE4W3>
JOULIN, A. et al. Bag of Tricks for Efficient Text Classification. (M. Lapata, P. Blunsom, A. Koller, Eds.)Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 2, Short Papers. Anais...Valencia, Spain: Association for Computational Linguistics, abr. 2017. Disponível em: <https://aclanthology.org/E17-2068/>
JOYCE, J. M. Kullback-Leibler Divergence. Em: LOVRIC, M. (Ed.). International Encyclopedia of Statistical Science. Berlin, Heidelberg: Springer Berlin Heidelberg, 2011. p. 720–722.
JURAFSKY, D.; MARTIN, J. H. Speech and Language Processing: An Introduction to Natural Language Processing, Computational Linguistics, and Speech Recognition. 3rd. ed. USA: Prentice Hall PTR, 2023.
KALAI, A. T. et al. Why language models hallucinate. [s.l.] OpenAI, 2025.
KALOULI, A.-L. et al. Curing the SICK and Other NLI Maladies. Computational Linguistics, v. 49, n. 1, p. 199–243, mar. 2023.
KAMRUZZAMAN, M.; SHOVON, MD.; KIM, G. Investigating Subtler Biases in LLMs: Ageism, Beauty, Institutional, and Nationality Bias in Generative Models. (L.-W. Ku, A. Martins, V. Srikumar, Eds.)Findings of the Association for Computational Linguistics: ACL 2024. Anais...Bangkok, Thailand: Association for Computational Linguistics, ago. 2024. Disponível em: <https://aclanthology.org/2024.findings-acl.530/>
KARAMOLEGKOU, A. et al. Copyright Violations and Large Language Models. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. Anais...Singapore: Association for Computational Linguistics, 2023. Disponível em: <https://aclanthology.org/2023.emnlp-main.458>
KE, Z. et al. Continual Pre-training of Language Models., 2023. Disponível em: <https://arxiv.org/abs/2302.03241>
KERNER, T. Domain-Specific Pretraining of Language Models: A Comparative Study in the Medical Field., 2024. Disponível em: <https://arxiv.org/abs/2407.14076>
KHATTAB, O.; ZAHARIA, M. ColBERT: Efficient and Effective Passage Search via Contextualized Late Interaction over BERT., 2020. Disponível em: <https://arxiv.org/abs/2004.12832>
KIANPOUR, M.; WEN, S.-F. Timing Attacks on Machine Learning: State of the Art. Intelligent Systems Conference. Anais...Springer, 2020.
KIM, D. K. et al. Analyzing Offensive Language Dataset Insights from Training Dynamics and Human Agreement Level. (O. Rambow et al., Eds.)Proceedings of the 31st International Conference on Computational Linguistics, COLING 2025, Abu Dhabi, UAE, January 19-24, 2025. Anais...Association for Computational Linguistics, 2025. Disponível em: <https://aclanthology.org/2025.coling-main.653/>
KIM, S. et al. Prometheus: Inducing Fine-grained Evaluation Capability in Language Models. arXiv preprint arXiv:2310.08491, 2023.
KLENK, M. How Do Technological Artefacts Embody Moral Values? Philosophy & Technology, v. 34, n. 3, p. 525–544, set. 2021.
KNUTH, D. E. Fundamental Algorithms. The Art of Computer Programming. 3. ed. [s.l.] Addison-Wesley, 1997. v. 1
KOGKALIDIS, K.; CHATZIKYRIAKIDIS, S. On Tables with Numbers, with Numbers. (S. Truong et al., Eds.)Proceedings of the 1st Workshop on Language Models for Underserved Communities (LM4UC 2025). Anais...Albuquerque, New Mexico: Association for Computational Linguistics, 2025. Disponível em: <https://aclanthology.org/2025.lm4uc-1.12/>
KOJIMA, T. et al. Large Language Models are Zero-Shot Reasoners. NeurIPS. Anais...2022. Disponível em: <http://papers.nips.cc/paper\_files/paper/2022/hash/8bb0d291acd4acf06ef112099c16f326-Abstract-Conference.html>
KRAKOVNA, V. et al. Specification Gaming: The Flip Side of AI Ingenuity. DeepMind Safety Research Blog, 2020.
KREUTZER, J. et al. Quality at a Glance: An Audit of Web-Crawled Multilingual Datasets. Transactions of the Association for Computational Linguistics, v. 10, p. 50–72, jan. 2022.
KWIATKOWSKI, T. et al. Natural Questions: a Benchmark for Question Answering Research. Transactions of the Association of Computational Linguistics, 2019.
LAI, J. et al. Large language models in law: A survey. AI Open, v. 5, p. 181–196, 2024.
LAN, Z. et al. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations. 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. Anais...OpenReview.net, 2020. Disponível em: <https://openreview.net/forum?id=H1eA7AEtvS>
LASOTA, L. Regulating Corporate Behaviour in Digital Ecosystems: Increasing Fairness and Contestability of Digital Markets with Free Software. MIC 2023: Toward Green, Inclusive, and Digital Growth. Anais...a2023.
LASOTA, L. The European Union’s AI act from the perspective of Open Technologies. Em: ALEGRE, M.; FÜRST, H. (Eds.). Advocacia & Bioética: Novas Tecnologias. https://www.editoracasadodireito.com.br/produto/novas-tecnologias. São Paulo: Casa do Direito, 2023b.
LASOTA, L. The CRA and the Challenges of Regulating Cybersecurity in Open Environments: The Case of Free and Open Source Software. Em: Digital Decade: How the EU shapes digitalisation. Berlin: Weizenbaum Institute for the Networked Society - The German Internet Institute, 2025.
LASOTA, L.; SINGHAL, N. Free Software and AI openness: Overcoming challenges in the licensing world. [s.l.] Zenodo, abr. 2024. Disponível em: <https://zenodo.org/doi/10.5281/zenodo.10966136>.
LAZZARINI MORETTI, J.; MALTESE ZUFFO, M. LGPD e inteligência artificial: Um estudo comparado. Revista de Direito Internacional e Globalização Econômica, v. 13, n. 13, p. 21–42, 2025.
LEE, J. et al. A survey of large language models in finance (finllms). arXiv preprint arXiv:2402.02315, 2024.
LEE, K. et al. Deduplicating Training Data Makes Language Models Better., 2022. Disponível em: <https://arxiv.org/abs/2107.06499>
LEIDNER, J. L.; PLACHOURAS, V. Ethical by Design: Ethics Best Practices for Natural Language Processing. Proceedings of the First ACL Workshop on Ethics in Natural Language Processing. Anais...Valencia, Spain: Association for Computational Linguistics, abr. 2017. Disponível em: <https://aclanthology.org/W17-1604>
LEIKE, J. et al. Scalable Agent Alignment via Reward Modeling: a Research Direction. arXiv preprint arXiv:1811.07871. Anais...2018.
LESTER, B.; AL-RFOU, R.; CONSTANT, N. The Power of Scale for Parameter-Efficient Prompt Tuning. (M.-F. Moens et al., Eds.)Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021. Anais...Association for Computational Linguistics, 2021. Disponível em: <https://doi.org/10.18653/v1/2021.emnlp-main.243>
LEWIS, M. et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. (D. Jurafsky et al., Eds.)Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, July 5-10, 2020. Anais...Association for Computational Linguistics, a2020. Disponível em: <https://doi.org/10.18653/v1/2020.acl-main.703>
LEWIS, P. S. H. et al. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. (H. Larochelle et al., Eds.)Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual. Anais...b2020. Disponível em: <https://proceedings.neurips.cc/paper/2020/hash/6b493230205f780e1bc26945df7481e5-Abstract.html>
LI, D. et al. From Generation to Judgment: Opportunities and Challenges of LLM-as-a-judge. CoRR, v. abs/2411.16594, 2024.
LI, H. et al. CaseGen: A benchmark for multi-stage legal case documents generation., 2025. Disponível em: <https://arxiv.org/abs/2502.17943>
LI, P. et al. Making AI Less "Thirsty": Uncovering and Addressing the Secret Water Footprint of AI Models. arXiv preprint arXiv:2304.03271, a2023.
LI, Q. Parameter Efficient Fine-Tuning on Selective Parameters for Transformer-Based Pre-Trained Models. 2024 IEEE International Conference on Multimedia and Expo (ICME). Anais...2024.
LI, R. et al. StarCoder: may the source be with you! CoRR, v. abs/2305.06161, b2023.
LI, W. W. et al. BERT Is Not The Count: Learning to Match Mathematical Statements with Proofs. (A. Vlachos, I. Augenstein, Eds.)Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2023, Dubrovnik, Croatia, May 2-6, 2023. Anais...Association for Computational Linguistics, c2023. Disponível em: <https://aclanthology.org/2023.eacl-main.260>
LI, X. L.; LIANG, P. Prefix-Tuning: Optimizing Continuous Prompts for Generation. (C. Zong et al., Eds.)Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers), Virtual Event, August 1-6, 2021. Anais...Association for Computational Linguistics, 2021. Disponível em: <https://doi.org/10.18653/v1/2021.acl-long.353>
LI, Y. et al. A comparative study of pretrained language models for long clinical text. Journal of the American Medical Informatics Association, v. 30, n. 2, p. 340–347, nov. 2022.
LIANG, X. et al. Contrastive Demonstration Tuning for Pre-trained Language Models. (Y. Goldberg, Z. Kozareva, Y. Zhang, Eds.)Findings of the Association for Computational Linguistics: EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7-11, 2022. Anais...Association for Computational Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.findings-emnlp.56>
LIN, C.-Y. ROUGE: A Package for Automatic Evaluation of Summaries. Text Summarization Branches Out. Anais...Barcelona, Spain: Association for Computational Linguistics, jul. 2004. Disponível em: <https://aclanthology.org/W04-1013>
LIN, S.; HILTON, J.; EVANS, O. TruthfulQA: Measuring How Models Mimic Human Falsehoods. (S. Muresan, P. Nakov, A. Villavicencio, Eds.)Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Anais...Dublin, Ireland: Association for Computational Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.acl-long.229/>
LIRA, T. et al. Aroeira: A Curated Corpus for the Portuguese Language with a Large Number of Tokens. Anais da XXXIV Brazilian Conference on Intelligent Systems. Anais...Porto Alegre, RS, Brasil: SBC, 2024. Disponível em: <https://sol.sbc.org.br/index.php/bracis/article/view/33560>
LITA, A.-C.; COSTEA, C.-E.; IONESCU, B. Chunking for RAG: A Comprehensive Evaluation of Chunking Strategies for Retrieval-Augmented Generation. Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing. Anais...2024.
LIU, Y. et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach., a2019. Disponível em: <https://arxiv.org/abs/1907.11692>
LIU, Y. et al. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692, b2019.
LIU, Y. et al. Multilingual Denoising Pre-training for Neural Machine Translation. Trans. Assoc. Comput. Linguistics, v. 8, p. 726–742, 2020.
LIU, Y. et al. HD-Eval: Aligning Large Language Model Evaluators Through Hierarchical Criteria Decomposition. (L.-W. Ku, A. Martins, V. Srikumar, Eds.)Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2024, Bangkok, Thailand, August 11-16, 2024. Anais...Association for Computational Linguistics, 2024. Disponível em: <https://doi.org/10.18653/v1/2024.acl-long.413>
LIU, Z. et al. A Robustly Optimized BERT Pre-Training Approach with Post-Training. Chinese Computational Linguistics: 20th China National Conference, CCL 2021, Hohhot, China, August 13–15, 2021, Proceedings. Anais...Berlin, Heidelberg: Springer-Verlag, 2021. Disponível em: <https://doi.org/10.1007/978-3-030-84186-7_31>
LOBO, P. Profiling na Lei Geral de Proteção de Dados: O Livre Desenvolvimento da Personalidade em Face da Governamentalidade Algorítmica. 1. ed. [s.l.] Editora Foco, 2022.
LONGPRE, S. et al. A Pretrainers Guide to Training Data: Measuring the Effects of Data Age, Domain Coverage, Quality, & Toxicity. (K. Duh, H. Gomez, S. Bethard, Eds.)Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers). Anais...Mexico City, Mexico: Association for Computational Linguistics, jun. 2024. Disponível em: <https://aclanthology.org/2024.naacl-long.179/>
LONGPRE, S. et al. The Responsible Foundation Model Development Cheatsheet: A Review of Tools & Resources., 2025. Disponível em: <https://arxiv.org/abs/2406.16746>
LUONG, T.; PHAM, H.; MANNING, C. D. Effective Approaches to Attention-based Neural Machine Translation. (L. Màrquez et al., Eds.)Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, EMNLP 2015, Lisbon, Portugal, September 17-21, 2015. Anais...The Association for Computational Linguistics, 2015. Disponível em: <https://doi.org/10.18653/v1/d15-1166>
LV, K. et al. Full Parameter Fine-tuning for Large Language Models with Limited Resources., 2024. Disponível em: <https://arxiv.org/abs/2306.09782>
MALEKI, N.; PADMANABHAN, B.; DUTTA, K. AI Hallucinations: A Misnomer Worth Clarifying. 2024 IEEE Conference on Artificial Intelligence (CAI). Anais...Singapore, Singapore: IEEE, jun. 2024. Disponível em: <https://ieeexplore.ieee.org/document/10605268/>
MARKOV, A. A. The theory of algorithms. Trudy Matematicheskogo Instituta Imeni VA Steklova, v. 42, p. 3–375, 1954.
MATHUR, N.; BALDWIN, T.; COHN, T. Tangled up in BLEU: Reevaluating the Evaluation of Automatic Machine Translation Evaluation Metrics. Proceedings of the Conference on Empirical Methods in Natural Language Processing. Anais...Association for Computational Linguistics, 2020. Disponível em: <https://aclanthology.org/2020.acl-main.448/>
MATTHEWS, B. W. Comparison of the predicted and observed secondary structure of T4 phage lysozyme. Biochimica et Biophysica Acta (BBA) - Protein Structure, v. 405, n. 2, p. 442–451, 1975.
MCCANN, B. et al. Learned in Translation: Contextualized Word Vectors. Proceedings of the 31st International Conference on Neural Information Processing Systems. Anais...: NIPS’17.Red Hook, NY, USA: Curran Associates Inc., 2017.
MCMILLAN-MAJOR, A.; BENDER, E. M.; FRIEDMAN, B. Data Statements: From Technical Concept to Community Practice. ACM Journal on Responsible Computing, p. 3594737, 2023.
MCQUILLAN, D. Resisting AI: an anti-fascist approach to artificial intelligence. Bristol, UK: [s.n.].
MECKLENBURG, N. et al. Injecting New Knowledge into Large Language Models via Supervised Fine-Tuning., 2024. Disponível em: <https://arxiv.org/abs/2404.00213>
MEJIAS, U. A.; COULDRY, N. Datafication. Internet Policy Review, v. 8, n. 4, nov. 2019.
MEVA, DR. D.; KUKADIYA, H. Performance Evaluation of Large Language Models: A Comprehensive Review. International Research Journal of Computer Science, v. 12, p. 109–114, mar. 2025.
MICELI, M. et al. Who Trains the Data for European Artificial Intelligence? The Left, DiPLab, Weizenbaum Institute; DAIR Institute., 2024. Disponível em: <https://hal.science/hal-04662589v1>
MIIKKULAINEN, R.; DYER, M. G. Natural Language Processing With Modular Pdp Networks and Distributed Lexicon. Cognitive Science, v. 15, n. 3, p. 343–399, 1991.
MILLER, B. Is Technology Value-Neutral? Science, Technology, & Human Values, v. 46, n. 1, p. 53–80, jan. 2021.
MINAEE, S. et al. Large language models: A survey. arXiv preprint arXiv:2402.06196, 2024.
MINAEE, S. et al. Large Language Models: A Survey., 2025. Disponível em: <https://arxiv.org/abs/2402.06196>
MITCHELL, M. et al. Model Cards for Model Reporting. Proceedings of the Conference on Fairness, Accountability, and Transparency. Anais...Atlanta GA USA: ACM, jan. 2019. Disponível em: <https://dl.acm.org/doi/10.1145/3287560.3287596>
MIZRAHI, D. et al. Language Models Improve When Pretraining Data Matches Target Tasks., 2025. Disponível em: <https://arxiv.org/abs/2507.12466>
MOHAMMAD, S. Ethics Sheets for AI Tasks. Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Anais...Dublin, Ireland: Association for Computational Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.acl-long.573>
MOREIRA, J.; VIANNA, D. Challenges of Generative AI in Legal Practice: A Jusbrasil Perspective. Proceedings of the 40th Brazilian Symposium on Databases (SBBD 2025). Anais...Brazil: Brazilian Computer Society (SBC), 2025. Disponível em: <https://sbbd.org.br/2025/wp-content/uploads/2025/10/SBBD25-Jusbrasil.pdf>
MORESCHI, B.; PEREIRA, G.; COZMAN, F. G. The Brazilian Workers in Amazon Mechanical Turk: Dreams and realities of ghost workers. Revista Contracampo, v. 39, n. 1, abr. 2020.
MOROZOV, E. To save everything, click here : the folly of technological solutionism. New York, NY: PublicAffairs, 2014.
MUNGER, K. Chatbots for Good and Evil. EACL via Underline Science Inc., 2023. Disponível em: <https://underline.io/lecture/72154-chatbots-for-good-and-evil>
NAIR, P. A. et al. Matryoshka Quantization. Sparsity in LLMs (SLLM): Deep Dive into Mixture of Experts, Quantization, Hardware, and Inference. Anais...2025.
NARAYAN, S.; COHEN, S. B.; LAPATA, M. Don’t Give Me the Details, Just the Summary! Topic-Aware Convolutional Neural Networks for Extreme Summarization. Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Anais...Brussels, Belgium: 2018.
NATIONAL INSTITUTE OF STANDARDS AND TECHNOLOGY. A Plan for Global Engagement on AI Standards. [s.l.] U.S. Department of Commerce, National Institute of Standards; Technology, jul. 2024. Disponível em: <https://nvlpubs.nist.gov/nistpubs/ai/NIST.AI.100-5.pdf>.
NEPOMUCENO, J. et al. Agentic AI Meets Fundamentalist Financial Analysis: Limits and Possibilities. Proceedings of the 14th Brazilian Conference on Intelligent Systems (BRACIS 2025). Anais...2025.
NGO, R.; CHAN, L.; MINDERMANN, S. The Alignment Problem from a Deep Learning Perspective. arXiv preprint arXiv:2303.16200, 2023.
NIJKAMP, E. et al. ProGen2: Exploring the Boundaries of Protein Language Models. CoRR, v. abs/2206.13517, 2022.
NOVOBILSKÁ, L. Free and Open Source Software Licensing Requirements and Copyright Infringement Involving Artificial Intelligence Technologies. 2023.
O’NEIL, C. Weapons of math destruction : how big data increases inequality and threatens democracy. First edition ed. New York: [s.n.].
OLIVEIRA, L. L. Inteligência artificial e desigualdade social: o impacto do colonialismo digital nas políticas públicas. Internet & Sociedade, v. 5, n. 1, 2024.
OUYANG, L. et al. Training language models to follow instructions with human feedback. (A. H. Oh et al., Eds.)Advances in Neural Information Processing Systems. Anais...2022. Disponível em: <https://openreview.net/forum?id=TG8KACxEON>
OVERWIJK, A.; XIONG, C.; CALLAN, J. ClueWeb22: 10 Billion Web Documents with Rich Information. (E. Amigó et al., Eds.)SIGIR ’22: The 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, July 11 - 15, 2022. Anais...ACM, 2022. Disponível em: <https://doi.org/10.1145/3477495.3536321>
PAIOLA, P. H. et al. Adapting llms for the medical domain in portuguese: A study on fine-tuning and model evaluation. arXiv preprint arXiv:2410.00163, 2024.
PAN, S. et al. Unifying large language models and knowledge graphs: A roadmap. IEEE Transactions on Knowledge and Data Engineering, v. 36, n. 7, p. 3580–3599, 2024.
PAN, Y. et al. On the Risk of Misinformation Pollution with Large Language Models. Findings of the Association for Computational Linguistics: EMNLP 2023. Anais...Singapore: Association for Computational Linguistics, 2023. Disponível em: <https://aclanthology.org/2023.findings-emnlp.97>
PANICKSSERY, A.; BOWMAN, S. R.; FENG, S. LLM Evaluators Recognize and Favor Their Own Generations. Proceedings of the 38th International Conference on Neural Information Processing Systems, 2024.
PAPINENI, K. et al. BLEU: A Method for Automatic Evaluation of Machine Translation. Proceedings of the 40th Annual Meeting on Association for Computational Linguistics. Anais...: ACL ’02.USA: Association for Computational Linguistics, 2002. Disponível em: <https://doi.org/10.3115/1073083.1073135>
PARK, K. et al. OffsetBias: Leveraging Debiased Data for Tuning Evaluators., 2024. Disponível em: <https://arxiv.org/abs/2407.06551>
PARMAR, M. et al. Don’t Blame the Annotator: Bias Already Starts in the Annotation Instructions. Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics. Anais...Dubrovnik, Croatia: Association for Computational Linguistics, 2023. Disponível em: <https://aclanthology.org/2023.eacl-main.130>
PASSINATO, E.; RIOS, W.; FILHO, A. G. Integração de Modelos de Linguagem e RAG na Criação de Chatbots Oftalmológicos. Anais do XXIV Simpósio Brasileiro de Computação Aplicada à Saúde. Anais...Porto Alegre, RS, Brasil: SBC, 2024. Disponível em: <https://sol.sbc.org.br/index.php/sbcas/article/view/28831>
PAULLADA, A. et al. Data and its (dis)contents: A survey of dataset development and use in machine learning research. Patterns, v. 2, n. 11, p. 100336, nov. 2021.
PETERS, M. E. et al. Deep Contextualized Word Representations. (M. A. Walker, H. Ji, A. Stent, Eds.)Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2018, New Orleans, Louisiana, USA, June 1-6, 2018, Volume 1 (Long Papers). Anais...Association for Computational Linguistics, 2018. Disponível em: <https://doi.org/10.18653/v1/n18-1202>
PEYRARD, M. Studying Summarization Evaluation Metrics in the Appropriate Scoring Range. Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP). Anais...Online: Association for Computational Linguistics, 2019. Disponível em: <https://aclanthology.org/P19-1502/>
PIĘKOS, P.; MALINOWSKI, M.; MICHALEWSKI, H. Measuring and Improving BERTs Mathematical Abilities by Predicting the Order of Reasoning. Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers). Anais...Online: Association for Computational Linguistics, ago. 2021. Disponível em: <https://aclanthology.org/2021.acl-short.49>
PIRES, R. et al. Sabiá: Portuguese Large Language Models. (M. C. Naldi, R. A. C. Bianchi, Eds.)Intelligent Systems. Anais...Cham: Springer Nature Switzerland, 2023.
PLOUG, T. The Right Not to Be Subjected to AI Profiling Based on Publicly Available DataPrivacy and the Exceptionalism of AI Profiling. Philosophy & Technology, v. 36, n. 1, p. 14, mar. 2023.
POLO, F. M. et al. Efficient Multi-Prompt Evaluation of LLMs. Proceedings of the 38th Conference on Neural Information Processing Systems (NeurIPS 2024), 2024.
POMBAL, J. et al. M-Prometheus: A Suite of Open Multilingual LLM Judges. CoRR, v. abs/2504.04953, 2025.
PROTOTYPEJAM. Lake Merritt: AI Evaluation Workbench. https://prototypejam.github.io/lake_merritt/, 2025.
PROVILKOV, I.; EMELIANENKO, D.; VOITA, E. BPE-Dropout: Simple and Effective Subword Regularization. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Anais...Online: Association for Computational Linguistics, jul. 2020.
RACKAUCKAS, Z. Rag-fusion: a new take on retrieval-augmented generation. arXiv preprint arXiv:2402.03367, 2024.
RADFORD, A. et al. Improving Language Understanding by Generative Pre-Training., 2018. Disponível em: <https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf>
RADFORD, A. et al. Language Models are Unsupervised Multitask Learners. a2019.
RADFORD, A. et al. Language models are unsupervised multitask learners. OpenAI blog, v. 1, n. 8, p. 9, b2019.
RADFORD, A. et al. Learning Transferable Visual Models From Natural Language Supervision., 2021. Disponível em: <https://arxiv.org/abs/2103.00020>
RADFORD, A.; NARASIMHAN, K. Improving Language Understanding by Generative Pre-Training. 2018.
RAE, J. W. et al. Scaling Language Models: Methods, Analysis & Insights from Training Gopher. CoRR, v. abs/2112.11446, 2021.
RAE, J. W. et al. Scaling Language Models: Methods, Analysis & Insights from Training Gopher., 2022. Disponível em: <https://arxiv.org/abs/2112.11446>
RAFAILOV, R. et al. Direct Preference Optimization: Your Language Model is Secretly a Reward Model. Advances in Neural Information Processing Systems (NeurIPS). Anais...2023.
RAFFEL, C. et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. Journal of Machine Learning Research, v. 21, n. 140, p. 1–67, a2020.
RAFFEL, C. et al. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res., v. 21, n. 1, p. 140:5485–140:5551, jan. b2020.
RAFFEL, C. et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer., 2023. Disponível em: <https://arxiv.org/abs/1910.10683>
RAJI, D. et al. AI and the Everything in the Whole Wide World Benchmark. (J. Vanschoren, S. Yeung, Eds.)Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks. Anais...Curran, 2021. Disponível em: <https://datasets-benchmarks-proceedings.neurips.cc/paper_files/paper/2021/file/084b6fbb10729ed4da8c3d3f5a3ae7c9-Paper-round2.pdf>
RAJPURKAR, P. et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text. (J. Su, K. Duh, X. Carreras, Eds.)Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. Anais...Austin, Texas: Association for Computational Linguistics, nov. 2016. Disponível em: <https://aclanthology.org/D16-1264>
REAL, L.; CARVALHO, A.; SILVA, A. DA. Avaliação de Grandes Modelos de Linguagem: Fundamentos, Métodos Tradicionais e Desafios Atuais. Em: CASELI, H. M.; NUNES, M. G. V. (Eds.). Processamento de Linguagem Natural: Conceitos, Técnicas e Aplicações em Português. 3. ed. [s.l.] BPLN, 2024.
REAL, L.; FONSECA, E.; GONÇALO OLIVEIRA, H. The ASSIN 2 Shared Task: A Quick Overview. Computational Processing of the Portuguese Language: 14th International Conference, PROPOR 2020, Evora, Portugal, March 2–4, 2020, Proceedings. Anais...Berlin, Heidelberg: Springer-Verlag, 2020. Disponível em: <https://doi.org/10.1007/978-3-030-41505-1_39>
REHAK, R. AI Narrative Breakdown. A Critical Assessment of Power and Promise. Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency. Anais...: FAccT ’25.New York, NY, USA: Association for Computing Machinery, 2025. Disponível em: <https://doi.org/10.1145/3715275.3732083>
REIMERS, N.; GUREVYCH, I. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing. Anais...Association for Computational Linguistics, nov. 2019. Disponível em: <https://arxiv.org/abs/1908.10084>
REIMERS, N.; GUREVYCH, I. Making Monolingual Sentence Embeddings Multilingual using Knowledge Distillation. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing. Anais...Association for Computational Linguistics, nov. 2020. Disponível em: <https://arxiv.org/abs/2004.09813>
RIBEIRO, M. T. et al. Beyond Accuracy: Behavioral Testing of NLP Models with CheckList. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Anais...Online: Association for Computational Linguistics, jul. 2020. Disponível em: <https://aclanthology.org/2020.acl-main.442>
RÍO, B. G. DEL; VAAHTIO, T. Improving LLM systems with A/B testing. https://www.flow-ai.com/blog/improving-llm-systems-with-a-b-testing, 2024.
RODRIGUES, J. et al. Advancing Neural Encoding of Portuguese with Transformer Albertina PT-. CoRR, v. abs/2305.06721, 2023.
RODRIGUES, R. C. et al. Portuguese Language Models and Word Embeddings: Evaluating on Semantic Similarity Tasks. (P. Quaresma et al., Eds.)Computational Processing of the Portuguese Language. Anais...Springer Nature Switzerland AG: Springer International Publishing, 2020.
ROGERS, A. Changing the World by Changing the Data. Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Anais...Online: Association for Computational Linguistics, 2021. Disponível em: <https://aclanthology.org/2021.acl-long.170>
ROGERS, A.; BALDWIN, T.; LEINS, K. Just What do You Think You’re Doing, Dave?” A Checklist for Responsible Data Use in NLP. Findings of the Association for Computational Linguistics: EMNLP 2021. Anais...Punta Cana, Dominican Republic: Association for Computational Linguistics, 2021. Disponível em: <https://aclanthology.org/2021.findings-emnlp.414>
ROMERA-PAREDES, B.; TORR, P. H. S. An embarrassingly simple approach to zero-shot learning. (F. R. Bach, D. M. Blei, Eds.)Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 6-11 July 2015. Anais...: JMLR Workshop e Conference Proceedings.JMLR.org, 2015. Disponível em: <http://proceedings.mlr.press/v37/romera-paredes15.html>
RUSSEL, S. Human Compatible Artificial Intelligence and the Problem of Control. [s.l.] Penguin Books, 2019.
SAI, A. B.; MOHANKUMAR, A. K.; KHAPRA, M. M. A Survey of Evaluation Metrics Used for NLG Systems. ACM Comput. Surv., v. 55, n. 2, p. 26:1–26:39, 2023.
SAID, H. 40 Large Language Model Benchmarks and The Future of LLMs. Arize AI; https://arize.com/blog/llm-benchmarks-mmlu-codexglue-gsm8k, 2025.
SAMPAIO, R. C.; SABBATINI, M.; LIMONGI, R. Diretrizes para o uso ético e responsável da inteligência artificial generativa: um guia prático para pesquisadores. Boletim Técnico do PPEC, v. 10, p. e025003–e025003, 2024.
SAMUYLOVA, E. LLM‑as‑a‑Judge: A Complete Guide to Using LLMs for Evaluations. https://www.evidentlyai.com/llm-guide/llm-as-a-judge, 2025.
SÁNCHEZ, L. C. Ethical Considerations and Best Practices in LLM Development. https://neptune.ai/blog/llm-ethical-considerations, 2024.
SANH, V. et al. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. CoRR, v. abs/1910.01108, 2019.
SANTOS, A. A. et al. O teste de Cloze na avaliação da compreensão em leitura. Psicologia: reflexão e crı́tica, v. 15, p. 549–560, 2002.
SANTOS, D. O projecto Processamento Computacional do Português: Balanço e perspectivas. (M. das Graças Volpe Nunes, Ed.)V Encontro para o processamento computacional da língua portuguesa escrita e falada (PROPOR 2000). Anais...São Paulo: ICMC/USP, 2000. Disponível em: <https://www.linguateca.pt/Diana/download/SantosPROPOR2000.pdf>
SANTOS, H.; WOLOSZYN, V.; VIEIRA, R. BlogSet-BR: A Brazilian Portuguese Blog Corpus. (N. Calzolari et al., Eds.)Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018). Anais...Miyazaki, Japan: European Language Resources Association (ELRA), 2018. Disponível em: <https://aclanthology.org/L18-1105/>
SARDINHA, T. B.; FILHO, J. L. M.; ALAMBERT, E. Manual Corpus Brasileiro. [s.l: s.n.].
SARDINHA, T. B.; KAUFFMANN, C.; ACUNZO, C. M. A multi-dimensional analysis of register variation in Brazilian Portuguese. Corpora, v. 9, n. 2, p. 239–271, 2014.
SARTHI, P. et al. RAPTOR: Recursive Abstractive Processing for Tree-Organized Retrieval. International Conference on Learning Representations (ICLR). Anais...2024.
SAWCZYN, A. et al. FactSelfCheck: Fact-Level Black-Box Hallucination Detection for LLMs. arXiv, 2025.
SCAO, T. L. et al. BLOOM: A 176B-Parameter Open-Access Multilingual Language Model. CoRR, v. abs/2211.05100, 2022.
SCHAAKE, M. The Tech Coup: How to Save Democracy from Silicon Valley. [s.l.] Princeton University Press, 2024.
SCHEUERMAN, M. K.; HANNA, A.; DENTON, E. Do Datasets Have Politics? Disciplinary Values in Computer Vision Dataset Development. Proceedings of the ACM on Human-Computer Interaction, v. 5, n. CSCW2, p. 1–37, out. 2021.
SCHICK, T.; SCHÜTZE, H. Exploiting Cloze-Questions for Few-Shot Text Classification and Natural Language Inference. (P. Merlo, J. Tiedemann, R. Tsarfaty, Eds.)Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, EACL 2021, Online, April 19 - 23, 2021. Anais...Association for Computational Linguistics, 2021. Disponível em: <https://doi.org/10.18653/v1/2021.eacl-main.20>
SCHIRRU, L. Direito autoral e Inteligência Artificial: autoria e titularidade nos produtos da IA. [s.l.] Dialetica, 2023.
SCHLANGEN, D. Targeting the Benchmark: On Methodology in Current Natural Language Processing Research. Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers). Anais...Online: Association for Computational Linguistics, ago. 2021. Disponível em: <https://aclanthology.org/2021.acl-short.85>
SCHMIDHUBER, J.; HEIL, S. Sequential neural text compression. IEEE Transactions on Neural Networks, v. 7, n. 1, p. 142–146, 1996.
SCHNABEL, T. et al. Evaluation methods for unsupervised word embeddings. (L. Màrquez, C. Callison-Burch, J. Su, Eds.)Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing. Anais...Lisbon, Portugal: Association for Computational Linguistics, set. 2015. Disponível em: <https://aclanthology.org/D15-1036/>
SCHULMAN, J. et al. Proximal Policy Optimization Algorithms. arXiv preprint arXiv:1707.06347. Anais...2017.
SCHUSTER, M.; NAKAJIMA, K. Japanese and Korean voice search. 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Anais...2012.
SCHWARTZ, L. Primum Non Nocere: Before working with Indigenous data, the ACL must confront ongoing colonialism. Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). Anais...Dublin, Ireland: Association for Computational Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.acl-short.82>
SELVAN, R. et al. Carbon Footprint of Selecting and Training Deep Learning Models for Medical Image Analysis. (L. Wang et al., Eds.)Medical Image Computing and Computer Assisted Intervention – MICCAI 2022. Anais...Cham: Springer Nature Switzerland, 2022. Disponível em: <https://doi.org/10.1007/978-3-031-16443-9_49>
SERRAS, F. R. et al. Exploring Computational Discernibility of Discourse Domains in Brazilian Portuguese within the Carolina Corpus. (P. Gamallo et al., Eds.)Proceedings of the 16th International Conference on Computational Processing of Portuguese - Vol. 1. Anais...Santiago de Compostela, Galicia/Spain: Association for Computational Lingustics, mar. 2024. Disponível em: <https://aclanthology.org/2024.propor-1.26/>
SHAH, D.; SCHWARTZ, H. A.; HOVY, D. Predictive Biases in Natural Language Processing Models: A Conceptual Framework and Overview. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Anais...Online: Association for Computational Linguistics, 2020. Disponível em: <http://arxiv.org/abs/1912.11078>
SHAH, R. et al. Goal Misgeneralization in Deep Reinforcement Learning. arXiv preprint arXiv:2210.01790, 2022.
SHANNON, C. E. Prediction and entropy of printed English. Bell System Technical Journal, v. 30, n. 1, p. 50–64, 1951.
SHAQIRI, M. et al. Differences between the correlation coefficients Pearson, Kendall and Spearman. 5th International Conference of Natural Science and Mathematics, nov. 2023.
SHENG, E. et al. The Woman Worked as a Babysitter: On Biases in Language Generation. (K. Inui et al., Eds.)Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Anais...Hong Kong, China: Association for Computational Linguistics, nov. 2019. Disponível em: <https://aclanthology.org/D19-1339/>
SHI, H. et al. Continual learning of large language models: A comprehensive survey. ACM Computing Surveys, 2024.
SHI, Z.; LIPANI, A. Don’t Stop Pretraining? Make Prompt-based Fine-tuning Powerful Learner., 2023. Disponível em: <https://arxiv.org/abs/2305.01711>
SHMUELI, B. et al. Beyond Fair Pay: Ethical Implications of NLP Crowdsourcing. Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Anais...Online: Association for Computational Linguistics, 2021. Disponível em: <https://aclanthology.org/2021.naacl-main.295>
SHUMAILOV, I. et al. AI models collapse when trained on recursively generated data. Nature, v. 631, n. 8022, p. 755–759, 2024.
ŞIMŞEK, C. AI resistance: Who says no to AI and why? Zenodo, 2025. Disponível em: <https://zenodo.org/doi/10.5281/zenodo.16893847>
ŞIMŞEK, C.; YASAR, A. G. From Rejection to Regulation: Mapping the Landscape of AI Resistance. 2025.
SØGAARD, A.; PLANK, B.; HOVY, D. Selection Bias, Label Bias, and Bias in Ground Truth. (Q. Liu, F. Xia, Eds.)Proceedings of COLING 2014, the 25th International Conference on Computational Linguistics: Tutorial Abstracts. Anais...Dublin, Ireland: Dublin City University; Association for Computational Linguistics, ago. 2014. Disponível em: <https://aclanthology.org/C14-3005/>
SOLAIMAN, I. et al. Evaluating the Social Impact of Generative AI Systems. Em: The Oxford Handbook of the Foundations and Regulation of Generative AI. [s.l.] Oxford University Press, 2025.
SOUZA, F.; NOGUEIRA, R.; LOTUFO, R. BERTimbau: pretrained BERT models for Brazilian Portuguese. (R. Cerri, R. C. Prati, Eds.)Proceedings of the 2020 Brazilian Conference on Intelligent Systems. Anais...Springer International Publishing, 2020.
STIENNON, N. et al. Learning to summarize with human feedback. (H. Larochelle et al., Eds.)Advances in Neural Information Processing Systems. Anais...Curran Associates, Inc., a2020. Disponível em: <https://proceedings.neurips.cc/paper_files/paper/2020/file/1f89885d556929e98d3ef9b86448f951-Paper.pdf>
STIENNON, N. et al. Learning to Summarize with Human Feedback. Advances in Neural Information Processing Systems (NeurIPS), b2020.
STRUBELL, E.; GANESH, A.; MCCALLUM, A. Energy and Policy Considerations for Deep Learning in NLP. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Anais...Florence, Italy: Association for Computational Linguistics, 2019. Disponível em: <https://www.aclweb.org/anthology/P19-1355>
STURZENEKER, M. et al. Carolina’s Methodology: building a large corpus with provenance and typology information. (C. Trojahn et al., Eds.)Proceedings of the Second Workshop on Digital Humanities and Natural Language Processing (2nd DHandNLP 2022) co-located with International Conference on the Computational Processing of Portuguese (PROPOR 2022). Anais...Fortaleza, Brazil: 2022. Disponível em: <https://ceur-ws.org/Vol-3128/paper4.pdf>
SUN, K.; DREDZE, M. Amuro & Char: Analyzing the Relationship between Pre-Training and Fine-Tuning of Large Language Models. (V. Adlakha et al., Eds.)Proceedings of the 10th Workshop on Representation Learning for NLP (RepL4NLP-2025). Anais...Albuquerque, NM: Association for Computational Linguistics, 2025. Disponível em: <https://aclanthology.org/2025.repl4nlp-1.11/>
SURESH, H.; GUTTAG, J. V. A Framework for Understanding Sources of Harm throughout the Machine Learning Life Cycle. Equity and Access in Algorithms, Mechanisms, and Optimization, p. 1–9, out. 2021.
SUTSKEVER, I.; VINYALS, O.; LE, Q. V. Sequence to Sequence Learning with Neural Networks. (Z. Ghahramani et al., Eds.)Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, December 8-13 2014, Montreal, Quebec, Canada. Anais...2014. Disponível em: <https://proceedings.neurips.cc/paper/2014/hash/a14ac55a4f27472c5d894ec1c3c743d2-Abstract.html>
TAMKIN, A. et al. Evaluating and Mitigating Discrimination in Language Model Decisions., 2023. Disponível em: <https://arxiv.org/abs/2312.03689>
TANG, Y. et al. Multilingual Translation with Extensible Multilingual Pretraining and Finetuning. CoRR, v. abs/2008.00401, 2020.
TAYLOR, R. et al. Galactica: A Large Language Model for Science. CoRR, v. abs/2211.09085, 2022.
TAYLOR, W. L. “Cloze procedure”: A new tool for measuring readability. Journalism quarterly, v. 30, n. 4, p. 415–433, 1953.
TEI CONSORTIUM. TEI P5: Guidelines for Electronic Text Encoding and Interchange. [s.l.] Retrieved May 20, 2021, 2021.
THAKUR, A. S. et al. JUDGING THE JUDGES: EVALUATING ALIGNMENT AND VULNERABILITIES IN LLMs-AS-JUDGES. arXiv preprint arXiv:2406.12624v5, 2025.
THOPPILAN, R. et al. LaMDA: Language Models for Dialog Applications. CoRR, v. abs/2201.08239, 2022.
THUN, M. VON; HANLEY, D. A. Stopping Big Tech from Becoming Big AI: A Roadmap for Using Competition Policy to Keep Artificial Intelligence Open for All. Open Markets Institute, 2024. Disponível em: <https://www.openmarketsinstitute.org/publications/report-stopping-big-tech-big-ai-roadmap>
TOLLES, J.; MEURER, W. J. Logistic Regression: Relating Patient Characteristics to Outcomes. JAMA, v. 316, n. 5, p. 533–534, ago. 2016.
TONIAZZO, D.; BARBOSA, T.; RUARO, R. O Direito à Explicação nas Decisões Automatizadas: uma Abordagem Comparativa Entre o Ordenamento Brasileiro e Europeu. Revista Internacional Consinter de Direito, v. 13, p. 55–69, dez. 2021.
TOUVRON, H. et al. LLaMA: Open and Efficient Foundation Language Models. CoRR, v. abs/2302.13971, a2023.
TOUVRON, H. et al. LLaMA: Open and Efficient Foundation Language Models., b2023. Disponível em: <https://arxiv.org/abs/2302.13971>
TRAUTMANN, D. et al. Measuring the Groundedness of Legal Question-Answering Systems. Proceedings of the Natural Legal Language Processing Workshop 2024. Anais...Singapore: Association for Computational Linguistics, 2024. Disponível em: <https://aclanthology.org/2024.nllp-1.14>
ULMER, D. et al. Experimental Standards for Deep Learning in Natural Language Processing Research. Findings of the Association for Computational Linguistics: EMNLP 2022. Anais...Abu Dhabi, United Arab Emirates: Association for Computational Linguistics, dez. 2022. Disponível em: <https://aclanthology.org/2022.findings-emnlp.196>
VAROQUAUX, G.; LUCCIONI, S.; WHITTAKER, M. Hype, Sustainability, and the Price of the Bigger-is-Better Paradigm in AI. Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency. Anais...Athens Greece: ACM, jun. 2025. Disponível em: <https://dl.acm.org/doi/10.1145/3715275.3732006>
VASWANI, A. et al. Attention is All you Need. (I. Guyon et al., Eds.)Advances in Neural Information Processing Systems. Anais...Curran Associates, Inc., a2017. Disponível em: <https://proceedings.neurips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html>
VASWANI, A. et al. Attention is all you need. Advances in neural information processing systems, v. 30, b2017.
VONGTHONGSRI, K. G-Eval Simply Explained: LLM-as-a-Judge for LLM Evaluation. Confident AI; https://www.confident-ai.com/blog/g-eval-the-definitive-guide, 2025.
WAGNER FILHO, J. A. et al. The brWaC Corpus: A New Open Resource for Brazilian Portuguese. (N. Calzolari et al., Eds.)Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018). Anais...Miyazaki, Japan: European Language Resources Association (ELRA), 2018. Disponível em: <https://aclanthology.org/L18-1686>
WANG, A. et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding. Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP. Anais...Brussels, Belgium: Association for Computational Linguistics, nov. 2018. Disponível em: <https://aclanthology.org/W18-5446/>
WANG, A. et al. SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems. Advances in Neural Information Processing Systems, v. 32, p. 3261–3275, b2019.
WANG, A. et al. SuperGLUE: a stickier benchmark for general-purpose language understanding systems. Em: Proceedings of the 33rd International Conference on Neural Information Processing Systems. Red Hook, NY, USA: Curran Associates Inc., 2019a.
WANG, B.; KOMATSUZAKI, A. GPT-J-6B: A 6 Billion Parameter Autoregressive Language Model. https://github.com/kingoflolz/mesh-transformer-jax, 2021.
WANG, R. et al. Can LLMs Replace Human Evaluators? An Empirical Study of LLM-as-a-Judge in Software Engineering. Proc. ACM Softw. Eng., v. 2, n. ISSTA, jun. 2025.
WANG, Y. et al. CodeT5: Identifier-aware Unified Pre-trained Encoder-Decoder Models for Code Understanding and Generation. (M.-F. Moens et al., Eds.)Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021. Anais...Association for Computational Linguistics, 2021. Disponível em: <https://doi.org/10.18653/v1/2021.emnlp-main.685>
WANG, Y. et al. PandaLM: An Automatic Evaluation Benchmark for LLM Instruction Tuning Optimization. The Twelfth International Conference on Learning Representations, ICLR 2024, Vienna, Austria, May 7-11, 2024. Anais...OpenReview.net, 2024. Disponível em: <https://openreview.net/forum?id=5Nn2BLV7SB>
WEBER, M. et al. RedPajama: an Open Dataset for Training Large Language Models., 2024. Disponível em: <https://arxiv.org/abs/2411.12372>
WEI, J. et al. Emergent Abilities of Large Language Models. Trans. Mach. Learn. Res., v. 2022, b2022.
WEI, J. et al. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. NeurIPS. Anais...a2022. Disponível em: <http://papers.nips.cc/paper\_files/paper/2022/hash/9d5609613524ecf4f15af0f7b31abca4-Abstract-Conference.html>
WEIDINGER, L. et al. Ethical and social risks of harm from Language Models. arXiv:2112.04359 [cs], dez. 2021.
WEIGEND, A. Data for the People: How to Make Our Post-Privacy Economy Work for You. [s.l.] Basic Books, 2017.
WEIZENBAUM, J. Computermacht und Gesellschaft : freie Reden / Joseph Weizenbaum. Hrsg. von Gunna Wendt .. Original-Ausgabe, 1. Auflage ed. Frankfurt am Main: [s.n.].
WENDEHORST, C. Liability for Artificial Intelligence: The Need to Address Both Safety Risks and Fundamental Rights Risks. Em: VOENEKY, S. et al. (Eds.). The Cambridge Handbook of Responsible Artificial Intelligence: Interdisciplinary Perspectives. Cambridge Law Handbooks. [s.l.] Cambridge University Press, 2022. p. 187–209.
WENZEK, G. et al. CCNet: Extracting High Quality Monolingual Datasets from Web Crawl Data. (N. Calzolari et al., Eds.)Proceedings of the Twelfth Language Resources and Evaluation Conference. Anais...Marseille, France: European Language Resources Association, 2020. Disponível em: <https://aclanthology.org/2020.lrec-1.494/>
WERBOS, P. J. Backpropagation through time: what it does and how to do it. Proc. IEEE, v. 78, n. 10, p. 1550–1560, 1990.
WESTENBERGER, J.; SCHULER, K.; SCHLEGEL, D. Failure of AI projects: understanding the critical factors. Procedia Computer Science, v. 196, p. 69–76, 2022.
WHITTAKER, M. et al. Disability, bias, and AI. AI Now Institute, v. 8, n. 11, 2019.
WHITTAKER, M. The steep cost of capture. Interactions, v. 28, n. 6, p. 50–55, nov. 2021.
WIEGREFFE, S.; PINTER, Y. Attention is not not Explanation. (K. Inui et al., Eds.)Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Anais...Hong Kong, China: Association for Computational Linguistics, nov. 2019. Disponível em: <https://aclanthology.org/D19-1002>
WIELING, M.; RAWEE, J.; VAN NOORD, G. Reproducibility in Computational Linguistics: Are We Willing to Share? Computational Linguistics, v. 44, n. 4, p. 641–649, dez. 2018.
WOLF, T. et al. Transformers: State-of-the-Art Natural Language Processing. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. Anais...Online: Association for Computational Linguistics, out. 2020. Disponível em: <https://www.aclweb.org/anthology/2020.emnlp-demos.6>
WOLFRAM, W. Variation and Language: Overview. Em: BROWN, K. (Ed.). Encyclopedia of Language & Linguistics (Second Edition). Second Edition ed. Oxford: Elsevier, 2006. p. 333–341.
WRIGHT, B. Manufacturing Reality: Slavoj Zizek and the Reality of the Virtual. LondonBen Wright Film Productions, 2004.
WU, T. The attention merchants : from the daily newspaper to social media : how our time and attention is harvested and sold. London: [s.n.].
WU, X.-K. et al. LLM Fine-Tuning: Concepts, Opportunities, and Challenges. Big Data and Cognitive Computing, v. 9, n. 4, p. 87, abr. 2025.
WU, Y. et al. Memorizing Transformers. The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. Anais...OpenReview.net, 2022. Disponível em: <https://openreview.net/forum?id=TrjbxzRcnf->
XIE, S. M. et al. An Explanation of In-context Learning as Implicit Bayesian Inference. The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. Anais...OpenReview.net, 2022. Disponível em: <https://openreview.net/forum?id=RdJVFCHjUMI>
XIONG, K. et al. Com2 : A Causal-Guided Benchmark for Exploring Complex Commonsense Reasoning in Large Language Models. (W. Che et al., Eds.)Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Anais...Vienna, Austria: Association for Computational Linguistics, jul. 2025. Disponível em: <https://aclanthology.org/2025.acl-long.785/>
XIONG, R. et al. On Layer Normalization in the Transformer Architecture. Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event. Anais...: Proceedings of Machine Learning Research.PMLR, 2020. Disponível em: <http://proceedings.mlr.press/v119/xiong20b.html>
XU, A. et al. Detoxifying Language Models Risks Marginalizing Minority Voices. (K. Toutanova et al., Eds.)Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Anais...Online: Association for Computational Linguistics, jun. 2021. Disponível em: <https://aclanthology.org/2021.naacl-main.190/>
XU, H.; ASHLEY, K. A question-answering approach to evaluating legal summaries., 2023. Disponível em: <https://arxiv.org/abs/2309.15016>
XU, Q.; HE, X. Security Challenges in Natural Language Processing Models. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: Tutorial Abstracts. Anais...Singapore: Association for Computational Linguistics, 2023. Disponível em: <https://aclanthology.org/2023.emnlp-tutorial.2>
XU, W.; RUDNICKY, A. Can artificial neural networks learn language models? Proc. 6th International Conference on Spoken Language Processing (ICSLP 2000). Anais...2000.
XU, Y. et al. Hard Sample Aware Prompt-Tuning. (A. Rogers, J. L. Boyd-Graber, N. Okazaki, Eds.)Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, Toronto, Canada, July 9-14, 2023. Anais...Association for Computational Linguistics, 2023. Disponível em: <https://aclanthology.org/2023.acl-long.690>
XUE, L. et al. mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer. (K. Toutanova et al., Eds.)Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021, Online, June 6-11, 2021. Anais...Association for Computational Linguistics, 2021. Disponível em: <https://doi.org/10.18653/v1/2021.naacl-main.41>
YAMAGUCHI, A. et al. Frustratingly Simple Pretraining Alternatives to Masked Language Modeling. Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Anais...Online; Punta Cana, Dominican Republic: Association for Computational Linguistics, nov. 2021. Disponível em: <https://aclanthology.org/2021.emnlp-main.249>
YAN, S.-Q. et al. Corrective Retrieval Augmented Generation., 2024. Disponível em: <https://arxiv.org/abs/2401.15884>
YANG, T. et al. Ethics of Data Work. Principles for Academic Data Work Requesters. Weizenbaum Institute, 2025. Disponível em: <https://www.weizenbaum-library.de/handle/id/920>
YANG, Z. et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding. (H. M. Wallach et al., Eds.)Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada. Anais...2019. Disponível em: <https://proceedings.neurips.cc/paper/2019/hash/dc6a7e655d7e5840e66733e9ee67cc69-Abstract.html>
YAO, X. et al. NLP From Scratch Without Large-Scale Pretraining: A Simple and Efficient Framework., 2022. Disponível em: <https://arxiv.org/abs/2111.04130>
YE, Y. et al. LIMO: Less is More for Reasoning. arXiv, jul. 2025. Disponível em: <http://arxiv.org/abs/2502.03387>. Acesso em: 7 ago. 2025
YEH, Y.-T.; ESKÉNAZI, M.; MEHRI, S. A Comprehensive Assessment of Dialog Evaluation Metrics. ArXiv, v. abs/2106.03706, 2021.
YEO, W. et al. UniversalRAG: Retrieval-Augmented Generation over Corpora of Diverse Modalities and Granularities. arXiv preprint arXiv:2504.20734, 2025.
YUAN, W.; NEUBIG, G.; LIU, P. BARTScore: Evaluating Generated Text as Text Generation. Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, virtual. Anais...2021. Disponível em: <https://proceedings.neurips.cc/paper/2021/hash/e4d2b6e6fdeca3e60e0f1a62fee3d9dd-Abstract.html>
ZHANG, D.; XU, Z.; ZHAO, W. LLMs and Copyright Risks: Benchmarks and Mitigation Approaches. (M. Lomeli, S. Swayamdipta, R. Zhang, Eds.)Proceedings of the 2025 Annual Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 5: Tutorial Abstracts). Anais...Albuquerque, New Mexico: Association for Computational Linguistics, a2025. Disponível em: <https://aclanthology.org/2025.naacl-tutorial.7/>
ZHANG, K. et al. CitaLaw: Enhancing LLM with Citations in Legal Domain. (W. Che et al., Eds.)Findings of the Association for Computational Linguistics: ACL 2025. Anais...Vienna, Austria: Association for Computational Linguistics, jul. b2025. Disponível em: <https://aclanthology.org/2025.findings-acl.583/>
ZHANG, S. et al. Instruction Tuning for Large Language Models: A Survey., c2025. Disponível em: <https://arxiv.org/abs/2308.10792>
ZHANG, T. et al. BERTScore: Evaluating Text Generation with BERT. 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. Anais...OpenReview.net, 2020. Disponível em: <https://openreview.net/forum?id=SkeHuCVFDr>
ZHANG, Y. et al. 🧜Siren’s Song in the AI Ocean: A Survey on Hallucination in Large Language Models. Computational Linguistics, v. 51, n. 4, p. 1373–1418, dez. d2025.
ZHAO, W. X. et al. A Survey of Large Language Models. CoRR, v. abs/2303.18223, 2023.
ZHAO, Y. et al. One Token to Fool LLM-as-a-Judge., 2025. Disponível em: <https://arxiv.org/abs/2507.08794>
ZHOU, C. et al. LIMA: Less Is More for Alignment. CoRR, v. abs/2305.11206, a2023.
ZHOU, H. et al. A survey of large language models in medicine: Progress, application, and challenge. arXiv preprint arXiv:2311.05112, b2023.
ZIEGLER, D. M. et al. Fine-Tuning Language Models from Human Preferences. CoRR, v. abs/1909.08593, 2019.
ZUBOFF, S. The age of surveillance capitalism : the fight for a human future at the new frontier of power. First edition ed. New York: [s.n.].