Referências
ABADJI, J. et al. Towards a Cleaner Document-Oriented
Multilingual Crawled Corpus., a2022. Disponível em: <https://arxiv.org/abs/2201.06642>
ABADJI, J. et al. Towards a Cleaner Document-Oriented
Multilingual Crawled Corpus. Proceedings of the Thirteenth
Language Resources and Evaluation Conference.
Anais...Marseille, France: European Language Resources
Association, jun. b2022. Disponível em: <https://aclanthology.org/2022.lrec-1.463>
ABDALLA, M. et al. The Elephant in the
Room: Analyzing the Presence of
Big Tech in Natural
Language Processing
Research. Proceedings of the 61st
Annual Meeting of the Association
for Computational Linguistics
(Volume 1: Long Papers).
Anais...Toronto, Canada: Association for Computational
Linguistics, 2023. Disponível em: <https://aclanthology.org/2023.acl-long.734>
ABERCROMBIE, G. et al. Mirages. On
Anthropomorphism in Dialogue
Systems. Proceedings of the 2023
Conference on Empirical Methods
in Natural Language Processing.
Anais...Singapore: Association for Computational
Linguistics, 2023. Disponível em: <https://aclanthology.org/2023.emnlp-main.290>
ABID, A.; FAROOQI, M.; ZOU, J. Persistent
Anti-Muslim Bias in
Large Language Models.
Proceedings of the 2021 AAAI/ACM
Conference on AI, Ethics, and
Society. Anais...Virtual Event USA: ACM,
jul. 2021. Disponível em: <https://dl.acm.org/doi/10.1145/3461702.3462624>
AGHAJANYAN, A.; GUPTA, S.; ZETTLEMOYER, L. Intrinsic
Dimensionality Explains the Effectiveness of Language Model
Fine-Tuning. (C. Zong et al., Eds.)Proceedings of the 59th
Annual Meeting of the Association for Computational Linguistics and the
11th International Joint Conference on Natural Language Processing,
ACL/IJCNLP 2021, (Volume 1: Long Papers), Virtual Event,
August 1-6, 2021. Anais...Association for Computational
Linguistics, 2021. Disponível em: <https://doi.org/10.18653/v1/2021.acl-long.568>
AGIRRE, E. Cross-Lingual Word
Embeddings. Computational Linguistics, v.
46, n. 1, p. 245–248, mar. 2020.
ALBALAK, A. et al. A Survey on Data Selection for Language
Models., 2024. Disponível em: <https://arxiv.org/abs/2402.16827>
ALKHAMISSI, B. et al. Investigating Cultural Alignment of Large
Language Models. (L.-W. Ku, A. Martins, V. Srikumar,
Eds.)Proceedings of the 62nd Annual Meeting of the Association for
Computational Linguistics (Volume 1: Long Papers).
Anais...Bangkok, Thailand: Association for
Computational Linguistics, ago. 2024. Disponível em: <https://aclanthology.org/2024.acl-long.671/>
AMARAL, V. DO et al. A RAG-Powered Academic Chatbot with
Ontology-Driven Factual Verification. Anais Estendidos do XL
Simpósio Brasileiro de Bancos de Dados. Anais...Porto
Alegre, RS, Brasil: SBC, 2025. Disponível em: <https://sol.sbc.org.br/index.php/sbbd_estendido/article/view/37615>
AMODEI, D. et al. Concrete Problems in AI Safety. arXiv
preprint arXiv:1606.06565. Anais...2016.
ANISUZZAMAN, D. et al. Fine-tuning large language models for specialized
use cases. Mayo Clinic Proceedings: Digital Health, v.
3, n. 1, p. 100184, 2025.
ARVAN, M.; PINA, L.; PARDE, N. Reproducibility in
Computational Linguistics: Is
Source Code Enough?
Proceedings of the 2022 Conference on
Empirical Methods in Natural
Language Processing.
Anais...Abu Dhabi, United Arab Emirates: Association
for Computational Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.emnlp-main.150>
ASAI, A. et al. Self-rag: Learning to retrieve, generate, and critique
through self-reflection. 2024.
ASSUNÇÃO, I. V.; JANSON, S. F. Afinal,
o que é privacidade? Um panorama histórico do direito à privacidade no
ordenamento constitucional brasileiro. Internet &
Sociedade, v. 5, n. 1, 2024.
BAHDANAU, D.; CHO, K.; BENGIO, Y. Neural Machine Translation by
Jointly Learning to Align and Translate. (Y. Bengio, Y. LeCun,
Eds.)3rd International Conference on Learning Representations,
ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference
Track Proceedings. Anais...San Diego, California.:
2015. Disponível em: <http://arxiv.org/abs/1409.0473>
BAI, J. et al. Qwen technical report. arXiv preprint
arXiv:2309.16609, 2023.
BAI, X. et al. Explicitly unbiased large
language models still form biased associations. Proceedings
of the National Academy of Sciences, v. 122, n. 8, p.
e2416228122, fev. 2025.
BANERJEE, S.; LAVIE, A. METEOR: An Automatic Metric
for MT Evaluation with Improved Correlation with Human
Judgments. (J. Goldstein et al., Eds.)Proceedings of the
ACL Workshop on Intrinsic and Extrinsic Evaluation Measures
for Machine Translation and/or Summarization.
Anais...Ann Arbor, Michigan: Association for
Computational Linguistics, jun. 2005. Disponível em: <https://aclanthology.org/W05-0909>
BARRETT, A. M. et al. Benchmark Early and Red Team Often: A
framework for assessing and managing dual-hazards of AI
foundational models. UC Berkeley Center for
Long-Term Cybersecurity, 2024. Disponível em: <https://cltc.berkeley.edu/wp-content/uploads/2024/05/Dual-Use-Benchmark-Early-Red-Team-Often.pdf>
BARRY LYNN, M. VON T.; MONTOYA, K. AI in the Public
Interest: Confronting the Monopoly Threat. Open Markets
Institute, 2023. Disponível em: <https://www.openmarketsinstitute.org/publications/report-ai-in-the-public-interest-confronting-the-monopoly-threat>
BELINKOV, Y.; GEHRMANN, S.; PAVLICK, E. Interpretability and
Analysis in Neural NLP.
Proceedings of the 58th Annual Meeting of the
Association for Computational
Linguistics: Tutorial Abstracts.
Anais...Online: Association for Computational
Linguistics, 2020. Disponível em: <https://www.aclweb.org/anthology/2020.acl-tutorials.1>
BELINKOV, Y.; GLASS, J. Analysis Methods in Neural
Language Processing: A Survey. Transactions of the
Association for Computational Linguistics, v. 7, p. 49–72,
2019.
BELTAGY, I.; PETERS, M. E.; COHAN, A. Longformer: The Long-Document
Transformer. CoRR, v. abs/2004.05150, 2020.
BELZ, A. et al. A Systematic Review of Reproducibility Research
in Natural Language Processing. Proceedings of the 16th
Conference of the European Chapter of the Association for Computational
Linguistics: Main Volume. Anais...Online: Association
for Computational Linguistics, abr. 2021. Disponível em: <https://aclanthology.org/2021.eacl-main.29>
BENCKE, L. et al. Can we trust LLMs as relevance
judges? Anais do XXXIX Simpósio Brasileiro de Bancos de Dados.
Anais...Porto Alegre, RS, Brasil: SBC, 2024. Disponível
em: <https://sol.sbc.org.br/index.php/sbbd/article/view/30724>
BENDER, E. M. On
achieving and evaluating language-independence in NLP.
Linguistic Issues in Language Technology, v. 6, 2011.
BENDER, E. M. et al. On the Dangers of Stochastic Parrots: Can
Language Models Be Too Big? 🦜. Proceedings of the 2021 ACM
Conference on Fairness, Accountability, and Transparency.
Anais...: FAccT ’21.New York, NY, USA: Association for
Computing Machinery, 2021. Disponível em: <https://doi.org/10.1145/3442188.3445922>
BENDER, E. M. Resisting
Dehumanization in the Age of
“AI”. Current Directions in
Psychological Science, v. 33, n. 2, p. 114–120, abr. 2024.
BENDER, E. M.; FRIEDMAN, B. Data Statements for Natural
Language Processing: Toward Mitigating System Bias and Enabling Better
Science. Transactions of the Association for Computational
Linguistics, v. 6, p. 587–604, 2018.
BENDER, E. M.; HANNA, A. The AI Con:
How to fight big tech’s hype and create the future we
want. [s.l.] Random House, 2025.
BENDER, E. M.; KOLLER, A. Climbing towards NLU:
On Meaning, Form, and Understanding in the Age of
Data. Proceedings of the 58th Annual Meeting of the Association
for Computational Linguistics. Anais...Online:
Association for Computational Linguistics, jul. 2020. Disponível em:
<https://aclanthology.org/2020.acl-main.463>
BENGIO, Y. et al. A Neural Probabilistic Language Model. J.
Mach. Learn. Res., v. 3, n. null, p. 1137–1155, mar. 2003.
BENOTTI, L.; BLACKBURN, P. Ethics consideration sections in
natural language processing papers. Proceedings of the 2022
Conference on Empirical Methods
in Natural Language Processing.
Anais...Abu Dhabi, United Arab Emirates: Association
for Computational Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.emnlp-main.299>
BERTSCH, A. et al. Unlimiformer:
Long-Range Transformers with Unlimited Length Input.
CoRR, v. abs/2305.01625, 2023.
BERWICK, R. C.; CHOMSKY, N. Por que apenas nós?
Linguagem e evolução. [s.l.]
SciELO-Editora UNESP, 2017.
BIANCHI, F.; HOVY, D. On the Gap between Adoption and
Understanding in NLP. Findings of the Association
for Computational Linguistics: ACL-IJCNLP 2021.
Anais...Online: Association for Computational
Linguistics, ago. 2021. Disponível em: <https://aclanthology.org/2021.findings-acl.340>
BIBAL, A. et al. Is Attention Explanation? An Introduction to
the Debate. Proceedings of the 60th Annual Meeting of the
Association for Computational Linguistics (Volume 1: Long Papers).
Anais...Dublin, Ireland: Association for Computational
Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.acl-long.269>
BIBER, D. Variation across Speech and Writing.
Cambridge: Cambridge University Press, 1988.
BIBER, D.; CONRAD, S. Register, Genre, and Style.
Cambridge: Cambridge University Press, 2009.
BIBER, D.; FINEGAN, E. Sociolinguistic
Perspectives On Register. [s.l.] Oxford University Press,
1994.
BINZ, M. et al. How
should the advancement of large language models affect the practice of
science? Proceedings of the National Academy of
Sciences, v. 122, n. 5, p. e2401227121, fev. 2025.
BIRD, S. Decolonising Speech and Language Technology.
Proceedings of the 28th International Conference on Computational
Linguistics. Anais...Barcelona, Spain (Online):
International Committee on Computational Linguistics, dez. 2020.
Disponível em: <https://aclanthology.org/2020.coling-main.313>
BIRD, S. Local Languages, Third
Spaces, and other High-Resource
Scenarios. Proceedings of the 60th
Annual Meeting of the Association
for Computational Linguistics
(Volume 1: Long Papers).
Anais...Dublin, Ireland: Association for Computational
Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.acl-long.539>
BLODGETT, S. L. et al. Language (Technology) is
Power: A Critical
Survey of “Bias” in
NLP. Proceedings of the 58th Annual
Meeting of the Association for
Computational Linguistics.
Anais...Online: Association for Computational
Linguistics, a2020. Disponível em: <https://www.aclweb.org/anthology/2020.acl-main.485>
BLODGETT, S. L. et al. Language (Technology) is Power: A
Critical Survey of “Bias” in NLP. (D.
Jurafsky et al., Eds.)Proceedings of the 58th Annual Meeting of the
Association for Computational Linguistics.
Anais...Online: Association for Computational
Linguistics, jul. b2020. Disponível em: <https://aclanthology.org/2020.acl-main.485/>
BLOM, J. D. A dictionary of hallucinations. [s.l.]
Springer, 2010.
BOSTROM, N. Superintelligence: Paths, Dangers,
Strategies. Oxford, UK: Oxford University Press, 2014.
BOWMAN, S. R.; DAHL, G. What Will it Take to Fix Benchmarking in
Natural Language Understanding? (K. Toutanova et al.,
Eds.)Proceedings of the 2021 Conference of the North American Chapter of
the Association for Computational Linguistics: Human Language
Technologies. Anais...Online: Association for
Computational Linguistics, jun. 2021. Disponível em: <https://aclanthology.org/2021.naacl-main.385/>
BRANDES, N. et al. ProteinBERT: a
universal deep-learning model of protein sequence and function.
Bioinform., v. 38, n. 8, p. 2102–2110, 2022.
BRENNAN, K.; KAK, A.; WEST, S. M. The
AGI Mythology: The Argument to End All Arguments. Em:
Artificial Power: 2025 Landscape Report. [s.l.] AI Now
Institute, 2025.
BRENNEIS, A. Assessing dual use
risks in AI research: necessity, challenges and mitigation
strategies. Research Ethics, v. 21, n. 2, p.
302–330, 2025.
BROUSSARD, M. More Than a Glitch: Confronting Race, Gender, and
Ability Bias in Tech. 1. ed. Cambridge, Massachusetts: MIT
Press, 2023.
BROWN, T. B. et al. Language Models are Few-Shot
Learners. (H. Larochelle et al., Eds.)Advances in Neural
Information Processing Systems. Anais...Curran
Associates, Inc., 2020. Disponível em: <https://proceedings.neurips.cc/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html>
BRUM, H.; NUNES, M. DAS G. V. Building a Sentiment Corpus
of Tweets in Brazilian Portuguese. (N. C. (Conference
chair) et al., Eds.)Proceedings of the Eleventh International Conference
on Language Resources and Evaluation (LREC 2018).
Anais...Miyazaki, Japan: European Language Resources
Association (ELRA), mar. 2018.
BURCHELL, L. et al. An Expanded Massive Multilingual Dataset for
High-Performance Language Technologies (HPLT). (W.
Che et al., Eds.)Proceedings of the 63rd Annual Meeting of the
Association for Computational Linguistics (Volume 1: Long Papers).
Anais...Vienna, Austria: Association for Computational
Linguistics, jul. 2025. Disponível em: <https://aclanthology.org/2025.acl-long.854/>
CARLSMITH, J. Is Power-Seeking AI an Existential Risk?
[s.l.] Open Philanthropy, 2022.
CARMO, D. et al. PTT5: Pretraining
and validating the T5 model on Brazilian Portuguese
data. CoRR, v. abs/2008.09144, 2020.
CASTILHO, S. et al. DELA Corpus - A Document-Level
Corpus Annotated with Context-Related Issues. Proceedings of
the Sixth Conference on Machine Translation.
Anais...Online: Association for Computational
Linguistics, nov. 2021. Disponível em: <https://aclanthology.org/2021.wmt-1.63>
CENTRE FOR INFORMATION POLICY LEADERSHIP, C. The Limitations of
Consent as a Legal Basis for Data Processing in the Digital
Society. Washington DC, London, BrusselsCentre for
Information Policy Leadership & Hunton Andrews Kurth
LLP; Bae Kim & Lee, 2024. Disponível em: <https://www.informationpolicycentre.com/uploads/5/7/1/0/57104281/cipl_bkl_limitations_of_consent_legal_basis_data_processing_dec24.pdf>
CERVANTES, M. DE. D.
Quixote de La Mancha — Primeira Parte. [s.l.] eBooksBrasil,
1605.
CHALAMALASETTI, K. et al. clembench: Using Game Play to Evaluate
Chat-Optimized Language Models as Conversational Agents. (H.
Bouamor, J. Pino, K. Bali, Eds.)Proceedings of the 2023 Conference on
Empirical Methods in Natural Language Processing.
Anais...Singapore: Association for Computational
Linguistics, dez. 2023. Disponível em: <https://aclanthology.org/2023.emnlp-main.689>
CHANDRA, M. et al. From Lived
Experience to Insight: Unpacking
the Psychological Risks of Using
AI Conversational
Agents. Proceedings of the 2025 ACM
Conference on Fairness,
Accountability, and Transparency.
Anais...Athens Greece: ACM, jun. 2025. Disponível em:
<https://dl.acm.org/doi/10.1145/3715275.3732063>
CHEN, G. H. et al. Humans or
LLMs as the Judge? A Study on Judgement
Bias. Proceedings of the 2024 Conference on Empirical
Methods in Natural Language Processing, a2024.
CHEN, K. et al. How Susceptible are
Large Language Models to
Ideological Manipulation? Proceedings
of the 2024 Conference on Empirical
Methods in Natural Language
Processing. Anais...Miami, Florida, USA:
Association for Computational Linguistics, b2024. Disponível em: <https://aclanthology.org/2024.emnlp-main.952>
CHEN, T. et al. CopyBench: Measuring
Literal and Non-Literal Reproduction of Copyright-Protected Text in
Language Model Generation. (Y. Al-Onaizan, M. Bansal, Y.-N.
Chen, Eds.)Proceedings of the 2024 Conference on Empirical Methods in
Natural Language Processing. Anais...Miami, Florida,
USA: Association for Computational Linguistics, nov. c2024. Disponível
em: <https://aclanthology.org/2024.emnlp-main.844/>
CHENG, J. et al. Dated Data: Tracing Knowledge Cutoffs in Large
Language Models., 2024. Disponível em: <https://arxiv.org/abs/2403.12958>
CHILD, R. et al. Generating
Long Sequences with Sparse Transformers. CoRR, v.
abs/1904.10509, 2019.
CHO, K. et al. Learning Phrase Representations using
RNN Encoder-Decoder for Statistical Machine
Translation. (A. Moschitti, B. Pang, W. Daelemans,
Eds.)Proceedings of the 2014 Conference on Empirical Methods in Natural
Language Processing, EMNLP 2014, October 25-29, 2014, Doha,
Qatar, A meeting of SIGDAT, a Special Interest Group of the
ACL. Anais...ACL, 2014.
Disponível em: <https://doi.org/10.3115/v1/d14-1179>
CHOMSKY, N. Aspects of the Theory of Syntax. Cambridge,
MA: MIT Press, 1965.
CHOWDHERY, A. et al. PaLM: Scaling Language
Modeling with Pathways. CoRR, v. abs/2204.02311,
2022.
CHRISTIANO, P. F. et al. Deep Reinforcement Learning from Human
Preferences. (I. Guyon et al., Eds.)Advances in Neural
Information Processing Systems 30: Annual Conference on Neural
Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA,
USA. Anais...2017. Disponível em: <https://proceedings.neurips.cc/paper/2017/hash/d5e2c0adad503c91f91df240d0cd4e49-Abstract.html>
CLARK, K. et al. ELECTRA: Pre-training Text
Encoders as Discriminators Rather Than Generators. 8th
International Conference on Learning Representations, ICLR
2020. Anais...Addis Ababa, Ethiopia: OpenReview.net,
abr. 2020. Disponível em: <https://openreview.net/forum?id=r1xMH1BtvB>
COHEN, K. B. et al. Three Dimensions of Reproducibility in
Natural Language Processing. Proceedings of the Eleventh
International Conference on Language Resources and Evaluation
(LREC 2018). Anais...Miyazaki, Japan:
European Language Resources Association (ELRA), 2018. Disponível em:
<https://aclanthology.org/L18-1025>
COLLOBERT, R.; WESTON, J. A unified architecture for natural
language processing: deep neural networks with multitask
learning. (W. W. Cohen, A. McCallum, S. T. Roweis, Eds.)Machine
Learning, Proceedings of the Twenty-Fifth International Conference
(ICML 2008), Helsinki, Finland, June 5-9, 2008.
Anais...: ACM International Conference
Proceeding Series.ACM, 2008. Disponível em: <https://doi.org/10.1145/1390156.1390177>
COMMON CRAWL. Common Crawl – Open Repository of Web Crawl
Data. https://commoncrawl.org, [s.d.].
CONNEAU, A.; LAMPLE, G. Cross-Lingual Language Model Pretraining. Em:
Proceedings of the 33rd International Conference on Neural
Information Processing Systems. Red Hook, NY, USA: Curran
Associates Inc., 2019.
CORMACK, G. V.; CLARKE, C. L.; BUETTCHER, S. Reciprocal rank
fusion outperforms condorcet and individual rank learning
methods. Proceedings of the 32nd international ACM SIGIR
conference on Research and development in information retrieval.
Anais...2009.
CORMEN, T. et al. Introduction to Algorithms. Em: 2. ed. [s.l.] MIT
Press; McGraw-Hill, 2001.
CORRÊA, N. K. et al. Tucano:
Advancing Neural Text Generation for Portuguese.
Patterns, 2025.
CORTES, C.; VAPNIK, V. Support-Vector
Networks. Machine Learning, v. 20, n. 3, p.
273–297, set. 1995.
CORTIZ, D. et al. A Weakly Supervised Dataset of Fine-Grained
Emotions in Portuguese. Anais do XIII Simpósio Brasileiro de
Tecnologia da Informação e da Linguagem Humana.
Anais...Porto Alegre, RS, Brasil: SBC, 2021. Disponível
em: <https://sol.sbc.org.br/index.php/stil/article/view/17786>
COULDRY, N.; MEJIAS, U. A. The Costs of
Connection: How Data Is Colonizing Human Life and Appropriating It for
Capitalism. [s.l.] Stanford University Press, 2019.
CRESPO, M. C. R. M. et al. Carolina: a General Corpus of
Contemporary Brazilian Portuguese with Provenance, Typology and
Versioning Information., 2023. Disponível em: <https://arxiv.org/abs/2303.16098>
CRONBACH, L. J. Studies of
acquiescence as a factor in the true–false test. Journal of
Educational Psychology, v. 33, p. 401–415, 1942.
CUEVA, R. et al. Inteligência Artificial e Desafios
Regulatórios. [s.l.] Forense, 2026.
CURRY, A.; CERCAS CURRY, A. Computer says “No”: The
Case Against Empathetic Conversational AI. (A.
Rogers, J. Boyd-Graber, N. Okazaki, Eds.)Findings of the Association for
Computational Linguistics: ACL 2023. Anais...Toronto,
Canada: Association for Computational Linguistics, jul. 2023. Disponível
em: <https://aclanthology.org/2023.findings-acl.515/>
D’ARCY, A.; BENDER, E. M. Ethics
in Linguistics. Annual Review of
Linguistics, v. 9, n. 1, p. 49–69, jan. 2023.
DAI, Z. et al. Transformer-XL: Attentive Language Models beyond
a Fixed-Length Context. (A. Korhonen, D. R. Traum, L. Màrquez,
Eds.)Proceedings of the 57th Conference of the Association for
Computational Linguistics, ACL 2019, Florence, Italy, July
28- August 2, 2019, Volume 1: Long Papers.
Anais...Association for Computational Linguistics,
2019. Disponível em: <https://doi.org/10.18653/v1/p19-1285>
DARCIO, L. et al. LattesRex: Building ChatBots for
Semi-Structured Documents. Proceedings of the 14th Brazilian
Symposium in Information and Human Language Technology (STIL 2025).
Anais...Brazil: 2025. Disponível em: <https://github.com/Lucas-Darcio/LattesRex>
DAVIES, M.; FERREIRA, M. J. Corpus do Português:
Web/Dialects. https://www.corpusdoportugues.org/web-dial/, 2016.
DAVIES, M.; FERREIRA, M. J. Corpus do Português:
NOW. https://www.corpusdoportugues.org/now/, 2018.
DETTMERS, T. et al. QLoRA: Efficient Finetuning of Quantized LLMs.
arXiv preprint arXiv:2305.14314, 2023.
DEVLIN, J. et al. BERT: Pre-training of Deep
Bidirectional Transformers for Language Understanding. (J.
Burstein, C. Doran, T. Solorio, Eds.)Proceedings of the 2019 Conference
of the North American Chapter of the Association for Computational
Linguistics: Human Language Technologies, NAACL-HLT 2019.
Anais...Minneapolis, MN, USA: Association for
Computational Linguistics, 2019. Disponível em: <https://doi.org/10.18653/v1/n19-1423>
DIERK, C.; HEALEY, J.; DOGAN, M. D. Evaluating LLMs in
Experiential Context: Insights from a Survey of Recent CHI
Publications. Human-centered Evaluation and Auditing of
Language Models Workshop (HEAL), CHI ’25.
Anais...Yokohama, Japan: ACM, 2025. Disponível em:
<https://heal-workshop.github.io/chi2025_papers/43_Evaluating_LLMs_in_Experien.pdf>
DING, N. et al. Parameter-efficient
fine-tuning of large-scale pre-trained language models.
Nature machine intelligence, v. 5, n. 3, p. 220–235,
2023.
DONG, Q. et al. A
Survey for In-context Learning. CoRR, v.
abs/2301.00234, 2023.
DOU, G. et al. Avoiding Copyright Infringement via Large
Language Model Unlearning. (L. Chiruzzo, A. Ritter, L. Wang,
Eds.)Findings of the Association for Computational Linguistics: NAACL
2025. Anais...Albuquerque, New Mexico: Association for
Computational Linguistics, abr. 2025. Disponível em: <https://aclanthology.org/2025.findings-naacl.288/>
EDITORIAL. Emotional risks of
AI companions demand attention. Nature Machine
Intelligence, v. 7, n. 7, p. 981–982, jul. 2025.
ES, S. et al. Ragas: Automated evaluation of retrieval augmented
generation. Proceedings of the 18th Conference of the European
Chapter of the Association for Computational Linguistics: System
Demonstrations. Anais...2024.
ETHAYARAJH, K.; JURAFSKY, D. Utility is in the Eye
of the User: A Critique of
NLP Leaderboards. Proceedings of the
2020 Conference on Empirical
Methods in Natural Language
Processing (EMNLP).
Anais...Online: Association for Computational
Linguistics, 2020. Disponível em: <https://www.aclweb.org/anthology/2020.emnlp-main.393>
FAN, A.; LEWIS, M.; DAUPHIN, Y. Hierarchical Neural Story
Generation. Proceedings of the 56th Annual Meeting of the
Association for Computational Linguistics (Volume 1: Long Papers).
Anais...Melbourne, Australia: Association for
Computational Linguistics, jul. 2018. Disponível em: <https://aclanthology.org/P18-1082>
FÄRBER, F. et al. IaraMed: A Women’s Healthcare Chatbot for
Portuguese Speakers. Anais do XXV Simpósio Brasileiro de
Computação Aplicada à Saúde. Anais...Porto Alegre, RS,
Brasil: SBC, 2025. Disponível em: <https://sol.sbc.org.br/index.php/sbcas/article/view/35558>
FEIJÓ, D. DE V.; MOREIRA, V. P. Mono vs Multilingual
Transformer-based Models: a Comparison across Several Language
Tasks. CoRR, v. abs/2007.09757, 2020.
FINGER, M. et al. Carolina: General Corpus of Contemporary
Brazilian Portuguese with Provenance and Typology Information.,
2020. Disponível em: <https://c4ai.inova.usp.br/portal/carolina>
FIRTH, J. R. A synopsis of linguistic theory 1930–1955.
[s.l.] Blackwell, 1957. p. 1–32
FISHER, J. et al. Biased LLMs can Influence
Political Decision-Making. (W. Che et al., Eds.)Proceedings of
the 63rd Annual Meeting of the Association for Computational Linguistics
(Volume 1: Long Papers). Anais...Vienna, Austria:
Association for Computational Linguistics, jul. 2025. Disponível em:
<https://aclanthology.org/2025.acl-long.328/>
FLORIDI, L. Translating
Principles into Practices of
Digital Ethics: Five
Risks of Being Unethical.
Philosophy & Technology, v. 32, n. 2, p. 185–193,
jun. 2019.
FOKKENS, A. et al. Offspring from Reproduction Problems: What
Replication Failure Teaches Us. Proceedings of the 51st Annual
Meeting of the Association for Computational Linguistics (Volume 1: Long
Papers). Anais...Sofia, Bulgaria: Association for
Computational Linguistics, ago. 2013. Disponível em: <https://aclanthology.org/P13-1166>
FORT, K.; ADDA, G.; COHEN, K. B. Amazon
Mechanical Turk: Gold
Mine or Coal Mine?
Computational Linguistics, v. 37, n. 2, p. 413–420,
jun. 2011.
FRANCISCO, P. A. P.; HUREL, L. M.; RIELLI, M. M. Regulação do
Reconhecimento Facial no Setor Público: Avaliação de Experiências
Internacionais. Instituto Igarapé - DataPrivacyBR,
2020. Disponível em: <https://www.dataprivacybr.org/wp-content/uploads/2021/11/regulacao-do-reconhecimento-facial-no-setor-publico.pdf>
FYFE, S. et al. Apophenia, theory of mind and schizotypy: perceiving
meaning and intentionality in randomness. Cortex, v.
44, n. 10, p. 1316–1325, 2008.
GABRIEL, I. Artificial
Intelligence, Values and Alignment. Minds and
Machines, v. 30, n. 3, p. 411–437, 2020.
GAJULAMANDYAM, D. K. et al. Domain Specific
Finetuning of LLMs Using PEFT Techniques. 2025 IEEE 15th
Annual Computing and Communication Workshop and Conference (CCWC).
Anais...2025.
GAO, M. et al. LLM-based NLG
Evaluation: Current Status and Challenges. ArXiv, 2024.
Disponível em: <https://arxiv.org/abs/2402.01383>
GAO, T.; YAO, X.; CHEN, D. SimCSE: Simple Contrastive Learning
of Sentence Embeddings. (M.-F. Moens et al., Eds.)Proceedings
of the 2021 Conference on Empirical Methods in Natural Language
Processing, EMNLP 2021, Virtual Event / Punta Cana,
Dominican Republic, 7-11 November, 2021.
Anais...Association for Computational Linguistics,
2021. Disponível em: <https://doi.org/10.18653/v1/2021.emnlp-main.552>
GARCIA, E. A. S. et al.
RoBERTaLexPT:
A Legal RoBERTa Model pretrained with
deduplication for Portuguese. (P. Gamallo et al.,
Eds.)Proceedings of the 16th International Conference on Computational
Processing of Portuguese - Vol. 1. Anais...Santiago de
Compostela, Galicia/Spain: Association for Computational Lingustics,
mar. 2024. Disponível em: <https://aclanthology.org/2024.propor-1.38>
GATT, A.; KRAHMER, E. Survey
of the State of the Art in Natural Language Generation: Core tasks,
applications and evaluation. Journal of Artificial
Intelligence Research, v. 61, n. 1, p. 65–170, 2018.
GEBRU, T. et al. Datasheets
for datasets. Communications of the ACM, v. 64, n.
12, p. 86–92, dez. 2021.
GEBRU, T.; BENDER, E. M.; MCMILLAN-MAJOR, A. Statement from the
listed authors of Stochastic Parrots on the
“AI pause” letter., 2023. Disponível
em: <https://www.dair-institute.org/blog/letter-statement-March2023/>
GEBRU, T.; TORRES, E. P. The TESCREAL bundle:
Eugenics and the promise of utopia through artificial general
intelligence. First Monday, abr. 2024.
GEHMAN, S. et al.
RealToxicityPrompts:
Evaluating Neural Toxic Degeneration in Language Models. (T.
Cohn, Y. He, Y. Liu, Eds.)Findings of the Association for Computational
Linguistics: EMNLP 2020. Anais...Online: Association
for Computational Linguistics, nov. 2020. Disponível em: <https://aclanthology.org/2020.findings-emnlp.301/>
GEVA, M.; GUPTA, A.; BERANT, J. Injecting Numerical Reasoning
Skills into Language Models. (D. Jurafsky et al.,
Eds.)Proceedings of the 58th Annual Meeting of the Association for
Computational Linguistics, ACL 2020, Online, July 5-10,
2020. Anais...Association for Computational
Linguistics, 2020. Disponível em: <https://doi.org/10.18653/v1/2020.acl-main.89>
GOLUMBIA, D. Cyberlibertarianism:
The Right-Wing Politics of Digital Technology. [s.l.]
University of Minnesota Press, 2024.
GONG, Z. et al. Continual Pre-training of Language Models for
Math Problem Understanding with Syntax-Aware Memory Network.
(S. Muresan, P. Nakov, A. Villavicencio, Eds.)Proceedings of the 60th
Annual Meeting of the Association for Computational Linguistics (Volume
1: Long Papers), ACL 2022, Dublin, Ireland, May 22-27,
2022. Anais...Association for Computational
Linguistics, 2022. Disponível em: <https://doi.org/10.18653/v1/2022.acl-long.408>
GONZALEZ-CABELLO, M. et al. Fairness in
crowdwork: Making the human AI supply chain more humane.
Business Horizons, v. 68, n. 5, p. 645–657, 2025.
GOODFELLOW, I.; BENGIO, Y.; COURVILLE, A. Deep
Learning. [s.l.] MIT Press, 2016. v. 1
GOOGLE CLOUD. Best Practices with Large Language
Models. https://cloud.google.com/vertex-ai/generative-ai/docs/learn/prompt-best-practices?hl=en,
2025.
GRAY, M. L.; SURI, S. Ghost work: How to stop
Silicon Valley from building a new global
underclass. [s.l.] Harper Business, 2019.
GRAY WIDDER, D.; WEST, S.; WHITTAKER, M. Open (For Business): Big
Tech, Concentrated Power, and the Political Economy of Open AI.
SSRN Electronic Journal, 2023.
GU, Y. et al. Domain-Specific
Language Model Pretraining for Biomedical Natural Language
Processing. ACM Trans. Comput. Healthcare, v. 3, n.
1, out. 2021.
GURURANGAN, S. et al. Don’t Stop Pretraining: Adapt Language
Models to Domains and Tasks. Proceedings of the 58th Annual
Meeting of the Association for Computational Linguistics.
Anais...Online: Association for Computational
Linguistics, jul. 2020. Disponível em: <https://aclanthology.org/2020.acl-main.740>
HADDAD, I. Artificial Intelligence and Data in Open
Source. Linux Foundation, 2022. Disponível em: <https://www.linuxfoundation.org/hubfs/LF%20Research/Artificial%20Intelligence%20and%20Data%20in%20Open%20Source%20-%20Report.pdf?hsLang=en>
HANNIGAN, T. R.; MCCARTHY, I. P.; SPICER, A. Beware of botshit:
How to manage the epistemic risks of generative chatbots.
Business Horizons, v. 67, n. 5, p. 471–486, 2024.
HAVENS, L. et al. Situated Data,
Situated Systems: A
Methodology to Engage with Power
Relations in Natural Language
Processing Research. (M. R.
Costa-jussà et al., Eds.)Proceedings of the Second
Workshop on Gender Bias in
Natural Language Processing.
Anais...Barcelona, Spain (Online): Association for
Computational Linguistics, dez. 2020. Disponível em: <https://aclanthology.org/2020.gebnlp-1.10/>
HE, K. et al. Deep Residual Learning for Image
Recognition. 2016 IEEE Conference on Computer
Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV,
USA, June 27-30, 2016. Anais...IEEE
Computer Society, 2016. Disponível em: <https://doi.org/10.1109/CVPR.2016.90>
HE, P. et al. Deberta: decoding-Enhanced Bert with Disentangled
Attention. 9th International Conference on Learning
Representations, ICLR 2021, Virtual Event, Austria, May
3-7, 2021. Anais...OpenReview.net, 2021. Disponível em:
<https://openreview.net/forum?id=XPZIaotutsD>
HEDDERICH, M. A.; OULASVIRTA, A. Explaining
crowdworker behaviour through computational rationality.
Behaviour & Information Technology, v. 44, n. 3, p.
552–573, 2025.
HELLER, J. Legal AI benchmarking:
CoCounsel – from code to courtroom: The meticulous testing
of CoCounsel’s professional-grade
AI., 23 out. 2024. Disponível em: <https://www.thomsonreuters.com/en-us/posts/innovation/legal-ai-benchmarking-cocounsel/>.
Acesso em: 20 ago. 2025
HENDREN, S. All
technology is assistive: Six design rules on
disability. Em: SAYERS, J. (Ed.). Making Things and Drawing
Boundaries: Experiments in the Digital Humanities. [s.l.]
University of Minnesota Press, Minneapolis, MN, 2014.
HENDRYCKS, D. et al. Unsolved Problems in ML Safety. arXiv
preprint arXiv:2109.13916, 2021.
HERSHCOVICH, D. et al. Towards Climate
Awareness in NLP
Research. Proceedings of the 2022
Conference on Empirical Methods
in Natural Language Processing.
Anais...Abu Dhabi, United Arab Emirates: Association
for Computational Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.emnlp-main.159>
HICKS, M. T.; HUMPHRIES, J.; SLATER, J. ChatGPT
is bullshit. Ethics and Information Technology, v.
26, n. 2, p. 38, jun. 2024.
HOCHREITER, S. Untersuchungen zu dynamischen neuronalen Netzen.
Diploma, Technische Universität
München, v. 91, n. 1, p. 31, 1991.
HOCHREITER, S.; SCHMIDHUBER, J. Long Short-Term
Memory. Neural Computation, v. 9, n. 8, p.
1735–1780, nov. 1997.
HOFFMANN, J. et al. Training
Compute-Optimal Large Language Models. CoRR, v.
abs/2203.15556, 2022.
HOFMANN, V. et al. AI
generates covertly racist decisions about people based on their
dialect. Nature, v. 633, n. 8028, p. 147–154, set.
2024.
HOLTZMAN, A. et al. The Curious Case of Neural Text
Degeneration. ICLR. Anais...OpenReview.net,
2020. Disponível em: <http://dblp.uni-trier.de/db/conf/iclr/iclr2020.html#HoltzmanBDFC20>
HOOKER, S. Moving
beyond “algorithmic bias is a data problem”.
Patterns, v. 2, n. 4, p. 100241, abr. 2021.
HORNIK, K.; STINCHCOMBE, M. B.; WHITE, H. Multilayer
feedforward networks are universal approximators. Neural
Networks, v. 2, n. 5, p. 359–366, 1989.
HOUAMEGNI, L. R. P.; GEDIKLI, F. Evaluating the Effectiveness of
Large Language Models in Automated News Article Summarization.,
2025. Disponível em: <https://arxiv.org/abs/2502.17136>
HOULSBY, N. et al. Parameter-Efficient Transfer Learning for
NLP. (K. Chaudhuri, R. Salakhutdinov,
Eds.)Proceedings of the 36th International Conference on Machine
Learning, ICML 2019, 9-15 June 2019, Long Beach,
California, USA. Anais...: Proceedings of
Machine Learning Research.PMLR, 2019. Disponível em: <http://proceedings.mlr.press/v97/houlsby19a.html>
HOWARD, J.; RUDER, S. Universal Language Model Fine-tuning for
Text Classification. Proceedings of the 56th Annual Meeting of
the Association for Computational Linguistics (Volume 1: Long Papers).
Anais...Melbourne, Australia: Association for
Computational Linguistics, jul. a2018. Disponível em: <^5^>
HOWARD, J.; RUDER, S. Universal Language
Model Fine-tuning for Text
Classification. (I. Gurevych, Y. Miyao,
Eds.)Proceedings of the 56th Annual Meeting of
the Association for Computational
Linguistics (Volume 1: Long
Papers). Anais...Melbourne, Australia:
Association for Computational Linguistics, jul. b2018. Disponível em:
<https://aclanthology.org/P18-1031/>.
Acesso em: 25 jul. 2025
HU, E. J. et al. LoRA: Low-Rank Adaptation of Large
Language Models., 2021. Disponível em: <https://arxiv.org/abs/2106.09685>
HU, E. J. et al. LoRA: Low-Rank Adaptation of Large Language
Models. The Tenth International Conference on Learning
Representations, ICLR 2022, Virtual Event, April 25-29,
2022. Anais...OpenReview.net, 2022. Disponível em:
<https://openreview.net/forum?id=nZeVKeeFYf9>
HUANG, J.; YANG, D.; POTTS, C. Demystifying Verbatim
Memorization in Large Language Models. (Y. Al-Onaizan, M.
Bansal, Y.-N. Chen, Eds.)Proceedings of the 2024 Conference on Empirical
Methods in Natural Language Processing. Anais...Miami,
Florida, USA: Association for Computational Linguistics, nov. a2024.
Disponível em: <https://aclanthology.org/2024.emnlp-main.598/>
HUANG, L. et al. A Survey on
Hallucination in Large Language Models: Principles, Taxonomy,
Challenges, and Open Questions. ACM Transactions on
Information Systems, v. 43, n. 2, p. 1–58, b2024.
II, S. M. W. GSM8K Benchmark. Klu; https://klu.ai/glossary/GSM8K-eval, 2025.
IP, J.; VONGTHONGSRI, K. deepeval., ago.
2025. Disponível em: <https://github.com/confident-ai/deepeval>
IVGI, M.; SHAHAM, U.; BERANT, J. Efficient Long-Text
Understanding with Short-Text Models. Transactions of the
Association for Computational Linguistics, v. 11, p. 284–299,
2023.
JACKSON, P.; MOULINIER, I. Natural Language Processing for
Online Applications – Text retrieval, extraction and
categorization. [s.l.] John Benjamins, 2002.
JAIN, S.; WALLACE, B. C. Attention is not Explanation.
Proceedings of the 2019 Conference of the North American
Chapter of the Association for Computational Linguistics: Human Language
Technologies, Volume 1 (Long and Short Papers).
Anais...Minneapolis, Minnesota: Association for
Computational Linguistics, 2019. Disponível em: <https://aclanthology.org/N19-1357>
JAKUBČEK, M. et al. The TenTen corpus family.
Proceedings of the 7th International Corpus Linguistics Conference (CL).
Anais...2013.
JAN, E. et al. Multitask-Bench: Unveiling and Mitigating Safety
Gaps in LLMs Fine-tuning. (O. Rambow et al.,
Eds.)Proceedings of the 31st International Conference on Computational
Linguistics. Anais...Abu Dhabi, UAE: Association for
Computational Linguistics, jan. 2025. Disponível em: <https://aclanthology.org/2025.coling-main.606/>
JAPKOWICZ, N.; BOUKOUVALAS, Z. Machine Learning Evaluation:
Towards Reliable and Responsible AI. [s.l.] Cambridge
University Press, 2024.
JENSEN, K. N.; PLANK, B. Fine-tuning vs From Scratch: Do Vision
& Language Models Have Similar Capabilities on
Out-of-Distribution Visual Question Answering? (N. Calzolari et
al., Eds.)Proceedings of the Thirteenth Language Resources and
Evaluation Conference. Anais...Marseille, France:
European Language Resources Association, jun. 2022. Disponível em:
<https://aclanthology.org/2022.lrec-1.161/>
JEONG, S. et al. Adaptive-rag: Learning to adapt retrieval-augmented
large language models through question complexity. arXiv
preprint arXiv:2403.14403, 2024.
JI, Z. et al. Survey of
Hallucination in Natural Language Generation. ACM Comput.
Surv., v. 55, n. 12, mar. 2023.
JIANG, L.; WAGNER, C. How Low is Low? Crowdworker Perceptions of
Microtask Payments in Work versus Leisure Situations.
Proceedings of the 2024 CHI Conference on Human Factors in Computing
Systems. Anais...: CHI ’24.New York, NY, USA:
Association for Computing Machinery, 2024. Disponível em: <https://doi.org/10.1145/3613904.3642601>
JIAO, J. et al. Navigating LLM
Ethics: Advancements, Challenges, and Future Directions.
ArXiv, v. abs/2406.18841, 2024.
JIN, X. et al. Lifelong Pretraining: Continually Adapting
Language Models to Emerging Corpora. Proceedings of the 2022
Conference of the North American Chapter of the Association for
Computational Linguistics: Human Language Technologies.
Anais...Seattle, United States: Association for
Computational Linguistics, jul. 2022. Disponível em: <https://aclanthology.org/2022.naacl-main.351>
JIN, Z. et al. How Good Is
NLP? A Sober Look at
NLP Tasks through the Lens of
Social Impact. Findings of the
Association for Computational
Linguistics: ACL-IJCNLP 2021.
Anais...Online: Association for Computational
Linguistics, 2021. Disponível em: <https://aclanthology.org/2021.findings-acl.273>
JOHNSTON, S. F. Alvin
Weinberg and the Promotion of the Technological Fix.
Technology and Culture, v. 59, n. 3, p. 620–651, 2018.
JOSHI, M. et al. TriviaQA: A Large Scale Distantly
Supervised Challenge Dataset for Reading Comprehension. (R.
Barzilay, M.-Y. Kan, Eds.)Proceedings of the 55th Annual Meeting of the
Association for Computational Linguistics, ACL 2017,
Vancouver, Canada, July 30 - August 4, Volume 1: Long Papers.
Anais...Association for Computational Linguistics,
2017. Disponível em: <https://doi.org/10.18653/v1/P17-1147>
JOSHI, M. et al. SpanBERT:
Improving Pre-training by Representing and Predicting Spans.
Transactions of the Association for Computational
Linguistics, v. 8, p. 64–77, 2020.
JOSHI, N.; TALY, A.; MUPPALLA, D. LLM-Cite: Cheap
Fact Verification with Attribution via URL
Generation., 2025. Disponível em: <https://openreview.net/forum?id=qb2QRoE4W3>
JOULIN, A. et al. Bag of Tricks for Efficient Text
Classification. (M. Lapata, P. Blunsom, A. Koller,
Eds.)Proceedings of the 15th Conference of the European
Chapter of the Association for Computational Linguistics: Volume 2,
Short Papers. Anais...Valencia, Spain: Association for
Computational Linguistics, abr. 2017. Disponível em: <https://aclanthology.org/E17-2068/>
JOYCE, J. M. Kullback-Leibler
Divergence. Em: LOVRIC, M. (Ed.). International Encyclopedia
of Statistical Science. Berlin, Heidelberg: Springer Berlin
Heidelberg, 2011. p. 720–722.
JURAFSKY, D.; MARTIN, J. H. Speech and Language Processing: An
Introduction to Natural Language Processing, Computational Linguistics,
and Speech Recognition. 3rd. ed. USA: Prentice Hall PTR, 2023.
KALAI, A. T. et al. Why language models hallucinate.
[s.l.] OpenAI, 2025.
KALOULI, A.-L. et al. Curing the SICK
and Other NLI Maladies. Computational
Linguistics, v. 49, n. 1, p. 199–243, mar. 2023.
KAMRUZZAMAN, M.; SHOVON, MD.; KIM, G. Investigating Subtler
Biases in LLMs: Ageism, Beauty, Institutional, and
Nationality Bias in Generative Models. (L.-W. Ku, A. Martins,
V. Srikumar, Eds.)Findings of the Association for Computational
Linguistics: ACL 2024. Anais...Bangkok, Thailand:
Association for Computational Linguistics, ago. 2024. Disponível em:
<https://aclanthology.org/2024.findings-acl.530/>
KARAMOLEGKOU, A. et al. Copyright Violations and
Large Language Models.
Proceedings of the 2023 Conference on
Empirical Methods in Natural
Language Processing.
Anais...Singapore: Association for Computational
Linguistics, 2023. Disponível em: <https://aclanthology.org/2023.emnlp-main.458>
KE, Z. et al. Continual Pre-training of Language
Models., 2023. Disponível em: <https://arxiv.org/abs/2302.03241>
KERNER, T. Domain-Specific Pretraining of Language Models: A
Comparative Study in the Medical Field., 2024. Disponível em:
<https://arxiv.org/abs/2407.14076>
KHATTAB, O.; ZAHARIA, M. ColBERT: Efficient and Effective
Passage Search via Contextualized Late Interaction over BERT.,
2020. Disponível em: <https://arxiv.org/abs/2004.12832>
KIANPOUR, M.; WEN, S.-F. Timing Attacks on Machine Learning:
State of the Art. Intelligent Systems Conference.
Anais...Springer, 2020.
KIM, D. K. et al. Analyzing Offensive Language Dataset Insights
from Training Dynamics and Human Agreement Level. (O. Rambow et
al., Eds.)Proceedings of the 31st International Conference on
Computational Linguistics, COLING 2025, Abu Dhabi, UAE,
January 19-24, 2025. Anais...Association for
Computational Linguistics, 2025. Disponível em: <https://aclanthology.org/2025.coling-main.653/>
KIM, S. et al. Prometheus: Inducing Fine-grained Evaluation Capability
in Language Models. arXiv preprint arXiv:2310.08491,
2023.
KLENK, M. How
Do Technological Artefacts
Embody Moral Values?
Philosophy & Technology, v. 34, n. 3, p. 525–544,
set. 2021.
KNUTH, D. E. Fundamental Algorithms. The Art of Computer
Programming. 3. ed. [s.l.] Addison-Wesley, 1997. v. 1
KOGKALIDIS, K.; CHATZIKYRIAKIDIS, S. On Tables with Numbers,
with Numbers. (S. Truong et al., Eds.)Proceedings of the 1st
Workshop on Language Models for Underserved Communities (LM4UC 2025).
Anais...Albuquerque, New Mexico: Association for
Computational Linguistics, 2025. Disponível em: <https://aclanthology.org/2025.lm4uc-1.12/>
KOJIMA, T. et al. Large Language Models are Zero-Shot
Reasoners. NeurIPS. Anais...2022. Disponível
em: <http://papers.nips.cc/paper\_files/paper/2022/hash/8bb0d291acd4acf06ef112099c16f326-Abstract-Conference.html>
KRAKOVNA, V. et al. Specification Gaming: The Flip Side of AI Ingenuity.
DeepMind Safety Research Blog, 2020.
KREUTZER, J. et al. Quality at a Glance: An
Audit of Web-Crawled Multilingual Datasets. Transactions of
the Association for Computational Linguistics, v. 10, p. 50–72,
jan. 2022.
KWIATKOWSKI, T. et al. Natural Questions: a Benchmark for Question
Answering Research. Transactions of the Association of
Computational Linguistics, 2019.
LAI, J. et al. Large language models in law: A survey. AI
Open, v. 5, p. 181–196, 2024.
LAN, Z. et al. ALBERT: A Lite
BERT for Self-supervised Learning of Language
Representations. 8th International Conference on Learning
Representations, ICLR 2020, Addis Ababa, Ethiopia, April
26-30, 2020. Anais...OpenReview.net, 2020. Disponível
em: <https://openreview.net/forum?id=H1eA7AEtvS>
LASOTA, L. Regulating Corporate
Behaviour in Digital Ecosystems: Increasing Fairness and Contestability
of Digital Markets with Free Software. MIC 2023: Toward
Green, Inclusive, and Digital Growth. Anais...a2023.
LASOTA, L. The
European Union’s AI act from the
perspective of Open Technologies. Em:
ALEGRE, M.; FÜRST, H. (Eds.). Advocacia &
Bioética: Novas
Tecnologias.
https://www.editoracasadodireito.com.br/produto/novas-tecnologias. São
Paulo: Casa do Direito, 2023b.
LASOTA, L. The
CRA and the Challenges of
Regulating Cybersecurity in Open
Environments: The Case of
Free and Open Source
Software. Em: Digital Decade:
How the EU shapes digitalisation.
Berlin: Weizenbaum Institute for the Networked Society - The German
Internet Institute, 2025.
LASOTA, L.; SINGHAL, N. Free Software and
AI openness: Overcoming challenges in the
licensing world. [s.l.] Zenodo, abr. 2024. Disponível em:
<https://zenodo.org/doi/10.5281/zenodo.10966136>.
LAZZARINI MORETTI, J.; MALTESE ZUFFO, M. LGPD e
inteligência artificial: Um estudo comparado. Revista de
Direito Internacional e Globalização Econômica, v. 13, n. 13,
p. 21–42, 2025.
LEE, J. et al. A survey of large language models in finance (finllms).
arXiv preprint arXiv:2402.02315, 2024.
LEE, K. et al. Deduplicating Training Data Makes Language Models
Better., 2022. Disponível em: <https://arxiv.org/abs/2107.06499>
LEIDNER, J. L.; PLACHOURAS, V. Ethical by Design: Ethics Best
Practices for Natural Language Processing. Proceedings of the
First ACL Workshop on Ethics in Natural Language
Processing. Anais...Valencia, Spain: Association for
Computational Linguistics, abr. 2017. Disponível em: <https://aclanthology.org/W17-1604>
LEIKE, J. et al. Scalable Agent Alignment via Reward Modeling: a
Research Direction. arXiv preprint arXiv:1811.07871.
Anais...2018.
LESTER, B.; AL-RFOU, R.; CONSTANT, N. The Power of Scale for
Parameter-Efficient Prompt Tuning. (M.-F. Moens et al.,
Eds.)Proceedings of the 2021 Conference on Empirical Methods in Natural
Language Processing, EMNLP 2021, Virtual Event / Punta
Cana, Dominican Republic, 7-11 November, 2021.
Anais...Association for Computational Linguistics,
2021. Disponível em: <https://doi.org/10.18653/v1/2021.emnlp-main.243>
LEWIS, M. et al. BART: Denoising
Sequence-to-Sequence Pre-training for Natural Language Generation,
Translation, and Comprehension. (D. Jurafsky et al.,
Eds.)Proceedings of the 58th Annual Meeting of the Association for
Computational Linguistics, ACL 2020, Online, July 5-10,
2020. Anais...Association for Computational
Linguistics, a2020. Disponível em: <https://doi.org/10.18653/v1/2020.acl-main.703>
LEWIS, P. S. H. et al. Retrieval-Augmented Generation for
Knowledge-Intensive NLP Tasks. (H. Larochelle et
al., Eds.)Advances in Neural Information Processing Systems 33: Annual
Conference on Neural Information Processing Systems 2020, NeurIPS 2020,
December 6-12, 2020, virtual. Anais...b2020. Disponível
em: <https://proceedings.neurips.cc/paper/2020/hash/6b493230205f780e1bc26945df7481e5-Abstract.html>
LI, D. et al. From
Generation to Judgment: Opportunities and Challenges of
LLM-as-a-judge. CoRR, v. abs/2411.16594, 2024.
LI, H. et al. CaseGen: A benchmark for multi-stage
legal case documents generation., 2025. Disponível em: <https://arxiv.org/abs/2502.17943>
LI, P. et al. Making AI Less "Thirsty": Uncovering and Addressing the
Secret Water Footprint of AI Models. arXiv preprint
arXiv:2304.03271, a2023.
LI, Q. Parameter
Efficient Fine-Tuning on Selective Parameters for Transformer-Based
Pre-Trained Models. 2024 IEEE International Conference on
Multimedia and Expo (ICME). Anais...2024.
LI, R. et al. StarCoder: may the
source be with you! CoRR, v. abs/2305.06161, b2023.
LI, W. W. et al. BERT Is Not The Count: Learning to
Match Mathematical Statements with Proofs. (A. Vlachos, I.
Augenstein, Eds.)Proceedings of the 17th Conference of the European
Chapter of the Association for Computational Linguistics,
EACL 2023, Dubrovnik, Croatia, May 2-6, 2023.
Anais...Association for Computational Linguistics,
c2023. Disponível em: <https://aclanthology.org/2023.eacl-main.260>
LI, X. L.; LIANG, P. Prefix-Tuning: Optimizing Continuous
Prompts for Generation. (C. Zong et al., Eds.)Proceedings of
the 59th Annual Meeting of the Association for Computational Linguistics
and the 11th International Joint Conference on Natural Language
Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers),
Virtual Event, August 1-6, 2021. Anais...Association
for Computational Linguistics, 2021. Disponível em: <https://doi.org/10.18653/v1/2021.acl-long.353>
LI, Y. et al. A
comparative study of pretrained language models for long clinical
text. Journal of the American Medical Informatics
Association, v. 30, n. 2, p. 340–347, nov. 2022.
LIANG, X. et al. Contrastive Demonstration Tuning for
Pre-trained Language Models. (Y. Goldberg, Z. Kozareva, Y.
Zhang, Eds.)Findings of the Association for Computational Linguistics:
EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7-11,
2022. Anais...Association for Computational
Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.findings-emnlp.56>
LIN, C.-Y. ROUGE: A Package for Automatic
Evaluation of Summaries. Text Summarization Branches Out.
Anais...Barcelona, Spain: Association for Computational
Linguistics, jul. 2004. Disponível em: <https://aclanthology.org/W04-1013>
LIN, S.; HILTON, J.; EVANS, O.
TruthfulQA: Measuring How Models Mimic
Human Falsehoods. (S. Muresan, P. Nakov, A. Villavicencio,
Eds.)Proceedings of the 60th Annual Meeting of the Association for
Computational Linguistics (Volume 1: Long Papers).
Anais...Dublin, Ireland: Association for Computational
Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.acl-long.229/>
LIRA, T. et al. Aroeira: A Curated Corpus for the Portuguese
Language with a Large Number of Tokens. Anais da XXXIV
Brazilian Conference on Intelligent Systems.
Anais...Porto Alegre, RS, Brasil: SBC, 2024. Disponível
em: <https://sol.sbc.org.br/index.php/bracis/article/view/33560>
LITA, A.-C.; COSTEA, C.-E.; IONESCU, B. Chunking for
RAG: A Comprehensive Evaluation of Chunking Strategies for
Retrieval-Augmented Generation. Proceedings of the 2024
Conference on Empirical Methods in Natural Language Processing.
Anais...2024.
LIU, Y. et al. RoBERTa: A Robustly Optimized BERT Pretraining
Approach., a2019. Disponível em: <https://arxiv.org/abs/1907.11692>
LIU, Y. et al. Roberta: A robustly optimized bert pretraining approach.
arXiv preprint arXiv:1907.11692, b2019.
LIU, Y. et al. Multilingual Denoising
Pre-training for Neural Machine Translation. Trans. Assoc.
Comput. Linguistics, v. 8, p. 726–742, 2020.
LIU, Y. et al. HD-Eval: Aligning Large Language Model Evaluators
Through Hierarchical Criteria Decomposition. (L.-W. Ku, A.
Martins, V. Srikumar, Eds.)Proceedings of the 62nd Annual Meeting of the
Association for Computational Linguistics (Volume 1: Long Papers),
ACL 2024, Bangkok, Thailand, August 11-16, 2024.
Anais...Association for Computational Linguistics,
2024. Disponível em: <https://doi.org/10.18653/v1/2024.acl-long.413>
LIU, Z. et al. A Robustly Optimized BERT Pre-Training Approach
with Post-Training. Chinese Computational Linguistics: 20th
China National Conference, CCL 2021, Hohhot, China, August 13–15, 2021,
Proceedings. Anais...Berlin, Heidelberg:
Springer-Verlag, 2021. Disponível em: <https://doi.org/10.1007/978-3-030-84186-7_31>
LOBO, P. Profiling
na Lei Geral de Proteção de Dados: O Livre Desenvolvimento da
Personalidade em Face da Governamentalidade Algorítmica. 1.
ed. [s.l.] Editora Foco, 2022.
LONGPRE, S. et al. A Pretrainer’s Guide to Training
Data: Measuring the Effects of Data Age, Domain Coverage, Quality,
& Toxicity. (K. Duh, H. Gomez, S. Bethard,
Eds.)Proceedings of the 2024 Conference of the North American Chapter of
the Association for Computational Linguistics: Human Language
Technologies (Volume 1: Long Papers). Anais...Mexico
City, Mexico: Association for Computational Linguistics, jun. 2024.
Disponível em: <https://aclanthology.org/2024.naacl-long.179/>
LONGPRE, S. et al. The Responsible Foundation Model Development
Cheatsheet: A Review of Tools & Resources., 2025.
Disponível em: <https://arxiv.org/abs/2406.16746>
LUONG, T.; PHAM, H.; MANNING, C. D. Effective Approaches to
Attention-based Neural Machine Translation. (L. Màrquez et al.,
Eds.)Proceedings of the 2015 Conference on Empirical Methods in Natural
Language Processing, EMNLP 2015, Lisbon, Portugal,
September 17-21, 2015. Anais...The Association for
Computational Linguistics, 2015. Disponível em: <https://doi.org/10.18653/v1/d15-1166>
LV, K. et al. Full Parameter Fine-tuning for Large Language
Models with Limited Resources., 2024. Disponível em: <https://arxiv.org/abs/2306.09782>
MALEKI, N.; PADMANABHAN, B.; DUTTA, K. AI
Hallucinations: A Misnomer
Worth Clarifying. 2024
IEEE Conference on Artificial
Intelligence (CAI).
Anais...Singapore, Singapore: IEEE, jun. 2024.
Disponível em: <https://ieeexplore.ieee.org/document/10605268/>
MARKOV, A. A. The theory of algorithms. Trudy Matematicheskogo
Instituta Imeni VA Steklova, v. 42, p. 3–375, 1954.
MATHUR, N.; BALDWIN, T.; COHN, T. Tangled up in
BLEU: Reevaluating the Evaluation of Automatic Machine
Translation Evaluation Metrics. Proceedings of the Conference
on Empirical Methods in Natural Language Processing.
Anais...Association for Computational Linguistics,
2020. Disponível em: <https://aclanthology.org/2020.acl-main.448/>
MATTHEWS, B. W. Comparison of the
predicted and observed secondary structure of T4 phage lysozyme.
Biochimica et Biophysica Acta (BBA) - Protein
Structure, v. 405, n. 2, p. 442–451, 1975.
MCCANN, B. et al. Learned in Translation: Contextualized Word
Vectors. Proceedings of the 31st International Conference on
Neural Information Processing Systems. Anais...:
NIPS’17.Red Hook, NY, USA: Curran Associates Inc., 2017.
MCMILLAN-MAJOR, A.; BENDER, E. M.; FRIEDMAN, B. Data Statements:
From Technical Concept to
Community Practice. ACM Journal on
Responsible Computing, p. 3594737, 2023.
MCQUILLAN, D. Resisting AI: an anti-fascist approach to
artificial intelligence. Bristol, UK: [s.n.].
MECKLENBURG, N. et al. Injecting New Knowledge into Large
Language Models via Supervised Fine-Tuning., 2024. Disponível
em: <https://arxiv.org/abs/2404.00213>
MEJIAS, U. A.; COULDRY, N. Datafication.
Internet Policy Review, v. 8, n. 4, nov. 2019.
MEVA, DR. D.; KUKADIYA, H. Performance
Evaluation of Large Language Models: A Comprehensive Review.
International Research Journal of Computer Science, v.
12, p. 109–114, mar. 2025.
MICELI, M. et al. Who Trains the Data for European Artificial
Intelligence? The Left, DiPLab, Weizenbaum Institute; DAIR
Institute., 2024. Disponível em: <https://hal.science/hal-04662589v1>
MIIKKULAINEN, R.; DYER, M. G. Natural Language
Processing With Modular Pdp Networks and Distributed Lexicon.
Cognitive Science, v. 15, n. 3, p. 343–399, 1991.
MILLER, B. Is
Technology Value-Neutral?
Science, Technology, & Human Values, v. 46, n. 1,
p. 53–80, jan. 2021.
MINAEE, S. et al. Large language models: A survey. arXiv
preprint arXiv:2402.06196, 2024.
MINAEE, S. et al. Large Language Models: A Survey.,
2025. Disponível em: <https://arxiv.org/abs/2402.06196>
MITCHELL, M. et al. Model Cards for
Model Reporting. Proceedings of the
Conference on Fairness,
Accountability, and Transparency.
Anais...Atlanta GA USA: ACM, jan. 2019. Disponível em:
<https://dl.acm.org/doi/10.1145/3287560.3287596>
MIZRAHI, D. et al. Language Models Improve When Pretraining Data
Matches Target Tasks., 2025. Disponível em: <https://arxiv.org/abs/2507.12466>
MOHAMMAD, S. Ethics Sheets for AI
Tasks. Proceedings of the 60th Annual
Meeting of the Association for
Computational Linguistics (Volume
1: Long Papers).
Anais...Dublin, Ireland: Association for Computational
Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.acl-long.573>
MOREIRA, J.; VIANNA, D. Challenges of Generative AI
in Legal Practice: A Jusbrasil Perspective.
Proceedings of the 40th Brazilian Symposium on Databases
(SBBD 2025). Anais...Brazil: Brazilian
Computer Society (SBC), 2025. Disponível em: <https://sbbd.org.br/2025/wp-content/uploads/2025/10/SBBD25-Jusbrasil.pdf>
MORESCHI, B.; PEREIRA, G.; COZMAN, F. G. The
Brazilian Workers in Amazon
Mechanical Turk: Dreams and
realities of ghost workers. Revista Contracampo, v.
39, n. 1, abr. 2020.
MOROZOV, E. To save everything, click here : the folly of
technological solutionism. New York, NY: PublicAffairs, 2014.
MUNGER, K. Chatbots for Good and
Evil. EACL via Underline Science Inc., 2023.
Disponível em: <https://underline.io/lecture/72154-chatbots-for-good-and-evil>
NAIR, P. A. et al. Matryoshka Quantization. Sparsity in
LLMs (SLLM): Deep Dive into Mixture of Experts, Quantization, Hardware,
and Inference. Anais...2025.
NARAYAN, S.; COHEN, S. B.; LAPATA, M. Don’t Give Me the Details,
Just the Summary! Topic-Aware Convolutional Neural Networks
for Extreme Summarization. Proceedings of the 2018 Conference
on Empirical Methods in Natural Language Processing.
Anais...Brussels, Belgium: 2018.
NATIONAL INSTITUTE OF STANDARDS AND TECHNOLOGY. A Plan for
Global Engagement on AI Standards. [s.l.] U.S. Department of
Commerce, National Institute of Standards; Technology, jul. 2024.
Disponível em: <https://nvlpubs.nist.gov/nistpubs/ai/NIST.AI.100-5.pdf>.
NEPOMUCENO, J. et al. Agentic AI Meets Fundamentalist Financial
Analysis: Limits and Possibilities. Proceedings of the 14th
Brazilian Conference on Intelligent Systems (BRACIS 2025).
Anais...2025.
NGO, R.; CHAN, L.; MINDERMANN, S. The Alignment Problem from a Deep
Learning Perspective. arXiv preprint arXiv:2303.16200,
2023.
NIJKAMP, E. et al. ProGen2: Exploring the
Boundaries of Protein Language Models. CoRR, v.
abs/2206.13517, 2022.
NOVOBILSKÁ, L. Free and Open
Source Software Licensing Requirements and Copyright Infringement
Involving Artificial Intelligence Technologies. 2023.
O’NEIL, C. Weapons of math destruction : how big data increases
inequality and threatens democracy. First edition ed. New York:
[s.n.].
OLIVEIRA, L. L. Inteligência
artificial e desigualdade social: o impacto do colonialismo digital nas
políticas públicas. Internet & Sociedade, v. 5,
n. 1, 2024.
OUYANG, L. et al. Training language models to follow
instructions with human feedback. (A. H. Oh et al.,
Eds.)Advances in Neural Information Processing Systems.
Anais...2022. Disponível em: <https://openreview.net/forum?id=TG8KACxEON>
OVERWIJK, A.; XIONG, C.; CALLAN, J. ClueWeb22: 10 Billion Web
Documents with Rich Information. (E. Amigó et al.,
Eds.)SIGIR ’22: The 45th International ACM
SIGIR Conference on Research and Development in Information
Retrieval, Madrid, Spain, July 11 - 15, 2022.
Anais...ACM, 2022. Disponível em: <https://doi.org/10.1145/3477495.3536321>
PAIOLA, P. H. et al. Adapting llms for the medical domain in portuguese:
A study on fine-tuning and model evaluation. arXiv preprint
arXiv:2410.00163, 2024.
PAN, S. et al. Unifying large language models and knowledge graphs: A
roadmap. IEEE Transactions on Knowledge and Data
Engineering, v. 36, n. 7, p. 3580–3599, 2024.
PAN, Y. et al. On the Risk of
Misinformation Pollution with
Large Language Models.
Findings of the Association for Computational
Linguistics: EMNLP 2023.
Anais...Singapore: Association for Computational
Linguistics, 2023. Disponível em: <https://aclanthology.org/2023.findings-emnlp.97>
PANICKSSERY, A.; BOWMAN, S. R.; FENG, S. LLM Evaluators
Recognize and Favor Their Own Generations. Proceedings of
the 38th International Conference on Neural Information Processing
Systems, 2024.
PAPINENI, K. et al. BLEU: A Method for Automatic
Evaluation of Machine Translation. Proceedings of the
40th Annual Meeting on Association for Computational Linguistics.
Anais...: ACL ’02.USA: Association for Computational
Linguistics, 2002. Disponível em: <https://doi.org/10.3115/1073083.1073135>
PARK, K. et al. OffsetBias: Leveraging Debiased Data for Tuning
Evaluators., 2024. Disponível em: <https://arxiv.org/abs/2407.06551>
PARMAR, M. et al. Don’t Blame the Annotator: Bias Already Starts
in the Annotation Instructions. Proceedings of the 17th
Conference of the European Chapter of the Association for Computational
Linguistics. Anais...Dubrovnik, Croatia: Association
for Computational Linguistics, 2023. Disponível em: <https://aclanthology.org/2023.eacl-main.130>
PASSINATO, E.; RIOS, W.; FILHO, A. G. Integração de Modelos de
Linguagem e RAG na Criação de Chatbots Oftalmológicos. Anais do
XXIV Simpósio Brasileiro de Computação Aplicada à Saúde.
Anais...Porto Alegre, RS, Brasil: SBC, 2024. Disponível
em: <https://sol.sbc.org.br/index.php/sbcas/article/view/28831>
PAULLADA, A. et al. Data and its
(dis)contents: A survey of dataset development and use in
machine learning research. Patterns, v. 2, n. 11,
p. 100336, nov. 2021.
PETERS, M. E. et al. Deep Contextualized Word
Representations. (M. A. Walker, H. Ji, A. Stent,
Eds.)Proceedings of the 2018 Conference of the North American Chapter of
the Association for Computational Linguistics: Human Language
Technologies, NAACL-HLT 2018, New Orleans, Louisiana, USA,
June 1-6, 2018, Volume 1 (Long Papers).
Anais...Association for Computational Linguistics,
2018. Disponível em: <https://doi.org/10.18653/v1/n18-1202>
PEYRARD, M. Studying Summarization Evaluation Metrics in the
Appropriate Scoring Range. Proceedings of the Conference on
Empirical Methods in Natural Language Processing (EMNLP).
Anais...Online: Association for Computational
Linguistics, 2019. Disponível em: <https://aclanthology.org/P19-1502/>
PIĘKOS, P.; MALINOWSKI, M.; MICHALEWSKI, H. Measuring and
Improving BERT’s Mathematical Abilities by
Predicting the Order of Reasoning. Proceedings of the 59th
Annual Meeting of the Association for Computational Linguistics and the
11th International Joint Conference on Natural Language Processing
(Volume 2: Short Papers). Anais...Online: Association
for Computational Linguistics, ago. 2021. Disponível em: <https://aclanthology.org/2021.acl-short.49>
PIRES, R. et al. Sabiá: Portuguese Large Language
Models. (M. C. Naldi, R. A. C. Bianchi, Eds.)Intelligent
Systems. Anais...Cham: Springer Nature Switzerland,
2023.
PLOUG, T. The
Right Not to Be
Subjected to AI Profiling
Based on Publicly Available
Data—Privacy and the
Exceptionalism of AI
Profiling. Philosophy &
Technology, v. 36, n. 1, p. 14, mar. 2023.
POLO, F. M. et al. Efficient
Multi-Prompt Evaluation of LLMs. Proceedings of
the 38th Conference on Neural Information Processing Systems (NeurIPS
2024), 2024.
POMBAL, J. et al. M-Prometheus:
A Suite of Open Multilingual LLM Judges.
CoRR, v. abs/2504.04953, 2025.
PROTOTYPEJAM. Lake Merritt: AI Evaluation Workbench. https://prototypejam.github.io/lake_merritt/, 2025.
PROVILKOV, I.; EMELIANENKO, D.; VOITA, E. BPE-Dropout:
Simple and Effective Subword Regularization. Proceedings of
the 58th Annual Meeting of the Association for Computational
Linguistics. Anais...Online: Association for
Computational Linguistics, jul. 2020.
RACKAUCKAS, Z. Rag-fusion: a new take on retrieval-augmented generation.
arXiv preprint arXiv:2402.03367, 2024.
RADFORD, A. et al. Improving Language Understanding
by Generative Pre-Training., 2018. Disponível em:
<https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf>
RADFORD, A. et al. Language Models are Unsupervised Multitask
Learners. a2019.
RADFORD, A. et al. Language models are unsupervised multitask learners.
OpenAI blog, v. 1, n. 8, p. 9, b2019.
RADFORD, A. et al. Learning Transferable Visual Models From
Natural Language Supervision., 2021. Disponível em: <https://arxiv.org/abs/2103.00020>
RADFORD, A.; NARASIMHAN, K. Improving Language Understanding by
Generative Pre-Training. 2018.
RAE, J. W. et al. Scaling
Language Models: Methods, Analysis & Insights from
Training Gopher. CoRR, v. abs/2112.11446, 2021.
RAE, J. W. et al. Scaling Language Models: Methods, Analysis
& Insights from Training Gopher., 2022. Disponível em:
<https://arxiv.org/abs/2112.11446>
RAFAILOV, R. et al. Direct Preference Optimization: Your
Language Model is Secretly a Reward Model. Advances in Neural
Information Processing Systems (NeurIPS). Anais...2023.
RAFFEL, C. et al. Exploring the Limits of
Transfer Learning with a Unified Text-to-Text Transformer.
Journal of Machine Learning Research, v. 21, n. 140, p.
1–67, a2020.
RAFFEL, C. et al. Exploring the limits of transfer learning with a
unified text-to-text transformer. J. Mach. Learn. Res.,
v. 21, n. 1, p. 140:5485–140:5551, jan. b2020.
RAFFEL, C. et al. Exploring the Limits of Transfer Learning with
a Unified Text-to-Text Transformer., 2023. Disponível em:
<https://arxiv.org/abs/1910.10683>
RAJI, D. et al. AI and the Everything in the Whole Wide World
Benchmark. (J. Vanschoren, S. Yeung, Eds.)Proceedings of the
Neural Information Processing Systems Track on Datasets and Benchmarks.
Anais...Curran, 2021. Disponível em: <https://datasets-benchmarks-proceedings.neurips.cc/paper_files/paper/2021/file/084b6fbb10729ed4da8c3d3f5a3ae7c9-Paper-round2.pdf>
RAJPURKAR, P. et al. SQuAD: 100,000+
Questions for Machine Comprehension of Text. (J. Su, K. Duh, X.
Carreras, Eds.)Proceedings of the 2016 Conference on Empirical Methods
in Natural Language Processing. Anais...Austin, Texas:
Association for Computational Linguistics, nov. 2016. Disponível em:
<https://aclanthology.org/D16-1264>
REAL, L.; CARVALHO, A.; SILVA, A. DA. Avaliação
de Grandes Modelos de Linguagem: Fundamentos, Métodos
Tradicionais e Desafios Atuais. Em: CASELI, H. M.; NUNES, M. G. V.
(Eds.). Processamento de Linguagem Natural: Conceitos,
Técnicas e Aplicações em
Português. 3. ed. [s.l.] BPLN, 2024.
REAL, L.; FONSECA, E.; GONÇALO OLIVEIRA, H. The ASSIN 2 Shared
Task: A Quick Overview. Computational Processing of the
Portuguese Language: 14th International Conference, PROPOR 2020, Evora,
Portugal, March 2–4, 2020, Proceedings. Anais...Berlin,
Heidelberg: Springer-Verlag, 2020. Disponível em: <https://doi.org/10.1007/978-3-030-41505-1_39>
REHAK, R. AI Narrative Breakdown. A Critical Assessment of Power
and Promise. Proceedings of the 2025 ACM Conference on
Fairness, Accountability, and Transparency. Anais...:
FAccT ’25.New York, NY, USA: Association for Computing Machinery, 2025.
Disponível em: <https://doi.org/10.1145/3715275.3732083>
REIMERS, N.; GUREVYCH, I. Sentence-BERT: Sentence Embeddings
using Siamese BERT-Networks. Proceedings of the 2019 Conference
on Empirical Methods in Natural Language Processing.
Anais...Association for Computational Linguistics, nov.
2019. Disponível em: <https://arxiv.org/abs/1908.10084>
REIMERS, N.; GUREVYCH, I. Making Monolingual Sentence Embeddings
Multilingual using Knowledge Distillation. Proceedings of the
2020 Conference on Empirical Methods in Natural Language Processing.
Anais...Association for Computational Linguistics, nov.
2020. Disponível em: <https://arxiv.org/abs/2004.09813>
RIBEIRO, M. T. et al. Beyond Accuracy: Behavioral Testing of
NLP Models with
CheckList. Proceedings of the 58th
Annual Meeting of the Association for Computational Linguistics.
Anais...Online: Association for Computational
Linguistics, jul. 2020. Disponível em: <https://aclanthology.org/2020.acl-main.442>
RÍO, B. G. DEL; VAAHTIO, T. Improving LLM systems with A/B
testing. https://www.flow-ai.com/blog/improving-llm-systems-with-a-b-testing,
2024.
RODRIGUES, J. et al. Advancing Neural
Encoding of Portuguese with Transformer Albertina PT-.
CoRR, v. abs/2305.06721, 2023.
RODRIGUES, R. C. et al. Portuguese Language
Models and Word Embeddings: Evaluating on Semantic Similarity
Tasks. (P. Quaresma et al., Eds.)Computational Processing
of the Portuguese Language. Anais...Springer Nature
Switzerland AG: Springer International Publishing, 2020.
ROGERS, A. Changing the World by
Changing the Data. Proceedings of the
59th Annual Meeting of the
Association for Computational
Linguistics and the 11th International
Joint Conference on Natural
Language Processing (Volume 1:
Long Papers). Anais...Online:
Association for Computational Linguistics, 2021. Disponível em: <https://aclanthology.org/2021.acl-long.170>
ROGERS, A.; BALDWIN, T.; LEINS, K. “Just
What do You Think
You’re Doing, Dave?”
A Checklist for Responsible
Data Use in NLP.
Findings of the Association for Computational
Linguistics: EMNLP 2021.
Anais...Punta Cana, Dominican Republic: Association for
Computational Linguistics, 2021. Disponível em: <https://aclanthology.org/2021.findings-emnlp.414>
ROMERA-PAREDES, B.; TORR, P. H. S. An embarrassingly simple
approach to zero-shot learning. (F. R. Bach, D. M. Blei,
Eds.)Proceedings of the 32nd International Conference on Machine
Learning, ICML 2015, Lille, France, 6-11 July 2015.
Anais...: JMLR Workshop e Conference
Proceedings.JMLR.org, 2015. Disponível em: <http://proceedings.mlr.press/v37/romera-paredes15.html>
RUSSEL, S. Human Compatible Artificial Intelligence and the
Problem of Control. [s.l.] Penguin Books, 2019.
SAI, A. B.; MOHANKUMAR, A. K.; KHAPRA, M. M. A Survey of Evaluation Metrics
Used for NLG Systems. ACM Comput.
Surv., v. 55, n. 2, p. 26:1–26:39, 2023.
SAID, H. 40 Large Language Model Benchmarks and The Future of
LLMs. Arize AI; https://arize.com/blog/llm-benchmarks-mmlu-codexglue-gsm8k,
2025.
SAMPAIO, R. C.; SABBATINI, M.; LIMONGI, R. Diretrizes
para o uso ético e responsável da
inteligência artificial generativa: um guia
prático para pesquisadores. Boletim
Técnico do PPEC, v. 10, p. e025003–e025003, 2024.
SAMUYLOVA, E. LLM‑as‑a‑Judge: A
Complete Guide to Using LLMs for
Evaluations. https://www.evidentlyai.com/llm-guide/llm-as-a-judge,
2025.
SÁNCHEZ, L. C. Ethical Considerations and Best Practices in LLM
Development. https://neptune.ai/blog/llm-ethical-considerations,
2024.
SANH, V. et al. DistilBERT, a
distilled version of BERT: smaller, faster, cheaper and
lighter. CoRR, v. abs/1910.01108, 2019.
SANTOS, A. A. et al. O teste de Cloze na
avaliação da compreensão em
leitura. Psicologia: reflexão e
crı́tica, v. 15, p. 549–560, 2002.
SANTOS, D. O projecto Processamento
Computacional do Português:
Balanço e perspectivas. (M. das Graças Volpe
Nunes, Ed.)V Encontro para o processamento
computacional da língua portuguesa escrita e falada (PROPOR
2000). Anais...São Paulo: ICMC/USP, 2000. Disponível
em: <https://www.linguateca.pt/Diana/download/SantosPROPOR2000.pdf>
SANTOS, H.; WOLOSZYN, V.; VIEIRA, R.
BlogSet-BR: A
Brazilian Portuguese Blog Corpus. (N.
Calzolari et al., Eds.)Proceedings of the Eleventh International
Conference on Language Resources and Evaluation (LREC
2018). Anais...Miyazaki, Japan: European Language
Resources Association (ELRA), 2018. Disponível em: <https://aclanthology.org/L18-1105/>
SARDINHA, T. B.; FILHO, J. L. M.; ALAMBERT, E. Manual Corpus
Brasileiro. [s.l: s.n.].
SARDINHA, T. B.; KAUFFMANN, C.; ACUNZO, C. M. A multi-dimensional
analysis of register variation in Brazilian Portuguese.
Corpora, v. 9, n. 2, p. 239–271, 2014.
SARTHI, P. et al. RAPTOR: Recursive Abstractive Processing for
Tree-Organized Retrieval. International Conference on Learning
Representations (ICLR). Anais...2024.
SAWCZYN, A. et al. FactSelfCheck:
Fact-Level Black-Box Hallucination Detection for
LLMs. arXiv, 2025.
SCAO, T. L. et al. BLOOM:
A 176B-Parameter Open-Access Multilingual Language
Model. CoRR, v. abs/2211.05100, 2022.
SCHAAKE, M. The
Tech Coup: How to Save Democracy from Silicon Valley.
[s.l.] Princeton University Press, 2024.
SCHEUERMAN, M. K.; HANNA, A.; DENTON, E. Do Datasets
Have Politics? Disciplinary
Values in Computer Vision
Dataset Development. Proceedings
of the ACM on Human-Computer Interaction, v. 5, n. CSCW2, p.
1–37, out. 2021.
SCHICK, T.; SCHÜTZE, H. Exploiting Cloze-Questions for Few-Shot
Text Classification and Natural Language Inference. (P. Merlo,
J. Tiedemann, R. Tsarfaty, Eds.)Proceedings of the 16th Conference of
the European Chapter of the Association for Computational Linguistics:
Main Volume, EACL 2021, Online, April 19 - 23, 2021.
Anais...Association for Computational Linguistics,
2021. Disponível em: <https://doi.org/10.18653/v1/2021.eacl-main.20>
SCHIRRU, L. Direito
autoral e Inteligência Artificial: autoria e titularidade nos produtos
da IA. [s.l.] Dialetica, 2023.
SCHLANGEN, D. Targeting the Benchmark: On Methodology in Current
Natural Language Processing Research. Proceedings of the 59th
Annual Meeting of the Association for Computational Linguistics and the
11th International Joint Conference on Natural Language Processing
(Volume 2: Short Papers). Anais...Online: Association
for Computational Linguistics, ago. 2021. Disponível em: <https://aclanthology.org/2021.acl-short.85>
SCHMIDHUBER, J.; HEIL, S. Sequential neural text
compression. IEEE Transactions on Neural Networks,
v. 7, n. 1, p. 142–146, 1996.
SCHNABEL, T. et al. Evaluation methods for unsupervised word
embeddings. (L. Màrquez, C. Callison-Burch, J. Su,
Eds.)Proceedings of the 2015 Conference on Empirical Methods in Natural
Language Processing. Anais...Lisbon, Portugal:
Association for Computational Linguistics, set. 2015. Disponível em:
<https://aclanthology.org/D15-1036/>
SCHULMAN, J. et al. Proximal Policy Optimization
Algorithms. arXiv preprint arXiv:1707.06347.
Anais...2017.
SCHUSTER, M.; NAKAJIMA, K. Japanese and Korean
voice search. 2012 IEEE International Conference on
Acoustics, Speech and Signal Processing (ICASSP).
Anais...2012.
SCHWARTZ, L. Primum Non Nocere:
Before working with Indigenous data, the
ACL must confront ongoing colonialism. Proceedings
of the 60th Annual Meeting of the
Association for Computational
Linguistics (Volume 2: Short
Papers). Anais...Dublin, Ireland:
Association for Computational Linguistics, 2022. Disponível em: <https://aclanthology.org/2022.acl-short.82>
SELVAN, R. et al. Carbon Footprint of Selecting and Training
Deep Learning Models for Medical Image Analysis. (L. Wang et
al., Eds.)Medical Image Computing and Computer Assisted Intervention –
MICCAI 2022. Anais...Cham: Springer Nature Switzerland,
2022. Disponível em: <https://doi.org/10.1007/978-3-031-16443-9_49>
SERRAS, F. R. et al. Exploring Computational Discernibility of
Discourse Domains in Brazilian Portuguese
within the Carolina Corpus. (P. Gamallo et al.,
Eds.)Proceedings of the 16th International Conference on Computational
Processing of Portuguese - Vol. 1. Anais...Santiago de
Compostela, Galicia/Spain: Association for Computational Lingustics,
mar. 2024. Disponível em: <https://aclanthology.org/2024.propor-1.26/>
SHAH, D.; SCHWARTZ, H. A.; HOVY, D. Predictive
Biases in Natural Language
Processing Models: A
Conceptual Framework and
Overview. Proceedings of the 58th
Annual Meeting of the Association
for Computational Linguistics.
Anais...Online: Association for Computational
Linguistics, 2020. Disponível em: <http://arxiv.org/abs/1912.11078>
SHAH, R. et al. Goal Misgeneralization in Deep Reinforcement Learning.
arXiv preprint arXiv:2210.01790, 2022.
SHANNON, C. E. Prediction and entropy of printed English. Bell
System Technical Journal, v. 30, n. 1, p. 50–64, 1951.
SHAQIRI, M. et al. Differences between the correlation coefficients
Pearson, Kendall and Spearman.
5th International Conference of Natural Science and
Mathematics, nov. 2023.
SHENG, E. et al. The Woman Worked as a Babysitter: On Biases in
Language Generation. (K. Inui et al., Eds.)Proceedings of the
2019 Conference on Empirical Methods in Natural Language Processing and
the 9th International Joint Conference on Natural Language Processing
(EMNLP-IJCNLP). Anais...Hong Kong, China: Association
for Computational Linguistics, nov. 2019. Disponível em: <https://aclanthology.org/D19-1339/>
SHI, H. et al. Continual learning of large language models: A
comprehensive survey. ACM Computing Surveys, 2024.
SHI, Z.; LIPANI, A. Don’t Stop Pretraining? Make Prompt-based
Fine-tuning Powerful Learner., 2023. Disponível em: <https://arxiv.org/abs/2305.01711>
SHMUELI, B. et al. Beyond Fair Pay:
Ethical Implications of NLP
Crowdsourcing. Proceedings of the 2021
Conference of the North American
Chapter of the Association for
Computational Linguistics: Human
Language Technologies.
Anais...Online: Association for Computational
Linguistics, 2021. Disponível em: <https://aclanthology.org/2021.naacl-main.295>
SHUMAILOV, I. et al. AI models collapse
when trained on recursively generated data. Nature,
v. 631, n. 8022, p. 755–759, 2024.
ŞIMŞEK, C. AI resistance: Who says no to AI and why?
Zenodo, 2025. Disponível em: <https://zenodo.org/doi/10.5281/zenodo.16893847>
ŞIMŞEK, C.; YASAR, A. G. From Rejection to
Regulation: Mapping the Landscape of AI Resistance. 2025.
SØGAARD, A.; PLANK, B.; HOVY, D. Selection Bias, Label Bias, and
Bias in Ground Truth. (Q. Liu, F. Xia, Eds.)Proceedings of
COLING 2014, the 25th International Conference on
Computational Linguistics: Tutorial Abstracts.
Anais...Dublin, Ireland: Dublin City University;
Association for Computational Linguistics, ago. 2014. Disponível em:
<https://aclanthology.org/C14-3005/>
SOLAIMAN, I. et al. Evaluating
the Social Impact of Generative AI Systems. Em: The Oxford
Handbook of the Foundations and Regulation of Generative AI.
[s.l.] Oxford University Press, 2025.
SOUZA, F.; NOGUEIRA, R.; LOTUFO, R. BERTimbau: pretrained BERT
models for Brazilian Portuguese. (R. Cerri, R. C. Prati,
Eds.)Proceedings of the 2020 Brazilian Conference on Intelligent
Systems. Anais...Springer International Publishing,
2020.
STIENNON, N. et al. Learning to summarize with human
feedback. (H. Larochelle et al., Eds.)Advances in Neural
Information Processing Systems. Anais...Curran
Associates, Inc., a2020. Disponível em: <https://proceedings.neurips.cc/paper_files/paper/2020/file/1f89885d556929e98d3ef9b86448f951-Paper.pdf>
STIENNON, N. et al. Learning to Summarize with Human Feedback.
Advances in Neural Information Processing Systems
(NeurIPS), b2020.
STRUBELL, E.; GANESH, A.; MCCALLUM, A. Energy and
Policy Considerations for Deep
Learning in NLP. Proceedings of the
57th Annual Meeting of the
Association for Computational
Linguistics. Anais...Florence, Italy:
Association for Computational Linguistics, 2019. Disponível em: <https://www.aclweb.org/anthology/P19-1355>
STURZENEKER, M. et al. Carolina’s Methodology: building a large
corpus with provenance and typology information. (C. Trojahn et
al., Eds.)Proceedings of the Second Workshop on Digital Humanities
and Natural Language Processing (2nd DHandNLP 2022) co-located with
International Conference on the Computational Processing of Portuguese
(PROPOR 2022). Anais...Fortaleza, Brazil: 2022.
Disponível em: <https://ceur-ws.org/Vol-3128/paper4.pdf>
SUN, K.; DREDZE, M. Amuro & Char: Analyzing the
Relationship between Pre-Training and Fine-Tuning of Large Language
Models. (V. Adlakha et al., Eds.)Proceedings of the 10th
Workshop on Representation Learning for NLP (RepL4NLP-2025).
Anais...Albuquerque, NM: Association for Computational
Linguistics, 2025. Disponível em: <https://aclanthology.org/2025.repl4nlp-1.11/>
SURESH, H.; GUTTAG, J. V. A Framework
for Understanding Sources of Harm
throughout the Machine Learning
Life Cycle. Equity and Access in
Algorithms, Mechanisms, and Optimization, p. 1–9, out. 2021.
SUTSKEVER, I.; VINYALS, O.; LE, Q. V. Sequence to Sequence
Learning with Neural Networks. (Z. Ghahramani et al.,
Eds.)Advances in Neural Information Processing Systems 27: Annual
Conference on Neural Information Processing Systems 2014, December 8-13
2014, Montreal, Quebec, Canada. Anais...2014.
Disponível em: <https://proceedings.neurips.cc/paper/2014/hash/a14ac55a4f27472c5d894ec1c3c743d2-Abstract.html>
TAMKIN, A. et al. Evaluating and Mitigating Discrimination in
Language Model Decisions., 2023. Disponível em: <https://arxiv.org/abs/2312.03689>
TANG, Y. et al. Multilingual
Translation with Extensible Multilingual Pretraining and Finetuning.
CoRR, v. abs/2008.00401, 2020.
TAYLOR, R. et al. Galactica:
A Large Language Model for Science.
CoRR, v. abs/2211.09085, 2022.
TAYLOR, W. L. “Cloze procedure”: A new tool for measuring
readability. Journalism quarterly, v. 30, n. 4, p.
415–433, 1953.
TEI CONSORTIUM. TEI
P5: Guidelines for Electronic Text Encoding and
Interchange. [s.l.] Retrieved May 20, 2021, 2021.
THAKUR, A. S. et al. JUDGING THE JUDGES: EVALUATING
ALIGNMENT AND VULNERABILITIES IN LLMs-AS-JUDGES.
arXiv preprint arXiv:2406.12624v5, 2025.
THOPPILAN, R. et al. LaMDA:
Language Models for Dialog Applications. CoRR, v.
abs/2201.08239, 2022.
THUN, M. VON; HANLEY, D. A. Stopping Big Tech from Becoming Big
AI: A Roadmap for Using Competition Policy to Keep
Artificial Intelligence Open for All. Open Markets Institute,
2024. Disponível em: <https://www.openmarketsinstitute.org/publications/report-stopping-big-tech-big-ai-roadmap>
TOLLES, J.; MEURER, W. J. Logistic Regression:
Relating Patient Characteristics to Outcomes. JAMA,
v. 316, n. 5, p. 533–534, ago. 2016.
TONIAZZO, D.; BARBOSA, T.; RUARO, R. O Direito à
Explicação nas Decisões Automatizadas: uma Abordagem Comparativa Entre o
Ordenamento Brasileiro e Europeu. Revista Internacional
Consinter de Direito, v. 13, p. 55–69, dez. 2021.
TOUVRON, H. et al. LLaMA: Open and
Efficient Foundation Language Models. CoRR, v.
abs/2302.13971, a2023.
TOUVRON, H. et al. LLaMA: Open and Efficient Foundation Language
Models., b2023. Disponível em: <https://arxiv.org/abs/2302.13971>
TRAUTMANN, D. et al. Measuring the Groundedness of Legal
Question-Answering Systems. Proceedings of the Natural Legal
Language Processing Workshop 2024. Anais...Singapore:
Association for Computational Linguistics, 2024. Disponível em: <https://aclanthology.org/2024.nllp-1.14>
ULMER, D. et al. Experimental Standards for Deep Learning in
Natural Language Processing Research. Findings of the
Association for Computational Linguistics: EMNLP 2022.
Anais...Abu Dhabi, United Arab Emirates: Association
for Computational Linguistics, dez. 2022. Disponível em: <https://aclanthology.org/2022.findings-emnlp.196>
VAROQUAUX, G.; LUCCIONI, S.; WHITTAKER, M. Hype,
Sustainability, and the Price of the
Bigger-is-Better Paradigm in
AI. Proceedings of the 2025 ACM
Conference on Fairness,
Accountability, and Transparency.
Anais...Athens Greece: ACM, jun. 2025. Disponível em:
<https://dl.acm.org/doi/10.1145/3715275.3732006>
VASWANI, A. et al. Attention is All you Need. (I. Guyon
et al., Eds.)Advances in Neural Information Processing Systems.
Anais...Curran Associates, Inc., a2017. Disponível em:
<https://proceedings.neurips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html>
VASWANI, A. et al. Attention is all you need. Advances in neural
information processing systems, v. 30, b2017.
VONGTHONGSRI, K. G-Eval Simply Explained:
LLM-as-a-Judge for LLM
Evaluation. Confident AI; https://www.confident-ai.com/blog/g-eval-the-definitive-guide,
2025.
WAGNER FILHO, J. A. et al. The brWaC
Corpus: A New Open Resource for Brazilian
Portuguese. (N. Calzolari et al., Eds.)Proceedings
of the Eleventh International Conference on Language Resources and
Evaluation (LREC 2018). Anais...Miyazaki,
Japan: European Language Resources Association (ELRA), 2018. Disponível
em: <https://aclanthology.org/L18-1686>
WANG, A. et al. GLUE: A Multi-Task Benchmark and
Analysis Platform for Natural Language Understanding.
Proceedings of the 2018 EMNLP Workshop
BlackboxNLP: Analyzing and Interpreting Neural
Networks for NLP. Anais...Brussels,
Belgium: Association for Computational Linguistics, nov. 2018.
Disponível em: <https://aclanthology.org/W18-5446/>
WANG, A. et al. SuperGLUE: A Stickier Benchmark for General-Purpose
Language Understanding Systems. Advances in Neural Information
Processing Systems, v. 32, p. 3261–3275, b2019.
WANG, A. et al. SuperGLUE: a stickier benchmark for general-purpose
language understanding systems. Em: Proceedings of the 33rd
International Conference on Neural Information Processing
Systems. Red Hook, NY, USA: Curran Associates Inc., 2019a.
WANG, B.; KOMATSUZAKI, A. GPT-J-6B: A 6 Billion Parameter
Autoregressive Language Model. https://github.com/kingoflolz/mesh-transformer-jax,
2021.
WANG, R. et al. Can
LLMs Replace Human Evaluators? An Empirical Study of
LLM-as-a-Judge in Software Engineering. Proc.
ACM Softw. Eng., v. 2, n. ISSTA, jun. 2025.
WANG, Y. et al. CodeT5: Identifier-aware Unified Pre-trained
Encoder-Decoder Models for Code Understanding and Generation.
(M.-F. Moens et al., Eds.)Proceedings of the 2021 Conference on
Empirical Methods in Natural Language Processing, EMNLP
2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November,
2021. Anais...Association for Computational
Linguistics, 2021. Disponível em: <https://doi.org/10.18653/v1/2021.emnlp-main.685>
WANG, Y. et al. PandaLM: An Automatic Evaluation Benchmark for
LLM Instruction Tuning Optimization. The Twelfth
International Conference on Learning Representations, ICLR
2024, Vienna, Austria, May 7-11, 2024.
Anais...OpenReview.net, 2024. Disponível em: <https://openreview.net/forum?id=5Nn2BLV7SB>
WEBER, M. et al. RedPajama: an Open Dataset for Training Large
Language Models., 2024. Disponível em: <https://arxiv.org/abs/2411.12372>
WEI, J. et al. Emergent Abilities of
Large Language Models. Trans. Mach. Learn. Res., v.
2022, b2022.
WEI, J. et al. Chain-of-Thought Prompting Elicits Reasoning in
Large Language Models. NeurIPS. Anais...a2022.
Disponível em: <http://papers.nips.cc/paper\_files/paper/2022/hash/9d5609613524ecf4f15af0f7b31abca4-Abstract-Conference.html>
WEIDINGER, L. et al. Ethical
and social risks of harm from Language
Models. arXiv:2112.04359 [cs], dez.
2021.
WEIGEND, A. Data for the People: How
to Make Our Post-Privacy Economy Work for You. [s.l.] Basic
Books, 2017.
WEIZENBAUM, J. Computermacht und Gesellschaft : freie Reden /
Joseph Weizenbaum. Hrsg. von Gunna Wendt .. Original-Ausgabe,
1. Auflage ed. Frankfurt am Main: [s.n.].
WENDEHORST, C. Liability for Artificial Intelligence: The Need to
Address Both Safety Risks and Fundamental Rights Risks. Em: VOENEKY, S.
et al. (Eds.). The Cambridge Handbook of Responsible Artificial
Intelligence: Interdisciplinary Perspectives. Cambridge Law
Handbooks. [s.l.] Cambridge University Press, 2022. p. 187–209.
WENZEK, G. et al. CCNet: Extracting High Quality
Monolingual Datasets from Web Crawl Data. (N. Calzolari et al.,
Eds.)Proceedings of the Twelfth Language Resources and Evaluation
Conference. Anais...Marseille, France: European
Language Resources Association, 2020. Disponível em: <https://aclanthology.org/2020.lrec-1.494/>
WERBOS, P. J. Backpropagation
through time: what it does and how to do it. Proc.
IEEE, v. 78, n. 10, p. 1550–1560, 1990.
WESTENBERGER, J.; SCHULER, K.; SCHLEGEL, D. Failure of
AI projects: understanding the critical factors.
Procedia Computer Science, v. 196, p. 69–76, 2022.
WHITTAKER, M. et al. Disability,
bias, and AI. AI Now Institute, v. 8,
n. 11, 2019.
WHITTAKER, M. The steep cost
of capture. Interactions, v. 28, n. 6, p. 50–55,
nov. 2021.
WIEGREFFE, S.; PINTER, Y. Attention is not not
Explanation. (K. Inui et al., Eds.)Proceedings of the 2019
Conference on Empirical Methods in Natural Language Processing and the
9th International Joint Conference on Natural Language Processing
(EMNLP-IJCNLP). Anais...Hong Kong, China: Association
for Computational Linguistics, nov. 2019. Disponível em: <https://aclanthology.org/D19-1002>
WIELING, M.; RAWEE, J.; VAN NOORD, G. Reproducibility in
Computational Linguistics: Are
We Willing to Share?
Computational Linguistics, v. 44, n. 4, p. 641–649,
dez. 2018.
WOLF, T. et al. Transformers: State-of-the-Art Natural Language
Processing. Proceedings of the 2020 Conference on Empirical
Methods in Natural Language Processing: System Demonstrations.
Anais...Online: Association for Computational
Linguistics, out. 2020. Disponível em: <https://www.aclweb.org/anthology/2020.emnlp-demos.6>
WOLFRAM, W. Variation and
Language: Overview. Em: BROWN, K. (Ed.). Encyclopedia of
Language & Linguistics (Second Edition). Second Edition ed.
Oxford: Elsevier, 2006. p. 333–341.
WRIGHT, B. Manufacturing Reality: Slavoj Zizek and the Reality
of the Virtual. LondonBen Wright Film Productions, 2004.
WU, T. The attention merchants : from the daily newspaper to
social media : how our time and attention is harvested and
sold. London: [s.n.].
WU, X.-K. et al. LLM
Fine-Tuning: Concepts,
Opportunities, and Challenges. Big
Data and Cognitive Computing, v. 9, n. 4, p. 87, abr. 2025.
WU, Y. et al. Memorizing Transformers. The Tenth
International Conference on Learning Representations, ICLR
2022, Virtual Event, April 25-29, 2022.
Anais...OpenReview.net, 2022. Disponível em: <https://openreview.net/forum?id=TrjbxzRcnf->
XIE, S. M. et al. An Explanation of In-context Learning as
Implicit Bayesian Inference. The Tenth International Conference
on Learning Representations, ICLR 2022, Virtual Event,
April 25-29, 2022. Anais...OpenReview.net, 2022.
Disponível em: <https://openreview.net/forum?id=RdJVFCHjUMI>
XIONG, K. et al. Com2 : A Causal-Guided Benchmark for
Exploring Complex Commonsense Reasoning in Large Language
Models. (W. Che et al., Eds.)Proceedings of the 63rd
Annual Meeting of the Association for Computational Linguistics (Volume
1: Long Papers). Anais...Vienna, Austria: Association
for Computational Linguistics, jul. 2025. Disponível em: <https://aclanthology.org/2025.acl-long.785/>
XIONG, R. et al. On Layer Normalization in the Transformer
Architecture. Proceedings of the 37th International Conference
on Machine Learning, ICML 2020, 13-18 July 2020, Virtual
Event. Anais...: Proceedings of Machine Learning
Research.PMLR, 2020. Disponível em: <http://proceedings.mlr.press/v119/xiong20b.html>
XU, A. et al. Detoxifying Language Models Risks Marginalizing
Minority Voices. (K. Toutanova et al., Eds.)Proceedings of the
2021 Conference of the North American Chapter of the Association for
Computational Linguistics: Human Language Technologies.
Anais...Online: Association for Computational
Linguistics, jun. 2021. Disponível em: <https://aclanthology.org/2021.naacl-main.190/>
XU, H.; ASHLEY, K. A question-answering approach to evaluating
legal summaries., 2023. Disponível em: <https://arxiv.org/abs/2309.15016>
XU, Q.; HE, X. Security Challenges in
Natural Language Processing
Models. Proceedings of the 2023
Conference on Empirical Methods
in Natural Language Processing:
Tutorial Abstracts.
Anais...Singapore: Association for Computational
Linguistics, 2023. Disponível em: <https://aclanthology.org/2023.emnlp-tutorial.2>
XU, W.; RUDNICKY, A. Can artificial
neural networks learn language models? Proc. 6th
International Conference on Spoken Language Processing (ICSLP 2000).
Anais...2000.
XU, Y. et al. Hard Sample Aware Prompt-Tuning. (A.
Rogers, J. L. Boyd-Graber, N. Okazaki, Eds.)Proceedings of the 61st
Annual Meeting of the Association for Computational Linguistics (Volume
1: Long Papers), ACL 2023, Toronto, Canada, July 9-14,
2023. Anais...Association for Computational
Linguistics, 2023. Disponível em: <https://aclanthology.org/2023.acl-long.690>
XUE, L. et al. mT5: A Massively Multilingual
Pre-trained Text-to-Text Transformer. (K. Toutanova et al.,
Eds.)Proceedings of the 2021 Conference of the North American Chapter of
the Association for Computational Linguistics: Human Language
Technologies, NAACL-HLT 2021, Online, June 6-11, 2021.
Anais...Association for Computational Linguistics,
2021. Disponível em: <https://doi.org/10.18653/v1/2021.naacl-main.41>
YAMAGUCHI, A. et al. Frustratingly Simple Pretraining
Alternatives to Masked Language Modeling. Proceedings of the
2021 Conference on Empirical Methods in Natural Language Processing.
Anais...Online; Punta Cana, Dominican Republic:
Association for Computational Linguistics, nov. 2021. Disponível em:
<https://aclanthology.org/2021.emnlp-main.249>
YAN, S.-Q. et al. Corrective Retrieval Augmented
Generation., 2024. Disponível em: <https://arxiv.org/abs/2401.15884>
YANG, T. et al. Ethics of Data Work. Principles for Academic
Data Work Requesters. Weizenbaum Institute, 2025. Disponível
em: <https://www.weizenbaum-library.de/handle/id/920>
YANG, Z. et al. XLNet: Generalized Autoregressive Pretraining
for Language Understanding. (H. M. Wallach et al.,
Eds.)Advances in Neural Information Processing Systems 32: Annual
Conference on Neural Information Processing Systems 2019, NeurIPS 2019,
December 8-14, 2019, Vancouver, BC, Canada.
Anais...2019. Disponível em: <https://proceedings.neurips.cc/paper/2019/hash/dc6a7e655d7e5840e66733e9ee67cc69-Abstract.html>
YAO, X. et al. NLP From Scratch Without Large-Scale Pretraining:
A Simple and Efficient Framework., 2022. Disponível em: <https://arxiv.org/abs/2111.04130>
YE, Y. et al. LIMO: Less is
More for Reasoning. arXiv, jul. 2025.
Disponível em: <http://arxiv.org/abs/2502.03387>.
Acesso em: 7 ago. 2025
YEH, Y.-T.; ESKÉNAZI, M.; MEHRI, S. A
Comprehensive Assessment of Dialog Evaluation Metrics.
ArXiv, v. abs/2106.03706, 2021.
YEO, W. et al. UniversalRAG: Retrieval-Augmented Generation over Corpora
of Diverse Modalities and Granularities. arXiv preprint
arXiv:2504.20734, 2025.
YUAN, W.; NEUBIG, G.; LIU, P. BARTScore: Evaluating Generated
Text as Text Generation. Advances in Neural Information
Processing Systems 34: Annual Conference on Neural Information
Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, virtual.
Anais...2021. Disponível em: <https://proceedings.neurips.cc/paper/2021/hash/e4d2b6e6fdeca3e60e0f1a62fee3d9dd-Abstract.html>
ZHANG, D.; XU, Z.; ZHAO, W. LLMs and Copyright
Risks: Benchmarks and Mitigation Approaches. (M. Lomeli, S.
Swayamdipta, R. Zhang, Eds.)Proceedings of the 2025 Annual Conference of
the Nations of the Americas Chapter of the Association for Computational
Linguistics: Human Language Technologies (Volume 5: Tutorial Abstracts).
Anais...Albuquerque, New Mexico: Association for
Computational Linguistics, a2025. Disponível em: <https://aclanthology.org/2025.naacl-tutorial.7/>
ZHANG, K. et al. CitaLaw: Enhancing
LLM with Citations in Legal Domain. (W. Che et
al., Eds.)Findings of the Association for Computational Linguistics: ACL
2025. Anais...Vienna, Austria: Association for
Computational Linguistics, jul. b2025. Disponível em: <https://aclanthology.org/2025.findings-acl.583/>
ZHANG, S. et al. Instruction Tuning for Large Language Models: A
Survey., c2025. Disponível em: <https://arxiv.org/abs/2308.10792>
ZHANG, T. et al. BERTScore: Evaluating Text Generation with
BERT. 8th International Conference on Learning
Representations, ICLR 2020, Addis Ababa, Ethiopia, April
26-30, 2020. Anais...OpenReview.net, 2020. Disponível
em: <https://openreview.net/forum?id=SkeHuCVFDr>
ZHANG, Y. et al. 🧜Siren’s
Song in the AI Ocean: A Survey on Hallucination in Large Language
Models. Computational Linguistics, v. 51, n. 4, p.
1373–1418, dez. d2025.
ZHAO, W. X. et al. A
Survey of Large Language Models. CoRR, v.
abs/2303.18223, 2023.
ZHAO, Y. et al. One Token to Fool
LLM-as-a-Judge., 2025. Disponível em: <https://arxiv.org/abs/2507.08794>
ZHOU, C. et al. LIMA: Less
Is More for Alignment. CoRR, v. abs/2305.11206,
a2023.
ZHOU, H. et al. A survey of large language models in medicine: Progress,
application, and challenge. arXiv preprint
arXiv:2311.05112, b2023.
ZIEGLER, D. M. et al. Fine-Tuning Language Models from
Human Preferences. CoRR, v. abs/1909.08593, 2019.
ZUBOFF, S. The age of surveillance capitalism : the fight for a
human future at the new frontier of power. First edition ed.
New York: [s.n.].