Ethical and social risks of harm from language models

L Weidinger, J Mellor, M Rauh, C Griffin… - arXiv preprint arXiv …, 2021 - arxiv.org
This paper aims to help structure the risk landscape associated with large-scale Language
Models (LMs). In order to foster advances in responsible innovation, an in-depth …

Taxonomy of risks posed by language models

L Weidinger, J Uesato, M Rauh, C Griffin… - Proceedings of the …, 2022 - dl.acm.org
Responsible innovation on large-scale Language Models (LMs) requires foresight into and
in-depth understanding of the risks these models may pose. This paper develops a …

Identifying and mitigating privacy risks stemming from language models: A survey

V Smith, AS Shamsabadi, C Ashurst… - arXiv preprint arXiv …, 2023 - arxiv.org
Rapid advancements in language models (LMs) have led to their adoption across many
sectors. Alongside the potential benefits, such models present a range of risks, including …

Language generation models can cause harm: So what can we do about it? an actionable survey

S Kumar, V Balachandran, L Njoo… - arXiv preprint arXiv …, 2022 - arxiv.org
Recent advances in the capacity of large language models to generate human-like text have
resulted in their increased adoption in user-facing settings. In parallel, these improvements …

Security and privacy challenges of large language models: A survey

BC Das, MH Amini, Y Wu - arXiv preprint arXiv:2402.00888, 2024 - arxiv.org
Large Language Models (LLMs) have demonstrated extraordinary capabilities and
contributed to multiple fields, such as generating and summarizing text, language …

Guardrails for trust, safety, and ethical development and deployment of Large Language Models (LLM)

A Biswas, W Talukdar - Journal of Science & Technology, 2023 - thesciencebrigade.com
The AI era has ushered in Large Language Models (LLM) to the technological forefront,
which has been much of the talk in 2023, and is likely to remain as such for many years to …

Privacy in large language models: Attacks, defenses and future directions

H Li, Y Chen, J Luo, Y Kang, X Zhang, Q Hu… - arXiv preprint arXiv …, 2023 - arxiv.org
The advancement of large language models (LLMs) has significantly enhanced the ability to
effectively tackle various downstream NLP tasks and unify these tasks into generative …

Attacks, defenses and evaluations for llm conversation safety: A survey

Z Dong, Z Zhou, C Yang, J Shao, Y Qiao - arXiv preprint arXiv:2402.09283, 2024 - arxiv.org
Large Language Models (LLMs) are now commonplace in conversation applications.
However, their risks of misuse for generating harmful responses have raised serious societal …

Assessing language model deployment with risk cards

L Derczynski, HR Kirk, V Balachandran… - arXiv preprint arXiv …, 2023 - arxiv.org
This paper introduces RiskCards, a framework for structured assessment and
documentation of risks associated with an application of language models. As with all …

On protecting the data privacy of large language models (llms): A survey

B Yan, K Li, M Xu, Y Dong, Y Zhang, Z Ren… - arXiv preprint arXiv …, 2024 - arxiv.org
Large language models (LLMs) are complex artificial intelligence systems capable of
understanding, generating and translating human language. They learn language patterns …