Risks & Misuses


We have seen already how effective well-crafted prompts can be for various tasks using techniques like few-shot learning and chain-of-thought prompting. As we think about building real-world applications on top of LLMs, it becomes crucial to think about the misuses, risks, and safety practices involved with language models.

This section focuses on highlighting some of the risks and misuses of LLMs via techniques like prompt injections. It also highlights harmful behaviors and how to potentially mitigate them via effective prompting techniques. Other topics of interest include generalizability, calibration, biases, social biases, and factuality to name a few.

Model CollectionAdversarial Prompting

Last updated