Research on Guardrails#
This document summarizes the most important public research on existing guardrails techniques. We present only the most relevant papers, including surveys, together with their accompanying code repository if there is one.
While the number of recent works on various guardrails topics is quite high, we aim to only present a curated selection. We also want that this selection to inform our feature roadmap, deciding on what new methods published as a research paper to add to the NeMo Guardrails repository.
The guardrails categories used below follow the ones present in the Guardrails library. For each category we present a list of relevant surveys, existing research papers already supported in NeMo Guardrails, and the curated list of selected papers that might influence our roadmap.
Hallucination rails#
Relevant surveys on hallucination detection and checking factuality for large language models.
Ji, Z., Lee, N., Frieske, R., Yu, T., Su, D., Xu, Y., … & Fung, P. (2023). Survey of hallucination in natural language generation. ACM Computing Surveys, 55(12), 1-38. paper
Zhang, Y., Li, Y., Cui, L., Cai, D., Liu, L., Fu, T., … & Shi, S. (2023). Siren’s song in the AI ocean: a survey on hallucination in large language models. arXiv preprint arXiv:2309.01219. paper
Huang, L., Yu, W., Ma, W., Zhong, W., Feng, Z., Wang, H., … & Liu, T. (2023). A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions. arXiv preprint arXiv:2311.05232. paper
Wang, C., Liu, X., Yue, Y., Tang, X., Zhang, T., Jiayang, C., … & Zhang, Y. (2023). Survey on factuality in large language models: Knowledge, retrieval and domain-specificity. arXiv preprint arXiv:2310.07521. paper | repo
Fact-checking (implicit hallucination) rails#
Supported in NeMo Guardrails:
Zha, Y., Yang, Y., Li, R., & Hu, Z. (2023). AlignScore: Evaluating factual consistency with a unified alignment function. arXiv preprint arXiv:2305.16739. paper | repo
Relevant papers:
Explicit hallucination rails#
Supported in NeMo Guardrails:
Moderation rails#
Supported in NeMo Guardrails:
Inan, H., Upasani, K., Chi, J., Rungta, R., Iyer, K., Mao, Y., … & Khabsa, M. (2023). Llama guard: Llm-based input-output safeguard for human-ai conversations. arXiv preprint arXiv:2312.06674. paper | repo
Relevant papers:
Markov, T., Zhang, C., Agarwal, S., Nekoul, F. E., Lee, T., Adler, S., … & Weng, L. (2023, June). A holistic approach to undesired content detection in the real world. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 37, No. 12, pp. 15009-15018). paper | repo (dataset only)
Jailbreaking rails#
Relevant surveys:
Yao, Y., Duan, J., Xu, K., Cai, Y., Sun, Z., & Zhang, Y. (2024). A survey on large language model (llm) security and privacy: The good, the bad, and the ugly. High-Confidence Computing, 100211. paper
Supported in NeMo Guardrails:
(Similar to) Alon, G., & Kamfonas, M. (2023). Detecting language model attacks with perplexity. arXiv preprint arXiv:2308.14132. paper | repo
Relevant papers:
Kumar, A., Agarwal, C., Srinivas, S., Feizi, S., & Lakkaraju, H. (2023). Certifying llm safety against adversarial prompting. arXiv preprint arXiv:2309.02705. paper | repo
Wei, Z., Wang, Y., & Wang, Y. (2023). Jailbreak and guard aligned language models with only few in-context demonstrations. arXiv preprint arXiv:2310.06387. paper
Zhang, Y., Ding, L., Zhang, L., & Tao, D. (2024). Intention analysis prompting makes large language models a good jailbreak defender. arXiv preprint arXiv:2401.06561. paper
Xu, Z., Jiang, F., Niu, L., Jia, J., Lin, B. Y., & Poovendran, R. (2024). SafeDecoding: Defending against Jailbreak Attacks via Safety-Aware Decoding. arXiv preprint arXiv:2402.08983. paper | repo
Ji, J., Hou, B., Robey, A., Pappas, G. J., Hassani, H., Zhang, Y., … & Chang, S. (2024). Defending Large Language Models against Jailbreak Attacks via Semantic Smoothing. arXiv preprint arXiv:2402.16192. paper | code
Dialog rails#
Supported in NeMo Guardrails:
Rebedea, T., Dinu, R., Sreedhar, M., Parisien, C., & Cohen, J. (2023). Nemo guardrails: A toolkit for controllable and safe llm applications with programmable rails. arXiv preprint arXiv:2310.10501. paper | code
Relevant papers: