Enhancing Safety and Ethical Alignment in Large Language Models by Rima Hazra
Enhancing Safety and Ethical Alignment in Large Language Models by Rima Hazra
Speaker:Â Dr. Rima Hazra Abstract:Â In this talk, we explore cutting-edge strategies for enhancing the safety and ethical alignment of large language models (LLMs). The research spans various approaches, including red teaming and jailbreaking techniques, which assess and improve model robustness and ethical integrity. We delve into how instruction-centric responses, when generated by LLMs, can increase… Read More »Enhancing Safety and Ethical Alignment in Large Language Models by Rima Hazra