🔬 Research Summary by Jayanth Yetukuri, a final year Ph.D. student at UCSC, advised by Professor Yang Liu, where his research focuses on improving the trustworthiness of Machine Learning models. [Original paper … [Read more...] about Towards User-Guided Actionable Recourse
Core Principles of Responsible AI
LLM Platform Security: Applying a Systematic Evaluation Framework to OpenAI’s ChatGPT Plugins
🔬 Research Summary by Umar Iqbal, an Assistant professor at Washington University in St. Louis, researching computer security and privacy. [Original paper by Umar Iqbal (Washington University in St. Louis), … [Read more...] about LLM Platform Security: Applying a Systematic Evaluation Framework to OpenAI’s ChatGPT Plugins
Robust Distortion-free Watermarks for Language Models
🔬 Research Summary by Rohith Kuditipudi, a third year Ph.D. student at Stanford University advised by John Duchi and Percy Liang. [Original paper by Rohith Kuditipudi, John Thickstun, Tatsunori Hashimoto, and … [Read more...] about Robust Distortion-free Watermarks for Language Models
Bias Propagation in Federated Learning
🔬 Research Summary by Hongyan Chang, a sixth-year Ph.D. student at the National University of Singapore, focuses on algorithmic fairness and privacy, particularly their intersection, and is also invested in advancing … [Read more...] about Bias Propagation in Federated Learning
GenAI Against Humanity: Nefarious Applications of Generative Artificial Intelligence and Large Language Models
🔬 Research Summary by Emilio Ferrara, a professor at the Thomas Lord Department of Computer Science of the University of Southern California. [Original paper by Emilio Ferrara] Overview: This paper delves … [Read more...] about GenAI Against Humanity: Nefarious Applications of Generative Artificial Intelligence and Large Language Models