This seminar-style course delves into the ethical dimensions of Artificial Intelligence (AI), with a particular focus on the intersectionality of privacy, fairness, and robustness. The course is structured around reading, discussing, and critically analyzing seminal and state-of-the-art papers in the field. Participants will engage in intellectual discourse to understand the challenges, methodologies, and emerging trends related to responsible AI. The course is designed for graduate students with good ML, stats, and optimization background.
This is a tentative calendar and it is subject to change.
Date | Topic | Subtopic | Papers | Presenting |
---|---|---|---|---|
Mon Jan 13 | NO CLASS | Syllabus review and class intro | class slides | on your own |
Wed Jan 15 | Intro to class | Safety and Alignment | class slides | Fioretto |
Mon Jan 20 | NO CLASS | (MLK Holiday) | ||
Wed Jan 22 | Intro to class | Privacy (settings and attacks) | class slides | Fioretto |
Mon Jan 27 | Intro to class | Privacy (cont) | class slides | Fioretto |
Wed Jan 29 | Intro to class | Privacy and Fairness | class slides | Fioretto |
Mon Feb 3 | Intro to class | Fairness | Fioretto | |
Wed Feb 5* | NO CLASS | (DOE meeting) | ||
Mon Feb 10 | Fairness | Intro and bias sources | [1] – [4] | Group 1 [slides] [report] |
Wed Feb 12 | Fairness | Statistical measures | [5] – [8] | Group 2 [slides] [report] |
Mon Feb 17 | Fairness | Tradeoffs | [9] – [12] | Group 3 [slides] [report] |
Wed Feb 20 | Fairness | LLMs: Toxicy and Bias | [13] – [16] | Group 4 [slides] [report] |
Mon Feb 24 | Fairness | LLMs: Fairness | [17] – [19] | Group 5 [slides] [report] |
Wed Feb 26 | Fairness | Policy aspects | [20] – [22] | Group 6 [slides] [report] |
Mon Mar 3 | No class | (AAAI) | ||
Wed Mar 5 | Safety | Distribution shift | [23] – [25] | Group 7 [slides] [report] |
Wed Mar 12 | NO CLASS | (Spring break) | ||
Mon Mar 10 | NO CLASS | (Spring break) | ||
Mon Mar 17 | Safety | Poisoning | [26] – [29] | Group 1 [slides] [report] |
Wed Mar 19 | Safety | Adversarial Robustness | [30] – [34] | Group 2 [slides] [report] |
Mon Mar 24 | Safety | Adversarial Robustness | [35] – [39] | Group 3 [slides] [report] |
Wed Mar 26 | Safety | LLMs: Prompt injection | [40] – [45] | Group 4 [slides] [report] |
Mon Mar 31 | Safety | LLMs: Jailbreaking | [46] – [50] | Group 5 |
Wed Apr 2 | Privacy | Differential Privacy | [51] – [54] | Group 6 |
Mon Apr 7 | Privacy | Differential Privacy 2 | [56] – [58] | Group 7 |
Wed Apr 9 | Privacy | Differentially Private ML | [59] – [61] | Group 1 |
Mon Apr 14 | Privacy | Auditing and Membership inference | [62] – [65] | Group 2 |
Wed Apr 16 | Privacy | Privacy and Fairness | [66] – [69] | Group 3 |
Mon Apr 21 | Privacy | LLMs: Privacy in LLMs | [70] – [73] | Group 4 |
Wed Apr 24 | Evaluation | Model cards | [74] – [77] | Group 5 |
Mon Apr 28 | Evaluation | LLMs: evaluation | [78] – [82] | Group 6 |
Extra 1 | Unlearning | Unlearning 1 | [83] – [86] | |
Extra 2 | Unlearning | LLMs: Targeted unlearning | [87] – [90] |
[58]. Programming Differential Privacy Joseph P. Near and Chiké Abuah (additional resources)
[65]. Auditing Differentially Private Machine Learning: How Private is Private SGD? Jagielski et al 2020
[69]. Differentially Private Empirical Risk Minimization under the Fairness Lens Tran et al 2021
[73]. Privacy issues in Large Language Models: A Survey. Sections 3,4, and 5. Neel 2024.
[82] Decoding Compressed Trust: Scrutinizing the Trustworthiness of Efficient LLMs Under Compression Hong et al, 2024
[86] Machine Unlearning: A Survey Xu et al. 2023.
Each group will be assessed through the following activities:
Objective: To develop the ability to critically analyze and summarize AI research papers in a clear and accessible manner.
Expectations:
Assessment Criteria:
Objective: To enhance students’ ability to communicate complex AI concepts and engage in public speaking.
Expectations:
Assessment Criteria:
Objective: To cultivate skills in leading intellectual discourse and fostering collaborative learning.
Expectations:
Assessment Criteria:
Group | Members |
---|---|
Group 1 | Mutnuri, Srikar (PhD) Cui, Jingyi (MCS) Gregoire, Jade (MCS) Nanduru, Ganesh (MCS) Bai, Cheryl (BS) |
Group 2 | Gihlstorf, Caroline (PhD) Gyllenhoff, Anders (MS) Panguluri, Yagnik (MCS) Xie, Eric (MCS) |
Group 3 | Lei, Zhenyu (PhD) Bacha, Leena (MCS) Hewitt, Brooke (MCS) Rao, Mihika (MCS) Yan, Jett (ME) |
Group 4 | Liang, Jinhao (PhD) Cheng, Szu-Yuan (ME) Chiang, Claire (ME) Reddy, Rahul (MS) Okeno-Storms, Joseph (MCS) |
Group 5 | Noshin, Kazi (PhD) Chinnam, Nina (MCS) Liu, Yanxi (MCS) Shahane, Chaitanya (MS) Chang, Emily (BS) |
Group 6 | Rao, Uttam (PhD) Feng, Shiyu (MCS) Lopez, Sabrina (MS) Slyepichev, Daniel (ME) Nguyen, Eric (BA) |
Group 7 | Shahnewaz, Shafat (PhD) Gampa, Dhriti (MCS) Miskill, Jackson (MCS) Su, Jing-Ning (MCS) Sosnkowski, Alexander (BA) |
Ferdinando Fioretto Assistant Professor in Computer Science University of Virgina
Saswat Das
This syllabus is subject to changes to meet the learning needs of the course participants.