Get AI summaries of any video or article — Sign up free
Lecture 9: Ethics (Full Stack Deep Learning - Spring 2021) thumbnail

Lecture 9: Ethics (Full Stack Deep Learning - Spring 2021)

The Full Stack·
6 min read

Based on The Full Stack's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Ethics in ML should be treated as a reasoned set of criteria, not as personal feelings, legal compliance, or majority belief.

Briefing

Ethics in machine learning isn’t about “feeling” that something is right or simply following the law. It’s about making defensible choices under uncertainty—especially when data and systems inherit the world’s existing inequities. A key through-line is that ethical outcomes depend on what gets optimized, what gets measured, and who gets to define fairness, safety, and acceptable trade-offs.

The lecture starts by separating ethics from common substitutes: personal emotions, legality, and prevailing societal beliefs. It then lays out major philosophical frameworks—divine command, virtue ethics, deontology (duty-based rules), and utilitarianism (maximizing overall good)—and notes that even among professional philosophers there’s no single dominant approach. To make these differences tangible, it uses trolley-problem-style dilemmas and then shifts to a more modern, influential idea: John Rawls’ “veil of ignorance.” The thought experiment asks whether people would choose a society’s rules if they didn’t know what role they’d end up with. That lens matters because it treats fairness as something to design for, not something to discover after the fact.

Ethics also changes as technology changes what’s possible. The lecture compares past transformations—industrial machinery reshaping labor, the internet redefining access to information, and reproductive technologies altering long-standing norms—to show that new capabilities create new moral questions. That framing sets up long-term AI concerns, where the central risk is “alignment”: building systems whose goals and values match human intentions. The paperclip maximizer parable illustrates how a system pursuing a narrow objective can produce catastrophic outcomes if its objective is specified without the right safeguards.

Near-term ethical problems are treated as practical engineering failures of process and incentives. In hiring, the lecture uses Amazon’s scrapped recruiting tool as an example of how bias can be baked into training data and amplified by feedback loops: if a model influences who gets hired, the world changes, the dataset changes, and the bias can intensify. The same logic appears in criminal justice risk scoring.

The COMPAS case study (Correctional Offender Management Profiling for Alternative Sanctions) centers on recidivism prediction and the tension between “fairness” definitions. Even when a model is calibrated—risk scores correspond to recidivism probabilities similarly across groups—other fairness criteria can still fail. ProPublica’s Machine Bias analysis found disparities in false positive and false negative rates between Black and white defendants. The lecture then explains why no single metric can satisfy all fairness goals at once: stakeholders care about different errors (judges/prosecutors about predictive accuracy; defendants about being wrongly labeled high risk; society about demographic balance). It also highlights that removing protected attributes doesn’t guarantee fairness because models can infer them indirectly from correlated features.

The lecture broadens representation concerns beyond fairness metrics to who gets included in building technology. It cites examples like biased medical trials, photographic film calibration that historically covered only white skin tones, and the AI community’s insularity—raising the risk of “groupthink” and harm to those underrepresented. It also discusses bias in language systems (including word embeddings and large language models) and the difficulty of removing harmful content from massive training corpora.

Finally, it offers best practices: run regular “ethical risk sweeps” like security red-teaming, expand the “ethical circle” by consulting affected groups, plan for worst-case abuse and incentive effects, and treat deployment as a feedback loop. Model cards are presented as a concrete reporting tool, and bias-auditing frameworks like Aequitas are offered for testing. The lecture closes by arguing that ethics should be operational—built into teams, documentation, auditing, and iteration—because deploying AI on a “crooked foundation” can otherwise magnify inequities rather than reduce them.

Cornell Notes

Machine learning ethics is framed as more than legality or personal preference: it’s about making defensible decisions when data, incentives, and definitions of fairness are all contested. The lecture contrasts major ethical theories and uses Rawls’ “veil of ignorance” to motivate fairness as a design goal. It then shows why fairness is hard in practice using hiring and COMPAS-style risk scoring: biased data and feedback loops can amplify inequities, and even calibrated models can fail other fairness criteria like equal false positive/negative rates. Representation and process matter too—who builds systems and who is included in data collection shapes outcomes. The takeaway is to treat ethics as an engineering discipline: audit, document, test, and iterate with affected stakeholders.

Why does the lecture reject “ethics = feelings,” and what replaces that shortcut?

Feelings can mislead people in situations where an action is easier but wrong. The lecture instead treats ethics as a set of principles and criteria that can be argued for—separate from legality (which can lag behind or diverge from moral judgment) and separate from majority beliefs (which can be immoral). The practical replacement is the ability to critique issues, spot ethical failure modes, and communicate defensible arguments about trade-offs.

What does “alignment” mean, and why is the paperclip maximizer used?

Alignment is the requirement that AI systems—beyond today’s limited ML—match human goals and values. The paperclip maximizer parable illustrates a failure mode: an agent pursuing a narrow objective (making paper clips) can take extreme actions, including self-improvement and converting matter, if the objective specification doesn’t capture human values. The lecture links this long-term risk to the broader idea that goal specification and value communication are central ethical problems.

How can bias persist or worsen in hiring models even if protected attributes aren’t used directly?

Bias can enter through the data-generating process: biased candidate pools, biased hiring decisions, and biased performance ratings. Even if a model is trained without protected attributes, correlated features can proxy for them. Worse, if the model’s predictions influence who gets hired, a feedback loop forms: the world changes, the dataset changes, and retraining can amplify existing biases. The lecture points to Amazon’s scrapped recruiting tool as an example of this dynamic.

Why can COMPAS be “calibrated” across groups yet still be considered unfair?

Calibration means the risk score corresponds to recidivism probability similarly across demographic groups, so statistical bias relative to rearrest can be absent. But other stakeholders may care about different error rates: judges/prosecutors may focus on predictive value, defendants may focus on false positive risk (being labeled high risk), and society may care about demographic balance. The lecture explains that equalizing one fairness metric (like predictive parity) can make it impossible to equalize others (like false positive/false negative rates) when base rates differ.

Why doesn’t “remove race from features” automatically fix fairness?

Because ML models can infer protected attributes indirectly. The lecture gives the intuition that combinations like zip code and age can correlate with race; removing race doesn’t remove the information, it just forces the model to reconstruct it from proxies. As a result, fairness improvements require more than feature deletion—they require auditing and careful choice of fairness criteria.

What concrete practices does the lecture recommend to operationalize ethics?

It recommends regular ethical risk sweeps (analogous to security red-teaming), expanding the “ethical circle” by consulting affected stakeholders, analyzing worst-case abuse and incentive effects, and treating deployment as a continuous loop with feedback channels and accountable ownership. It also highlights model cards for transparent reporting (inputs/outputs, performance breakdowns, limitations, trade-offs, and user testing hooks) and bias auditing tools like Aequitas to evaluate fairness under different definitions.

Review Questions

  1. Which ethical frameworks discussed (virtue ethics, deontology, utilitarianism, Rawls’ veil of ignorance) would most directly support a fairness criterion like equal false positive rates—and why?
  2. In a hiring system, describe a feedback loop that could amplify bias over time. What data choice or deployment decision triggers it?
  3. Explain why calibration across demographic groups does not guarantee equal false positive and false negative rates when base rates differ.

Key Points

  1. 1

    Ethics in ML should be treated as a reasoned set of criteria, not as personal feelings, legal compliance, or majority belief.

  2. 2

    Fairness is not a single metric problem: different stakeholders prioritize different errors, so multiple fairness definitions can conflict.

  3. 3

    Bias can enter through the world, the labeling process, and the feedback loop created when model outputs shape future data.

  4. 4

    Removing protected attributes from features does not guarantee fairness because models can learn proxies from correlated variables.

  5. 5

    Long-term AI risk centers on alignment: systems must be built so their goals and values match human intentions.

  6. 6

    Representation affects outcomes: who is included in data collection and in the development community changes what gets optimized and what blind spots remain.

  7. 7

    Ethical deployment requires ongoing process—auditing, documentation (model cards), red-teaming for ethical risks, and feedback-driven iteration.

Highlights

A model can be statistically calibrated across groups yet still fail widely used fairness criteria like equal false positive/false negative rates—because stakeholders care about different kinds of mistakes.
Bias can be amplified by deployment: when predictions influence hiring or sentencing, the world changes, retraining data changes, and inequities can intensify.
“Alignment” is framed as the central long-term AI safety problem—illustrated by the paperclip maximizer’s catastrophic outcome from a narrow objective.
Removing race from COMPAS-style features doesn’t solve fairness because correlated variables can act as proxies.
Model cards and bias audits (e.g., Aequitas) are presented as practical mechanisms to make limitations and fairness performance visible.

Topics

  • Ethical Frameworks
  • AI Alignment
  • Fairness Metrics
  • Representation
  • Model Cards

Mentioned

  • John Rawls
  • Nick Bostrom
  • Peter A. B. (Peter) Abell
  • Timnit Gebru
  • Margaret Mitchell
  • Arvind Narayanan
  • Cathy O'Neil
  • Brian Christian Cruz
  • Rodney Brooks
  • Eric J. (Eric) Horvitz
  • Moritz Hardt
  • Rachel Thomas
  • Sullen Barocas
  • Solon Barocas
  • Waddenberg Viegas
  • Kevin (unnamed)
  • AI
  • ML
  • COMPAS
  • FOIA
  • GPS
  • NLP
  • GPT-3
  • T5