Want to use Sway with your students?

Back to Reports

Students Debate Algorithmic Bias: Hiring, Governance, and Recidivism

Timelines
Opinion deadline:
2025-09-29 00:00:00+00:00
Completion deadline:
2025-10-02 05:59:00+00:00
Launch deadline:
2025-09-25 19:00:00+00:00
Info
Instructor:
[Redacted]
Min. chat time:
30 minutes
Created on:
2025-09-25 18:42:50.844749+00:00
Chat threads:
38
Topics
Algorithmic Decision-Making and Bias
Algorithmic decision-making is less biased than human decision-making.

Algorithms in Governance
Algorithms should make decisions about who is elected, and which laws are passed.

Algorithms in Hiring
Algorithms should make hiring decisions.

COMPAS
The COMPAS recidivism algorithm is/was biased against Blacks.

Disparate Impact
We should have laws prohibiting disparate impact.

Hedden's Paper
Hedden's 'Perfectly Fair Algorithm' really is perfectly fair.

Pessimism
There can be no such thing as a fair or unbiased algorithm.
Written Feedback
Selected optional written feedback responses from students.
Opinion Distribution
Algorithmic Decision-Making and Bias
Algorithmic decision-making is less biased than human decision-making.
30
20
10
0
-3
-2
-1
0
1
2
3
Mean: 0.75 (95% confidence interval: 0.37 to 1.13)
Algorithms in Governance
Algorithms should make decisions about who is elected, and which laws are passed.
60
40
20
0
-3
-2
-1
0
1
2
3
Mean: -2.12 (95% confidence interval: -2.39 to -1.85)
Algorithms in Hiring
Algorithms should make hiring decisions.
30
20
10
0
-3
-2
-1
0
1
2
3
Mean: -1.25 (95% confidence interval: -1.61 to -0.89)
COMPAS
The COMPAS recidivism algorithm is/was biased against Blacks.
30
20
10
0
-3
-2
-1
0
1
2
3
Mean: 0.51 (95% confidence interval: 0.11 to 0.91)
Disparate Impact
We should have laws prohibiting disparate impact.
30
20
10
0
-3
-2
-1
0
1
2
3
Mean: 0.75 (95% confidence interval: 0.36 to 1.14)
Hedden's Paper
Hedden's 'Perfectly Fair Algorithm' really is perfectly fair.
30
20
10
0
-3
-2
-1
0
1
2
3
Mean: -0.25 (95% confidence interval: -0.58 to 0.08)
Pessimism
There can be no such thing as a fair or unbiased algorithm.
20
15
10
5
0
-3
-2
-1
0
1
2
3
Mean: 0.51 (95% confidence interval: 0.07 to 0.95)
Instructor Report

Students explored whether algorithms—especially risk-assessment tools like COMPAS—can be fair, what "bias" means when applied to automated systems, and whether laws should prohibit disparate impact. Across 38 discussions, most groups grappled with overlapping questions from Brian Hedden's work on "perfectly fair" algorithms, ProPublica's investigation of COMPAS, and broader debates about algorithmic decision-making in hiring, criminal sentencing, and governance. The conversations ranged from tightly focused technical arguments about calibration and error rates to wide-open normative disputes about whether fairness is even definable, let alone achievable.

Themes

  • The single most recurring debate was whether "bias" means biased rules or biased outcomes—and nearly every group stumbled on this distinction before making progress. Many students initially argued that COMPAS cannot be biased because race is not an explicit input, then gradually accepted that proxy variables (neighborhood, employment history, prior arrests) can encode racial disparities rooted in historical policing, housing discrimination, and poverty. This shift from intent-based to impact-based reasoning was the most common arc across threads, though the speed and depth of the shift varied considerably.

  • Students repeatedly wrestled with the mathematical impossibility of satisfying calibration and equalized error rates simultaneously when base rates differ across groups. Several groups discovered this tradeoff organically—sometimes prompted by Guide, sometimes through their own reading of Hedden—and it consistently reframed their debates. Once students understood that choosing a fairness metric is partly a moral and political decision rather than a purely technical one, conversations moved from "is COMPAS biased?" toward "what kind of unfairness are we willing to tolerate and why?" As one student put it, the question becomes whether unfairness lies "in the metric itself or in how society uses accurate-but-unequal information."

  • A strong majority of groups gravitated toward hybrid models—algorithms for screening or benchmarking, humans for final decisions—but several groups noticed an internal contradiction in that position. Students who had earlier argued that human bias is the core problem then endorsed giving humans the last word, and Guide frequently flagged this tension. One student captured the unresolved worry cleanly: "If human bias is the problem algorithms solve, why trust humans with the final call?" Groups that engaged this contradiction most directly tended to produce sharper conclusions about what each component (algorithm, human) actually contributes and where each fails.

Guide's role

  • Guide consistently forced students to engage with the strongest version of their partner's argument rather than letting them talk past each other. When students drifted toward agreement on vague principles ("it needs improvement," "we need transparency"), Guide redirected them toward concrete decision rules—who overrides whom when a judge and COMPAS disagree, what counts as "clean" data, how an ethics board would function. This pattern of pushing from slogans to mechanisms was Guide's most visible and productive tendency across threads.

  • Guide was especially effective at catching definitional drift and calling out mid-conversation position shifts. In multiple threads, Guide flagged when a student quietly moved from "algorithms can't be unbiased" to "algorithms need human oversight," or from "COMPAS is fair" to "COMPAS reflects biased reality"—then pressed students to acknowledge and justify the shift. This kept discussions honest and forced students to identify what specific argument or example had changed their thinking.

  • Guide occasionally overcorrected or provoked friction, but generally maintained rigor without derailing conversations. In one thread, a student explicitly told the group to "ignore the AI," and in a few others students expressed impatience with Guide's persistent questioning near the end of sessions. These moments were rare and did not escalate into personal attacks. More commonly, Guide's challenges—like asking whether a proposed "universal truth" algorithm could avoid embedding human judgment, or whether "corrective bias" is distinguishable from ideological bias—produced the sharpest turns in student reasoning.

Common ground

  • Nearly every group agreed that race-neutral inputs do not guarantee race-neutral outcomes, even if they arrived at that conclusion from different starting points. Students who initially defended COMPAS on the grounds that race is not an explicit variable consistently came around once proxy variables were introduced—whether through their partner's arguments, Guide's prompts, or their own research mid-conversation. The lending and hiring analogies (e.g., penalizing career gaps that correlate with maternity leave, or using zip codes that correlate with race) were especially effective at making this point concrete.

  • Most groups converged on the view that "perfect" algorithmic fairness is unattainable, though they disagreed on whether this makes algorithms useless or merely imperfect tools. Students who engaged Hedden's framework closely tended to distinguish between theoretical possibility (an algorithm can satisfy calibration in a toy model) and practical relevance (real-world data is too entangled with historical injustice for any single metric to guarantee fairness). Several groups landed on a pragmatic formulation: algorithms may still be preferable to unaided human judgment if they are transparent, auditable, and used as one input among many.

  • On disparate impact, groups that initially split between "ban it" and "don't regulate it" reliably converged on a process-based middle ground. Rather than blanket prohibition, most pairs endorsed something like mandatory review, justification requirements, and escalating consequences for repeated failures to address unjustified disparities. One group sharpened this into a clean insight: once you attach penalties to repeated refusal, you have effectively created a law—so the real question is what triggers enforcement and who bears the burden of proof.

Persistent disagreements

  • The deepest unresolved split was whether an algorithm that faithfully reflects an unjust system is itself biased or merely a mirror. Some students held firm that if COMPAS applies the same rules to everyone and achieves calibration within groups, the algorithm is not biased—the world is. Their partners argued that reliably producing racially disparate harms makes the tool biased regardless of internal consistency. Guide pushed hard on this in multiple threads, asking whether there is a meaningful difference between "the system is biased" and "the system reliably produces biased outcomes," but several groups ended without resolution.

  • Students could not agree on whether correcting for group-level disparities at the algorithm level creates new unfairness. One striking example: a student proposed reweighting COMPAS scores by race to equalize false-positive rates, then reversed course after Guide pointed out this would mean assigning different scores to two people with identical circumstances based solely on race. The tension between group-level equity and individual-level equal treatment surfaced in many threads and was never cleanly resolved—students recognized the dilemma but could not find a principle that satisfied both values simultaneously.

  • Several groups stalled on whether removing predictive but bias-laden variables (like criminal history shaped by over-policing) would reduce fairness or improve it. Students defending COMPAS argued that dropping strong predictors would reduce accuracy and potentially misclassify risk in ways that harm everyone, including the groups they aim to protect. Students challenging COMPAS countered that accuracy measured against a biased baseline is not true accuracy—it is precision at predicting an unjust system. Guide helped sharpen this framing in most threads, but neither side had a satisfying answer to the other's core objection.

Insights

  • One group proposed a creative "COMPAS+" model: keep a baseline recidivism score, then generate a second score estimating recidivism risk under improved social supports. The idea was that comparing the two scores would make visible how much of the predicted risk is attributable to lack of opportunity rather than individual propensity—potentially giving judges actionable information without simply rigging the model. This was the most original policy proposal across all threads.

  • Several students independently reframed algorithmic bias correction as a permanent maintenance problem rather than a solvable one. One student called it "an endless game of whack-a-mole"—as new data enters and social conditions shift, new biases emerge, and corrections chase them indefinitely. This framing resonated because it challenged both the optimistic view (we can debias algorithms) and the pessimistic view (bias makes algorithms unusable), replacing both with a more realistic picture of ongoing, imperfect work.

  • A student's attempt to defend COMPAS by claiming that Black defendants are flagged more often because they "statistically commit" more serious offenses prompted one of the most important clarifying exchanges. Guide pressed directly on whether this reflects actual offending rates or arrest and conviction rates—a distinction the student had not considered. The question went partially unanswered, but it exposed how easily students can treat criminal-justice data as a neutral record of behavior rather than as an artifact of enforcement decisions.

Possible misconceptions

  • Multiple students initially treated the absence of race as an explicit input as sufficient proof that an algorithm cannot be racially biased. This "race-blind means race-neutral" assumption appeared in roughly a third of threads and was consistently challenged—by partners, by Guide, or by both—but it often took sustained pressure before students fully grasped how proxy variables work. Instructors may want to address this directly, since students seemed to find it genuinely counterintuitive rather than merely convenient.

  • Several students conflated overall accuracy with fairness, treating similar accuracy rates across groups as evidence that COMPAS treats everyone equally. This overlooks the possibility that aggregate accuracy can mask very different error patterns (e.g., higher false positives for one group, higher false negatives for another). Guide corrected this in most threads where it appeared, but it suggests the distinction between accuracy, calibration, and error-rate parity may need more scaffolding before discussion.

  • At least one student claimed that differential base rates in offending straightforwardly justify differential treatment by COMPAS, without accounting for how those base rates might themselves be products of biased measurement. The difference between "commits more crime" and "is arrested and convicted more often" is arguably a critical empirical distinction that students need to understand before they can reason well about whether algorithmic predictions built on criminal-justice data are legitimate.

Lessons

  • The COMPAS case worked exceptionally well as a discussion anchor. It gave students a concrete, high-stakes example that connected technical fairness concepts (calibration, error rates, proxy variables) to real human consequences. Groups that engaged COMPAS deeply tended to produce more nuanced reasoning than groups that stayed at the level of abstract claims about algorithms and bias.

  • Hedden's "perfectly fair algorithm" framework generated productive confusion. Students who tried to evaluate Hedden's claim were forced to distinguish between theoretical possibility and practical relevance, between necessary and sufficient conditions for fairness, and between calibration and broader fairness criteria. However, several groups spent significant time debating what Hedden "really meant" without fully resolving it, which suggests the reading may benefit from a clearer framing question that directs students toward the specific trade-off argument rather than the headline claim.

  • Discussions about disparate impact law tended to converge too quickly once students realized they agreed on the middle ground (process-based review rather than blanket prohibition). These threads often ran out of substantive disagreement before running out of time, suggesting the prompt might benefit from a sharper forced-choice structure or additional edge cases that make the middle ground harder to hold.

Chat Threads
Pessimism
  • Students started with a definitional clash: Student 1 argued fairness is achievable via a fixed ruleset, while Student 2 argued bias is inevitable when algorithms sort people. Their early back-and-forth used examples like a coin flip and YouTube recommendations to separate “bias from user input” (Student 1’s view) from “bias baked into design and social context” (Student 2’s view).
  • The discussion narrowed into a pragmatic distinction between “simple/context-free” algorithms and real-world decision systems about people. Student 2 conceded that simplified algorithms can be unbiased in specific contexts (e.g., “math equation” type cases), while maintaining that algorithms applied to populations inherit bias from societal patterns and discriminatory data; Student 1 partially shifted, agreeing that complexity and human entanglement make unbiasedness much harder.
  • Guide pushed them from vague claims to operational details, repeatedly forcing Student 1 to specify mechanisms (weights, data, and definitions of “universal truth”). The pivotal moment was when Student 1 proposed an algorithm to determine guilt based on “universal truth,” and Guide pressed on what counts as evidence and how it would be weighed; Student 1 then acknowledged the need for training/comparative case data and backed away from the court example, landing on a more limited claim that unbiased algorithms may exist only where human input isn’t necessary.

Student 1, you need to be more specific. What is this universal truth that would determine guilt? Evidence? Testimony? Prior convictions? Someone has to program what the algorithm considers and how it weighs different factors. Those choices introduce human judgment - and potentially bias - right from the start. How does your universal truth avoid this problem?

Hedden's Paper
  • Students’ core dispute centered on whether Hedden’s “perfectly fair algorithm” can count as perfectly fair if it mainly works in an idealized model. Student 1 began doubtful, arguing the algorithm’s heavy assumptions and potential misalignment with real-world conditions undermine claims of “perfect” fairness, but later softened to a conditional endorsement: it can be “perfectly fair” as a theoretical framework while still falling short of lived, practical fairness.
  • Student 2 defended Hedden by treating the paper as a proof-of-concept that isolates what fairness looks like once messy contextual factors are stripped away. As the exchange progressed, Student 2 granted that real-world fairness involves real experiences, yet maintained the algorithm can still be perfectly fair “on its own terms”; Student 1 remained uneasy about generalizing from a constructed case to dismiss many other fairness metrics in general settings.
  • Guide helped keep the conversation on the central theory-vs-practice tension by pushing each student to respond directly to the other’s strongest premise and by correcting its own over-interpretation of Student 2’s earlier claim. The chat also repeatedly drifted into meta-discussion about Sway’s “Progress” indicator and how “substantive” messages are judged, including attempts to elicit prompts to extend the dialogue, which occasionally displaced deeper engagement with Hedden’s specific metric arguments.

Student 2, Student 1 just argued that fairness requires real-world application because it involves real experiences. This directly challenges your view that theoretical fairness is sufficient. If you're right that Hedden's algorithm is perfectly fair in principle, how do you address Student 1's point that fairness without real-world impact might be meaningless? This seems like the crux of your disagreement.

COMPAS
  • Students split the “COMPAS is biased against Blacks” claim into two questions: whether the model explicitly uses race versus whether it reproduces racial disparities through correlated inputs. Student 1 argued the tool isn’t inherently biased because it doesn’t include race and largely mirrors inequities embedded in policing/arrest data, while Student 2 emphasized that outcomes and disparate errors matter more than intent or variable choice.

  • The discussion shifted from “bias” to “fitness for use,” with Student 2 focusing on harm from a moderately accurate tool and Student 1 emphasizing comparators and governance. Student 2 cited reporting that COMPAS accuracy was only around 66% and argued that deploying a system that’s wrong in roughly one out of three cases can institutionalize group-level generalizations, whereas Student 1 countered that judges are also inconsistent and less auditable—so a constrained, continuously audited tool could still improve on the status quo.

  • Guide repeatedly forced clarification of claims, then steered them toward explicit success criteria and a conditional evaluation framework. It challenged vague assertions about “group averages,” pressed for sourcing of accuracy claims, and later got both students to articulate standards (calibration across races, transparency, lower error than judges, limited use as one input). One notable friction point: Student 1 grew impatient near the end (“we’re running out of time”), but the exchange remained largely focused on substance rather than personal attacks.

So you both want transparency and calibration across races. But Student 2, you said earlier COMPAS has failed and causes more harm than good. If it met the standards you and Student 1 just described — calibrated, transparent, lower error than judges — would you support using it? Or is there something deeper about algorithmic risk assessment that makes it unacceptable even if technically improved?

COMPAS
  • Students started by separating “algorithmic bias” from “biased society,” then gradually collapsed that distinction under scrutiny. Student 1 initially argued COMPAS wasn’t biased because race wasn’t an input and accuracy was “acceptable,” while Student 2 argued biased training data makes the system biased in effect. As they discussed over-policing and ProPublica’s findings on unequal error types, Student 1 shifted to saying COMPAS is biased because it amplifies existing inequities.

  • The group’s most substantive turn came when they debated what fairness standard actually matters (overall accuracy vs. error patterns vs. calibration). Student 2 drew on Hedden to suggest algorithms may not be responsible for fixing societal unfairness and initially read this as supporting COMPAS, but both students then reconsidered once “recidivism” was framed as potentially contaminated by biased enforcement. By the end, Student 1 favored removing COMPAS until substantial systemic improvements occur, while Student 2 left room for conditional use if the tool becomes transparent and not a “blind” decision aid.

  • Guide drove the conversation from abstract claims to concrete fairness tradeoffs and actionable policy questions. It challenged Student 1’s reliance on overall accuracy by highlighting race-skewed false positives/negatives, then pressed Student 2 to confront whether calibration matters if the target label (rearrest/recidivism) is itself biased. Guide also pushed beyond “fix bias first” rhetoric by asking what that would mean in practice given COMPAS is already deployed, which led to clearer proposals (removal vs. transparency and safeguards).

Student 1, your self-driving car analogy doesn't quite work. Roads aren't unfair systems in the same way criminal justice is — there's no systemic bias built into road conditions that disadvantages specific groups. The key question Hedden raises is whether algorithms should be expected to correct for systemic injustice, not whether they can operate in complex environments. Student 2, you say Hedden favors COMPAS because equal risk scores mean equal recidivism rates across races. But doesn't that ignore Student 1's earlier point about over-policing? Does calibration within groups actually solve the fairness problem if the outcome being predicted is itself contaminated by bias?

COMPAS
  • Students argued that COMPAS is biased because it produces racially unequal harms, especially higher false-positive risk labels for Black defendants. Student 2 and Student 3 leaned on disparate impact, proxy variables (e.g., neighborhood, employment), and a Kantian “treating people as means” framing; they also introduced a “self-fulfilling prophecy” feedback loop where higher scores can worsen life conditions and future recidivism risk.

  • Student 1 (assigned devil’s advocate) defended COMPAS by emphasizing that race is not an explicit input and that differing error rates can be a mathematical trade-off when base rates differ. Their position evolved into a “the data/system is biased, not the algorithm” argument, plus a skepticism that the feedback loop claim needs causal evidence beyond plausible mechanism and a comparison point (judges may be worse).

  • Guide kept the debate anchored to concrete fairness definitions and forced the group to confront the central incompatibility claim (calibration vs equalized error rates). It pushed students to clarify whether “bias” is about intent, process, or outcomes, but the conversation ended with logistics/assignment-completion chatter (scheduling, quizzes, hitting 100%) before students answered Guide’s final “what should we do if the math constraint is real?” question.

But you left Guide's question unanswered: If the math makes different error rates inevitable, what should we do? Accept unequal false positives? Abandon calibration? Not use algorithms at all? This is the heart of the fairness debate. Want to tackle it, or are you satisfied wrapping up here?

COMPAS
  • Students started with the ProPublica framing (higher false positives for Black defendants) and then drifted toward a “biased data, not biased code” explanation. Student 1 initially treated disparate outcomes as direct evidence of algorithmic bias, but after Student 2 emphasized that COMPAS doesn’t use race explicitly, Student 1 reclassified COMPAS as a “messenger” reflecting deeper social inequities and possibly bias in social-circumstance variables or criminal-history inputs.
  • Student 2 argued that lack of race as an input means the system isn’t biased, but repeatedly leaned on a shaky causal story about group differences in offending. They claimed Black defendants are flagged more because they “statistically commit” more serious offenses, then partially softened to “unless creators built it to be biased” and eventually agreed it might be better described as bias tied to social circumstances; the key distinction between actual offending vs arrest/conviction rates remained largely unresolved.
  • Guide consistently pressed them to separate input neutrality from outcome disparity and to confront proxy variables and biased upstream data. It challenged Student 1’s midstream reversal (is disparate impact “bias” or not?) and forced Student 2 to grapple with how unequal false positive rates could arise even without race as a feature, including the possibility that criminal-history data and social factors encode race-correlated signals.

You're both circling something important but haven't fully worked it out. Student 2, you didn't answer my question: Do Black people actually commit more crimes, or are they just arrested/convicted more? This matters for your argument. Student 1, you say it's biased about social circumstances instead of race — but if social circumstances correlate strongly with race due to systemic factors, does that distinction actually matter? Is using proxies for race different from using race directly?

Algorithmic Decision-Making and Bias
  • Students split on whether algorithms are less biased, with Student 1 emphasizing consistency and fixability while Student 2 stressed that bias is often imported from coders and historical data. The exchange moved from “humans are emotional” vs. “algorithms mirror humans” into a more technical back-and-forth about whether bias is easier to detect and remove in code than in people.

  • The debate sharpened around scale, auditability, and interpretability, with Student 1 arguing that logging enables systematic fairness testing and rapid updates, and Student 2 arguing that black-box models make those logs hard to turn into causal explanations or targeted fixes. Student 2 repeatedly pushed the point that knowing outcomes (e.g., under-hiring women) isn’t the same as knowing why, while Student 1 countered that counterfactual tests and emerging interpretability tools still make algorithms more measurable and correctable than human judgment.

  • Guide kept the discussion rigorous by forcing each student to engage the other’s strongest objection (scale harms, opacity, scope of comparison, and whether “fixes” are durable). Its prompts also exposed a late-stage convergence: even if corrections are possible, Student 2 suggested they may be reactive and perpetual (new biases entering new datasets), reframing “bias reduction” as an ongoing maintenance problem rather than a solvable one.

Student 2, Student 1 claims that algorithmic bias can be systematically measured and corrected at scale while human bias resists correction. But if algorithmic bias often amplifies human bias from training data, doesn't correcting it just mean we're constantly chasing after human biases that keep getting baked into new datasets? Are we really solving bias, or just playing an endless game of whack-a-mole?

Algorithmic Decision-Making and Bias
  • Students started with a theoretical-vs-current-capability split, then converged on the idea that context and stakes drive how bias shows up. Student 1 argued algorithms can be designed as neutral procedures (even using a “recipe” analogy), while Student 2 initially rejected that framing and emphasized that today’s systems inherit human limitations and untested correlations. As the chat progressed, Student 2 clarified they were mainly thinking about high-stakes domains (e.g., bail), and Student 1 partially conceded complexity makes “controversial” decisions harder.

  • The discussion surfaced two different notions of bias—statistical/social bias versus “any systematic tendency”—and that semantic mismatch briefly derailed the exchange. Student 2 suggested an algorithm that consistently returns the “right answer” is inherently biased, implying true “unbiasedness” would look like randomness; Student 1 countered that bias should mean an unjustified or incorrect distortion, not merely consistency. They returned to practical concerns when Student 2 pointed out biased training data in court contexts, while Student 1 proposed checking for biases against protected categories without fully resolving how that works when proxies are involved.

  • Guide productively steered the conversation toward concrete mechanisms (data, proxies, and indirect encoding) and repeatedly pushed students to specify definitions and real-world implementation details. Its strongest interventions forced Student 1 to operationalize “show it’s not biased” and pushed Student 2 to articulate what they meant by “right answer” in bail decisions. One notable moment of friction: Student 1 explicitly dismissed Guide’s line of questioning (“ignore the AI”), after which the conversation ended abruptly due to a platform interruption (being kicked to a quiz), leaving the key proxy-variable issue only partially addressed.

Student 1, let's get back to the core issue. You mentioned that for bail decisions, it would be sufficient to show that it doesn't have biases against race, class, etc. But how do you actually ensure this in practice? If the algorithm uses factors like zip code, employment history, or education level - couldn't these still indirectly encode racial or class bias?

COMPAS
  • Students began with a “race-blind inputs imply no racial bias” claim, then moved toward a proxy-variable account of how bias can enter anyway. Student 1 initially argued COMPAS can’t be biased if it doesn’t explicitly use race, while Student 2 countered that variables like neighborhood and poverty can function as indirect stand-ins given historical housing discrimination; Student 1 ultimately accepted that disparate outcomes can justify calling the system biased.

  • The conversation shifted from intent-based definitions of bias to outcome-based evaluation, which helped produce a quick consensus. Student 2 framed the moral assessment as largely consequentialist (COMPAS as an “implicit ethical agent” whose harms matter more than designers’ motives), and Student 1 explicitly updated their view to align with an outcomes-focused standard for bias.

  • Guide repeatedly pressured them to resolve concrete policy tradeoffs rather than stopping at agreement that “it needs improvement.” It pushed on whether the source of bias matters to those harmed, whether reducing racial disparity requires sacrificing predictive accuracy, and whether “human oversight” is coherent without specifying what happens when judge and algorithm conflict—driving the discussion from general fixes (audits, transparency, using COMPAS as one input) toward decision-rule and governance questions.

You've both converged on mutual checks and balances. But here's the problem: If COMPAS and the judge disagree, who wins? You can't have it both ways. Either COMPAS overrides human bias (giving it real power), or judges override COMPAS (making it just expensive window dressing). Student 1, when they conflict, which should prevail? And if your answer is the judge, what's COMPAS actually adding?

Algorithmic Decision-Making and Bias
  • Students moved from initial confusion about the platform to a substantive debate over whether algorithms can be less biased than humans. Student 1 argued that algorithms can be “more fair” in some domains (e.g., facial recognition with an objective target), while Student 2 pressed the idea (via Hedden) that even “fair” algorithms can still produce harmful or unequal impacts across groups.

  • A key tension emerged between “better training data + diverse developers” versus deeper limits of fairness and interpretability. Student 2 emphasized that bias can enter through unrepresentative data and subtle societal background factors, and also questioned whether we can reliably diagnose why a model is biased (especially for complex systems like LLMs). Student 1 increasingly framed the advantage as auditability—code and performance can be inspected and updated—while Student 2 countered that even with transparency, identifying the real causal source of disparate impact may be difficult.

  • Guide kept the discussion productive by repeatedly forcing students to define what they mean by bias and fairness, and by translating disagreements into sharper questions that each student had to answer. It highlighted when Student 1’s argument shifted from “algorithms can be unbiased” to “algorithms may be biased but still less biased than humans,” and pushed Student 2 to give concrete examples of equal treatment producing unequal effects. This helped Student 2 concede partial ground (diverse teams can mitigate bias) without dropping the broader skepticism about achieving robust fairness across groups.

If it's hard to tell whether an outcome comes from bias or from a fair algorithm, how would developers even know they need to fix something? What would make you confident that what looks like algorithmic fairness isn't actually hidden bias?

COMPAS
  • Students initially split on whether COMPAS itself is biased versus merely reflecting biased criminal-justice realities, then converged on a “both can be true” framing. Student 1 started by leaning on class slides and arguing accuracy parity suggests the tool isn’t inherently biased, while Student 2 emphasized ProPublica’s findings (higher false positives for Black defendants; higher false negatives for white defendants) as evidence of bias in outputs; Student 1 later researched and explicitly revised their stance toward the fairness-tradeoff view (satisfying one fairness criterion can violate another).

  • They sharpened the debate by distinguishing intent, mechanism, and impact: even if not intentionally biased, COMPAS can still produce racially disparate outcomes via data and design choices. Student 2 pushed back on the idea that “programming choices” are separate from “the algorithm,” and Student 1 ultimately agreed that biased reality and biased predictions are linked—while also raising a broader concern about a future of automated justice and whether “more accurate” AI should replace human judges.

  • Guide drove the conversation from vague claims to clearer definitions and tradeoffs, repeatedly forcing students to address objections directly and consider practical implications. It pressed Student 1 to define what it would mean for an algorithm to be “not biased,” then later challenged Student 2’s proposed fixes by asking what “better data” could mean if the underlying system is biased; Guide also prompted a pragmatic discussion about whether such tools should be abandoned or used cautiously while reforms unfold.

Student 1, can you clarify what you mean by the actual algorithm is not biased by itself? You seem to be drawing a distinction between the algorithm having bias versus the algorithm reinforcing existing disparities. But if COMPAS systematically produces different risk scores for Black and white defendants with similar criminal histories, wouldn't that be algorithmic bias rather than just a reflection of external systemic issues? Put another way: if the algorithm's outputs consistently disadvantage one racial group, does it matter whether that bias originated from the training data, the features selected, or the scoring methodology?

Hedden's Paper
  • Students treated the claim of “perfect fairness” as hinging on what counts as fairness, quickly zeroing in on calibration versus other statistical parity criteria. Student 1 argued calibration is only one among many fairness metrics and emphasized equal false-positive/false-negative rates as more morally meaningful; Student 2 initially defended calibration as Hedden’s precise standard and the only coherent, non-contradictory one when base rates differ.

  • The discussion evolved from a simple agree/disagree split into a shared “trade-off” framing that separated mathematical coherence from moral fairness. Student 2 increasingly conceded that calibration can look like statistical accuracy that may mirror biased historical patterns, while Student 1 softened to calling calibration potentially necessary but not foundational, especially when training data encode prejudice. They ended by debating when it’s acceptable to relax calibration to reduce disparate harms via balanced error rates, especially in high-stakes domains.

  • Guide consistently pushed them to specify what calibration can and cannot guarantee, and to confront the practical consequences of prioritizing one fairness target over another. It forced Student 2 to reconcile endorsing “perfect fairness” with admitting calibration may be insufficient morally, and pressed Student 1 to name a concrete baseline criterion and defend it. Guide’s strongest moves highlighted the interpretability stakes of calibration (what risk scores mean) and whether unfairness lies in the metric itself or in how society uses accurate-but-unequal information.

If you say calibration works only in an ideal world with no corrupted or biased data, consider this: even with biased historical data, if Group A truly has a higher base rate for some outcome, a calibrated algorithm may still be giving accurate information. Is the problem really with calibration itself, or with how we use that information in a world shaped by historical injustice?

Pessimism
  • Students split on whether algorithmic fairness is logically impossible or achievable under the right definition. Student 1 argued that “fair” implies freedom from bias and structural inequality, so any algorithm trained on human-generated data and contested fairness standards can’t truly be fair; Student 2 leaned on Hedden’s framing to claim fairness is possible in principle, even if not fully realized yet.

  • The discussion pivoted from “can fair algorithms exist?” to “can they be used fairly?” and then back to necessary vs. sufficient conditions. Student 2 tried to preserve the possibility of a perfectly fair algorithm by separating the algorithm’s performance from biased base rates and human implementation, while Student 1 pressed that real-world inputs and usage corrupt the process and later sharpened the critique by noting Hedden treats calibration as necessary but not sufficient—so calibration alone can’t establish perfect fairness.

  • Guide repeatedly forced conceptual clarity by flagging definitional drift and pushing each student to answer the other’s strongest objection. It highlighted when the debate slipped into a different question (existence vs. deployment) and later helped Student 1 articulate the key logical point that Hedden’s example doesn’t prove fairness if it satisfies only a necessary condition; this also nudged Student 2 to qualify their stance as provisional pending a more complete fairness criterion.

Yes, absolutely relevant! The original statement is about whether fair algorithms can exist — but if they can't be used fairly, that's a crucial challenge to the whole concept. Student 1, you've made a powerful point — even if Student 2's perfect algorithm exists, human implementation introduces bias. But does this mean the algorithm itself becomes unfair, or just that fairness gets corrupted in practice? This seems to cut to the heart of whether we're talking about theoretical vs. practical fairness.

Algorithms in Governance
  • Students quickly converged on rejecting the literal claim that algorithms should decide elections and laws, reframing the debate as “algorithmic support” vs. “algorithmic influence.” Student 2 walked back from the initial pro position to argue for algorithms as tools that organize information and detect bias, while Student 1 insisted that even “narrowing choices” covertly shapes outcomes and erodes democratic accountability.

  • The core disagreement became whether algorithms can be meaningfully transparent and accountable in a mass democracy. Student 2 argued that publishing code and publicly testing it could make assumptions more visible than human judgment, while Student 1 countered that most citizens can’t interpret code, so power shifts to technical elites and contested concepts (like bias/fairness) get “locked in” through implementation details.

  • Guide repeatedly pushed both students from abstractions (bias, fairness) toward concrete governance mechanisms (e.g., gerrymandering detection) and back to the hard boundary problem. This prompted Student 2 to propose algorithms enforcing democratically voted rules, while Student 1 sharpened the point that translating principles into code is itself a value-laden step that can subtly determine outcomes even if the “high-level” principle was voted on.

Student 1, Student 2 is now saying algorithms would only enforce principles that citizens have already voted on democratically. Does that address your concern about programmers setting the rules? Or does the problem remain even if the high-level principles come from democratic votes? Think about the gap between a voted principle like fairness and the specific code that implements it.

Hedden's Paper
  • Students started split on whether Hedden’s “Perfectly Fair Algorithm” could literally be perfectly fair, then converged on the view that “perfect” fairness is unattainable. Student 1 argued that fairness definitions conflict and training data embeds historical bias, while Student 2 initially leaned on calibration-within-groups as a key fairness condition before backing off the “perfect” wording and reframing algorithms as potentially “fairer than humans,” not flawless.

  • The discussion sharpened around tradeoffs between fairness metrics and the limits of calibration as a gold standard. Student 1 noted that even well-calibrated systems can still produce unequal error rates across groups, so meeting one metric can worsen another; Student 2 accepted this and shifted toward a pragmatic stance emphasizing algorithms as decision-support with checks rather than stand-alone arbiters.

  • Guide drove the main conceptual progress by pressing students to distinguish rhetorical “perfection” from Hedden’s technical claim, and by pushing them to test their critiques against calibration specifically. Late in the chat, students added concerns about opacity and accountability (e.g., COMPAS-style secrecy) and how incentives could corrupt fairness claims, though one student also shared off-platform contact info, which is a privacy risk and a distraction from the academic thread.

Student 2, if you agree that datasets need to be clean and bias-free, doesn't that actually support Student 1's skepticism? How can we ever guarantee truly clean data when, as Student 1 points out, all human data carries historical bias? Student 1, can you engage more specifically with Hedden's argument about calibration within groups? Does this particular fairness criterion address your concerns, or does it still fall short of perfect fairness?

COMPAS
  • Students started from a common headline claim—higher false positive rates for Black defendants—but quickly uncovered that “bias” can mean different things (algorithmic rules vs. downstream outcomes). Student 1 initially argued COMPAS was biased because equal overall accuracy still produced harsher real-world consequences via false positives, while Student 2 mostly pressed on whether fixing disparities would necessarily reduce accuracy. As the exchange progressed, the focus shifted from “COMPAS is biased” to what fairness target is even coherent when base rates differ.

  • Student 1’s stance evolved from advocating error-rate equalization (including group-specific penalties) to rejecting explicit race-based score adjustments once the equal-treatment implications were made vivid. A key turning point was realizing that disparate false positive rates do not automatically mean “identical cases” are treated differently, and that equalizing group error rates can require race-conditioned decisions that conflict with individual-level equal treatment. By the end, Student 1 landed on a nuanced position: COMPAS may yield disparate outcomes because of biased policing and structural inequality in the data, without the scoring rule itself being intrinsically biased.

  • Guide drove most of the intellectual progress by forcing precision about fairness metrics, highlighting internal contradictions, and repeatedly returning to what can/can’t be fixed at the algorithm level. It challenged Student 1’s proposed “data reweighting” as effectively race-based differential treatment, then clarified the meaning of false positive rates versus calibration and why base-rate differences matter. Guide also pushed the conversation into the “so what” territory: whether “biased outcomes” implies algorithmic wrongdoing, and whether the best description is that COMPAS reflects a biased system rather than being biased in isolation.

So you're proposing we intentionally treat identical cases differently based on race? Let's think through the implications here. If two people — one Black, one white — have the exact same criminal history, same circumstances, same risk factors... your approach would assign them different risk scores purely because of their race. How do you square that with principles of equal treatment under the law? And more practically: what happens when defendants learn the algorithm is explicitly race-adjusting their scores?

Pessimism
  • Students converged on a pragmatic view that algorithms can be made substantially fairer than human decision-making, even if “perfect” neutrality is elusive. Student 2 began with a strict claim—biased data inevitably yields biased outputs and bias-free data is impossible—while Student 1 argued that auditing, reweighting, and targeted data collection (e.g., oversampling underrepresented groups) can push systems toward fairness in practice, using scholarships/admissions as examples.
  • A key tension was whether fairness is about perfect impartiality or about aligning criteria with a stated goal (e.g., merit or impact) while managing tradeoffs. Student 1 reframed “unbiased” as goal-aligned and context-sensitive (e.g., adjusting for rural students’ fewer course offerings), while Student 2 pressed that such adjustments risk over/under-compensation and depend on contested choices about what counts as merit or utility.
  • Guide strengthened the discussion by repeatedly forcing students to define terms and reconcile shifts in their claims. Its interventions highlighted that Student 2 drifted from “no unbiased algorithm exists” toward “algorithms are fair enough but need human oversight,” and pushed both students to separate bias problems from scope/measurement limits (e.g., resume screening vs. interviews) and to assess what, if anything, remains of the original absolutist position.

Student 2, you've shifted from algorithms can't be unbiased to algorithms need human oversight. Student 1's arguing that algorithms can achieve fairness through refinement and testing. Your interview example suggests humans add value, but does that mean the algorithm itself is biased, or just incomplete? If an algorithm accurately assesses resumes but misses something interviews capture, is that a bias problem or just a limitation of scope?

Disparate Impact
  • Students started split on whether laws should prohibit disparate impact, using hiring requirements as the key test case. Student 1 initially defended “neutral” criteria like heavy-lifting standards even if they disproportionately exclude women, while Student 2 argued the point of disparate impact doctrine is to force justification and weed out arbitrary barriers (e.g., requiring a driver’s license for a desk job).

  • The discussion shifted as Student 1 moved from skepticism to conditional support once the “job-related and necessary” idea was clarified. Student 2 pushed on how vague screens like culture fit or personality can mask bias, and Student 1 ended up endorsing a relevance-checking framework, then proposed concrete enforcement tools (documented evidence for requirements and periodic audits), with Student 2 agreeing but emphasizing practical limits.

  • Guide played an active role by pressing students to name what changed their minds and to confront implementation trade-offs, especially around enforcement scope and cost. It repeatedly turned tentative agreement into more specific policy reasoning by challenging whether complaint-based or size-scaled audits would create loopholes, and by forcing students to articulate what level of scrutiny small businesses should face versus large employers.

Student 2, wouldn't regular government audits be expensive and bureaucratic? Some might argue this creates unnecessary regulatory burden on businesses. Is there a balance between preventing discrimination and avoiding over-regulation?

Disparate Impact
  • Students started by aligning on a basic definition of disparate impact as neutral-seeming rules that disproportionately harm a group, then quickly moved into whether that should be illegal. Student 1 initially favored prohibitive laws (partly framing them as deterrence and education), while Student 2 argued that hard legal definitions can be rigid, hard to measure, and inappropriate when intent is neutral—preferring transparency, impact reviews, and remediation.

  • The discussion converged toward a process-based enforcement model rather than a blanket ban, with Student 1 notably shifting positions midstream. Student 2 proposed coaching, timelines, public metrics, and escalating penalties only after notice and repeated refusal; Student 1 accepted most of this and suggested that continued harm after notice starts to look like intentional discrimination rather than “true” disparate impact.

  • Guide played a central role in turning broad intuitions into concrete policy tradeoffs (gaming, enforceability, and what counts as a “law”). Its prompts forced the pair to address how a system could be manipulated and to reconcile the idea of “regulations without laws” with the reality that penalties create legal obligations, which helped crystallize their shared endpoint: enforceable processes without automatically treating disparate outcomes as illegal per se.

Student 1, you've raised an important tension: how do you enforce regulations without legal backing? If there are penalties for repeated refusal (as Student 2 suggested), aren't those penalties themselves laws? What's the meaningful difference between laws prohibiting disparate impact and regulations with legal penalties for creating disparate impact after being warned?

Algorithmic Decision-Making and Bias
  • Students quickly agreed that both humans and algorithms carry bias, but disagreed on whether algorithms meaningfully reduce it. Student 1 emphasized that algorithms inherit programmer and historical-data bias and can look less biased mainly because their reasoning is opaque, while Student 2 argued that diverse design teams and oversight could make algorithms comparatively less biased than individual judges (even if perfect neutrality is unrealistic).
  • The discussion shifted from “which is less biased” to practical governance questions about transparency, accountability, and hybrid systems in criminal sentencing. Student 2 moved toward a blended approach where algorithms provide benchmarks to reduce unnecessarily harsh outcomes and improve efficiency, while Student 1 increasingly stressed auditability, unclear standards for when judges should defer to tools, and the risk that implementation and experimentation could worsen backlogs or create new errors.
  • Guide repeatedly sharpened the students’ claims by forcing them to define mechanisms (scrutiny, retraining, oversight) rather than relying on general intuitions about technology vs. humans. It pushed Student 1 to clarify the “masking bias” idea and challenged the assumption that algorithms lack accountability by pointing to public scrutiny of tools like COMPAS; it also pressed Student 2 on who sets the norms for “updated reasoning” and how to ethically validate systems without harming defendants. The back-and-forth culminated in a concrete proposal from Student 2 (run the algorithm in parallel “in the background” and review divergences), which Student 1 met with pointed questions about discretion, legal standards, and cost.

Student 2, you're suggesting algorithms need real-world deployment to improve, but Student 1 worries about experimenting on actual defendants. How do you balance the need for algorithmic development against the risk of getting someone's sentence wrong during that learning process? Is there an ethical way to beta test on people's lives?

Pessimism
  • Students split on whether algorithmic fairness is conceptually impossible or just hard to achieve in real life. Student 1 argued that biased training data and contested definitions of fairness make “unbiased” outputs unattainable, then sharpened this into a deeper claim that even technically “fair” calibration can entrench injustice (e.g., recidivism tools built on biased policing data).

  • The debate pivoted from data bias to whether “technical fairness” can be separated from moral and political consequences. Student 2 leaned on Hedden’s distinction between an algorithm’s internal evaluation and unfair background conditions, using low-stakes cases (coin flips) to defend the possibility of unbiased algorithms; Student 1 countered that calibration can become accuracy-at-predicting an unjust system and that designers who prioritize calibration while foreseeing unequal error burdens are morally complicit. By the end, Student 2 partially conceded that high-stakes human classification is likely biased in practice, retreating to the claim that unbiased algorithms may exist only in narrow, low-complexity or non-human domains.

  • Guide consistently forced students to address each other’s strongest objections and to clarify what kind of “possibility” was being defended. It pressed Student 2 to respond directly to the recidivism/real-world complexity critique, then helped Student 1 isolate “calibration to injustice” as the key hinge of their argument before testing whether Student 2’s separation thesis survives the “complicity” worry. Guide’s final move pushed the discussion toward a clean distinction between theoretical existence and practical relevance in the cases that matter most.

Student 2, can you clarify what you mean by theoretical and on a very small scale? Are you conceding that unbiased algorithms are impossible in practice for consequential human decisions, but defending only their theoretical possibility in trivial cases? If so, does that theoretical possibility matter if it can't apply to the contexts where algorithmic fairness actually matters?

Algorithmic Decision-Making and Bias
  • Students split on whether algorithms are inherently less biased, converging on a “depends on design + context” view. Student 2 started from the idea that machines lack personal beliefs and can be optimized for objectivity, while Student 1 argued that designers’ assumptions and “fairness” definitions (via Hedden’s “Perfectly Fair Algorithm”) inevitably shape outcomes and may privilege procedural fairness over equitable outcomes.

  • The discussion sharpened around real-world mechanisms of bias: historical data, “simple” decision rules, and proxy variables. Student 1 used COMPAS and then an automated-customs scenario to argue that systems can replicate past profiling even without intent, concluding that truly unbiased training data may be impossible; Student 2 narrowed the customs case to a non-ML, rule-checking decision tree and suggested that excluding protected traits could avoid discrimination, prompting Student 1 to push back on indirect discrimination via correlated inputs (nationality, travel patterns, socioeconomic indicators). A brief side moment involved Student 2 asking about grading/progress mechanics, and Student 1 mentioning illness, but they returned to the conceptual dispute.

  • Guide consistently improved the exchange by demanding concrete examples, clarifying what type of “algorithm” was being discussed, and forcing each student to directly address the other’s strongest objection. Its prompts moved the conversation from abstract claims about objectivity into specific failure modes (amplification of discrimination in data, bias in “simple” contexts) and then into the technical distinction between ML systems and hand-coded rules, which was central to their apparent mismatch.

Student 1, Student 2's suggesting we could simply exclude protected characteristics like race and ethnicity from the algorithm entirely. Does this solve the bias problem you're worried about, or could bias still creep in through other variables?

COMPAS
  • Students framed “bias” in COMPAS differently: one treated it as biased outcomes driven by biased policing, while the other treated it as a race-neutral tool applied consistently to “objective” inputs. Student 2 argued that arrest history, “noncompliance,” and recorded violence are themselves shaped by unequal policing, so the algorithm inherits those distortions; Student 1 countered that COMPAS doesn’t choose the inputs’ social origins and that disparities can come from differing base rates rather than discriminatory model design.

  • The discussion shifted from “does it use race?” to a deeper fight over fairness vs. predictive performance and what counts as acceptable evidence. Student 1 leaned on validation claims and the idea that removing predictive variables could reduce accuracy and create new unfairness (e.g., misclassifying risk), while Student 2 questioned the reliability of recidivism data itself and emphasized critiques that simple/transparent models can match COMPAS, making opacity harder to justify.

  • Guide kept the exchange focused by forcing each side to address the other’s strongest causal and normative claims rather than talking past them. It repeatedly reframed the debate into crisp dilemmas—whether racially neutral proxies can still be “biased,” and whether high-stakes sentencing requires transparency beyond third-party validation—prompting Student 1 to defend accountability mechanisms and Student 2 to clarify whether sacrificing accuracy is warranted or avoidable.

Student 2 is arguing that the high stakes of criminal sentencing create a special obligation for transparency — that defendants have a right to challenge the specific reasoning behind their risk scores, especially when COMPAS makes mistakes ~35% of the time. Does the potential impact on someone's freedom change whether black-box algorithms are acceptable? Or can validation studies provide sufficient accountability even in high-stakes contexts?

Pessimism
  • Students largely agreed that human choices seep into most algorithmic decisions, but they disagreed on whether that makes fairness impossible in principle. Student 1 argued that because humans select, label, and frame inputs, bias is inevitable and will show up in outcomes; Student 2 started from “slightly disagree,” claiming some algorithms can be fair when they rely on objective measurements rather than socially loaded data.

  • The discussion sharpened around “accuracy vs fairness” and the COMPAS example, exposing a key split about what fairness should mean. Student 1 emphasized disparate impact (an algorithm can be accurate yet still unfair if it burdens certain groups), while Student 2 emphasized uniform procedure and individualized assessment, even if the world producing the data is unequal; this evolved into an explicit “fair but not just” distinction.

  • Guide consistently redirected the conversation from abstractions to definitions and real-world implications, pushing students to confront hidden assumptions. It challenged Student 2’s weather example as mostly about accuracy, then pressed both students on whether equal rules in an unequal society count as fairness, culminating in a debate over equal opportunity vs outcome-based adjustments and what that implies for the possibility of unbiased algorithms.

Student 2, that's a crucial distinction you're making — fair but not just. But think about what you're saying: if an algorithm applies the same rules to everyone but systematically disadvantages certain groups because of past injustice, is it really fair? Or are you just describing equal treatment in an unequal world? Can fairness exist independently from justice?

COMPAS
  • Students converged on the idea that historical policing patterns can distort the “objective” signal COMPAS learns, but they split on whether that distortion counts as bias. Student 1 framed the algorithm as inheriting systemic over-policing and thus reinforcing inequality, while Student 2 treated “bias” as unjustified prejudice and argued that observed patterns—however produced—can still be legitimately predictive for public safety purposes.
  • The debate shifted from definitions to institutional role: should courts prioritize neutrality in procedure or fairness in outcomes when tools produce racially skewed results? Student 2 leaned on a “courtroom neutrality” model (similar standards for all; fix inequality elsewhere) and worried that adjusting scores by race would introduce disparate treatment, while Student 1 argued that using a tool built on skewed baselines is not truly neutral given the high-stakes consequences of sentencing and supervision.
  • Guide’s most productive move was pushing them past “equity vs neutrality” slogans and into the empirical question of calibration/accuracy under biased base rates and opaque measurement. This pressure led Student 2 to retreat to uncertainty (proprietary definitions of recidivism; unknown data cleanliness) as a reason not to change the system, while Student 1 flipped that uncertainty into an argument against relying on COMPAS at all—ending with Student 2 beginning (but not completing) a response about needing to trust COMPAS “to some degree.”

Student 1, Student 2 is arguing that adjusting COMPAS to help Black defendants would create a new unfairness against other groups. His logic seems to be: biased data + neutral application = unfair but consistent, while biased data + adjustment = unfair and inconsistent. But does his argument actually hold? If the data systematically over-predicts recidivism for Black defendants due to over-policing, wouldn't correcting that make the algorithm more accurate for everyone, not less fair to others?

Disparate Impact
  • Students converged quickly on the idea that blanket laws against disparate impact are too rigid and should be replaced (or supplemented) by case-by-case review. Student 1 argued that “neutral” processes can still yield uneven outcomes without being unfair, while Student 2 started pro-law but moved toward a model where laws would mandate ethics-board review rather than outright prohibition.
  • The discussion became more concrete when Student 1 introduced employment testing and lending as examples where neutral, risk-based criteria can generate disparities. Student 2 repeatedly framed ethics boards as a way to assess validity, intent, and alternatives without forcing binary legal outcomes, and ultimately endorsed allowing “justified” disparate impact when it can’t be avoided.
  • Guide repeatedly pressed them to clarify thresholds, contexts, and the unresolved normative question of when disparity should be illegal versus permissible. When Student 2’s stance shifted, Guide called out the pivot and pushed for context distinctions (research vs. hiring/lending), which helped surface that both students ultimately reject prohibition even outside research; the conversation also showed some frustration from Student 1 (e.g., calling the AI “annoying”), but no sustained personal attacks between students.

You both started with opposing views and now you're agreeing — but I'm not sure you've actually resolved the core tension. The original statement was about laws prohibiting disparate impact. Now you're both agreeing on mandatory ethics review — but that's not the same as prohibition. Here's what's unresolved: If an ethics board reviews a policy and says yes, this will have disparate impact, but it's justified — should that policy be legal or illegal? Student 2, under your view, can justified disparate impact ever be allowed? Or must it always be prevented/corrected?

Pessimism
  • Students began with a technical carve-out—arguing an algorithm can be unbiased in narrow, non-ethical “pass/fail” domains—but quickly confronted how values enter upstream through definitions and thresholds. Student 1 initially pointed to factory robots, then shifted after Student 2 highlighted that even “objective” measures (e.g., seatbelt stress tests) encode human choices and can reproduce historical design assumptions (like testing around an average male body).

  • The conversation evolved from “is bias present?” to “when is bias acceptable or even useful?”, with both students converging on the view that strict unbiasedness is impossible. Student 2 floated “corrective” bias (bias that counteracts real-world imbalances), while Student 1 reframed bias as sometimes appropriate for user-selected goals (e.g., vegetarian or high-protein meal plans), ultimately agreeing that “objective fairness” is unattainable given human subjectivity and value-laden data.

  • Guide played a decisive role by repeatedly tightening definitions and forcing the students to reconcile apparent contradictions (technical neutrality vs fairness, and corrective bias vs imposed values). It pushed Student 2 to produce a clearer example of corrective bias, then pressed both students on whether they were now effectively agreeing with the original claim and on the unresolved governance question: who gets to label a bias “corrective” rather than ideological or discriminatory.

But here's the challenge: Who decides which biases are corrective and which are just imposing someone's values? For example, affirmative action algorithms correct historical discrimination — but critics argue they create new unfairness. A vegetarian meal plan serves user choice — but what about an algorithm that prioritizes certain news sources or job candidates? Student 1, how do we distinguish between justified bias and problematic bias?

COMPAS
  • Students zeroed in on whether “bias” should mean biased code/inputs or biased outcomes, and they talked past each other a bit because of that framing gap. Student 1 argued COMPAS is not biased because it does not explicitly use race and is largely reflecting base-rate differences captured in the training data; Student 2 countered that race-blind design can still yield racially disparate false positives via proxy variables and historically skewed data.

  • The conversation sharpened around an accuracy-versus-fairness tradeoff, with Student 2 pushing the normative stakes more directly. Student 1 emphasized that removing race-correlated predictors could reduce predictive performance and that algorithms may still be preferable to potentially biased judges, while Student 2 argued that legitimacy and harm in the justice system make disparate impact ethically unacceptable even if predictive accuracy is high.

  • Guide played an assertive role by challenging Student 1’s narrow definition of bias and repeatedly forcing engagement with disparate impact and the provenance of “reality” in the data. Guide used analogies (e.g., hiring algorithms) and direct questions to press whether statistical accuracy in an unjust system merely automates injustice; Student 1 ultimately went silent, leaving Student 2’s argument unaddressed and the discussion unresolved on the key objection.

If COMPAS systematically produces higher false positive rates for Black defendants, doesn't that make it a biased tool — regardless of whether race is an explicit input? Think about it this way: If a hiring algorithm never looks at gender but consistently rejects qualified women because it penalizes career gaps (which correlate with maternity leave), is that algorithm unbiased just because gender isn't in the code? Your argument seems to be: COMPAS reflects reality, not bias. But what if the reality it reflects is itself the product of biased policing, biased arrests, and biased social conditions?

COMPAS
  • Students’ core disagreement centered on what it even means for COMPAS to be biased: biased rules vs biased outcomes. Student 1 argued that since race is not an input and the same rules apply to everyone, COMPAS is not “itself” biased and is mainly reflecting inequities embedded in criminal-history data; Student 2 pushed that consistent rules can still yield biased results (e.g., higher false positives for one group), so the system is biased in effect even if not intentionally.

  • The discussion evolved from “race isn’t used so it can’t be racist” to a more nuanced split between intent, inputs, and impact. Student 2 conceded that common variables (prior arrests/convictions, age of first arrest) can be “fair” inputs, but emphasized base-rate differences, reinforcement of existing injustices, and concern about non-public features; Student 1 resisted assumptions about hidden inputs and leaned on equal application and similar accuracy across groups as evidence of fairness.

  • Guide consistently pressured both students to confront the practical stakes of their definitions and translate them into policy implications. It repeatedly asked what difference there is between “reflecting” societal bias and “being” biased, and used a sentencing-use question to move the debate from semantics to consequences; the students ended up converging that COMPAS could still be used, but only as one tool, because human judges may be even more biased and inconsistent.

Student 2, Student 1 is drawing a sharp line between the system's design and its outputs. He's saying: biased data → unbiased system → biased results = system isn't biased. You seem to be saying: if the system systematically produces biased outcomes, the system itself is biased. Is there a meaningful difference between the system is biased and the system reliably produces biased outcomes? What matters more for the people affected by COMPAS scores?

COMPAS
  • Students framed the COMPAS bias question as a clash between “equal treatment” and “equitable outcomes.” Student 1 argued COMPAS is “inherently fair” because it does not use race and applies the same inputs to everyone, while Student 2 countered that “race-neutral” variables can still reproduce racial disparities because they are shaped by historical and structural inequities.

  • The conversation evolved from whether COMPAS is racist to whether courts should try to correct for inequity at all—and whether “more data” makes prediction fairer. Student 2 pressed a possible circularity (assuming fairness to justify including correlated variables) and proposed a “COMPAS+” approach: keep a baseline score but add an additional analysis estimating recidivism risk under improved social supports, claiming this could reduce inequity without simply “rigging” the model. Student 1 maintained that equity-oriented adjustments don’t belong in parole decisions, but softened when confronted with legal parallels (race-neutral lending rules) and explicitly conceded a key analogy after learning about disparate-impact liability.

  • Guide repeatedly forced students to confront internal inconsistencies and directly answer each other’s strongest objections. It highlighted how proxy variables (zip code, income, education) can act as conduits for historic discrimination, and later used Student 1’s own mortgage analogy to show how “neutral inputs” can still be legally and ethically problematic. Guide also kept the exchange substantive by summarizing Student 2’s points for Student 1 and then demanding a direct response to the “double standard” between what humans vs. algorithms may consider in high-stakes decisions.

Student 1, you've made an analogy to bank lending, but Student 2 might push back here: banks using credit scores that reflect historical discrimination have been found liable for discriminatory lending practices. Courts have ruled this violates fair lending laws. If discriminatory outcomes in lending are legally problematic even when using race-neutral factors, why should COMPAS be different? Your mortgage analogy might actually support Student 2's position rather than yours.

COMPAS
  • Students started from near-zero knowledge of COMPAS and quickly zeroed in on proxy variables as the main mechanism for racial disparity. Student 1 initially asked how an algorithm could “use skin color,” then accepted that inputs like neighborhood, employment, and criminal history can indirectly reproduce racial patterns tied to segregation, housing policy, and concentrated poverty.

  • They wrestled with the fairness dilemma: strong evidence of disparate outcomes versus disagreement about what “fair” even means. Student 2 argued ProPublica’s false-positive disparities were concerning enough to demand action, while also stressing that competing fairness metrics can be simultaneously rigorous yet conflict—making the core dispute partly moral/political rather than purely technical.

  • Guide repeatedly pushed them from broad reactions toward concrete governance options, while the students periodically signaled frustration and a desire to “be done.” The discussion cycled through proposals (government investigation/oversight vs. independent academic/nonprofit audits) and then back to skepticism about bias in all institutions; Guide kept forcing them to confront the implications of that skepticism rather than letting the conversation end in vague calls for “transparency.”

Student 1, Student 2 just said choosing a fairness definition is ultimately a moral and political choice, not just a technical one. But you argued earlier that scrutiny can lead to right and just endings. Can there be objectively right answers to moral and political questions, or is Student 2 correct that it's fundamentally a choice? This gets at whether COMPAS bias is a fact to discover or a value judgment to make.

Disparate Impact
  • Students began by aligning on a definition of disparate impact and quickly moved from slogans about fairness to concrete criteria for judging policies. Student 2 initially argued that justice requires focusing on outcomes (using COMPAS as the core example), while Student 1 pushed back that “equal results” can’t be the only standard and introduced safety-based cases (e.g., driver’s license requirements affecting disabled people).

  • The discussion evolved into a shared framework: disparate impacts can be acceptable when tightly connected to a legitimate purpose, and problematic when they reflect bias or failure to meet the policy’s aims. Student 1 tested edge cases (true positive rate differences vs. false positives/negatives) and emphasized intent/purpose; Student 2 conceded outcome differences aren’t automatically unjust but argued laws are still needed to force justification and avoid “neutral intent” being used as a shield.

  • Guide’s main contribution was pressure-testing the “ban disparate impact” claim by surfacing tradeoffs (burden of proof, contested justifications) and pushing students to clarify the law’s scope. This helped prompt Student 1’s key reframing—that the original prompt sounded like a prohibition on all disparate impact—and led both students to converge on supporting laws targeting unjustified disparate impact rather than any disparate impact whatsoever.**

Algorithms in Hiring
  • Students started polarized on whether algorithms should make hiring decisions, then converged on a hybrid model. Student 1 initially argued algorithms reduce unconscious human bias and can handle high application volume fairly, while Student 2 stressed the need for human judgment, values, and nuance; by the end, Student 1 shifted to endorsing human involvement for context and authentication, with algorithms handling scale and early screening.

  • A key tension was “bias vs. accountability,” with students disagreeing on which system is easier to correct and trust. Student 1 claimed algorithmic bias is more detectable and fixable via audits and transparency, whereas Student 2 countered that auditing depends on company incentives and that humans can be questioned and held accountable in ways algorithmic decisions often are not. The discussion also broadened into practical risks like applicants exaggerating credentials and the role of interviews in verifying claims.

  • Guide drove the most productive turns by pushing students to confront contradictions and directly answer each other’s strongest objections. It pressed Student 1 on how “consistent” algorithms might simply reproduce historical discrimination, then challenged Student 2 to explain why human judgment wouldn’t reintroduce bias after algorithmic screening. Near the end, Guide highlighted the group’s emerging inconsistency—endorsing human final decisions despite earlier concern about human bias—prompting Student 1 to articulate a more careful “shared flaws, balanced strengths” conclusion.

You both now agree humans should make final decisions — but doesn't that contradict Student 1's original concern about unconscious bias? If human bias is the problem algorithms solve, why trust humans with the final call? Also worth exploring: What specific qualities or contexts make human judgment irreplaceable? Student 2 mentioned qualitative data and authentication — are these things algorithms could never do well, or just things they can't do yet?

Hedden's Paper
  • Students split “perfect fairness” into two competing definitions: procedural consistency vs. fairness of the overall system (including inputs). Student 1 initially argued the algorithm is fair in Hedden’s toy setup because it uses only “relevant” information (coin bias) and applies the same rule to everyone, while conceding real-world deployment would be harder. Student 2 pushed that equal rules don’t guarantee fairness when individuals start with different biases, so the system can remain unfair even if the procedure is impartial.

  • The discussion’s main turning point was whether an algorithm can be called “perfectly fair” if it faithfully propagates unfair or luck-based starting conditions. Student 1 gradually acknowledged that if you evaluate the whole system (where coin biases may reflect unfair origins), “perfect” fairness becomes questionable or impossible—yet still defended fairness “within the algorithm” as consistent treatment given the data. Student 2 sharpened the critique by emphasizing that luck-based inputs (coin bias) make outcomes unfair even if the rule is uniform, and ended by framing perfect fairness as unrealistic for most algorithms because inputs are typically environmentally influenced.

  • Guide drove the conversation by repeatedly challenging each student’s fairness criterion and forcing them to confront implications of their view. It pressed Student 1 on why coin bias counts as permissible “relevant” information while room membership does not, and then pressed Student 2 on whether rejecting environmentally-shaped inputs would make prediction impossible. These interventions helped both students articulate more nuanced positions (Student 1 distinguishing procedural vs. systemic fairness; Student 2 allowing some past-influenced information but rejecting “perfect” fairness claims).

If an algorithm uses unfair inputs, can its outputs really be perfectly fair just because it applied consistent rules? You're essentially saying the algorithm is fair given the biased data it receives. But doesn't perfectly fairness require evaluating the whole system — including whether using that biased data was the right choice in the first place?

Algorithms in Hiring
  • Students split on whether algorithms should decide hires, with Student 1 arguing for “algorithm-assisted, human-final” hiring and Student 2 defending eventual algorithmic replacement. Student 1 emphasized that models inherit historical and structural bias, overvalue proxies (keywords, GPA, years), and miss context-dependent traits like judgment and growth; Student 2 countered with debiasing via preprocessing/oversampling and using essays to quantify “soft skills,” then later reframed the claim as “not today, but in a few decades.”

  • The discussion surfaced a sharp philosophical divide about what human qualities are and whether empathy is a feature or a bug in high-stakes decisions. Student 2 argued that traits like empathy and motivation are ultimately reducible to neurochemistry and could be programmed, and even suggested a “heartless” algorithm might outperform humans by firing an error-prone doctor immediately—an unexpectedly rigid stance that Student 1 pushed back on by framing empathy as part of long-run safety and fairness (support, recovery, context). Student 1 also challenged Student 2’s proposed safeguards (essays, reference letters) as new channels for gaming and social-network bias rather than true measures of merit.

  • Guide’s strongest contributions were targeted challenges that forced Student 2 to confront the limits of “debiasing” and the claim that AI is inherently neutral. It pressed on whether dataset balancing can entrench stereotypes when opportunity itself has been unequal, and it highlighted Student 2’s partial concession (algorithms not ready now) to pivot the debate toward deeper assumptions about where bias comes from and whether automation can remove it rather than repackage it.

Student 1, Student 2 just made an interesting concession — he's agreeing that algorithms aren't ready today but should eventually replace human hiring a few decades down the road. But he's also making a broader claim: that humans are fundamentally to blame for bias, while AI is inherently neutral. Do you think this framing is accurate? Is bias really just a human problem that AI can solve, or might there be something deeper at stake about the nature of human judgment itself?

Disparate Impact
  • Students started with a tentative “yes” to banning disparate impact, but quickly converged on a more conditional view that distinguishes harmful neutral policies from equity programs. Student 2 initially opposed the idea due to a misread (thinking it would outlaw affirmative action), then reversed after clarifying the definition and argued for protections against unfair workplace rules while preserving targeted equity measures.

  • Their main sticking point became implementation: how to measure impact and who decides when disparities are justified or “too far.” They floated statistical benchmarks (e.g., comparing hiring patterns to relevant baselines), then worried about biased statistics and fell back on human judgment—prompting Student 1 to briefly “give up” before accepting that imperfect governance may still beat the status quo.

  • Guide drove the conversation by repeatedly forcing them to specify decision rules, burden-of-proof timing (pre-implementation vs. post-complaint), and real-world boundaries around “professionalism” as a potentially biased rationale. It also provided concrete Canadian examples (e.g., clean-shaven rules, “Canadian experience,” mandatory minimums) and pressed them to test their principles against cases where safety might be real but appearance norms can mask discrimination, culminating in a sharp correction of a flawed analogy.

That's a false equivalence. Not affording a shirt is about economic access to a universal requirement. Religious beards are about requiring someone to violate their identity for a job. A better comparison: Should Sikh men have to remove turbans because a company decides turbans aren't professional? Do you think religious/cultural practices deserve different treatment than personal choices? And if so, why?

Algorithms in Governance
  • Students treated full algorithmic governance as a high-stakes tradeoff between reducing corruption and creating new, harder-to-fix failure modes. Student 1 began from a “capture and abuse” frame (analogizing to gerrymandering) and focused on accountability, while Student 2 argued that bias and corruption already exist and that sufficiently “neutral” systems could improve fairness even if perfection is impossible.
  • The discussion sharpened around two concrete fault lines: minority rights under utilitarian optimization and the legitimacy of who builds/updates the system. Student 1 introduced a pointed scenario where majority welfare gains could justify suppressing minority religion/culture, and argued that endless edge cases plus “infallible machine” rhetoric could make harm both easier to excuse and harder to reverse. Student 2 conceded strict utilitarianism is limited, reframed the status quo as already failing minorities and being captured by wealthy elites, and floated continuous democratic updating—leading Student 1 to soften toward using algorithms as tools “one day,” but not as a near-term replacement for human governance.
  • Guide consistently improved rigor by forcing definitions and making students answer each other’s strongest objections rather than staying abstract. It pressed Student 1 to specify what “neutrality” would even mean, then compelled Student 2 to directly address the minority-rights objection instead of sidestepping it, and later tested whether Student 1’s accountability critique applies equally to constitutions and amendment processes. That sequencing helped move the exchange from vague optimism/pessimism to institutional details (capture, updates, accountability, and value conflicts).

Student 1, you've argued that algorithmic systems would lack accountability and prevent moral progress. But how is that different from constitutional systems? Constitutions constrain present-day decision-making based on past values, yet we update them through amendment processes. Student 2's suggesting something similar for algorithms — periodic democratic updates. Your concern seems to be about who designs and updates these systems. But isn't that also true for our current laws and constitutions? What makes algorithmic governance uniquely vulnerable to capture by elites?

COMPAS
  • Students dug into whether COMPAS bias should be judged by statistical parity (error rates) or by broader social consequences tied to race-correlated inputs. Student 1 leaned on ProPublica’s disparate false-positive/false-negative rates and the idea that “race-blind” features (e.g., neighborhood, prior contact with the system) can still encode structural inequality, while Student 2 emphasized that calibration within groups can hold even when error rates differ.

  • A key turning point was Student 2’s insistence on the fairness tradeoff: with different base rates, you cannot generally satisfy both calibration and equalized error rates at the same time. Student 1 accepted the math but argued that this is exactly why calibration alone is an incomplete fairness target if it yields systematically more harmful errors for Black defendants; they proposed reframing the goal from predicting reoffense to identifying who needs support to avoid reoffending.

  • Guide kept the discussion from stalling in slogans by repeatedly forcing a direct response to the other student’s strongest point and sharpening definitions of bias. Its prompts pushed the pair to separate “algorithmic performance criteria” (calibration, accuracy) from “justice-relevant impact,” and Student 2 ultimately conceded a meaningful distinction between “technically unbiased” and “practically biased” systems when proxies effectively penalize people for context (neighborhood, income) rather than individual risk.

Student 2, Student 1 seems to be wrapping up, but there's still a key disagreement here. You've argued the algorithm itself isn't biased if it meets calibration standards. She's arguing that using racially correlated proxies makes it biased regardless of calibration. Do you think there's a meaningful distinction between technically unbiased and practically biased algorithms? Or is calibration sufficient to call something fair?

Post-Chat Survey Data

72

Total Survey Responses

37

Threads With Surveys

97.4%

Response Rate

Pre/Post Opinion Change by Topic
Shows opinion distributions before and after discussion for students who provided both pre-chat and post-chat responses. Only topics with at least 5 matched responses are shown.
COMPAS
The COMPAS recidivism algorithm is/was biased against Blacks.
Strongly
agree
Moderately
agree
Slightly
agree
No idea
Slightly
disagree
Moderately
disagree
Strongly
disagree
3
2
1
0
-1
-2
-3
Pre-chat
Post-chat
Wilcoxon signed-rank: W = 31, p = 0.058
Hodges-Lehmann Δ = 1.50 (95% CI: 0.00 to 2.50)
Algorithmic Decision-Making and Bias
Algorithmic decision-making is less biased than human decision-making.
Strongly
agree
Moderately
agree
Slightly
agree
No idea
Slightly
disagree
Moderately
disagree
Strongly
disagree
3
2
1
0
-1
-2
-3
Pre-chat
Post-chat
Wilcoxon signed-rank: W = 6, p = 0.812
Hodges-Lehmann Δ = 0.00 (95% CI: -1.00 to 1.00)
Disparate Impact
We should have laws prohibiting disparate impact.
Strongly
agree
Moderately
agree
Slightly
agree
No idea
Slightly
disagree
Moderately
disagree
Strongly
disagree
3
2
1
0
-1
-2
-3
Pre-chat
Post-chat
Wilcoxon signed-rank: W = 12, p = 0.812
Hodges-Lehmann Δ = 0.50 (95% CI: -3.50 to 4.50)
Pessimism
There can be no such thing as a fair or unbiased algorithm.
Strongly
agree
Moderately
agree
Slightly
agree
No idea
Slightly
disagree
Moderately
disagree
Strongly
disagree
3
2
1
0
-1
-2
-3
Pre-chat
Post-chat
Wilcoxon signed-rank: W = 10, p = 0.312
Hodges-Lehmann Δ = 1.00 (95% CI: -0.50 to 1.50)
Hedden's Paper
Hedden's 'Perfectly Fair Algorithm' really is perfectly fair.
Strongly
agree
Moderately
agree
Slightly
agree
No idea
Slightly
disagree
Moderately
disagree
Strongly
disagree
3
2
1
0
-1
-2
-3
Pre-chat
Post-chat
Wilcoxon signed-rank: W = 3, p = 0.625
Hodges-Lehmann Δ = -1.00 (95% CI: -3.00 to 2.00)
Survey Response Distributions
Scale: –2 = Strongly disagree, 0 = Neutral, +2 = Strongly agree. Post-chat surveys sample a subset of the total survey items, so response counts vary across some items.
How was your chat?
🔥 Awesome 21 (29%)
👍 Good 32 (44%)
😐 It's OK 15 (21%)
👎 Not a fan 2 (3%)
💩 Hated it 2 (3%)
mean = 0.94 (95% confidence interval: 0.73–1.16)
Guide's contributions improved the discussion
Strongly agree 21 (29%)
Agree 34 (47%)
Neutral 8 (11%)
Disagree 4 (6%)
Strongly disagree 5 (7%)
mean = 0.86 (95% confidence interval: 0.60–1.12)
I felt comfortable sharing my honest opinions with my partner
Strongly agree 19 (50%)
Agree 17 (45%)
Neutral 1 (3%)
Disagree 0 (0%)
Strongly disagree 1 (3%)
mean = 1.39 (95% confidence interval: 1.14–1.65)
It was valuable to chat with a student who did NOT share my perspective
Strongly agree 12 (33%)
Agree 17 (47%)
Neutral 5 (14%)
Disagree 1 (3%)
Strongly disagree 1 (3%)
mean = 1.06 (95% confidence interval: 0.74–1.37)
I was not offended by my partner's perspective
Strongly agree 25 (81%)
Agree 3 (10%)
Neutral 1 (3%)
Disagree 1 (3%)
Strongly disagree 1 (3%)
mean = 1.61 (95% confidence interval: 1.26–1.96)
My partner was respectful
Strongly agree 30 (77%)
Agree 8 (21%)
Neutral 1 (3%)
Disagree 0 (0%)
Strongly disagree 0 (0%)
mean = 1.74 (95% confidence interval: 1.58–1.91)
My partner was genuinely trying to understand my perspective
Strongly agree 20 (62%)
Agree 8 (25%)
Neutral 2 (6%)
Disagree 2 (6%)
Strongly disagree 0 (0%)
mean = 1.44 (95% confidence interval: 1.12–1.75)
My partner had better reasons for their views than I expected
Strongly agree 11 (29%)
Agree 15 (39%)
Neutral 9 (24%)
Disagree 3 (8%)
Strongly disagree 0 (0%)
mean = 0.89 (95% confidence interval: 0.59–1.20)
Guide treated me and my partner with equal respect
Strongly agree 17 (39%)
Agree 20 (45%)
Neutral 5 (11%)
Disagree 0 (0%)
Strongly disagree 2 (5%)
mean = 1.14 (95% confidence interval: 0.85–1.43)
It would be good if more students and classes used Sway
Strongly agree 5 (15%)
Agree 7 (21%)
Neutral 13 (38%)
Disagree 7 (21%)
Strongly disagree 2 (6%)
mean = 0.18 (95% confidence interval: -0.21–0.57)
Guide contributed the right amount
Agree 47 (65%)
Neutral 19 (26%)
Disagree 6 (8%)
mean = 0.57 (95% confidence interval: 0.42–0.72)
Sway helped me articulate my thoughts/feelings better
Strongly agree 7 (21%)
Agree 15 (45%)
Neutral 7 (21%)
Disagree 3 (9%)
Strongly disagree 1 (3%)
mean = 0.73 (95% confidence interval: 0.37–1.08)