Building Trustworthy Products: Ethical AI in Product Management
AI features that harm users aren't just an engineering problem — they are a product failure. Here is the PM's playbook for building AI that users can actually trust.
The Product Manager's Responsibility in the AI Age
When Amazon's internal AI recruiting tool started systematically downgrading resumes from women, that wasn't just an ML failure. It was a product failure. Someone decided to build that system, someone signed off on deploying it, and no one caught the bias before real candidates were harmed.
That someone — in the role of defining what the system should do, what data it should use, and what "good" looks like — was, functionally, a product manager.
Ethical AI isn't a checkbox. It's a product competency.
What the Regulations Actually Require
The EU AI Act (effective 2024–2026) creates binding obligations for AI systems based on risk level:
Unacceptable risk (banned): Social scoring systems, real-time biometric surveillance, systems that exploit psychological vulnerabilities.
High risk (strict requirements): AI in hiring, credit scoring, healthcare, law enforcement, education. These require mandatory conformity assessments, bias audits, and human oversight mechanisms.
Limited/minimal risk (transparency obligations): Chatbots must disclose they're AI. Deepfakes must be labelled.
If you're building any of these systems — or integrating AI into workflows that touch these areas — you're operating in regulated territory. The PM owns that compliance scope, not the legal team.
The Four Ethical Pillars Every AI PM Must Know
Fairness: Does your AI treat all user groups equitably? Bias can enter through training data, feature selection, or proxy variables. A credit model that uses zip code as a feature can be racially discriminatory even without ever seeing race.
Transparency: Can users understand why the AI made a decision? Explainability isn't just a technical property — it's a UX requirement. "We think you might like this" is transparent. "A black-box model scored you 3.2" is not.
Privacy: What data is the model trained on? Is it used with consent? Can a user opt out? GDPR and its equivalents give users rights that your product must accommodate.
Accountability: When the AI is wrong — and it will be wrong — who is responsible? Who investigates? Who gets notified? These aren't engineering questions. They're product design questions.
The PM's Ethical AI Playbook
Before You Build
- Map harm vectors: who could be negatively affected by this system, and how?
- Define fairness metrics upfront — not after you see results
- Establish data contracts: what data can be used, collected, stored, and for how long?
- Include ethics review in your definition of "done" for AI features
While You Build
- Require diversity in training data — check for underrepresented groups
- Add explainability requirements to the acceptance criteria
- Build in human override mechanisms for high-stakes decisions
- Test across demographic segments, not just overall accuracy
After Launch
- Monitor model performance per user segment, not just in aggregate
- Set up bias alert dashboards alongside standard product metrics
- Create a clear escalation path when the model behaves unexpectedly
- Schedule regular bias audits — model behavior drifts as the world changes
Real Cases Where PM Decisions Mattered
Healthcare algorithm bias (2019): A widely-used hospital risk algorithm assigned lower care priority to Black patients because it used healthcare spending as a proxy for health needs — ignoring historical inequities in healthcare access. The product team defined the proxy variable. That was a product decision.
Recommendation radicalization: Multiple platforms have faced regulatory scrutiny for recommendation algorithms that pushed users toward increasingly extreme content. The engagement metric the PM optimized for was the root cause.
Resume screening bias: Amazon's recruiting tool, mentioned above, trained on historical hiring data that reflected past biases. The PM who defined the success metric ("matches our historical best performers") inadvertently baked discrimination into the system.
In each case, the harm was preventable at the product definition stage.
Tools and Frameworks Worth Knowing
- Model Cards (Google): Standardized documentation format for ML models that includes intended use, limitations, and evaluation results across subgroups
- Datasheets for Datasets (Microsoft Research): Documentation standard for training data
- IBM AI Fairness 360: Open-source toolkit for bias detection and mitigation
- LIME / SHAP: Explainability libraries that make model decisions interpretable
You don't need to use all of these. But knowing they exist — and requiring your engineering team to justify why they're not needed — is part of the PM's job.
Ethical AI Is a Competitive Advantage
This isn't just about avoiding regulatory fines. Users are paying attention.
Products that are transparent about AI, that give users control, that behave consistently across different groups — these are products users trust. Trust is a retention driver. Trust is a growth driver.
The companies building AI carelessly are accumulating a liability. The companies building AI responsibly are accumulating an asset.
You get to decide which kind of product you're building.
Free Newsletter
Enjoyed this? Get more like it every week.
Frameworks, teardowns, and AI tools for PMs — free on Substack.