9 min left 0%

ai ethics & responsible ai

If you can focus on the responsible AI piece — ethical, privacy-first, security-first products — you will not only make good money, you will be ahead before everybody else catches up.
Talvinder Singh, from a Pragmatic Leaders lecture on responsible PM

Every AI PM will tell you they care about ethics. Very few can point to a decision where they traded a metric for a principle. That gap — between intention and action — is where this page lives.

AI ethics is not a seminar you attend and forget. It is a set of calls you make every sprint: what data to train on, what failures to tolerate, who bears the cost when the model is wrong, and what you refuse to build even when the business case is strong.

The five principles of responsible AI — fairness, transparency, privacy, accountability, and safety — are easy to print on a slide deck. The hard part is applying them when your model’s accuracy is 92% on one demographic and 71% on another, and your launch date is next Thursday.

Bias is not a bug you fix once

Bias in AI products is not a single defect you patch and move on. It is a category of failures that enter your system through five distinct doors — and if you do not know which door is open, you cannot close it.

// scene:

Model review meeting. The team is evaluating a loan approval model before launch.

Data Scientist: “Accuracy is 94% overall. We're good to ship.”

AI PM: “Break it down by demographic. What's the accuracy for tier-2 city applicants versus metro?”

Data Scientist: “Metro is 96%. Tier-2 is... 78%.”

AI PM: “So one in five tier-2 applicants gets a wrong decision. And our training data — how much of it comes from tier-2?”

Data Scientist: “About 12% of the dataset. Most of our historical data is metro.”

AI PM: “Then we are not launching a loan approval model. We are launching a metro loan approval model and hoping it works elsewhere. That is not the same product.”

Launch was delayed two weeks. The team augmented the training data with tier-2 samples. The second version was 91% accurate across both segments.

// tension:

A 94% overall accuracy hid a 78% accuracy for the users who needed fair treatment most.

The five bias categories every AI PM needs to know:

Dataset bias. Your model is only as fair as the data it learned from. If your training data over-represents one population, the model performs well for that population and poorly for everyone else. In India, this shows up constantly — models trained on English-language data that fail on Hinglish inputs. Credit models trained on metro salaried employees that misjudge gig workers in tier-2 cities. Healthcare models trained on hospital data that miss patterns common in rural primary care.

Association bias. The model learns correlations in the data that encode societal prejudice. A resume screening model trained on a decade of hiring data learns that male candidates are preferred — not because men are better, but because men were historically hired more. The model reproduces the bias it was trained on. Amazon built and scrapped exactly this system.

Automation bias. Users over-trust AI outputs because they come from a machine. A doctor who would question a colleague’s diagnosis accepts an AI’s diagnosis without scrutiny. A recruiter who would push back on a human’s “not a fit” verdict accepts an AI’s rejection score. The PM’s job: design the interface so the human stays in the loop, not rubber-stamping.

Interaction bias. Users shape the model through their behaviour. Microsoft’s Tay chatbot learned from Twitter users and started producing racist content within 24 hours. Recommendation engines that show users what they engage with most create filter bubbles. The model reflects the worst of its users unless you design guardrails.

Confirmation bias. The team interprets ambiguous model results as confirming what they already believe. The model is 94% accurate? Ship it. Nobody asks for whom it is 94% accurate until a PM forces the disaggregated analysis.

Hallucinations are a product problem, not a model problem

When a large language model confidently generates false information — citing papers that do not exist, inventing statistics, fabricating legal precedents — the ML team calls it a hallucination. But for the PM, “the model hallucinates” is not an explanation. It is a product failure you need to manage.

The user does not care why the model was wrong. They care that your product told them something false with the same confidence it tells them something true. The trust damage is identical whether the cause is training data, retrieval failure, or the model’s tendency to complete patterns.

// thread: #ai-product — After a customer complaint about incorrect medical information
Support Lead Customer used our health assistant to check drug interactions. It said ibuprofen and warfarin are safe together. They're not. Patient's doctor caught it.
ML Engineer That's a known hallucination pattern. The model generates plausible-sounding interactions.
AI PM I don't care what it's called. A user got wrong medical advice from our product. What's our guardrail?
ML Engineer We could add a disclaimer saying 'not medical advice.'
AI PM A disclaimer doesn't prevent harm. It prevents lawsuits. Different thing. We need retrieval-augmented generation against a verified drug database. If the model can't cite a verified source, it says 'I don't know.' Ship nothing else.
ML Engineer That'll reduce our response coverage by 30%. 😬 2
AI PM Good. I'd rather cover 70% correctly than 100% with some of it dangerous.

What PMs should do about hallucinations

1. Classify the risk. Not all hallucinations are equal. A chatbot inventing a restaurant recommendation is annoying. A chatbot inventing a drug interaction is dangerous. A chatbot inventing a legal precedent is liability. Classify your use case: is a wrong answer embarrassing, costly, or harmful? Your guardrail intensity should match.

2. Design for uncertainty. The model does not know what it does not know. Your product must. Use retrieval-augmented generation to ground responses in verified sources. Show confidence indicators where appropriate. And when the model cannot answer reliably, make it say so — “I don’t have enough information” is a better product experience than a confident lie.

3. Build feedback loops. Users who flag wrong answers are doing your QA for free. Make it trivially easy to report errors. Route those reports to a human review queue. Use them to improve retrieval, fine-tune the model, and identify systematic failure patterns.

4. Test adversarially. Before launch, try to make your model fail. Feed it edge cases, contradictory inputs, questions outside its domain. If your QA team cannot break it, your users will.

Privacy in the age of AI is different

The Ethics in Product page covers general data privacy. AI adds three specific dimensions that change the calculus.

Training data consent. When you collect user data for analytics, the user understands (roughly) what is happening. When you use their data to train a model that serves other users, the consent boundary shifts. Did your user agree to their conversations training a model that serves a competitor’s employee? Your privacy policy probably says yes. Your user probably did not read it.

Inference reveals what users never shared. AI models can infer sensitive attributes from seemingly innocuous data. Purchase patterns reveal health conditions. Typing patterns reveal emotional states. Location data reveals religious affiliation. In India, where caste, religion, and economic status carry real social weight, inference-based profiling is not a theoretical concern — it is a weapon.

The Indian regulatory landscape. The DPDPA gives Indian users rights over their personal data — consent, access, correction, erasure. But AI complicates every one of these rights. How do you erase someone’s data from a trained model? How do you explain an AI decision to a user who has the right to ask? These questions do not have clean answers yet, which means the PM who builds thoughtful defaults today has a structural advantage when enforcement tightens.

And then there is Aadhaar. Over 1.3 billion people enrolled. Biometric data linked to identity. If your AI product touches Aadhaar data — directly or through a partner — you are handling the most sensitive dataset in Indian history. Masked Aadhaar exists. Virtual IDs exist. Use them. If your engineering team says “it’s easier to just store the full number,” that is a PM failure, not an engineering decision.

// exercise: · 15 min
AI privacy impact assessment

Pick an AI feature you are building (or one you use daily). Answer these questions:

  1. Training data: What data trains the model? Did users consent specifically to model training, or is it buried in a general privacy policy?
  2. Inference risk: What can the model infer about a user beyond what they explicitly shared? List three sensitive attributes that could be derived from the input data.
  3. Right to explanation: If a user asks “why did the AI make this decision about me,” can your team answer? In under 5 minutes? With something the user can understand?
  4. Right to erasure: If a user asks to delete their data, can you remove their contribution from the trained model? If not, what is your mitigation?
  5. Indian-language data: If your product serves users in Hindi, Tamil, Bengali, or other Indian languages — is the training data representative? Or does performance degrade for non-English users?

If you cannot answer #3, your AI product has an explainability gap. If you cannot answer #4, you have a compliance gap. Both will become expensive.

// learn the judgment

You are the PM for CRED's AI-powered spend insights feature. The feature analyses users' credit card statements to generate personalised observations like 'you spent 40% more on food delivery this month than your peer group.' Internal testing shows the feature is accurate for metro users with regular spending patterns. But your data team flags that the peer group comparison uses income proxies derived from credit limit and spend volume — which means users in lower credit tiers are compared against a peer group that skews toward lower-income demographics. A subset of users — 8% of the feature's audience — are receiving insights that inaccurately categorise their spending as 'above average' when they are actually in the normal range for their real peer group, because the proxy is miscalibrated for gig workers and freelancers with irregular income. The VP of Product wants to launch in two weeks.

The call: Do you launch on schedule with the current peer comparison model, or delay to fix the miscalibration?

// practice for score

You are the PM for CRED's AI-powered spend insights feature. The feature analyses users' credit card statements to generate personalised observations like 'you spent 40% more on food delivery this month than your peer group.' Internal testing shows the feature is accurate for metro users with regular spending patterns. But your data team flags that the peer group comparison uses income proxies derived from credit limit and spend volume — which means users in lower credit tiers are compared against a peer group that skews toward lower-income demographics. A subset of users — 8% of the feature's audience — are receiving insights that inaccurately categorise their spending as 'above average' when they are actually in the normal range for their real peer group, because the proxy is miscalibrated for gig workers and freelancers with irregular income. The VP of Product wants to launch in two weeks.

The call: Do you launch on schedule with the current peer comparison model, or delay to fix the miscalibration?

0 chars (min 80)

The responsible AI checklist for PMs

Principles are useless without a process. Here is a checklist — not exhaustive, but practical — that you can apply before any AI feature ships.

Before training:

  • Is the training data representative of your actual user population? Not just your current users — the users you intend to serve.
  • Have you checked for label bias? Who labelled the training data, and what assumptions did they carry?
  • Have you documented what the model should not do? Negative requirements matter more than positive ones in AI.

Before launch:

  • Have you tested accuracy across demographic segments, not just in aggregate?
  • Have you defined what happens when the model is wrong? Not a disclaimer — an actual fallback experience.
  • Can a user understand why the model made a specific decision about them?
  • Is there a human escalation path for high-stakes decisions?

After launch:

  • Are you monitoring for drift — is the model’s performance degrading for specific segments over time?
  • Are you tracking feedback and error reports by category, not just volume?
  • Do you have a kill switch? If the model starts producing harmful outputs, can you shut it down in minutes, not days?

The PM who treats this checklist as a launch gate — not a nice-to-have — builds products that survive regulatory scrutiny, earn user trust, and avoid the front-page headline nobody wants.

Accountability: when the model is wrong, who pays?

This is the question most AI products dodge. When a human makes a wrong decision, there is a person to hold accountable. When an AI makes a wrong decision, accountability diffuses — the data team blames the training data, the ML team blames the model architecture, the PM blames the requirements, and the user is left holding the consequence.

In India, this plays out in specific ways. A credit scoring model denies a loan to a self-employed shopkeeper in Jaipur because it was trained on salaried employee data from Bengaluru. Whose fault is it? The shopkeeper does not care about your architecture diagram. They care that a machine said no, and nobody can explain why.

The PM’s job is to define accountability before launch, not after the first lawsuit:

  • Who reviews edge cases? Not the model. A human.
  • Who is authorized to override the model? Define this role explicitly.
  • What is the appeal process? If a user disagrees with an AI decision, what do they do? “Contact support” is not an appeal process. It is a dead end.
  • Who is responsible when the model causes harm? The answer is the product team. Not the model. Not the algorithm. The people who chose to ship it.

Test yourself

// interactive:
The Indian language model dilemma

You are the PM for a customer support chatbot at a large Indian fintech. The chatbot handles 50,000 queries daily in English and Hindi. Your team has built a new model that also supports Tamil, Telugu, and Bengali — but testing shows accuracy in these languages is 68% compared to 91% in English and 85% in Hindi. Marketing wants to announce 'Now in 5 languages!' at the upcoming product launch event, which is in 10 days.

Your VP of Product says: 'The regional language users are currently getting zero support from the chatbot. Even 68% is better than 0%. Let's launch all five.' The head of CX says: 'If we launch at 68% accuracy, we'll get a flood of complaints in languages our support team doesn't speak.' What do you do?

Where to go next

ai ethics & responsible ai 0%
9 min left