Ethics & Responsible AI
Ethics questions are increasingly common in AI/ML interviews, especially at companies like Google, Microsoft, Meta, and OpenAI. These 8 questions test your ability to recognize ethical issues, make principled decisions under pressure, and balance business goals with societal impact. Interviewers want to see mature judgment, not textbook definitions.
Q1: Tell me about a time you discovered bias in an ML model. What did you do?
Situation: While conducting a routine fairness audit on our resume screening model, I discovered that the model was 23% less likely to advance resumes from candidates who attended historically Black colleges and universities (HBCUs) compared to candidates with similar qualifications from other institutions. The model had been in production for 8 months.
Task: I needed to quantify the bias, determine the root cause, fix the model, and communicate the issue to leadership — all while being sensitive to the fact that this meant our hiring process may have unfairly disadvantaged candidates for 8 months.
Action: I immediately escalated to my manager and our head of HR, framing it as an urgent issue rather than waiting for the next review cycle. I then conducted a root cause analysis. The training data was historical hiring decisions that reflected human biases — the model had learned that candidates from certain schools were less likely to be hired, not because they were less qualified, but because the historical hiring process was biased. I took three actions: (1) Removed institution name as a feature entirely and replaced it with objective education metrics (degree level, GPA, relevant coursework), (2) Applied equalized odds constraints during retraining to ensure consistent advancement rates across demographic groups, and (3) Built a fairness monitoring dashboard that tracked model outcomes by demographic dimensions weekly, with automated alerts if disparities exceeded thresholds. I also recommended that HR review the 8 months of historical decisions to identify candidates who may have been unfairly filtered out.
Result: The debiased model maintained 94% of its original accuracy while eliminating the demographic disparity. HR identified and re-reviewed 47 candidates who may have been affected, resulting in 12 additional interview invitations. The fairness dashboard became a standard requirement for all HR-related ML models. Leadership issued an internal memo about the importance of fairness audits, and I was asked to create a bias detection checklist for the organization.
Q2: Describe a time you had to push back on a stakeholder who wanted to use AI in a way you believed was unethical.
Situation: Our marketing VP wanted to build an ML model that would identify users most susceptible to impulse purchases and target them with aggressive promotional notifications during vulnerable moments (late at night, after paycheck deposits, during emotional events detected from social media activity).
Task: I needed to communicate why this was ethically problematic without being preachy or dismissive of the legitimate business goal of improving marketing conversion rates.
Action: I prepared a structured argument organized around three pillars. First, ethical harm: I cited research on dark patterns and manipulative design, showing that targeting vulnerable moments causes genuine financial harm to users and erodes trust. Second, business risk: I showed examples of companies that faced regulatory action and PR crises for similar practices (the FTC's targeting of predatory advertising algorithms). Third, better alternative: I proposed a model that optimized for "right offer, right time" based on user intent signals (browsing behavior, wishlist activity, search patterns) rather than vulnerability signals. This approach would improve conversions by matching offers to genuine interest, not by exploiting emotional states. I presented this privately to the VP first, not in a group meeting, to avoid putting her on the defensive. I framed it as: "I want to help you hit your conversion targets. Here is an approach that does that without the risks."
Result: The VP agreed to the intent-based approach. It achieved a 19% improvement in notification-driven conversions — better than internal projections for the vulnerability-targeting approach. Users who received intent-matched notifications had 40% lower unsubscribe rates. The VP later thanked me for redirecting the approach, noting that a competitor was publicly criticized for a similar vulnerability-targeting strategy 3 months later.
Q3: Tell me about a time you had to make a data privacy decision that impacted your ML model's performance.
Situation: Our health tracking app collected detailed location data, which was the strongest predictor in our activity recommendation model. When GDPR enforcement intensified, our legal team determined that we needed explicit consent for location data use in ML, and initial testing showed only 35% of users would opt in. Losing location data for 65% of users would significantly degrade recommendation quality.
Task: I needed to maintain recommendation quality while complying with GDPR, respecting user privacy preferences, and not creating a two-tier experience where opted-in users got significantly better recommendations than opted-out users.
Action: I refused to take the shortcut of making the consent prompt manipulative (dark patterns to increase opt-in rates). Instead, I rebuilt the recommendation model with privacy by design. For opted-in users, the model used location data as before. For opted-out users, I developed proxy features that captured similar signals without location: time-of-day activity patterns, device motion sensor data (which users had consented to), and self-reported activity preferences collected during onboarding. I also implemented differential privacy for the location-based model to ensure that even opted-in users' individual locations could not be reconstructed from the model weights. I designed the consent prompt to be genuinely informative: explaining exactly what data would be used, how it would improve their experience, and making it equally easy to accept or decline.
Result: The privacy-preserving model retained 89% of recommendation quality for opted-out users compared to the location-based model. With honest consent prompts, our opt-in rate actually increased to 52% (users trusted us more when we were transparent). The differential privacy implementation was later cited as a best practice in our SOC 2 Type II audit. Apple featured our approach in their privacy best-practices showcase for health apps.
Q4: Describe a time you identified potential harm from an AI system before it was deployed.
Situation: Our team was about to launch an AI-powered content recommendation system for a news platform. During pre-launch testing, I noticed that the model heavily favored sensational and emotionally charged content because it maximized engagement metrics (clicks, time spent, shares). The model was working perfectly by its training objective — the problem was the objective itself.
Task: I needed to raise the concern about potential harm (filter bubbles, misinformation amplification, user wellbeing impact) without delaying a launch that had executive sponsorship and a hard deadline tied to advertiser commitments.
Action: I quantified the issue with data rather than making abstract ethical arguments. I analyzed the model's top recommendations across 1,000 test users and found: 73% of recommended articles had high emotional valence scores, political content was 4x more polarized than the platform's overall content distribution, and the model was creating reinforcing feedback loops where users who clicked one sensational article received increasingly extreme recommendations. I presented this analysis to the product lead and VP with a concrete proposal: add a "content quality" signal to the reward function alongside engagement, implement a diversity requirement ensuring recommendations spanned at least 3 content categories, and add a "rabbit hole detector" that would break recommendation chains if a user's content was becoming increasingly extreme. I estimated the timeline impact: 1 additional week for the quality signal and diversity requirement, 2 weeks for the rabbit hole detector (which could be added post-launch as a fast-follow).
Result: The team agreed to a 1-week delay for the quality signal and diversity requirement, with the rabbit hole detector shipped 2 weeks post-launch. Engagement metrics dipped 3% initially but recovered within a month as user retention improved (users stayed on the platform longer over time when recommendations were diverse and high-quality). The VP later credited the intervention with avoiding a potential PR crisis similar to what competitors faced around algorithmic amplification of harmful content.
Q5: Tell me about a time you had to decide whether to use user data for a purpose it was not originally collected for.
Situation: Our product team wanted to use customer support conversation transcripts (collected for quality assurance) to train a sentiment analysis model that would predict customer churn. The support transcripts contained rich emotional signals, but they were originally collected under a privacy policy that mentioned "quality improvement" not "predictive analytics."
Task: I needed to evaluate whether this data use was appropriate and recommend a path forward that balanced the business value of churn prediction with our ethical and legal obligations to customers.
Action: I collaborated with our legal and privacy teams to assess the situation. We concluded that using the transcripts directly violated the spirit of our privacy policy, even if a lawyer might argue it was technically defensible. I proposed three alternatives: (1) Update the privacy policy and re-consent existing users before using their data — the most ethical but slowest path. (2) Anonymize and aggregate the transcripts so that no individual conversation was used directly, only aggregate sentiment patterns. (3) Train the churn model on different signals entirely: product usage patterns, support ticket frequency, NPS scores (which users explicitly provided for analysis). I recommended option 3 as the primary approach with option 2 as a supplement, because it was both ethically clean and technically sound. I argued against option 1 for practical reasons: retroactive re-consent typically has very low response rates, creating a biased dataset.
Result: We built the churn model using product usage signals and aggregated support sentiment (option 3 + 2). It achieved 81% precision in identifying at-risk customers, which was strong enough to drive a successful retention campaign. When GDPR auditors reviewed our data practices 6 months later, they specifically praised our data use discipline. The product team initially pushed back on the "slower" approach, but the positive audit outcome demonstrated that doing things right is less costly than doing things over.
Q6: Describe a time you had to balance model accuracy against fairness.
Situation: Our insurance pricing model used zip code, credit score, and claim history as primary features. A fairness analysis showed that the model charged 18% higher premiums for customers in predominantly minority zip codes, even after controlling for actual claim rates in those areas. The model was accurate in predicting claims, but the zip code feature was encoding socioeconomic disparities, not risk.
Task: I needed to address the fairness issue while maintaining a model that was actuarially sound — the business could not afford to underprice risk — and navigate the internal tension between the actuarial team (who defended the model's accuracy) and the compliance team (who flagged the disparity).
Action: I facilitated a working session between the actuarial team, compliance, and our ML team to align on principles before solutions. We agreed that: (1) the model should price risk accurately, (2) it should not use proxies for protected characteristics, and (3) any fairness improvement should not create adverse selection risk. With these principles established, I implemented a three-step approach. First, I removed zip code and replaced it with property-level features (building age, construction type, local weather risk scores from public data) that captured actual risk factors without encoding demographic proxies. Second, I added explicit fairness constraints during model training using equalized odds. Third, I built a monitoring dashboard that tracked pricing disparity across demographic groups and flagged any regression. The actuarial team initially resisted removing zip code, so I ran a parallel test showing that the property-level features achieved comparable predictive accuracy while eliminating the demographic disparity.
Result: The fair model achieved within 2% of the original model's predictive accuracy while reducing the demographic pricing gap from 18% to 3%. The remaining 3% gap was attributable to legitimate risk factors (building characteristics) rather than demographic proxies. The state insurance regulator praised our approach during their annual review, and two competitors later adopted similar methodologies after facing regulatory pressure.
Q7: Tell me about a time you had to make a decision about AI transparency or explainability.
Situation: Our loan approval model was a gradient-boosted tree ensemble that was highly accurate but provided no explanation to applicants who were denied. Regulators were increasing scrutiny under the Equal Credit Opportunity Act, which requires lenders to provide specific reasons for adverse actions. Our legal team asked us to make the model "explainable" in 6 weeks.
Task: I needed to add meaningful, accurate explanations to every loan decision without significantly degrading model performance or adding latency to the real-time approval pipeline.
Action: I evaluated three explainability approaches: (1) Replacing the model with an inherently interpretable model (logistic regression or decision list), (2) Using post-hoc explanations (SHAP or LIME), or (3) Building a hybrid system. Option 1 would sacrifice 8 points of AUC. Option 2 was computationally expensive for real-time use (SHAP took 3 seconds per prediction). I chose option 3: keep the complex model for predictions but build a companion explanation model. I trained a logistic regression model on the same features, using the complex model's decisions as labels. This "mimic model" agreed with the complex model 96% of the time and could provide coefficient-based explanations instantly. For the 4% of cases where the models disagreed, I used cached SHAP values computed in a batch process. I worked with our legal team to translate technical feature importance scores into consumer-friendly language: "Your debt-to-income ratio exceeds our threshold" rather than "Feature DTI_ratio had SHAP value -0.34."
Result: We deployed the explanation system 2 days ahead of the 6-week deadline. The regulator reviewed our approach and found it satisfactory. Customer complaints about unexplained denials dropped by 67%. Unexpectedly, the explanation system also helped our credit analysts: they started using the feature importance scores to give applicants specific advice on how to improve their application for future submissions, which improved reapplication success rates by 24%.
Q8: Describe a time you raised an ethical concern that was unpopular or inconvenient.
Situation: Our team was building an employee productivity monitoring tool that used computer vision to analyze webcam feeds, screen activity, and keystroke patterns to generate "productivity scores" for remote workers. The product was scheduled for a major client demo in 2 weeks, and the client (a Fortune 500 company) was excited about it. I had serious concerns about the invasiveness of the approach and its impact on employee wellbeing and trust.
Task: I needed to raise my concerns in a way that would be taken seriously by a team that was excited about a lucrative product opportunity, without being dismissive of the legitimate business problem (managing remote team productivity).
Action: I did not raise the concern casually in a stand-up. I prepared a written analysis and requested a dedicated 30-minute meeting with the product lead and engineering director. My argument had three parts. First, I presented research showing that invasive monitoring decreases employee trust, increases stress, and paradoxically reduces genuine productivity (the Hawthorne Effect reverses under surveillance). Second, I highlighted legal risks: several states were considering employee surveillance legislation, and the EU was already restricting this under GDPR Article 22. Third, I proposed an alternative: instead of monitoring individual behavior, build a team-level analytics tool that tracked outcomes (project completion, delivery quality, communication patterns at the team level) rather than individual surveillance. This would address the client's underlying need (understanding remote team effectiveness) without the ethical and legal risks. I acknowledged that my proposal would mean redesigning the demo and potentially losing this specific deal if the client insisted on individual monitoring.
Result: The leadership team was initially frustrated because of the timeline pressure, but after reviewing my analysis, they agreed to pivot the product positioning. We redesigned the demo around team-level analytics. The client actually preferred this approach — their HR team had been privately concerned about employee backlash from individual monitoring. We closed the deal at a higher contract value because the team analytics product served more stakeholders (HR, management, and employees). Two months later, a competitor launched the surveillance-style product I had argued against and faced immediate employee backlash and negative press coverage.
Key Themes Across Ethics Questions
- Proactive, not reactive: The best answers show you identifying ethical issues before they cause harm, not responding after damage is done.
- Data over ideology: Frame ethical concerns with quantified risk, concrete examples, and business impact — not moral lectures.
- Always provide alternatives: Never just say "we should not do this." Always propose what we should do instead that addresses the same business need ethically.
- Fairness is not optional: Treat fairness as a requirement, not a nice-to-have. Models that are biased are not "accurate" — they are accurately reflecting biased training data.
- Transparency builds trust: Being honest about model limitations, data use, and decision processes builds more long-term trust than opacity, even when transparency is uncomfortable.
- Think systemically: Consider second-order effects. An AI system that optimizes for engagement might amplify harmful content. An employee monitoring tool might destroy the trust it was meant to measure.
Lilly Tech Systems