Skip to main content
  1. Behavioral Interviews - 170+/

Tell Me About a Time You Set a Technical Direction That Turned Out to Be Wrong

Lakshay Jawa
Author
Lakshay Jawa
Sharing knowledge on system design, Java, Spring, and software engineering best practices.
Table of Contents
Leadership - This article is part of a series.
Part : This Article

S1 — What the Interviewer Is Really Probing
#

The exact scoring dimension here is technical accountability under authority — not whether you’ve been wrong, but whether you can hold the weight of being wrong cleanly. The interviewer wants to see that technical confidence and intellectual honesty coexist in you. Most engineering leaders have made a bad call; very few describe it without hedging, blame-diffusing, or skipping straight to the fix.

At the EM level, the bar is: you made the call, you felt the cost, you corrected it. The story should show that you have enough technical depth to diagnose your own mistake and enough psychological safety to say “I got this wrong” before saying “here’s what I did.” At the Director level, the bar shifts: the wrong direction affected multiple teams or quarters, the correction required org-level buy-in, and the reflection is about the system conditions that let a bad direction persist — not just the technical judgment error itself.

EM bar: “I decided X, it had Y consequence, I corrected it, and here is what changed in how I make these calls.” Director bar: “My direction shaped how three teams built for six months. Reversing it required a structured transition, re-alignment with product and finance, and a post-mortem that changed our RFC process.”

The failure mode is the candidate who either picks a trivial example where they were barely wrong, or picks a real mistake but softens it with “the team agreed,” “the data wasn’t available,” and “in hindsight it was reasonable.” The upgrade is naming the specific decision clearly, quantifying the cost, and then showing genuine learning — not just “I’d consult more people next time.”


S2 — STAR Breakdown
#

flowchart LR
  A["SITUATION\nSubscription platform\ngreenfield build"] --> B["TASK\nOwn service boundary\narchitecture"]
  B --> C["ACTION\nDecomposed into\nfine-grained Java services,\ndetected cost signal,\nbuilt reversal plan\n(60–70% of answer)"]
  C --> D["RESULT\nConsolidated to\n3 bounded services,\ncost and latency\nnormalised"]

SITUATION — Set enough context to make the stakes clear. Name the product area, the team size, and why you were the one making the architectural call. Avoid vague “we were building a platform” openings.

TASK — Be specific about what decision was yours. “I owned the service decomposition strategy” is better than “I led the architecture.” At the EM level, the decision is yours alone. At the Director level, the decision shaped a cross-team direction.

ACTION — This is 60–70% of the answer. Cover: what you decided and why it seemed right at the time; what early signals you noticed or rationalised away; the moment you accepted it was wrong; how you built the reversal plan; how you communicated the mistake to the team. Include one moment of doubt: “I remember the infra bill looking off at week eight and telling myself it would amortise — it didn’t.” Use “I,” not “we,” for decisions you owned.

RESULT — One metric. What the wrong direction cost — time, money, team bandwidth. What the corrected direction achieved. Close with one concrete process change that came out of it, not a platitude.


S3 — Model Answer: Engineering Manager
#

This answer draws from real experience.

[S] We were building a subscription management platform for a telecom ecommerce product — SIM plan upgrades, renewals, entitlement grants, billing triggers, notifications. It was a greenfield build, the team was eight engineers, and I owned the architecture. I had a strong prior: fine-grained microservices aligned to the single-responsibility principle would let us move independently, test in isolation, and scale the components that mattered most.

[T] I was responsible for the initial service boundary design and had to have something reviewable within two weeks so parallel development streams could begin.

[A] I decomposed the subscription domain into nine Java Spring Boot services: plan-catalog, subscription-lifecycle, entitlement, billing-trigger, renewal-scheduler, notification-dispatcher, proration-calculator, dunning, and a thin API gateway. The reasoning was clean on paper — each service had one job, one sub-team could own it, and we could scale renewal independently from notification. I ran the design through a review, got buy-in, and we started building.

Three months in, the infra bill was roughly three times our projection. Each Spring Boot JVM needed a minimum of 512 MB to start, and we were running two replicas per service across dev, staging, and production. The compute cost scaled linearly with service count in a way I had not modelled. Beyond cost, the inter-service call graph was dense. A single subscription renewal touched six services in sequence. P99 latency on the renewal flow was 1.4 seconds. Debugging a failed renewal meant correlating logs across six services. The team was spending more time on service mesh configuration and distributed tracing than on business logic.

I remember looking at that p99 number and realising the performance tuning pass I had been planning would not fix the fundamental problem — the seams were wrong. The services were too granular for the actual change velocity and call density of the domain. I could have rightsized containers and consolidated environments and called the architecture sound. I chose not to. I drafted a consolidation proposal: three services instead of nine — subscription-core covering lifecycle and entitlement, billing-engine covering trigger and proration and dunning, and comms-service covering notification and scheduling. I presented it to the team with the cost data and owned the original decision explicitly: “I overfit to SRP without modelling JVM memory overhead or inter-service call density.”

[R] Compute cost dropped 58% within a quarter. P99 on the renewal flow fell from 1.4 seconds to 340 milliseconds. On-call incidents related to inter-service timeouts dropped to near zero in the following cycle. What changed in how I make these calls: every service boundary RFC now requires a back-of-envelope infrastructure cost model that accounts for runtime overhead, and any Java service decomposition proposal must include a call-graph density analysis before it gets signed off.


S4 — Model Answer: Director / VP Engineering
#

[S] We were scaling a real-money gaming platform from a single fantasy sports product to four verticals — fantasy sports, poker, casual games, and a sportsbook. Each vertical had its own engineering team. I had just taken over as Director across sixty engineers in six squads. My first major architectural call was to mandate an event-driven microservices model across all verticals, replacing the existing service-per-vertical pattern. The rationale was coherent: decouple teams, enable independent deployment, let each vertical own its data stores. I had seen it work at a previous org at twice the scale, and I moved quickly — six weeks from RFC to mandate.

[T] I owned the architectural direction across all four product engineering teams and the shared platform team. This was an org-level decision that would shape hiring, tooling, and delivery for the next eighteen months.

[A] We built Kafka as the backbone. Each vertical published domain events; other verticals consumed them. Within three months we had forty-plus event schemas in production across teams. Within five months the problems were unmistakable. Schema evolution became a coordination tax — every schema change needed a migration window across all consumers. Debugging a failed contest payout required replaying Kafka offsets across four verticals and correlating six event types. On-call burden tripled. The sportsbook team, which was newest, was spending forty percent of sprint capacity on event plumbing rather than product features.

I had two choices: invest in better tooling — a schema registry, a saga orchestrator, distributed tracing — and try to make the pattern work, or accept that I had applied an architecture that did not fit our team maturity or our actual data sharing model. I ran a structured post-mortem, not just on the technology, but on my decision process. I had moved too fast, had not assessed whether our teams had the operational maturity to run a distributed event mesh, and had underestimated how much data the verticals shared — which made event choreography brittle rather than decoupling.

The reversal plan was a bounded-context consolidation: three macro-domains instead of forty fine-grained services, with inter-domain communication through a platform-owned API layer rather than raw event consumption. I took the proposal to the VP Product and CFO with a six-quarter cost-benefit model and took accountability for the twelve-week migration cost. I restructured the RFC process to require a maturity assessment and a dissenting-reviewer sign-off before any org-wide architectural mandate could proceed.

[R] By the end of the migration, deployment incidents fell sixty-two percent, the sportsbook shipped its first feature vertical on schedule for the first time, and the on-call rotation normalised across all verticals. The maturity assessment gate has since blocked two subsequent premature architectural mandates that would have had similar consequences.


S5 — Judgment Layer
#

Assertion 1: Naming the specific decision is non-negotiable. Vague ownership — “we moved in a direction that didn’t work” — disqualifies you at EM or Director level. The interviewer needs to hear “I decided X” before they can evaluate anything else. The trap: “The team agreed on the approach.” The upgrade: “I owned the service boundary design. The team built what I scoped.”

Assertion 2: The cost must be quantified, not described. “It caused some problems” is forgettable. “Compute cost was three times projected” or “on-call incidents tripled” is what sticks. Metrics signal that you operated with accountability, not just memory. The trap: “We realised it wasn’t working well.” The upgrade: “Compute cost was 58% above baseline within one quarter.”

Assertion 3: The moment of acceptance is the most important beat. Strong candidates describe the exact moment they stopped defending the direction. This is what separates intellectually honest leaders from ones who course-correct only when forced. The trap: Skipping straight to “so we changed it.” The upgrade: “I remember looking at the p99 data and realising the performance tuning pass I had been planning would not fix a seam problem.”

Assertion 4: The reversal plan should be yours, not the team’s reaction. The best answers show that you drove the correction — you did not wait for someone above you to notice, and you did not let the team quietly route around your direction. The trap: “The team started consolidating services on their own.” The upgrade: “I drafted the consolidation proposal, owned the communication, and ran the correction sprint.”

Assertion 5: Reflection must be process-level, not platitude-level. “I learned to consult more people” is a platitude. “I now require a back-of-envelope infrastructure cost model in any service boundary RFC” is a process change. Interviewers at this level have heard the platitudes. The trap: “I now listen more carefully before making big decisions.” The upgrade: “We added a call-graph density requirement to the service boundary RFC template.”

Assertion 6: The Director answer must show cross-functional correction, not just technical correction. A Director who reverses an architectural mistake without involving product, finance, or an executive layer is operating below bar. The correction is itself a stakeholder management exercise. The trap: Director tells a story where they retooled the tech quietly. The upgrade: “I presented a six-quarter cost-benefit model to the VP Product and CFO and took explicit accountability for the twelve-week migration cost.”

Assertion 7: The story should not be too old. If your example is from six or more years ago, the interviewer will wonder whether you have been in a protected environment since then. Aim for within the last three years. The trap: Opening with “early in my career…” for a senior role interview. The upgrade: A recent example that shows the lesson is current, not historical.


S6 — Follow-Up Questions
#

1. “How long did it take you to realise the direction was wrong?” Why they ask: Tests signal detection — did you notice early or only when it was undeniable? Model response: “The first cost signal appeared at week eight. I rationalised it for about three weeks before accepting it was not an anomaly. In hindsight I should have acted on week eight, not week eleven.” What NOT to do: Say “we noticed pretty quickly” without specifics — it sounds defensive.

2. “What would you have done differently at the design stage?” Why they ask: Tests whether the learning is superficial or structural. Model response: “I would have modelled JVM startup overhead explicitly in the RFC and stress-tested the inter-service call graph before committing to nine services. SRP was the right principle — the granularity was wrong.” What NOT to do: Say “I would have gotten more buy-in” — the problem was the design, not consensus.

3. “How did the team respond when you told them the direction was wrong?” Why they ask: Tests empathy and psychological safety — how do you hold authority and honesty simultaneously? Model response: “There was frustration — they had built to spec and now we were reversing it. I named that directly. I told them the design was mine, the call to reverse it was mine, and the consolidation sprint was extra work I was asking them to absorb because of my call. Naming it clearly landed better than I expected.” What NOT to do: Say “they were all very supportive” — it is not credible and suggests you are unaware of the friction your correction created.

4. “Was there anyone who flagged the issue before you accepted it was wrong?” Why they ask: Tests whether you create environments where dissent surfaces, or where it gets filtered out. Model response: “One engineer flagged the call graph density at week six. I treated it as a performance tuning problem at the time rather than a design signal. That was my mistake — I had categorised the concern before I had evaluated it.” What NOT to do: Say “no one flagged it” without reflection — it suggests either no one felt safe raising it or you are not remembering accurately.

5. [Scope amplifier — EM → Director reframe] “Imagine this had affected three teams instead of one. How would your approach to the reversal have changed?” Why they ask: Tests whether you can scale your thinking; used to differentiate EM from Director candidates. Model response: “The technical correction would be similar, but the process would change substantially. I would need a migration plan that minimised disruption across all three teams’ roadmaps, a business case for the reversal cost presented to product and finance, and a post-mortem on the RFC process that let a flawed mandate through without a dissenting-reviewer gate.” What NOT to do: Treat it as just a bigger version of the same problem — the interviewer wants to see that org-level correction is a different skill from team-level correction.

6. “How do you now evaluate technical direction proposals before committing?” Why they ask: Tests whether learning translated into a repeatable system. Model response: “Three gates in any architecture RFC: a back-of-envelope cost model that accounts for operational overhead, a call-graph density analysis for distributed systems, and a dissenting reviewer who has to explicitly sign off or document their objection.” What NOT to do: Describe a vague “I am more careful now” posture without mechanics.

7. “What is the difference between a wrong direction and a right direction that needs adjustment?” Why they ask: Tests calibration — not every course correction is a failure. Model response: “A wrong direction has a flawed premise — something in the design logic is incorrect from the start. An adjustment is a correct direction encountering new information mid-flight. In my case, the SRP premise was not wrong, but the granularity assumption was — which makes it a wrong direction, not just an adjustment.” What NOT to do: Blur the distinction — it suggests you would reframe future failures as adjustments rather than own them.


S7 — Decision Framework
#

flowchart TD
  A["Signal that technical\ndirection may be wrong"] --> B{"Fundamental design\nflaw or implementation\nissue?"}
  B --> C["Implementation issue"] --> D["Targeted fix,\nnormal process"]
  B --> E["Fundamental flaw"] --> F{"How far in\nare we?"}
  F --> G["Early\n< 2 months"] --> H["Correct fast,\ncommunicate narrowly,\ndocument premise error"]
  F --> I["Deep\n> 2 months\nor multi-team"] --> J["Quantify compounding\ncost of staying vs\nreversal cost"]
  J --> K{"Reversal cost\n< compounding cost?"}
  K --> L["Yes"] --> M["Own the call,\nbuild reversal plan,\nengage stakeholders"]
  K --> N["Not yet"] --> O["Mitigate, document,\nplan structured\nexit window"]
  M --> P["Name the mistake\nbefore the fix\nin all communications"]

S8 — Common Mistakes
#

Mistake Why It Fails What Good Looks Like
Picking a trivial example Signals no real accountability experience; wastes a high-signal question Choose a decision with measurable cost that required genuine correction
“We made this decision together” Diffuses ownership; the interviewer needs to see your judgment, not shared blame “I owned the service boundary design. I got it wrong.”
Skipping to the fix without naming the cost Sounds like damage control, not honest reflection Quantify what the wrong direction cost before describing the correction
Vague reflection (“I’d be more collaborative”) Platitude, not process; suggests the learning did not stick Name a specific gate or artefact you added to your decision process
Story is too old — six-plus years ago Raises the question of whether you have been in a protected environment since Use an example from the last three years
EM answers a Director question Story stays at single-team scope for a DIR-level role Raise scope: cross-team impact, cross-functional correction, process change at org level
Director answers an EM question Over-abstracts when the role needs hands-on technical judgment Directors still need to show personal technical reasoning, not just org response
No moment of acceptance Answer jumps from “we saw a problem” to “we fixed it” Name the specific moment you stopped rationalising and accepted the direction was wrong

S9 — Fluency Signals
#

Phrase What It Signals Example in Context
“The seams were wrong” Deep design vocabulary; distinguishes boundary error from implementation error “SRP was sound — the seams were wrong. Nine services for that call density was never going to work.”
“I modelled the wrong variable” Precise self-diagnosis; not defensive “I modelled team ownership boundaries. I did not model JVM startup overhead or inter-service call graph density.”
“The premise, not the execution, was flawed” Shows you can distinguish types of failure “This was not poor execution of a correct design. The premise — that fine-grained decomposition would reduce cost — was wrong for a JVM stack at this service count.”
“I stopped defending it at [specific signal]” Names the acceptance moment with precision “When p99 hit 1.4 seconds on a renewal flow touching six services, I stopped treating it as a tuning problem.”
“The correction was mine to own, not the team’s to absorb silently” Leadership vocabulary; signals psychological safety awareness “I named the mistake before asking the team to pick up the consolidation work.”
“What changed in my process” Shows learning translated into a system, not a feeling “We added a back-of-envelope infra cost model as a required RFC artefact. No service boundary proposal ships without it.”
“Compounding cost of staying versus reversal cost” Strategic vocabulary; shows you evaluated the trade-off rather than just reacting “At month three the compounding cost of staying on nine services was clearly exceeding the reversal cost. That is when I moved.”

S10 — Interview Cheat Sheet
#

Time target: 4–5 minutes. Do not rush past the cost or the acceptance moment — those are the substance of the answer.

EM vs Director calibration:

  • EM: one team, one module, technical correction plus team communication
  • Director: multi-team impact, cross-functional correction, RFC process change at org level

Opening formula: “I made a service decomposition call on [product area] that was grounded in [principle] but did not account for [specific variable]. Here is what happened and what it cost.”

The one thing that separates good from great on this question: most candidates describe a mistake and a fix. Great candidates describe the moment of acceptance — when they stopped rationalising and owned the call. That is the beat interviewers remember. Build your answer around that pivot, not the technical correction that follows it.

If you blank: Start with the cost — “The thing I got wrong cost us X” — and work backwards. Cost → what caused it → what I decided → why I thought it was right → when I accepted it was not. The reverse structure often unlocks the story faster than trying to start at the situation.

Leadership - This article is part of a series.
Part : This Article

Related

Give Me an Example of a Time You Identified a Problem No One Else Saw and Took Ownership of Fixing It

S1 — What the Interviewer Is Really Probing # The exact scoring dimension is proactive vigilance and unassigned ownership — the disposition to notice a signal others missed, run it to ground without being asked, and decide that fixing it is your job before anyone tells you it is. This is not a question about being a good citizen. It is a question about whether you create a different environmental outcome than someone with identical authority and identical information would create by default.

Tell Me About a Time You Disagreed With Your Manager's Direction but Still Had to Lead Your Team to Execute It

S1 — What the Interviewer Is Really Probing # The exact scoring dimension is disagree-and-commit discipline — the ability to hold your professional obligation cleanly separate from your personal conviction. This is one of the most important leadership tests in any panel because it exposes whether your integrity survives disagreement. Almost every candidate says they disagreed and still executed. Very few demonstrate that they executed without hedging, without signalling their displeasure to their team, and without protecting themselves by leaving a paper trail of “I told you so.”

Describe a Situation Where You Had to Make a High-Stakes Decision With Incomplete Information

S1 — What the Interviewer Is Really Probing # The exact scoring dimension is judgment under uncertainty — the interviewer is not testing whether you made the right call. They are testing whether you have a structured mental process for making consequential decisions when the information needed for certainty either doesn’t exist yet or can’t be obtained in time. This is the difference between a leader and an analyst: analysts wait for completeness; leaders learn to act on enough.

Led a Team Through a Significant Technical Change They Were Resistant To

S1 — What the Interviewer Is Really Probing # The scoring dimension here is change leadership under resistance — not change management in the HR-training sense, but your ability to drive conviction-led transformation while preserving the trust of the people who are pushing back. Interviewers care about whether you understand why resistance exists. Is it fear of irrelevance? A legitimate technical objection? Loss of ownership over something engineers spent years building? Leaders who conflate all resistance as “people being difficult” and steamroll it create short-term compliance and long-term resentment. Leaders who can diagnose and address root cause create lasting followership.