Trusted by teams at
In this article
In final-two vendor decisions, 61% of enterprise buyers report making the final call primarily on "fit" or "gut feel" because their evaluation criteria ran out of resolution. That outcome is avoidable with the right final-stage framework. Both vendors on your shortlist have already passed your filters. They have answered your questions, passed your reference checks, and impressed the people in the room. The problem is not that either vendor is obviously wrong. The problem is that your evaluation process was built to eliminate bad options, not to differentiate between good ones. This article gives you the framework to run the final stage correctly.
Key findings
Most final-two vendor decisions feel like a tie because the evaluation was designed to filter, not to differentiate - the criteria run out of resolution right when you need them most.
The asymmetry test reveals which vendor advantages actually map to outcomes your board will track, separating meaningful differences from surface-level ones.
The reference calls that matter most at the final stage are not the ones vendors suggest - they are the calls you initiate with clients who match your situation most closely.
Writing the decision memo before you decide is the single most reliable way to force the trade-offs into the open and produce a choice you can defend under scrutiny.
Why the final-two decision is harder than the longlist
Getting to the final two is the hardest the decision has been. The longlist phase had obvious eliminations: vendors who could not name a specific quality rate, agencies without enterprise experience in your industry, firms where the person in the pitch was not the person who would do the work. By the time you are choosing between two finalists, all of those signals are gone. Both vendors have cleared your bar on every dimension you set.
That creates a specific problem. Your evaluation framework was built to eliminate vendors who fail a minimum threshold. It was not built to rank two vendors who both exceed that threshold on every dimension. When buyers hit this point, the framework runs out of resolution and the decision defaults to subjective language: cultural fit, communication style, "just felt right." Those are not bad signals. They are also not the basis for a board-level decision on a multi-year mobile program.
The final stage requires a different tool. Instead of asking whether each vendor clears your bar, you need to ask which vendor is measurably better on the dimensions that will determine whether your program succeeds in the first 12 months. That is a narrower question with a more specific answer.
One more thing makes the final-two decision harder. Both vendors know they are competing. Their communication becomes more attentive. Their responses are faster. Their proposals are more tailored. You are seeing the best version of each vendor, not the working version. The framework below is designed to get past the sales posture and surface the operational reality.
The asymmetry test: what each vendor does better that actually matters
The asymmetry test is the right starting point because it reframes the question. The test separates vendor advantages that matter for your situation from vendor advantages that look good in a proposal.
Start by listing every dimension where one vendor outperformed the other in your evaluation. Include quality rate, delivery speed, how often the app ships, team stability, compliance experience, and communication track record. For each advantage, write one sentence that connects it to a specific outcome your board will track. If you cannot write that sentence, the advantage does not belong on your list.
An example. Vendor A has a higher quality rate on their reference apps - 99.5% versus 99.1%. That sounds significant. The question is whether that gap matters for your program. If your app serves 500 internal field operations staff on a closed network, both numbers are acceptable and the gap does not change the decision. If your app is about to go to 2 million consumers and a 0.4% difference means 8,000 more crash reports per day, the gap is decisive. The asymmetry test forces you to make that call explicitly rather than treating all advantages as equal.
Do the same exercise in reverse. For every advantage one vendor has, ask what it costs you from the other vendor. A vendor who ships faster but has a higher rate of regressions reaching users may be slower on a risk-adjusted basis. A vendor with deeper compliance experience may add four weeks to every scope change because their review process is more rigorous. Neither of those trade-offs is obvious from the proposals alone.
After the asymmetry test, you should have a short list of genuine differences that are tied to specific outcomes. If that list is empty - if both vendors are genuinely equivalent on every dimension that maps to a board outcome - you are in a different situation and the tie-breaker section at the end of this article applies.
How to run a structured debrief of your evaluation evidence
A structured debrief surfaces the differences that did not show up during the evaluation because you were not looking for them at the time. Most evaluation processes generate more signal than the buyer uses. The debrief is where you use it.
Pull every piece of evidence you collected on each vendor: the written proposals, the demo recordings, the notes from reference calls, the written responses to your due diligence questions. Read them in sequence for each vendor, looking for four things.
The first is internal consistency. Does the vendor say the same things in writing that they said in the demo? A proposal that describes a two-week release cycle, combined with a reference call where the client mentions quarterly releases, is a red flag that did not get flagged. Look for gaps between the pitch and the evidence.
The second is specificity. Which vendor gave you concrete numbers and verifiable examples, and which vendor described process? "We use automated testing" is not the same as "our crash-free rate on the logistics app we shipped last year was 99.3%." Specificity in the evaluation is the most reliable predictor of specificity in delivery. Vendors who cannot answer with data in the pitch will not have the data in the engagement either.
The third is how each vendor handled pushback. Did you ask either vendor a hard question - about a scope assumption, a timeline risk, a past failure - and how did they respond? The vendor who gave you an honest, specific answer with a clear resolution path is showing you their communication posture under pressure. The vendor who gave you a reassuring non-answer is showing you something different.
The fourth is what each vendor asked you. The questions a vendor asks during the evaluation reveal what they are actually worried about. A vendor who asked specific questions about your compliance requirements, your internal stakeholder process, or the current state of the app is gathering information they need to de-risk delivery. A vendor who asked mostly about budget and timeline is gathering information they need to close the deal.
The reference call you have not made yet
By the final stage, you have probably spoken to the references each vendor offered. Those calls are useful but structurally limited. The clients were selected because they are satisfied. The vendor coached them in advance. The most useful reference call at the final stage is one you initiate yourself.
For each finalist vendor, find a client they did not suggest. Most vendors list their clients on their website or on Clutch. Find a client from the last 18 months whose situation resembles yours most closely - similar industry, similar app complexity, similar team size, similar board mandate. Contact them directly and ask for 20 minutes.
The questions that generate the most useful signal at this stage are different from standard reference questions. You already know whether the vendor delivered. What you do not know is the texture of the working relationship under pressure.
Ask: "In month three of the engagement, when you had moved past the initial excitement, what did working with this team actually feel like day to day?" That question gets past the highlight reel and into the operational reality.
Ask: "Was there a moment where you were genuinely worried about the engagement? What happened and how did the vendor handle it?" Every real engagement has at least one difficult moment. A reference who cannot name one either had an unusually smooth engagement or is not being candid. Either way, the answer is informative.
Ask: "If you had to sign with one of two vendors on your final shortlist and this vendor was one of them, what would you tell someone who was on the fence?" That question shifts the reference from advocate to advisor, which is the only posture that gives you useful information at this stage.
Not ready to commit yet? A 30-minute call gives you a straight read on how Wednesday compares to the other vendor on your shortlist - including the trade-offs we would not recommend hiding.
Read more decision guides →How to weight the final criteria for your board
Before you decide, write the decision memo. This is not a summary of what you already know. It is a document that forces you to name the criteria, the weights, and the scores - and then defend them.
The memo has four parts. The first is the decision context: what you are buying, what it needs to do, and what your board will measure in the first 12 months. The second is the criteria and weights: the dimensions from your asymmetry test and the relative importance you assigned to each one. The third is the evidence: one paragraph per criterion per vendor, citing specific numbers or examples from your evaluation. The fourth is the decision: which vendor, and why the evidence points there.
The weights you assign to criteria should reflect your board situation, not generic vendor evaluation logic. Here is how to think about the three most common board mandates.
If your mandate is to move faster because your current vendor is slow, weight delivery speed and release rhythm at 40% combined. The other dimensions matter, but a vendor who ships reliably every two weeks solves the problem that brought you here. Require specific numbers from the last six months of each vendor's reference apps before assigning scores.
If your mandate is to add AI features because the board has set that direction, weight compliance posture and data handling experience at 35% combined. AI features that touch user data require vendors who have handled regulated data before. A vendor without that track record will learn on your engagement, and that learning is expensive.
If your mandate is to stabilize after a bad vendor experience - the previous team was inconsistent, communication broke down, the app shipped less reliably than promised - weight team stability and communication track record at 40% combined. The evidence that matters most is whether the same engineers who started an engagement are still on it 12 months later, and what happened the last time something went wrong.
The process of writing the memo does one more thing. It reveals whether your decision is actually supported by the evidence or whether you are rationalizing a preference. If you write the memo and find yourself reaching for subjective language - "they just seem more professional," "the communication felt easier" - you need more evidence before you commit.
What to do if you genuinely cannot decide
If you have run the asymmetry test, debriefed your evaluation evidence, made the reference calls, and written the memo - and the two vendors are still genuinely equivalent on every dimension that maps to a board outcome - you are not stuck. You are in a tie-breaker situation with a specific resolution.
The tie-breaker is not a coin flip. It is a structured answer to one question: which vendor is more recoverable from if you are wrong?
Think through what failure looks like in the first 90 days with each vendor. If the relationship breaks down, what does switching cost you? A vendor who has shipped an app similar to yours before - same platform, same industry, similar team size - will produce a more transferable starting point than a vendor who is entering your industry for the first time. The app built by the vendor with more relevant experience is more likely to be maintained by your next vendor without a full rebuild.
The second tie-breaker is the behavior at the contract stage. The vendor who gives you more flexibility in the contract - on notice periods, on team reassignment, on what happens if delivery milestones are missed - is showing you something real about how confident they are in their ability to deliver. A vendor who pushes back on every protective clause is not necessarily hiding something, but it is worth asking why.
The third tie-breaker is what happens when you ask each vendor to tell you where the other vendor is stronger. A vendor who answers that question honestly and specifically - "our competitor has more experience in healthcare, and if that compliance posture is your highest priority, you should weight that heavily" - is showing you a posture that will serve you well when something goes wrong in the engagement. A vendor who deflects or dismisses the question is showing you something different.
If you have worked through all three tie-breakers and the decision is still genuinely ambiguous, the answer is to start smaller. Run a four-week scoped project with one vendor before committing to a multi-year engagement. The cost of a pilot is almost always lower than the cost of signing the wrong vendor for three years.
Bring your vendor shortlist. In 30 minutes you will have a side-by-side read on how each vendor compares on the dimensions that matter for your board situation - and a straight answer on whether Wednesday is the right fit.
Book my 30-min call →Frequently asked questions
Not ready for a conversation yet? Browse cost analyses, vendor comparisons, and decision frameworks for every stage of the buying process.
Read more decision guides →About the author
Rameez Khan
LinkedIn →Head of Delivery, Wednesday Solutions
Rameez has shipped mobile products at scale across on-demand logistics, entertainment, and edtech, and has led enterprise AI enablement across multiple Wednesday engagements. As Head of Delivery at Wednesday Solutions, he oversees how every engagement is scoped, staffed, and run from first build to production.
30 minutes with an engineer. You leave with a squad shape, a monthly cost, and a start date.
Get your start date →Keep reading
Shipped for enterprise and growth teams across US, Europe, and Asia