Research-backed startup idea validation: how the ranking works
Each discovery runs a three-phase pipeline: market research that surfaces real products, pricing, user complaints, and competitor gaps; opportunity generation where every idea must cite specific evidence from that research; and competitor validation that confirms real alternatives exist with real pricing. Results are then ranked by a dual-score system — an insight score for signal quality and a business score for standalone viability — so the strongest directions surface first and weak ideas are automatically downranked.
Why this is different from generic AI idea lists
Generic AI idea lists usually start from common patterns in training data. Research-backed startup idea validation starts from market evidence: named competitors, visible pricing, user complaints, workflow friction, and signals that buyers already spend time or money on the problem. The ranking is meant to answer a practical question: which direction is most worth validating next?
The three-phase pipeline
When you enter a market or niche, the system runs three distinct phases before returning results:
- Research: Searches for real products, real pricing, specific user complaints, quantified market trends, and named competitor gaps. Every finding must include at least one concrete data point: a product name with a price, a complaint with a source, or a trend with a number.
- Generation: Produces 10 opportunities grounded in those findings. Each opportunity must cite a specific research finding in its evidence field, and the set is diversified across different problems and audience segments rather than collapsing into one angle.
- Validation: Searches for real competitors for each generated opportunity and retrieves their actual pricing. If no real competitor with confirmed pricing exists, the system says so instead of fabricating comparisons.
How ranking works
Every opportunity receives two scores. The insight score weights signal quality: speed to revenue, cost efficiency, evidence strength, pricing basis, and scalability. The business score adds viability adjustments: standalone monetization strength, packaging fit, moat, and pricing confidence. Ideas flagged as free tools, lead magnets, or indirect-monetization plays are downranked automatically, so they do not outrank genuinely monetizable standalone products unless the entire market is weak.
What operator fit means
If you specify a role or background, the system adjusts what types of opportunities it generates. Technical users see more SaaS, APIs, and automation tools. Non-technical users see more services, content products, and community plays. This is not cosmetic filtering: the generation prompt, packaging rules, and diversity constraints all change based on who is asking, so the ranked ideas are more likely to fit what the user can actually build and sell.
Limits and caveats
- Research quality depends on what is publicly available about a market. Thin or emerging markets produce weaker findings, and the system should say so rather than invent specifics.
- Scores are directional, not precise. A business score of 7.2 is meaningfully different from 3.1, but the difference between 6.8 and 7.0 is noise.
- Competitor validation confirms the market is active, not that a competitor is successful. Real pricing is evidence of willingness to pay, not proof of revenue scale.
- The system does not replace talking to customers. It gives you a stronger ranked starting point; the next step is always real validation with buyers.
Related questions
See ranked opportunities
See the methodology work on a real market. This opens the workspace with a research-heavy starting point so you can test the ranking flow on your own niche.
Run a research-backed scan →