Bevetu
Investor Scrutiny
Core investor questions and the company's current answers.
- No genuine answer yet
#01
Is the problem real, and is there evidence that users want this solution?
Question
The deck cites the UK CMA investigation to support the problem of information asymmetry in vet care. But where is the company's own evidence that pet owners want Bevetu specifically? How many user interviews were done? How many owners would actually pay £3/month for this?
Current answer
The deck shows that the market has a problem, but it does not prove demand for this product. The CMA report is valid context, but not product validation.
What would count as a real answer
- 30–50 user interviews
- waitlist or landing page conversion data
- pricing tests
- early paid pilots
- Partial answer only
#02
Is the product defensible, or is it just "structured questions + LLM + report"?
Question
Why is this not just a form plus ChatGPT? A prototype seems easy to build. What is the actual moat?
Current answer
The deck argues that the long-term moat is Anibrary, a vet-validated dataset. The near-term differentiation is structured intake, vet-designed flows, and a clinic-oriented report.
Why it matters
The workflow may be useful, but the technical stack as presented is not strongly defensible. The moat is deferred to a future dataset that does not exist yet.
What would count as a real answer
- a credible cold-start dataset plan
- evidence that the workflow outperforms free substitutes
- proof that data quality compounds over time
- No genuine answer yet
#03
Will vets actually use this?
Question
The whole product flow assumes vets will review the report, agree with it, or explain why they disagree. But what evidence is there that vets want this? Wouldn't many vets see it as extra work or a challenge to their judgment?
Current answer
There is no validation in the deck that vets will accept this workflow. It is presented as an assumption.
Why it matters
If vets ignore the report, the owner-facing value drops sharply and the clinic-side feedback loop weakens. This is not a minor issue — it is structural to the current model.
What would count as a real answer
- vet interviews
- design partners
- pilot clinics
- LOIs or workflow testing
- No genuine answer yet
#04
Is Anibrary a real moat, or a cold-start paradox?
Question
The deck presents Anibrary as the defensible asset. But the dataset requires user-submitted outcomes, vet validation, and scale. None of that exists yet. Isn't this just a future promise?
Current answer
Yes — today, Anibrary is more a strategic intention than an actual moat.
Why it matters
The pitch currently depends on a circular logic: users create data, data improves accuracy, accuracy attracts users. Without a seed strategy, this is a classic cold-start problem.
What would count as a real answer
- seeded cases from vet partners or schools
- licensed or public data sources
- retrospective anonymized case contributions
- a narrow-condition launch with expert-reviewed cases
- Partial answer only
#05
Is ChatGPT the real threat?
Question
Why would a user pay for Bevetu when ChatGPT is free and already on their phone? Isn't ChatGPT a substitute?
Current answer
The beginning of an answer exists: Bevetu is not trying to beat LLMs on raw intelligence, but on structured workflow, guided questioning, and trust design.
Why it matters
This is a positioning answer, not a durable defensibility answer. The follow-up will always be: "Why can't ChatGPT add this flow?"
What would count as a real answer
- proof the structured workflow materially improves behavior or outcomes
- trust or compliance advantage
- proprietary data or integrations that general tools lack
- No strong answer, but fixable
#06
Are you solving the right problem, or only a symptom?
Question
The CMA found structural issues in the veterinary market: consolidation, pricing opacity, prescribing conflicts. Bevetu gives owners a better pre-visit assessment. Does it actually solve the problem you are citing?
Current answer
The product does not solve the structural economics of the vet market. It addresses an owner-level symptom: poor preparation and low confidence before treatment discussions.
Why it matters
Better framing: "We are not fixing the veterinary market structure. We are helping owners navigate it better."
- Weak answer
#07
Is the market sizing credible?
Question
The deck uses dog and cat population to build TAM, SAM, and SOM, then assumes 3% of SAM after five years. Why is 3% credible? And should TAM be pets or owners?
Current answer
The market sizing is directionally illustrative, but not deeply grounded. The assumptions appear chosen to support a respectable forecast rather than derived from evidence.
What would count as a real answer
- owner-based segmentation
- usage frequency data
- channel-specific acquisition math
- bottom-up market sizing from real customer behavior
- No genuine answer yet
#08
Do the unit economics make sense?
Question
The model assumes £3/month, 25% paid conversion, and £11 per report for data licensing. But where are CAC, churn, LTV, payback, and buyer use cases?
Current answer
The revenue model is a set of assumptions, not a fully defensible operating model.
Why it matters
For a consumer subscription, churn and acquisition matter more than headline market size. For data licensing, buyer identity and data quality matter more than theoretical price per report.
What would count as a real answer
- scenario-based model
- sensitivity analysis
- early acquisition test results
- identified B2B buyers and use cases
- No genuine answer yet
#09
What is the regulatory and liability risk?
Question
This is medical-adjacent guidance involving animals. What happens if an owner delays care based on the assessment? What are the legal, insurance, and regulatory implications?
Current answer
The deck does not adequately address regulation, professional boundaries, or liability.
Why it matters
This is existential, not cosmetic. The product sits close to clinical judgment and could face questions around claims, disclaimers, liability coverage, and acceptability within veterinary practice norms.
What would count as a real answer
- legal review
- clear product boundary definition
- insurance plan
- documented compliance approach
- Legitimate gap
#10
Is the team complete for this kind of company?
Question
The team appears strong on software and product, but veterinary expertise is advisory, not full-time. Is that enough for a clinically sensitive workflow?
Current answer
The team is credible on product and execution, but there is no full-time veterinary operator in the core team.
What would count as a real answer
- a more engaged veterinary lead
- clinical advisory depth
- formal partnerships with vets or vet institutions
- No genuine answer yet
#11
Is the traction real traction?
Question
The deck cites SEIS/EIS approval, Azure credits, incubator programs, and support services. But where are users, revenue, pilots, or LOIs?
Current answer
The traction shown is ecosystem validation, not market traction.
Why it matters
These signals help credibility, but they do not prove demand.
- Weak answer
#12
Is the raise and valuation justified?
Question
The ask is £250K for 10% equity, implying a £2.5M post-money valuation, pre-launch and pre-revenue. Why is that justified?
Current answer
The deck does not provide enough evidence to strongly justify that price today.
What would count as a real answer
- real user evidence
- pilots
- stronger moat evidence
- narrower claims and a more credible de-risking path