
Haladir
Rooting Formal Methods into CodeGen RL & Data
About
RLVR changed the game for post-training. We're here to build and supply the next-generation of coding data and RL environments using model checking languages and formal verification. In a world where solid coding data is becoming increasingly sparse, Haladir provides the data and environments needed to build the next generation of coding models.
Founders
AI Research Report
Problem & Solution
Problem / Solution Report
Problem
Next‑generation AI coding models face two critical constraints: data scarcity and reliability. Public code corpora are increasingly filtered or restricted, and existing datasets lack machine‑checkable contracts or formal specifications. Consequently, models may generate unsafe or incorrect code, especially in safety‑critical domains, and evaluation remains limited to surface‑level tests.
Solution
Haladir’s proposition—Rooting Formal Methods into CodeGen RL & Data—addresses this gap through three pillars:
- Formally Verified Training Data – High‑quality code datasets enriched with contracts, weakest‑precondition‑driven specifications, and conversion pipelines that retro‑fit existing corpora with formal guarantees.
- RL Environments – Verification‑aware environments where coding agents receive rewards tied to formal spec compliance, enabling reinforcement‑learning‑based alignment beyond traditional test suites.
- Model Validation – Rigorous frameworks that apply formal correctness verification and safety‑property testing to AI‑generated code, delivering state‑of‑the‑art performance on correctness metrics.
Significance
Embedding model‑checking languages into both data and reward structures reduces hallucination‑induced defects and raises trust for AI‑assisted development, particularly in regulated or high‑assurance sectors such as infrastructure, embedded systems, and finance.
Value Proposition
- For model developers (labs, hyperscalers, dev‑tool vendors): access to formally verified datasets and RL environments that improve alignment and reduce downstream bugs.
- For enterprises deploying agentic coding: a defensible pathway to production with stronger correctness guarantees, lowering risk and accelerating adoption.
Haladir therefore positions itself as a critical infrastructure layer for the emerging ecosystem of AI‑driven software engineering.
Market & Competitors
Market and Competitors Report
Market Context & Trends
The AI Code Tools market is projected to grow from $4.86 B in 2023 to $26.03 B by 2030 (CAGR 27.1%). The AI‑in‑Software‑Development segment expands from $674.3 M in 2024 to $15.7 B by 2033 (CAGR 42.3%). The AI Training Dataset market rises from $2.82 B in 2024 to $9.58 B by 2029 (CAGR 27.7%). These macro trends, combined with $33.9 B of private generative‑AI investment in 2025, indicate strong demand for tooling that ensures correctness and safety.
Competitive Landscape
| Category | Notable Players | Relation to Haladir | |---|---|---| | Formal verification vendors | Certora, Runtime Verification, Galois, AdaCore SPARK | Offer verification services but do not provide AI‑focused training data or RL environments. | | Code analysis / security tools | GitHub CodeQL, Semgrep, Snyk Code | Provide static analysis and security checks; complement validation but lack formal‑spec RL reward loops. | | AI coding assistants | GitHub Copilot, Gemini Code Assist, Amazon CodeWhisperer, Sourcegraph Cody | Potential customers; may eventually build in‑house datasets but currently rely on generic code corpora. | | RL / evaluation infrastructure | Patronus AI (coding RL environments), Ragas, Evidently AI | Offer generic RL environments; Haladir differentiates by centering formal verification as the core reward and benchmark. | | Academic benchmarks | Clever (formal‑verified code generation benchmark) | Highlights emerging interest in formally verified generation; validates market need. |
Competitive Advantages
- Specialized focus on formally verified datasets and verification‑driven RL rewards, a niche not addressed by mainstream analysis or verification tools.
- End‑to‑end stack from data creation to validation, creating defensibility for safety‑critical customers.
- YC backing and early investors (SV Angel, Susa Ventures, Joshua Browder) provide credibility and access to enterprise networks.
Risks
- Market nascence: large AI labs may develop proprietary equivalents.
- Adoption barrier: customers need expertise in formal methods.
- Competition could evolve to integrate formal verification into their own pipelines.
Overall, Haladir operates in a rapidly expanding AI‑software market but occupies a differentiated, high‑value niche centered on formal guarantees.
Total Addressable Market
Quantitative TAM Report
Haladir operates at the intersection of three high‑growth segments: AI code‑generation tools, AI‑in‑software‑development, and AI training datasets. Market research provides the following baseline figures:
- AI Code Tools – $4.86 B in 2023, projected to $26.03 B by 2030 (CAGR 27.1%) [Grand View Research].
- AI in Software Development – $674.3 M in 2024, projected to $15.7 B by 2033 (CAGR 42.3%); 2025 estimate ≈ $933 M [Grand View Research].
- AI Training Dataset – $2.82 B in 2024, projected to $9.58 B by 2029 (CAGR 27.7%) [MarketsandMarkets].
Using a top‑down triangulation, we approximate the spend on formally verified datasets, specification authoring, RL environments and validation as 10‑25 % of the adjacent market spend. Applying this share to the 2025 combined adjacent spend (~$7.9 B + $933 M + ~$3.6 B ≈ $12.4 B) yields a 2025 TAM range of $1.2 B‑$3.1 B. Extending to 2029/30, the combined adjacent market reaches roughly $35.6 B, giving a 2029‑30 TAM range of $3.6 B‑$8.9 B.
Investment momentum further supports growth: the 2025 Stanford AI Index reports $33.9 B of private investment in generative AI, an 18.7 % increase from 2023, indicating robust capital flow into tooling and infrastructure that Haladir targets.
Methodology: 1) Extract base market sizes from credible reports. 2) Project intermediate years using reported CAGRs. 3) Apply a conservative share‑of‑spend factor (10‑25 %) to isolate the niche of formal‑methods‑enabled data and environments. 4) Validate the upside narrative with macro‑investment trends.
Founder Analysis
Founders and Background Report
Haladir was founded in 2025 and is part of Y Combinator’s Winter 2026 batch. The founding team listed by Y Combinator comprises Jibran Hutchins, Quan Huynh, Preston Schmittou, and Joseph Tso.
- Jibran Hutchins – Co‑founder. He is affiliated with Carnegie Mellon University (Tepper School of Business) as indicated on his LinkedIn profile and the YC listing. Investor commentary on X highlights his role in initiating Haladir’s pre‑seed round.
- Joseph Tso – Co‑founder. Identified on the YC page as “CS @ Princeton” and his LinkedIn profile confirms a computer‑science background from Princeton University.
- Quan Huynh – Co‑founder and Chief Information Officer. Listed on the YC company page and LinkedIn/X profiles as a founding member, though detailed employment history is not publicly visible.
- Preston Schmittou – Co‑founder. Described on YC as a “Freshman at UVA Wise,” indicating an early‑career contributor.
The team size of four is corroborated by the YC listing and the company’s LinkedIn page, which also enumerates the same founders as employees. Crunchbase confirms the company’s San Francisco location, active status, and pre‑seed funding, further validating the team composition.
The founders collectively bring strong academic credentials (CMU, Princeton) and early‑stage startup experience through Y Combinator, positioning them well to tackle formal‑methods‑driven AI coding tools.
Unlock Full AI Research Report
Enter your email to access the complete analysis.
We'll never spam you. Unsubscribe anytime.