
RunAnywhere
The default way of running On-Device AI at Scale
About
Edge AI is inevitable, but shipping it is painful: every device class behaves differently, runtimes vary, models are huge, and performance collapses under memory/power constraints. RunAnywhere turns that into an enterprise-ready workflow: one SDK to run models on-device, plus a control plane to manage models, enforce policies, and measure outcomes across thousands of devices.
Founders
AI Research Report
Problem & Solution
Problem/Solution Report
Problem: Shipping on‑device AI at production scale is hard. Devices vary (iOS/Android/PC/edge), runtimes and model formats differ (Core ML, MLX, ONNX, GGUF, etc.), models are large and memory/power constrained, and teams need ways to update, observe, and route workloads between device and cloud without sacrificing privacy, cost, or latency. As YC summarizes, “every device class behaves differently, runtimes vary, models are huge, and performance collapses under memory/power constraints.” Teams also need fleet‑level control: policy‑based routing, OTA model updates without app releases, analytics, and fallbacks.
Solution: RunAnywhere provides one SDK for iOS, Android, and other edge form factors with identical APIs and a cloud control plane to manage deployments. The platform supports on‑device runtimes and formats (GGUF/ONNX/CoreML/MLX), ships native mobile runtimes, and introduces policy‑based routing to decide per request whether to run locally or route to cloud, with configurable fallbacks. Its control plane offers OTA model updates (no App Store release needed), usage analytics, and rules/policy enforcement, all designed to keep user data private and minimize latency and cloud costs.
Developer experience and speed to value are emphasized: the homepage highlights a “3‑minute setup,” privacy‑first design, and “add the SDK in less than 5 lines of code,” with real‑time fleet analytics and dashboards. The Product Hunt launch describes RunAnywhere as “Ollama but for mobile, with a cloud fallback,” clarifying the positioning: a production‑grade mobile runtime plus a management/control layer for live apps.
Value proposition: (1) Lower latency and better UX from on‑device inference; (2) privacy by default with data staying on device; (3) improved cost control by avoiding cloud calls except when policies dictate; (4) faster iteration and safer rollout via OTA updates and analytics; and (5) multi‑platform parity (iOS/Android/edge) with one integration surface area.
Market & Competitors
Market and Competitors Report
Market context: On‑device AI is moving from niche to mainstream as AI‑capable smartphones and PCs scale rapidly. Gartner projects 369 M GenAI smartphones (2025) rising to 559 M (2026) and AI PCs at 77.8 M units (2025) and 143.1 M (2026). Grand View Research estimates the overall on‑device AI market at $8.6 B (2024) growing to $36.64 B (2030), with software gaining share. This acceleration reflects user preference for instant responses and privacy, and enterprise desire to control inference costs and reduce dependency on cloud latency/connectivity.
Competitive landscape: RunAnywhere competes across two layers: (a) on‑device runtimes/SDKs and (b) deployment/control planes for edge fleets. On the runtime side, platform‑native stacks (Apple Core ML/MLX, Google/Android NNAPI/Gemini Nano) and open frameworks (e.g., MLC/ggml/ONNX‑based engines) enable local inference but often lack an opinionated, cross‑platform, production‑grade SDK with identical APIs and a baked‑in routing/observability layer for mobile. On the fleet/control side, edge MLOps offerings (e.g., OTA, routing, and analytics at the device level) have been more common in IoT/industrial contexts; RunAnywhere adapts these capabilities to mobile/consumer apps with per‑request hybrid (device/cloud) policy routing, differential OTA model updates, and privacy‑preserving usage analytics.
Peer set: Tracxn classifies RunAnywhere among on‑device AI platform providers and lists emerging competitors such as EdgeRunner AI, EmbeDL, and SandLogic. Platform‑native pathways from Apple, Google, and Qualcomm (Core ML/MLX, Gemini Nano, AI‑capable SoCs/NPUs) also represent indirect competition.
Positioning and advantages: RunAnywhere’s strengths are (1) identical mobile SDK APIs across iOS/Android, (2) per‑request policy‑based routing between device and cloud, (3) OTA model/prompt/rule updates without app releases, and (4) privacy‑first analytics and fleet management. As device adoption scales and software budgets shift toward on‑device capabilities, vendors that pair runtime performance with enterprise‑grade control planes should capture growing share.
Risks: Platform‑native improvements (Apple Intelligence, Gemini Nano) may reduce the perceived need for third‑party SDKs; open‑source runtimes can compress willingness to pay if teams are comfortable operating them; and edge analytics/privacy requirements vary by region. Nonetheless, multi‑platform parity, fleet orchestration, and policy/observability continue to present integration and ops challenges that a specialized SDK + control plane can uniquely address.
Total Addressable Market
Quantitative TAM Report
There is no single “on‑device AI SDK/control‑plane” TAM published by a major analyst, but adjacent, quantifiable markets and device adoption curves provide defensible bounds for RunAnywhere’s opportunity. First, the overall on‑device AI market (hardware + software) was estimated at USD 8.6 billion in 2024 and is projected to reach USD 36.64 billion by 2030 (Grand View Research), implying a ~27.8 % CAGR. Within that, software is a substantial and faster‑growing component (hardware ~60 % share in 2024, software the remainder), suggesting a 2025 on‑device AI software segment on the order of USD 4‑6 billion and rising into the low‑ to mid‑teens of billions by 2030 as software’s share expands.
Second, device‑side adoption provides a powerful top‑down lens for serviceable market sizing. Gartner projects GenAI smartphone units of ~369 million in 2025 and ~559 million in 2026, with end‑user spending on GenAI smartphones reaching ~$298 billion in 2025 and ~$393 billion in 2026. In PCs, Gartner expects AI PCs to represent 31 % of PC shipments in 2025 (77.8 M units) and 55 % in 2026 (143.1 M units), while Canalys projects AI‑capable PCs to exceed 100 million shipments in 2025 and 205 million by 2028.
Third, triangulating an SDK/control‑plane TAM from these device pools: a conservative bottom‑up scenario assumes 2026 GenAI smartphones (~559 M) plus AI PCs (~143 M) for ~702 M new AI‑capable devices. If 2‑5 % of those devices are active monthly users of third‑party apps that integrate a commercial on‑device AI SDK with a lightweight control‑plane fee averaging $0.50‑$2.00 per active device per year, the annualized SDK/control‑plane revenue pool would range from roughly $7 M‑$70 M for that 2026 cohort alone. Extending to a cumulative installed base over multiple years and adding other edge form factors (tablets, wearables, embedded/industrial, automotive) increases the range significantly.
Finally, alternate analyst signals corroborate the growth vector: ComputerWeekly cites the on‑device AI market “topping $10 bn in 2024” and growing to ~$30.6 bn by 2029 (~25 % CAGR), and Counterpoint notes that cumulative GenAI smartphone shipments surpassed 500 million by October 2025. These data points support the view that on‑device AI is transitioning from early to mainstream, expanding the addressable market for tools that make on‑device AI deployable and operable at scale.
Founder Analysis
Founders and Background Report
RunAnywhere was founded by Sanchit Monga and Shubham Malhotra, who are building “the default way of running on‑device AI at scale” as part of Y Combinator’s Winter 2026 batch. The YC company profile lists both as active founders and places the company in San Francisco, reflecting its developer‑tools and enterprise focus around on‑device inference and fleet control at the edge.
Cofounder Sanchit Monga is a hands‑on software engineer and open‑source builder. His personal site and GitHub indicate he previously worked at Intuit on mobile products used by millions (QuickBooks Workforce and QuickBooks Solopreneur/Accounting). He also ships side projects (e.g., Prepend AI, Placem8, AppTrail) and leads the RunAnywhere open‑source SDKs (3.7k+ GitHub stars at the time of capture). Crunchbase also lists him as a founder of RunAnywhere, confirming his leadership role.
Cofounder Shubham Malhotra is an engineer with a formal software engineering background; his LinkedIn shows a Bachelor of Science in Software Engineering from Rochester Institute of Technology. His GitHub profile shows active contributions to RunAnywhere’s SDKs and references his San Francisco location, plus links to a Google Scholar profile, underscoring an engineering/research orientation.
In community launch notes on Product Hunt, the founders describe themselves as “Sanchit and Shubham (AWS/Microsoft)”, suggesting prior experience at those hyperscalers. While the YC profile and company site emphasize their mission and product, these public signals together paint a picture of two product‑minded, mobile‑first engineers who have shipped software used at scale and are now focused on making on‑device AI production‑ready for mobile and other edge form factors.
Unlock Full AI Research Report
Enter your email to access the complete analysis.
We'll never spam you. Unsubscribe anytime.