Island of Misfit Startups: Part IV (SkyVida)

Amazon Prime Air wants to run a thousand delivery flights a day from a single depot outside Dallas. Walmart’s DroneUp is targeting 1.5 million households from thirty stores. Zipline has done 1.7 million deliveries in Africa and is scaling into the U.S. Wing is flying suburban routes in Virginia. That’s four operators, in one country, doing one thing (delivery), and the coordination problem is already giving the FAA heartburn.

Now add the real estate surveyor. The insurance adjuster. The power line inspector. The bridge guy. The film crew. The agricultural drone. The hobbyist. The public safety drone. The news drone. And the ones nobody’s thought of yet, because that’s how proliferation works.

Let’s do some math on where this lands.

Delivery is the volume driver. Americans receive about 55-60 packages per capita per year today, but that’s the current logistics cost curve talking. When drone delivery drops the marginal cost of moving a 5-pound package across a metro to under a dollar, you unlock a pile of trips that don’t currently happen. Kid’s inhaler left at home. Four bolts from the hardware store. Lunch. Add pharmacy, grocery, food delivery, and the intra-business stuff (hospitals moving lab samples, construction sites moving tools), and a reasonable estimate for a drone-mature metro is about one drone flight leg per person per day.

One leg per capita per day in Dallas-Fort Worth is 8 million flights. Average flight time of 12 minutes means roughly 65,000 drones airborne simultaneously at steady state. But flights aren’t evenly distributed across 24 hours. Concentrate into the 6-hour peak window and you’re looking at 120,000-150,000 simultaneous drones during lunch rush. And those aren’t spread evenly across 9,000 square miles. They’re clustered around commercial districts, residential density, depot locations, and the corridors between them. The effective density in a hot corridor near an Amazon depot is orders of magnitude higher than the metro average.

That’s not a thought experiment. That’s just a lunchtime in Dallas, ten years from now. Heterogeneous missions nobody pre-coordinated, built by different manufacturers, running different firmware, operated by different companies who’ve never talked to each other.

How do they not hit things?

Two answers exist. One: build a bigger air traffic control system for drones (UTM). Two: put sensors on every drone and let it avoid obstacles on its own (onboard DAA). Both are real, both are getting built, and neither solves the problem that emerges when you combine density with heterogeneity. The gap between them is where drones start hitting things.

Let me steelman the incumbents first, so I won’t look lazy. (I’ve been guilty of lazy work. Ask anyone who’s read my code.) UAS Traffic Management (the FAA’s and NASA’s attempt to extend air traffic control downward into low-altitude airspace) is far from stupid. The architecture is layered: strategic deconfliction through UAS Service Suppliers before flight, tactical separation through a mix of ground-based and onboard systems in flight, and collision avoidance as the last resort through onboard detect-and-avoid. The FAA’s own ConOps already envisions drones using onboard DAA equipment plus “procedural rules of the road” to maintain separation. The naive objection that pairwise interactions grow at O(N²) is technically true but practically misleading. Real traffic systems partition: sectorized airspace, stratified altitude, spatially bounded interactions. With dynamic geofencing and corridor segmentation, a well-engineered federated UTM degrades toward O(N log N) or effectively O(N). Europe’s U-Space framework, Japan’s drone corridors over power grids, Singapore’s 5G urban canyon navigation… real work, real infrastructure, scaling further than the pessimists admit.

The layered architecture is correct in outline. The hole is in the middle.

The strategic layer (pre-flight deconfliction through USSs) works. The bottom layer (onboard DAA for individual collision avoidance) works, and the FAA assumes it will exist. But strategic deconfliction can’t help you in real time, and onboard DAA can’t coordinate across agents. Every centrally mediated safety decision in between requires an RF round-trip: telemetry up, processing, command down. On 4G/LTE, 50-200ms one way. On 5G, 10-50ms. Call it 100-300ms round-trip on a good day, worse in congested spectrum or urban canyons. Let’s say, generously, that the processing time is 3 seconds on top of the RF round-trip. Traditional radar updates every 4-12 seconds, which is fine for commercial jets with miles of separation. For small drones closing on each other at combined speeds of 60+ mph with meters of separation, a 3-second-old position fix covers more than 250 feet of closure. In a corridor, that’s a drone punching through the window on the 20th floor.

At the densities we projected for Dallas (150,000 simultaneous drones at peak, concentrated in corridors) the encounters centralized systems can’t resolve in time aren’t edge cases. They’re the encounters at sector boundaries, in handoff zones between service providers, where two drones from different operators converge and neither UTM system knows about the other until the round-trip completes. The GAO’s February 2026 report said the quiet part out loud: there’s no standardized two-way communication between drones and other airspace users, and the FAA has no timeline for building one. A well-partitioned federated UTM handles strategic coordination. It cannot guarantee safety in the tactical window where the round-trip exceeds the collision timeline.

The bottom layer does most of the work. Give every drone good enough onboard sensors and it avoids obstacles independently. The FAA already requires this for BVLOS waivers, and the DAA industry is building it. For the vast majority of encounters in open airspace, independent reactive avoidance is sufficient. Two drones see each other, both maneuver away, nobody dies.

But individual avoidance has failure modes that no amount of better sensors can fix. One is the hallway dance. You’re walking toward someone in a corridor. You step right. They step right. You step left. They step left. On foot, it’s awkward and you both laugh. At 30 mph in a constrained urban corridor with a third drone behind each of you, it’s a cascade. Symmetric geometry plus identical avoidance algorithms produce oscillation, and in dense traffic the oscillation propagates backward through the flow. Multi-vehicle freeway pileup, triggered not by a collision but by two reactive systems failing to break symmetry.

Cooperative intent sharing, even minimal (“I’m going right”), breaks the oscillation. That’s the argument for a coordination layer: not as a replacement for onboard sensing, but as a symmetry-breaker for the degenerate cases where reactive physics alone produces cascade. Everywhere else, reactive physics handles it.

Time to talk flock.

Craig Reynolds proved in 1986 that coherent flocking emerges from three local rules with no central coordinator, and this has been validated on physical drone swarms. The academic work is real. It also mostly assumes homogeneous agents doing the same thing. Not what Dallas airspace looks like when forty drones are doing deliveries, seventeen are surveying roofs, and one poor bastard is trying to film a wedding.

The unsolved problem is decentralized coordination among different robots doing different jobs with different flight envelopes and different priorities, none of which have ever met. The FAA’s architecture has a top layer (strategic deconfliction) and a bottom layer (individual onboard DAA). Between them: nothing standardized. No protocol for how drones negotiate with each other in real time. No certification path for multi-agent coordination logic. No way to prove that forty different manufacturers’ avoidance systems won’t produce cascading failures when they meet in a corridor.

SkyVida

Here’s the answer: push the coordination logic into the flight controller. Not into a cloud service. Not into a UTM provider. Into the firmware, in the airframe, at the edge. Not to replace onboard DAA, but to give it the missing coordination layer that turns individual avoidance into collective safety.

Every drone that flies carries a standardized coordination rule set, an open protocol for local interaction, that governs its behavior relative to neighboring agents, obstacles, and terrain. The rules are computationally cheap and manufacturer-agnostic. Any drone running SkyVida can safely share airspace with any other drone running SkyVida, regardless of who built it or what mission it’s on.

Think of it like the rules of the road. You stay right, you yield at intersections, you maintain following distance. Nobody files a “driving plan” with a central authority for their morning commute. The rules are baked into driver training (or in this case, firmware), and coherent traffic flow emerges from local compliance. The DMV doesn’t coordinate your lane change. You do, because you have eyes and rules.

The Rule Stack

Layer 1: Separation. Hard geometric constraints. Minimum distance from other agents, obstacles, terrain. Non-negotiable physics.

Layer 2: Right-of-way. Priority negotiation based on mission class, battery state, maneuverability, and urgency. The medical drone carrying a defibrillator outranks the roof surveyor. Negotiated locally, peer-to-peer, in milliseconds. This is the symmetry-breaker: two drones converging head-on in a corridor don’t oscillate because the protocol resolves who yields before avoidance maneuvers begin.

The bandwidth requirement for symmetry-breaking is tiny (maybe 8-12 bits of total state) and there are dead-simple ways to transmit that without spectrum allocation or infrastructure. The communication prerequisite is a smaller problem than it looks.

Layer 3: Flow. Soft alignment rules that produce efficient traffic patterns without mandated corridors. Drones heading the same direction in the same area naturally laminate into flow layers, because the local rules make it energetically favorable.

Layer 4: Degraded mode. Sensors fail. Communication drops. GPS gets jammed. Conservative defaults: slow down, increase separation, descend to a safe altitude, find a landing zone. Optical signaling is the last channel to fail, because it requires nothing except LEDs and a camera. Even in degraded mode, a drone broadcasting “I’m impaired” via blink pattern gives its neighbors the information they need to route around it.

Punchline: the SkyVida nav protocol doesn’t replace UTM or onboard DAA. It fills the gap between them. UTM handles the strategic layer: mission planning, airspace allocation, operator authentication, regulatory compliance. Onboard DAA handles the bottom layer: individual obstacle detection and last-resort collision avoidance. SkyVida handles the middle, the entire tactical layer that neither UTM nor onboard DAA can reach: real-time multi-agent coordination, local deconfliction, emergent flow management. UTM goes down? Drones running SkyVida still don’t hit each other. If onboard DAA is all you have, you get the hallway dance at density. If UTM is the only layer, you lose safety the moment the tactical window shrinks below the round-trip latency. The FAA gets to keep its oversight architecture. SkyVida makes it survivable at high (inevitable) density.

(Incidentally, this means DJI’s market position is even more interesting than it looks. They control roughly 70% of the global commercial drone market. If China stands up its own certification authority, Chinese-certified drones and NATO-certified drones have no shared trust layer even if they run compatible protocols. The drones can avoid each other. The insurers can’t underwrite each other. Sleep well.)

The Business of Simulation

TL;DR: SkyVida is the “crash test” authority for UAS, issuing scores by which liability is priced.

Think about how AI benchmarks work. MMLU, HumanEval, SWE-bench: public test suites that anyone can run. Every lab runs them, publishes scores, iterates. The benchmarks become the shared language for “how good is this model.” They’re indispensable. They also get gamed to death. Every lab optimizes against the published tests until the scores stop meaning what they were supposed to mean. The benchmarks that actually matter for deployment decisions are the private evals: undisclosed test sets, novel scenarios, real-world distributions the test-taker has never trained against.

Same structure. But with a twist the AI analogy doesn’t have.

The open benchmark suite is the foundation: reference protocol, closed-form physics models (sensor, turbulence, propulsor, aerodynamic interaction), a reference flight decision module, standardized test configurations, and a base simulator that ties it all together into a working development and benchmarking environment. Manufacturers fork it, run their tactical stack against the published scenarios, get scores, iterate, publish results, compare. Community contributes scenarios, physics models, and reference implementations. The benchmark suite becomes the shared language for “how safe is this drone at density.” All of this is genuinely useful software, not a demo. The community owns it.

The adoption path matters. A free toolkit that makes your product safer is something an engineering team pulls down on Tuesday. Nobody has to extract their IP, hand anything to a startup, or get legal approval: they run their own code in their own environment against published test configurations. The benchmark suite spreads because it’s useful, the way pytest spreads because it’s useful.

None of that is the business, and none of it is sufficient for certification.

The open benchmarks are software-in-the-loop. Your tactical stack runs on the sim server, against simulated physics, and the sim reports how your logic performed. This is fine for development. It is not fine for proving your drone won’t kill someone. A tactical stack that makes correct decisions in 2ms on a cloud instance might take 15ms on the actual STM32 in the flight controller, because the compute budget is completely different. RTOS scheduling jitter, bus contention between sensor drivers, thermal throttling at altitude, memory constraints forcing different algorithm paths. The sim tests the logic, not the logic running on the hardware that actually flies. The timing failures that produce collisions live in exactly that gap.

The automotive industry solved this decades ago. Hardware-in-the-loop testing. The sim runs the physics and the scenario on the server. The actual flight controller board sits in the loop, receiving simulated sensor data over its real interfaces and returning real commands at real timing on its real processor. You test the decision logic and the hardware it executes on, together, under realistic time pressure. A manufacturer whose logic passes every software benchmark but chokes on the actual chip under thermal load fails the HIL test. That’s the test that matters.

So the business has three layers.

Layer one: open benchmark suite. Software-in-the-loop. Free. Development tool. Adoption flywheel. The community owns it.

Layer two: private eval scenarios. Same sim architecture, but with undisclosed test configurations. Real operational telemetry from real corridors, contributed by operators under data agreements. Edge cases accumulated across years of certification engagements. The scenarios manufacturers have never seen, the same way KesslerGym’s Rot Engine uses undisclosed failure modes to prevent overfitting. If students have the answer key, the test is worthless.

Layer three: HIL certification. A physical testing facility where the manufacturer’s actual flight controller board runs against the private eval scenarios in real time, with standardized sensor injection rigs, calibrated timing measurement, and controlled thermal and RF conditions. This is the test an insurer trusts, and it’s what “SkyVida-certified” means.

The open benchmarks are the curriculum. The private eval scenarios are the exam questions nobody’s seen. The HIL facility is the proctored testing center where you take the exam on the hardware that actually ships.

What’s a flight decision module? The whole tactical stack. Onboard DAA, coordination protocol implementation, maneuver planning, degraded-mode behavior, sensor fusion. Everything between “here’s what the world looks like right now” and “here’s what the drone does next.” Two manufacturers can implement the same protocol and get very different scores, because a quadcopter with a $40 camera and a delivery hexacopter with LiDAR and ADS-B In make different tradeoffs in the same corridor, on different chips, under different thermal envelopes.

The call contract. For benchmark scores to mean anything across all three layers, every implementation has to be tested against the same interface: world state in, flight decision out. SkyVida defines and publishes that interface. The inputs are what a real airframe’s sensors would deliver, filtered through realistic noise models (or, in HIL, injected through actual sensor interfaces). The outputs are what a real flight controller would execute. SkyVida’s proprietary interface spec, not a community artifact; the community doesn’t get to vote on the call contract. That’s the universal socket, and SkyVida owns it.

Anyone can fork a software sim. However… a HIL testing facility with standardized sensor injection rigs for dozens of different flight controller boards, calibrated to aviation-grade measurement standards, with undisclosed scenario libraries built from real operational data, is a capital-intensive physical operation that a competitor can’t clone on GitHub. This is closer to a wind tunnel than a software platform. This is a very real moat (touching lots of atoms).

The automotive HIL testing market is worth roughly $740 million annually and growing. Over 500 OEMs and Tier-1 suppliers use real-time ECU simulation platforms, and a full automotive HIL test bench runs north of $500,000 per installation. The economics favor third-party testing: roughly 70% of automotive Tier-2 suppliers use outside HIL providers rather than building their own rigs. That’s SkyVida’s model. You don’t need every airframe maker to buy a bench. You need one facility that tests everyone’s boards. Flight controllers are smaller, cheaper hardware than automotive ECUs, running on simpler buses, which means lower per-bench cost. Scenario complexity runs higher (three-dimensional, multi-agent, weather-dependent), so the software investment is larger. The net: the physical facility is fundable, the third-party testing model is proven, and the capital requirement is the moat.

So, the elevator pitch: “FICO score” for drone accident risk. Tested on the actual hardware, against scenarios you’ve never seen, in a facility you don’t control.

What hasn’t been built is… any of it! And the canonical target is ambitious: 100,000 agents with continuous physics-based interactions, different missions, wind, turbulence, sensor failures, GPS denial, non-cooperative obstacles, and firmware-level decision logic in a realistic urban canyon. Building the open benchmark suite on open-source primitives is the most capital-efficient first step because it requires no hardware, no regulatory approval, and no airspace. The HIL facility comes later, funded by early revenue from consulting and private eval licensing.

NASA’s already simulated thousands of missions in the Dallas-Fort Worth area. The SkyNetUAM platform has run 100,000 missions per day and measured scheduling latency, state update performance, and throughput stability. Their finding: centralized scheduling works at 5,000 concurrent missions, but real-time execution must be event-driven and localized. That result validates the architecture. Centralized for planning. Edge-encoded for execution.

All the revenue flows from the score. If the private eval and HIL certification can demonstrate (and real-world claims eventually confirm) that a 90+ SkyVida score correlates with a significant drop in collision claims, insurers price that discount regardless of what any regulator recognizes. UL didn’t wait for the government to mandate appliance testing. Insurers noticed UL-tested products burned down fewer houses, and they priced the difference. The regulatory framework caught up to the insurance market, not the other way around.

The sequence: government research contract funds the sim, the sim becomes the open benchmark suite, open benchmarks gain adoption, manufacturers start publishing software-in-the-loop scores, early private eval customers pay for undisclosed-scenario testing, the HIL facility opens and produces hardware-validated scores, insurers notice that HIL scores correlate with safety outcomes, insurers start requiring SkyVida certification for dense metro coverage, airframe makers certify because they need the coverage, the FAA notices the entire industry is already using it, regulatory recognition follows. The insurer is the regulatory pathway. The FAA is the caboose. But the FAA is also the first check.

Four customer segments, in order of when they pay:

The FAA and NASA. They need a 100,000-agent simulation to stress-test UTM at density, and they don’t have one. NASA’s SkyNetUAM tested scheduling at 100,000 missions per day, but that’s logistics, not physics-based multi-agent interaction with realistic sensor models and heterogeneous tactical stacks. The FAA cannot validate its own three-layer ConOps without a sim that models what happens when the strategic layer’s pre-flight plan meets ten thousand real-time deviations simultaneously. An SBIR or research contract to build that sim funds the core infrastructure. The sim SkyVida builds to test others’ tactical stacks is the same sim the FAA uses to test its own architecture. One codebase, two customers, and the government pays first.

Insurers. They move faster than regulators and they have a simpler question: what’s the probability this thing causes a loss? A SkyVida score is a number backed by a methodology: 500,000 dense-corridor scenarios, run on the actual flight controller hardware, tested by an independent evaluator against undisclosed scenarios. That’s something an actuary can price. Once one major aviation insurer requires it for dense metro operations, every operator needs the score, which means every airframe maker needs to get certified. Insurers are the forcing function.

Airframe makers. During development they run the open benchmarks internally and iterate until they’re competitive. For certification they submit their board to the HIL facility. Their IP stays protected: SkyVida tests the board as a black box, measures what comes out, and never sees the source. They get a number that their customers’ insurers will accept.

Regulators. The FAA wants to say yes to dense operations. They need cover. “The operator’s fleet is SkyVida-certified” becomes the get-out-of-jail card for the official who approved the license. Shadow standards first. “We recommend…” language that hardens into requirements.

The consulting practice serves cities and nations designing drone airspace frameworks, bridging UTM policy and edge coordination physics. Small market, high value, additional near-term revenue alongside the government research contracts. Export controls and ITAR review apply to the HIL facility’s operational datasets and certification infrastructure, not to the open-source package. The protocol is for everyone. The score is the business.

Risks and Rivals

SkyVida doesn’t launch into a vacuum. The UTM providers (Wing’s OpenSky, ANRA Technologies) are building the managed airspace layer the FAA is already designing around. SkyVida is complementary, not competitive: their architecture doesn’t survive the tactical window at density, and SkyVida is the layer that makes it survivable. The insurer pathway operates independently of whether the FAA ever formally recognizes edge coordination.

Applied Intuition is the scariest name in the space. They built simulation credibility in autonomous vehicles, they’re moving into aerospace, and they have the capital to build a competing sim faster than any startup. Two counters. First, specialization: Applied Intuition’s generalist platform doesn’t accumulate the same edge-case library for heterogeneous drones in crowded urban airspace. That counter has killed a lot of startups who weren’t as special as they thought they were. Second, the HIL facility is a physical moat Applied Intuition hasn’t built for drones. Their automotive HIL expertise doesn’t trivially transfer to the flight controller ecosystem, which runs on different hardware, different RTOS architectures, and different sensor interfaces. But if they decide to build it, they have the capital and the institutional relationships to move fast.

The AI benchmark analogy cuts both ways. Public benchmarks get gamed. If the open benchmark suite becomes the shared standard, manufacturers will overfit to it, the same way every LLM lab overfits to MMLU. That’s a feature for the business model (it’s why the private eval and HIL layers exist) but a risk for the community’s trust in the open benchmarks. If published scores become meaningless, the pipeline from open benchmarks to certification adoption weakens. SkyVida needs to keep the open benchmarks honest enough to be useful while keeping the private eval hard enough to be credible. That’s a curation problem, and it requires ongoing investment in scenario design and physics fidelity that a small team may not sustain.

ASTM F38 coordination logic, MavLink extensions, proprietary stacks: none of them matter if you own the socket. The call contract is protocol-agnostic by design. Package your tactical stack against the interface, run the benchmarks. But here’s the risk: someone else publishes a competing benchmark suite first. The interface definition that wins is the one with the most forks, not the best physics. First-mover advantage in developer tooling is real, and SkyVida needs the open benchmark suite to be genuinely best-in-class to survive a well-funded competitor publishing their own.

The HIL facility itself is a risk. It’s capital-intensive, it requires physical space and specialized equipment, and it doesn’t produce revenue until manufacturers have a reason to certify, which doesn’t happen until insurers require the score, which doesn’t happen until the score correlates with outcomes, which requires drones flying at density and occasionally crashing. The flywheel needs real-world failure data to spin up. Government research contracts, consulting, and private eval licensing bridge the gap, but the path from “FAA-funded simulation tool” to “revenue-generating HIL facility” is long and requires the sim to prove its value before anyone will fund the physical plant.

The fragilities are real. Standardization is a political and geopolitical problem, and the only realistic forcing function is FAA rulemaking, which moves at the speed of lobbying. The academic foundations for heterogeneous coordination are thin and early. Emergent behavior is hard to certify because regulators want deterministic guarantees and local-rule coordination is probabilistic, which is why the sim is the first thing to build. And the whole operation requires credibility the founders may not have: real operational telemetry, institutional trust, and a realistic path through partnership with NASA, MITRE, or a national lab. Solo founders building this in a garage won’t get the FAA’s attention. Solo founders with a NASA SBIR to build the sim the FAA needs, or with Allianz as their first certification customer, might.

Why I’m Not Building This

I’ve never written firmware for a flight controller, my understanding of multi-rotor aerodynamics is enthusiastic but theoretical, and I don’t have relationships at the FAA or with the operators scaling delivery. And I’m busy. My wife pointed out I’ve published twelve essays in two weeks. “That seems like a lot.” She’s not wrong.

But the architecture is sound. The problem is real and getting realer by the quarter. In outline, the FAA’s layered approach is correct: strategic deconfliction at the top, onboard DAA at the bottom. Between those layers, where heterogeneous drones negotiate with each other in real time, nothing standardized exists. The alternative (edge-encoded coordination with a standardized open protocol, validated through an open benchmark suite, certified through hardware-in-the-loop testing) is technically feasible, eminently simulatable, and sitting there waiting.

The person who builds this needs embedded systems chops, enough aeronautics to get the physics right, the political stamina to drive an open standard through an industry that would rather build silos, and enough capital to build a simulation and HIL testing operation that regulators and insurers trust.

If that’s you, the window is now. The mega-operators are scaling. Part 108 is coming. The density is growing. Every drone flying today is coordinating through proprietary systems that can’t talk to each other, or not coordinating at all.

Someone should build this before the sky gets too crowded.



This is Part IV of “The Island of Misfit Startups.” Part I was LensReader, on fixing the thermodynamics of attention. Part II was KesslerGym, on training autonomous systems with messy reality. Part III was Mini Cricket, on edge AI for individual threat detection. The series explores startup architectures built on uncomfortable truths, for problems I can diagnose but won’t build myself. If you have the domain expertise and want to run with SkyVida, find me on LinkedIn or something.