What I Learned Building the Data Infrastructure Behind a £400m Natural Capital Portfolio Rich Stockdale PhD

Spring 2026

The most valuable model I’ve ever built doesn’t run on GPUs. It runs on soil data, species surveys, LiDAR scans, carbon pricing curves, and five years of proprietary field intelligence across 50,000 acres — and it has already underwritten the UK’s first institutional-scale natural capital portfolio worth more than £400m. If you’re still thinking about large-scale models as a purely digital phenomenon, you’re looking at the wrong landscape.

Everyone in capital markets is talking about large language models. Almost nobody is talking about the models that will actually reshape how we originate, underwrite, and manage an entirely new asset class. We call them Large-Scale Landscape Models — the ecological and financial equivalent of training a neural network on billions of datapoints, except the training data is rooted in soil, streams, species, and the complex interplay between what the land is today and what it could become tomorrow. These are Oxygen Conservation’s trade secrets, and they sit at the foundation of everything we do.

Here’s what most people in this market still get wrong: they treat natural capital like a traditional land play with an environmental overlay, applying conventional rural property valuations to assets whose primary value driver — ecological recovery potential — doesn’t appear anywhere in the Red Book framework. That’s not a minor oversight; it’s a systematic mispricing of an entire emerging asset class, and it’s precisely the kind of market inefficiency that creates asymmetric opportunity for investors willing to build the data infrastructure required to see what others can’t.

The gap between traditional valuations and natural capital valuations is where we operate.

Training Nature’s Model: Why Most Ecological Data Is Useless for Capital Allocation

Landscape modelling itself isn’t new — ecologists and planners have used maps to track habitat and biodiversity for decades. But here’s the critical distinction that separates what we’ve built from what’s come before: traditional ecological models were designed to inform conservation planning, not capital allocation. They were flat, static, and entirely disconnected from the financial modelling frameworks that institutional investors actually use to deploy capital. They answered the question “what lives here?” when the question that matters for investability is “what could this landscape become, over what timeline, at what cost, with what revenue profile, and at what risk?”

At Oxygen Conservation, our proprietary Large-Scale Landscape Models are designed from first principles to answer the second question — the investment question — while maintaining the ecological rigour that underpins credible, verifiable environmental outcomes. They empower our scientists, economists, and conservationists to target acquisitions where we can deliver maximum positive impact for nature, for people, and for investors simultaneously, because those three objectives are not in tension when the data infrastructure is built correctly.

Profit is the result of our work, not its purpose, but it is critical. Without investment-grade returns, conservation cannot scale.

And we are committed to Scaling Conservation.

The 5-Layer Natural Capital Intelligence Stack

The architecture of our landscape models follows what I think of as a five-layer intelligence stack — and understanding this architecture matters, because it explains both why most natural capital ventures remain sub-scale and why the ones that build this infrastructure correctly will compound their advantage over time.

Layer 1 — Open Foundation Data. More than 80 datasets form the backbone: habitat maps, soil classifications, floodplain assessments, environmental designations, topography, and climate projections. Think of this as the pre-training data in an LLM — openly available, essential for baseline capability, but entirely insufficient on its own to generate proprietary insight. Open data allows you to build quickly and at scale, but if open data is all you have, you’re competing on the same information as everyone else. There is no alpha in public datasets.

Layer 2 — Proprietary Field Intelligence. This is where competitive moats form. Using advanced drone technology, we capture LiDAR scans, photogrammetry, and thermal imagery — measuring canopy cover, herbivore numbers and movements, carbon stocks, and hydrology at a resolution far beyond anything available in public datasets. Every drone flight establishes a baseline; each return flight builds a progress report that enriches both our ecological understanding and our financial projections. Over months and years, we track how interventions — rewetting, planting, removing livestock — reshape the land in real time, generating proprietary training data that strengthens our predictive models with every cycle. This is the fine-tuning layer, and it’s extraordinarily difficult to replicate without owning and operating the land yourself.

Layer 3 — Financial and Market Integration. This is the layer that most conservation organisations don’t have and most investors don’t yet demand — but it’s the layer that makes natural capital investable. We integrate carbon pricing curves, biodiversity unit demand data, transaction volumes, rural property market intelligence, credit market dynamics, and ecosystem service valuations alongside our ecological models. This dual integration gives us something critical: the ability to compare traditional land-use returns with the emerging revenues from restoration and regeneration in the same analytical framework, using the same financial language that institutional allocators already speak. IRRs, NPVs, cashflow forecasts — we model them all, because these metrics are not optional decoration; they’re the language of capital allocation, and any natural capital platform that can’t speak this language fluently will remain permanently sub-scale.

Layer 4 — Scenario Simulation Engine. Our models don’t produce single-point forecasts — they simulate multiple scenarios to map how landscapes may evolve across different intervention pathways and time horizons. We can model how a drained peat bog might become a carbon sink again, how an abandoned hill farm might transform into temperate rainforest, or how the removal of sheep might unleash a surge of biodiversity — and critically, we can attach financial outcomes to each scenario with increasing confidence as our proprietary field data deepens the model’s training set. This is where the LLM analogy becomes most precise: just as language models improve through iterative training on better data, our landscape models improve every time we collect new field observations, verify a prediction, or recalibrate a financial assumption against actual market transactions.

Layer 5 — Decision and Execution Layer. Data at this scale risks becoming overwhelming, and a model is only as valuable as the actions it enables. Our data lives as a single source of truth in cloud-based platforms, with front-end dashboards that transform complexity into clarity for every stakeholder — from our internal acquisition team to our institutional investors to the communities living on and around our estates. Conservation decisions have historically been opaque, reserved for technical experts. We’ve deliberately inverted that by presenting analytical outputs through dashboards that build trust, accelerate decisions, and increase transparency across the entire value chain.

This five-layer stack is not a metaphor. It’s the operational architecture behind every acquisition, every nature recovery plan, and every investor report we produce.

The Invisible Investment Case: How Data Stacking Reveals Mispriced Assets

So how do we actually find opportunities that others can’t see?

By laying data over data — systematically, rigorously, across all five layers — until patterns emerge that are invisible to anyone operating on fewer dimensions. Each dataset functions like a transparent slide; alone, it tells part of the story. Combined, they snap into focus, revealing investment opportunities that we can access and realise for our investors in a market where most participants are still operating on two or three data layers at best.

This isn’t cartography. It’s strategic intelligence — a form of machine learning for landscapes that reveals what I can only describe as a hidden algorithm of the land itself. By combining ecological forecasts with financial modelling, we calculate both the market value of land today and the natural capital value it can deliver tomorrow.

The gap between the two is where we act.

This is how we’ve assembled the UK’s first institutional-scale natural capital portfolio across more than 50,000 acres in under five years — including a record carbon pricing deal at £125 per credit that the market said couldn’t be done. Each acquisition is underpinned by rigorous modelling, ensuring that we deliver both measurable ecological impact and credible financial returns. Investors trust us because our models give them visibility into asset performance. Communities trust us because our models show pathways to jobs, housing, and economic opportunity. Nature benefits because our models prioritise habitats where recovery will be fastest and most enduring, aligned to the very purpose of what we do.

People, Place, and Profit: What Gets Modelled Gets Protected

Too often, conservation stops at species counts and carbon tonnes. Our models go further, integrating social and cultural value into the analytical framework because landscapes that ignore people rarely last. We assess potential for renewable energy, ecotourism, community housing, green jobs, and regenerative agriculture. We model visitor flows to balance public access with ecological protection. We map local economic multipliers to ensure that our acquisitions don’t just restore ecosystems — they regenerate the communities that depend on them.

This matters for a reason that goes beyond social responsibility: conservation that cannot support livelihoods, communities, and long-term economic use is fragile by design. By embedding social and economic dynamics into our models from the outset, we build recovery that can be lived with, worked within, and sustained across generations. That’s how conservation moves from short-term success to long-term resilience.

People, place, and profit. Not in sequence. In parallel.

What Most People Get Wrong: Why Traditional Approaches Can’t Scale

Here’s the uncomfortable truth about the natural capital market as it stands today: the overwhelming majority of participants are using valuation frameworks, ecological survey methodologies, and investment structures that were designed for a world in which nature was either a free externality or a philanthropic concern. The Red Book valuation system — still the dominant framework for rural property in the UK — systematically undervalues ecological potential because it was never designed to price it. In practice, this means that the more precious the nature asset, the lower the traditional valuation. Read that again. The system that determines how we price rural land in this country actively penalises the very thing that will generate the most value over the next fifty years.

This is not a theoretical problem. It’s a structural mispricing that directly affects capital allocation, investment returns, and the pace at which conservation can scale. Our Large-Scale Landscape Models exist, in part, to correct this — to provide the data infrastructure that allows natural capital to be priced, compared, and traded with the same rigour that institutional investors already apply to infrastructure, real estate, and private equity.

The market will professionalise or it will remain sub-scale. There is no third option.

The Next Model: Where This Goes from Here

We are only at the beginning, and the trajectory of what’s possible is accelerating faster than most people in this sector appreciate. As AI capabilities mature, we’re exploring how machine learning can refine predictions, automate monitoring, and uncover correlations invisible to human analysts — not as a speculative exercise, but as a direct extension of the five-layer stack we’ve already built and deployed at scale. Blockchain-enabled verification may soon track carbon credits and biodiversity units with a transparency and auditability that current market infrastructure cannot provide. Crowdsourced science — including through platforms like our own Mosaic Earth — could feed live citizen-observer data into our models, enriching them further and creating the kind of network effects that compound data advantages over time.

The next model of conservation looks like this: precision at the centimetre scale, allowing us to place the right intervention in the right location with the kind of specificity that maximises both ecological and financial outcomes. Models that embrace complexity rather than simplifying it away, incorporating species interactions, soil chemistry, microhabitats, and hydrological systems. Digital twins of entire estates that let us test interventions visually and simulate ecological responses before committing capital. And sensor networks — monitoring soil moisture, water quality, pest activity, and temperature — that transform our landholdings into active data networks, enabling faster, smarter, more responsive stewardship.

Nature will become infrastructure. Data will make it investable. We intend to build both.

Closing the Loop

Proprietary Large-Scale Landscape Models are one of our competitive advantages — and they are reshaping natural capital investment itself. They form the bridge between ecology and economy, between what is scientifically possible and what is financially bankable. They allow us to price tomorrow’s natural capital today, to identify mispriced assets that others cannot yet see, and to act with the pace and precision that institutional investors demand.

We’ve already used them to assemble a £400m portfolio across 50,000 acres. Next, we will deploy them to Scale Conservation towards £1 billion AUM — building a diversified, institutional-grade asset class with risk-adjusted returns that compete with infrastructure, real estate, and private equity.

The question isn’t whether natural capital becomes an institutional asset class. It’s whether your data infrastructure can underwrite it at the speed the market is about to demand.


Reflection

I’m probably going to get in trouble for sharing this much. The models described in this article — including the underlying methodologies, data source inputs, processing techniques, and outputs — are proprietary and constitute confidential trade secret information of Oxygen Conservation Limited. Our team has built something genuinely differentiated, and there’s a reasonable argument that sharing the architecture publicly erodes our competitive advantage.

But here’s why we’re doing it anyway: we didn’t start Oxygen Conservation to dominate a niche. We started it to Scale Conservation. And Scaling Conservation means the entire sector needs to professionalise — not just our corner of it. If sharing how we think about data infrastructure helps other operators build better, more investable natural capital platforms, that grows the market for everyone. More institutional capital flowing into nature recovery is not a threat to our business; it’s validation of the thesis we’ve been building towards since day one.

The models are ours. The mission is shared.