Model Leeching shows how attackers can distill ChatGPT-class task knowledge into smaller models for about fifty dollars, then use them to tune follow on attacks.
Lewis Birch

When people talk about “model stealing,” the discussion often stops at theory. In the real world, defenders want to know something more concrete. Which models are easy to steal, which ones are harder, and what actually happens after an attacker walks away with a stolen copy?
The PINCH research project set out to answer exactly those questions. PINCH is an adversarial extraction framework that automates end to end model stealing campaigns across many different deep learning architectures, datasets, and deployment environments. Instead of focusing on a single handcrafted attack, the team built infrastructure to run hundreds of extraction scenarios and measure how they really behave.
From Mindgard’s perspective, this type of work is essential. If security teams are going to build realistic threat models for AI, they need data, not just intuition. PINCH gives us a repeatable way to see how extraction attacks scale, where they fail, and why “partial” theft can still be good enough for an attacker.
At its core, PINCH is a large automation engine for model stealing. The framework:
Under the hood, PINCH takes advantage of transfer learning and curated deployment repositories to spin up models quickly across domains such as image classification and time series. Once targets are live, it launches extraction attacks that try to recover architecture choices, parameter values, and hyperparameters.
Rather than stop at a few cherry picked examples, we pushed PINCH to evaluate extraction against different model architectures, spanning modern convolutional networks and other deep learning families.
Most prior extraction work has been narrow. A paper might showcase a powerful attack against a specific convolutional network on a specific dataset running on a single GPU stack. That proves feasibility, but it does not tell a CISO whether their own architecture choices or deployment environment make theft more or less likely.
By contrast, PINCH highlights which characteristics of a deep learning system actually move the needle:
This kind of cross cutting view is exactly what operators need when they are deciding which models should sit behind stricter access controls or stricter monitoring.
A key finding from PINCH is that “fully” stealing a model is not a requirement for real risk. Even when an extraction attack only partially recovers the victim’s behavior, the resulting stolen model can still be used to stage other attacks.
We used PINCH to show that partially successful extraction can support model inversion attacks. In other words, an attacker can use a stolen approximation of your model to infer sensitive training data, even if they have not perfectly cloned your architecture or parameters.
For security teams, this changes the threshold for concern. It is not enough to say “our model is complicated, so copying it exactly is hard.” If an adversary can get close enough to reconstruct sensitive inputs or stage more focused exploits, then the damage is already done.
Another important insight is that stolen models can match the target’s accuracy while still looking very different internally. PINCH found cases where two stolen models achieved equivalent performance to the victim, yet their internal learned representations and architectural details diverged.
This has two implications:
For defenders, the takeaway is simple. You should assume that an attacker who can heavily query your model can create a functionally equivalent clone that is good enough for practical misuse, even if the internals do not match line by line.
PINCH provides a realistic lower bound on what an adversary with time, infrastructure, and access can achieve. It shows that:
In Mindgard’s view, this reinforces several best practices for AI security:
PINCH also underscores the value of automated testing. Just as organizations use dynamic application security testing to probe web apps, AI owners will need frameworks that exercise models under realistic attack patterns and surface systemic weaknesses.
The research behind PINCH gives Mindgard a detailed view of how extraction behaves across the stack. It is one more reminder that AI security is not only about prompts and outputs. It is about the full system of architectures, datasets, hardware, and access patterns that adversaries can exploit.