Eduba Veteran-Owned

Your team has AI tools.
They're not sure what to do with them.

That's not a training problem. It's a translation problem. Someone needs to sit with your people, learn what they actually do all day, and show them exactly where these tools fit. Not in theory. In their workflows, with their data, on their deadlines.

That's what we do.

Jake Van Clief leading an AI training session
AI workflow training, 2025
Pacific Life Colgate-Palmolive KPMG UK IAG University of Edinburgh Correlation One British Council
95%
still using what we taught
after 30 days
~10%
industry average
for AI training adoption
340+
people trained
since May 2025

Most AI training looks the same. Someone shows your team a bunch of demos, everyone nods, and within a week nobody's changed how they work. The tools sit there. The licenses keep billing. The team goes back to what they were doing before.

It's not because the tools are bad. It's because nobody connected them to the actual work. The gap between "look what ChatGPT can do" and "here's how this saves you two hours on Thursday" is where most training falls apart.

01

We learn your workflows first

Before we train anyone, we spend time understanding what your team actually does. What tools they use, where time disappears, what's tedious, what's high-stakes. This is the part most vendors skip.

02

We train on what the audit uncovered

No generic curriculum. Every session is built around the specific opportunities we found. Your team works with their own tasks, their own data, in real time. They leave with workflows they can use the next morning.

03

We build what makes sense to automate

Some things should be automated. Most shouldn't. We help you tell the difference, then build the agents and workflows that actually earn back the investment. You get agent charters and a 30-day action plan.

Pacific Life & Colgate-Palmolive
300+ employees trained on AI workflow integration across multiple departments. Hands-on sessions built around each team's actual processes.
KPMG UK
40+ executives trained on strategic AI decision-making. Focused on when to invest in AI and when the answer is no.
Academy of Intl. Affairs NRW
AI ethics briefing to ambassadors and international affairs experts in Bonn. Presented research on AI governance and soft power measurement.
University of Edinburgh
40+ professors trained on AI integration. Published research on psychometric assessment of large language models.
Jake Van Clief

Jake Van Clief

Founder & CEO

I spent eight years in the Marine Corps working on cryptographic systems and fighter jet avionics, where "close enough" isn't a concept that exists. That turned into a master's degree in governance and AI policy at the University of Edinburgh, which turned into this: helping organizations figure out what to actually do with AI, not just what it can theoretically do.

The thing I've gotten good at is taking the biggest AI skeptics in a room and making their criticisms the most valuable part of the conversation. Healthy skepticism is an asset. We just need to point it at the right questions.

8 Years USMC MSc Future Governance, Edinburgh Veteran-Owned
David McDermott

David McDermott

Chief Architect

David brings over a decade of software engineering experience across defense, gaming, and enterprise systems. He holds an MS in Computer Science from the University of Iowa and has built production systems at BAE Systems, Epic Games, and Derivco.

He architects Eduba's technical infrastructure, including EdubaWare and the agent routing systems that underpin the training platform. David is the person who makes sure the things we teach people to build actually work at scale.

MS Computer Science, Iowa BAE Systems Epic Games

If your team has AI tools they're barely using,
that's the conversation I like having.

No pitch deck. No pressure. Just a conversation about what your team does and whether we can help.

The work speaks
for itself.

From Fortune 500 training floors to academic research labs, these are the engagements that built our reputation. Not because we told people what AI can do, but because we showed them what to do with it.

Training room at Pacific Life with employees at tables
Pacific Life & Colgate-Palmolive
Ongoing
In partnership with Correlation One  ·  May 2025 to present

What started as a training engagement has become something more like embedded consulting. Hundreds of employees across both organizations, from frontline teams all the way to executive leadership, trained through single and multi-day sessions built entirely around their existing workflows.

The sessions didn't just teach people how to use AI tools. They uncovered where automation actually made sense, and where it didn't. Out of the coursework itself, teams began building agents that are projected to save 2,000 to 3,000 hours annually. But the bigger outcome was what nobody planned for: new value-adds beyond just automation, places where AI could improve decision quality, not just speed.

600+
employees trained
and counting
6-9K
projected hours saved
annually from course-built agents
95%
sustained adoption
after 30 days
Workflow Integration Agent Development Executive & Frontline Multi-Day Sessions
Executive workshop with attendees at round tables
KPMG UK
Completed
Executive AI Strategy Workshop  ·  October 2025

Forty-plus senior executives needed something that most AI training doesn't offer: an honest conversation about when to invest in AI and when the answer is no. The engagement was built around strategic decision-making, not tool demos. We focused on helping leadership develop the judgment to evaluate AI investments against real operational needs, cutting through vendor narratives to get to what actually matters.

40+
senior executives trained
1
core question: when
should the answer be no?
Strategic AI Decision-Making Executive Workshop Global Consultancy
Academy of International Affairs NRW workshop group photo in Bonn
Academy of International Affairs NRW
Ongoing
AIA Fellow Workshop  ·  Bonn, Germany

An AI ethics and governance briefing for ambassadors and international affairs experts, focused on how AI tools are reshaping soft power and influence measurement across 25+ countries. The workshop brought together researchers from the British Council, the British Foreign Policy Group, Ben Gurion University, and other institutions to discuss what happens when the tools we use to measure influence start changing the influence itself.

Presented the Ethics Engine research alongside ongoing work with ICR and the British Council on international soft power, connecting the psychometric evaluation of AI models to the broader question of how governments and institutions should think about the tools they're adopting.

AI Ethics International Policy Ambassador Briefing Soft Power
International Airlines Group
Completed
AI Training for Innovation Team  ·  IAG (British Airways, Iberia, Vueling, Aer Lingus parent)

Training delivered to IAG's innovation division, helping one of the world's largest airline groups think through where AI fits within development, thought leadership, and customer experience operations. The kind of environment where getting it wrong isn't just expensive, it's felt by millions of passengers.

“A pleasure working with Jake and his team at Eduba.”

Lissa Pritchard  ·  Head of Innovation: Development, Thought Leadership & Customer Experience, IAG
Innovation Strategy Customer Experience Aviation
Vigilore  /  Armetour
Ongoing
AI Integration for Defense Operations  ·  Full product design & production by Eduba

Working with Armetour's defense and intelligence teams to integrate AI capabilities into operational workflows. This engagement draws on the same zero-defect thinking that comes from working on systems where failure isn't an abstraction. When the stakes are this high, computational orchestration matters: knowing which layer each problem belongs on, whether that's AI, traditional code, human judgment, or a decision not to build it at all.

VigilOre is the system that came out of this work: a multi-agent compliance platform for DRC mining operations. Four specialized agents handle document intake, aggregation, comparison, and reporting across multilingual regulatory frameworks. The entire product, from architecture to the demo video above, was designed and produced by Eduba.

160+
hours of manual compliance
work per cycle, before
<5
minutes to process the same
documents now

“Always ahead of the competition and ready to deliver and exceed expectations.”

Andrew Sdunich  ·  Armetour
Defense Intelligence Multi-Agent Systems Full Product Build
Ethics Engine interface showing psychometric assessment dashboard
The Ethics Engine
Published
Neuroplastics Lab, University of Edinburgh  ·  arxiv.org/abs/2510.11742

Most organizations choose AI tools based on marketing. We built a methodology to evaluate them using validated psychological instruments, the same psychometric frameworks applied to humans, now applied at scale to large language models. The Ethics Engine gives organizations something they've never had before: an empirical, repeatable way to assess the behavioral tendencies of the models they're trusting with their work.

Being able to evaluate these models at all was the hard part. Doing it at scale is what makes it useful. This research was also presented at UKICER 2025 under "AI Workflows: Honest, Transparent, Reusable Work" and formed the basis of the AI governance briefing at the Academy of International Affairs NRW.

Psychometric Assessment LLM Evaluation Published Research UKICER 2025 University of Edinburgh
ICR research roundtable discussion in Bonn
ICR Research
Published
British Council-Commissioned Research  ·  International Soft Power

Applying machine learning and NLP to map how ideas transform as they cross borders, for British Council-commissioned research on international soft power. The work uses AI to do what would take a research team months: trace influence patterns across languages, institutions, and policy frameworks.

The engagement went well enough that ICR made a decision they hadn't made in over a decade: they wanted to produce a hard copy of the study. That's not a metric you find on a dashboard, but it tells you something about the quality of the work. When a research organization breaks a ten-year pattern to put your findings in print, the work did its job.

British Council Soft Power Research NLP & Machine Learning First Hard Copy in 10+ Years
Eduba Labs advertisement at Edinburgh Futures Institute
Flagler College & Edinburgh Futures Institute
2022-2025
Faculty AI Workshops & Academic Research  ·  St. Augustine, FL & Edinburgh, UK

Before the Fortune 500 engagements, before the global consultancies, the work started here. After using AI to produce research that was accepted to Flagler's Saints Academic Review, faculty started asking how it was done. That turned into contracted AI workshops for the college, teaching professors how to think about these tools in the classroom.

It was the first time the tables turned from student to teacher, and it proved something that still holds: the people most skeptical of AI often become its most thoughtful adopters, if someone takes the time to connect it to their actual work. The work continued at the University of Edinburgh, where Eduba Labs ran experimental workshops at the Edinburgh Futures Institute for faculty exploring AI in education.

“Whoa, this is a really good use of AI. How do you do this?”

Faculty response, as reported by Flagler College News
Read the full Flagler College feature →
Higher Education Faculty Training Edinburgh Futures Institute Where It All Started

Your team could be the next one
that actually uses what they learned.

No pitch deck. No pressure. Just a conversation about what your team does and whether we can help.