In recent years, the capabilities, availability and acceptance of artificial intelligence (AI) have transformed products and markets beyond recognition. The integration of AI features in software or software controlled products has become essential for growth, profitability and company evaluations. However, this pressure is causing companies to adopt AI without building up the necessary expertise or preparing the organisation for the specifics of AI research and development (R&D) and maintenance.
The result is AI software products that are not fully developed or understood by the organisation, creating new risks (software acting like a black box) for the company and thus for potential investors. Especially within M&A and private equity transactions, a software due diligence of LLM- and GenAI-driven products— must evolve rapidly to address these still rather new and unique technology risks.
Traditional software due diligence focuses on code quality and architecture, automated R&D processes, people/team capabilities, and the scalability of the team and software, as well as its documentation. In part 5 of the blog series on "Software Due Diligence: The Key To Successful M&A Deals", we explore how due diligence should be conducted for AI-based software. We focus on technological areas (e.g. the AI stack and practices) that require additional attention, expertise, and tools.
Our insights aim to make investors aware of the additional risks associated with the development and use of AI features in software-based businesses. While AI features for sure drive the company valuation, one should be aware that it will also increase the levels of potential risk.
Before diving into AI software due diligence, it’s worth taking a step back to understand the different patterns of AI integration found in today’s software products. Not every “AI-powered” solution works the same way – and those differences have major implications for valuation, risk assessment, and scalability.
Broadly speaking, there are three main categories:
This category includes products and services that rely on custom-trained, in-house machine learning models. Their key characteristics are:
Takeaway: Proprietary ML systems are technically complex and costly to maintain, but they provide long-term defensibility through owned IP, data assets, and differentiation.
In this concept, companies integrate external Large Language Models (LLMs) via APIs from providers such as OpenAI, Anthropic, or Google Gemini. Common traits are:
Takeaway: Ideal for fast innovation and prototyping, but with clear dependency, compliance, and data privacy risks due to limited control over model behavior and data handling.
These models are not built from scratch, nor are they simply consumed via a static API.
Instead, they are fine-tuned versions of existing foundation models (LLMs/GenAI), trained further with proprietary data to fit specific use cases. Examples include:
Takeaway: This approach blends flexibility and customization with higher technical investment and ongoing model management needs.
For investors and technology assessors, understanding which AI integration pattern a target company uses is critical.
These differences call for distinct due diligence lenses – both when assessing technical robustness and when estimating long-term business value.
The first step in any diligence process involving LLM/GenAI-Driven Products is to determine how central the AI components are to the product’s core value proposition. There’s a major difference between using a generative model for auto-generating blog summaries and augmenting workflows with GPT-driven assistants or applications completely built on retrieval-augmented generation (RAG)..
Key questions to assess are the following:
Answering these questions helps establish whether the AI capability is a strategic differentiator — or an implementation detail that can be replaced or removed with minimal impact.
Large Language Models (LLMs) and generative AI systems are inherently non-deterministic — identical inputs may yield divergent outputs. This probabilistic behavior introduces operational and compliance risks that traditional QA frameworks cannot fully capture.
Due diligence must therefore assess observability, control, and resilience in real-world operation:
For products operating in regulated sectors such as finance, legal, healthcare, or defense, the evidentiary standard for monitoring, documentation, and fall-back management must be significantly higher.
LLM-driven software requires specialized expertise that extends beyond traditional ML or DevOps roles. A sustainable AI organization demonstrates depth in:
Key due diligence questions include:
In early-stage startups, key-person dependency is often the largest operational risk if know-how is undocumented or non-transferable.
AI systems expand the attack surface well beyond traditional software security. Due diligence must evaluate defensive depth against new vectors, including:
A robust security posture combines preventive controls, active monitoring, and clear incident-response playbooks for AI-specific threats.
Modern AI diligence extends beyond performance to ethical accountability.
Regulators and investors increasingly expect proactive governance around bias and transparency. Due diligence should confirm:
Failure to evidence these controls may expose the acquirer to reputational and regulatory risk — even when technical performance is strong.
Generative AI blurs the line between deterministic code and probabilistic reasoning.
This evolution requires a new form of software due diligence — one that combines technical depth, regulatory literacy, and ethical awareness.
For investors, the objective is not only to identify technology risks, but to recognize capabilities that create defensible value: robust data provenance, mature MLOps, explainable model behavior, and credible governance.
When executed rigorously, AI-focused due diligence becomes more than risk management — it becomes the foundation for confident investment, responsible innovation, and long-term competitive advantage in the AI-driven economy.
Act Now: Integrate AI Expertise Into Your Transaction Evaluation
Whether you're acquiring a tech company or investing in a data-driven business model – if AI plays a role, you need new criteria, new methods, and partners with hands-on experience.
👉 Cape of Good Code combines deep technology analysis with AI-specific expertise – delivering clear insights on the viability, scalability, and sustainability of your target technology in no time.
📞 Schedule a non-binding initial consultation or request our AI Software Due Diligence.