The Hidden Cost of AI Safety Nets

Author: Denis Avetisyan


As artificial intelligence increasingly assists in critical safety engineering tasks, a subtle degradation of human reasoning can occur – a phenomenon we call the ‘Competence Shadow’.

Human-AI collaboration for safety analysis manifests in four distinct structures, each defined by unique patterns of information flow and the implementation of active shadow mechanisms to ensure robust oversight.
Human-AI collaboration for safety analysis manifests in four distinct structures, each defined by unique patterns of information flow and the implementation of active shadow mechanisms to ensure robust oversight.

This review introduces the concept of the Competence Shadow and proposes structure-aware collaboration models to maximize the benefits of AI assistance while mitigating cognitive risks in hazard analysis.

While artificial intelligence promises to enhance safety engineering, a critical paradox emerges: assistance can subtly introduce systematic blind spots, potentially exacerbating rather than mitigating risk. This paper, ‘The Competence Shadow: Theory and Bounds of AI Assistance in Safety Engineering’, develops a formal framework revealing how AI-driven analysis can narrow human reasoning – termed the ‘competence shadow’ – and derives quantifiable performance bounds for various human-AI collaboration structures. The central finding is that AI’s impact hinges not on its capabilities, but on how it is integrated into workflows, demonstrating that degradation compounds multiplicatively with unchecked reliance. Can a shift from evaluating AI tools in isolation to qualifying collaborative structures unlock truly trustworthy Physical AI systems?


Navigating the Promise and Peril of AI-Driven Safety Assessments

The integration of artificial intelligence into safety analysis is generating substantial optimism regarding streamlined certification processes. Proponents suggest that AI-driven systems can automate many traditionally manual tasks – such as reviewing documentation, identifying potential hazards, and verifying compliance with standards – leading to projected efficiency gains of 60 to 80 percent. This acceleration stems from the AI’s capacity to rapidly process vast quantities of data, detect patterns, and flag anomalies with a speed and consistency exceeding human capabilities. Consequently, industries are actively exploring these technologies to reduce costs, shorten development cycles, and ultimately, bring safer products to market more quickly, though careful consideration of potential limitations remains crucial.

A critical challenge arises when safety analysis increasingly depends on artificial intelligence: the emergence of what is termed the ‘Competence Shadow’. This phenomenon describes a systematic reduction in the scope of hazard identification, stemming from the inherent limitations of even the most advanced AI systems. While promising significant efficiency gains, over-reliance on AI can inadvertently constrain the analytical process to patterns and scenarios present within its training data. This means potentially novel hazards, or those existing at the fringes of typical operational conditions, may be consistently overlooked. The Competence Shadow isn’t simply a matter of missed errors; it represents a narrowing of the very skillset-the breadth of critical thinking and anticipatory reasoning-needed to proactively ensure safety, creating a paradoxical vulnerability despite the appearance of thorough automated analysis.

Large language models, despite their demonstrated capabilities, are fundamentally limited by the scope of data used during their training. This creates a ‘competence shadow’ in AI-powered safety analysis, where the system’s hazard identification is constrained by previously encountered patterns. While adept at recognizing and flagging known risks, these models struggle with genuinely novel situations or ‘edge-case’ hazards – those falling outside the distribution of their training data. Consequently, a reliance on LLMs without supplemental human oversight risks a systematic narrowing of the safety assessment, potentially missing critical, previously unforeseen dangers. This isn’t a failure of intelligence, but rather a consequence of operating within defined boundaries, highlighting the need for a hybrid approach that leverages AI’s efficiency while retaining human intuition and the capacity for creative problem-solving.

A waterfall analysis reveals that compounding mechanisms-including scope framing, attention allocation, confidence asymmetry, and time compression-sequentially degrade performance from an idealized baseline of 0.948 to a final quality score of 0.680, representing a 20% reduction from human-level performance.
A waterfall analysis reveals that compounding mechanisms-including scope framing, attention allocation, confidence asymmetry, and time compression-sequentially degrade performance from an idealized baseline of 0.948 to a final quality score of 0.680, representing a 20% reduction from human-level performance.

Unveiling the Sources of the Competence Shadow: Cognitive and Systemic Limits

Automation Bias and Attention Allocation Bias significantly contribute to the competence shadow by altering human analytical processes. Automation Bias manifests as a demonstrated preference for suggestions generated by AI systems, even when those suggestions conflict with readily available, contradictory evidence. This preference isn’t necessarily due to a perceived correctness of the AI’s output, but rather a cognitive shortcut favoring machine-generated information. Complementing this, Attention Allocation Bias leads analysts to prioritize confirming AI outputs rather than independently exploring alternative possibilities or conducting more comprehensive investigations. This biased allocation of cognitive resources reduces the scope of analysis and increases the risk of overlooking critical information not identified by the AI system, effectively narrowing the considered hazard space.

Scope Framing and Confidence Asymmetry represent systemic limitations impacting comprehensive risk assessment. Scope Framing occurs because AI analysis operates based on a pre-defined, implicit ontology – a specific categorization of concepts and relationships – which inherently restricts the scope of considered hazards to those recognized within that framework. Confidence Asymmetry arises from a tendency to prioritize hazards identified by AI systems, even when those hazards are statistically less likely or lack corroborating evidence, while simultaneously downplaying or ignoring risks falling outside the AI’s identified scope. This combined effect results in a narrowed hazard space, reducing the overall effectiveness of risk mitigation efforts by limiting the consideration of potentially critical, yet unflagged, threats.

Perceived efficiency gains from automation can lead to Time Compression, wherein analytical processes are accelerated to the detriment of thoroughness; this reduction in dedicated analysis time directly correlates with increased error rates and missed anomalies. Simultaneously, Automation Complacency-a decreased level of critical oversight stemming from reliance on automated systems-diminishes operator vigilance and the proactive identification of potential issues. These effects are often synergistic, as compressed timelines further exacerbate reduced vigilance, creating a feedback loop that diminishes the overall quality of analysis and increases the risk of undetected hazards.

Structuring Collaboration for Robust Analysis: Amplifying Human Expertise

Current approaches to AI-assisted safety analysis demonstrate that simply adding AI tools is insufficient for robust mitigation of potential risks; significant gains in analysis quality are achieved through specific collaborative structures. Analyses employing Independent Analysis (π2), Tool Augmentation (π3), and Human-Initiated Exploration (π4) consistently yielded quality scores exceeding 88%, in contrast to the 68% achieved by Serial Dependency (π1). This data highlights that the structure of collaboration – how humans and AI interact – is a primary driver of analytical effectiveness, emphasizing the need to move beyond basic AI assistance and actively amplify human expertise within the analysis process.

Independent Analysis (π2), Tool Augmentation (π3), and Human-Initiated Exploration (π4) represent collaborative approaches to safety analysis that demonstrably outperform methods reliant on serial dependency. Quantitative evaluation reveals that Independent Analysis achieves a safety analysis quality of 88%, while both Tool Augmentation and Human-Initiated Exploration attain a higher quality score of 89.8%. These results indicate a significant correlation between the chosen collaboration structure and the resulting efficacy of the safety analysis process, suggesting that distributing analytical tasks and leveraging combined expertise yields more robust outcomes than sequential, dependent reviews.

Analysis employing a Serial Dependency (π1) collaboration structure yielded a safety analysis quality of 68%, significantly lower than other tested methods. This approach involves a single analyst reviewing the output of an AI system without independent verification. The comparatively low quality score demonstrates that relying solely on AI output, without subsequent human oversight from a separate, independent analyst, introduces substantial risk and negatively impacts the reliability of safety assessments. This outcome highlights the critical importance of collaboration structures that incorporate independent analysis to mitigate potential errors or biases introduced by AI systems.

The Non-Degradation Condition is a critical requirement for safe AI integration in safety analysis workflows. This condition stipulates that the introduction of AI must not reduce the overall quality of the analysis. Specifically, in Serial Dependency (π1) – where analysis relies heavily on sequential human-AI interaction – maintaining quality necessitates that the AI’s capability (qAI) meets or exceeds a defined threshold. This threshold is calculated as qAI ≥ qh(1-αeff<i>γ)/(1-αeff</i>γ*qh), where qh represents human capability, αeff is the effective error rate, and γ is a parameter representing the coverage of AI assistance. Failure to meet this condition in π1 results in a demonstrable decrease in analysis quality, highlighting the importance of ensuring AI competence matches or surpasses human performance in sequentially dependent workflows.

A high-performing team benefits from complementary skillsets-specifically, expertise in standards and judgment, domain knowledge, and operational experience-as illustrated by the combined coverage achieved when these individual competencies are integrated.
A high-performing team benefits from complementary skillsets-specifically, expertise in standards and judgment, domain knowledge, and operational experience-as illustrated by the combined coverage achieved when these individual competencies are integrated.

Towards Human-Centric Safety Intelligence: Beyond Automation

Human-centric safety intelligence signifies a fundamental shift in how artificial intelligence is applied to complex systems, moving beyond automation to focus on enhancing human capabilities. This approach doesn’t seek to replace the expertise of safety engineers, but rather to provide them with intelligent tools that extend their analytical reach and reduce cognitive load. By recognizing that nuanced judgment, contextual understanding, and ethical considerations remain firmly within the human domain, this framework champions a collaborative partnership between people and machines. The aim is not to create autonomous safety systems, but to empower human experts with AI-driven insights, allowing them to make more informed decisions, identify subtle risks, and ultimately build safer, more resilient technologies. This principled integration promises to unlock substantial improvements in safety outcomes while preserving the critical role of human oversight and accountability.

A robust approach to integrating artificial intelligence into safety-critical systems necessitates a detailed understanding of what constitutes true competence in safety engineering. This isn’t simply about possessing technical knowledge, but a multifaceted skillset rigorously defined by frameworks such as the Five-Dimensional Competence Framework. This framework identifies competence not only in technical capabilities, but also in critical thinking, situational awareness, communication, and professional conduct. By formally characterizing these dimensions, researchers can move beyond subjective assessments and develop quantifiable metrics for evaluating both human and artificial intelligence contributions to safety analysis. Such a structured understanding is crucial for identifying gaps – the ‘competence shadow’ – where AI can genuinely augment human expertise, and for ensuring that AI-driven insights are reliably integrated into the broader safety assessment process.

The pursuit of enhanced safety analysis hinges on a strategic reconciliation of human expertise and artificial intelligence, specifically addressing the phenomenon known as the competence shadow – the gap between possessed knowledge and its effective application. Rather than striving for full automation, a more fruitful approach involves actively minimizing this shadow through tools that augment human capabilities. By offloading repetitive tasks and providing rapid access to relevant data, artificial intelligence frees human analysts to focus on complex problem-solving, critical thinking, and nuanced judgment-areas where humans demonstrably excel. This synergistic collaboration not only improves the quality of safety assessments by reducing errors and biases, but also dramatically increases efficiency, allowing for more thorough analyses within constrained timeframes and resources. Ultimately, this human-centric approach promises a substantial leap forward in proactive hazard identification and risk mitigation, exceeding the capabilities of either system operating in isolation.

The study highlights a crucial point: simply adding AI to a workflow doesn’t guarantee improvement, and can even introduce hidden vulnerabilities. This echoes Brian Kernighan’s observation that “Debugging is twice as hard as writing the code in the first place. Therefore, if you write the code as cleverly as possible, you are, by definition, not smart enough to debug it.” The ‘Competence Shadow’ reveals a similar dynamic; an overly complex AI assistance system, however ingenious, can obscure the underlying reasoning process, making it difficult to identify errors. The paper’s emphasis on structure-aware evaluation isn’t just about assessing AI capability, but about ensuring the entire system-human and AI-remains comprehensible and maintainable. If the system survives on duct tape, it’s probably overengineered, and the same applies to safety-critical AI assistance.

Beyond the Horizon

The concept of the Competence Shadow suggests a fundamental, if somewhat unsettling, truth: simply adding intelligence to a system does not guarantee improved performance. Indeed, it invites a re-evaluation of how one assesses collaborative structures. The field has, for too long, focused on maximizing AI capability while treating human cognition as a static element. A more nuanced approach acknowledges that competence is not a fixed resource, but a landscape shaped by the tools at hand. The challenge, then, isn’t merely building ‘smarter’ AI, but designing systems where human reasoning is sustained, not subtly eroded.

Future work must move beyond evaluating AI assistance in isolation. The true metric of success will not be hazard identification rate, but the resilience of the overall analytical structure. Consider a city’s infrastructure: one does not demolish entire blocks to upgrade the plumbing. Instead, one refits and reinforces, preserving the essential framework. Similarly, AI integration in safety engineering demands a structure-aware evaluation, one that prioritizes the evolution of existing cognitive processes rather than their wholesale replacement.

The limitations of current methodologies are clear. The field requires new tools to map these ‘shadows’-to visualize how assistance subtly alters the cognitive load and biases of human analysts. Only then can one begin to design truly robust, and truly safe, collaborative systems. The pursuit of intelligent assistance must, paradoxically, focus less on intelligence itself and more on the preservation of human judgment.


Original article: https://arxiv.org/pdf/2603.25197.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-03-28 20:44