top of page
検索

The Limits of Medical AI in 2026 Are Not About Performance but About Admissibility Conditions — ADIC as an Implementation Candidate

As of 2026, the debate surrounding medical AI has moved beyond simple performance evaluations of "how smart the AI is." What clinical practice and regulatory systems are truly questioning is the operational framework: "Is it safe to admit this AI's output into clinical use?", "Who holds the responsibility?", "Can it be immediately stopped if deemed dangerous?", and "Can the decision-making process be audited later?"

The WHO (World Health Organization) strongly demands the assurance of safety, accountability, and human oversight in its guidance on the medical application of generative AI [1]. The U.S. FDA (Food and Drug Administration) also requires not only "submissions for evaluating safety and effectiveness" but also risk management across the Total Product Life Cycle (TPLC) for AI-enabled medical devices (SaMD) [2].

In other words, the limit of medical AI manifests itself not primarily in the model's intelligence, but in the unfixed nature of its operational admissibility conditions.



1. Where Medical AI is Stalled in 2026

The limitations currently facing medical AI can be categorized into the following four points.

1-1. High performance does not mean immediate usability

What is required in clinical settings is not localized high accuracy in a single task, but safety and stability during continuous operation. As evidenced by the FDA's premise of risk management across the entire lifecycle (TPLC) for AI-enabled devices, high initial model performance is merely the starting point for implementation [2].

1-2. How to handle AI that changes post-market

In Japan as well, discussions are advancing within the PMDA's (Pharmaceuticals and Medical Devices Agency) Expert Committee on AI-enabled SaMD, predicated on the nature of adaptive AI whose performance can change post-market due to machine learning. The central issues there are the monitoring of ML (machine learning) bias, the reuse of evaluation data in post-market learning, and the continuous development of clinical databases [3]. In short, "how to supervise the AI after it is deployed" has become the primary focus.

1-3. High hurdles in data utilization itself

Guidelines established by Japan's Ministry of Health, Labour and Welfare (MHLW) strictly outline the legal basis for each R&D stage and the practical operational procedures for pseudonymized information, with the protection of patients' rights and interests as a major premise [4]. This indicates that the barrier to medical AI implementation lies, even before model construction, in how to legitimately and safely handle data and integrate it into operational procedures.

1-4. Implementation often lacks continuous monitoring and evaluation frameworks

This issue is clearly pointed out at the forefront of implementation research. A 2025 paper on FAIR-AI argues that what healthcare institutions truly need is not initial evaluation but "rigorous evaluation and ongoing monitoring," and that existing frameworks lack practical guidance for the clinical workflow [5]. This strongly suggests that AI implementation is not a "deploy and forget" process.

Furthermore, a 2025 meta-analysis in npj Digital Medicine showed that the overall accuracy of generative AI in diagnostic tasks remained at 52.1%, and while there was no significant difference compared to non-specialists, it was significantly inferior to specialists [6]. This does not mean medical AI is worthless, but rather provides data backing that "supervised assistance and operation with conditional admissibility, rather than replacing independent judgment, is realistic."


2. The True Limit is Not "Lack of Performance" but "Lack of Admissibility Conditions"

As mentioned above, medical AI is stalled not simply because AI is not smart enough. It is because the "conditions under which outputs may be admitted into clinical workflow" are not fixed as a system. Specifically, this can be defined as a lack of the following four conditions:

  • Responsibility Boundary: The boundary that defines who holds responsibility for which judgments based on the AI's output.

  • Safe State Transition: The conditions to stop the AI, suspend judgment, and initiate a human handoff when it enters a danger zone.

  • Audit Trail: A recording mechanism that makes it possible to retrospectively verify "why" a particular AI output was authorized.

  • Human Oversight: The maintenance of a state where humans can intervene in the system and exercise real authority.

The WHO's guidance on medical generative AI aims in exactly this direction, strongly recommending the implementation of risk management and governance [1]. These four points are not abstract concepts, but core conditions organizing the requirements emphasized by existing regulatory authorities at the implementation level.


3. What Japan's Medical AI Implementation Needs Next

Japanese regulatory authorities are not trying to hinder the introduction of medical AI. As seen in "DASH for SaMD 2," they are aiming to promote practical application while advancing review and change management processes through initiatives like two-stage approval, clarifying concepts for general-purpose SaMD, and promoting the use of change plan confirmation procedures [7]. Looking at the discussions in the PMDA's Expert Committee on AI-enabled program medical devices, the focus has shifted from "whether to approve AI or not" to "how to manage it, permit changes, and retain evidence."

Therefore, what Japan's medical AI implementation urgently needs is not "rejecting AI," but the explicit documentation and system implementation of the following operational conditions:

  • Pre-deployment Release criteria

  • Post-deployment Monitoring criteria

  • Fail-safe / Human handoff criteria during anomalies


4. Where Does ADIC Fit In?

For this issue of "unfixed admissibility conditions," the introduction of "ADIC" serves as one solution approach. Below, GhostDrift's ADIC is proposed as an implementation candidate for the institutional and operational requirements mentioned above.

ADIC is not an AI model that replaces diagnostics itself. Nor is it a black box that magically enhances the accuracy of existing medical AI. ADIC is positioned as a gate layer that determines whether an AI's output should be admitted into clinical practice, and as an evidence layer that externalizes and records the rationale for that determination.

Specifically, it functions as the following architecture:

  • Release Gate: Determines whether the AI's output meets the established thresholds for medical validity, safety, and scope of application, and judges whether to pass this output to the clinical setting.

  • Safety Gate: Immediately stops or suspends processing and reverts to a human (physician) verification phase when the AI's behavior reaches out-of-distribution (OOD) areas, performance degradation, input deficiencies, or dangerous conditions.

  • Audit Gate: Retains metadata such as the reasons for authorization/blocking, thresholds, decision logs, and human intervention history regarding why the output was authorized (or blocked), enabling retrospective verification and auditing.

This structure can be positioned as a candidate that provides connection points at the implementation level for lifecycle management [2], handling of post-market changes [3], human oversight [1], and auditability emphasized by the FDA, PMDA, and WHO.


5. What Can ADIC Improve?

By deploying ADIC as a gate layer, medical institutions and developers can possess more explicit control and recording mechanisms for at least the following challenges:

  1. Reducing the risk of post-hoc diffusion of responsibility: It makes it easier to clarify responsibility boundaries through the recording of admissibility conditions.

  2. Embedding the "right to stop" into the implementation: It makes it easier to embed transitions to stopping, suspending, or human verification during emergencies as design conditions.

  3. Demonstrating "on what grounds it was authorized" during audits: During an audit, it makes it easier to demonstrate the rationale for authorization/blocking decisions based on external criteria.

In short, ADIC is not a technology that makes medical AI omnipotent, but rather a technology that brings medical AI closer to an "admissible form" that can withstand the legal and ethical requirements of clinical practice.




6. Conclusion

The limits of medical AI in 2026 cannot be explained merely by accuracy competitions or a lack of model performance. The true bottleneck lies in the fact that responsibility boundaries, explicit stopping conditions, secured audit trails, and human oversight have not been fixed as "operational admissibility conditions."

ADIC is positioned as a robust implementation candidate to address this deficiency.

The next competition in medical AI is not just about creating smarter models. It is a competition of "whether the conditions can be fixed" to safely and legitimately admit the increasingly complex AI outputs into the clinical workflow.


References

[1] World Health Organization. Ethics and governance of artificial intelligence for health: guidance on large multi-modal models. Geneva: World Health Organization; 2024. [2] U.S. Food and Drug Administration. Artificial Intelligence-Enabled Device Software Functions: Lifecycle Management and Marketing Submission Recommendations. Draft Guidance. 2025. [3] Pharmaceuticals and Medical Devices Agency (PMDA). Expert Committee on AI-Enabled Program Medical Devices. [3a] Pharmaceuticals and Medical Devices Agency (PMDA). Report on AI-Enabled Program Medical Devices. August 28, 2023. [4] Ministry of Health, Labour and Welfare (MHLW). Guidelines for the Utilization of Medical Digital Data for AI Research and Development. March 31, 2024. [5] Wells BJ, et al. A practical framework for appropriate implementation and review of artificial intelligence (FAIR-AI) in healthcare. npj Digital Medicine. 2025. [6] Takita H, Kabata D, Ueda D. A systematic review and meta-analysis of diagnostic performance comparison between generative AI and physicians. npj Digital Medicine. 2025. [7] Ministry of Economy, Trade and Industry (METI) & Ministry of Health, Labour and Welfare (MHLW). Strategy Package for Promoting the Practical Application of Program Medical Devices 2 (DASH for SaMD 2). 2023.

 
 
 

コメント


bottom of page