Blogs & News
AI is reshaping healthcare research, but the need for rigorous data governance hasn’t changed. Sensitive patient records, clinical trials, and genomic datasets require environments that protect confidentiality while enabling innovation. With this in mind, Aridhia proudly unveils the AI Research Assistant Framework (Aira), a new capability within the Digital Research Environment (DRE) that brings the value of secure AI to healthcare and life sciences organisations without compromising on the provenance of the data driving the model outputs.
Healthcare datasets such as clinical records, omics data, and trial metadata are subject to strict data-sharing agreements, regulatory controls, and patient confidentiality standards. Public cloud-based AI services like OpenAI do not provide the required level of assurance, as they rely on external data transfer and lack robust legal privilege frameworks.
It’s not only the datasets themselves that are sensitive, but also the prompts, instructions, and other inputs researchers provide, as well as the outputs of models. These can include identifiable information, confidential research data, or intellectual property that must remain within a secure environment.
OpenAI CEO Sam Altman has recently acknowledged:
“There’s no legal confidentiality when using ChatGPT… If you talk to a therapist or a doctor, there’s legal privilege. We haven’t figured that out yet for AI.”
Sam Altman, CEO OpenAI
This establishes the need for offline AI models that can work within a safe, secure analysis sandbox, such as Aridhia’s Workspaces, that in turn operate entirely within the security perimeter. These models must be able to comply with data use conditions – whether that’s specified by a single organisation, multi-party collaboration, or a particular research topic where there may be multiple data controllers granting access to their datasets.
As custodians of health data, data owners hold the responsibility of protecting subject privacy, honouring agreements, and upholding ethical research standards.
Offline modelling ensures that:
Whether mandated by legal contracts, institutional policy, or regulatory bodies, offline deployment is not optional, it is essential for safe and scalable AI in healthcare.
The regulatory environment for AI is evolving quickly, with different countries and even individual states introducing new requirements around transparency, explainability, and data residency. Navigating these can be complex, for example across the EU where the EU AI Act and European Health Data Space initiative present further challenges when working with patient data and AI technologies as described in this article from the European Law Blog.
For healthcare organisations operating across borders, this patchwork of rules adds another layer of complexity. Deploying AI entirely within a governed, compliant environment like the Aridhia DRE ensures you meet current obligations and are ready to adapt as regulations change.
It’s unlikely that existing data sharing agreements have considered the potential of the unsafe use of AI technologies, even in what are considered “trusted research environments” or TREs. It is increasingly important that they do, and in some cases, that existing agreements are re-visited and updated accordingly.
LLMs are excellent at generating code and interpreting natural language, but they do not perform mathematics, statistical reasoning, or validation of scientific results. Their outputs can include hallucinations: plausible but incorrect information, which pose serious risks in clinical and scientific domains. Researchers should always use verified, peer-reviewed code for statistical calculations. LLMs may assist with code generation but are not reliable sources of numerical truth.
Aridhia’s DRE also provides several technical and governance safeguards to review code and output:
Designed for flexible, secure AI experimentation, the framework includes:
Aira enables customers to experiment with, fine-tune, and deploy models, whether open source, commercial, or in-house, all within the secure boundary of the DRE. This ensures model innovation can occur without any data leaving the environment. We recognise that newer models are not always better, as recent examples such as OpenAI’s ChatGPT-5 regressions have shown.
A primary focus for Aira is on secure, assisted code generation for analytics in R, Python, and SQL. This capability helps researchers rapidly develop and refine analysis pipelines, query data, and create visualisations, all within the DRE environment. Every code snippet produced can be peer-reviewed, version-controlled, and validated before execution, ensuring reproducibility and scientific integrity.
However, beyond code generation, Aira provides support for a wide range of high-value use cases, including:
Together, these capabilities allow organisations to test, train, and deploy models without compromising the confidentiality of datasets, prompts, or outputs.
Aridhia’s AI Research Assistant Framework delivers the most compliant and capable platform for working with LLMs in life sciences. It puts full control in the hands of researchers and data owners, supporting both innovation and integrity.
Contact us today if you would like to learn more about The AI Research Assistant Framework.
August 13, 2025
Scott joined Aridhia in March 2022 with over 25 years’ experience in software development within small start-ups and large global enterprises. Prior to Aridhia, Scott was Head of Product at Sumerian, a data analytics organisation acquired by ITRS in 2018. As CPO, he is responsible for product capabilities and roadmap, ensuring alignment with customer needs and expectations.