Clinical Trials - Aventior https://aventior.com Thu, 11 Sep 2025 14:17:27 +0000 en-US hourly 1 https://wordpress.org/?v=6.8.1 https://aventior.com/wp-content/uploads/2021/11/Avetior-Favicon-100x100.png Clinical Trials - Aventior https://aventior.com 32 32 The Next Big Thing in AI in Clinical Trials https://aventior.com/clinical-trials/the-next-big-thing-in-ai-in-clinical-trials/ Mon, 20 Feb 2023 07:07:53 +0000 https://aventior.com/?p=6086

Pharma and Life Science companies face issues like complex procedures, compliance, patient recruitment, and high cost while conducting clinical trials. Most of the said companies have moved into digital mode and the usage of Artificial Intelligence (AI) technology has gained momentum. 

The top trends in the clinical trials market are AI, Data management, Wearables, Decentralized clinical trials, cloud computing, cybersecurity, 5G, and Blockchain. AI will have a 22% impact on the above trends. 

The Next Big Thing in AI in Clinical Trials

How does AI tech in clinical trials replace traditional studies?

Initially, the patients would visit the clinic and fill out the forms. The patient recruitment would be done basis of such data received. Incorrect patient recruitment leads to trial failure and negative ROI. The FDA approvals for successful clinical trials take time, tons of paperwork, and, investment.

AI in clinical trials is paramount as it can:

1) Optimize trial design

The trial design is the most important factor and having the right data is the key to success. The data has to be organized and analyzed in a way so that researchers can cite a meaningful pattern for the trial. Optimal trial design will help to lower the cost and time involved and it leads to the success of the trial. AI technology-backed tools like Natural language processing (NLP) and Machine learning (ML) help in data interpretation and optimize the trial design. 

2) Identify & recruit suitable patients for trials 

AI can speed up the process to identify and recruit eligible patients and create subgroups for trials. The data collected from EMR and social media platforms can be analyzed accurately and simplify eligibility criteria for cohorts as well. Aventior offers IT solutions to the healthcare domain. This allows them to focus on business goals instead of server hosting and its maintenance. Aventior delivers purpose-based AI in clinical trials to Pharma companies. It helps to digitize the clinical trial data collection process. It allows sponsors and CROs to capture data from participants remotely through one unified platform through a Data-driven analytics dashboard.

3) Ensure patient retention throughout their trial

It has been observed around 30% to 40% of recruits either drop out or turn non-adherent during the trial. To overcome this issue companies are opting for AI for clinical developmentAventior enhances the patient experience involved in the clinical trial. The use of eCOA solutions permits the patient to easily submit the samples from their homes. The wearables segment and smartphones backed by eCOA solutions help in accurate data capturing and reporting. Data captured makes it easier to understand patients’ behavior during studies. The effectiveness and efficiency of eCOA solutions make it a rational approach for the future of clinical trials.

4) Help in data management and reduces bias in it

AI in clinical trials helps doctors and scientists in their decision-making process. It improves the accuracy of the trial and its outcome. Aventior-backed eCOA solutions ensure the transcriptional error is minimized and data accuracy is improved. It gives consistent and reliable data. Any issues related to data can be corrected in real-time.

5) Adherence to compliance 

Regulatory bodies have laid down stringent rules for data collection techniques and methods. The data collected has to be reported as per protocol requirements. Aventior ensures AI backed eCOA platform meets the regulatory guidelines.

6) Enhance the cost-effectiveness of clinical trials

Due to the above usage of AI in clinical trials, the cost of clinical trials is lessened and the effectiveness of outcome is enhanced. The usage of AI in Pharma and Life science is here to stay and evolve in the coming years.

Aventior had been involved with a clinical research automation company in the process to overcome the difficulties in digitizing the clinical trial data collection process. They wanted one unified platform to decentralize studies. As this will allow the CROs and sponsors to procure data from it. The company wanted to use serverless technologies as the base platform architecture. For this Aventior developed a purpose-led application on open-source technology. It was hosted on Amazon Web services with scalable infrastructure. GCP and 21 CFR Part 11 of compliance were taken into consideration and adequate privacy and security were built into the application. Aventior was able to deliver a secure platform that digitized clinical studies data collection with outcome visibility in real-time.

The advancement of AI in clinical trials has gathered momentum post-covid and this will continue to boom. Pharma and bioscience would focus on the execution and outcome of clinical trials and the IT solutions will be outsourced to third-party AI companies to handle. They can develop tailor-made efficient solutions and platforms to get positive ROI. 

For more information about our solutions for AI in clinical trials, write to us at info@aventior.com

To download the poster presentation, kindly enter the following details :

[contact-form-7]

The post The Next Big Thing in AI in Clinical Trials first appeared on Aventior.

]]>
Top Ways Artificial Intelligence Is Transforming Digital Pathology https://aventior.com/clinical-trials/top-ways-artificial-intelligence-is-transforming-digital-pathology/ Tue, 27 Apr 2021 12:47:18 +0000 https://aventior.com/?p=4355

Way back in the late 1960s, telepathology, wherein images captured from the microscope to photographic plates, was invented. Later we saw the refined version of the same in digital pathology. The virtual microscope is used on digitized specimen slides and data is generated for evaluation. The entire information is available on a single slide. Whereas, in the traditional method, different tissues had to be taken on different slides and scrutinized under a microscope.

The major benefit of digital pathology is its ability to detect potential diseases that can develop in the future. Usage of Artificial Intelligence (AI) in digital pathology has gained popularity in recent times. It addresses the shortcomings of digital pathology – like, shortage of experienced pathologists, the processing of a huge amount of data available through patient care, and its classification and management. AI-based pathology data management reduces the errors caused in diagnosis & classification to a large extent.

AI-based computational pathology is transforming the face of Digital Pathology.

Let us see the top ways in which AI has brought about the transformation in Digital Pathology:

Patient-centric Treatments

AI can analyze data quickly. This helps doctors to take accurate decisions promptly with regards to prognosis & the line of treatment. AI is a life-saver when it comes to patients afflicted by cancer. Since there are various types of cancer with all possible diagnoses, taking a well-informed decision is often tough. Using AI in pathology diagnosis of common types of cancer is must faster. This leads to speedy treatment and chances of survival are enhanced. The doctor can also make a patient-centric evaluation on a case-to-case basis using AI-based pathology reports.

Faster results & saves time

Certain slides are labor-intensive. These slides require more time while analyzing. With AI such slides can be handled effectively. This gives the pathologists to use their time & efforts to focus on other challenging work. The perfect example is the Google-developed AI algorithm called Lymph Node Assistant. This helps to identify the metastatic breast cancer slides with 99% accuracy in half the time.

Ease of work for Pathologists

Artificial Intelligence transforming digital pathology

For the overworked and understaffed lab, AI is a boon. AI can flag cases that need detailed study and further investigation by the pathologist. The pathologists need not waste time detecting problem-free cases. For example, cases of biopsies – few of the biopsies are problem-free. AI can spot such cases swiftly and mark them as normal. The pathologists can focus on slides that need their attention on a priority basis. This eases the workflow of pathologists and improves their efficiency.

Data Management

The huge load of data generated is easily managed, analyzed, classified, and stored. The same can be easily retrieved for further studies. The Electronic Health Records information can be shared in real-time. AI manages huge datasets. This improves the efficiency of doctors and pathologists to a great extent. When it comes to saving lives, a large amount of handy detailed information is crucial.

Research & Development 

AI is capable of capturing subtle patterns or any changes & give out information that could be missed by the human mind. Humans have limited visualization power. However, it is easy for AI to visualize these patterns or changes. It can also predict the effects through visualization. This leads to new learnings & discoveries. All this is a huge help in the development of the healthcare system.

Consistent Task Completion

AI technology can handle massive data with accuracy and publish reports in a detailed manner. The performance is purely based on software technology. The output is consistent and 100% accurate. There is no room for error due to overwork. Whereas in the case of pathologists, due to overwork or any other human shortcomings, they may tend to miss out on minor points. This is completely avoidable by the usage of AI in Digital pathology.

Conclusion

Each technology comes with pros & cons. Nevertheless, AI has its shortcomings like storage space limitations, the cost of incorporating AI in digital pathology, and more. With passing time, the cons will be outweighed. The cost of AI-based technology will be lower. In the future, we only see its advancement in digital pathology. The patients can avoid visiting doctors for initial diagnosis. He can add his data images in the app and get them checked by AI. The amount of time saved hugely increases the chances of quick recovery. The doctor can help the patient with accurate diagnosis and treatment. AI in digital pathology is beneficial to doctors, patients, and the entire healthcare system.

Aventior offers solutions for Metadata management for digital pathology. This is accurately done by using Artificial Intelligence technologies. Their Label Extraction Solution uses a combination of OCR technologies and AI to comprehend & store data from pathology slides. For more information regarding our Digital pathology and AI services, please feel free to contact us at info@aventior.com.

Tell Us for more about your requirements here

 

[contact-form-7]

The post Top Ways Artificial Intelligence Is Transforming Digital Pathology first appeared on Aventior.

]]>
Biotech Batch Record Digitization: Challenges and Role of AI https://aventior.com/clinical-trials/cpv-batch-record-digitization-challenges-and-role-of-ai/ Wed, 01 Jul 2020 14:23:00 +0000 http://52.15.134.29/?p=1125 Abstract: In drug manufacturing, keeping track of data is crucial for the drug’s approval from...

The post Biotech Batch Record Digitization: Challenges and Role of AI first appeared on Aventior.

]]>
Abstract:

In drug manufacturing, keeping track of data is crucial for the drug’s approval from the FDA. The “Continued Process Verification” (CPV) data needs to be maintained for ensuring that the product outputs are within predetermined quality limits. In spite of rising demand for the creation of digital data directly at the source itself, some companies follow the traditional methods of documenting the processes parameters on paper, on designed forms. This leads to data being inaccessible for others unless it is again digitized by someone. The traditional way of achieving this is to have someone enter the data into a computer system by shuffling around the pages in the document. The manual process consumes a lot of time and leaves very little time for the data entered to be validated. This article briefly describes the possible methods of automating the manual data entry process, and how the upcoming technologies can be used for this work.

Introduction

Converting handwritten and typed data from papers into digital formats is one of the most commonly faced challenges across industries today. Keeping data on papers has its own set of limitations like limited accessibility, searchability, using data for analytics, etc. The digital transformation of such documents is necessary. Over a while, companies have adopted various methods of converting this data into a digital structured format. Some companies hire interns to manually enter the data from the document into excel sheets or word documents, while some companies need the scientists working on the projects to manually enter this data into digital documents. The manual data entry process consumes valuable time and effort from the scientists while explaining the entire process of data entry to new interns consumes time from the team. With advancements in the software industry, there have been multiple attempts to solve these problems, but each solution comes with its own set of limitations. Robotic Process Automation (RPA) is one of the closest successful solutions in helping companies convert their data from papers to structured digital formats. RPA relies on the rules that the documents might follow. The papers are scanned and processed as an image in the RPA software. The software tries to identify the set of parameters on the image, which need to be translated into the structured database. The set of parameters is searched based on certain rules of the document, which could be the sequence of pages, the sequence of words on the pages, or some other form of a landmark for identifying the parameters on the paper. This approach mostly fails if the paper documents do not follow any template, or there exist multiple pages with similar contents, or if most of the contents on the paper are handwritten. Considering the dynamic nature of the documents, it becomes difficult for the software to define rules based on which the process can be completely automated. Many solutions/software rely on Optical Character Recognition (OCR) engines as one of the primary components in their toolbox, which is further combined with techniques from the Natural Language Processing (NLP) domain, to try to make sense of the extracted texts. But many solutions fail due to the inability of the OCR to provide accurate results on scanned pages containing hand-written texts, special symbols, marking, notes, etc. This leads to breaking the flow of a possible fully automated solution.

The following sections talk about various possible OCR engines from the leading firms in the market and try to explore and evaluate the performances of the OCRs specifically on hand-written texts. Further, an experiment is performed to try to integrate the OCRs with custom-built software that can use the OCR output and try to structure the data from scanned pages into a database. The pros and cons of this approach are explained in subsequent sections. Furthermore, to overcome the shortfalls of OCR technology, an alternate approach is suggested, using Speech-to-text for data extraction. The speech-to-text approach is also integrated within a custom-built software to evaluate the efficiency of data extraction, in terms of speed and accuracy. The speech-to-text based solution is further investigated on its scalability aspect, and how much time and effort would be needed per batch records are calculated.

Experimental Setup

The main objective of the experiments was to study and analyze the best possible approach in automating the data extraction process from scanned documents archived in PDF formats, which would also include the handwritten texts.

Data: The experiment is performed on a set of sample scanned PDF files. Each PDF file would represent one batch-record. Multiple batch records are obtained from the same CMO and for the same drug, which would make sure that all the PDFs have the same parameters, which can be extracted from them. Also, this ensures that all the PDFs try to closely follow a certain template, which can be thus used in automated software.

OCR Evaluations: The experiment started with the evaluation of the performance of various state-of-the-art commercial and open-source OCR engines, and the implementation of various methods to achieve better performances using image processing techniques. The OCRs are evaluated based on factors like ease of use in the BioTech and Pharmaceutical Industry; and the accuracy of performance. Commercial OCRs included in this experiment include Google Vision APIs, Amazon Textract, and Abbyy FineReader, while Tesseract OCR will be used under the open-source OCR category. Based on the performance achieved from the OCRs and its applicability in the BioTech and Pharmaceuticals industry, the next steps would be decided.

Speech-to-text / Speech Recognition: The Speech-to-text solution is analyzed for its applicability in the industry, and the solution is evaluated based on the performance achieved in terms of accuracy of data extraction and speed of data extraction by performing a time study. For evaluating the achievable speed via speech-to-text based data extraction system, the software is created for the workflow which allows to user to choose the PDF files and load extraction metadata, and auto navigates using an intelligent page finding algorithm, simultaneously allowing the user to record his words for data entry into the database against a parameter, thus automating the data structuring process. The time study is performed on five batch records, where each batch record is split into three PDFs. Each PDF denotes a stage in the drug manufacturing process.

OCR Assessment

The OCRs selected for the experimental study included Google Vision, Amazon Textract, Abbyy FineReader, and Google’s open-source Tesseract OCR engines. To evaluate the performance of the OCRs, 20 pages from the batch record PDFs in the form of images, were sent to each OCR after cropping the images to sections containing only handwritten information and the obtained text was compared to the ground truth texts which were recorded manually for these pages. The standard text-to-text comparison algorithms were used, in which two strings are compared with each other based on character-wise similarities. The results from the first evaluation are tabulated below.

It can be observed that Google Vision’s  API service provides the best results on the handwritten texts but is capable of giving results with only 66.7% accuracy. In the BioTech and Pharmaceuticals domain, the achieved accuracy is much lower than desired, as the information is very sensitive to safety factors. To improve the accuracy of Google’s options, i.e. Vision API and Tesseract, we can perform image enhancement techniques, which would provide better images for the OCR, and hence the OCR output accuracy can be improved.

The image processing techniques involved with document processing were implemented as a pre-processing step. Techniques like image sharpening, contrast and brightness adjustments, noise removal, etc. were implemented, and the same images were again passed on to the OCRs. These image improvement techniques proved beneficial for the Tesseract OCR but reduced the accuracy of Vision API. The accuracy from Google Vision API reduced from 66.7% to 60%, but for Tesseract OCR the accuracy increased from 33.5% to 44.1%. This showed that one cannot rely solely on the image processing techniques, but the improvements in the OCR engine itself were required. As Google’s Vision API is a third party solution, one has no control over the training of the system or even with data privacy. Hence, for the next round, only Tesseract OCR was considered. It was trained further on handwritten data. The handwritten sections from the PDFs were split into train and test. The training data was used for further training of Tesseract OCR, while the test data was used for evaluating the OCR performance while training. 

The following two images show a sample result of image processing techniques applied which resulted in the improvement of Tesseract OCR performance.


A sample of a handwritten section, before and after image processing

Tesseract OCR’s performance was measured across all the handwritten sections in the entire 5 batch records resulting in 15 PDFs of varying sizes. The average performance of the OCR on the entire documents set was observed to be 34.35% before training, while after training the Tesseract OCR results improved to 47.1%. Using OCRs which could utmost provide maximum 50% accuracy on handwritten texts cannot be used for developing an automated system.  

The OCR technology has yet to reach its full maturity. There are multiple instances of the document and handwritten data variations observed in our daily activities. No doubt that the OCR technology has seen tremendous improvements in the past 5 years, but it is yet to become independent of variations like different handwritings, font sizes, shades of ink, image noise and associated garbage values, bordered and borderless tables, etc. It is due to these limitations of the OCR technology, that a completely or partially automated solution cannot be developed.

Speech-to-Text

The speech-to-text conversion software is matured enough and can produce results with 90-95% accuracy in the achieved text outputs. The performance can be further improved up to 100% if the speech recording is made in a silent environment and the pace of speech and accent is maintained. The Google Cloud Speech service was initially used but was found to be slowing down the process of data extraction, as it needs the audio to be first recorded and saved in a file, which is then sent to the cloud service and it returns a text output. The text output is provided to the data entry system, which takes care of the text being allocated to the correct parameter. The audio file needs to be recorded for each data point, which makes the entire operation of the data fill-up process very slow.

As an alternative, Speech-to-text software running locally can be used. The workflow is created to integrate the speech-to-text method into data extraction and data entry software. The following diagram denotes the workflow of the software for new batch-record processing.

The speech-to-text approach is integrated into software that can take care of the remaining workflow for data entry into a database format. The input to the workflow is the PDF and a list of parameters that need to be extracted from PDF. The pages in the PDF are converted into images for ease of processing. A mapping file is created which can map the parameters found in the PDFs to the associated page within the PDF. The pages of interest are brought to the user for each parameter so that the user can start reading the information on the page for the parameter. The speech-to-text software running in the background keeps listening to the voice and starts creating the equivalent text in a text file. Once all the parameters are recorded, the text file can be processed within the software to automatically fill-up the information for each parameter in a structured format. A validation step is also introduced within the software so as to allow the user to check the processed key-value pairs of information and correct the data if any errors are detected.

The results for time study are recorded over multiple iterations for all the five batch records. On an average, around 120 minutes were needed for each batch record, containing 3 PDFs each, which consisted of around 200 pages per batch and around 150 parameters to be extracted from those PDFs per batch. While an initial setup time of 5 hours was needed to create and validate the mapping file, which is a one time process for each type of CMO and Drug combination. The manual data entry operation can process one or two batch records per day, resulting in 5 to 10 batch records per week. The speed achieved via the above solution is trice of that of the manual data entry process.
Considering the above setup, and two roles involved in this process, one for data entry using speech-to-text technology, while the other for validation, we could scale-up the software to process around 30 drug manufacturing batch-records for a single CMO and Drug combination in a week. 

The speech-to-text system results in nearly 100% accuracy and the system can be scaled up to work much faster than manual data entry and data validation process.

Conclusion

After comparing the various types of OCR solutions, it can be concluded that OCR solutions are not effective against the variations in handwritten texts. The maximum of 50% accuracy was achieved in the trials performed on the handwritten information in the PDFs, while the Speech-to-text solution achieved nearly 100% accuracy of data extraction using the human-in-loop method for the data validation. Speech-to-text seems to be a much viable solution in the industry, as it can outperform the current manual data entry approaches. Also, a combination of OCR (for pre-processing) and speech recognition yields an average of 5X improvement in overall processing speeds with zero/near-zero error rates.

The post Biotech Batch Record Digitization: Challenges and Role of AI first appeared on Aventior.

]]>