Google AI Releases ‘CXR Foundation Tool’ to Allow Researchers to Jump-Start CXR modeling Efforts Using Simpler Transfer Learning Methods
Close to a billion chest X-ray pictures are taken worldwide each year to help identify and treat health issues varying from collapsed lungs to infections. These chest X-Rays are often less expensive and easier to obtain than other types of medical imaging. Existing obstacles, however, continue to obstruct the optimum usage of CXRs. For example, qualified radiologists who can adequately interpret CXR pictures are in low numbers in certain places. Furthermore, expert interpretation variability, workflow variations across institutions, and the prevalence of unusual illnesses known only to subspecialists all contribute to the difficulty of making high-quality CXR interpretations.
Recent research has used machine learning to investigate possible answers to some of these problems. Deep learning algorithms that detect anomalies in CXRs and greater access, precision, and efficiency in identifying illnesses and ailments affecting the lungs and heart are being developed with much attention and effort. Building solid CXR models, on the other hand, need enormous annotated training datasets, which may be prohibitively costly and time-consuming to develop. Limited data is accessible in certain circumstances, such as when dealing with underrepresented communities or researching uncommon medical disorders. Furthermore, the quality of CXR pictures varies between populations, regions, and organizations, making it challenging to create robust models that function well internationally.
Let us discuss how Google Health uses advanced machine learning (ML) methods to generate pre-trained CXR networks having the capability to convert CXR pictures to information-rich mathematical vectors, which enables the development of CXR models with lesser data and fewer computational resources in “Simplified Transfer Learning for Chest Radiography Models Using Less Data,” published in the journal Radiology. In this journal, the researchers demonstrate that, while using fewer data and computing, their strategy achieves performance equivalent to impeccable deep learning models across various prediction tasks. The researchers also took pleasure in introducing the CXR Foundation, a tool that lets developers construct bespoke embeddings for their CXR pictures using a CXR-specific network of researchers. This study will hopefully assist in speeding up the evolution of CXR models, which will aid in illness identification and contribute to more equal health access all across the globe.
To create such a network of Chest X-rays, pre-training a model on a generic task using non-medical datasets is done, and then the model is refined on a target medical task. By applying natural image understanding to medical images, this transfer learning process may improve target work performance or, at the very least, accelerate convergence. However, transfer learning could still require massive labeled medical datasets for the refinement step.
Using a three-step model training setup consisting of a generic image pre-training based on traditional transfer learning, a CXR-specific pre-training, and task-specific training, the team of researchers created a CXR-specific image classifier using supervised contrastive learning (SupCon). SupCon joins representations of images with the same label and separates representations of pictures with different labels. This model has been pre-trained on de-identified CXR datasets of more than 800,000 images generated in collaboration with Northwestern Medicine and Apollo Hospitals in the United States and India.
According to the findings, adding the second pre-training stage allows high-quality models to be trained with up to 600 times fewer data than traditional transfer learning approaches that use pre-trained models on generic, non-medical datasets. This was discovered to be true regardless of the model architecture or dataset used for natural image pre-training. This approach can help Researchers and Developers significantly reduce their dataset size requirements.
The team said their work on public datasets like ChestX-ray14 and CheXpert significantly and consistently improved the data-accuracy tradeoff for models created throughout a range of training dataset sizes. When evaluating the tool’s ability to develop tuberculosis models, for example, data efficiency improvements were more striking; models trained on the embeddings of only 45 images outperformed the radiologists in detecting tuberculosis on external validation data.
The researchers announced that the CXR Foundation tool is being released, along with scripts to train linear and non-linear classifiers to help speed up CXR modeling endeavors with limited data and computational requirements. Through these embeddings, this tool will help scientists jump-start CXR modeling attempts using relatively simple transfer learning methods.
This Article is written as a research summary article by Marktechpost Research Staff based on the research paper 'Simplified Transfer Learning for Chest Radiography Models Using Less Data'. All Credit For This Research Goes To Researchers on This Project. Checkout the paper and Google blog. Please Don't Forget To Join Our ML Subreddit
Credit: Source link
Comments are closed.