Synthetic generation and tagging of CT and pathology images
Our similar previous work has focussed on using pre-trained embedding models for chest x-rays (https://github.com/nhsx/txt-ray-align) and variational auto-encoders (https://github.com/nhsx/SynthVAE) - see reports within the repos. This investigation would not seek to make a final dataset, but rather explore the methodology and the level of fidelity and quality that these synthetically generated data could obtain with current methods. The methodology and report would be made open through a GitHub repository at the project end but with no training or synthetic data published.
- Quantification of the accuracy obtained when using pre-trained models to represent the text and image embeddings for different healthcare images
- Demonstration of generation of image from text and text from image for established multi-modal datasets.
- Investigation into the best algorithm to use for generating synthetic healthcare images in terms of fidelity, privacy, fairness, and explainability.
- Possible extension into visual question and answering search algorithm
NHS England Transformation Directorate