Artificial intelligence for bone cancer imaging


Research highlights

  • AI can integrate clinical information to stratify patients to personalized imaging procedures.

  • AI can aid radiologists in the detection of tumors in other organ systems.

  • Due to sparse data, the characterization of bone tumors is limited to date.

  • AI can enable ultralow-dose imaging.

  • AI can facilitate treatment monitoring and surveillance.

Introduction

Imaging tests, such as magnetic resonance imaging (MRI), computed tomography (CT), and positron emission tomography (PET), are essential for diagnosing bone cancers and monitoring tumor response to therapy. Recent evidence suggests that artificial intelligence (AI) can assist with tumor detection and treatment monitoring [ ]. AI is broadly defined as intelligence demonstrated by machines in contrast to intelligence displayed by humans or animals. Machine learning (ML), a subset of AI, focuses on training machines to use mathematical algorithms and statistical models to generate predictions and decisions from inputs without using explicit instructions. For instance, if a machine is fed a list of input features of bone lesions (e.g., location, size, homogeneity, etc.), it could predict if the lesion is benign or malignant and provide an estimate of the prediction's accuracy. Recently, a particular method within ML called deep learning has gained popularity. Deep learning identifies patterns in data and takes actions to reach preset goals without specific programming. One of the most commonly studied deep learning techniques in medical imaging is convolutional neural networks (CNNs). CNNs have been used for many medical imaging applications, such as bone age estimations [ ] and diagnosis of pathological diseases [ ]. However, applications for bone tumor imaging have been limited thus far. This chapter provides an overview of the current status of AI in bone cancer imaging and its future potential with regard to deep-CNN applications. We also highlight technical concepts important for understanding some methodologies used in these studies.

AI technologies

AI attempts to complement human intelligence in cognitive functions such as speech recognition, visual perception, and decision-making. AI algorithms can integrate complex clinical and imaging data, substantially accelerate quantitative measures, and are constantly available. While AI shows increasing promise in medical imaging, human intelligence is still critical in adding common sense and ethical considerations. Hence, AI combined with human expertise can increase the safety, efficiency, and accuracy of bone tumor diagnosis and treatment.

Supervised versus unsupervised learning

There are two main learning styles in ML: supervised and unsupervised learning. Supervised learning involves datasets with known outputs. For instance, each entry within a dataset could have two features, such as tumor size and location, and a known output (called the label ), such as benign or malignant diagnosis. Supervised learning uses this dataset (called training data ) to train the model, generating a mapping function that connects input to output through pattern recognition, penalizing incorrect predictions, and correcting the model until a desired accuracy is achieved. In contrast, unsupervised learning involves datasets with no labels, where the computer extrapolates information and performs pattern recognition to generate outputs/predictions.

ML algorithms

Support vector machines (SVMs) are examples of supervised ML algorithms which create a decision boundary that separates two classes of data points in N-dimensional space (where N is the number of features in the training data). For instance, in a training dataset of 500 examples with each example having two features (e.g., tumor size and location), an SVM would consider a 2D space with 500 data points and segregate (e.g., benign and malignant tumors) by a decision boundary. A decision tree classifier is a classification algorithm that uses features to perform branching until it sufficiently segregates the training data. For instance, a branch point could be a tumor size >8 cm. This branch point creates a data subset which will be inputs to the next branch node. Random forest (RF) is another example of a classification algorithm, where individual decision trees become an ensemble. Each decision tree gives its prediction and the tree with the most votes becomes the prediction for the model. Naive Bayes classifier is a supervised classification algorithm based on Bayes' theorem that returns a probabilistic prediction given the inputs, assuming all features are independent.

Neural networks (NNs) are modeled after the architecture of the brain. Each neuron (or node) has incoming connections each associated with an assigned weight. When data come in through the connection, they are scaled by the corresponding weight. The node integrates the input values via an activation function, which determines whether to “fire” off the data to the next node. With this, input data are not evaluated independently but with the consideration of multiple different variables. Deep learning incorporates NNs with many more processing layers in the network.

CNNs are popular deep learning architectures for image recognition. This is because CNNs are effective at dimension reduction and images have high dimensionality because each pixel is considered a feature. CNNs use mathematical operations to process input data through multiple layers of interconnected networks, similar to the operations within the visual system of the human brain. When we look at an image, we look for certain characteristics to narrow down what the image can be. For a bone tumor image, we have a list of predefined features of what constitutes a malignant tumor, such as ill-defined borders, cortical destruction, aggressive periosteal reaction, etc. Recognizing these features, we classify the image as malignant bone tumor.

Neuroscientists David Hubel and Torsten Wiesel found that subpopulations of neurons in the visual cortex have different functions [ , ]: while some neurons detect edge orientation, other neurons determine different visual features. Together, these multiple layers of processing allow comprehensible image formation. To translate this to machine language, the machine uses pixels to delineate edges and curves within an image while CNNs process this information to outline specific structures, such as bones or tumors. CNNs provide image classifications with minimal data preprocessing and assign learnable weights to their multiple interconnected algorithms to define the importance of a specific feature, such as cortical destruction or bone matrix formation. Multiple layers in the algorithm can extract features from the input data such that the algorithm “learns” to predict a prespecified output, such as a certain bone tumor. When many convolutional layers are stacked together, they are called deep-CNNs.

Deep-CNNs use a variation of multilayer perceptrons (“neurons”) with shared-weights architecture and translation invariance characteristics to reconstruct high-level features from incomplete input data. Deep-CNNs can alleviate the need for explicit feature extraction, simplify and accelerate the feature selection process, and integrate the processes of feature extraction, selection, and supervised classification in the same deep architecture (called end-to-end learning ) [ ]. With these advantages, deep-CNNs are more accurate, faster, vendor independent, and less prone to “hallucination artifacts” (i.e., enhanced random signal effects) when compared to previous approaches. Even so, there are some limitations. Because CNNs are trained using overlapping patches of images across iterations, computational redundancy occurs and only local features in the image subsections can be extracted [ ]. Therefore, CNNs can be computationally expensive and miss global image semantics. Many enhancements, notably LeNet [ ], U-Net [ ], ResNet [ ], AlexNet [ ], and fully convolutional network (FCN) [ ], have been proposed to address some of CNNs' limitations. CNNs are continuously being refined and have shown promise in most, if not all, aspects of bone cancer imaging.

AI-assisted tumor detection and segmentation

The information from medical image data provides crucial information for the management of bone tumors. For example, the only cure for a nonmetastasized osteosarcoma to date is complete tumor resection. If an AI algorithm would underestimate tumor extent, a subtotal tumor resection could lead to local tumor recurrence and death of a patient who might have otherwise survived. If an AI algorithm would overestimate tumor extent, an unnecessarily large resection could lead to unnecessary morbidities. Hence, AI needs to detect and delineate bone tumors accurately and measure the tumor size, volume, and/or metabolism according to standardized criteria.

You're Reading a Preview

Become a Clinical Tree membership for Full access and enjoy Unlimited articles

Become membership

If you are a member. Log in here