The Future of AI in Radiation Therapy: LLM-Powered Multimodal Models for Precision Target Contouring

 Imagine a world where radiation therapy no longer relies solely on time-consuming manual contouring by oncologists—where AI can seamlessly analyze imaging data, interpret clinical records, and define treatment regions with precision that rivals human expertise.

For decades, radiation therapy has depended on specialists to manually outline target volumes—a process that is both labor-intensive and prone to interobserver variability. Even with the rise of AI-powered segmentation models, most solutions have relied only on imaging data, failing to consider crucial clinical information such as tumor staging, pathology reports, and patient history. This limitation has hindered automation and consistency in treatment planning, leaving oncologists burdened with manual adjustments.

This is where LLMSeg comes in—a groundbreaking LLM-driven multimodal AI model designed to transform target volume contouring. By integrating both imaging and clinical text data, LLMSeg achieves unprecedented accuracy and contextual awareness, bridging the gap between AI automation and expert decision-making.

This article explores the research behind "LLM-drivenMultimodal Target Volume Contouring in Radiation Oncology" (Nature communications, 24 October 2024), a study that introduces LLMSeg as a solution to one of the most persistent challenges in AI-assisted radiation therapy. We will delve into the limitations of conventional AI models, how LLMSeg overcomes these barriers, and what this innovation means for the future of medical AI.

🚀 How does LLMSeg redefine the future of radiation therapy? Let’s explore.

Futuristic AI-powered radiation therapy system integrating CT scans and clinical data for precision oncology. The image showcases an advanced digital interface where an AI model overlays tumor contours on a CT scan while simultaneously analyzing pathology reports and patient treatment history. A medical professional reviews AI-generated segmentation on a holographic screen in a high-tech hospital setting. The interface text is clear, bold, and highly legible, enhancing medical imaging readability. This represents the future of AI-driven multimodal cancer treatment and automated radiation therapy planning

Unveiling the Challenges: Why Traditional AI Falls Short in Radiation Therapy

1. Advancements in AI-Based Medical Image Analysis

Artificial intelligence (AI) has undergone rapid advancements in recent years, revolutionizing various fields of healthcare, particularly in medical image analysis. The integration of deep learning models has enabled automated diagnosis and treatment planning, significantly enhancing clinical decision-making.

Traditionally, AI models in medical imaging have been single-modality systems, relying on either visual data (e.g., CT, MRI, and X-ray images) or text-based clinical records (e.g., electronic medical records, pathology reports). These unimodal models have shown success in specific tasks, such as tumor segmentation and anomaly detection. However, they fail to capture the full complexity of clinical decision-making, where multimodal data—including imaging, histopathological findings, genomic markers, and treatment history—are crucial.

As a result, recent research has shifted toward multimodal AI models that integrate diverse sources of patient data to achieve more accurate, personalized, and context-aware predictions. This shift is particularly significant in radiation oncology, where treatment planning depends on both imaging and clinical insights.

 

2. Limitations of Unimodal AI Models

While AI has significantly improved image processing capabilities, existing unimodal AI models face critical limitations when applied to real-world clinical workflows. These limitations arise primarily from their inability to synthesize non-visual clinical information, which is essential for accurate radiation therapy planning.

2.1 Limitations of Image-Based AI Models

  • Traditional segmentation models can accurately identify anatomical structures in CT and MRI scans, but they do not incorporate clinical data, such as tumor stage, genetic markers, and treatment history.
  • Tumors with similar visual characteristics may require different treatment plans based on histological and pathological factors, which are not discernible from imaging alone.
  • Variability in imaging protocols and scanner settings across institutions can cause significant performance drops in AI models that rely solely on image-based segmentation.

 

2.2 Limitations of Text-Based AI Models

  • NLP-based models trained on clinical text can extract patient information from medical records, but they lack the ability to directly correlate this information with imaging data.
  • While text-based AI can assist in treatment planning by analyzing guidelines and clinical notes, it cannot automatically generate precise tumor contouring without visual input.

 

As a result, current AI models remain highly dependent on human oversight, requiring clinicians to manually integrate AI-generated outputs with other patient-specific data. This hinders full automation and limits AI’s impact in clinical practice.

 

3. Importance of Target Volume Contouring in Radiation Oncology

In radiation oncology, one of the most crucial steps in treatment planning is target volume contouring—the process of defining the exact regions where radiation should be delivered while sparing healthy tissues.

The key components of target volume contouring include:

  1. Computed Tomography (CT) Simulation
    • A pre-treatment CT scan is performed to visualize the tumor and surrounding organs.
  2. Manual Delineation of Target Volumes and Organs-at-Risk (OARs)
    • Radiation oncologists manually segment the tumor and OARs using imaging data and patient history.
    • This step is highly time-consuming, taking hours to complete, and is prone to interobserver variability.
  3. Radiation Treatment Planning
    • Based on the delineated target volume, a personalized radiation dose is assigned to optimize tumor control while minimizing damage to healthy tissues.
  4. Treatment Delivery and Monitoring
    • Radiation therapy is administered over multiple sessions, with periodic assessments to adjust for changes in tumor size or patient response.

 

Since target volume definition directly impacts treatment success, any inaccuracies in contouring can lead to:
Under-treatment, where insufficient radiation is delivered to the tumor, increasing the risk of recurrence.
Over-treatment, where excessive radiation affects healthy tissues, leading to unnecessary side effects.

Thus, precise and consistent contouring is critical for maximizing therapeutic outcomes in radiation oncology.

 

4. Challenges of Traditional Contouring Methods

Currently, target volume delineation is performed manually by radiation oncologists or with the aid of traditional AI segmentation models. However, both approaches have inherent challenges.

4.1 Manual Contouring by Experts

  • Time-Consuming: The process requires hours of meticulous work, increasing the workload of specialists.
  • Variability Between Experts: Different oncologists may define target volumes differently, leading to interobserver variability.
  • Complex Decision-Making: Oncologists must consider tumor histology, stage, lymph node involvement, and prior treatments, making the task highly complex.

 

4.2 Traditional AI-Based Auto-Segmentation

  • Limited to Imaging Data: Most AI models rely solely on CT or MRI scans, ignoring crucial clinical variables such as tumor grade and genetic markers.
  • Poor Generalization: Models trained on data from a single institution often perform poorly on external datasets from different hospitals, requiring extensive retraining.
  • Lack of Clinical Context: AI-generated contours do not adapt to different treatment protocols, as they lack awareness of the oncologist’s decision-making process.

 

These limitations underscore the need for an AI model that integrates both imaging and clinical data to improve accuracy and consistency in target volume delineation.

 

5. The Need for an LLM-Driven Multimodal Approach

To address these challenges, researchers have begun exploring Large Language Models (LLMs) in combination with multimodal AI to create more sophisticated and clinically relevant solutions. The LLMSeg model, introduced in this study, represents a novel approach that integrates both visual and textual clinical information to enhance target volume contouring.

5.1 Key Advantages of LLM-Driven Multimodal AI

  1. Combining Imaging and Clinical Data
    • Unlike traditional AI models, LLMSeg incorporates not only CT images but also clinical text data (e.g., tumor stage, pathology reports, and treatment history).
  2. Simulating Oncologists' Decision-Making
    • Instead of segmenting tumors based on visual appearance alone, LLMSeg mimics the decision-making process of expert oncologists by considering multimodal patient data.
  3. Improved Generalization and Robustness
    • By integrating textual clinical knowledge, the model achieves higher accuracy across diverse datasets, making it more adaptable to different hospitals and imaging protocols.
  4. Enhanced Data Efficiency
    • Traditional AI models require large datasets for training, but LLMSeg can achieve strong performance even with limited data, improving feasibility in real-world clinical settings.

 

The introduction of LLMSeg marks a significant step toward more intelligent and clinically aware AI systems in radiation oncology. It bridges the gap between unimodal AI limitations and the need for context-driven automation in treatment planning.

 

The evolution of AI in medical imaging has led to breakthroughs in automated segmentation and diagnosis. However, traditional unimodal AI models remain limited in clinical decision-making, as they lack the ability to integrate patient-specific textual data alongside imaging data.

In radiation oncology, target volume contouring is a complex, multimodal task, requiring input from both imaging and clinical history. Existing AI models fail to capture this complexity, necessitating a novel approach that integrates LLMs with multimodal data.

The LLMSeg model, proposed in this study, represents a groundbreaking advancement in AI-driven target volume delineation. By leveraging large language models and cross-attention mechanisms, LLMSeg enables context-aware segmentation, closely mirroring oncologists' expertise.

 

Introducing LLMSeg: A Multimodal AI Breakthrough for Precision Target Contouring

1. Overview of LLMSeg: A Multimodal AI Model for Target Volume Contouring

The study introduces LLMSeg, a large language model (LLM)-driven multimodal AI designed to enhance the accuracy and efficiency of target volume contouring (TVC) in radiation oncology. Unlike conventional unimodal AI models, which rely solely on imaging data, LLMSeg integrates both textual clinical information and imaging data to improve context-aware segmentation.

Overview of proposed LLMSeg

1.1 Key Features of LLMSeg

Multimodal Integration: Combines CT images and patient-specific clinical data (tumor stage, pathology reports, treatment history).
LLM-Driven Decision Support: Utilizes large language models to interpret clinical text and guide segmentation.
Bidirectional Feature Alignment: Employs cross-attention mechanisms to align image features with textual information.
Generalization and Robustness: Achieves high accuracy across different datasets and institutions, overcoming data distribution shifts.
Data-Efficient Training: Maintains strong performance even with limited training data, making it suitable for real-world applications.

 

2. Technical Breakdown of LLMSeg Model Architecture

2.1 How LLMSeg Works: The Multimodal Learning Process

LLMSeg is built on a deep learning architecture that combines:

  1. A 3D Image Encoder: Extracts spatial and structural features from CT scans.
  2. A Pre-trained Large Language Model (LLM): Processes and understands clinical text.
  3. A Cross-Attention Alignment Module: Fuses textual and imaging data to generate accurate target volume contours.

2.1.1 Interactive Alignment Mechanism

The core innovation of LLMSeg lies in its ability to align text and image features bidirectionally:

  • Text-to-Image Alignment: Extracted text embeddings influence image segmentation decisions.
  • Image-to-Text Alignment: Features from CT scans guide the contextual interpretation of clinical data.
  • Multi-Level Cross-Attention: Ensures deep feature fusion at multiple network layers, improving segmentation accuracy.

 

2.2 Model Training and Optimization

  • Pre-training Strategy: LLMSeg utilizes a pre-trained LLM (LLaMA-7B) fine-tuned for medical data processing.
  • Loss Function: Combines Cross-Entropy (CE) loss and Dice loss to balance classification accuracy and shape consistency.
  • Optimization Algorithm: Uses AdamW optimizer for stable and efficient training.

 

3. Performance Comparison: LLMSeg vs. Conventional AI Models

To validate LLMSeg, the researchers conducted extensive experiments on breast cancer and prostate cancer datasets. The model was evaluated against existing unimodal AI models, including:

  • 3D U-Net
  • SegMamba
  • UNETR
  • HIPIE (State-of-the-art vision-language segmentation model)
  • ConTEXTualNet (Multimodal segmentation model for 3D images)

3.1 Accuracy and Robustness in Target Volume Contouring

3.1.1 Dice Coefficient and IoU Metrics

The Dice coefficient and Intersection over Union (IoU) are commonly used to measure segmentation accuracy:

Model

Internal Test

(Dice ↑)

External Test #1

(Dice ↑)

External Test #2

(Dice ↑)

3D U-Net

0.807

0.731

0.444

HIPIE

0.743

0.736

0.617

ConTEXTualNet

0.819

0.815

0.826

LLMSeg (Ours)

0.829

0.822

0.844

LLMSeg outperformed all baseline models across internal and external test datasets, demonstrating its superior generalization ability.

 

3.1.2 Generalization Across Different Institutions

  • Unimodal AI models showed a sharp drop in performance when tested on external datasets due to variations in CT scanners, acquisition protocols, and patient demographics.
  • LLMSeg maintained high performance across all datasets, proving its robustness in real-world clinical settings.

 

4. Expert Evaluation: Clinical Validation of LLMSeg

4.1 Clinician-Based Assessment Rubrics

To assess clinical usability, radiation oncologists evaluated LLMSeg's segmentation quality using five expert-defined rubrics:

  1. Laterality (correctly identifying tumor side)
  2. Surgery Type Consideration (distinguishing between mastectomy and breast-conserving surgery)
  3. Volume Definition (accurate delineation of target volume)
  4. Coverage (ensuring complete treatment area)
  5. Integrity (absence of unnecessary regions in segmentation)

 

4.2 Expert Scoring Results

Model

Laterality (1pt)

Surgery Type (1pt)

Volume Definition (1.5pt)

Coverage (1pt)

Integrity (0.5pt)

Total (5pt)

Vision-Only AI

0.786

0.887

0.900

0.478

0.216

3.267

LLMSeg (Ours)

0.990

0.987

1.142

0.602

0.253

3.973

  • LLMSeg outperformed vision-only AI models in all categories, particularly in laterality, volume definition, and surgery type recognition, highlighting its clinical relevance.

 

5. Data Efficiency: LLMSeg’s Performance with Limited Training Data

One of the key challenges in medical AI is the scarcity of high-quality annotated datasets. The researchers tested LLMSeg’s data efficiency by progressively reducing the training dataset size.

5.1 Effect of Data Reduction on Dice Score

Training Data Size

Vision-Only AI (Dice ↑)

LLMSeg (Dice ↑)

100%

0.807

0.829

40%

0.700

0.801

20%

0.500

0.793

  • Even with only 40% of the training data, LLMSeg maintained a Dice score above 0.8, while the vision-only AI model suffered a sharp performance drop.
  • At 20% of the dataset, vision-only AI failed to perform accurate segmentation, while LLMSeg still delivered clinically acceptable results.

 

6. LLMSeg’s Adaptability to Different Cancer Types

Beyond breast cancer radiotherapy, the researchers tested LLMSeg’s performance on prostate cancer cases.

6.1 Expert Scoring in Prostate Cancer Cases

Model

Primary Site (1pt)

Volume Definition (1.5pt)

Coverage (1pt)

Integrity (0.5pt)

Total (4pt)

Vision-Only AI

0.470

0.717

0.313

0.171

1.670

LLMSeg (Ours)

0.583

0.951

0.379

0.249

2.162

  • LLMSeg achieved significantly higher scores in expert-based evaluation, proving its potential for broader application in radiation oncology.

 

The findings demonstrate that LLMSeg surpasses traditional AI models in:

Accuracy and robustness across different datasets.
Expert validation confirming clinical relevance.
Data efficiency, requiring fewer training samples.
Adaptability to multiple cancer types.

 

The Future of AI-Driven Radiation Therapy: Predictions and Emerging Trends

1. The Expanding Potential of LLM-Driven Multimodal AI

The LLMSeg model, proposed in this study, represents a breakthrough in integrating clinical (text) and imaging data to enhance target volume contouring (TVC) accuracy. However, its potential extends beyond radiation oncology, offering the possibility of transforming multiple areas of medicine through multimodal AI integration.

1.1 Expanding AI Applications Beyond Radiation Oncology

Pathology: AI-driven cancer diagnosis by integrating histopathological findings with genomic mutations
Precision Medicine: Personalized treatment plans based on clinical history, genomic markers, and therapy response
Surgical Planning AI: Combining preoperative imaging and treatment history for optimal surgical strategies
Electronic Medical Record (EMR) Analysis: Automated clinical documentation summarization and decision support using LLMs

The core concept of multimodal AI suggests that its potential is not limited to radiation therapy, but can redefine clinical workflows across various medical disciplines.

 

2. The Evolution of Medical AI and Its Clinical Applications

LLMSeg and similar multimodal AI models signal a shift toward a more context-aware, physician-assisted AI ecosystem. The future of medical AI will focus on the following key developments:

2.1 Integration and Automation of Multimodal Data

AI is evolving from processing a single data type to integrating comprehensive patient information
Seamless analysis of imaging, genomics, clinical data, and pharmacological responses in a unified system
Potential collaboration with automated clinical decision support (CDS) systems

For example, instead of AI merely analyzing CT scans,
🩸 It could integrate genomic mutation analysis → interpret pathology reports → recommend personalized oncology treatments
This multi-step AI-driven diagnostic and treatment process is expected to become a reality.

 

2.2 Enhancing Collaboration Between AI and Healthcare Professionals

AI is not meant to replace medical professionals but to augment their decision-making by providing enhanced analytical support.

In radiation oncology, AI-based contouring models assist oncologists in finalizing target volumes
AI can analyze clinical data and suggest treatment plans, while physicians retain control over final decisions
The rise of Explainable AI (XAI) will enhance AI trustworthiness and transparency

Thus, instead of an "AI-dominant decision-making process where physicians approve results,"
🩺 The future will see "AI assisting physicians, while humans make the final call."

 

2.3 The Fusion of Medical AI and Large Language Models (LLMs)

The integration of LLMs into medical AI paves the way for next-generation intelligent clinical models.

Traditional Medical AI

LLM-Integrated Medical AI

Processes only imaging data

Integrates multimodal data (imaging + clinical records)

Limited to structured data

Capable of analyzing free-text clinical notes

Performance declines with small datasets

Leverages LLM capabilities to achieve robust results even with limited data

Supports simple decision-making

Provides personalized treatment recommendations

Thus, the convergence of LLMs and medical AI is likely to revolutionize clinical decision-making rather than merely automating specific tasks.

 

3. The Future of LLM-Driven AI in Radiation Oncology

Radiation oncology is one of the most promising fields for the early adoption of LLM-based AI due to several reasons:

3.1 The High Data Dependency of Radiation Therapy

Radiation therapy requires a complex interplay of tumor size, staging, genomic markers, and radiation dosage parameters
Conventional AI models that analyze only CT scans struggle to incorporate essential clinical variables
Multimodal AI can seamlessly integrate all these factors, optimizing treatment planning

 

3.2 Potential for Automated Radiation Therapy Planning

Current radiation therapy planning is time-consuming, often taking hours or days
AI could automatically generate target volume contours and recommend optimal radiation dosages
This could lead to Automated Radiation Planning (ARP), minimizing manual workload

 

4. Technological and Ethical Challenges & Solutions

For AI to achieve widespread adoption in healthcare, certain technological and ethical challenges must be addressed.

4.1 Technological Challenges

Data Bias Issues

  • AI models trained on limited datasets from specific institutions may perform poorly on diverse patient populations
  • Solution: Implementing Federated Learning to train models on multi-institutional data without centralized data storage

Uncertainty in AI Decision-Making

  • AI models must be designed to account for uncertainty in clinical scenarios
  • Solution: Utilizing Explainable AI (XAI) techniques to improve model transparency

 

4.2 Ethical Considerations

Can AI-generated decisions be trusted?

  • AI models should clearly explain their decision-making processes, allowing physicians to validate and adjust recommendations
  • Solution: Implementing explainability features such as confidence maps

Data Privacy and Security

  • Patient medical data is subject to strict privacy regulations such as GDPR (EU) and HIPAA (US)
  • Solution: AI systems must comply with privacy-preserving protocols and anonymization techniques

 

5. Conclusion: A Leap Toward Multimodal AI-Driven Medical Innovation

The LLMSeg model proposed in this study demonstrates a novel approach to target volume contouring in radiation oncology by integrating clinical and imaging data.

It overcomes the limitations of unimodal AI models by considering clinical data alongside imaging data
Achieves superior data efficiency, strong generalization performance, and expert validation
Has potential applications beyond radiation oncology, including pathology, precision medicine, and surgical planning
Marks the convergence of LLMs and medical AI, paving the way for AI-driven precision medicine

 

Revolutionizing Cancer Treatment: Key Takeaways and Final Thoughts

1. Summary of Key Findings

In this study, LLMSeg, a Large Language Model (LLM)-driven multimodal AI, was introduced to revolutionize target volume contouring (TVC) in radiation oncology. Unlike conventional unimodal AI models, which rely solely on imaging data, LLMSeg integrates both clinical text data and imaging information to provide more accurate and context-aware segmentation.

1.1 Overcoming the Limitations of Unimodal AI

Traditional AI models for segmentation struggle with handling textual clinical data, leading to poor generalization in real-world scenarios.
LLMSeg introduces cross-attention mechanisms to align textual clinical knowledge with imaging data, simulating expert decision-making.
It demonstrates high generalization ability, maintaining robust performance across varied datasets from different hospitals.

 

1.2 Performance Superiority of LLMSeg

  • LLMSeg outperforms traditional segmentation models, achieving a higher Dice coefficient, improved IoU, and lower HD-95 scores.
  • It is validated by expert evaluations, receiving significantly higher clinical relevance scores compared to vision-only AI models.
  • The model proves to be data-efficient, maintaining high performance even when trained on smaller datasets.

 

1.3 Expanding the Scope of Multimodal AI in Medicine

  • Beyond radiation oncology, LLMSeg can be adapted for pathology, precision medicine, and AI-driven clinical decision support.
  • LLM-driven multimodal AI has the potential to transform traditional medical workflows by integrating diverse patient data sources.
  • The study highlights the critical role of Explainable AI (XAI) in ensuring trust and transparency in AI-assisted decision-making.

 

2. Clinical Implications of LLMSeg

The introduction of LLMSeg marks a significant advancement in the field of AI-driven medical imaging and radiation therapy.

2.1 Impact on Radiation Oncology

Time Efficiency: LLMSeg reduces the need for extensive manual contouring, significantly saving time for oncologists.
Consistency & Accuracy: The AI model minimizes interobserver variability, ensuring standardized treatment planning.
Automation Potential: LLMSeg paves the way for Automated Radiation Planning (ARP), where AI-driven contouring could become a clinical standard.

 

2.2 AI-Augmented Medical Decision-Making

Multimodal AI provides a holistic view of patient data, combining imaging, pathology reports, and genomics for comprehensive decision-making.
Physicians retain control over final treatment plans, using AI-generated insights to enhance clinical judgment rather than replace human expertise.
LLMs improve data accessibility, allowing AI to interpret and summarize unstructured clinical text efficiently.

 

3. Challenges & Future Directions

While LLMSeg demonstrates remarkable progress, several challenges remain for its widespread adoption in clinical settings.

3.1 Data Availability & Generalization

Access to diverse and well-annotated medical datasets remains a challenge.
Solution: Expanding training datasets through federated learning across multiple institutions to improve model generalizability.

 

3.2 Explainability & Trustworthiness

Clinicians may hesitate to trust AI-generated contours without proper justification.
Solution: Enhancing Explainable AI (XAI) capabilities, providing visual confidence maps and rationale for AI-driven recommendations.

 

3.3 Regulatory & Ethical Considerations

Medical AI must comply with strict data privacy regulations such as GDPR and HIPAA.
Solution: Implementing privacy-preserving AI techniques, ensuring secure data handling and ethical deployment of AI in healthcare.

 

4. The Future of LLM-Driven AI in Medicine

The success of LLMSeg highlights a broader shift in medical AI—toward truly multimodal, context-aware AI systems.

4.1 The Path to AI-Powered Personalized Medicine

Future AI systems will not only analyze medical images but also integrate genomic data, EMR records, and patient histories.
AI-powered precision medicine will enable tailored treatment strategies, improving patient outcomes and minimizing side effects.

 

4.2 Expanding AI Applications in Healthcare

Automated cancer diagnostics, where AI assists in early detection by analyzing pathology slides and genetic markers.
Surgical planning AI, providing real-time recommendations based on multimodal patient data.
AI-assisted treatment response prediction, using patient-specific data to adjust therapies dynamically.

 

4.3 The Role of LLMs in Future AI Models

Large Language Models (LLMs) will continue to revolutionize medical AI, enabling automated clinical documentation, enhanced decision support, and improved patient care.
Future advancements in multimodal learning will enable AI models to process complex real-world data more effectively, leading to improved diagnostic accuracy.

 

5. Conclusion: A Paradigm Shift in AI-Driven Medical Imaging

The LLMSeg model is a significant milestone in AI-driven radiation oncology, demonstrating how multimodal AI can outperform conventional approaches.

🔹 By integrating clinical text with imaging data, LLMSeg enables more precise, context-aware segmentation.
🔹 Its robust performance across datasets confirms its potential for real-world clinical applications.
🔹 The success of LLM-driven AI marks the beginning of a new era in medical imaging and AI-assisted decision-making.

The future of AI in medicine lies in multimodal integration, automation, and personalization. As AI technologies continue to evolve, LLM-driven multimodal models like LLMSeg will play a crucial role in shaping the next generation of medical AI systems—enhancing efficiency, accuracy, and ultimately, patient care.

 

What kind of new future did this article inspire you to imagine? Feel free to share your ideas and insights in the comments! I’ll be back next time with another exciting topic. Thank you! 😊

Comments

Popular posts from this blog

AI-Based Electronic Polymer Manufacturing: How Polybot is Revolutionizing Next-Generation Thin Film Technology

The Future of Sustainable Protein: How Single-Cell Protein (SCP) Can Revolutionize Food Production