The future of mathematical oncology in the age of AI.
1/5 보강
This perspective article discusses emerging advances at the interface of mechanistic modeling and data-driven machine learning, highlighting opportunities for AI to accelerate discovery, improve predi
APA
Rockne RC, Andersen M, et al. (2026). The future of mathematical oncology in the age of AI.. NPJ systems biology and applications, 12(1), 22. https://doi.org/10.1038/s41540-026-00656-9
MLA
Rockne RC, et al.. "The future of mathematical oncology in the age of AI.." NPJ systems biology and applications, vol. 12, no. 1, 2026, pp. 22.
PMID
41588010 ↗
Abstract 한글 요약
This perspective article discusses emerging advances at the interface of mechanistic modeling and data-driven machine learning, highlighting opportunities for AI to accelerate discovery, improve predictive modeling, and enhance clinical decision-making. We address critical limitations of current AI approaches and propose a perspective on a future where AI augments mechanistic rigor, clinical relevance, and human creativity under the umbrella of a redefined understanding of Mathematical Oncology.
🏷️ 키워드 / MeSH 📖 같은 키워드 OA만
📖 전문 본문 읽기 PMC JATS · ~36 KB · 영문
Introduction
Introduction
In May 2025, early-career researchers and established investigators in Mathematical Oncology and artificial intelligence (AI) met to discuss the current and future impact of AI technologies on the field. The focus of the meeting was on immersive learning, spirited debate, and collaborative visioning, moving away from a typical lecture-heavy format. The two-day meeting was held in Siracusa, Sicily, and was structured to help engagement and dialogue, with each day beginning with presentations led by early-career researchers focused on a specific topic, followed by group discussions. Several themes emerged from the discussions that form the basis of this perspective on the future of Mathematical Oncology in the age of AI.
In May 2025, early-career researchers and established investigators in Mathematical Oncology and artificial intelligence (AI) met to discuss the current and future impact of AI technologies on the field. The focus of the meeting was on immersive learning, spirited debate, and collaborative visioning, moving away from a typical lecture-heavy format. The two-day meeting was held in Siracusa, Sicily, and was structured to help engagement and dialogue, with each day beginning with presentations led by early-career researchers focused on a specific topic, followed by group discussions. Several themes emerged from the discussions that form the basis of this perspective on the future of Mathematical Oncology in the age of AI.
AI and the future of scientific discovery
AI and the future of scientific discovery
A central question was whether AI could eventually outpace human researchers in generating new knowledge and ideas. Several scenarios included whether AI could design viable molecules that don’t exist in nature, or discover new physical laws or mathematical conjectures. While some viewed this as speculative, others among us pointed to real-world examples where AI has already begun to do exactly that.
A central question was whether AI could eventually outpace human researchers in generating new knowledge and ideas. Several scenarios included whether AI could design viable molecules that don’t exist in nature, or discover new physical laws or mathematical conjectures. While some viewed this as speculative, others among us pointed to real-world examples where AI has already begun to do exactly that.
AI-driven discovery
AI-driven discovery
For instance, in drug discovery, a generative AI platform successfully designed a novel molecule for lung disease, which has since progressed to clinical trials1. Recent works have demonstrated how data-driven approaches can elucidate explicit mathematical formulations of the biological mechanisms governing tumor growth and treatment response from longitudinal series of data2,3. In continuum mechanics, data-driven computational methods based on regression and neural networks have been demonstrated to enable discovery of mathematical models representing the constitutive mechanical behavior of many different materials4,5. Additionally, an AI model identified new physics governing many-body systems in so called dusty plasma, correcting long-held assumptions. Even in experimental quantum optics, an algorithm called Theseus demonstrated its capacity for explainable AI by designing new, more efficient experimental setups and novel concepts that human physicists have later interpreted and applied6.
Meanwhile, in mathematics, AlphaGeometry demonstrated its ability to solve complex geometry problems at a level comparable to a human gold medalist at the International Mathematical Olympiad, showcasing a capacity for complex reasoning and proof generation7. Mathematics researchers have integrated AI with proof assistance softwares such as Lean to automate the process of generating and proving new mathematical results8,9. Perhaps the most famous example is the generation of protein structure from amino acid sequences generated by Alphafold10. While in most examples, AI is still assisting researchers, rather than working autonomously, it is envisioned that it will soon do so.
At the same time, AI excels at analysis of high-dimensional data, such as the in-depth characterization based on multi-omic profiles: AI and machine learning (ML) are transforming genetic analysis by providing effective tools to handle the large amounts of data generated by genome sequencing. These technologies are proving crucial for detecting and interpreting genetic variants, particularly in oncology. AI algorithms are trained on wide datasets of sequenced genomes to identify the presence of single nucleotide polymorphisms, indels, and structural variations that can discriminate true variants among sequencing errors with a high degree of accuracy, as well as perform analysis and interpretation for clinical use11,12.
Many at the conference noted that AI’s capabilities are bound by the data it learns from, with severely constrained out-of-distribution reasoning13. In domains where data is sparse or non-existent, human intuition and creativity remain irreplaceable. Think, for instance, of the creative leap of Albert Einstein in formulating the theory of relativity based on the state of knowledge at the beginning of the 20th century. This discussion led to a broader reflection: perhaps the future role of researchers is not to compete with AI, but to use AI as an accelerating tool to work at the frontier of knowledge and science, where data is scarce, uncertainty is high, new paradigms are needed, and where AI will be inherently limited. This acceleration would allow us to check data patterns, make predictions, and test, if not necessarily validate and verify, ideas nearly instantly, and is a way to spend more time understanding the results and asking better questions.
Indeed, the central goal of science is identifying and asking the next question, a task that currently AI is not well equipped to address. The medical sector, and in particular oncology, is one such application field that historically has been strongly supported by human experience and intuition. Changing treatment paradigms limit purely data-driven innovation, as data from contemporary patient cohorts will always be a step behind the next emerging therapy or clinical trial.
For instance, in drug discovery, a generative AI platform successfully designed a novel molecule for lung disease, which has since progressed to clinical trials1. Recent works have demonstrated how data-driven approaches can elucidate explicit mathematical formulations of the biological mechanisms governing tumor growth and treatment response from longitudinal series of data2,3. In continuum mechanics, data-driven computational methods based on regression and neural networks have been demonstrated to enable discovery of mathematical models representing the constitutive mechanical behavior of many different materials4,5. Additionally, an AI model identified new physics governing many-body systems in so called dusty plasma, correcting long-held assumptions. Even in experimental quantum optics, an algorithm called Theseus demonstrated its capacity for explainable AI by designing new, more efficient experimental setups and novel concepts that human physicists have later interpreted and applied6.
Meanwhile, in mathematics, AlphaGeometry demonstrated its ability to solve complex geometry problems at a level comparable to a human gold medalist at the International Mathematical Olympiad, showcasing a capacity for complex reasoning and proof generation7. Mathematics researchers have integrated AI with proof assistance softwares such as Lean to automate the process of generating and proving new mathematical results8,9. Perhaps the most famous example is the generation of protein structure from amino acid sequences generated by Alphafold10. While in most examples, AI is still assisting researchers, rather than working autonomously, it is envisioned that it will soon do so.
At the same time, AI excels at analysis of high-dimensional data, such as the in-depth characterization based on multi-omic profiles: AI and machine learning (ML) are transforming genetic analysis by providing effective tools to handle the large amounts of data generated by genome sequencing. These technologies are proving crucial for detecting and interpreting genetic variants, particularly in oncology. AI algorithms are trained on wide datasets of sequenced genomes to identify the presence of single nucleotide polymorphisms, indels, and structural variations that can discriminate true variants among sequencing errors with a high degree of accuracy, as well as perform analysis and interpretation for clinical use11,12.
Many at the conference noted that AI’s capabilities are bound by the data it learns from, with severely constrained out-of-distribution reasoning13. In domains where data is sparse or non-existent, human intuition and creativity remain irreplaceable. Think, for instance, of the creative leap of Albert Einstein in formulating the theory of relativity based on the state of knowledge at the beginning of the 20th century. This discussion led to a broader reflection: perhaps the future role of researchers is not to compete with AI, but to use AI as an accelerating tool to work at the frontier of knowledge and science, where data is scarce, uncertainty is high, new paradigms are needed, and where AI will be inherently limited. This acceleration would allow us to check data patterns, make predictions, and test, if not necessarily validate and verify, ideas nearly instantly, and is a way to spend more time understanding the results and asking better questions.
Indeed, the central goal of science is identifying and asking the next question, a task that currently AI is not well equipped to address. The medical sector, and in particular oncology, is one such application field that historically has been strongly supported by human experience and intuition. Changing treatment paradigms limit purely data-driven innovation, as data from contemporary patient cohorts will always be a step behind the next emerging therapy or clinical trial.
Agentic AI in mathematical oncology
Agentic AI in mathematical oncology
Agentic AI, or the orchestration of multiple large language models (LLMs) as ‘agents’ to execute multi-component workflows, was identified as an advance that necessitates a paradigm shift in how researchers in the field of Mathematical Oncology operate. The group discussed the inevitable development of workflows that mimic the process of mathematical modeling: take experimental or clinical data, construct a family of plausible mechanistic mathematical models, simulate top candidate models, and generate computational codes to reproduce the results in a matter of seconds. Such agentic workflows could be a transformative change to the way Mathematical Oncology research is conducted14.
Beyond speed and efficiency, agentic AI offers opportunities for reproducibility and transparency by encoding each step of the modeling pipeline in machine-readable form, potentially reducing human error and facilitating collaborative science across institutions. Moreover, these systems could act as adaptive research assistants, autonomously designing computational experiments, integrating new data streams, or proposing model refinements in response to emerging therapies or clinical outcomes. Early demonstrations in adjacent fields underscore the feasibility of such approaches: multi-agent systems have been used for automated software development, hypothesis generation in the physical sciences, and self-directed optimization of scientific writing15–17. In oncology, embedding agentic AI into the modeling pipeline could accelerate the translation of mechanistic insights into clinical strategies, bridging the gap between abstract models and actionable patient-specific predictions.
Nevertheless, challenges remain in ensuring scientific validity, interpretability, and alignment with ethical principles, as agentic AI systems gain autonomy in shaping research trajectories. Addressing these issues will require not only technical safeguards but also new frameworks for oversight, education, and collaborative engagement between human and AI researchers18. The importance of integrating AI approaches with mechanistic understanding was also raised as a potential opportunity. By embedding structural and biochemical data into AI systems, researchers can envision multi-scale models that connect atomic-level changes to tissue-level dynamics and patient outcomes. For instance, the concept of temperature is an emergent property that is not defined for an individual molecule. This multi-scale abstraction not only enhances predictive accuracy but also improves interpretability, an essential feature in clinical decision-making. As several participants noted, the challenge with the effective use of agentic AI in Mathematical Oncology is educational: how do we train the next generation to think across scales, disciplines, and leverage AI as an accelerating tool in the development of mathematical models in oncology?
Agentic AI, or the orchestration of multiple large language models (LLMs) as ‘agents’ to execute multi-component workflows, was identified as an advance that necessitates a paradigm shift in how researchers in the field of Mathematical Oncology operate. The group discussed the inevitable development of workflows that mimic the process of mathematical modeling: take experimental or clinical data, construct a family of plausible mechanistic mathematical models, simulate top candidate models, and generate computational codes to reproduce the results in a matter of seconds. Such agentic workflows could be a transformative change to the way Mathematical Oncology research is conducted14.
Beyond speed and efficiency, agentic AI offers opportunities for reproducibility and transparency by encoding each step of the modeling pipeline in machine-readable form, potentially reducing human error and facilitating collaborative science across institutions. Moreover, these systems could act as adaptive research assistants, autonomously designing computational experiments, integrating new data streams, or proposing model refinements in response to emerging therapies or clinical outcomes. Early demonstrations in adjacent fields underscore the feasibility of such approaches: multi-agent systems have been used for automated software development, hypothesis generation in the physical sciences, and self-directed optimization of scientific writing15–17. In oncology, embedding agentic AI into the modeling pipeline could accelerate the translation of mechanistic insights into clinical strategies, bridging the gap between abstract models and actionable patient-specific predictions.
Nevertheless, challenges remain in ensuring scientific validity, interpretability, and alignment with ethical principles, as agentic AI systems gain autonomy in shaping research trajectories. Addressing these issues will require not only technical safeguards but also new frameworks for oversight, education, and collaborative engagement between human and AI researchers18. The importance of integrating AI approaches with mechanistic understanding was also raised as a potential opportunity. By embedding structural and biochemical data into AI systems, researchers can envision multi-scale models that connect atomic-level changes to tissue-level dynamics and patient outcomes. For instance, the concept of temperature is an emergent property that is not defined for an individual molecule. This multi-scale abstraction not only enhances predictive accuracy but also improves interpretability, an essential feature in clinical decision-making. As several participants noted, the challenge with the effective use of agentic AI in Mathematical Oncology is educational: how do we train the next generation to think across scales, disciplines, and leverage AI as an accelerating tool in the development of mathematical models in oncology?
Mechanistic and data-driven modeling on real-world clinical data
Mechanistic and data-driven modeling on real-world clinical data
In cancer applications, the very structure of the data often dictates model design. In practice, clinical data may be longitudinal, sparse, irregularly sampled, and high-dimensional, combining patient covariates, imaging, multi-omics, and circulating biomarkers. While each snapshot can be rich, uneven time points and privacy-driven restrictions on data sharing mean that most longitudinal clinical datasets are “small” in AI terms. Tens, hundreds to at best a few thousand patient trajectories are common, given the data origins from either routinely acquired data, large-scale phase 3 clinical trials, or, in most cases, pooled specific samples across institutions. Merging data from heterogeneous sources, such as different treatment centers, harbors additional challenges regarding data harmonization and comparable protocols, while clinical application of trained frameworks requires extensive evaluation of model robustness and generalization in light of limited sample numbers, and possible data bias. Evolving clinical protocols and instrumentation further add to this challenge, implying that data sparsity and irregular sampling will remain persistent challenges in the clinical data domain without major regulatory and political paradigm shifts.
Purely data-driven models may struggle to digest these complex real-world data realistically, with a possible high chance of overfitting, given the flexibility of these universal function approximators. Mechanistic models, grounded in biological laws, such as tumor growth equations, pharmacokinetics/pharmacodynamics (PK/PD), and tissue mechanics, offer complementary strengths. Physical constraints such as mass conservation and nonnegativity avoid nonsensical extrapolations. Perhaps the most critical feature of mechanistic models is predicting beyond the data, to infer never seen before treatment strategies based on mechanistic knowledge and insights. However, classical mechanistic models often lack the flexibility to integrate unstructured, high-dimensional data or to adapt to unanticipated biology, for example, the emergence of resistant cell populations.
Understanding causal and mechanistic interactions remains a major challenge for AI/ML methods. Temporally dense longitudinal datasets present a significant opportunity to move beyond static associations, enabling the capture of dynamic relationships that evolve over time. However, a particularly critical scenario arises when the assumption of stationarity is violated, a common occurrence in biological systems, where the focus is often on studying transitions between states, such as from health to disease or during stages of development. Identifying the key genetic or epigenetic events that trigger these transitions is especially challenging given their transient and dynamic nature. Importantly, the events that cause state transitions are, by definition, non-stationary and may not align with those exhibiting the greatest differences between steady states. At the molecular level, this is a frequent scenario, with the primary interest often centered on elucidating the sequence of molecular events responsible for transitions such as protein folding/unfolding or activation/inactivation processes.
In cancer applications, the very structure of the data often dictates model design. In practice, clinical data may be longitudinal, sparse, irregularly sampled, and high-dimensional, combining patient covariates, imaging, multi-omics, and circulating biomarkers. While each snapshot can be rich, uneven time points and privacy-driven restrictions on data sharing mean that most longitudinal clinical datasets are “small” in AI terms. Tens, hundreds to at best a few thousand patient trajectories are common, given the data origins from either routinely acquired data, large-scale phase 3 clinical trials, or, in most cases, pooled specific samples across institutions. Merging data from heterogeneous sources, such as different treatment centers, harbors additional challenges regarding data harmonization and comparable protocols, while clinical application of trained frameworks requires extensive evaluation of model robustness and generalization in light of limited sample numbers, and possible data bias. Evolving clinical protocols and instrumentation further add to this challenge, implying that data sparsity and irregular sampling will remain persistent challenges in the clinical data domain without major regulatory and political paradigm shifts.
Purely data-driven models may struggle to digest these complex real-world data realistically, with a possible high chance of overfitting, given the flexibility of these universal function approximators. Mechanistic models, grounded in biological laws, such as tumor growth equations, pharmacokinetics/pharmacodynamics (PK/PD), and tissue mechanics, offer complementary strengths. Physical constraints such as mass conservation and nonnegativity avoid nonsensical extrapolations. Perhaps the most critical feature of mechanistic models is predicting beyond the data, to infer never seen before treatment strategies based on mechanistic knowledge and insights. However, classical mechanistic models often lack the flexibility to integrate unstructured, high-dimensional data or to adapt to unanticipated biology, for example, the emergence of resistant cell populations.
Understanding causal and mechanistic interactions remains a major challenge for AI/ML methods. Temporally dense longitudinal datasets present a significant opportunity to move beyond static associations, enabling the capture of dynamic relationships that evolve over time. However, a particularly critical scenario arises when the assumption of stationarity is violated, a common occurrence in biological systems, where the focus is often on studying transitions between states, such as from health to disease or during stages of development. Identifying the key genetic or epigenetic events that trigger these transitions is especially challenging given their transient and dynamic nature. Importantly, the events that cause state transitions are, by definition, non-stationary and may not align with those exhibiting the greatest differences between steady states. At the molecular level, this is a frequent scenario, with the primary interest often centered on elucidating the sequence of molecular events responsible for transitions such as protein folding/unfolding or activation/inactivation processes.
Hybrid mechanistic learning
Hybrid mechanistic learning
To leverage both mechanistic knowledge and high-dimensional data strengths, hybrid mechanistic learning frameworks overlay data-driven adaptability into an equation-based structure2,19–24. Hybrid methods reduce sample complexity by incorporating known dynamics, making irregular and missing data less problematic, and allowing simulation under unseen conditions. Perhaps the most well known example is Physics-Informed Neural Networks (PINNs) that embed governing equations in the loss function to fit sparse data while satisfying known equations25.
Similarly, Neural ordinary differential equation (ODE) frameworks learn continuous-time vector fields parameterized by neural networks and can integrate pharmacological insights to discover growth laws from truncated tumor-volume data, improving extrapolation and survival predictions26. Operator learning methods, such as DeepONet and Fourier Neural Operators, learn mappings between function spaces, enabling rapid prediction of tumor growth or treatment response in quantitative systems pharmacology27,28. In survival analysis, hybrid approaches incorporate mechanistic descriptors, for example, hazard functions derived from tumor growth laws or pharmacokinetic-pharmacodynamic (PK/PD) models, into deep learning (DL) models, improving interpretability and calibration in small-sample contexts29.
These hybrid strategies mitigate data sparsity by reducing the sample complexity needed to infer meaningful laws and by making irregular sampling less problematic. For instance, PINNs enforce derivative constraints directly, avoiding the need for evenly spaced grids, and neural ODEs treat each observation as a constraint on the learned vector field. This hybrid thinking forms a conceptual loop: data acquisition informs ML residuals, which refine mechanistic models; these in turn generate verifiable predictions, and the cycle continues. Mechanistic model-driven training, which uses mechanistic models to match patient dynamics as training tools for machine learning models or for cohort generation, circumvents the issue of limited patient training data and potentially allows gaining insights into the patient decisions that impact outcome30.
Realizing the potential of these hybrid approaches demands cross-disciplinary fluency. Expertise in biology, mathematics, and machine learning is essential, and training the next generation of researchers to think seamlessly across these domains is as critical as the algorithms themselves. In this integrated view, the most promising advances will likely come from methods that blend the hypothesis-driven rigor of mechanistic modeling, the adaptability of AI, and the geometric insights of topological analysis, all while grounded in the realities of data sparsity.
To leverage both mechanistic knowledge and high-dimensional data strengths, hybrid mechanistic learning frameworks overlay data-driven adaptability into an equation-based structure2,19–24. Hybrid methods reduce sample complexity by incorporating known dynamics, making irregular and missing data less problematic, and allowing simulation under unseen conditions. Perhaps the most well known example is Physics-Informed Neural Networks (PINNs) that embed governing equations in the loss function to fit sparse data while satisfying known equations25.
Similarly, Neural ordinary differential equation (ODE) frameworks learn continuous-time vector fields parameterized by neural networks and can integrate pharmacological insights to discover growth laws from truncated tumor-volume data, improving extrapolation and survival predictions26. Operator learning methods, such as DeepONet and Fourier Neural Operators, learn mappings between function spaces, enabling rapid prediction of tumor growth or treatment response in quantitative systems pharmacology27,28. In survival analysis, hybrid approaches incorporate mechanistic descriptors, for example, hazard functions derived from tumor growth laws or pharmacokinetic-pharmacodynamic (PK/PD) models, into deep learning (DL) models, improving interpretability and calibration in small-sample contexts29.
These hybrid strategies mitigate data sparsity by reducing the sample complexity needed to infer meaningful laws and by making irregular sampling less problematic. For instance, PINNs enforce derivative constraints directly, avoiding the need for evenly spaced grids, and neural ODEs treat each observation as a constraint on the learned vector field. This hybrid thinking forms a conceptual loop: data acquisition informs ML residuals, which refine mechanistic models; these in turn generate verifiable predictions, and the cycle continues. Mechanistic model-driven training, which uses mechanistic models to match patient dynamics as training tools for machine learning models or for cohort generation, circumvents the issue of limited patient training data and potentially allows gaining insights into the patient decisions that impact outcome30.
Realizing the potential of these hybrid approaches demands cross-disciplinary fluency. Expertise in biology, mathematics, and machine learning is essential, and training the next generation of researchers to think seamlessly across these domains is as critical as the algorithms themselves. In this integrated view, the most promising advances will likely come from methods that blend the hypothesis-driven rigor of mechanistic modeling, the adaptability of AI, and the geometric insights of topological analysis, all while grounded in the realities of data sparsity.
Interpretable machine learning methods
Interpretable machine learning methods
There is a tremendous effort to understand how AI models generate specific predictions. While they can yield highly accurate results, the internal decision-making process is often hard to explain, and the models difficult to interpret. The lack of transparency raises concerns about trust, fairness and accountability in critical domains such as medicine and finance, where decisions can directly impact lives. In this context, interpretable and transparent ML techniques, e.g., Bayesian Networks (BNs), offer a distinct advantage31–33.
BNs are probabilistic, acyclic graphical models that capture high-order, non-transitive relationships. Unlike neural networks, each relationship is explicitly represented in the topology of the graph. Combined with conditional probability tables that explicitly quantify how variables interact, this structure makes the reasoning process inherently traceable. BNs can represent and test causal relationships, enabling “what-if” scenario analyses that can be especially well-suited for the clinical and scientific settings. This interpretability extends to high dimensional omics data as well; rather than relying on abstract latent features, BNs reveal biologically meaningful and interpretable multiscale nodes and modules.
Broadly applied in oncology, interpretable network modeling extends to the analysis of cellular signaling cascades, therapy-induced state changes, and the emergence of adaptive resistance. Dynamic and temporally resolved BN modeling can reveal the sequence of events that precede cellular transformation, thereby clarifying underlying causal pathways and generating clinically actionable, interpretable hypotheses. Importantly, these data-driven approaches can operate in parallel, or be integrated with mathematical and mechanistic models, facilitating the prioritization of parameters and regimes for targeted investigation in oncological and other biological systems34,35.
There is a tremendous effort to understand how AI models generate specific predictions. While they can yield highly accurate results, the internal decision-making process is often hard to explain, and the models difficult to interpret. The lack of transparency raises concerns about trust, fairness and accountability in critical domains such as medicine and finance, where decisions can directly impact lives. In this context, interpretable and transparent ML techniques, e.g., Bayesian Networks (BNs), offer a distinct advantage31–33.
BNs are probabilistic, acyclic graphical models that capture high-order, non-transitive relationships. Unlike neural networks, each relationship is explicitly represented in the topology of the graph. Combined with conditional probability tables that explicitly quantify how variables interact, this structure makes the reasoning process inherently traceable. BNs can represent and test causal relationships, enabling “what-if” scenario analyses that can be especially well-suited for the clinical and scientific settings. This interpretability extends to high dimensional omics data as well; rather than relying on abstract latent features, BNs reveal biologically meaningful and interpretable multiscale nodes and modules.
Broadly applied in oncology, interpretable network modeling extends to the analysis of cellular signaling cascades, therapy-induced state changes, and the emergence of adaptive resistance. Dynamic and temporally resolved BN modeling can reveal the sequence of events that precede cellular transformation, thereby clarifying underlying causal pathways and generating clinically actionable, interpretable hypotheses. Importantly, these data-driven approaches can operate in parallel, or be integrated with mathematical and mechanistic models, facilitating the prioritization of parameters and regimes for targeted investigation in oncological and other biological systems34,35.
Temporal dynamics
Temporal dynamics
Temporal dynamics in biomedical systems and in the mathematical sciences often unfold in high-dimensional spaces where available data are sparse, irregularly sampled, and multimodal. This is particularly evident in cancer applications, where clinical time-series combine imaging, multi-omics, and circulating biomarkers, yet are typically limited to a few hundred or thousand patient trajectories. Privacy-driven restrictions, evolving clinical protocols, and changes in instrumentation ensure that sparsity and irregular sampling remain persistent challenges.
Designing effective modeling frameworks for oncology is uniquely challenging due to the intrinsic nature of the available data. Because tumors evolve according to Darwinian principles, patient responses are dynamic, resulting in clinical time-series that are typically sparse, irregularly sampled, and composed of high-dimensional, multi-modal features such as imaging, multi-omics, and circulating biomarkers36,37. While individual data points can be information-rich, this temporal inconsistency hinders the application of standard deep learning methods that rely on regular sampling. Furthermore, oncology remains a “small data” domain, a reality driven by several persistent factors including data privacy concerns, a lack of dedicated infrastructure for structuring clinical information, significant inter- and intra-tumor heterogeneity, and a constantly shifting landscape of treatments38,39. Consequently, even as global data collection efforts expand, the ability of purely data-driven models to accurately capture long-term tumor evolution and treatment effects remains fundamentally constrained by these practical realities.
When confronted with these challenges in the available clinical data, the differences between ML and mechanistic models emerge: ML models are less adept in capturing temporal dynamics. While ML excels at pattern recognition, it often assumes static or dynamic but stationary data - an unrealistic assumption in oncology40. Dynamic ML/DL approaches that do not necessarily assume stationarity (such as RNNs, Temporal CNNs, BERT-type Transformers) tend to be sensitive to distributional shifts and computationally demanding for real-world data. Such methods rely on representativeness of prior observations and are therefore inherently limited in extrapolating to new or changing situations often encountered in scientific and clinical research. Mechanistic models, by contrast, are inherently dynamic and hypothesis-driven, grounded in our growing understanding of tumor biology, pharmacokinetics (PK/PD), and systems physiology41,42. Digital twins, for instance, aim to integrate clinical data with mechanistic modeling to produce patient-specific, and data-driven predictions21,43,44. As such, mechanistic models are ideally suited towards data-sparse regimes and extrapolation, including the modelling of counterfactuals. Notably, many such models (e.g., Latent ODEs / Neural stochastic differential equations (SDEs)) are explicitly designed to handle non-stationarity; oftentimes, it is their raison d’être. Importantly, however, mechanistic descriptors inherently remain limited by the assumptions they are grounded in and often lack flexibility to address unstructured data of known mechanistic connection to the trajectory of interest45.
While some promising approaches have recently been presented in other application domains, for example physics-informed diffusion models46, which embed biological and physical constraints into the generative process, first examples of mechanistic learning solutions, either sequential or as intrinsic combinations have addressed the modelling of tumor response dynamics on imaging as a combination of generative stable diffusion models and mechanistic ODEs describing governing dynamics of tumor size, and the modelling of outcomes, such as progression-free or overall survival47. Recent efforts have also focused on combining mechanistic survival models, e.g., based on hazard functions informed by tumor growth laws or PK/PD dynamics, with deep learning models trained on heterogeneous clinical and imaging data to enable biologically plausible extrapolation of survival predictions under different combination treatment scenarios48.
A hybrid strategy could reduce dependence on large longitudinal clinical datasets that are costly or infeasible to collect by preserving mechanisms and hypotheses to be studied. Such mechanistic learning solutions maintain an ability to abstract and approximate data regimes of limited mechanistic understanding, harnessing unstructured, deep data.
The same—and additional—challenges apply to all trajectory-indexed data, not just time series: spatial, spatiotemporal, pseudotime, and other biological coordinates. In principle, many time-series-centric statistical/ML methods can be generalized to arbitrary trajectory-indexed settings. However, beyond the issues noted above, especially the computational burden of handling irregular, fine-grained non-stationarity, further obstacles include branching and uncertain trajectories; spatial structure with directionalities, sections, borders, and anisotropy; multiple coordinates/mixed manifolds; spatial periodicity; and assigning orientation without explicit (time) lags. As with temporally indexed data, the field typically attains either fine-grained non-stationarity or scalability—but not both49–53.
In our view, the primary thrust should be to develop methodologies that capture interpretable, fine-grained non-stationarity (temporal, spatial, or otherwise trajectory-indexed) without prohibitive computational cost. This can be achieved by implementing dedicated, trajectory-type-specific frameworks that incorporate appropriate biological knowledge and constraints, rather than attempting a general-purpose solution for all trajectory-indexed data out of the box.
Temporal dynamics in biomedical systems and in the mathematical sciences often unfold in high-dimensional spaces where available data are sparse, irregularly sampled, and multimodal. This is particularly evident in cancer applications, where clinical time-series combine imaging, multi-omics, and circulating biomarkers, yet are typically limited to a few hundred or thousand patient trajectories. Privacy-driven restrictions, evolving clinical protocols, and changes in instrumentation ensure that sparsity and irregular sampling remain persistent challenges.
Designing effective modeling frameworks for oncology is uniquely challenging due to the intrinsic nature of the available data. Because tumors evolve according to Darwinian principles, patient responses are dynamic, resulting in clinical time-series that are typically sparse, irregularly sampled, and composed of high-dimensional, multi-modal features such as imaging, multi-omics, and circulating biomarkers36,37. While individual data points can be information-rich, this temporal inconsistency hinders the application of standard deep learning methods that rely on regular sampling. Furthermore, oncology remains a “small data” domain, a reality driven by several persistent factors including data privacy concerns, a lack of dedicated infrastructure for structuring clinical information, significant inter- and intra-tumor heterogeneity, and a constantly shifting landscape of treatments38,39. Consequently, even as global data collection efforts expand, the ability of purely data-driven models to accurately capture long-term tumor evolution and treatment effects remains fundamentally constrained by these practical realities.
When confronted with these challenges in the available clinical data, the differences between ML and mechanistic models emerge: ML models are less adept in capturing temporal dynamics. While ML excels at pattern recognition, it often assumes static or dynamic but stationary data - an unrealistic assumption in oncology40. Dynamic ML/DL approaches that do not necessarily assume stationarity (such as RNNs, Temporal CNNs, BERT-type Transformers) tend to be sensitive to distributional shifts and computationally demanding for real-world data. Such methods rely on representativeness of prior observations and are therefore inherently limited in extrapolating to new or changing situations often encountered in scientific and clinical research. Mechanistic models, by contrast, are inherently dynamic and hypothesis-driven, grounded in our growing understanding of tumor biology, pharmacokinetics (PK/PD), and systems physiology41,42. Digital twins, for instance, aim to integrate clinical data with mechanistic modeling to produce patient-specific, and data-driven predictions21,43,44. As such, mechanistic models are ideally suited towards data-sparse regimes and extrapolation, including the modelling of counterfactuals. Notably, many such models (e.g., Latent ODEs / Neural stochastic differential equations (SDEs)) are explicitly designed to handle non-stationarity; oftentimes, it is their raison d’être. Importantly, however, mechanistic descriptors inherently remain limited by the assumptions they are grounded in and often lack flexibility to address unstructured data of known mechanistic connection to the trajectory of interest45.
While some promising approaches have recently been presented in other application domains, for example physics-informed diffusion models46, which embed biological and physical constraints into the generative process, first examples of mechanistic learning solutions, either sequential or as intrinsic combinations have addressed the modelling of tumor response dynamics on imaging as a combination of generative stable diffusion models and mechanistic ODEs describing governing dynamics of tumor size, and the modelling of outcomes, such as progression-free or overall survival47. Recent efforts have also focused on combining mechanistic survival models, e.g., based on hazard functions informed by tumor growth laws or PK/PD dynamics, with deep learning models trained on heterogeneous clinical and imaging data to enable biologically plausible extrapolation of survival predictions under different combination treatment scenarios48.
A hybrid strategy could reduce dependence on large longitudinal clinical datasets that are costly or infeasible to collect by preserving mechanisms and hypotheses to be studied. Such mechanistic learning solutions maintain an ability to abstract and approximate data regimes of limited mechanistic understanding, harnessing unstructured, deep data.
The same—and additional—challenges apply to all trajectory-indexed data, not just time series: spatial, spatiotemporal, pseudotime, and other biological coordinates. In principle, many time-series-centric statistical/ML methods can be generalized to arbitrary trajectory-indexed settings. However, beyond the issues noted above, especially the computational burden of handling irregular, fine-grained non-stationarity, further obstacles include branching and uncertain trajectories; spatial structure with directionalities, sections, borders, and anisotropy; multiple coordinates/mixed manifolds; spatial periodicity; and assigning orientation without explicit (time) lags. As with temporally indexed data, the field typically attains either fine-grained non-stationarity or scalability—but not both49–53.
In our view, the primary thrust should be to develop methodologies that capture interpretable, fine-grained non-stationarity (temporal, spatial, or otherwise trajectory-indexed) without prohibitive computational cost. This can be achieved by implementing dedicated, trajectory-type-specific frameworks that incorporate appropriate biological knowledge and constraints, rather than attempting a general-purpose solution for all trajectory-indexed data out of the box.
Ethics, uncertainty, and abstraction
Ethics, uncertainty, and abstraction
Critical ethical and philosophical questions were also discussed. As AI systems increasingly rely on large datasets, for example from national health registries, and begin to be applied to highly temporally resolved patient-specific wearable data, how do we ensure patient autonomy and data sovereignty54–56? What frameworks are needed to govern the use of sensitive health data in model training? The protection of sensitive data is an issue even without AI. A solution may be training an AI/LLM on one set of sensitive data, and then apply the trained model elsewhere, for example, via federated learning57. This could be a way to benefit from multiple sensitive data sets without directly sharing or combining them.
Further concerns arise as we move toward AI Scientists–AI systems that act not only as tools but as autonomous agents in scientific discovery58. Currently, there is a significant implementation gap: while AI systems can generate innovative ideas59, they still often lack the ability to follow through with verification, robust experimental design, validation, and interpretation58,60. If such systems begin to generate scientific results without clear human oversight, questions emerge regarding scientific accountability and epistemic transparency: who is responsible for the outputs of these AI Scientists?
Finally, as AI systems gain increasing autonomy, there is a growing danger that they could pursue research directions misaligned with ethical standards in healthcare, i.e., autonomy, beneficence, nonmaleficence, and justice61–63. Without explicit ethical constraints, autonomous AI agents might prioritize technical novelty or performance metrics over transparency, clinical relevance or patient well-being. Similarly, AI systems might identify therapeutic targets or genetic interventions that are technically feasible but raise concerns about germline editing or reproductive manipulation. In the absence of human judgment and contextual awareness, such systems risk amplifying biomedical research agendas that are not aligned with equity, safety, or patient autonomy. Addressing these risks will require the integration of ethics-by-design principles64, domain-specific governance frameworks65, and international regulation66.
The group also wrestled with the philosophical concept of probability in medicine. Prompted by the provocative article from Spiegelhalter in 2024, the group discussed whether probability exists for single events such as tumor progression67. For example, if tumor progression is the predicted outcome with 90% probability, is the prediction wrong if the tumor does not progress? Is the framing of the question as a “right” or “wrong” prediction appropriate in Mathematical Oncology? Is probability a real property of biological systems, or merely a tool for managing uncertainty? While simplification is necessary for modeling, excessive abstraction can obscure critical biological complexity and ignore the reality that probabilities are a necessary means of describing model predictions. The discussion was brought back to practical reality by determining that models must strike a balance between tractability and fidelity, and a recognition that models will never have perfect, complete information, and therefore will inherently have uncertainty, which is best represented as probability.
Critical ethical and philosophical questions were also discussed. As AI systems increasingly rely on large datasets, for example from national health registries, and begin to be applied to highly temporally resolved patient-specific wearable data, how do we ensure patient autonomy and data sovereignty54–56? What frameworks are needed to govern the use of sensitive health data in model training? The protection of sensitive data is an issue even without AI. A solution may be training an AI/LLM on one set of sensitive data, and then apply the trained model elsewhere, for example, via federated learning57. This could be a way to benefit from multiple sensitive data sets without directly sharing or combining them.
Further concerns arise as we move toward AI Scientists–AI systems that act not only as tools but as autonomous agents in scientific discovery58. Currently, there is a significant implementation gap: while AI systems can generate innovative ideas59, they still often lack the ability to follow through with verification, robust experimental design, validation, and interpretation58,60. If such systems begin to generate scientific results without clear human oversight, questions emerge regarding scientific accountability and epistemic transparency: who is responsible for the outputs of these AI Scientists?
Finally, as AI systems gain increasing autonomy, there is a growing danger that they could pursue research directions misaligned with ethical standards in healthcare, i.e., autonomy, beneficence, nonmaleficence, and justice61–63. Without explicit ethical constraints, autonomous AI agents might prioritize technical novelty or performance metrics over transparency, clinical relevance or patient well-being. Similarly, AI systems might identify therapeutic targets or genetic interventions that are technically feasible but raise concerns about germline editing or reproductive manipulation. In the absence of human judgment and contextual awareness, such systems risk amplifying biomedical research agendas that are not aligned with equity, safety, or patient autonomy. Addressing these risks will require the integration of ethics-by-design principles64, domain-specific governance frameworks65, and international regulation66.
The group also wrestled with the philosophical concept of probability in medicine. Prompted by the provocative article from Spiegelhalter in 2024, the group discussed whether probability exists for single events such as tumor progression67. For example, if tumor progression is the predicted outcome with 90% probability, is the prediction wrong if the tumor does not progress? Is the framing of the question as a “right” or “wrong” prediction appropriate in Mathematical Oncology? Is probability a real property of biological systems, or merely a tool for managing uncertainty? While simplification is necessary for modeling, excessive abstraction can obscure critical biological complexity and ignore the reality that probabilities are a necessary means of describing model predictions. The discussion was brought back to practical reality by determining that models must strike a balance between tractability and fidelity, and a recognition that models will never have perfect, complete information, and therefore will inherently have uncertainty, which is best represented as probability.
Education and training the next generation
Education and training the next generation
A recurring theme was the need to rethink the training of the next generation scientists and in particular Mathematical Oncologists. These include multi-scale thinking and the ability to navigate from molecular biology to systems-level modeling, and from theoretical abstraction to clinical application. Cancer research often requires scientists to move between these different scales and domains, ensuring that the knowledge in one domain can inform learning in others. Realizing these hybrid workflows in practice demands fluency in biology, mathematics, and machine learning. Training the next generation of mathematical oncologists to think seamlessly across these scales is as critical as the algorithms themselves. Another related, important skill for a successful integration of modeling in cancer research is the competence in interdisciplinary collaboration and effective communication with biologists and clinicians. Scientists must be able to communicate effectively across different fields, each of which has its own terminology, priorities, and methods.
Concerns regarding the misuse of AI tools by students and researchers who lack a foundational understanding were also raised. As AI becomes more accessible, the risk of over-reliance grows, along with the illusion of understanding; the tendency to mistake a fluent, plausible-sounding output for a correct or meaningful one. This can lead to uncritical acceptance of results, especially when users lack the domain knowledge or reasoning skills to evaluate them. Therefore, AI literacy and data science should be another important aspect of training. A reasonable foundation in machine learning, data analysis techniques, and the ethical use of AI tools, more than just familiarity with software, is needed to understand the underlying algorithms and models, as well as their strengths and limitations68. AI literacy is not just about using AI tools but critically assessing their outputs. Scientists must be able to evaluate the quality of AI predictions, understand the biases embedded in datasets and AI-generated models, and consider the broader ethical implications of AI decisions. Developing this skill requires courses in both theoretical and practical aspects of AI, paired with hands-on experience in real-world research contexts69.
Critical thinking and logical reasoning, in particular, must be deliberately cultivated70–72. Just as Mathematical Oncologists are trained to critically evaluate experimental and clinical data by questioning protocol details, choosing appropriate statistical tests, and recognizing selection bias or image manipulation, an analogous set of skills is required to determine when to trust AI-based assessments. Students must learn to examine whether an AI model was trained on appropriate and representative data, question model assumptions, interpret uncertainty, and evaluate the limitations of AI-generated results. This ability also extends to recognizing when AI-generated outputs may reflect biases in the data or in the models themselves. The goal is not just to accept results at face value but to constantly ask, “What is missing here?” and “What assumptions are overlooked?” Exercises that involve critiquing flawed outputs, whether from a mathematical model or an LLM, can help students build healthy skepticism and scientific rigor. This could involve reviewing AI predictions or evaluating the robustness of models. By engaging with examples of flawed science, whether from human error or model misapplication, students can develop the necessary skills for high-quality research while using AI.
Students must understand that, ultimately, the responsibility for research lies with the human scientist, not the AI tools they use. While AI can aid in processing data and generating insights, it is the researcher who must apply critical judgment, interpret results, and make informed decisions. This awareness of responsibility is key to developing both research maturity and a strong sense of the value they bring to the scientific process. It is essential for students to recognize that the central goal of science is not simply to find answers but to continuously ask the next question. This process of inquiry, which is the drive to explore what comes next, remains at the heart of scientific advancement and, for now, is a uniquely human endeavor. AI may assist in refining or accelerating this journey, but the spark of curiosity, critical thinking, and the formulation of new questions is something only human researchers can contribute73.
A recurring theme was the need to rethink the training of the next generation scientists and in particular Mathematical Oncologists. These include multi-scale thinking and the ability to navigate from molecular biology to systems-level modeling, and from theoretical abstraction to clinical application. Cancer research often requires scientists to move between these different scales and domains, ensuring that the knowledge in one domain can inform learning in others. Realizing these hybrid workflows in practice demands fluency in biology, mathematics, and machine learning. Training the next generation of mathematical oncologists to think seamlessly across these scales is as critical as the algorithms themselves. Another related, important skill for a successful integration of modeling in cancer research is the competence in interdisciplinary collaboration and effective communication with biologists and clinicians. Scientists must be able to communicate effectively across different fields, each of which has its own terminology, priorities, and methods.
Concerns regarding the misuse of AI tools by students and researchers who lack a foundational understanding were also raised. As AI becomes more accessible, the risk of over-reliance grows, along with the illusion of understanding; the tendency to mistake a fluent, plausible-sounding output for a correct or meaningful one. This can lead to uncritical acceptance of results, especially when users lack the domain knowledge or reasoning skills to evaluate them. Therefore, AI literacy and data science should be another important aspect of training. A reasonable foundation in machine learning, data analysis techniques, and the ethical use of AI tools, more than just familiarity with software, is needed to understand the underlying algorithms and models, as well as their strengths and limitations68. AI literacy is not just about using AI tools but critically assessing their outputs. Scientists must be able to evaluate the quality of AI predictions, understand the biases embedded in datasets and AI-generated models, and consider the broader ethical implications of AI decisions. Developing this skill requires courses in both theoretical and practical aspects of AI, paired with hands-on experience in real-world research contexts69.
Critical thinking and logical reasoning, in particular, must be deliberately cultivated70–72. Just as Mathematical Oncologists are trained to critically evaluate experimental and clinical data by questioning protocol details, choosing appropriate statistical tests, and recognizing selection bias or image manipulation, an analogous set of skills is required to determine when to trust AI-based assessments. Students must learn to examine whether an AI model was trained on appropriate and representative data, question model assumptions, interpret uncertainty, and evaluate the limitations of AI-generated results. This ability also extends to recognizing when AI-generated outputs may reflect biases in the data or in the models themselves. The goal is not just to accept results at face value but to constantly ask, “What is missing here?” and “What assumptions are overlooked?” Exercises that involve critiquing flawed outputs, whether from a mathematical model or an LLM, can help students build healthy skepticism and scientific rigor. This could involve reviewing AI predictions or evaluating the robustness of models. By engaging with examples of flawed science, whether from human error or model misapplication, students can develop the necessary skills for high-quality research while using AI.
Students must understand that, ultimately, the responsibility for research lies with the human scientist, not the AI tools they use. While AI can aid in processing data and generating insights, it is the researcher who must apply critical judgment, interpret results, and make informed decisions. This awareness of responsibility is key to developing both research maturity and a strong sense of the value they bring to the scientific process. It is essential for students to recognize that the central goal of science is not simply to find answers but to continuously ask the next question. This process of inquiry, which is the drive to explore what comes next, remains at the heart of scientific advancement and, for now, is a uniquely human endeavor. AI may assist in refining or accelerating this journey, but the spark of curiosity, critical thinking, and the formulation of new questions is something only human researchers can contribute73.
Conclusion
Conclusion
In conclusion, Mathematical Oncology stands at a critical inflection point, shaped by the accelerating integration of AI. The convergence of mechanistic modeling with AI-driven methods offers opportunities to overcome long-standing challenges of sparse, irregular, and high-dimensional clinical datasets and simultaneously interpretable, clinically relevant predictions. We advocate for revolutionizing the training of future scientists, emphasizing interdisciplinary skills, critical thinking, and logical reasoning, while upholding conceptual depth, ethics, and scientific rigor. AI’s capacity for accelerating scientific discovery, as evidenced in drug design, the elucidation of biological mechanisms, and the discovery of new physical laws, positions it as a transformative tool. Agentic AI, in particular, offers a paradigm shift by orchestrating multi-component workflows that mimic mathematical modeling processes, thereby enhancing speed, efficiency, reproducibility, and transparency in research.
Furthermore, hybrid mechanistic learning frameworks, which overlay data-driven adaptability onto equation-based structures, are crucial for addressing data sparsity and irregular sampling in clinical data, allowing for more robust predictions and insights. Interpretable machine learning methods, such as Bayesian Networks, provide transparent decision-making processes, which are essential for trust and accountability in medicine. By integrating these areas, we can effectively leverage AI to bridge the gap between abstract models and actionable patient-specific predictions, ultimately accelerating the translation of mechanistic insights into clinical strategies and advancing the field of oncology. The future of Mathematical Oncology lies not in replacing human creativity but in augmenting it, using AI to ask sharper questions, test bold hypotheses, and translate insights into scientific discoveries and ultimately improved patient care.
This perspective was produced with assistance from Microsoft Copilot and ChatGPT 5.0, based on a blog post including a conceptual illustration and summary of a meeting in Syracuse, Sicily in May 202574. The authors provided all prompts, content, and approval of the final version.
In conclusion, Mathematical Oncology stands at a critical inflection point, shaped by the accelerating integration of AI. The convergence of mechanistic modeling with AI-driven methods offers opportunities to overcome long-standing challenges of sparse, irregular, and high-dimensional clinical datasets and simultaneously interpretable, clinically relevant predictions. We advocate for revolutionizing the training of future scientists, emphasizing interdisciplinary skills, critical thinking, and logical reasoning, while upholding conceptual depth, ethics, and scientific rigor. AI’s capacity for accelerating scientific discovery, as evidenced in drug design, the elucidation of biological mechanisms, and the discovery of new physical laws, positions it as a transformative tool. Agentic AI, in particular, offers a paradigm shift by orchestrating multi-component workflows that mimic mathematical modeling processes, thereby enhancing speed, efficiency, reproducibility, and transparency in research.
Furthermore, hybrid mechanistic learning frameworks, which overlay data-driven adaptability onto equation-based structures, are crucial for addressing data sparsity and irregular sampling in clinical data, allowing for more robust predictions and insights. Interpretable machine learning methods, such as Bayesian Networks, provide transparent decision-making processes, which are essential for trust and accountability in medicine. By integrating these areas, we can effectively leverage AI to bridge the gap between abstract models and actionable patient-specific predictions, ultimately accelerating the translation of mechanistic insights into clinical strategies and advancing the field of oncology. The future of Mathematical Oncology lies not in replacing human creativity but in augmenting it, using AI to ask sharper questions, test bold hypotheses, and translate insights into scientific discoveries and ultimately improved patient care.
This perspective was produced with assistance from Microsoft Copilot and ChatGPT 5.0, based on a blog post including a conceptual illustration and summary of a meeting in Syracuse, Sicily in May 202574. The authors provided all prompts, content, and approval of the final version.
출처: PubMed Central (JATS). 라이선스는 원 publisher 정책을 따릅니다 — 인용 시 원문을 표기해 주세요.
🏷️ 같은 키워드 · 무료전문 — 이 논문 MeSH/keyword 기반
- A Phase I Study of Hydroxychloroquine and Suba-Itraconazole in Men with Biochemical Relapse of Prostate Cancer (HITMAN-PC): Dose Escalation Results.
- Self-management of male urinary symptoms: qualitative findings from a primary care trial.
- Clinical and Liquid Biomarkers of 20-Year Prostate Cancer Risk in Men Aged 45 to 70 Years.
- Diagnostic accuracy of Ga-PSMA PET/CT versus multiparametric MRI for preoperative pelvic invasion in the patients with prostate cancer.
- Comprehensive analysis of androgen receptor splice variant target gene expression in prostate cancer.
- Clinical Presentation and Outcomes of Patients Undergoing Surgery for Thyroid Cancer.