66
Views
0
CrossRef citations to date
0
Altmetric
Editorial

Artificial Intelligence in Cancer Clinical Research: I. Introduction

ORCID Icon &

“insights that are available to anyone who can think logically with understanding and imagination lie beyond anything that can be formalized in a set of rules. Rules can sometimes be a partial substitute for understanding but they can never replace it entirely.”

- Roger Penrose

Human intelligence

In this first of the series on Artificial Intelligence in Cancer Clinical Research, it is important to attempt to define the concept of intelligence and the important features we associate with human intelligence. Intelligence is a construct that we associate most notably with human beings and our understanding of human intelligence including perception, cognition, thought, conceptualization, pattern recognition, symbolic processing, creativity, and problem solving. While we acknowledge apparent elements of intelligence in other living species, we most commonly think of intelligence as a manifestation of our conscious awareness and ability to form concepts along with our capacity for symbolic processing, the development of language and ultimately our ability to reason, make decisions and use judgement as we navigate our world (Citation1–4). At the same time, while consciousness and self-awareness seem fundamental and unique to the human experience, these concepts remain profound and elusive mysteries. How can the complex array of atoms and molecules that ultimately behave the physical laws of motion, along with the intricate network of neurons in the central nervous system, give rise to our internal conscious experience and ability to think and reason along with experiences such as joy, sadness, love and beauty? While conceptual processes may be amenable to machine learning (ML), the human intellect is capable of generating new, abstract concepts that further organize his or her environment and, importantly, discover abstract ideas. Such processes provide new insights beyond the actual world and, in a sense, create his or her own world of meaningful symbolic systems with an active curiosity striving for meaning and true understanding of the world around us (Citation2). While a sign represents a characteristic of the external world amenable to algorithmic analysis and mimicking, a symbol evokes an abstract internal response separate from a particular concrete event (Citation2). Human symbolic systems are not only used for communication of cognitive information enabling language but the communication of emotional information. In fact, many of humanity’s most important and highest qualities relate to the development and creative use of symbolic systems resulting in the development of culture, myth, idealism and empathy. While a sign can be thought of as a context free characteristic of an object or situation, a symbol refers to experiences within ourselves. These experiences gain meaning within a context of emotions and interests resulting from the intersection of our prior experience and memory and the present situation to generate feelings or emotions within ourselves as we participate as active contributors to the creation of our world (Citation1).

Artificial intelligence

Advances in computing technology have emerged with astounding and accelerating speed. This has given rise to the question of our ability to replicate or even exceed the capacity of our living neural circuitry to think and achieve features we associate with intelligence and the concepts of ‘artificial intelligence’, ‘machine learning’ and ‘large language models’ capable of responses we would normally associate with thought, learning and perhaps even consciousness. Features of artificial intelligence are emerging and are now blossoming across virtually every aspect of human activity including medicine and medical research. Artificial intelligence (AI) represents a broad array of considerations of automated processing of tasks previously relegated to humans. Machine learning (ML) is that aspect of AI related to the learning of patterns in data usually in the form of supervised learning based on mathematical algorithms that can learn from previous results over time performing future tasks using mathematical optimization methods. These can range from more routine repetitive tasks to generative AI neural networks. ML approaches are generally differentiated as either ‘supervised’ or ‘unsupervised’ techniques. In supervised approaches, labeled data are used to train a model and often used for classification and in prognostic or prediction models. In unsupervised methods, the modeling attempts to discover patterns of structure within the data without any prior labelling. ML has been applied in oncology to image analysis, natural language processing (NLP), clinical decision support and large real world data analysis among others. While ML broadly relates to methods of automated learning with minimal human involvement, deep AI or deep learning utilizes advanced neural networks in an attempt to simulate the function and responses of the human brain. Deep learning (DL) based on deep neural networks often based on billions of parameters has gained considerable attention for diagnostic and predictive modeling of treatment effects and toxicities and gene sequencing among many others.

Large language models (LLM) represent deep machine learning models previously trained on extremely large datasets transformed by neural networks working in parallel. These attempt to extract meaningful relationships and establish a grammar from sequences of text through unsupervised learning representing words as multidimensional vectors with statistical relationships. In essence, such models attempt to ‘understand’ and ultimately generate output similar and largely indistinguishable from human language. Although the output clearly does not represent actual thought as we experience it, it may appear to do so to an outside observer.

The term strong AI is often used for the aspirational goal of computer responses based on algorithms that could not be distinguished from those of a human being. A debate continues as to whether such a computer program, if achieved, would actually be conscious and self-aware in the same sense that the human mind is. Those supporting the affirmative argue that our sense of consciousness and self-awareness arise naturally from the immense complexity of interconnected neurons and resulting impulses that exist within the human brain. Therefore, they argue, all human thought and awareness is ultimately computational and results from the countless computations occurring simultaneously in the brain. Based on this presumption, strong AI seems destined, at some point in the future to have the ability to replicate human thought and consciousness. At that point, the behavior of strong AI software would become indistinguishable from human cognition and responsiveness.

However, the likelihood of this scenario emerging is challenged by the Incompleteness Theorem of Kurt Godel who demonstrated that there are computations that no algorithm can yield and truths that no formal system can provide. Likewise, Turing has demonstrated in the ‘Halting Problem’ that no computer algorithm can reveal when it will go into an infinite loop or prevent an infinite loop. The mathematician Roger Penrose has concluded that insights available to anyone thinking logically lie beyond what can be formalized as a set of rules or calculations and, therefore, aspects of higher intelligence and insight represent a non-algorithmic function of the brain (Citation4). Aspects of higher intelligence associated with the human mind appear to be beyond anything reducible to a true formal or algorithmic system. This would lead one to conclude that human consciousness is beyond that which is simulated by a computer algorithm and, therefore, no algorithmic computer can ever completely demonstrate conscious thought or self-awareness.” (Citation5)

General Limitations of AI: Admittedly, the rate of progress and output associated with advances in AI technology have exceeded nearly all expectations. However, like all rapidly advancing technologies, important limitations have been identified that are not always fully appreciated and may impact on the validity and accuracy of the applications and their output. LLMs have been successful in accessing huge amounts of clinical data and often perform quite well in relatively structured scenarios based on increasingly sophisticated algorithms. Nevertheless, challenges remain concerning their accuracy with both false positive and false negative results providing convincing but factually erroneous information reinforcing the need for continued clinical oversight (Citation6,Citation7). Despite the remarkable rate of progress and often impressive results with current LLMs, careful evaluation of LLMs continues to reveal frequent factual errors and even fabricated information suggesting that much additional development and validation is needed before clinicians, patients, or policy makers can, if ever, rely entirely on the output of such models (Citation8). A joint Working Group of the German Society of Hematology and Oncology, the German Association for Medical Informatics, Biometry and Epidemiology and the German Informatics Society have put forward a roadmap for the development and implementation of AI in Hematology and Oncology (Citation9). They highlight several of the more apparent general limitations of modern AI including data security and privacy, The potential for access to personal data has necessitated enhanced criteria for ethical review, informed consent and prior approval for the use and application of patient-related data (Citation9). Also highlighted, is the impact of potential biases of data-driven supervised ML approaches limiting the robustness and generalizability of results built on patterns learned from training data sets and their specific inherent limitations. It has been demonstrated by several authors that the performance of AI-generated predictive models declines in other data sets that differ from those used for development or early validation studies (Citation10,Citation11). If the population represented in the training data are not representative of the broader population to which the model is to be applied or certain variables are defined or collected in a biased fashion due to technical or logical errors, the conclusions drawn for other populations may not be valid. As Ghassemi and colleagues have shown, AI algorithms applied to chest x-rays display underdiagnosis bias in underserved patient populations different than the populations used in the training of the algorithm thus exacerbating already existing care biases (Citation12).

Another important caveat for complex modeling based on AI may lie in an inability to adequately explain the results and confirm model validity, as the needed statistical modeling methodology, which ensures the scientific rigor of results, tend not to be considered by these models. Importantly, the suggestion that AI may largely replace clinicians in providing the needed emotional support and empathy expected along with the important clinical knowledge patients expect appears inherently flawed. In a recent study by Ayers and colleagues, they report that chatbot generated quality and empathetic responses to patient questions posed in an online forum actually surpassed those from physicians (Citation13). However, these conclusions have been challenged by other authors who point out that the assessment of the quality and empathy reflected in those encounters were provided by physicians and not by patients. In the clinician-patient interaction, it is essential that the clinician truly feels empathy for a patient as opposed to receiving an optimized empathetic-appearing response from an algorithm (Citation14). Cadiente and colleagues argue that the proper evaluation of the patient experience lies with the patient and not with the clinician. Clearly, patients should be involved in the evaluation process and judge whether their experience was understood, shared and satisfactorily acted upon.

A remaining challenge in assessing AI models lies with the complex ‘black box’ nature of AI data and algorithms. At the same time, there appears to be continued pervasiveness of digital and AI illiteracy among the population including clinicians and patients. Moving forward, it is essential that the integration of AI methods into clinical practice and research be primarily driven by close collaboration between knowledgeable data clinician scientists and modeling methodologists in order for clinical modeling efforts to identify and attempt to mitigate the major challenges involved. The benefits and limitations of utilizing AI-driven approaches must be evaluated in well designed, comparative, large-scale population-base clinical trials before such models can be safely used for guiding patient care or forging healthcare policies in the future (Citation9).

Conclusions

For clinicians, the most critical challenge will not be that of designing or analyzing clinical research studies based on ML but will rather be truly understanding, critically appraising and applying the results of such studies. Liu and colleagues from Google Health have presented a ‘user’s guide’ on how to read articles that use ML (Citation15). The authors lay out the new language of AI related to ML methods and the principles of evaluating and applying the results of modeling studies while contrasting traditional decision methods with those emerging from ML approaches. The authors provide a checklist for reading manuscripts based on ML studies to make clinically relevant predictions. Nevertheless, a fundamental challenge for all developers and users of AI including ML studies is that the results nearly always lie somewhere between always being correct or often being incorrect. In reality, such studies appear to provide increasingly accurate albeit imperfect and sometimes completely erroneous predictions. Experienced clinicians and scientists will continue to be called upon to critically appraise and hopefully refine such models in order to improve their own imperfect predictions based on training and experience. While being cautiously optimistic and hopeful, we must remain vigilant of the limitations of AI methods for clinical research. It is absolutely essential that AI methods, like conventional research methods, be carefully scrutinized and utilized appropriately so as to complement rather than replace the essential role of human intelligence with its important intuitive, ethical, and empathetic abilities. It is imperative that clinicians and investigators limit the potential inappropriate or mistaken use of these methods no matter how enticing (Citation3). In the end, the goal must be that of carefully, responsibly and appropriately utilizing the best tools available including clinical acumen and experience to make the very best decisions for our patients and society.

Declaration of interest

The authors report no conflict of interest. The authors alone are responsible for the content and writing of the article.

Additional information

Funding

The author(s) reported there is no funding associated with the work featured in this article.

References

  • Lyman GH. Perception, cognition and thought: Part II symbolic processing and language. Cancer Invest. 2023;41(7):617–620. doi:10.1080/07357907.2023.2233317.
  • Lyman GH. Perception, cognition and thought: Part I nature, evolution and development of conceptual and symbolic processes. Cancer Invest. 2023;41(6):535–538. doi:10.1080/07357907.2023.2229539.
  • Lyman GH, Kuderer NM. Perception, cognition and thought: Part III: reasoning, judgement and decision-making. Cancer Invest. 2023;41(8):699–703. doi:10.1080/07357907.2023.2238944.
  • Lyman GH, Lyman CH, Kuderer NM. Perception, cognition and thought: Part IV consciousness, awareness, and "I". Cancer Invest. 2023;41(9):751–756. doi:10.1080/07357907.2023.2248785.
  • Penrose R. Shadows of the Mind. Oxford (UK): Oxford Unitversity Press; 1994.
  • Menz BD, Modi ND, Sorich MJ, Hopkins AM. Health disinformation use case highlighting the urgent need for artificial intelligence vigilance: weapons of mass disinformation. JAMA Intern Med. 2024;184(1):92–96. doi:10.1001/jamainternmed.2023.5947.
  • De Angelis L, Baglivo F, Arzilli G, Privitera GP, Ferragina P, Tozzi AE, et al. ChatGPT and the rise of large language models: the new AI-driven infodemic threat in public health. Front Public Health. 2023;11:1166120. doi:10.3389/fpubh.2023.1166120.
  • Harrer S. Attention is not all you need: the complicated case of ethically using large language models in healthcare and medicine. EBioMedicine. 2023;93:104672. doi:10.1016/j.ebiom.2023.104672.
  • Rösler W, Altenbuchinger M, Baeßler B, Beissbarth T, Beutel G, Bock R, et al. An overview and a roadmap for artificial intelligence in hematology and oncology. J Cancer Res Clin Oncol. 2023;149(10):7997–8006. doi:10.1007/s00432-023-04667-5.
  • Gichoya JW, Thomas K, Celi LA, et al. AI pitfalls and what not to do: mitigating bias in AI. Br J Radiol. 2023;96:20230023.
  • Obermeyer Z, Powers B, Vogeli C, Mullainathan S. Dissecting racial bias in an algorithm used to manage the health of populations. Science. 2019;366(6464):447–453. doi:10.1126/science.aax2342.
  • Seyyed-Kalantari L, Zhang H, McDermott MBA, Chen IY, Ghassemi M. Underdiagnosis bias of artificial intelligence algorithms applied to chest radiographs in under-served patient populations. Nat Med. 2021;27(12):2176–2182. doi:10.1038/s41591-021-01595-0.
  • Ayers JW, Poliak A, Dredze M, Leas EC, Zhu Z, Kelley JB, et al. Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum. JAMA Intern Med. 2023;183(6):589–596. doi:10.1001/jamainternmed.2023.1838.
  • Cadiente A, Chen J, Pilkington B. Machine-made empathy? Why medicine still needs humans. JAMA Intern Med. 2023;183(11):1278–1279. doi:10.1001/jamainternmed.2023.4386.
  • Liu Y, Chen P-HC, Krause J, Peng L. How to read articles that use machine learning: users’ guides to the medical literature. JAMA. 2019;322(18):1806–1816. doi:10.1001/jama.2019.16489.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.