8,799
Views
9
CrossRef citations to date
0
Altmetric
Research Article

Chatbots for future docs: exploring medical students’ attitudes and knowledge towards artificial intelligence and medical chatbots

ORCID Icon, ORCID Icon, ORCID Icon, ORCID Icon, ORCID Icon & ORCID Icon
Article: 2182659 | Received 15 Dec 2022, Accepted 16 Feb 2023, Published online: 28 Feb 2023

ABSTRACT

Artificial intelligence (AI) in medicine and digital assistance systems such as chatbots will play an increasingly important role in future doctor – patient communication. To benefit from the potential of this technical innovation and ensure optimal patient care, future physicians should be equipped with the appropriate skills. Accordingly, a suitable place for the management and adaptation of digital assistance systems must be found in the medical education curriculum. To determine the existing levels of knowledge of medical students about AI chatbots in particular in the healthcare setting, this study surveyed medical students of the University of Luebeck and the University Hospital of Tuebingen. Using standardized quantitative questionnaires and qualitative analysis of group discussions, the attitudes of medical students toward AI and chatbots in medicine were investigated. From this, relevant requirements for the future integration of AI into the medical curriculum could be identified. The aim was to establish a basic understanding of the opportunities, limitations, and risks, as well as potential areas of application of the technology. The participants (N = 12) were able to develop an understanding of how AI and chatbots will affect their future daily work. Although basic attitudes toward the use of AI were positive, the students also expressed concerns. There were high levels of agreement regarding the use of AI in administrative settings (83.3%) and research with health-related data (91.7%). However, participants expressed concerns that data protection may be insufficiently guaranteed (33.3%) and that they might be increasingly monitored at work in the future (58.3%). The evaluations indicated that future physicians want to engage more intensively with AI in medicine. In view of future developments, AI and data competencies should be taught in a structured way during the medical curriculum and integrated into curricular teaching.

Introduction

The healthcare system is undergoing a digital transformation, and artificial intelligence (AI) will play a significant role in defining everyday medical practice [Citation1]. The location- and time-independence of digital applications have created new opportunities for medicine and health communication that are also changing the doctor – patient relationship [Citation2]. The growing importance of e-health applications, wearables and AI applications such as chatbots can empower patients to collect their own health data [Citation3,Citation4].

Furthermore, the digital networking of patients, hospitals, physicians and other healthcare services is enabling a shift from a physician-centric approach to more patient-centered treatment [Citation5]. To exploit the potential of this technical innovation and ensure optimized care for patients, future doctors must be equipped with the appropriate skills [Citation6]. Future physicians will not only need to be flexible in responding to different healthcare contexts but will also require the competence to adequately deal with procedures and applications involving AI and the accompanying big data [Citation7]. The growing complexity of medicine and increasing specialization of knowledge require the integration of AI as well as the interaction with digital assistance systems already in the curriculum of medical studies [Citation8–10]. According to current literature, although AI competencies are essential for medical practice, they are not comprehensively taught in medical education [Citation7,Citation11,Citation12].

Medical curriculum in Germany

A look at the national competence-based learning objectives catalog for medicine (NKLM) [Citation13] shows that the teaching of competencies in the area of medical apps and artificial intelligence is still underrepresented. The national competence-based learning objectives catalog for medicine is currently being further developed on the basis of the ‘Master Plan for Medical Studies 2020’ [Citation14]. It is to be made compulsory at all medical faculties when the new Medical Licensing Regulations come into force. The focus is on the question of which competencies junior doctors should acquire as part of the core curriculum of their medical studies, including medical communication skills, interprofessional teamwork, scientific work and digital competencies [Citation15]. The importance of integrating AI into medical studies is already being explored extensively in research and literature, but curricular implementation [Citation16,Citation17] is in its early stages. The emerge of medical chatbots could have the potential to improve the efficiency, quality and accessibility of healthcare services by providing quick information and connecting patients to healthcare providers and will become increasingly important for both doctors and patients during the treatment process. However, there are also raising concerns about chatbots potentially eroding diagnostic practice, being driven by business logic, and having limitations and incompleteness that may harm patients in emergency situations [Citation18,Citation19]. Understanding how these technologies work and how they can be used can help medical students better understand the broader landscape of healthcare and be better prepared for the future of medicine [Citation20]. As far as the authors know, the experience and attitude towards the use of chatbots in medicine among medical students has not yet been recorded in Germany.

Therefore, the question arose for us, how to successfully prepare medical professionals for this new scenario during their medical training. This study is intended as basic research relating to AI in the medical curriculum. We aimed to provide an overview of the current knowledge and attitudes of medical students regarding AI technologies to determine what is necessary for future implementation in the curriculum. Therefore, the following questions arose regarding the implementation of AI in medical education:

  • What attitudes do medical students have about artificial intelligence and the use of chatbots in medicine?

  • Does their view of chatbots change after learning more about medical chatbots in the course?

Material and methods

Study design and participants

This study followed the Standards for Reporting Qualitative Research [Citation21] and presents a mixed-method study as part of a project supported by the Federal Ministry of Education and Research, Germany (BMBF). It was carried out at the Medical Faculty of Tuebingen with the support of the Institute for Neuro- and Bioinformatics, University of Luebeck. A hybrid course named ‘Chatbots for Future Docs’ was developed for medical students of all semesters and was offered as an elective course between January and March 2022. N = 12 medical students learned about conditions of doctor – patient communication in general, possible uses of chatbots in healthcare, the ethical framework of AI, how chatbots work in general and what it takes to create a ready-to-go bot. The course combined classical teaching of theoretical knowledge (represented by asynchronous digital short inputs) and practical competence acquisition in the form of exercises and group work during a synchronous online format via Zoom.

The attitudes of participants regarding the topic of AI in medicine were quantitatively assessed by means of a standardized questionnaire before and after the course via the learning management software ILIAS. Qualitative methods such as group discussions supplemented the detailed, subjective and individual attitudes of the medical students to the topic.

Ethics

The study received ethical approval from the Ethics Committee of Tuebingen Medical Faculty (no. 824/2021BO2). Participation was voluntary, and students provided their informed consent and received no compensation. All responses and data were kept anonymous.

Measurement and process of study

We combined quantitative and qualitative research methods in this study to achieve an in-depth analysis of the collected data [Citation22].

To conduct the quantitative survey, the medical students were asked to complete a self-developed standardized questionnaire at the beginning of the course to assess subjective attitudes toward AI in medicine. The questionnaire was divided into three subsections, with items including five-point Likert scales and multiple-choice questions. The first subsection aimed to gather demographic data including age, gender and general attitudes toward technology and AI. Attitudes toward technology use were assessed using a shortened version of the validated short scale for assessing individual differences in the willingness of technology, as per Neyer et al. [Citation23] (technology commitment). The second subsection examined the openness of medical students regarding the future use of AI in healthcare in particular, which in the future may be able to answer health-related questions, perform certain tests or examinations, diagnose health conditions and suggest and apply treatments. The third subsection addressed the attitudes of the medical students toward chatbots in medicine. Students were asked to revisit two questions from the questionnaire on chatbots in medicine after the course, as this was the main interest of the course.

The material for qualitative content analysis was based on group work conducted by the students, which involved discussing questions about the use of mental health chatbots in medicine (see A1, appendix).

Data analysis

Quantitation

Statistical analysis of the questionnaire was performed using IBM SPSS version 27. Mean values (M), standard deviations (SD), sum values, frequencies and percentages of the relevant items were obtained. The Pearson correlation coefficient was calculated to capture linear relationships between the relevant variables. We used the Wilcoxon test to analyze the statistical significance of the differences between the two repeated items after the course; the level of significance was p < .05.As one student did not complete the course, we used the mean imputation procedure for the missing values after testing for normal distribution using the Kolmogorov-Smirnov test procedure [Citation24,Citation25].

Qualitative analysis

The results of the students’ group work were recorded, transcribed and coded by three different authors (JAM, TFW, AM). Following discussions in regular meetings, findings were summarized and a category system consisting of main and subcategories (according to Mayring’s qualitative content analysis) was agreed upon [Citation26]. Selected text passages were used as quotations to illustrate each category [Citation26]. Inductive category formation was performed to reduce the content of the material to its essentials (bottom-up process). In the first step, the goal of the analysis and the theoretical background were determined. From this, questions about chatbots in medicine were developed and presented to the students for discussion in a group exercise. Two main topics were identified, namely positive and negative attitudes toward chatbots in medicine. In the second step, we worked through the individual statements of the students systematically and derived various categories: user group, technical implementation, acceptance, and use in medicine.

Results

The course was attended by 12 medical students from the clinical and preclinical study sections, who were able (at least partly) to give a broad explanation of AI. For further information, please see .

Table 1. Students’ previous knowledge about AI and demographic data.

Main results questionnaire

Part 1: general attitudes toward technology, and socio demography

The participating students stated that they could explain the main features and applications of AI (50.0%) and that they could at least give a broad definition of AI (50.0%). Evaluation of the technology commitment according to Neyer et al. 2016 showed that the study participants displayed a moderate willingness to use technology (M = 21; SD = 3.1, min. 18, max. 31).

Part 2: attitudes toward AI in medicine

Many students were not afraid of the future changes that AI could bring. The majority (83.3%) were not afraid of losing their jobs because of AI or of being overwhelmed by using AI (83.3%). However, some students agreed that they were afraid of increasing surveillance at their future workplaces (58.3% agreement vs. 25.0% disagreement) and decreasing transparency regarding the use of personal data (33.3% agreement vs. 33.3% disagreement) ( Plot A).

Figure 1. Attitudes of medical students toward AI in medicine (fears about AI in various areas of medicine).

Figure 1. Attitudes of medical students toward AI in medicine (fears about AI in various areas of medicine).

As shown in Plot B, the majority of students supported the use of AI in various areas of medicine. Nevertheless, there was less support for its use in human-centered areas such as therapy or follow-up treatment, while the use of AI in more technical areas such as diagnostics or administrative tasks was seen as mostly beneficial.

Students saw the use of AI in medicine primarily as an opportunity to reduce the administrative burden on physicians. At the same time, however, they also believed that the new technology would be able to make diagnoses faster and more accurate in the future and ultimately also make access to medical advice more sustainable ( Plot A).

Figure 2. Attitudes of medical students toward AI in medicine (statements about the use of AI and chatbots).

Figure 2. Attitudes of medical students toward AI in medicine (statements about the use of AI and chatbots).

Part 3: attitudes toward chatbots in medicine

At the beginning of the course 692% of the students never have used a medical chatbot before. However, the students had great confidence in chatbots, especially for organizational tasks. In task areas with high levels of responsibility, such as diagnosis or treatment suggestions, opinions tended to diverge and were more critical, but the students were not completely opposed to the use of AI in general ( Plot B).

As shown in Tables A2 and A3 in the appendix, no major differences in attitudes held before and after the course were detected. However, students were significantly (p = 0.02) less critical about privacy and the use of medical chatbots after the course. Prior to the course, 500% of students believed that data privacy and security could not be fully guaranteed, 41.7% were undecided, and only 8.3% did not believe this. After the course, only 36.4% were still critical of data protection, 18.2% were undecided and 45.5% did not believe that data protection cannot be guaranteed with med. Chatbots (see Table A3). Only a few statements showed a clear positive or negative tendency. For example, the majority agreed that using chatbots saved time (83.3% agreement pre; 72.7% agreement post) and money (75.0% agreement pre; 54.5% post). However, medical students felt that chatbots were not yet sufficiently established and that long-term success had yet to materialize (81.8%). There was also a fear of communication problems (81.8%) and loss of personal contact with patients (63.6%) because of the lack of maturity of the technology.

Findings of the qualitative content analysis

The main findings of the present study concern the students’ views on chatbots in medicine. Four main themes/categories were identified: user group, technical implementation, acceptance, and use in medicine ().

Table 2. Overview of categories, subcategories and corresponding examples based on qualitative analysis.

Theme 1: user group

The medical students were positive that chatbots are accessible to a broad user group as a result of their time- and location-independent availability. They also believed that possible language barriers or other hurdles could also be overcome. Nevertheless, they were concerned that certain groups of people (elderly people, visually impaired people) may not be familiar with modern technology or chatbots. The medical students also proposed a minimum age of 18 years and an alternative language mode specifically aimed at children.

Theme 2: technical implementation

The students felt that with the use of emojis, artificial delays, small talk and customization, there was an opportunity to provide the most human-like and realistic conversation possible with a chatbot. Furthermore, the development of the chatbot should take into account the context in which it would be used (administrative or personal assistant). However, according to the students’ evaluations, technical implementation was not sufficiently mature or flexible because of the limited and rigid algorithms used.

Theme 3: acceptance and trust

The use of chatbots as therapy tools or interactive diaries for patients was considered by the students to be a good opportunity to reduce anonymity, shyness and shame about disclosing personal and painful information. However, it was thought that doctor – patient communication could deteriorate as a result of non-verbal communication with a chatbot and interpersonal information may get lost. Thus, the first personal conversation between doctor and patient was considered indispensable.

Use in medicine

The students perceived the use of therapy chatbots in medicine as potentially supportive and complementary diagnostic tools that could be used independently by patients but did not understand them as a substitute for therapy. As an example, the students mentioned the possibility of technical errors or the risk of increasing social withdrawal in depressed persons due to the lack of personal contact. Therefore, human contact and empathy were considered essential for the success of therapy, which could be reduced by the use of chatbots.

Teaching evaluation

The participants enjoyed gaining a fundamental insight into the topic of chatbots and AI, which was completely new for some of them. In addition, the interactive design and hybrid format contributed to the positive evaluation of the course. In general, the students gained a new perspective on chatbots and their associated problems, as well as a basic understanding of how chatbots function, where they can be used and the effort involved in implementing them.

Discussion

In the context of digitalization in healthcare, applications that use AI are becoming increasingly important [Citation12]. This study contributed to the understanding of the current level of knowledge and attitudes regarding AI, particularly chatbots, in medicine.

Medical chatbots are an example of how artificial intelligence and technology are being integrated into healthcare. As such, it would be beneficial for medical students in Germany to learn about this technology as part of their medical curriculum. This will help ensure that they are well-equipped to work with and utilize medical chatbots in their future practice, in order to provide high-quality and efficient care to patients. Additionally, learning about medical chatbots would provide medical students with a better understanding of the role that technology plays in healthcare and how it is likely to continue to shape the medical profession in the future.

This study revealed that the attitudes and expectations of medical students were generally optimistic about the use of AI and chatbots in relation to a variety of purposes in medicine, including in areas such as administration, research and diagnostic imaging techniques (, Plot B). In particular, the students would trust the chatbots to perform recurring and supportive activities, such as answering simple questions, arranging appointments, and providing basic information. The majority were certain that they would not be replaced by AI in their jobs or that they would be less valued in their roles as future doctors, as other authors have stated in similar studies [Citation27]. However, they were more critical of the use of these new technologies in core tasks, such as carrying out personal counseling and specific medical examinations. The participants also expressed concerns that data protection and privacy may no longer be adequately guaranteed. Finally, there was a fear that personal contact with patients could be lost if the patients were increasingly engaging with technical systems rather than human personnel.

As other studies have demonstrated, it cannot be assumed that the generation of people who have grown up with digital technologies and are proficient in their use (similar to our cohort) are also aware of all the options and ethical consequences of the use of new technologies in their professional field. However, this is not synonymous with the simultaneous development of digital competencies in the professional field [Citation28,Citation29]. The areas in which AI can be applied in medicine are diverse and, with the development of smartphone apps, have reached not only the healthcare system but also the private sphere – and there will likely be more in the future [Citation30]. Accordingly, to remain empowered, future physicians must be able to understand how AI works, as well as how to interpret results, in order to meaningfully support patients with digital tools at the same time as critically monitoring AI [Citation31]. Digital learning opportunities and the development of AI skills are essential in medical education and may help to meet the vast need for qualifications [Citation32]. We found a great deal of uncertainty and skepticism regarding chatbots due to the lack of integration of AI topics into the medical curriculum (as yet), as well as a lack of knowledge about the basic conditions and legal and ethical requirements of AI use [Citation33], reflecting findings from other studies [Citation34–36].

While students saw a great deal of potential for the use of chatbots in medicine, they had many concerns about using them in areas that went beyond organizational activities such as making patient appointments. Above all, they believed that the technology was not yet sufficiently developed and that in the context of healthcare, patients needed a human counterpart. Neutral attitudes to chatbots were also evident from many statements in the questionnaire, which confirmed the thesis that there is not yet enough knowledge about the topic of AI in medicine for the study participants to have developed distinct opinions. In the questions about chatbots, in particular, which were repeated after the course, we could not identify any significant changes in attitudes. However, our study was able to give medical students, as non-computer scientists, a good initial overview of how a chatbot works, a basic understanding of how much data needs to be provided for an AI, and also potential future uses of AI in medicine and medical chatbots.

The perceived challenges and concerns of students relating to the use of AI and chatbots in healthcare must be addressed and taken seriously before future physicians are exposed to such tools [Citation37]. AI is still underrepresented in the medical curriculum, and students lack the opportunity to engage more intensively with the topic of AI and develop the required expertise [Citation11,Citation38,Citation39]. Therefore, for us, it was important not only to teach digital competencies and knowledge about AI theoretically but also to incorporate them practically into the teaching unit. Thus, digitization was both included in the teaching and incorporated as a learning objective.

Limitations

Although the students were very interested in the topic of chatbots in medicine, and the topic of AI is also gaining increasing importance in medicine in general, the results of our study are limited in terms of representing the perspectives of the student population due to the small number of participants. Also, the course duration of three months was too short for sustainable changes in viewpoints or information gain to occur. However, the goal of this work was not to draw representative conclusions for all German medical students, but rather to understand the state of knowledge and perceptions of medical students regarding AI and chatbots in medicine, which we believe was achieved with our sample. The next step will be to investigate which AI competencies should be included in the curriculum.

Conclusion

This study indicated that future physicians in Germany are willing to engage more intensively with AI in medicine. In our study, they were able to develop a basic understanding of how AI and chatbots will affect their future daily work. Although their basic attitude toward the use of clinical AI was positive, medical students also had concerns, especially with regard to the lack of data protection and declining personal contact with patients. With a view to future developments in the workplace, we can only emphasize once again how urgently medical curricula need to be supplemented with these new core competencies so that doctors can help to shape the technological course of patient treatment in an informed and self-confident manner.

Ethical approval

The study received ethical approval from the Ethics Committee of Tuebingen Medical Faculty (no. 824/2021BO2). Participation was voluntary, and students provided their written informed consent. All responses and data were kept anonymous.

Contributorship

JAM was responsible for the study design and implementation and data collection. AMM was responsible for the technical information of the chatbot, AHW for the didactical design of the teaching course. JAM analysed the research material and conducted the writing for the manuscript. TFW participated in data analysis and interpretation and critically revised the manuscript. AHW and AMM made significant contributions to the study design and critically revised the manuscript. KN and WF revised the manuscript and have given final approval of the version to be published. All authors approved the final version of the manuscript and agreed to be responsible for all aspects found therein.

Acknowledgments

We would like to thank our study assistants Marie Thiwissen and Felix Pfeiffer for their help.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

We acknowledge support via financing publication fees from ‘Deutsche Forschungsgemeinschaft’ and the Open Access Publishing Fund of the University of Tuebingen. We also thank the Federal Ministry of Education and Research, Germany (BMBF) for supporting this project (16DHBQP041/16DHBQP042).

References

  • Liebrich F. Digitale Medienprodukte in Der Arzt-Patienten-Kommunikation. Wiesbaden: Springer Fachmedien; 2017.
  • Kundu S. How will artificial intelligence change medical training? Communications Medicine. 2021;1(1).
  • Lin B, Wu S. Digital transformation in personalized medicine with artificial intelligence and the internet of medical things. OMICS. 2022 Feb;26(2):77–11. PubMed PMID: 33887155.
  • Bates M. Health care chatbots are here to help. IEEE Pulse. 2019 May-Jun;10(3):12–14. PubMed PMID: 31135345.
  • Rajpurkar P, Chen E, Banerjee O, et al. AI in health and medicine. Nat Med. 2022 Jan;28(1):31–38. PubMed PMID: 35058619.
  • Ng KH, Wong JHD. A clarion call to introduce artificial intelligence (AI) in postgraduate medical physics curriculum. Phys Eng Sci Med. 2022 Mar;45(1):1–2. PubMed PMID: 35006576.
  • Han ER, Yeo S, Kim MJ, et al. Medical education trends for future physicians in the era of advanced technology and artificial intelligence: an integrative review. BMC Med Educ. 2019 Dec 11;19(1):460. PubMed PMID: 31829208; PubMed Central PMCID: PMCPMC6907217.
  • Moldt J-A, Festl-Wietek T, Mamlouk AM, et al. Assessing medical students’ perceived stress levels by comparing a chatbot-based approach to the Perceived Stress Questionnaire (PSQ20) in a mixed-methods study. Digit Health. 2022;8:20552076221139092.
  • Wiljer D, Hakim Z. Developing an artificial intelligence-enabled health care practice: rewiring health care professions for better care. J Med Imaging Radiat Sci. 2019 Dec;50(4 Suppl 2):S8–14. PubMed PMID: 31791914.
  • Wartman SA, Combs CD. Reimagining medical education in the age of AI. AMA J Ethics. 2019 Feb 1;21(2):E146–152. PubMed PMID: 30794124.
  • Chan KS, Zary N. Applications and challenges of implementing artificial intelligence in medical education: integrative review. JMIR Med Educ. 2019 Jun 15;5(1):e13930. PubMed PMID: 31199295; PubMed Central PMCID: PMCPMC6598417.
  • Wood EA, Ange BL, Miller DD. Are we ready to integrate artificial intelligence literacy into medical school curriculum: students and faculty survey. J Med Educ Curric Dev. 2021 Jan-Dec;8:23821205211024078. PubMed PMID: 34250242; PubMed Central PMCID: PMCPMC8239949.
  • Fischer MR, Bauer D, Mohn K, et al. Finally finished! national competence based catalogues of learning objectives for undergraduate medical education (NKLM) and dental education (NKLZ) ready for trial. GMS Z Med Ausbild. 2015;32(3):Doc35. PubMed PMID: 26677513; PubMed Central PMCID: PMCPMC4580444.
  • Erschens R, Herrmann-Werner A, Schaffland TF, et al. Association of professional pre-qualifications, study success in medical school and the eligibility for becoming a physician: a scoping review. PLoS ONE. 2021;16(11):e0258941. ( PubMed PMID: 34762678; PubMed Central PMCID: PMCPMC8584759. eng).
  • Plange N, Feltgen N, Arbeitsgemeinschaft L. The “Nationaler Kompetenzbasierter Lernzielkatalog Medizin NKLM 2.0”-Implications for medical education in ophthalmology. Ophthalmologie. 2022 Nov 21. PubMed PMID: 36409352.
  • Kuhn S. Transformation durch Bildung. Deutsches Ärzteblatt. 2018;2018:115.
  • Nagy M, Radakovich N, Nazha A. Why machine learning should be taught in medical schools.Med Sci Educator. 2022 Apr 1];32(2):529–532.
  • Nadarzynski T, Miles O, Cowie A, et al. Acceptability of artificial intelligence (AI)-led chatbot services in healthcare: a mixed-methods study. Digit Health. 2019 Jan-Dec;5:2055207619871808. PubMed PMID: 31467682; PubMed Central PMCID: PMCPMC6704417.
  • Parviainen J, Rantala J. Chatbot breakthrough in the 2020s? an ethical reflection on the trend of automated consultations in health care.Medicine, Health Care and Philosophy. 2022 Mar 1;25(1):61–71.
  • Pryss R, Kraft R, Baumeister H, et al. Using chatbots to support medical and psychological treatment procedures challenges, opportunities, technologies, reference architecture. In: Montag C Baumeister Heditors. Digital Phenotyping and Mobile Sensing. Studies in Neuroscience, Psychology and Behavioral Economics. Cham: Springer International Publishing; 2023. p. 335–346.
  • O’brien BC, Harris IB, Beckman TJ, et al. Standards for reporting qualitative research: a synthesis of recommendations. Acad Med. 2014 Sep;89(9):1245–1251. PubMed PMID: 24979285; eng.
  • Kaplan B, Duchon D. combining qualitative and quantitative methods in information systems research: a case study. MIS Quarterly. 1988;12(4):571–586.
  • Neyer F, Felber J, Gebhardt C. Kurzskala zur Erfassung von Technikbereitschaft [Short scale for recording technology willingness]. In: GESIS - Leibniz Institute for Social Sciences. Zusammenstellung sozialwissenschaftlicher Items und Skalen (ZIS); 2016.
  • Baur N, Blasius J. Handbuch Methoden der empirischen Sozialforschung. 1 ed. Wiesbaden: Springer VS; 2014.
  • Göthlich S. E. Zum Umgang mit fehlenden Daten in großzahligen empirischen Erhebungen. Albers, S., Klapper, D., Konradt, U., Walter, A., Wolf, J., eds. In: Methodik der empirischen Forschung. Wiesbaden: Gabler Verlag; 2007. p. 119–134.
  • Mayring P. Qualitative Inhaltsanalyse. Grundlagen und Techniken. Weinheim: Deutscher Studien Verlag; 2000.
  • Dos Santos D P, Giese D, Brodehl S, et al. Medical students’ attitude towards artificial intelligence: a multicentre survey. Eur Radiol. 2019 Apr;29(4):1640–1646. PubMed PMID: 29980928.
  • Kirschner PA, De Bruyckere P. The myths of the digital native and the multitasker. In Teaching and Teacher EducationVol. 67 p. 135–142; 2017 Oct. PubMed PMID: WOS:000411543800013; English.
  • Kuhn S, Kadioglu D, Deutsch K, et al. Data Literacy in der Medizin. Der Onkologe. 2018;24(5):368–377.
  • Ryan P, Luz S, Albert P, et al. Using artificial intelligence to assess clinicians’ communication skills. BMJ. 2019 Jan 18;364:l161. PubMed PMID: 30659013.
  • Wallis C. How artificial intelligence will change medicine. Nature. 2019 Dec;576(7787):S48. PubMed PMID: 31853072; eng.
  • Mosch L, Back D-A, Balzer F, et al. Lernangebote zu Künstlicher Intelligenz in der Medizin. 2021. https://zenodo.org/record/5497668
  • Blum A, Bosch S, Haenssle H, et al. Künstliche intelligenz und smartphone-programm-applikationen (Apps). Der Hautarzt. 2020;71(9):691–698.
  • Sit C, Srinivasan R, Amlani A, et al. Attitudes and perceptions of UK medical students towards artificial intelligence and radiology: a multicentre survey. Insights Imaging. 2020 Feb 5;11(1):14. PubMed PMID: 32025951; PubMed Central PMCID: PMCPMC7002761. eng.
  • Chen M, Zhang B, Cai Z, et al. Acceptance of clinical artificial intelligence among physicians and medical students: a systematic review with cross-sectional survey. Front Med. 2022;9:990604. PubMed PMID: 36117979; PubMed Central PMCID: PMCPMC9472134. eng.
  • Valikodath NG, Cole E, Ting DSW, et al. Impact of artificial intelligence on medical education in ophthalmology. Transl Vis Sci Technol. 2021 Jun 1;10(7):14. PubMed PMID: 34125146; PubMed Central PMCID: PMCPMC8212436.
  • Palanica A, Flaschner P, Thommandram A, et al. physicians’ perceptions of chatbots in health care: cross-sectional web-based survey. J Med Internet Res. 2019 Apr 5;21(4):e12887. PubMed PMID: 30950796; PubMed Central PMCID: PMCPMC6473203. eng.
  • Bisdas S, Topriceanu CC, Zakrzewska Z, et al. Artificial Intelligence in medicine: a multinational multi-center survey on the medical and dental students’ perception. Front Public Health. 2021;9:795284. PubMed PMID: 35004598; PubMed Central PMCID: PMCPMC8739771.
  • Buhmann J, Felix J, Gächter T, et al. Digitalisierung der Medizin: Konsequenzen für die Ausbildung.Bulletin des Médecins Suisses. 2018 Oct 17.

Appendices

A1. Questions from the group work about mental health chatbots in medicine

  • How would you rate chatbots as therapy tools, for example, Woebot? What opportunities do they offer and what are their limitations?

  • How could they influence the doctor – patient relationship? In your opinion, what would chatbots need to do in order for both patients and healthcare professionals to benefit from them in a meaningful way?

  • Which aspects of the use of chatbots in medicine do you see as particularly critical and why?

  • How do you think a chatbot should be used in terms of dialog design? How relevant are factors such as emoji use, artificial delays, language and small talk?

  • What aspects should be considered in terms of linguistic style? Should the dialog be designed differently regardless of the target group?

  • How would you describe the optimal personality of the chatbot? Do you think a chatbot should have personality and/or show emotion? How far can a chatbot be humanized in your view?

A2. I have a positive attitude towards the application and development of chatbots in medicine, because…

A3. I am critical of the use of chatbots in medicine, because…