Skip to main content

Danish cardiologists' attitude towards clinical AI support: a survey study

Søren A. Rand1, Anna Kirstine Ringgaard1, Anders Christian Riis-Jensen2, Henning Bundgaard1, 3, Iben Mundbjerg Gjødsbøl4 & Johan Skov Bundgaard1, 5

9. sep. 2025
8 min.

Abstract

Over 50 years ago, an article in the New England Journal of Medicine predicted that computers would eventually “…augment and, in some cases, largely replace the intellectual functions of the clinician” [1]. Today, that has become the reality; AI already assists clinical practice (i.e. electrocardiography interpretation, magnetic resonance imaging optimisation) [2-4] and demonstrates immense potential in clinical applications - from surveillance of infectious diseases to electroencephalography pattern recognition and cancer treatment assistance [5-7].

In Region Zealand and the Capital Region of Denmark, the PM-Heart project – a network-based machine-learning algorithm predicting mortality in patients with ischaemic heart disease (IHD) – is being evaluated in a randomised controlled trial [8]. While AI advances rapidly, key questions remain about its role in healthcare: Are clinicians ready to implement and use AI? A central concern is whether clinicians and patients are willing to trust AI-assisted decision-making. Studies show that clinicians are highly interested in AI but often doubt its reliability [9]. Initiatives that neglect both technical progress and human apprehensions may struggle to gain traction [10], and, notably, most clinicians lack formal training in AI [11]. Public perception of AI in medicine is mixed – about half of patients view AI positively, even so “AI-driven” advice is often seen as less trustworthy [12-14].

This study explored Danish cardiologists’ perceptions of AI-based clinical decision support and provided insights into their readiness to adopt and embrace AI in clinical practice.

Methods

Throughout 2024, we conducted anonymous, non-digital, cross-sectional surveys distributed by convenience sampling at cardiology departments across hospitals on Zealand. The surveys targeted cardiologists and clinicians who participated as treating physicians in the PM-Heart trial to evaluate their attitudes towards AI support in clinical decision-making. Clinicians rated their agreement with ten statements using a five-point Likert scale (1 = strongly disagree, 3 = neutral, 5 = strongly agree):

1) Outcomes: AI will improve outcomes for patients with IHD

2) Accuracy: I trust AI to deliver accurate predictions

3) Support: AI will improve my clinical decision-making

4) Interest: I am interested in using AI in the treatment of my patients

5) Time: AI will save me time in decision-making processes

6) Safety: AI will improve the safety of handling patients

7) Optimism: I am optimistic about AI’s role in the future of cardiology

8) Training: I am willing to invest time in learning and adopting AI technologies in the clinic

9) Workflow: AI tools will fit into my current workflow

10) Adoption: I believe my colleagues will adopt AI earlier than I will.

Statistical analysis

Descriptive statistics summarised survey responses. Mean scores and agreement levels (4-5 = high) across ten statements were compared between junior (with < 10 years’ experience) and senior (≥ 10 years’ experience) cardiologists using the Mann-Whitney U test. No imputation was performed. We also assessed intra-individual differences, response variability and visualised findings in a heatmap. Analyses were conducted in RStudio (v4.3.3).

Trial registration: not relevant.

Results

Sixty clinicians at eight Danish cardiology departments completed the survey (50% with ≥ 10 years’ experience). Complete responses were obtained for 59/60, with only a single response missing. The highest agreement rates were seen for future optimism (68%, mean: 3.7), willingness to invest time in training (65%, mean: 3.8) and interest in using AI for their own patients (63%, mean: 3.6; Figure 1). The highest disagreement rates concerned confidence in accuracy (33%, mean: 3.0), improvement of patient safety (27%, mean: 3.1) and time-saving in decision-making (25%, mean: 3.3; Figure 1). Interestingly, the mean score for whether respondents believed peers would adopt AI earlier than themselves was below neutral (mean: 2.8), suggesting above-average self-expectations.

No statistically significant differences were found between junior and senior clinicians. However, the largest numerical gaps were observed for willingness to invest time in training (mean 4.0 versus 3.5), belief that AI will save time in decision-making (mean 3.5 versus 3.1) and interest in using AI for own patients (mean 3.8 versus 3.4).

Intra-individual differences

A heatmap of intra-individual deltas (Figure 2) illustrates consistency in agreement for each respondent. Overall, responses were relatively uniform, with mean deltas ranging from 0.3 to 1.7. High similarity was seen between willingness to invest time in training and both interest in using AI for their own patients and future optimism about AI. In contrast, responses varied more for the statement on whether peers would adopt AI earlier, showing greater divergence from other statements.

Discussion

This study explored the attitudes of Danish cardiologists in Eastern Denmark towards AI in clinical decision-making, using IHD as a case example. While optimism about AI was generally moderate to high, key barriers emerged – namely, concerns about accuracy, patient safety and workflow integration –mirroring previous findings [12-16]. Notably, responses on AI integration aligned with clinicians’ interest in AI (Figure 2), and both have been cited as major implementation barriers to digital technology adoption [17]. These findings align with earlier studies that emphasise the need to address both technical and human factors for successful AI adoption [10, 18].

Junior cardiologists (< 10 years’ of experience) showed greater willingness to invest in training and a stronger interest in using AI for patient care, though differences were not statistically significant. Similarly, a recent study found that 83% of medical students viewed AI knowledge as important for their future roles as physicians, despite limited prior experience [19].

Similarly, concerns about AI reliability – also reported in studies showing that advice labelled as ‘AI-generated’ is perceived as less trustworthy – are reflected in our findings, where only a third of cardiologists expressed high confidence in an AI clinical-decision-making tool [12-14, 20]. The strong link between training and optimism supports the idea that clinicians who are optimistic about the future use of AI are also more willing to invest time in AI training/education.

The survey was conducted at PM-Heart inclusion sites, which may have influenced responses. The questionnaire was informed by literature and clinical experts to ensure relevance, though psychometric methods – commonly used to assess survey validity and reliability – were not applied. Due to the dynamic clinical setting, a formal response rate was not calculated, though most present cardiologists participated.

Our survey study reinforces the duality of addressing both technical capabilities and human factors to ensure the successful adoption of AI in clinical practice. In more practical terms, education and training in AI are needed for the successful implementation of AI in healthcare [19].

Conclusions

Danish cardiologists view AI as a promising tool for clinical decision-making and enhancing patient outcomes. However, confidence in the accuracy, time-saving potential and patient safety benefits of AI must be addressed for successful adoption in clinical practice. Overall, there was a high degree of willingness to invest time in training/education on AI.

Correspondence Søren A. Rand. E-mail: soeren.albertsen.rand.02@regionh.dk

Accepted 11 June 2025

Published 9 September 2025

Conflicts of interest HB reports financial support from or interest in Amgen, Sanofi-Evensis, MSD and BMS. All authors have submitted the ICMJE Form for Disclosure of Potential Conflicts of Interest. These are available together with the article at ugeskriftet.dk/dmj

References can be found with the article at ugeskriftet.dk/dmj

Cite this as Dan Med J 2025;72(10):A01250019

doi 10.61409/A01250019

Open Access under Creative Commons License CC BY-NC-ND 4.0

Referencer

  1. Haug CJ, Drazen JM. Artificial intelligence and machine learning in clinical medicine. N Engl J Med. 2023;388(13):1201‑1208. https://doi.org/10.1056/NEJMra2302038
  2. Siontis KC, Noseworthy PA, Attia ZI, Friedman PA. Artificial intelligence enhanced electrocardiography in cardiovascular disease management. Nat Rev Cardiol. 2021;18(7):465‑478. https://doi.org/10.1038/s41569-020-00503-2
  3. Shimron E, Perlman O. AI in MRI: computational frameworks for a faster, optimized, and automated imaging workflow. Bioengineering (Basel). 2023;10(4):492. https://doi.org/10.3390/bioengineering10040492
  4. Holzinger A. The next frontier: AI we can really trust. In: Kamp M, et al. Machine learning and principles and practice of knowledge discovery in databases. ECML PKDD 2021. Communications in Computer and Information Science, vol. 1524. Springer, 2021. https://doi.org/10.1007/978-3-030-93736-2_33
  5. Brownstein JS, Rader B, Astley CM, Tian H. Advances in artificial intelligence for infectious‑disease surveillance. N Engl J Med. 2023;388(17):1597‑1607. https://doi.org/10.1056/NEJMra2119215
  6. Barnett AJ, Guo Z, Jing J, et al. Improving clinician performance in classifying EEG patterns on the ictal–interictal injury continuum using interpretable machine learning. N Engl J Med AI. 2024;1(6):e2300331. https://doi.org/10.1056/AIoa2300331
  7. Spratt DE, Tang S, Sun Y, et al. Artificial intelligence predictive model for hormone therapy use in prostate cancer. N Engl J Med Evid. 2023;2(8):e2300023. https://doi.org/10.1056/EVIDoa2300023
  8. Holm P, Haue AD, Westergaard D, et al. PMHnet‑alpha: development and validation of a neural network-based discrete‑time survival model for mortality prediction in ischemic heart disease. Eur Heart J Digit Health. 2022;3(4):2785. https://doi.org/10.1093/ehjdh/ztac076.2785
  9. Gebauer S, Eckert C. Survey of US physicians’ attitudes and knowledge of AI. BMJ Evid Based Med. 2024;29(4):279-281. https://doi.org/10.1136/bmjebm-2023-112726
  10. Allen MR, Webb S, Mandvi A, et al. Navigating the doctor‑patient‑AI relationship: a mixed‑methods study of physician attitudes toward artificial intelligence in primary care. BMC Prim Care. 2024;25(1):42. https://doi.org/10.1186/s12875-024-02282-y
  11. Banerjee A, Sarangi PK, Kumar S. Medical doctors’ perceptions of artificial intelligence (AI) in healthcare. Cureus. 2024;16(9):e70508. https://doi.org/10.7759/cureus.70508
  12. Fritsch SJ, Blankenheim A, Wahl A, et al. Attitudes and perception of artificial intelligence in healthcare: a cross‑sectional survey among patients. Digit Health. 2022;8:20552076221116772. https://doi.org/10.1177/20552076221116772
  13. Witkowski K, Okhai R, Neely SR. Public perceptions of artificial intelligence in healthcare: ethical concerns and opportunities for patient‑centered care. BMC Med Ethics. 2024;25(1):74. https://doi.org/10.1186/s12910-024-01066-4
  14. Reis M, Reis F, Kunde W. Influence of believed AI involvement on the perception of digital medical advice. Nat Med. 2024;30(11):3098‑3100. https://doi.org/10.1038/s41591-024-03180-7
  15. Ratwani RM, Bates DW, Classen DC. Patient safety and artificial intelligence in clinical care. JAMA Health Forum. 2024;5(2):e2305514. https://doi.org/10.1001/jamahealthforum.2023.5514
  16. Jongsma KR, Sand M, Milota M. Why we should not mistake accuracy of medical AI for efficiency. NPJ Digit Med. 2024;7(1):57. https://doi.org/10.1038/s41746-024-01047-2
  17. Abdullayev K, Chico TJA, Canson J, et al. Exploring stakeholder perspectives on the barriers and facilitators of implementing digital technologies for heart disease diagnosis: qualitative study. JMIR Cardio. 2025;9:e66464. https://doi.org/10.2196/66464
  18. Gjødsbøl IM, Ringgaard AK, Holm PC, et al. The robot butler: how and why should we study predictive algorithms and artificial intelligence (AI) in healthcare? Digit Health. 2024;10:20552076241241674. https://doi.org/10.1177/20552076241241674
  19. Gualda‑Gea JJ, Barón‑Miras LE, Bertran MJ, et al. Perceptions and future perspectives of medical students on the use of artificial intelligence-based chatbots: an exploratory analysis. Front Med (Lausanne). 2025;12:1529305. https://doi.org/10.3389/fmed.2025.1529305
  20. Han R, Acosta JN, Shakeri Z, et al. Randomised controlled trials evaluating artificial intelligence in clinical practice: a scoping review. Lancet Digit Health. 2024;6(5):e367‑e373. https://doi.org/10.1016/S2589-7500(24)00047-5