Published on in Vol 13 (2024)

This is a member publication of Imperial College London (Jisc)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/52349, first published .
Impact of Responsible AI on the Occurrence and Resolution of Ethical Issues: Protocol for a Scoping Review

Impact of Responsible AI on the Occurrence and Resolution of Ethical Issues: Protocol for a Scoping Review

Impact of Responsible AI on the Occurrence and Resolution of Ethical Issues: Protocol for a Scoping Review

Protocol

1Centre for Health Technology, School of Nursing and Midwifery, University of Plymouth, Plymouth, United Kingdom

2Translational and Clinical Research Institute, Newcastle University, Newcastle-upon-Tyne, United Kingdom

3Department of Primary Care and Public Health, School of Public Health, Imperial College London, London, United Kingdom

4Faculty of Life Sciences and Medicine, King's College London, London, United Kingdom

Corresponding Author:

Edward Meinert, MA, MSc, MBA, MPA, CEng, EUR ING, PhD

Translational and Clinical Research Institute

Newcastle University

Campus for Ageing and Vitality

Westgate Road

Newcastle-upon-Tyne, NE4 6BE

United Kingdom

Phone: 44 01912336161

Email: edward.meinert@newcastle.ac.uk


Background: Responsible artificial intelligence (RAI) emphasizes the use of ethical frameworks implementing accountability, responsibility, and transparency to address concerns in the deployment and use of artificial intelligence (AI) technologies, including privacy, autonomy, self-determination, bias, and transparency. Standards are under development to guide the support and implementation of AI given these considerations.

Objective: The purpose of this review is to provide an overview of current research evidence and knowledge gaps regarding the implementation of RAI principles and the occurrence and resolution of ethical issues within AI systems.

Methods: A scoping review following Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews (PRISMA-ScR) guidelines was proposed. PubMed, ERIC, Scopus, IEEE Xplore, EBSCO, Web of Science, ACM Digital Library, and ProQuest (Arts and Humanities) will be systematically searched for articles published since 2013 that examine RAI principles and ethical concerns within AI. Eligibility assessment will be conducted independently and coded data will be analyzed along themes and stratified across discipline-specific literature.

Results: The results will be included in the full scoping review, which is expected to start in June 2024 and completed for the submission of publication by the end of 2024.

Conclusions: This scoping review will summarize the state of evidence and provide an overview of its impact, as well as strengths, weaknesses, and gaps in research implementing RAI principles. The review may also reveal discipline-specific concerns, priorities, and proposed solutions to the concerns. It will thereby identify priority areas that should be the focus of future regulatory options available, connecting theoretical aspects of ethical requirements for principles with practical solutions.

International Registered Report Identifier (IRRID): PRR1-10.2196/52349

JMIR Res Protoc 2024;13:e52349

doi:10.2196/52349

Keywords



Background

The development of artificial intelligence (AI) is expanding across industries and its implementation has created ethical apprehensions, centered on the responsible creation and use of AI systems. Various guidelines, regulations, and frameworks have emerged with the goal of delivering responsible application and progression of AI [1]. The notable increase in instances of ethical harm has been caused by the misuse of technology (eg, voter manipulation, facial recognition surveillance, and unauthorized mass data collection) or by design flaws in the technology itself (eg, bias in recidivism, loan rejection, and medical misdiagnosis) [2]. Other ethical concerns include the transparent and equitable usage of data, safeguarding data privacy across platforms and systems, ensuring data availability and quality, endorsing data integration for interoperability, and addressing intellectual property concerns [1]. Responsible AI (RAI) promotes the adoption of ethical frameworks and the incorporation of RAI’s 3 main pillars of accountability (ie, the need to explain and validate decisions), responsibility (ie, users interpreting results and identifying errors), and transparency (ie, the need to describe, examine, and duplicate mechanisms) to foster moral responsibility in emerging technology [3].

Rationale

AI ethics is an emerging field that falls within the broader domain of digital ethics and addresses concerns arising from the development and implementation of new digital technologies [4]. Due to AI’s transformative potential, the discourse of guiding values and principles for its development and implementation highlights the need for resolving concerns for AI for data privacy, accountability, and inadvertently fostering bias. Governance against unregulated uses, privacy violations, and algorithm biases requires resilience for security and the preparation for potential attacks. By delivering and maintaining accessibility, safety, accuracy, and fairness, these challenges can be addressed through standardized RAI principles that involve ethical consideration and stakeholder participation [4-8].

The collaboration among diverse stakeholders in developing RAI principles and policies demonstrates the potential for developing robust ethical guidance. RAI principles could have a tangible impact on reducing both the frequency and intensity of ethical issues that arise in various contexts. The implementation of RAI principles exerts a positive influence on the overall ethical landscape, leading to a discernible decrease in the occurrence of ethical dilemmas and mitigating the severity of those that do emerge. The adoption of RAI principles could foster a more accountable AI ecosystem with heightened attention to conducting responsible research and innovation [8].

Recent literature advocates for continued efforts to adhere to RAI principles as a means of fostering a more ethical and trustworthy AI landscape and suggests the incorporation of RAI principles as a vital safeguard, not only in curbing the frequency of ethical issues but also in minimizing their impact. Academia, government, and corporate institutions have published guidelines aimed at fostering the responsible development and deployment of AI technology, encompassing beneficence, nonmaleficence, autonomy, justice or fairness, and accountability, as crucial considerations in ensuring RAI practices [2,9,10]. Various entities, both public and private, offer diverse RAI frameworks, contributing to democratization but lacking consensus and standardization in ethical values. Practical support for AI practitioners is limited, with few frameworks covering all phases of the software development life cycle (SDLC), which ensures the completion of all functionalities, user requirements, objectives, and end goals when creating, planning, and maintaining the development of a software project. The SDLC enhances the overall quality of software programs and the software development process and establishes a structured approach that comprises 7 distinct stages of planning, requirement gathering and analysis, design, coding, testing, deployment, and maintenance [8]. Theoretical frameworks are currently deficient in practical assistance encompassing all SDLC phases in developing, testing, and deploying RAI applications, including practical validation techniques for theoretical principles, as well as supporting stakeholders during the implementation and auditing phases. The synthesis of positive effects of RAI mitigation strategies through the application of RAI principles for minimizing ethical harm and the lack of standardization reveal the shortcomings of complete, uniform, and user-friendly RAI frameworks in current literature. Research value can be added that supports stakeholders across the entire SDLC and is accessible to different stakeholders (both technical and nontechnical), is streamlined, and expedites the adoption of RAI practices [8]. There is a need for both an update and an overview of effective and standardized RAI frameworks to reveal the current state of RAI principles driving positive change and mitigating ethical risks [2,9-14].

Given the evidence of the feasible effectiveness of RAI mitigation strategies for ethical issues in a broader sense, it is necessary to develop a comprehensive and multidisciplinary approach throughout the AI pipeline by embracing a practical and cohesive approach to RAI implementation [14-21]. Current literature on RAI primarily focuses on ethical principles like transparency and fairness but lacks detailed guidance on their practical implementation across the AI development lifecycle [19]. The main gap in the literature lies in transitioning from dynamic and changing theoretical frameworks to actionable strategies for embedding ethical principles in real-world AI applications [19].

Aim and Research Questions

This scoping review will offer an overview and summarize the state of the field of ethical issues inherent to AI systems, linked RAI practices, their strengths and weaknesses, the studies evaluating them, and gaps in the literature. This will help to inform directions of the present ethical issues and current practices to reduce the occurrence and resolve them through RAI principles (Table 1). This study aims to assess the impact of RAI principles on resolving ethical challenges inherent to AI systems. This review will be based on our research question: How does the use of RAI practices and ethical frameworks impact the occurrence and resolution of ethical issues within AI systems?

Table 1. Types of ethical issues and responding RAIa principles [2].
Ethical issue (categorized)DescriptionLinked RAI code (categorized)
Bias or discriminationAIb systems can exhibit bias and discriminate against individuals or groups based on factors such as race, gender, or socioeconomic status. This can lead to unfair treatment and perpetuate existing inequalitiesPrinciple of fairness and nondiscrimination: ensuring that AI systems do not exhibit bias or discriminate against individuals or groups based on protected characteristics
Lack of transparency and explainabilityAI systems often operate as black boxes, hindering understanding of their decision-making process and raising concerns about biases and errorsPrinciple of XAIc: techniques such as interpretable machine learning, model explanations, and rule-based systems are used to provide insights into how AI systems arrive at their decisions, increasing transparency
Privacy and data protectionAI systems rely on personal data, necessitating safeguards and measures to address privacy breaches and unauthorized usePrinciple of privacy: using privacy-enhancing technologies, anonymization techniques, data minimization practices, and secure data handling protocols to protect individuals’ privacy rights and prevent unauthorized access or misuse of personal data
Accountability and responsibilityAI systems can have unintended consequences and significant socioeconomic impacts, requiring comprehensive assessments and proactive measures to mitigate risks and address socioeconomic effectsPrinciple of auditing: conducting rigorous risk assessments, including ethical impact assessments, can help identify and mitigate potential risks and unintended consequences of AI systems. This may involve using frameworks such as the AI Ethics Impact Assessment Tool kit and involving multidisciplinary teams to evaluate the societal, environmental, and economic implications of AI applications

aRAI: responsible artificial intelligence.

bAI: artificial intelligence.

cXAI: explainable artificial intelligence.


Overview of the Study Design

Scoping reviews represent a specialized form of literature review that focuses on charting available literature across a wide-ranging subject area. Scoping reviews are well-suited for outlining present research pinpointing essential insights and areas of understanding that warrant deeper investigation [22]. The Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews (PRISMA-ScR) and Population, Intervention, Comparator, Outcome, and Studies (PICOS) framework were used to build the search strategy (Textbox 1) and provide a methodological framework for the scoping review.

Textbox 1. Population, intervention, comparator, outcome, and studies (PICOS) framework.

Population

Organizations developing artificial intelligence (AI) systems.

Intervention

Implementation of ethical frameworks and responsible artificial intelligence (RAI) principles in practice.

Comparator

Not applicable.

Outcomes

The focus is to examine the impact of RAI principles on ethical issues in AI systems. Secondary goals involve assessing the strengths, weaknesses, benefits, and limitations of current RAI principles in resolving ethical concerns.

Study types

Qualitative, quantitative, and mixed methods studies and literature that present any form of RAI principles for the implementation and resolution of ethical concerns in AI systems are eligible for inclusion. Protocols, reviews, abstracts, and meta-analyses older than 10 years were excluded from this review.

Search Strategy

The review will search 8 databases, including, PubMed, CORE, Scopus, IEEE Xplore, EBSCO, Web of Science, the ACM Digital Library, and ProQuest (Arts and Humanities). Based on a preliminary review of the literature, Medical Subject Headings (MeSH) terms and keywords were identified and grouped into 4 categories. They will be put together in the following way when searching the databases: RAI (MeSH OR Keywords) AND Ethical Issues (MeSH OR Keywords; Table 2).

Table 2. Search terms.
CategoryMeSHaKeywords (in title or abstract)
Ethical issues within AIbEthical challenges in artificial intelligence“Artificial Intelligence” OR “ethical challenges in AI” OR “ethical issues in AI” OR “Machine Learning” OR “AI governance”
RAIcResponsible AI OR machine learning OR deep learning“Responsible AI” OR “RAI principles” OR “supervised deep learning” OR “RAI” OR “supervised machine learning” OR “trustworthy AI” OR “Explainable AI” OR “XAI”
Ethical frameworksAI ethics OR AI governance“AI ethics” OR “ethical AI principles” OR “moral frameworks in AI” OR “ethical frameworks in AI”
RAI managementResponsible AI management“Ethical intervention” OR “Responsible AI Management” OR “RAI management”

aMeSH: Medical Subject Headings.

bAI: artificial intelligence.

cRAI: responsible AI.

Inclusion Criteria

All literature and research proposing RAI principles will be included and studies presenting theoretical impact and suggestions of ethical frameworks will be identified and analyzed independently as well. Interventions aimed at mitigating ethical issues will be included given that they support all main pillars of RAI principles; no comparator is required, and all study types will be eligible for inclusion.

Exclusion Criteria

Studies that do not evaluate RAI principles or responses to ethical issues within the development and use of AI systems such as protocols, reviews, and abstracts will be excluded. Studies that are not published in the English language and studies published before 2013 will not be eligible for inclusion. Considering the rapid technological advancements and the evolving consideration of AI ethics and RAI, expectations in literature older than 10 years may not provide the current insights and information required for a relevant analysis in this field [9].

Screening and Article Selection

Article references will be stored, and duplicates will be removed using the citation management software EndNote X9 (Clarivate). The EndNote X9 search function will also be used to conduct an initial screening of the references based on keywords from the search strategy. The remaining titles and abstracts will be screened, and a full-text review will be conducted by the author (SB) to determine final eligibility.

Data Extraction

Data will be extracted from included studies into a predeveloped charting list, highlighting the data that will be in focus for this scoping review (see Textbox 2).

Textbox 2. Full data charting list.

Title

Title of publication

Author

Full name of first author

Year

Year of publication

Country

The country in which the study is conducted

Quality or quantity

Whether the study entails qualitative, quantitative, or mixed methods research. Added “other” category that may fall outside of categories

Aims or purpose

Stated aims and objectives the research seeks to meet

Source of evidence

Whether the study is primary or secondary research, evidence synthesis, conference abstract, discussion article, etc

Study design

Description of methods and activities

Type of artificial intelligence (AI)–based technology

Notes on specific AI systems, that is, machine learning (ML), deep neural network (DNN), clinical decision support system (CDSS), and AI augmentation

Ethical issues

Inconclusive evidence, inscrutable evidence, misguided evidence, unfair outcomes, transformative effects, traceability, unjustified actions, and opacity

Responsible artificial intelligence (RAI) principles

Respect for human autonomy, prevention of harm, fairness, explicability, and patient privacy

Level of impact (high, medium, and low)

Data on evidence of the impact of RAI principles on the resolution of ethical issues

Probability (high, medium, and low)

Data on the expected likelihood of successful mitigation

Benefits of RAI principles

Notes on the benefits of RAI principles

Limitations of RAI principles and their impact

Notes on the theoretical and practical restrictions of RAI guidelines including affected impact

Strengths and limitations of RAI principles

Notes on the strengths and weaknesses of the use of RAI principles

Data Analysis and Synthesis

A descriptive analysis will be used given the expected large variety of study types. A meta-analysis will be attempted based on the outcome measures of the impact and influence of RAI principles on the occurrence and resolution of ethical issues, which may also indicate the primary outcome of the meta-analysis. The meta-analysis will aggregate findings on how RAI principles impact ethical issue management in AI systems, focusing on effectiveness metrics and stakeholder perceptions. Once all data were extracted, we would assess the feasibility of a meta-analysis. Given the expected variety of outcomes, a quantitative meta-analysis is unlikely to be possible, but we will synthesize findings of the differential impacts of various RAI principles on diverse ethical concerns and applications. There is a possibility that this may not be feasible due to the expected heterogeneity of studies. The extracted data will be summarized in a narrative synthesis to bring together findings relating to implementation challenges and successes of RAI principles, their influence on policy making, and trends in ethical AI research. This approach aims to provide insights into the operationalization of RAI principles and highlight areas for future research.


The results will be included in the full systematic review, which is expected to start in June 2024 and be completed for the submission of publication by the end of 2024.


Expected Findings

This scoping review will provide an overview of the state of the literature regarding the influence of RAI principles on the resolution of ethical issues within AI systems. This section will use the data extracted from studies to explore the conclusions that can be drawn, the limitations of the scoping review, and key areas for future research. Special focus will be placed on relevant stakeholders involved in the ethical AI ecosystem, and studies investigating interventions for RAI development will be summarized and discussed in a subsection. A summary of the current principles, their strengths and weaknesses, and the studies evaluating them will help to inform the development of functional RAI frameworks for mitigating ethical issues for the progress of AI development and use.

Strengths and Limitations

A key strengths of the proposed study is its use of the PRISMA-ScR guidelines to ensure transparent and replicable reporting of the review and its interdisciplinary approach. By including data across different disciplines, the review will be able to develop a nuanced understanding of RAI principles and their implementation in various contexts. We expect this to lead to a comprehensive, cross-disciplinary dialog that enriches the AI community’s perspective on ethical considerations.

The limitations of the scoping review will be discussed in detail in the full review; key limitations are expected to include the scope and the use of only 1 main reviewer. Although the broad scope is a strength in that the findings will not be limited to a specific discipline, including a broad scope will increase the difficulty of synthesizing various findings and determining lessons that are generalizable. It also means that, although we have selected a variety of databases to obtain good coverage of the literature for the search, there is a possibility of missing relevant studies meeting our inclusion criteria. Another limitation is that the literature search, screening, and data analysis will be executed by only 1 reviewer, which has an increased risk of overlooking relevant research and potential biases due to the lack of diverse perspectives and independent verification mechanisms. It is not expected to be possible for a second reviewer to conduct an independent screening due to time and resource constraints, although this will be reassessed when conducting the review. Additionally, the research team does not have proficiency in other languages, which will hinder our capability to review any literature beyond data sets available in English.

Future Directions and Dissemination

Future research could build upon findings from this review by conducting in-depth analyses to identify the gaps in ethical AI frameworks, particularly focusing on stakeholder engagement and interventions for RAI development. Additionally, there is a need for longitudinal studies to assess the effectiveness of emerging principles and interventions in mitigating ethical issues throughout the lifecycle of AI systems. The dissemination plan for this paper will involve publishing the findings in a peer-reviewed journal dedicated to computer sciences, social sciences, philosophy, and ethics. A concise policy brief will be drafted for review to inform and guide regulatory frameworks, ensuring accessibility to legislative audiences. The dissemination will leverage digital platforms, including academic social networks and research repositories, to maximize reach and impact across various sectors.

Conclusions

This scoping review is expected to provide a vital synthesis of research on RAI principles and guidance for efforts to chart a course for the balanced development of AI systems, where innovation is matched with ethical integrity. The scoping review’s findings will likely prompt the development of more nuanced ethical frameworks for AI, influencing both emerging technologies and policy making by highlighting the importance of aligning AI innovation with ethical responsibility. Future research may focus on creating adaptive, real-time monitoring tools to continuously evaluate and guide the ethical implementation of AI as technologies and societal norms evolve.

Acknowledgments

This study was supported by external funding from Parkinson’s UK Non-Drug Awards Programme (grant NDA-21). The views expressed in this review are those of the authors and are not funding agencies or their employers. For this study, no generative artificial intelligence was used.

Data Availability

The authors confirm that the data supporting the findings of this study are available within the paper and appendices (Multimedia Appendices 1 and 2).

Authors' Contributions

The review protocol was conceived and supervised by MM-I and EM. SB drafted the protocol including the research question, use of the population, intervention, comparator, outcome, and studies (PICOS) framework, and the search strategy. All authors contributed revisions.

Conflicts of Interest

EM is the editor-in-chief of JMIRx Med. All other authors have no conflicts of interest to declare.

Multimedia Appendix 1

Search strings.

DOCX File , 15 KB

Multimedia Appendix 2

Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews (PRISMA-ScR) checklist.

DOCX File , 25 KB

  1. Chowdhury NZ, Rahman S. AI and ethics: ensuring responsible and accountable AI development and deployment. ResearchGate. 2023. URL: https://tinyurl.com/49e6vzc [accessed 2024-05-07]
  2. Mikalef P, Conboy K, Lundström JE, Popovič A. Thinking responsibly about responsible AI and 'the dark side' of AI. Eur J Inf Sys. 2022;31(3):257-268. [FREE Full text] [CrossRef]
  3. Stahl BC. Ethical issues of AI. In: Artificial Intelligence for a Better Future. SpringerBriefs in Research and Innovation Governance. Cham, Switzerland. Springer; 2021:35-53.
  4. Dennehy D, Griva A, Pouloudi N, Dwivedi YK, Mäntymäki M, Pappas IO. Artificial intelligence (AI) and information systems: perspectives to responsible AI. Inf Syst Front. 2022;25(1):1-7. [FREE Full text] [CrossRef]
  5. Kazim E, Koshiyama AS. A high-level overview of AI ethics. Patterns (N Y). 2021;2(9):100314. [FREE Full text] [CrossRef] [Medline]
  6. Thiebes S, Lins S, Sunyaev A. Trustworthy artificial intelligence. Electron Markets. 2020;31(2):447-464. [FREE Full text] [CrossRef]
  7. Anagnostou M, Karvounidou O, Katritzidaki C, Kechagia C, Melidou K, Mpeza E, et al. Characteristics and challenges in the industries towards responsible AI: a systematic literature review. Ethics Inf Technol. 2022;24(3):37. [CrossRef]
  8. Schiff D, Rakova B, Ayesh A, Fanti A, Lennon M. Principles to practices for responsible AI: closing the Gap. arXiv. [FREE Full text]
  9. Haug CJ, Drazen JM. Artificial intelligence and machine learning in clinical medicine, 2023. N Engl J Med. 2023;388(13):1201-1208. [CrossRef] [Medline]
  10. Barletta VS, Caivano D, Gigante D, Ragone A. A rapid review of responsible AI frameworks: how to guide the development of ethical AI. 2023. Presented at: Proceedings of the 27th International Conference on Evaluation and Assessment in Software Engineering; June 14-16, 2023:358-367; Oulu, Finland. [CrossRef]
  11. Dwivedi N, Katiyar D, Goel G. A comparative study of various software development lfe cycle (SDLC) models. Int J Res Eng Sci Manag. 2022;5(3):141-144. [FREE Full text]
  12. Oprescu AM, Miró-Amarante G, García-Díaz L, Rey VE, Chimenea-Toscano A, Martínez-Martínez R, et al. Towards a data collection methodology for responsible artificial intelligence in health: a prospective and qualitative study in pregnancy. Inform Fusion. 2022;83-84:53-78. [FREE Full text] [CrossRef]
  13. Linardatos P, Papastefanopoulos V, Kotsiantis S. Explainable AI: a review of machine learning interpretability methods. Entropy (Basel). 2020;23(1):18. [FREE Full text] [CrossRef] [Medline]
  14. Buchholz J, Lang B, Vyhmeister E. The development process of responsible AI: the case of ASSISTANT*. IFAC-PapersOnLine. 2022;55(10):7-12. [FREE Full text] [CrossRef]
  15. Dignum V. Responsible artificial intelligence—from principles to practice. arXiv. Preprint posted online on May 22, 2022. [CrossRef]
  16. Tsamados A, Aggarwal N, Cowls J, Morley J, Roberts H, Taddeo M, et al. The ethics of algorithms: key problems and solutions. AI & Soc. 2021;37(1):215-230. [FREE Full text] [CrossRef]
  17. Vassileva B, Zwilling M. Responsible AI and Ethical Issues for Businesses and Governments. Hershey, PA. IGI Global; 2020.
  18. Floridi L, Cowls J, Beltrametti M, Chatila R, Chazerand P, Dignum V, et al. AI4People-an ethical framework for a good AI society: opportunities, risks, principles, and recommendations. Minds Mach (Dordr). 2018;28(4):689-707. [FREE Full text] [CrossRef] [Medline]
  19. Lu Q, Zhu L, Xu X, Whittle J. Responsible-AI-by-design: a pattern collection for designing responsible artificial intelligence systems. IEEE Softw. 2023;40(3):63-71. [CrossRef]
  20. Srivastava S, Sinha K. From bias to fairness: a review of ethical considerations and mitigation strategies in artificial intelligence. IJRASET. 2023;11(III):2247-2251. [FREE Full text] [CrossRef]
  21. Benjamins R, Barbado A, Sierra D. Responsible AI by design in practice. arXiv. Preprint posted online on December 20, 2019. [CrossRef]
  22. Rumrill PD, Fitzgerald SM, Merchant WR. Using scoping literature reviews as a means of understanding and interpreting existing literature. Work. 2010;35(3):399-404. [FREE Full text] [CrossRef] [Medline]


AI: artificial intelligence
MeSH: Medical Subject Headings
PICOS: population, intervention, comparator, outcome, and studies
PRISMA-ScR: Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews
RAI: responsible artificial intelligence
SDLC: software development life cycle


Edited by S Ma; submitted 01.09.23; peer-reviewed by B Chiyangwa, J McNealy; comments to author 05.04.24; revised version received 10.04.24; accepted 15.04.24; published 05.06.24.

Copyright

©Selina Boege, Madison Milne-Ives, Edward Meinert. Originally published in JMIR Research Protocols (https://www.researchprotocols.org), 05.06.2024.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Research Protocols, is properly cited. The complete bibliographic information, a link to the original publication on https://www.researchprotocols.org, as well as this copyright and license information must be included.