The principle "Key ethical principles for use of artificial intelligence for health" has mentioned the topic "responsibility" in the following places:

    4 Foster responsibility and accountability

    4 Foster responsibility and accountability

    4 Foster responsibility and accountability

    Although AI technologies perform specific tasks, it is the responsibility of human stakeholders to ensure that they can perform those tasks and that they are used under appropriate conditions.

    4 Foster responsibility and accountability

    responsibility can be assured by application of “human warranty”, which implies evaluation by patients and clinicians in the development and deployment of AI technologies.

    4 Foster responsibility and accountability

    The goal is to ensure that the algorithm remains on a machine learning development path that is medically effective, can be interrogated and is ethically responsible; it involves active partnership with patients and the public, such as meaningful public consultation and debate (101).

    4 Foster responsibility and accountability

    The use of AI technologies in medicine requires attribution of responsibility within complex systems in which responsibility is distributed among numerous agents.

    4 Foster responsibility and accountability

    The use of AI technologies in medicine requires attribution of responsibility within complex systems in which responsibility is distributed among numerous agents.

    4 Foster responsibility and accountability

    When medical decisions by AI technologies harm individuals, responsibility and accountability processes should clearly identify the relative roles of manufacturers and clinical users in the harm.

    4 Foster responsibility and accountability

    Institutions have not only legal liability but also a duty to assume responsibility for decisions made by the algorithms they use, even if it is not feasible to explain in detail how the algorithms produce their results.

    4 Foster responsibility and accountability

    To avoid diffusion of responsibility, in which “everybody’s problem becomes nobody’s responsibility”, a faultless responsibility model (“collective responsibility”), in which all the agents involved in the development and deployment of an AI technology are held responsible, can encourage all actors to act with integrity and minimize harm.

    4 Foster responsibility and accountability

    To avoid diffusion of responsibility, in which “everybody’s problem becomes nobody’s responsibility”, a faultless responsibility model (“collective responsibility”), in which all the agents involved in the development and deployment of an AI technology are held responsible, can encourage all actors to act with integrity and minimize harm.

    4 Foster responsibility and accountability

    To avoid diffusion of responsibility, in which “everybody’s problem becomes nobody’s responsibility”, a faultless responsibility model (“collective responsibility”), in which all the agents involved in the development and deployment of an AI technology are held responsible, can encourage all actors to act with integrity and minimize harm.

    4 Foster responsibility and accountability

    To avoid diffusion of responsibility, in which “everybody’s problem becomes nobody’s responsibility”, a faultless responsibility model (“collective responsibility”), in which all the agents involved in the development and deployment of an AI technology are held responsible, can encourage all actors to act with integrity and minimize harm.

    4 Foster responsibility and accountability

    To avoid diffusion of responsibility, in which “everybody’s problem becomes nobody’s responsibility”, a faultless responsibility model (“collective responsibility”), in which all the agents involved in the development and deployment of an AI technology are held responsible, can encourage all actors to act with integrity and minimize harm.