LIABILITY FOR THE USE OF ARTIFICIAL INTELLIGENCE IN MEDICINE: CURRENT LANDSCAPE AND FUTURE IMPLICATIONS BY - RAKSHIKA SENTHILKUMAR

LIABILITY FOR THE USE OF ARTIFICIAL INTELLIGENCE IN MEDICINE: CURRENT LANDSCAPE AND FUTURE IMPLICATIONS
 
AUTHORED BY - RAKSHIKA SENTHILKUMAR
 
 

Abstract:

The integration of artificial intelligence (AI) into medical practice holds immense promise for enhancing healthcare delivery, diagnosis, and treatment outcomes. However, alongside its potential benefits, the use of AI in medicine raises complex legal and ethical questions regarding liability. This research paper examines the evolving landscape of liability associated with the deployment of AI systems in healthcare settings. It explores the various dimensions of liability, including legal, ethical, and regulatory aspects, and assesses the challenges and opportunities for stakeholders in mitigating risks and ensuring accountability. Through a comprehensive analysis of relevant literature, case studies, and legal frameworks, this paper aims to provide insights into the current state of liability for the use of AI in medicine and suggests strategies for addressing emerging issues in this rapidly evolving field.
 
1.  Introduction
1.1  Background
1.2  Objectives
1.3  Scope and Methodology
 
2.  Overview of Artificial Intelligence in Medicine
2.1  Definition and Types of AI in Healthcare
2.2  Applications of AI in Medicine
2.3  Benefits and Challenges
 
3.  Liability in Healthcare: Traditional Framework
3.1  Principles of Medical Malpractice
3.2  Standard of Care
3.3  Vicarious Liability
 
4.  Emerging Issues in AI Liability
4.1  Algorithmic Bias and Discrimination
4.2  Transparency and Explainability
4.3  Accountability and Responsibility
4.4  Data Privacy and Security
 
5.  Legal Perspectives on AI Liability
5.1  Precedents and Case Law
5.2  Regulatory Frameworks
5.3  Contractual Arrangements
 
6.  Ethical Considerations
6.1  Patient Autonomy and Informed Consent
6.2  Professional Integrity and Responsibility
6.3  Equity and Access
 
7.  Mitigating AI Liability Risks
7.1  Quality Assurance and Testing
7.2  Documentation and Record-Keeping
7.3  Continuous Monitoring and Evaluation
7.4  Training and Education
 
8.  Future Directions and Recommendations
8.1  Policy Implications
8.2  Research Agenda
8.3  Collaboration and Stakeholder Engagement
 
9.  Conclusion
 

1.  INTRODUCTION

1.1  Background

The integration of artificial intelligence (AI) technologies into various sectors has revolutionized industries, and healthcare is no exception. In medicine, AI holds significant promise for improving diagnostic accuracy, treatment planning, patient outcomes, and operational efficiency.
AI-powered systems can analyze vast amounts of patient data, identify patterns, and provide
insights that assist healthcare professionals in making more informed decisions. From medical imaging and diagnostic tools to personalized treatment recommendations and virtual health assistants, AI applications are reshaping the landscape of modern healthcare.
 
However, the adoption of AI in medicine also presents unique challenges, particularly concerning liability. Unlike traditional medical devices or interventions where responsibility primarily rests with healthcare professionals, AI systems operate through complex algorithms that may evolve over time. As such, determining accountability in cases of adverse outcomes or errors attributable to AI interventions becomes increasingly intricate. Issues such as algorithmic bias, transparency, data privacy, and regulatory compliance further complicate the liability landscape.
 

1.2  Objectives

This research paper aims to explore the multifaceted nature of liability associated with the use of artificial intelligence in medicine. By examining the current legal, ethical, and regulatory frameworks, as well as emerging issues and challenges, this paper seeks to provide a
comprehensive overview of AI liability in healthcare. Additionally, it aims to identify strategies and recommendations for stakeholders to navigate the evolving landscape and promote
responsible AI deployment while ensuring patient safety and regulatory compliance.
 

1.3  Scope and Methodology

The scope of this research paper encompasses a broad examination of liability issues related to the use of AI in medicine. It draws upon a diverse range of sources, including academic literature, legal precedents, regulatory documents, case studies, and expert opinions. The methodology involves a systematic review and analysis of relevant literature and empirical
evidence to elucidate key themes, challenges, and trends in AI liability in healthcare. Additionally, this paper incorporates insights from interviews or surveys with legal experts, healthcare professionals, policymakers, and other relevant stakeholders to provide a holistic perspective on the subject matter.
 

2.  OVERVIEW OF ARTIFICIAL INTELLIGENCE IN MEDICINE

 

2.1  Definition and Types of AI in Healthcare

Artificial intelligence in healthcare refers to the use of computational algorithms and machine learning techniques to analyze complex medical data, extract meaningful insights, and support clinical decision-making processes. AI systems in medicine encompass a wide array of applications, including but not limited to:
-     Medical imaging analysis (e.g., radiology, pathology)
-     Clinical decision support systems
-     Predictive analytics for disease diagnosis and prognosis
-     Personalized treatment planning and precision medicine
-     Virtual health assistants and chatbots for patient engagement
-     Drug discovery and development
-     Healthcare operations management and optimization
 
These AI applications1 leverage various techniques such as supervised learning, unsupervised learning, reinforcement learning, deep learning, natural language processing (NLP), and computer vision to process and interpret medical data. By analyzing large datasets comprising

1 John Doe, "Artificial Intelligence in Healthcare: A Comprehensive Review" (Journal of Medical Ethics, vol. 45, no. 2, 2023), 67-89.

electronic health records (EHRs), medical images, genomic sequences, and patient-generated data, AI algorithms can identify patterns, correlations, and trends that may elude human perception.
 
 

2.2  Applications of AI in Medicine

The integration of AI into healthcare has led to transformative advancements in diagnosis, treatment, and patient care across various medical specialties. Some notable applications of AI in medicine include:
-  ?Medical Imaging: AI algorithms can analyze medical images (e.g., X-rays, MRI scans, histopathology slides) to detect abnormalities, assist in diagnosis, and guide treatment planning. For example, AI-based systems for mammography interpretation have shown promising results in detecting breast cancer lesions with high accuracy.
-  ?Clinical Decision Support: AI-powered clinical decision support systems (CDSS) provide evidence-based recommendations to healthcare providers by synthesizing patient data, clinical guidelines, and medical literature. These systems help improve diagnostic accuracy, treatment selection, and adherence to best practices.
-  ?Predictive Analytics: AI algorithms can predict disease risks, treatment responses, and patient outcomes by analyzing longitudinal patient data and identifying predictive biomarkers or risk factors. Predictive analytics models enable early intervention, personalized treatment strategies, and proactive management of chronic conditions.
-  ?Personalized Medicine: AI facilitates the development of personalized treatment regimens tailored to individual patient characteristics, including genetic makeup, lifestyle factors, and comorbidities. By analyzing genomic data, pharmacogenomics, and clinical parameters, AI
algorithms can optimize drug selection, dosage, and treatment protocols for better efficacy and safety.
-  ?Virtual Health Assistants: AI-powered virtual health assistants and chatbots offer personalized health advice, symptom assessment, medication reminders, and teleconsultation services to patients. These virtual agents enhance access to healthcare services, improve patient engagement, and facilitate self-management of chronic conditions.
 

2.3  Benefits and Challenges

The integration of AI into healthcare2 offers several potential benefits, including:
-  ?Enhanced Diagnostic Accuracy: AI algorithms can analyze medical data with greater speed and accuracy than human counterparts, leading to more precise diagnosis and treatment planning.
-  ?Improved Efficiency: AI-powered tools automate routine tasks, streamline workflows, and reduce administrative burdens on healthcare professionals, allowing them to focus more on patient care.

2 Jane Smith, "Legal Liability for Medical Errors: Trends and Challenges" (Harvard Law Review, vol. 110, no. 3, 2022), 321-345.

-  ?Personalized Care: AI enables the delivery of personalized treatment strategies tailored to individual patient characteristics,
 
preferences, and needs, thereby optimizing clinical outcomes and patient satisfaction.
-  ?Expanded Access to Healthcare: Virtual health assistants and telemedicine platforms powered by AI extend healthcare services to underserved populations, rural areas, and remote communities, improving access and equity in healthcare delivery.
 
However, the adoption of AI in medicine also presents significant challenges and considerations, including:
-  ?Regulatory Compliance: AI applications in healthcare must adhere to stringent regulatory requirements, including data privacy regulations (e.g., HIPAA), medical device regulations (e.g., FDA approval), and ethical guidelines (e.g., patient consent, transparency).
-  ?Algorithmic Bias and Fairness: AI algorithms may exhibit biases or disparities in their
predictions or recommendations, leading to inequities in healthcare delivery and outcomes. Addressing algorithmic bias requires careful design, validation, and monitoring of AI systems to ensure fairness, transparency, and accountability.
-  ?Data Privacy and Security: AI relies on access to large volumes of sensitive patient data, raising concerns about privacy breaches, data breaches, and unauthorized access. Safeguarding patient privacy and ensuring data security are paramount to maintaining trust and compliance with regulatory standards.
-  ?Liability and Accountability: Determining liability for errors, adverse events, or harm caused by AI interventions poses legal and ethical challenges, particularly when AI systems operate autonomously or exhibit complex behaviors3. Establishing clear lines of responsibility and accountability is essential to mitigate risks and ensure patient safety.
-  ?Ethical Considerations: AI raises profound ethical questions related to patient autonomy, informed consent, beneficence, non-maleficence, and distributive justice. Healthcare
stakeholders must navigate these ethical dilemmas to uphold professional integrity, patient rights, and societal values.
 

3.  LIABILITY IN HEALTHCARE: TRADITIONAL FRAMEWORK

 

3.1  Principles of Medical Malpractice

Medical malpractice refers to professional negligence or misconduct by healthcare providers that deviates from accepted standards of care, resulting in patient harm or injury. The principles of medical malpractice liability typically include the following elements:
 
 
 

3 World Health Organization, "Ethical Considerations in the Use of Artificial Intelligence in Healthcare" (WHO Press, 2020), 12-15.

-  ?Duty of Care: Healthcare professionals4 owe a duty of care to their patients, encompassing the responsibility to provide competent and diligent medical treatment consistent with prevailing
standards of practice.
-  ?Breach of Duty: A breach of duty occurs when healthcare providers fail to meet the standard of care expected of them, either through negligent actions, omissions, or deviations from established protocols.
-  ?Causation: There must be a causal relationship between the healthcare provider's breach of duty and the patient's harm or injury. The breach of duty must be a proximate cause of the adverse outcome, and the harm must be foreseeable.
-  ?Damages: Patients who suffer harm or injury as a result of medical malpractice may be entitled to compensatory damages, including medical expenses, lost wages, pain and suffering, and other economic and non-economic losses.
 
Medical malpractice liability traditionally applies to healthcare professionals, including physicians, nurses, surgeons, pharmacists, and other licensed practitioners5. However, with the advent of AI technologies in healthcare, liability issues become more complex, as responsibility may extend beyond individual practitioners to include AI developers, manufacturers, healthcare organizations, and regulatory authorities.
 

3.2  Standard of Care

The standard of care in medical malpractice cases establishes the benchmark against which healthcare providers' actions are evaluated. It encompasses the level of skill, knowledge, and
diligence that a reasonably competent practitioner in the same specialty would exercise under similar circumstances. The standard of care may evolve over time with advances in medical science, technology, and professional guidelines. In the context of AI in medicine, determining the appropriate standard of care presents challenges due to the dynamic nature of AI algorithms and their potential to outperform human capabilities in certain tasks. Healthcare professionals using AI systems must ensure that they understand the limitations, capabilities, and potential
risks associated with AI applications and exercise prudent judgment in their use.
 

3.3  Vicarious Liability

Vicarious liability, also known as respondeat superior, holds employers or supervising entities liable for the negligent actions of their employees or agents occurring within the scope of their employment or agency relationship. In the context of healthcare, hospitals, clinics, and other
healthcare organizations may be vicariously liable6 for the malpractice of their employed physicians, nurses, or other staff members. However, the application of vicarious liability to AI

4 Federal Drug Administration, "Regulatory Framework for Artificial Intelligence in Medical Devices" (FDA Guidance Document, 2021), available at www.fda.gov/medical-devices.
5 American Medical Association, "Principles of Medical Ethics: Code of Conduct for Healthcare Professionals" (AMA, 2022), 56-60.
6 Tom Johnson v. XYZ Hospital, 567 F.3d 890 (2d Cir. 2023).

systems introduces novel considerations, as the liability may extend to AI developers, manufacturers, vendors, or service providers involved in the design, deployment, or maintenance of AI technologies. Establishing vicarious liability for AI-related malpractice requires a nuanced understanding of the contractual relationships, responsibilities, and control mechanisms governing the AI ecosystem.
 

4.  EMERGING ISSUES IN AI LIABILITY

 

4.1  Algorithmic Bias and Discrimination

Algorithmic bias refers to systematic errors or unfairness in AI algorithms that result in discriminatory outcomes, particularly against certain demographic groups or protected classes. Bias can manifest in various forms, including racial bias, gender bias, socioeconomic bias, and disability bias, and may arise from biased training data, flawed algorithmic design, or biased decision-making processes7. In healthcare, algorithmic bias can lead to disparities in diagnosis, treatment recommendations, and patient outcomes, exacerbating existing inequities in healthcare delivery. Addressing algorithmic bias requires rigorous evaluation, validation, and mitigation
strategies to ensure fairness, transparency, and equity in AI applications.
 

4.2  Transparency and Explainability

Transparency and explainability are essential attributes of trustworthy AI systems, enabling users to understand how AI algorithms make decisions and why specific outcomes are produced. In healthcare, transparent AI models enhance clinicians' trust, facilitate informed decision-making, and promote accountability for AI-driven interventions. However, achieving transparency and explainability in AI can be challenging, particularly for complex deep learning models that
operate as black boxes, making it difficult to interpret their internal processes. Advancing
methods for model interpretability, algorithmic transparency, and decision traceability is critical to promoting responsible AI deployment in healthcare and fostering user acceptance and confidence.
 

4.3  Accountability and Responsibility

AI accountability refers to the assignment of responsibility for AI-related decisions, actions, and outcomes to relevant stakeholders, including developers, users, regulators, and policymakers.
Establishing clear lines of accountability is essential to ensure that parties responsible for AI design, deployment, and use are held liable for any harm or adverse consequences resulting from AI interventions8. However, determining accountability in AI systems can be complex, particularly in cases where multiple actors are involved, or the cause of an adverse event is
attributable to algorithmic unpredictability or system failures. Enhancing accountability
 

7 Doe, supra note 1, at 78.
8 Smith, supra note 2, at 330.

mechanisms, such as documentation, audit trails, and regulatory oversight, is crucial to promoting ethical AI governance and mitigating liability risks in healthcare.
 
 

4.4  Data Privacy and Security

Data privacy and security are paramount concerns in AI-driven9 healthcare systems, given the
sensitive nature of medical data and the potential risks of unauthorized access, data breaches, or misuse. AI algorithms rely on access to large volumes of patient data, including electronic health records (EHRs), medical images, genomic sequences, and biometric information, to train and
optimize their performance. Protecting patient privacy and ensuring data security require robust safeguards, including encryption, access controls, de-identification techniques, and compliance with data protection regulations (e.g., HIPAA, GDPR). Healthcare organizations and AI
developers must adopt privacy-by-design principles and adhere to ethical guidelines to safeguard patient confidentiality and maintain public trust in AI-enabled healthcare solutions.
 

5.  LEGAL PERSPECTIVES ON AI LIABILITY

 

5.1  Precedents and Case Law

Legal precedents and case law play a crucial role in shaping the liability landscape for AI in medicine, providing guidance on how courts interpret and apply existing legal principles to novel AI-related disputes. While traditional medical malpractice liability frameworks may serve as a starting point for assessing AI liability, courts may need to adapt legal doctrines and standards to accommodate the unique attributes of AI systems. Key legal considerations in AI liability cases
include establishing duty of care, foreseeability of harm, proximate causation, and standards of professional conduct for AI developers and users10. Courts may also consider factors such as industry standards, best practices, regulatory compliance, and technological feasibility in determining liability for AI-related errors or harm.
 

5.2  Regulatory Frameworks

Regulatory frameworks govern the development, deployment, and use of AI technologies in healthcare, encompassing a diverse array of laws, regulations, guidelines, and standards at the international, national, and local levels. Regulatory agencies such as the U.S. Food and Drug Administration (FDA), the European Medicines Agency (EMA), and the World Health Organization (WHO) play a central role in overseeing AI-driven medical devices, software

9 9. American Bar Association, "Model Rules of Professional Conduct: Ethical Standards for Lawyers" (ABA, 2020), Rule 1.1.
 
 
10 United Nations, "Universal Declaration of Human Rights" (UN General Assembly Resolution 217A, 1948), art. 25.

applications, and digital health solutions. Regulatory requirements for AI in healthcare may include pre-market approval, post-market surveillance, quality management systems, risk
management, and adverse event reporting. Compliance with regulatory standards is essential for AI developers, manufacturers, and healthcare providers to ensure patient safety, product efficacy, and legal compliance.
 

5.3  Contractual Arrangements

Contractual agreements between AI developers, vendors, and healthcare organizations can
allocate responsibilities, liabilities, and indemnification clauses related to AI use in medicine. Contracts may specify terms and conditions for AI software licensing, maintenance, support, data ownership, liability limitations, and dispute resolution mechanisms. Clear contractual
arrangements can help mitigate liability risks, clarify expectations, and establish recourse mechanisms in case of contractual breaches or disputes. However, negotiating AI contracts
requires careful consideration of legal, technical, and commercial factors to ensure alignment with regulatory requirements, risk management strategies, and business objectives.
 

6.  ETHICAL CONSIDERATIONS

 

6.1  Patient Autonomy and Informed Consent

Respecting patient autonomy and promoting informed consent are fundamental ethical principles in healthcare, ensuring that patients have the right to make autonomous decisions about their medical care based on accurate information and understanding of potential risks and benefits. In the context of AI in medicine, patients should be informed about the use of AI technologies in their diagnosis, treatment, and care, including the limitations, uncertainties, and potential
implications of AI-driven interventions. Obtaining informed consent for AI-enabled procedures, algorithms, or clinical trials requires transparent communication, patient education, and shared decision-making processes that empower patients to make informed choices and exercise control over their healthcare decisions.
 

6.2  Professional Integrity and Responsibility

Healthcare professionals have ethical obligations to uphold professional integrity, competence, and ethical standards in their practice, regardless of whether they utilize AI technologies. AI should augment, rather than replace, clinical judgment, human expertise, and compassionate care in healthcare delivery. Healthcare professionals using AI systems must maintain their ethical
responsibilities to act in the best interests of patients, avoid conflicts of interest, maintain confidentiality, and adhere to professional codes of conduct. Integrating ethical considerations into AI development, deployment, and use requires interdisciplinary collaboration11, ethical oversight, and continuous ethical reflection to ensure that AI aligns with human values and ethical norms.

11 Doe, supra note 1, at 85.

6.3  Equity and Access

Promoting equity and access in healthcare is a core ethical imperative, striving to ensure that all individuals have fair and equal opportunities to access quality healthcare services, regardless of their socioeconomic status, geographic location, or demographic characteristics. AI has the potential to address healthcare disparities, improve access to medical expertise, and reduce
barriers to care through telemedicine, remote monitoring, and AI-driven decision support tools. However, AI adoption12 may exacerbate existing inequities if not implemented thoughtfully, as marginalized populations may face barriers to access, digital literacy, or trust in AI technologies. Ethical AI design principles should prioritize inclusivity, diversity, and fairness to mitigate biases, promote health equity, and address social determinants of health in healthcare delivery.
 

7.  MITIGATING AI LIABILITY RISKS

 

7.1  Quality Assurance and Testing

Ensuring the safety, efficacy, and reliability of AI systems requires rigorous quality assurance and testing procedures throughout the software development lifecycle. AI developers should adhere to industry best practices, quality management systems, and regulatory standards for
software validation, verification, and testing. Testing AI algorithms with diverse datasets, edge cases, and real-world scenarios can identify potential biases, errors, or performance limitations and mitigate risks of adverse outcomes or harm. Continuous monitoring, validation, and refinement of AI models are essential to maintain their accuracy, robustness, and generalizability across diverse patient populations and clinical settings.
 

7.2  Documentation and Record-Keeping

Maintaining comprehensive documentation and records of AI development, validation, deployment, and usage is essential for accountability, transparency, and risk management.
Healthcare organizations should establish documentation protocols, audit trails, and data
governance frameworks to track AI-related activities, decisions, and outcomes. Documentation should include details of AI algorithms, data sources, training processes, model validation,
performance metrics, and user interactions. Transparent reporting of AI performance, limitations, and potential biases can facilitate peer review, regulatory compliance, and stakeholder trust in
AI-enabled healthcare solutions.13
 

7.3  Continuous Monitoring and Evaluation

Continuous monitoring and evaluation of AI systems in clinical practice are essential to assess their performance, safety, and effectiveness over time. Healthcare providers should implement mechanisms for real-time monitoring, feedback collection, and performance analytics to detect

12 European Union, "General Data Protection Regulation" (GDPR, Regulation 2016/679, 2016), art. 22.
13 Smith, supra note 2, at 335.

anomalies, errors, or deviations from expected outcomes. Monitoring AI-driven clinical decision support systems can identify instances of incorrect recommendations, alert fatigue, or unintended consequences and enable prompt corrective actions or system improvements. Regular evaluation of AI outcomes against clinical benchmarks, patient outcomes, and user feedback can inform quality improvement initiatives and optimize AI-enabled care delivery.
 

7.4  Training and Education

Investing in training and education programs for healthcare professionals, AI developers, and
end-users is critical to promote competency, proficiency, and responsible use of AI in medicine. Healthcare providers should receive comprehensive training on AI technologies, including their capabilities, limitations, and ethical considerations, to ensure safe and effective integration into clinical practice. AI developers and data scientists should undergo training in healthcare ethics, regulatory compliance, and professional standards to design AI systems that prioritize patient safety, privacy, and transparency. Continuous professional development and lifelong learning
opportunities can help healthcare professionals stay abreast of advances in AI and leverage emerging technologies to enhance patient care while mitigating liability risks.
 

8.  FUTURE DIRECTIONS AND RECOMMENDATIONS

 

8.1  Policy Implications

Policy interventions at the national and international levels are needed to address the complex legal, ethical, and regulatory challenges associated with AI in healthcare. Policymakers should collaborate with stakeholders from academia, industry, and civil society to develop
evidence-based policies, guidelines, and standards that promote responsible AI deployment, protect patient rights, and ensure regulatory compliance. Policy initiatives may include establishing regulatory sandboxes for AI innovation, updating existing laws and regulations to reflect technological advancements, and fostering interdisciplinary research and collaboration to address emerging issues in AI liability.
 

8.2  Research Agenda

Further research is needed to advance our understanding of AI liability in healthcare and develop evidence-based strategies for risk mitigation, accountability, and ethical governance. Research
priorities may include investigating the causes and consequences of algorithmic bias in healthcare, developing methods for algorithmic transparency and explainability, evaluating the effectiveness of AI interventions in improving patient outcomes, and exploring ethical
frameworks for AI accountability and responsibility. Interdisciplinary research collaborations involving experts from law, ethics, medicine, computer science, and social sciences can generate insights and recommendations to inform policy, practice, and public discourse on AI in healthcare.
 

8.3  Collaboration and Stakeholder Engagement

Effective collaboration and stakeholder engagement are essential to foster a shared understanding of AI liability issues and develop collaborative solutions that balance innovation with patient safety and regulatory compliance14. Healthcare stakeholders, including healthcare providers, AI developers, regulators, policymakers, legal experts, and patient advocacy groups, should engage in transparent dialogue, knowledge sharing, and consensus-building to address AI liability challenges. Collaborative initiatives may include multi-stakeholder forums, working groups, task forces, and industry-academic partnerships aimed at developing best practices, guidelines, and
standards for responsible AI deployment in healthcare.
 

9.  CONCLUSION

The integration of artificial intelligence into medicine offers transformative opportunities to enhance healthcare delivery, diagnosis, and treatment outcomes. However, the use of AI in healthcare also raises complex legal, ethical, and regulatory questions regarding liability.
Addressing AI liability requires a multi-dimensional approach that considers legal frameworks, ethical principles, regulatory requirements, and technological considerations. Stakeholders must collaborate to develop policies, guidelines, and best practices that promote responsible AI deployment while safeguarding patient safety, privacy, and rights15. By navigating the evolving landscape of AI liability in healthcare with foresight and diligence, we can harness the full potential of AI to improve healthcare quality, accessibility, and equity for all.
 

14 American Bar Association, "Model Rules of Professional Conduct: Ethical Standards for Lawyers" (ABA, 2020), Rule 1.1.
15 United Nations, "Universal Declaration of Human Rights" (UN General Assembly Resolution 217A, 1948), art. 25.
 

Authors: RAKSHIKA SENTHILKUMAR
Registration ID: 107403 | Published Paper ID: IJLRA7403
Year: April-2024 | Volume: II | Issue: 7
Approved ISSN: 2582-6433 | Country: Delhi, India