AI And Legal Liability: Who Is Responsible When Machines Go Wrong?
- IJLLR Journal
- 1 day ago
- 1 min read
Shreya Goyal, NIMS University Jaipur Rajasthan
ABSTRACT
Artificial Intelligence (AI) has experienced swift growth across various sectors, such as healthcare, transportation, finance, and legal practices, introducing both distinct opportunities and challenges. A key legal concern that emerges is the issue of accountability when AI systems fail or inflict harm. Conventional legal frameworks, including tort, product, and contract law, often struggle to address the complexities introduced by autonomous decision-making, algorithmic opacity, and the evolution of machine learning technologies. This research paper conducts a comprehensive analysis of the legal principles that currently dictate liability for damages caused by AI, assesses significant case studies, and explores comparative approaches across multiple jurisdictions, including the US, India, and the EU. It also examines the viability of ideas such as digital personhood and frameworks of collective responsibility. The piece proposes the development of a tailored legal framework that integrates human responsibility with algorithmic accountability, ensuring both progress in technology and comprehension of the law. It recommends a united strategy of responsibility and regulatory oversight that can more effectively meet the demands of a future shaped by AI.
Keywords: Artificial Intelligence (AI), Legal Liability, Accountability, Autonomous Decision-Making, Algorithmic Opacity, Machine Learning, Tort Law, Product Liability, Contract Law, Comparative Jurisprudence, Digital Personhood, Collective Responsibility, Human Responsibility, Regulatory Oversight, Tailored Legal Framework, Technological Innovation, AI Governance, EU AI Act, India AI Policy, US AI Regulation