Open Access Research Article

AN IN-DEPTH ANALYSIS OF THE ETHICAL IMPLICATIONS & INDIAS ADVANCEMENT ON ARTIFICIAL INTELLIGENCE (AI) REGULATIONS - THE WAY FORWARD

Author(s):
AKASH RAGHAV PRATEEK KUMAR RAGHAV VAIBHAV PANDEY
Journal IJLRA
ISSN 2582-6433
Published 2025/03/28
Access Open Access
Issue 7

Published Paper

PDF Preview

Article Details

AN IN-DEPTH ANALYSIS OF THE ETHICAL IMPLICATIONS & INDIA’S ADVANCEMENT ON ARTIFICIAL INTELLIGENCE (AI) REGULATIONS – THE WAY FORWARD
 
AUTHORED BY - AKASH RAGHAV,
PRATEEK KUMAR RAGHAV & VAIBHAV PANDEY
 
 
Abstract:
ARTIFICIAL INTELLIGENCE, or AI, plays a pivotal role in shaping the skills of human beings, obtaining insights, generating businesses, and addressing the various concerns of society like environmental concerns and many more. The utility of applications of Artificial Intelligence (AI) has acquired a significant place within a short period by bringing out several conveniences and benefits. Nevertheless, the expansion of AI has also paved the way for ethical concerns, viz., breach of privacy, algorithmic discrimination, reliability & security concerns, transparency, and some other unintentional effects. AI-deployed structure is significantly developed by the standard groups and societies, organizations dealing with the technology, groups of the analysts, and the governmental agencies. This paper deals with the main theme of ethical implications and comprehensive analysis of AI regulations in India by examining perspectives across government, industry & civil society stakeholders and the significance of sustainability in the era of artificial intelligence. In the recent few years, a lot of ethical guidelines have been released. These principles and the proposed recommendations to exercise the ‘disruptive’ ability of the emerging artificial intelligence technology aim to be included. This paper explores various guidelines, and further highlights the various overlaps. The paper aims to provide a detailed framework of the ethics of AI and India’s advancement on AI regulations & also examines how these principles & values are implemented in the application & development of the AI systems.
 
Keywords: Artificial Intelligence, Ethics, Guidelines, Implementation, Machine Learning.
 
 
 
Introduction
The emergence of AI technology has altered sectors worldwide, resulting in extraordinary achievements. Along with these benefits, ethical quandaries have arisen, demanding sophisticated regulatory structures. India, as a booming technology hub, faces the difficult issue of reconciling AI innovation with ethical concerns. Artificial intelligence (AI) has rapidly expanded from a niche technological concept to a pervasive presence that influences many aspects of modern life. Its integration covers across industries, including healthcare, finance, education, and entertainment, providing unparalleled efficiencies and new solutions to complicated challenges. Artificial intelligence systems, for example, have helped with early disease detection, individualized learning experiences, and supply chain logistics optimization. However, in addition to these gains, AI presents a number of ethical problems that must be carefully considered and managed. One major ethical problem is the ability of AI systems to propagate bias and discrimination. When taught on historical data tainted with societal prejudices, AI can unintentionally reinforce existing disparities, resulting in unfair treatment of specific populations. This issue is especially concerning in crucial sectors such as hiring processes, loan approvals, and criminal justice, where biased AI conclusions can have serious consequences for people's lives. Privacy & monitoring are another essential ethical factor. The massive data collecting inherent in AI functionalities raises concerns about individuals control over their personal information and the possibility of intrusive monitoring. The balance between using data for technical innovation and protecting individual privacy is tricky and sometimes riddled with complications. Accountability and transparency in AI decision-making procedures additionally constitute important ethical considerations. The "black box" character of many AI models implies that their internal operations are difficult to decipher, even by their authors. This opacity can make it difficult to assign blame when AI systems break or create unforeseen outcomes, weakening confidence in AI applications.
 
Concerning these multiple ethical problems, nations around the world are attempting to build proper legislative frameworks to oversee AI creation implementation. India, a developing technical superpower, is actively participating in this discussion. The country acknowledges the need to promote AI innovation while also implementing measures to address ethical concerns. This requires a comprehensive strategy that takes into account India's particular socioeconomic situation, a variety of cultures, and development aspirations.
 
 
This study will investigate the ethical concerns of AI and critically analyze India's progress in developing regulatory systems to address these challenges. By investigating both ethical quandaries & legislative remedies, we hope to provide a thorough overview of the existing situation & prospects for the governance of artificial intelligence in India.
 
2. Ethical Implications of AI
2.1 Bias & Discrimination
Artificial Intelligence systems trained on skewed data might reinforce existing preconceptions, resulting in discriminating decisions. For example, AI algorithms employed in employment procedures may prefer particular demographics over others, aggravating socioeconomic disparities. An AI mechanism, especially those built on machine learning, depends significantly upon data to function properly. Nevertheless, when this data is derived through historical records as well as human-generated inputs, it may contain unintentional biases reflecting societal disparities. This problem is especially prevalent in sensitive industries such as recruiting, law enforcement, medical care, & financial services. A major instance of AI prejudice was seen in hiring algorithms utilized by large organizations, where the system favored male applicants over female prospects. This happened because the artificial intelligence (AI) system was trained upon previous hiring data that primarily showed male employees, causing it to exacerbate existing gender inequities rather than erase them. Likewise, the technology of the facial recognition is being criticized for exhibiting racial biases, with research indicating that it misidentifies people with darker complexions at a considerably higher rate compared to those with lighter complexions.
 
Bias in AI-powered systems is particularly significant in India due to the country's complex social fabric, including discrepancies due to gender, caste, economic status, & linguistic variations. If AI algorithms are utilized in public welfare activities, banking, and legal systems are not thoroughly tested for bias, they may adversely impact marginalized people. For example, an AI-powered credit evaluation system that evaluates the eligibility of loans based on previous financial records may erroneously reject those from lower socioeconomic strata who have historically had limited access to official banking services.
 
In addition, there is rising worry about AI prejudice in law enforcement. Predictive law enforcement algorithms, which examine crime data to forecast probable criminal conduct, might exacerbate systemic inequality by disproportionately targeting certain communities. If previous crime data is biased towards certain ethnic or socioeconomic classes, AI models would keep highlighting these areas as high-risk regions, continuing the process of discrimination instead of addressing the underlying causes of crime. To address these concerns, it is critical to use fairness-aware models for machine learning, increase dataset variety, and incorporate strong auditing methods to detect and correct prejudices in AI systems. Furthermore, regulatory actions must require openness regarding AI decision-making processes; so that prejudiced results can be discovered & changed before they produce harm.
 
2.2 Privacy & Surveillance
The use of AI in data gathering & processing carries major privacy implications. Unauthorized access & exploitation of private data can lead to intrusive surveillance, weakening individual rights. Privacy constitutes one of the most significant ethical issues around AI, as the technology relies largely on gathering, analyzing, and processing massive amounts of personal information. AI-powered systems continuously monitor user conduct, preferences & actions in a variety of contexts, including social media platforms, e-commerce sites, healthcare applications, and government databases. While data-driven approaches improve user experiences and allow for personalized services, they also raise significant worries about the privacy of individuals, potential misuse & data security.
 
Data Collection & Consent
One of the most important privacy issues raised by artificial intelligence is the absence of qualified consent whenever data has been acquired. Many AI-powered systems collect user data without their explicit approval or rely upon complex & vague terms of service that customers sometimes do not fully comprehend. As a result, folks inadvertently provide personal data to corporations & governments, potentially exposing themselves to privacy violations. For instance, AI-powered recommendation algorithms on e-commerce & streaming sites personalize content based on user actions, search history, & preferences. However, users are frequently ignorant of the scope of data gathering and the way their information is maintained and shared. In India, the lack of thorough rules of the data protection will heighten concerns about how AI corporations handle personal information.
 

Mass Surveillance & Facial Recognition

Governments around the world, including India, are increasingly using AI-powered surveillance tools for law enforcement, safety of the public, national security. One of the more contentious applications is FRT[1], which is used for identifying people in public places, airports, and at border crossings. In India, the government is extending the use of AI-based surveillance technologies through initiatives like the National Automated Facial Recognition System (AFRS). While these technologies are intended to increase security and aid during criminal investigations, they additionally raise worries about mass surveillance, invasion of privacy, and the possibility of abuse by authorities. Without defined regulatory structures, facial recognition technology may be used to monitor persons without their consent, resulting in a breach of fundamental rights. Numerous researcheshave shown that the problems connected with facial recognition, such as mistakes in recognizing specific demographics, notably those from minority groups. This has resulted in more erroneous arrests & discrimination in policing. In the absence of strict monitoring, AI-powered surveillance may disproportionately affect underprivileged people, violating their freedoms and rights.
 
Data Security & AI-powered Cyber Threats
AI has additionally highlighted new vulnerabilities to cybersecurity. With an increased reliance on AI based data analysis, organizations and governments must protect sensitive information from cyber threats. AI-powered cyber attacks, that include deep fake frauds, phishing assaults, & automated hacking, have advanced, posing major hazards to individuals as well as businesses. Concerns about data security in India have grown in response to several incidents of massive data breaches.AI-powered systems, such as financial products & services, smart city initiative & digital healthcare records require strong security protocols to prevent illegal access and data leakage. The proposed Digital Personal Data Protection (DPDP) Bill, which is projected to be the nation's first comprehensive data protection law, aims to deal with some of these issues by governing how corporations gather, keep, and handle user information. However, its execution & effectiveness yet to be determined.
 
2.3 Accountability & Transparency
AI systems' decisions-making procedures are generally opaque, rendering it difficult to assign accountability for errors or unforeseen consequences. The lack of transparency poses questions about accountability for AI-driven behaviors. As artificial intelligence (AI) becomes more integrated into how decisions are made across businesses, the lack of accountability and transparency in AI-powered systems has become a serious ethical problem.AI models, specifically the ones based on deep learning, are frequently used as "black boxes," which means that humans cannot simply interpret their decision-making processes. The lack of explain ability complicates in auditing, understanding & ultimately holding AI systems responsible for their activities.
 
Challenges in AI Transparency
One among the primary concerns about AI transparency involves the fact that many powerful machine learning algorithms does not provide specific justifications for their results. AI models learn trends from data and generate predictions on the basis of complicated statistical relationships, as opposed to traditional software, which requires explicit programming of rules & logic. This transparency makes it challenging to see:
 
Identify errors or biases: Whenever the AI system made a mistake like an incorrect medical diagnosis or a wrongful loan rejection stakeholder& users generally struggle to know about whythe system has failed.
?       Ensure fairness: AI algorithms that affect key decisions, like recruiting, law enforcement, or credit authorizations, may unintentionally favor one group over another. Without being transparent, it would be difficult to identify & overcome these prejudices.
?       Building trust: Individuals who are affected by AI-driven judgments may be hesitant to put their faith in the technology since they don't comprehend the procedure for making decisions. For instance, in India, AI is rapidly being employed in court decision-making to help courts for analyzing the legal matters. However, if judges & legal experts don't completely comprehend how an AI system makes its recommendations, the likelihood of biased verdicts grows
 

Accountability in AI Decision-Making

Accountability means the ability to delegate accountability whenever AI systems inflict harm or failed to perform properly. This is especially important in industries wherever AI impacts high-stakes selections such as healthcare, banking, as well as autonomous mobility.
 

Who is Responsible When AI Fails?

One of the biggest challenges in AI accountability is determining who should be held accountable when an AI system produces harmful or unintended consequences. Possible entities that could be held accountable include:
1.      Developers and Data Scientists: If an AI model is trained on biased or flawed data, the responsibility may lie with the developers who built it.
2.      Organizations Using AI: Companies and government agencies deploying AI systems must ensure proper oversight and human intervention where necessary.
3.      Regulatory Bodies: Policymakers must create clear guidelines to define accountability for AI failures, ensuring legal protections for affected individuals.
For instance, if an AI-powered recruitment system rejects a qualified candidate due to biased training data, should the responsibility fall on the HR team using the system, the software developers, or the AI model itself? This ambiguity highlights the urgent need for accountability mechanisms.
 
The Need for Explainable AI (XAI)
To address transparency concerns, researchers & policymakers are advocating for Explainable AI (XAI) a set of frameworks& techniques designed to make AI decision-making more understandable to humans. The goal of XAI is to ensure that AI systems can provide justifications for their predictions, allowing stakeholders to assess the fairness & reliability of theirdecisions.
Some of the key approaches to enhancing AI transparency include:
?       Interpretable Machine Learning Models: Using easier models, like decision trees, where the reasoning behind assumptions is more understandable.
?       Feature Importance Analysis: Identifying which data points better influenced an AI model’s decision.
?       Audit Trails & Documentation: Maintaining expended records of AI decision-making processes for compliance& accountability.
For example, in India, AI is increasingly being used in financial services to assess credit risk & detect fraud. If a bank’s AI model denies a loan application, it’d be able to explain the reasoning- whether it was due to income level, credit history, or other factors so that applicants understand & can contest unfair decisions.
 
2.4 Job Displacement
Automation through AI threatens employment in various sectors, leading to job displacement & economic disparities. The substitution of human labor with machines necessitates discussions on the future of work & income distribution.The increasingly adoption of Artificial Intelligence (AI) around industries has sparked discussions regarding its impact on employment & economic structures. While AI brings significant benefits in terms of efficiency, innovation & automation. It too raises concerns about job displacement, wage polarization, & widening inequality in economic standards. As AI-driven automation substituted traditional jobs, societies must address these challenges of workforce adaptation & the potential risks involved regarding the economic disparity.
 
Impact of AI on Employment
AI & automation technologies are molding the labor market by substituting the repetitive, and administrative& manual tasks with machine-driven processes. Several sectors are witnessing this shift:
?       Manufacturing: AI-powered robotics & automated assembly lines hv’been significantly reduced the basic requirement for human labor in factories. Tasks that alreadyrequired well skilled workers, like welding, quality control& Packaging, are now operated by AI-driven machines.
?       Retail &Customer Service: Chat bots, automated checkout systems& virtual assistant are substituting human customer service representatives& cashiers, in order to reduce the number of employment at the entry-level.
?       Finance &Banking: AI algorithmis used for fraud detection, automated trading &risk assessment, in order to reduce the need for traditional financial analysts &clerical staff.
?       Healthcare: AI-driven diagnostics, virtual health assistants &robotic surgeries are changing the aerana medical services have delivered, reduced the demand for certain specific healthcare employment while rising the need of AI-trained experts.
?       Transportation: Autonomous vehicles & AI-powered logistics systems is emerging as a threat to the employment of trucking, public transportation &delivery services.
While AI enhances operational efficiency &productivity, the replacement of workers particularly those in repetitive jobs & low-skilled raises concerns regarding social instability & unemployment.
 
Job Polarization & the Skills Gap
The excessive use of AI has created a polarized labor market, in which the jobs at both the low-skilled ends& high skilled ends are rising, whereas the middle-skilled employment is declining. This concept is socalledjob polarization & is categorized by:
?       Growth in High-Skilled Unemployment: The demand for data scientists has been raised by AI, AI engineers, machine learning specialists &software developers. Such jobs demandadvanced expertise in the technical field &initiatingprospects for highly learned professionals.
?       Deterioration in Middle-Skilled Unemployment: There are a lot of traditional white-collar occupations, like administrative assistants, telemarketers &accountants,are at menace of automation& reducing openings for employees in these characters.
?       Development in Low-Skilled Service Unemployment: While AI is programmingnumerous jobs, it also generates demand for roles theneeded human communication, likecare giving, personal services &hospitality. However, suchemployments often pose lower wages & limited career development.
The change towards AI-driven wealth highlights a risingskills gap- a mismatch concerning the skills workers presently possess & the knowledge required in AI-centric businesses. Workforceswho lackdigital literacy & technical skills might struggle to find jobs, exacerbating economic disparity.
 
3. India's Advancement in AI Regulations
3.1 Current Regulatory Landscape
As of now, India is having no specific legislation which exclusively governs AI development &usage. The Ministry of Electronics &Information Technology (MeitY) acts as the principal agency administering AI-related policies, having recognized committees to draft the policy agendas relating to it. As Artificial Intelligence (AI) remains to transform industries &economies at the global level, India is being recognized the requirement for a structured regulatory draft to govern AI expansion, ethical consideration &deployment. However, the country has been taking the initiative in leveraging AI for the innovation & the economic growth, it had also recognized the requirement of responsible AI strategies to address issues like privacy, bias, accountability & job displacement.
 
The approach of the Indian Government regarding AI regulation is being evolved through numerous government initiatives, sector specific principles &policy reports which are aiming at the fostering invention while guaranteeing ethical acquiescence.
 
Initiatives of the Government on AI Regulation
There are several key governmental bodies in India who has actively involved in AI policy formulation & regulation, comprising the NITI Aayog, the Ministry of Electronics & Information Technology (MeitY), & the Reserve Bank of India (RBI).
 
1. Ministry of Electronics & Information Technology (MeitY) & AI Governance
The Ministry of Electronics &Information Technology (MeitY) has been proactively working on AI policies, mainly concerning data protection, AI ethics &cybersecurity. Some of MeitY’s crucial initiatives incorporate:
?       Artificial Intelligence Advisory Council: It is a task force which has been set up to stipulatesuggestions on AI governance, industry best practices & ethical AI deployment.
?       AI-Specific Standards & Guidelines: MeitY is working with industrial bodies to describe technical specifications for AI systems to ensure transparency, security &interoperability.
?       AI Mission of India: An anticipated initiative which aimed at developing the AI infrastructure at the national level, comprising AI compute resources, regulatory sandboxes for AI startups &open data frameworks.
 
2. Digital Personal Data Protection (DPDP) Act, 2023
One of the utmost significant regulatory developments touching AI in India is the Digital Personal Data Protection (DPDP) Act, 2023.[2]While mainly engrossed on data privacy, this legislation has lot of other implications for the governance of AI as AI systems depend profoundly on data for decision-making& training.
The following are the key provisions of the DPDP Act which are relevant to AI:
?       The User’s Consent for Data Processing: AI systems must get unambiguous consent from individuals before processing or collecting the personal data of him.
?       Right to Data Erasure& Correction: Individuals have the right to request the corrections & deletions of their data which have been used in AI models.
?       Limitations on Cross-Border Data Transfers: This Act controls how AI companies keep & process data crosswise international borders, affecting global AI relationships.
Even though the DPDP Act doesn’t directly control AI algorithms, it institutes ground for ethical AI procedures by enforcing privacy & the responsible data handling.
 
3. RBI & AI Regulations in the Financial Sector
After analyzing the use of AI in financial services at a broader level: -
The Reserve Bank of India (RBI) has laid down certain guidelines to control AI-driven financial applications.
The Chief regulatory methods incorporate:
?       Fair &Transparent AI Lending Practices: AI-powered loan authorization techniques need to be explainable & free from bias to prevent discriminatory lending methods.
?       AI-Driven Fraud Detection& Risk Management: AI applications in banking must beful filled with fraud prevention& cyber security norms.
?       Protection of Data in AI-Based Financial Services: Financial organizations which are using AI for credit scoring& customer profiling must follow data protection regulations.
RBI's proactive methodology guarantees that AI in the financial field stays secure, accountable & transparent.
 
3.2  National Strategy for Artificial Intelligence
In 2018, NITI Aayog released the "National Strategy for Artificial Intelligence," highlighting the requirement for data protection & sector-specific regulatory outlines. The document stipulates for the adoption of international ideals to guarantee ethical AI deployment.
 
1. NITI Aayog’s National Strategy for AI (2018)
NITI Aayog, India's premier policy think tank, released the National Strategy for AI in 2018, delineating India’s tactic to AI governance& development. The approach emphases leveraging AI in important sectors like healthcare, education, agriculture, finance & smart cities, while advocating regulatory& ethical challenges.
The important highlights of the approach incorporate:
?       AI for All: A vision to make AI accessible &inclusive for all inhabitants, guaranteeing equitable delivery of AI assistances.
?       Responsible AI Development: Emphasis on the ethical principles of AI, including transparency, fairness &accountability.
?       Data Protection &Privacy: Commendations for developing AI structures which adhere to security standards& privacy.
?       Research &Innovation: Creation of Centers of Excellence (COEs) to encourage AI research & development in the country.
Although the strategy stipulates a broad visualization for AI implementation, it doesn’t recommend a devoted legal outline for AI governance, leaving scope for additional regulatory growths.
 
2. Establishment of Regulatory Bodies
Debates are ongoing to formed voted regulatory authorities for over sighting of AI. The Ministry of Commerce & Industry had recognized an 'Artificial Intelligence Task Force'for exploring the formation of such authorities, pointing to guarantee ethical AI development & application.
 
Conclusion
The ethical implications of AI requires a robust regulatory framework to reduce the potential risks. India's proactive approaches, including strategic initiatives & proposed legislation, reflects a commitment to addressing these challenges. As AI continues to develop, continuous evaluation & adoption of regulatory methods will play a crucial role in harmonizing ethical standards &technological innovation. The massive integration of Artificial Intelligence (AI) into different sectors places both significant challenges of ethical concerns & transformative opportunities. Addressing concerns like bias, accountability, privacy& its impact on human innovation requires a comprehensive regulatory modal that balances creativity with ethical considerations.
 
India's way of approach to AI regulation highlights a nuanced understanding of such complexities. Initiatives like the proposed Digital India Act (DIA) aim to provide the robust legal basis for emerging technologies, including AI, by providing high-risk systems & delineating specific "no-go" ways for companies who are involved in utilizing the AI in consumer-facing applications.
 
The launch of the India’s AI Mission, with the allocation of more than $1.2 billion, undervalues the country's target in advancing its AI field. This project focuses on securing necessary infrastructure, aiding local startups, ensuring high-quality datasets, expanding AI education, & developing the state-of-the-art AI models, placing India as a significant entity in the global AI world.
 
In Collaborations with the international organizations, like UNESCO, to develop AI policies aligned with the global ethical standards & demonstrate India's dedication to the adoption of responsible AI. The exploration of establishing an Artificial Intelligence Safety Institute (AISI) further deals with the focus on formulating guidelines & standards which is ensuring the development of ethical AI without any stifling innovation.
 
However, challenges persist, particularly relating to data privacy, biasness, workforce replacement &accountability. The comprehensive data protection framework provides the way to reduce the citizen’s privacy risks & also advocating the need of the strict policies for safeguarding the individual interest and rights.
 
In conclusion, India's proactive approach in AI development & regulation signifies a systematic effort for harnessing the AI's potential responsibly. By advocating the fostering innovations &ethical considerations, India aims to provide that AI technologies contribute positively in order to develop the society, collaborating with the country’s broader agenda of technological advancement & inclusive growth.
 
REFERENCES: -
Books
1.      Binns, R."Ethics in Artificial Intelligence: A Framework for Thinking about the Ethical Challenges of AI"(2018)
2.      Tegmark, M. (2017), "Life 3.0: Being Human in the Age of Artificial Intelligence", Alfred A. Knopf.
3.      Russell, S., Dewey, D., & Tegmark, M. (2015), "Research Priorities for Robust and Beneficial Artificial Intelligence", AI & Society, 30(4), 99-106.
Journal Articles
1.      Chakraborty, A., & Paliwal, P. (2020),"Artificial Intelligence in India: A Case for Ethical Regulations"International Journal of Law and Information Technology, 28(3), 225-245.
2.      Binns, R. (2018), "The Ethics of Artificial Intelligence and Robotics: A Systematic Review of the Literature", AI & Society, 33(3), 285-306.
3.      Kshetri, N. (2019), "Artificial Intelligence, Blockchain, and the Regulation of AI in India", Technology in Society, 58, 101143.
4.      Gupta, R., & Chaturvedi, A. (2021), "Ethics of Artificial Intelligence: Regulatory Challenges and Global Responses",International Journal of Technology and Human Interaction, 17(1), 12-27.
5.      Nandakumar, M., & Radhakrishnan, P. (2022), "AI Governance and Regulation: Global and Indian Perspectives", Journal of Indian Law and Technology, 14(2), 199-218.
Reports and Articles
1.      Government of India. (2018), “National Strategy for Artificial Intelligence: AIforAll”
NITI Aayog, Government of India.
2.      The Future of Life Institute (2017), "The Ethics of Artificial Intelligence: Opportunities and Threats"AI & Ethics, 1(1), 1-13.
3.      Rajeev, P., & Mehta, K. (2020), "AI Regulation in India: Challenges and Path Forward"Economic and Political Weekly.
4.      Binns, R., & Palmer, R. (2020), "Artificial Intelligence Regulation: The Indian Approach and International Comparison", Law and Technology Review, 12(2), 55-77.
5.      World Economic Forum (2020),"Shaping the Future of Artificial Intelligence in India: Policy and Regulatory Considerations".
Web Articles and Blogs
1.      Iyer, A. (2022),"The Need for Ethical Guidelines in AI Regulation in India"
The Economic Times.
2.      Choudhury, P. (2021),"AI Regulations in India: Progress and Prospects"
India Today.
3.      Bhattacharyya, S. (2023), "AI Governance: India’s Ethical Roadmap for Artificial Intelligence", The Hindu Business Line.
4.      Sharma, K. (2020), "The Ethics of AI: How India is Shaping Its Regulatory Future"
YourStory.
Government and Institutional Frameworks
1.      NITI Aayog (2021),"Artificial Intelligence for All: Regulatory and Ethical Framework"
Government of India, NITI Aayog.
2.      Indian Ministry of Electronics and Information Technology (MeitY), (2023),"Regulating Artificial Intelligence in India: A Path Forward", Government of India.


[1] Facial Recognition Technique
[2] https://pib.gov.in/PressReleasePage.aspx?PRID=2090271

About Journal

International Journal for Legal Research and Analysis

  • Abbreviation IJLRA
  • ISSN 2582-6433
  • Access Open Access
  • License CC 4.0

All research articles published in International Journal for Legal Research and Analysis are open access and available to read, download and share, subject to proper citation of the original work.

Creative Commons

Disclaimer: The opinions expressed in this publication are those of the authors and do not necessarily reflect the views of International Journal for Legal Research and Analysis.