The Telecom Regulatory Authority of India (the “TRAI”) has recently released recommendations on “Leveraging Artificial Intelligence and Big Data in the Telecommunication Sector” and has suggested that the Indian government must adopt a regulatory framework for AI that applies across sectors. In this update, we analyze the significance of the proposed framework and its implications on India’s position on AI.
The impact of artificial intelligence (“AI”) can be seen today across the globe and in various sectors, and has the potential to profoundly influence industries across the spectrum, including finance, healthcare, customer services, etc. The functioning of AI is premised on processing large volumes of data using advanced algorithms to recognize patterns, learn from experiences and make predictions.
The development and deployment of AI technologies is susceptible to various risks, such as perpetuating and even amplifying existing biases, which can lead to unfair or discriminatory outcomes when making decisions or predictions. Additionally, the creation of inaccurate models can generate misleading insights, impacting critical areas such as healthcare diagnoses or financial forecasts. Therefore, the use of AI has sparked discussions on ethics and biases, emphasizing the need for transparent and accountable AI development with policymakers, businesses and technologists playing a crucial role in shaping AI’s path to ensure it aligns with human values.
Jurisdictions across the world have begun drafting policies and regulations and are taking up positions in the AI space. Governments appear to be pivoting towards a policy approach that harnesses and regulates AI, with varying levels of priority.
People’s Republic of China: China will soon become the first country to have an operational AI law, with the Chinese AI regulations (the “PRC Regulations”) scheduled to become operative from August 15, 2023. The provisional draft of the PRC Regulations was published jointly by seven (7) different regulatory bodies. Under the PRC Regulations, the most prominent directive is with respect to the content. AI service providers are prohibited from generating any content that is violent, obscene or against national interest, and AI chatbots are prohibited from generating false and harmful information. AI service providers must also select training data in a way that avoids discrimination. They must also report any users using AI to produce illegal content and restrict minors from becoming addicted to such services.
European Union: As with the General Data Protection Regulation, the European Union (the “EU”) has taken a regulation heavy approach with the EU Artificial Intelligence Act (the “EU Act”) proposing a comprehensive regulatory framework which aims to ensure the responsible development and deployment of AI within the EU. The EU Act proposes a risk-based approach categorizing AI systems as low, high, or very high based on potential harm and level of autonomy. High-risk AI systems, such as those in healthcare and transportation, face strict requirements on data transparency, human oversight and robustness. The EU Act also bans certain AI practices that pose an unacceptable risk to an individual’s rights, such as social credit scoring. It emphasizes transparency by mandating clear labelling of AI-generated content, safeguards against biased algorithms, and establishes a European Artificial Intelligence Board to oversee the implementation. Overall, the focus remains on ethical implications, aimed at fostering trust and accountability in AI technologies across the EU.
United States of America: The US Federal Trade Commission has published a set of five (5) principles which must be followed by companies when using AI for consumer protection. Companies are required to be: (i) be transparent with consumers about their interaction with AI tools; (ii) clearly explain decisions that result from AI; (iii) ensure that decisions are fair; (iv) ensure that the data and models being used are robust and empirically sound; and (v) hold themselves accountable for compliance, ethics, fairness and non-discrimination. An AI Bill of Rights has also been proposed, which is a non-binding roadmap for the responsible use of AI.
United Kingdom: The UK has taken a “pro-innovation” approach to AI regulation. The UK’s proposed AI framework does not seek to assign rules or risk levels to entire sectors or technologies, but instead regulates them on the basis of outcomes. Instead of hard-set rules, principles such as safety, security, robustness, fairness, accountability, redress, etc., will govern the use of AI, and individual sector regulators with domain-specific expertise will customize the implementation of the foregoing principles to specific contexts.
Other jurisdictions: Saudi Arabia has taken a more relaxed stance on regulation and created a Saudi Data & AI Authority to prepare and execute the national data and AI agenda. Another body known as the National Center for Artificial Intelligence has been set up to promote research, development and advancement in AI. The approach focuses on innovation rather than regulation. Australia has proposed a set of eight (8) “AI Ethics Principles” which organizations may voluntarily consider. Singapore is also in the process of creating a voluntary model framework which provides implementable guidance to private sector organizations to address key ethical and governance issues. Further, a national AI strategy has been unveiled by Canada, with an emphasis on talent development, research, and industry cooperation. Japan’s Society 5.0 initiative seeks to incorporate AI into every element to create a human-centred society.
Therefore, differing approaches are being taken towards AI and its regulation, with some countries providing a high degree of freedom for innovation and others seeking to actively regulate AI.
India is also emerging as a key player in the AI regulation space, and the path taken by India must be closely monitored due to the impact it may have on the AI space.
The AI landscape in India
India seems to be taking a mixed approach to AI and is trying to balance between innovation and regulation. On the innovation agenda, India has taken steps to encourage the responsible and inclusive development of AI for economic progress. The National Strategy on AI document of 2018, released by India’s NITI Aayog, covers training, research and development, centres of excellence, data accessibility, and high computing infrastructure. India’s objective for social transformation through responsible AI was the focus of the Responsible AI for Social Empowerment (RAISE) event in 2020, which attracted participants from academia, business, and governments worldwide. India has partnered with other developed nations like the United States of America, the United Kingdom, the European Union, Australia, Canada, France, Germany, Italy, Japan, Mexico, New Zealand, Republic of Korea, and Singapore as founding members of the Global Partnership on Artificial Intelligence (the “GPAI”). GPAI is a global initiative supporting ethical AI development based on inclusivity, diversity, creativity, and economic prosperity. Additionally, the Indian government has also established the “National Artificial Intelligence Portal,” a national hub for AI-based projects.
On the regulatory front, the TRAI has released recommendations on July 20, 2023 (the “TRAI Recommendations”) that cover a diverse range of topics related to leveraging big data and artificial intelligence in the telecommunication sector. AI, its definition, new threats, moral standards, and the necessity for regulation have been discussed, along with limitations on AI adoption and the requirement for a data governance structure. The TRAI Recommendations also examine global perspectives and advice on the regulatory approach for India. The focus of the TRAI Recommendations extends beyond the telecommunication industry, and the TRAI acknowledges that AI will have an impact across many sectors, including healthcare, banking, transportation, education, and agriculture. Due to the widespread influence of AI across industries, including telecommunication, a standard framework for all industries has been suggested (a more detailed analysis of the recommendations is affixed as an Annexure).
This strategy aligns with the dynamic nature of AI technology and its potential effects on people, businesses, and society. Adopting such a uniform framework will promote ethical AI practices and will guarantee uniform standards across many industries, while promoting innovation and advancement in the AI field. The government’s readiness to implement this idea and coordinate initiatives across many businesses to address AI’s difficulties and prospects successfully will, however, determine the proposal’s ultimate success.
One of the key recommendations made by the TRAI is to set up an Artificial Intelligence and Data Authority of India, (more on this in the Annexure), which will oversee the creation of responsible AI and control its applications in India. Given the dynamic nature of AI and its imminent exponential growth, it will be paramount that the regulatory authority is agile, responsive and fast moving. In this regard, while the recommendation to set up a dedicated body is well-intentioned, such a body will need to show a high level of efficiency in comparison to some of the other regulators in India. India, with its young average population, is adding millions to its workforce every year, and therefore, the Indian government’s AI policy may need to incorporate certain socialist principles to ensure job safety, which may lead to interesting differences in the policy approach of the Indian government towards AI vis-a-vis that of other developed nations.
To what extent the TRAI Recommendations will be implemented, and the ultimate approach taken by India remains to be seen. However, the balanced path that has been proposed augurs well for the future of the AI industry in India.
The TRAI Recommendations propose a comprehensive framework for the regulation of AI in India. Its salient features are listed below.
Recommendation on the regulatory framework
The recommendations provide that a comprehensive regulatory framework applicable across all sectors must be created, which is essential to foster responsible AI development in India. This framework should prioritize risk-based regulation, where high-risk AI use cases directly affecting humans are subject to legally binding obligations.
Three crucial elements make up the regulatory framework that the TRAI has suggested. The first is the establishment of an independent statutory body called the “Artificial Intelligence and Data Authority of India” (the “AIDAI”) to oversee responsible AI development and regulate its usage in the country. The second is the creation of a unique “Multi-Stakeholder Body” (the “MSB”) which will support the AIDAI in an advisory capacity to ensure thorough and competent decision-making. Diverse persons from business, academia, and other organizations will be represented on the MSB. The final element is the categorization of AI use-cases according to their dangers and governance by responsible AI principles. This strategy promotes the appropriate development and application of AI for the well-being of society, by striking a balance between innovation and risk mitigation.
The data authority
The TRAI has also referred to its “Regulatory Framework for Promoting Data Economy” recommendations of November 18, 2022, in which it had suggested the creation of the Data Digitization and Monetization Council (the “DDMC”). The DDMC was going to be in charge of tasks relating to the broad area of data, including data digitization, AI and associated technology, and was going to research on how developing technologies may affect data ethics. However, in light of the focus being placed on AI, the TRAI Recommendations now provide that the previously proposed DDMC be renamed AIDAI to reflect this. The TRAI Recommendations also state that as the creation of too many statutory bodies may cause confusion, the duties of the AIDAI be taken up by the TRAI by making suitable amendments to the TRAI Act, 1997, in order to promote synergy and a harmonious functioning.
The TRAI has recommended that the government should create the MSB to serve as an advisory body to the AIDAI. The members of the MSB will be appointed from various government agencies, such as the Ministry of Home Affairs, the Ministry of Electronics and Information Technology (the “MeitY”), the Department for Promotion of Industry and Internal Trade, and the Ministry of Telecommunications, Information and Broadcasting. Additionally, it will include representatives from academic and research institutions, four members from industry, one cybersecurity expert, one AI legal expert, and up to two other suitable individuals. When necessary, the MSB may issue special invitations to representatives of pertinent Ministries/Departments of the Central/State Governments.
Cross-border impact of AI
Given that AI has a worldwide influence cutting across national boundaries, the TRAI has recommended that the Indian government work with other nations and international organizations to create a global agency. This agency is envisaged as the leading worldwide entity for AI’s advancement, harmonization, and responsible application. India is urged to take the lead in establishing international AI standards and governance frameworks.
Privacy, security and ownership of the data in the telecommunication sector
The recommendations cover a range of topics related to data protection in the telecommunication industry. The TRAI reiterates its earlier suggestions from July 2018 regarding “Privacy, Security, and Ownership of Data in the Telecom Sector”. The TRAI suggests that one of the AIDAI’s responsibilities should be to advise the government on a uniform framework for sharing data held by the government and its entities with businesses, educational institutions, research and development centres, startups, etc., per pertinent privacy laws and policies.
Recommendations pertaining to the effective use of AI
- The Department of Telecommunication (the “DoT”) and Access Service Providers must work together, to improve data analysis capabilities utilizing AI and Machine Learning (“AI/ML”) to compare the overall number of mobile connections to predetermined restrictions.
- AI/ML-based tools must be implemented in the DoT Telecom Security Operation Centre to produce security alerts and reduce network security risks.
- AI/ML and new technologies must be effectively used to resolve telecom consumers’ complaints by sending data from service providers’ networks to the Department of Transportation’s Public Grievance Unit.
- The growth of AI must be promoted in a range of communication and broadcasting-related industries in cooperation with stakeholders.
Centre of Excellence for Artificial Intelligence
A Centre of Excellence for Artificial Intelligence (“CoE-AI”) should be established in each state and union territory to assist businesses, startups, researchers, and other organizations create and display technological capabilities. These facilities should provide high bandwidth, powerful computing power, and data sets for training AI models. Additionally, they must work with the anticipated 5G/6G labs to exchange resources and information. In order to undertake research and develop cutting-edge applications in industries including agriculture, healthcare, education, smart cities, and smart mobility, CoE-AIs should promote collaborations between big business, startups, and academia. Further, one CoE-AI should provide a common portal to allow all CoE-AIs to exchange data about facilities that are accessible, access policies, costs, and other pertinent information.
Committee to study and recommend aspects related to AI
The TRAI Recommendations provide that the DoT should work with the the MeitY and the Ministry of Education (the “MoE”) to form a group that includes representatives from these organizations and the Ministry of Skill Development, All India Council for Technical Education (“AICTE”), and the business community. The committee should periodically research and provide recommendations regarding the following AI-related topics:
- Analyse India’s need for qualified workers and AI/ML professionals, both at present and in the future.
- Suggest technical programs to boost the number of such individuals and re-train the current workforce for AI-based prospects.
- Create a framework for collaboration between the business world and academia to create and upgrade AI-related curricula that match the necessary skill requirements.
- Create a system for team research projects and establish internship and business training opportunities.
Training and education on the basic concepts of AI
- The TRAI has suggested that the DoT should collaborate with AICTE to make a course on the ethical use of AI mandatory for students in all technical institutes.
- Additionally, the DoT should work with the MoE to introduce similar courses or modules on ethical AI in non-technical institutes and schools, starting from primary education levels.
- The TRAI has also recommended the DoT to quickly develop appropriate courses on the Ethical Use of AI and Basic Concepts of AI for government officers and employees through the National Telecommunications Institute for Policy Research, Innovations & Training (the “NTIPRIT”). These programmes should be available through Mission Karma Yogi.
- Furthermore, the DoT and the Department of Personnel and Training (the “DoPT”) should work together to make these courses a requirement for all government officers and staff as part of their training program.
Digital Communication Innovation Square Scheme and AI
The Authority recommends that the DoT use the “Digital Communication Innovation Square” scheme to help startups and organizations run AI/ML events, like challenge and bounty programmes, to present their ideas, work with stakeholders, and improve their solutions/products. Additionally, the DoT has been advised to use the Inter-Ministerial Working Group to coordinate projects related to other sectoral ministries and departments.
About the Authors: