Logical Reasoning Questions for CLAT | QB Set 19

Even though Generative AI (GAI) stands as a transformative force, wielding power to revolutionise society in ground-breaking ways, existing legal frameworks and judicial precedents that have been designed for a pre-AI world may struggle to effectively govern this rapidly-evolving technology.

Safe harbour and liability fixation

One of the most persistent and contentious issues in Internet governance has been the fixing of liability on “intermediaries” for content hosted by them. The landmark Shreya Singhal judgment addressed this by upholding Section 79 of the IT Act which grants intermediaries ‘safe harbour’ protection against hosting content, contingent upon meeting the due diligence requirements outlined in Section 3(1)(b) of the Information Technology (Intermediaries Guidelines) Rules. However, its application to Generative AI tools remains challenging.

There are contrasting views on the role of GAI tools. Some argue that they should be considered intermediaries since they are used almost like a search engine even though they do not host links to third-party websites. Others argue that they are mere “conduits” for user prompts, where altering the prompt leads to changes in output – essentially making the generated content akin to third-party speech, and, therefore, attracting lesser liability for the content generated.

In Christian Louboutin Sas vs Nakul Bajaj and Ors (2018), the Delhi High Court held that safe harbour protection applies solely to “passive” intermediaries, referring to entities functioning as mere conduits or passive transmitters of information. However, in the context of Large Language Models (LLMs), making a distinction between user-generated and platform-generated content is increasingly challenging. Additionally, liability in the case of AI chatbots arises once the information is reposted on other platforms by the user; mere response to a user prompt is not considered dissemination.

Generative AI outputs have already led to legal conflicts in various jurisdictions. In June 2023, a radio host in the United States filed a lawsuit against Open AI, alleging that Chat GPT had defamed them. The ambiguity in classifying GAI tools, whether as intermediaries, conduits, or active creators, will complicate the ability of courts to assign liability, particularly in user reposts.

The copyright conundrum

Section 16 of Indian Copyright Act 1957 specifically provides that “no person” shall be entitled to protection of copyright except by the provisions of the Act. As in India, reluctance persists regarding the provisions of copyright protection to works generated by AI globally.

The critical questions are: should existing copyright provisions be revised to accommodate AI? If AI-generated works gain protection, would co-authorship with a human be mandatory? Should recognition extend to the user, the programmer itself, and by extension, the programme, or both? The 161st Parliamentary Standing Committee Report found that the Copyright Act of 1957 is “not well equipped to facilitate authorship and ownership by Artificial Intelligence”.

Under current Indian law, a copyright owner can take legal action against anyone who infringes on his/her work with remedies such as injunctions and damages. However, the question of who is responsible for copyright infringement by AI tools remains unclear. As previously argued, classifying GAI tools, whether as intermediaries, conduits, or active creators, will complicate the courts’ ability to assign liability. ChatGPT’s “Terms of Use” attempt to shift liability to the user for the illegal output. But the enforceability of such terms in India is uncertain.

The landmark K.S. Puttaswamy judgment (2017) by the Supreme Court of India established a strong foundation for privacy jurisprudence in the country, leading to the enactment of the Digital Personal Data Protection Act, 2023 (DPDP). While traditional data aggregators or consent managers raise privacy concerns during the collection and distribution of personal information, Generative AI introduces a new layer of complexity.

The DPDP Act introduces the “right to erasure” as well as “right to be forgotten”. However, once a GAI model is trained on a dataset, it cannot truly “unlearn” the information it has already absorbed. This raises a critical question. How can individuals exercise control over their personal information when it is woven into the very fabric of a powerful AI model?

Steps to pursue

First, learning by doing. Consider granting GAI platforms temporary immunity from liability following a sandbox approach. This approach allows responsible development while gathering data to identify legal issues that could inform future laws and regulations.

Second, data rights and responsibilities. The process of data acquisition for GAI training requires an overhaul. Developers must prioritise legal compliance by ensuring proper licensing and compensation for the intellectual property used in training models. Solutions could include revenue-sharing or licensing agreements with data owners.

Third, licensing challenges. Licensing data for GAI is complex as web-data lacks a centralised licensing body similar to copyright societies in the music industry. A potential solution is the creation of centralised platforms, akin to stock photo websites such as Getty Images, which simplify licensing, streamline access to necessary data for developers and ensure data integrity against historical bias and discrimination.

The jurisprudence around Generative AI (GAI) is hazy and yet to be evolved. It demands a comprehensive re-evaluation of existing digital jurisprudence. A holistic, government-supported approach and judicious interpretations by the constitutional courts are essential to maximise the benefits of this powerful technology, by safeguarding individual rights and protecting them against unwelcome harm all the while.

Question -1)

Statement: Existing legal frameworks struggle to effectively govern Generative AI due to its rapid evolution.

Question: Which of the following would most weaken the argument presented in the statement above?

A. Legal precedents can be updated through judicial interpretation to accommodate new technologies.
B. Generative AI models are inherently unpredictable and cannot be regulated effectively.
C. Governments worldwide have established stringent regulations specifically targeting AI technologies.
D. Generative AI technology is primarily used in non-sensitive sectors where legal oversight is less critical.

Question -2)

Statement: The ambiguity in classifying Generative AI tools complicates the assignment of liability in legal disputes.

Question: Which of the following scenarios would further complicate the classification of Generative AI tools?

A. A court ruling that Generative AI tools should be treated as intermediaries under existing legal frameworks.
B. Adoption of new international guidelines that clearly define the roles and responsibilities of Generative AI developers.
C. Emergence of a new software update that enables Generative AI tools to independently generate content without user prompts.
D. Establishment of a new regulatory body solely dedicated to overseeing the development and deployment of Generative AI technologies.

Question -3)

Statement: The Indian Copyright Act of 1957 is not equipped to handle authorship and ownership issues arising from AI-generated works.

Question: Which of the following statements, if true, would most strengthen the argument in the statement above?

A. Many countries have already amended their copyright laws to explicitly recognize AI as co-authors of works.
B. The Indian government has proposed a new amendment to the Copyright Act specifically addressing AI-generated content.
C. International treaties have unanimously agreed to exempt AI-generated content from copyright protection.
D. AI-generated content has consistently been deemed as falling under fair use exceptions in copyright law.

Question -4)

Statement: The DPDP Act introduced the “right to be forgotten,” but it poses challenges when applied to Generative AI.

Question: Which of the following statements would most effectively counter the challenges posed by the “right to be forgotten” in relation to Generative AI?

A. Generative AI models can be programmed to automatically erase any personal data upon user request.
B. The DPDP Act does not apply to AI-generated data, thus circumventing the “right to be forgotten.”
C. Courts can issue injunctions to prevent Generative AI models from accessing personal data for future use.
D. Users can avoid sharing personal information with Generative AI models to prevent data retention.

Question -5)

Statement: The classification of Generative AI tools as intermediaries, conduits, or active creators complicates liability assignment.

Question: Which of the following legal principles could potentially resolve the ambiguity in the classification of Generative AI tools?

A. Establishing a new legal category specifically for AI technologies under existing frameworks.
B. Treating Generative AI tools as passive transmitters of information under the existing safe harbor provisions.
C. Requiring AI developers to obtain explicit consent from users for any content generated by the AI.
D. Holding AI developers strictly liable for any content generated by their AI tools, regardless of user input.

Question -6)

Statement: The enforceability of ChatGPT’s “Terms of Use” in India remains uncertain.

Question: Which of the following would most weaken the uncertainty surrounding the enforceability of ChatGPT’s “Terms of Use” in India?

A. Implementation of a new regulatory framework that explicitly recognizes AI-generated content.
B. The Supreme Court of India’s recent ruling explicitly supporting the enforceability of similar terms of use agreements.
C. A government directive requiring all AI platforms to revise their terms of use to comply with Indian law.
D. An international treaty mandating uniform terms of use agreements for all AI platforms worldwide.


Calling all law aspirants!

Are you exhausted from constantly searching for study materials and question banks? Worry not!

With over 15,000 students already engaged, you definitely don't want to be left out.

Become a member of the most vibrant law aspirants community out there!

It’s FREE! Hurry!

Join our WhatsApp Groups (Click Here) and Telegram Channel (Click Here) today, and receive instant notifications.

CLAT Buddy
CLAT Buddy