The world of artificial intelligence is rife with growing concerns about the release and use of AI models such as OpenAI’s GPT-4. The Center for AI and Digital Policy (CAIDP) has filed a complaint with the Federal Trade Commission (FTC) against OpenAI’s release of the GPT-4 model. The CAIDP believes that this release undermines transparent and trustworthy AI standards and violates Section 5 of the FTC Act.
Furthermore, CAIDP has called for a variety of regulatory measures, including independent assessments of GPT AI and the establishment of baseline standards for the generative AI market. Civil society groups are also calling for authorities to protect people from the potential threats posed by OpenAI’s GPT and ChatGPT AI models.
The concerns over OpenAI’s GPT models are not limited to the United States. European regulators are being urged to investigate ChatGPT due to concerns about biases and information fabrication. The CAIDP is advocating for independent assessments and easier incident reporting mechanisms for interactions with GPT-4.
In response to these concerns, EU lawmakers are planning to regulate the AI industry through an Artificial Intelligence Act. However, some proposals may be outdated and may not address the specific risks posed by emerging technologies.
It’s not just advocacy groups that are sounding the alarm about GPT-4. Prominent figures such as Elon Musk and Steve Wozniak have also voiced support for CAIDP’s criticism of OpenAI and its GPT models. These figures have highlighted the potential dangers posed by AI models and the need for strict regulation to protect consumers.
Overall, the concerns raised by CAIDP and other civil society groups highlight the urgent need for regulatory measures to protect consumers from potential threats posed by AI models. Given the growing risks, there is a pressing need for independent assessment and monitoring of AI models, as well as the establishment of clear standards and incident reporting mechanisms.