Hackers have gathered at DEF CON, an annual hacking conference in Las Vegas, for a first-of-its-kind hacking contest. The competition aims to test the security of AI chatbots, revealing potential risks associated with these systems. Over the next four days, more than 3,000 hackers will attempt to infiltrate and potentially compromise leading generative artificial intelligence systems. This event marks the largest-ever public exercise focused on discovering security weaknesses in large language models.
Generative AI systems have gained significant popularity, with the widespread use of tools like ChatGPT. However, these systems are not immune to vulnerabilities, as hackers have already found ways to circumvent their security controls and exploit mainstream models. The red-teaming exercise at DEF CON’s AI Village aims to engage America’s leading hackers to identify security flaws and biases embedded in these large language models, shedding light on potential harms to society.
Rumman Chowdhury, an AI ethicist and researcher and one of the organizers of the event, emphasizes that most harmful incidents associated with large language models occur in everyday use. These incidents can include disinformation, racial bias, inconsistent responses, and the manipulation of AI models to produce undesirable outputs. By allowing hackers to examine the vulnerabilities of AI systems from leading labs, the organizers hope to demonstrate the possibility of creating independent and inclusive AI governance solutions.
The event is also seen as an opportunity to address the lack of inclusivity in AI security discussions. As AI policy is being shaped, it is crucial to involve a wider range of stakeholders to ensure comprehensive governance. Kellee Wicker, the director of the Science and Technology Innovation Program at the Wilson Center, emphasizes the importance of including diverse perspectives in AI security discussions and policymaking.
During the event, participants will be randomly assigned a model from one of the participating firms and provided with a list of challenges. These challenges vary across five categories, including prompt hacking, security, information integrity, internal consistency, and societal harm. Any problematic material identified by the participants will be submitted to judges for evaluation.
The winners of the event are expected to be announced on the final day of the conference. However, the full results of the red-teaming exercise will not be released until February. Red-teaming has gained importance in evaluating AI systems, and leading AI companies, as part of recent voluntary security commitments secured by the White House, have pledged to subject their products to external security testing. While AI safety remains a complex and evolving discipline, exercises like this provide valuable insights into the risks posed by large language models.
The merger of cybersecurity and AI safety in a red-teaming event like DEF CON represents a unique approach to address the risks associated with the rapidly proliferating use of AI. By learning from past experiences in securing computer systems, the cybersecurity community can contribute to mitigating potential harms to society at an early stage.
In conclusion, the hacking contest at DEF CON has brought together thousands of hackers to challenge the security of AI chatbots. The event aims to uncover vulnerabilities and biases in generative AI systems, shedding light on potential risks to society. By involving leading hackers and merging the disciplines of cybersecurity and AI safety, this event provides valuable insights into AI governance and the need for diverse perspectives in shaping AI policy.