![]() The conference features a variety of talks and interactive activities focused on various aspects of cybersecurity, including cryptography, malware analysis, network security, technology hardware, and more. ![]() ![]() 2023 has over 1200 attendees with 5 tracks of 75 speakers. It brings together security experts, researchers, hackers, and privacy enthusiasts from around the state to share knowledge, learn from each other, and discuss emerging trends and challenges in the cybersecurity industry. Paying hackers a “bounty” if they uncover a security bug is commonplace in the cybersecurity industry - but it was a newer concept to researchers studying harmful AI bias.CypherCon is an annual spring hacker / cybersecurity conference that takes place downtown Milwaukee, Wisconsin. In another example, searching for Chowdhury using an early version of Microsoft’s Bing search engine chatbot - which is based on the same technology as ChatGPT but can pull real-time information from the internet - led to a profile that speculated Chowdhury “loves to buy new shoes every month” and made strange and gendered assertions about her physical appearance.Ĭhowdhury helped introduce a method for rewarding the discovery of algorithmic bias to DEF CON’s AI Village in 2021 when she was the head of Twitter’s AI ethics team - a job that has since been eliminated upon Elon Musk’s October takeover of the company. ![]() In one example, known as the “grandma exploit,” users were able to get chatbots to tell them how to make a bomb - a request a commercial chatbot would normally decline - by asking it to pretend it was a grandmother telling a bedtime story about how to make a bomb. “What happens now is kind of a scattershot approach where people find stuff, it goes viral on Twitter,” and then it may or may not get fixed if it’s egregious enough or the person calling attention to it is influential, Chowdhury said. Many others are hobbyists showing off humorous or disturbing outputs on social media until they get banned for violating a product’s terms of service. Some are official “red teams” authorized by the companies to “prompt attack” the AI models to discover their vulnerabilities. There’s already a community of users trying their best to trick chatbots and highlight their flaws. government officials in March at the South by Southwest festival in Austin, Texas, where Sven Cattell, founder of DEF CON’s long-running AI Village, and Austin Carson, president of responsible AI nonprofit SeedAI, helped lead a workshop inviting community college students to hack an AI model.Ĭarson said those conversations eventually blossomed into a proposal to test AI language models following the guidelines of the White House’s Blueprint for an AI Bill of Rights - a set of principles to limit the impacts of algorithmic bias, give users control over their data and ensure that automated systems are used safely and transparently. The idea of a mass hack caught the attention of U.S. These systems, built on what’s known as large language models, also emulate the cultural biases they’ve learned from being trained upon huge troves of what people have written online. ![]() “We need a lot of people with a wide range of lived experiences, subject matter expertise and backgrounds hacking at these models and trying to find problems that can then go be fixed.”Īnyone who’s tried ChatGPT, Microsoft’s Bing chatbot or Google’s Bard will have quickly learned that they have a tendency to fabricate information and confidently present it as fact. “This is why we need thousands of people,” said Rumman Chowdhury, a coordinator of the mass hacking event planned for this summer’s DEF CON hacker convention in Las Vegas that’s expected to draw several thousand people. Some of the things they’ll be looking to find: How can chatbots be manipulated to cause harm? Will they share the private information we confide in them to other users? And why do they assume a doctor is a man and a nurse is a woman? No sooner did ChatGPT get unleashed than hackers started “jailbreaking” the artificial intelligence chatbot - trying to override its safeguards so it could blurt out something unhinged or obscene.īut now its maker, OpenAI, and other major AI providers such as Google and Microsoft, are coordinating with the Biden administration to let thousands of hackers take a shot at testing the limits of their technology. ![]()
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |