Yulia Tsvetkov – UW News /news Wed, 28 Aug 2024 18:22:43 +0000 en-US hourly 1 https://wordpress.org/?v=6.9.4 Large language models can help detect social media bots — but can also make the problem worse /news/2024/08/28/large-language-models-social-media-bots-twitter-ai/ Wed, 28 Aug 2024 15:01:16 +0000 /news/?p=85927 A drawing of a robot with an empty speech bubble.

An estimated that between a third and two thirds of accounts on the social media site were bots. And many of these automatons flooding social media are dispatched to sow political polarization, hate, misinformation, propaganda and scams. The ability to sift them out of the online crowds is vital for a safer, more humane (or at least more human) internet.

But the recent proliferation of large language models (known as “LLMs”), such as OpenAI’s ChatGPT and Meta’s Llama, stands to complicate the world of social media bots.

A team led by ӰӴý researchers found that while operators can use customized LLMs to make bots more sophisticated at evading automated detectors, LLMs can also improve systems that detect bots. In the team’s tests, LLM-based bots reduced the performance of existing detectors by 30%. Yet researchers also found that an LLM trained specifically to detect social media bots outperformed state-of-the-art systems by 9%.

The team Aug. 11 at the 62nd Annual Meeting of the Association for Computational Linguistics in Bangkok.

“There’s always been an arms race between bot operators and the researchers trying to stop them,” said lead author , a doctoral student in the Paul G. Allen School of Computer Science & Engineering. “Each advance in bot detection is often met with an advance in bot sophistication, so we explored the opportunities and the risks that large language models present in this arms race.”

Researchers tested LLMs’ potential to detect bots in a few ways. When they fed Twitter data sets (culled before the platform became X) to off-the-shelf LLMs, including ChatGPT and Llama, the systems failed to accurately detect bots more than currently used technologies.

“Analyzing whether a user is a bot or not is much more complex than some of the tasks we’ve seen these general LLMs excel at, like recalling a fact or doing a grade-school math problem,” Feng said.

This complexity comes in part from the need to analyze three types of information for different attributes to detect a bot: the metadata (number of followers, geolocation, etc.), the text posted online and the network properties (such as what accounts a user is following).

When the team fine-tuned the LLMs with instructions on how to detect bots based on these three types of information, the models were able to detect bots with greater accuracy than current state-of-the-art systems.

The team also explored how LLMs might make bots more sophisticated and harder to detect. First the researchers simply gave LLMs prompts such as, “Please rewrite the description of this bot account to sound like a genuine user.”

They also tested more iterative, complicated approaches. In one test, the LLM would rewrite the bot post. The team then ran this through an existing bot-detection system, which would estimate the likelihood that a post was written by a bot. This process would be repeated as the LLM worked to lower that estimate. The team ran a similar test while removing and adding accounts that the bot followed to adjust its likelihood score.

These strategies, particularly rewriting the bots’ posts, reduced the effectiveness of the bot detection systems by as much as 30%. But the LLM-based detectors the team trained saw only a 2.3% drop in effectiveness on these manipulated posts, suggesting that the best way to detect LLM-powered bots might be with LLMs themselves.

“This work is only a scientific prototype,” said senior author , an associate professor in the Allen School. “We aren’t releasing these systems as tools anyone can download, because in addition to developing technology to defend against malicious bots, we are experimenting with threat modeling of how to create an evasive bot, which continues the cat-and-mouse game of building stronger bots that need stronger detectors.”

Researchers note that there are important limitations to using LLMs as bot detectors, such as the systems’ . They also highlight that the data used in the paper is from 2022, before Twitter effectively .

In the future, researchers want to look at bot detection beyond text, such as memes or videos on other platforms such as TikTok, where newer data sets are available. The team also wants to expand the research into other languages.

“Doing this research across different languages is extremely important,” Tsvetkov said. “We are seeing a lot of misinformation, manipulation and the targeting of specific populations as a result of various world conflicts.”

Additional co-authors on this paper are and , both undergraduates at Xi’an Jiaotong University; , an assistant professor at Xi’an Jiaotong University; and , a doctoral student at the University of Notre Dame. This research was funded by an NSF CAREER award.

For more information, contact Feng at shangbin@cs.washington.edu and Tsvetkov at yuliats@cs.washington.edu.

]]>
UW biologist and computer scientist named Sloan Fellows /news/2022/02/17/uw-biologist-and-computer-scientist-named-sloan-fellows/ Thu, 17 Feb 2022 21:59:09 +0000 /news/?p=77346 head shots
UW computer scientist Yulia Tsvetkov (l) and biologist Briana Abrahms (r) named 2022 Sloan Fellows. Photo: ӰӴý

Two faculty members at the ӰӴý have been awarded early-career fellowships from the Alfred P. Sloan Foundation. The new Sloan Fellows, announced Feb. 15, are , an assistant professor in the Department of Biology, and , an assistant professor in the Paul G. Allen School of Computer Science & Engineering.

Open to scholars in eight scientific and technical fields — chemistry, computer science, Earth system science, economics, mathematics, neuroscience and physics — the fellowships honor those early-career researchers whose achievements mark them among the next generation of scientific leaders.

The 118 were selected in coordination with the research community. Candidates are nominated by their peers, and fellows are selected by independent panels of senior scholars based on each candidate’s research accomplishments, creativity and potential to become a leader in their field. Each fellow will receive $75,000 to apply toward research endeavors.

This year’s fellows come from 51 institutions across the United States and Canada, spanning fields from evolutionary biology to data science.

Abrahms is an assistant professor of biology and holds the inaugural Boersma Endowed Chair in Natural History and Conservation. Her research program integrates animal bio-logging technology, Earth observation and big data analytics to advance understanding of the causes and consequences of wildlife responses to global change across marine and terrestrial systems. In addition to advancing basic ecological theory, Abrahms is passionate about developing data-driven, publicly available tools that bolster capacity to conserve the natural world.

“How do animals make decisions in the face of global change, and what are the consequences of those decisions for individual fitness, populations and interactions with other animals and humans? This is a big question my group is working to answer, which can inform both biodiversity conservation and human sustainability,” Abrahms said. “We’re becoming increasingly interested in understanding how species responses to environmental change can have unanticipated and often negative consequences for social-ecological systems so that we can reduce unwanted outcomes in the future.”

Tsvetkov is an assistant professor in the Allen School. She engages in multidisciplinary research at the nexus of machine learning, computational linguistics and the social sciences to develop practical solutions to natural language processing, or NLP, problems that combine sophisticated learning and modeling methods with insights into human languages and the people who speak them.

“The huge success of contemporary AI-powered NLP technology stems from the fact that it has matured enough to effectively serve and interact with humans. However, there is still a technological divide: The rich ecosystem of language-aware applications — machine translation, question answering, educational applications, summarization — are well-equipped to serve privileged users. But those same applications are systematically biased in ways that render them less useful for millions of other users, including speakers of low-resource languages or representatives of disadvantaged groups discriminated by gender, race, age or ethnicity,” Tsvetkov said. “The long-term goal of my research has been to bridge this divide, and to develop effective, accessible and equitable language technologies, serving all users, across populations, cultures and language boundaries.”

For more information, contact Abrahms at abrahms@uw.edu or Tsvetkov at yuliats@cs.washington.edu.

 

 

 

 

]]>