|
1 | 1 | ---
|
2 | 2 | title: Cohere For AI Invited Talk
|
3 | 3 |
|
4 |
| -event: Cohere For AI - Guest Speaker Jason S. Lucas, Ph.D Student |
5 |
| -event_url: https://cohere.com/events/c4ai-Lucas-Jason-2024 |
| 4 | +event: Computing Research Association (CRA) CRA-WP Grad Cohort Workshop Lightening Talk |
| 5 | +event_url: https://cra.org/cra-wp/events/2024-grad-cohort-for-women-grad-cohort-for-ideals/ |
6 | 6 |
|
7 |
| -location: Online Presentation |
| 7 | +location: Alohilani Resort Waikiki Beach, Honolulu, HI |
8 | 8 | address:
|
9 | 9 | street:
|
10 |
| - region: Pensylvania |
11 |
| - city: University park |
12 |
| - postcode: 16802-2440 |
| 10 | + region: Central Pacific |
| 11 | + city: Honolulu |
| 12 | + postcode: 96815-2440 |
13 | 13 | country: United States
|
14 | 14 |
|
15 |
| -summary: In this talk we present research that tackles the misuse of large language models (LLMs) by introducing the Fighting Fire with Fire (F3) strategy, which uses GPT-3.5-turbo to generate and detect disinformation. By employing advanced techniques, we achieved a 68-72% accuracy in identifying deceptive content. We also address COVID-19 misinformation in low-resource regions, focusing on the Caribbean. Using US fact-checked claims, we trained models to detect misinformation in English, Spanish, and Haitian French. Our results highlight the limitations of current methods and the need for further multilingual research. |
| 15 | +summary: The widespread use and disruptive effects of large language models (LLMs) have led to concerns about their potential misuse, such as generating harmful and misleading content on a large scale. To address this risk, the authors propose a novel "Fighting Fire with Fire" (F3) strategy, which utilizes the generative and reasoning capabilities of modern LLMs to counter disinformation created by both humans and LLMs. |
16 | 16 |
|
17 | 17 |
|
18 | 18 |
|
19 |
| -abstract: |
| 19 | +abstract: 'Recent ubiquity and disruptive impacts of large language models (LLMs) have raised concerns about their potential to be misused (.i.e, generating large-scale harmful and misleading content). To combat this emerging risk of LLMs, we propose a novel “Fighting Fire with Fire” (F3) strategy that harnesses modern LLMs’ generative and emergent reasoning capabilities to counter human-written and LLM-generated disinformation. First, we leverage GPT-3.5-turbo to synthesize authentic and deceptive LLM-generated content through paraphrase-based and perturbation-based prefix-style prompts, respectively. Second, we apply zero-shot in-context semantic reasoning techniques with cloze-style prompts to discern genuine from deceptive posts and news articles. In our extensive experiments, we observe GPT-3.5-turbo’s zero-shot superiority for both in-distribution and out-of-distribution datasets, where GPT-3.5-turbo consistently achieved accuracy at 68-72%, unlike the decline observed in previous customized and fine-tuned disinformation detectors. Our codebase and dataset are available at https://github.com/mickeymst/F3.' |
20 | 20 |
|
21 | 21 | # Talk start and end times.
|
22 | 22 | # End time can optionally be hidden by prefixing the line with `#`.
|
23 |
| -date: '2024-05-28T09:00:00Z' |
24 |
| -date_end: '2024-05-28T09:20:00Z' |
| 23 | +date: '2024-04-11T09:00:00Z' |
| 24 | +date_end: '2024-04-13T09:20:00Z' |
25 | 25 | all_day: false
|
26 | 26 |
|
27 | 27 | # Schedule page publish date (NOT talk date).
|
28 |
| -publishDate: '2024-05-28T00:00:00Z' |
| 28 | +publishDate: '2024-04-11T00:00:00Z' |
29 | 29 |
|
30 | 30 | authors: [admin]
|
31 | 31 | tags: []
|
|
0 commit comments