Skip to content

Commit 6def2da

Browse files
committed
cohere
1 parent cb832bf commit 6def2da

File tree

1 file changed

+11
-11
lines changed
  • content/event/CoHere Invited Talk

1 file changed

+11
-11
lines changed

content/event/CoHere Invited Talk/index.md

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -1,31 +1,31 @@
11
---
22
title: Cohere For AI Invited Talk
33

4-
event: Cohere For AI - Guest Speaker Jason S. Lucas, Ph.D Student
5-
event_url: https://cohere.com/events/c4ai-Lucas-Jason-2024
4+
event: Computing Research Association (CRA) CRA-WP Grad Cohort Workshop Lightening Talk
5+
event_url: https://cra.org/cra-wp/events/2024-grad-cohort-for-women-grad-cohort-for-ideals/
66

7-
location: Online Presentation
7+
location: Alohilani Resort Waikiki Beach, Honolulu, HI
88
address:
99
street:
10-
region: Pensylvania
11-
city: University park
12-
postcode: 16802-2440
10+
region: Central Pacific
11+
city: Honolulu
12+
postcode: 96815-2440
1313
country: United States
1414

15-
summary: In this talk we present research that tackles the misuse of large language models (LLMs) by introducing the Fighting Fire with Fire (F3) strategy, which uses GPT-3.5-turbo to generate and detect disinformation. By employing advanced techniques, we achieved a 68-72% accuracy in identifying deceptive content. We also address COVID-19 misinformation in low-resource regions, focusing on the Caribbean. Using US fact-checked claims, we trained models to detect misinformation in English, Spanish, and Haitian French. Our results highlight the limitations of current methods and the need for further multilingual research.
15+
summary: The widespread use and disruptive effects of large language models (LLMs) have led to concerns about their potential misuse, such as generating harmful and misleading content on a large scale. To address this risk, the authors propose a novel "Fighting Fire with Fire" (F3) strategy, which utilizes the generative and reasoning capabilities of modern LLMs to counter disinformation created by both humans and LLMs.
1616

1717

1818

19-
abstract:
19+
abstract: 'Recent ubiquity and disruptive impacts of large language models (LLMs) have raised concerns about their potential to be misused (.i.e, generating large-scale harmful and misleading content). To combat this emerging risk of LLMs, we propose a novel “Fighting Fire with Fire” (F3) strategy that harnesses modern LLMs’ generative and emergent reasoning capabilities to counter human-written and LLM-generated disinformation. First, we leverage GPT-3.5-turbo to synthesize authentic and deceptive LLM-generated content through paraphrase-based and perturbation-based prefix-style prompts, respectively. Second, we apply zero-shot in-context semantic reasoning techniques with cloze-style prompts to discern genuine from deceptive posts and news articles. In our extensive experiments, we observe GPT-3.5-turbo’s zero-shot superiority for both in-distribution and out-of-distribution datasets, where GPT-3.5-turbo consistently achieved accuracy at 68-72%, unlike the decline observed in previous customized and fine-tuned disinformation detectors. Our codebase and dataset are available at https://github.com/mickeymst/F3.'
2020

2121
# Talk start and end times.
2222
# End time can optionally be hidden by prefixing the line with `#`.
23-
date: '2024-05-28T09:00:00Z'
24-
date_end: '2024-05-28T09:20:00Z'
23+
date: '2024-04-11T09:00:00Z'
24+
date_end: '2024-04-13T09:20:00Z'
2525
all_day: false
2626

2727
# Schedule page publish date (NOT talk date).
28-
publishDate: '2024-05-28T00:00:00Z'
28+
publishDate: '2024-04-11T00:00:00Z'
2929

3030
authors: [admin]
3131
tags: []

0 commit comments

Comments
 (0)