File size: 3,140 Bytes
8caeaa7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7bc0b3c
8caeaa7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7bc0b3c
 
 
 
 
 
 
 
 
 
8caeaa7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
---
language:
  - en
license: apache-2.0
tags:
  - text
pretty_name: Reddit DadJokes dataset
size_categories:
  - "100K<n<1M"
task_categories:
  - text-generation
dataset_info:
  config_name: default
  features:
  - name: input
    dtype: string
  - name: output
    dtype: string
  - name: instruction
    dtype: string
  splits:
    - name: train
      num_bytes: 19203073
      num_examples: 147753
train-eval-index:
  - config: default
    task: text-generation
    splits:
      train_split: train
configs:
- config_name: default
  data_files:
  - split: train
    path: "data/train/*"
---

# Reddit /r/DadJokes dataset

Dataset is based on a [semi-public Pushshift Reddit dataset](https://convokit.cornell.edu/documentation/subreddit.html).

## Source data

The actual extract of all top-level posts on `/r/DadJokes` can be found in the `src/reddit_dadjokes.csv.gz` file. The raw post contents are very noisy, so we did a couple of steps to clean things up:

* Used a `Llama-8B-instruct` to split each dad joke into the `intro` and a `punchline` parts. See `src/parse.py` script for details.
* Removed too short and too long posts. Also removed all punchlines being emojis or URLs.
* There are a lot of duplicate submissions present, so we used the `intfloat/e5-base-v2` model for semantic deduplication.

## Usage

Dataset has all the metadata to be used with the [HF Datasets] library:

```python
from datasets import load_dataset

data = load_dataset('shuttie/reddit-dadjokes', split='train')
```

## Stats

147753 jokes split to `intro` and `punchline` parts. Jokes are from 2014 till 2022. 

## Examples

To be Alpaca-format compatible we use the `Continue a dad joke:` instruction by default, but it's up to you for using something else.


```json
{"input": "A grizzly kept talking to me", "output": "He was unbearable", "instruction": "Continue a dad joke:"}
{"input": "I rubbed mayonnaise on my eyes", "output": "Oh fuck oh shit it hurts please help this is no joke it pains fuck help me already", "instruction": "Continue a dad joke:"}
{"input": "What do you say to encourage an asteroid?", "output": "Go little rockstar.", "instruction": "Continue a dad joke:"}
{"input": "They always ask me why my mood is always negative", "output": "Just multiply by -1 \ud83d\ude09", "instruction": "Continue a dad joke:"}
{"input": "Recently I started working with horses", "output": "It's a stable job.", "instruction": "Continue a dad joke:"}
{"input": "My favorite word is \"drool\"", "output": "Just rolls off the tongue.", "instruction": "Continue a dad joke:"}
{"input": "Where\u2019s a dogs favorite place to eat", "output": "At Woofle House", "instruction": "Continue a dad joke:"}
{"input": "Why don't oysters share their pearls?", "output": "They're shellfish \ud83d\udc1a\ud83d\ude02!", "instruction": "Continue a dad joke:"}
{"input": "What do you call two guys hanging out above a window?", "output": "Kurt & Rod", "instruction": "Continue a dad joke:"}
{"input": "Me: I'm not saying a word without my lawyer present!", "output": "So where's my present?!", "instruction": "Continue a dad joke:"}
```

## License

Apache 2.0