File size: 2,010 Bytes
ed565a6 4c7d7e7 6c626b7 4c7d7e7 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 | ---
license: cc0-1.0
---
# Dataset Card for Dataset Name
## Dataset Description
- **Homepage:** N/A
- **Repository:** N/A
- **Paper:** N/A
- **Leaderboard:** N/A
- **Point of Contact:** N/A
### Dataset Summary
Text from Reddit Sydney using convokit to obtain it.
### Supported Tasks and Leaderboards
N/A
### Languages
English. Typically Australian English. Will include swearing, profanity, slang and possibly offensive material, as it is taken from Reddit and has not been filtered.
## Dataset Structure
Plain text
### Data Instances
N/A
### Data Fields
N/A
### Data Splits
N/A. You need to do splits yourself
## Dataset Creation
Using this script:
```python
from convokit import Corpus, download # https://convokit.cornell.edu/documentation/subreddit.html
corpus = Corpus(filename=download("subreddit-sydney"))
textarr = []
for utt in corpus.iter_utterances():
if utt.text != "[deleted]":
textarr.append(utt.text)
text = '\n'.join(textarr);
text_file = open("input.txt", "w")
n = text_file.write(text)
text_file.close()
```
### Curation Rationale
I don't know what this means.
### Source Data
Reddit's Sydney subreddit.
#### Initial Data Collection and Normalization
See script
#### Who are the source language producers?
See script
### Annotations
#### Annotation process
N/A
#### Who are the annotators?
N/A
### Personal and Sensitive Information
Most likely.
## Considerations for Using the Data
### Social Impact of Dataset
There is unfettered discussion. It is probably horrible to release a LLM trained on just this with no safety precautions.
### Discussion of Biases
This is going to be full of biases. It is raw internet discussion.
### Other Known Limitations
None
## Additional Information
None
### Dataset Curators
Not curated
### Licensing Information
Public Domain for the Python script and this representation of Reddit data. Original authors and Reddit may have some rights.
### Citation Information
None
### Contributions
N/A
|