| | import time |
| | import praw |
| | import prawcore |
| | import pandas as pd |
| | import os |
| | from datetime import datetime |
| |
|
| | FILENAME_POSTS = 'output_posts_after2022.csv' |
| |
|
| |
|
| | def create_reddit_instance(): |
| | reddit = praw.Reddit( |
| | client_id=os.getenv('REDDIT_CLIENT_ID'), |
| | client_secret=os.getenv('REDDIT_CLIENT_SECRET'), |
| | user_agent=os.getenv('REDDIT_USER_AGENT') |
| | ) |
| | return reddit |
| |
|
| |
|
| | def get_subreddit(reddit, subreddit_name): |
| | subreddit = reddit.subreddit(subreddit_name) |
| | return subreddit |
| |
|
| |
|
| | def load_existing_data(file_name): |
| | if os.path.exists(file_name): |
| | df = pd.read_csv(file_name) |
| | existing_ids = df['id'].tolist() |
| | else: |
| | df = pd.DataFrame() |
| | existing_ids = [] |
| | return df, existing_ids |
| |
|
| |
|
| | def get_top_comments(submission): |
| | top_comments = [] |
| | submission.comments.sort_by = 'top' |
| | for comment in submission.comments[:5]: |
| | top_comments.append(comment.body) |
| | return top_comments |
| |
|
| |
|
| | def get_new_post_row(submission, top_comments): |
| | date = datetime.fromtimestamp(submission.created) |
| | new_row = { |
| | "id": submission.id, |
| | "url": submission.url, |
| | "score": submission.score, |
| | "title": submission.title, |
| | "body": submission.selftext, |
| | "top_comment1": top_comments[0] if len(top_comments) > 0 else None, |
| | "top_comment2": top_comments[1] if len(top_comments) > 1 else None, |
| | "top_comment3": top_comments[2] if len(top_comments) > 2 else None, |
| | "top_comment4": top_comments[3] if len(top_comments) > 3 else None, |
| | "top_comment5": top_comments[4] if len(top_comments) > 4 else None, |
| | "date": date |
| | } |
| | return new_row |
| |
|
| |
|
| | def save_data(df, file_name): |
| | df.to_csv(file_name, index=False) |
| |
|
| |
|
| | def main(): |
| | reddit = create_reddit_instance() |
| | subreddit = get_subreddit(reddit, 'tunisia') |
| | df_posts, existing_post_ids = load_existing_data(FILENAME_POSTS) |
| |
|
| | print('Starting to scrape posts') |
| |
|
| | |
| | new_posts = list(subreddit.new(limit=1000)) |
| |
|
| | for submission in new_posts: |
| | if submission.id in existing_post_ids: |
| | print(f'Skipped post {submission.id}') |
| | continue |
| | try: |
| | top_comments = get_top_comments(submission) |
| | new_row = get_new_post_row(submission, top_comments) |
| | df_posts = df_posts._append(new_row, ignore_index=True) |
| | except prawcore.exceptions.TooManyRequests: |
| | print("Hit rate limit, sleeping .....") |
| | time.sleep(60) |
| |
|
| | save_data(df_posts, FILENAME_POSTS) |
| | print('Finished scraping') |
| | print("Data saved to ", FILENAME_POSTS) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|