1
0
Fork 0
mirror of https://github.com/Lynnesbian/FediBooks/ synced 2024-11-01 06:08:59 +00:00
FediBooks/app/scrape.py

137 lines
3.8 KiB
Python
Raw Normal View History

#!/usr/bin/env python3
import MySQLdb
import requests
import json, re
import functions
cfg = json.load(open('config.json'))
def scrape_posts(account):
db = MySQLdb.connect(
host = cfg['db_host'],
user=cfg['db_user'],
passwd=cfg['db_pass'],
db=cfg['db_name'],
use_unicode=True,
charset="utf8mb4"
)
handle = account[0]
outbox = account[1]
2020-01-20 02:52:38 +00:00
# print("Scraping {}".format(handle))
c = db.cursor()
last_post = 0
c.execute("SELECT COUNT(*) FROM `posts` WHERE `fedi_id` = %s", (handle,))
count = c.fetchone()
if count is not None and int(count[0]) > 0:
# we've downloaded this user's posts before
# find out the most recently downloaded post of theirs
c.execute("SELECT `post_id` FROM `posts` WHERE `fedi_id` = %s ORDER BY `id` DESC LIMIT 1", (handle,))
last_post = c.fetchone()[0]
2020-03-18 05:06:03 +00:00
done = False
try:
r = requests.get(outbox, timeout = 10)
j = r.json()
# check for pleroma
pleroma = 'next' not in j
if pleroma:
if 'first' in j:
# backwards compatibility for older (pre-v1.0.7) pleroma instances
j = j['first']
else:
uri = "{}&min_id={}".format(outbox, last_post)
r = requests.get(uri, timeout = 10)
j = r.json()
except:
print("Couldn't load or parse outbox at URL {}".format(outbox))
done = True
# here we go!
# warning: scraping posts from outbox.json is messy stuff
while not done and len(j['orderedItems']) > 0:
for oi in j['orderedItems']:
if oi['type'] == "Create":
# this is a status/post/toot/florp/whatever
# first, check to see if we already have this in the database
post_id = re.search(r"([^\/]+)/?$", oi['object']['id']).group(1) # extract 123 from https://example.com/posts/123/
c.execute("SELECT COUNT(*) FROM `posts` WHERE `fedi_id` = %s AND `post_id` = %s", (handle, post_id))
count = c.fetchone()
if count is not None and int(count[0]) > 0:
# this post is already in the DB.
# we'll set done to true because we've caught up to where we were last time.
done = True
# we'll still iterate over the rest of the posts, though, in case there are still some new ones on this page.
continue
content = oi['object']['content']
# remove HTML tags and such from post
content = functions.extract_post(content)
if len(content) > 65535:
# post is too long to go into the DB
continue
try:
c.execute("INSERT INTO `posts` (`fedi_id`, `post_id`, `content`, `cw`) VALUES (%s, %s, %s, %s)", (
handle,
post_id,
content,
1 if (oi['object']['summary'] != None and oi['object']['summary'] != "") else 0
))
except:
#TODO: error handling
2020-03-18 05:05:32 +00:00
print("Failed to insert post {} for user {}".format(post_id, handle))
if not done:
try:
if pleroma:
if 'next' in j:
r = requests.get(j['next'], timeout = 10)
else:
done = True
2020-01-20 04:37:39 +00:00
else:
if 'prev' in j:
r = requests.get(j['prev'], timeout = 10)
else:
done = True
except requests.Timeout:
print("Timed out while loading next page for {}".format(handle))
except:
print("Encountered unknown error while getting next page for {}".format(handle))
if r.status_code == 429:
# we are now being ratelimited, move on to the next user
print("Hit rate limit while scraping {}".format(handle))
done = True
else:
j = r.json()
2019-09-11 05:28:50 +00:00
db.commit()
db.commit()
2020-01-20 02:52:38 +00:00
# print("Finished scraping {}".format(handle))
print("Establishing DB connection")
db = MySQLdb.connect(
host = cfg['db_host'],
user=cfg['db_user'],
passwd=cfg['db_pass'],
2019-09-10 12:21:22 +00:00
db=cfg['db_name'],
use_unicode=True,
charset="utf8mb4"
)
cursor = db.cursor()
print("Downloading posts")
cursor.execute("SELECT `handle`, `outbox` FROM `fedi_accounts` ORDER BY RAND()")
accounts = cursor.fetchall()
cursor.close()
db.close()
functions.do_in_pool(scrape_posts, accounts, timeout=60)
print("Done!")