#!/usr/bin/python3 import os.path import sys import feedparser from mastodon import Mastodon import json import requests import re import sqlite3 from datetime import datetime, date, time, timedelta if len(sys.argv) < 4: print("Usage: python3 tootbot.py twitter_account mastodon_login mastodon_passwd mastodon_instance") sys.exit(1) # sqlite db to store processed tweets (and corresponding toots ids) sql = sqlite3.connect('tootbot.db') db = sql.cursor() db.execute('''CREATE TABLE IF NOT EXISTS tweets (tweet text, toot text, twitter text, mastodon text, instance text)''') if len(sys.argv)>4: instance = sys.argv[4] else: instance = 'amicale.net' if len(sys.argv)>5: days = int(sys.argv[5]) else: days = 1 twitter = sys.argv[1] mastodon = sys.argv[2] passwd = sys.argv[3] mastodon_api = None d = feedparser.parse('http://twitrss.me/twitter_user_to_rss/?user='+twitter) for t in reversed(d.entries[0:5]): # check if this tweet has been processed db.execute('SELECT * FROM tweets WHERE tweet = ? AND twitter = ? and mastodon = ? and instance = ?',(t.id, twitter, mastodon, instance)) last = db.fetchone() # process only unprocessed tweets less than 1 day old if last is None and (datetime.now()-datetime(t.published_parsed.tm_year, t.published_parsed.tm_mon, t.published_parsed.tm_mday, t.published_parsed.tm_hour, t.published_parsed.tm_min, t.published_parsed.tm_sec) < timedelta(days=days)): if mastodon_api is None: # Create application if it does not exist if not os.path.isfile(instance+'.secret'): if Mastodon.create_app( 'tootbot', api_base_url='https://'+instance, to_file = instance+'.secret' ): print('tootbot app created on instance '+instance) else: print('failed to create app on instance '+instance) sys.exit(1) try: mastodon_api = Mastodon( client_id=instance+'.secret', api_base_url='https://'+instance ) mastodon_api.log_in( username=mastodon, password=passwd, scopes=['read', 'write'], to_file=mastodon+".secret" ) except: print("ERROR: First Login Failed!") sys.exit(1) #h = BeautifulSoup(t.summary_detail.value, "html.parser") c = t.title # if t.author != '(%s)' % twitter: # c = ("RT %s\n" % t.author[1:-1]) + c toot_media = [] # get the pictures... for p in re.finditer(r"https://pbs.twimg.com/[^ \xa0\"]*", t.summary): media = requests.get(p.group(0)) media_posted = mastodon_api.media_post(media.content, mime_type=media.headers.get('content-type')) toot_media.append(media_posted['id']) # replace t.co link by original URL m = re.search(r"http[^ \xa0]*", c) if m != None: l = m.group(0) r = requests.get(l, allow_redirects=False, headers={'Accept-Encoding': 'identity'}) if r.status_code in {301,302}: c = c.replace(l,r.headers.get('Location')) # remove pic.twitter.com links m = re.search(r"pic.twitter.com[^ \xa0]*", c) if m != None: l = m.group(0) c = c.replace(l,' ') # remove ellipsis c = c.replace('\xa0…',' ') if toot_media is not None: toot = mastodon_api.status_post(c, in_reply_to_id=None, media_ids=toot_media, sensitive=False, visibility='public', spoiler_text=None) if "id" in toot: db.execute("INSERT INTO tweets VALUES ( ? , ? , ? , ? , ? )", (t.id, toot["id"], twitter, mastodon, instance)) sql.commit() break