#!/usr/bin/python3 import os.path import sys import feedparser from mastodon import Mastodon import json import requests import re import sqlite3 from datetime import datetime, date, time, timedelta import time import random if len(sys.argv) < 4: print("Usage: python3 tootbot.py twitter_account mastodon_login mastodon_passwd mastodon_instance") sys.exit(1) random.seed() sleep_time = random.randint(0, 10 * 60) time.sleep(sleep_time) # sqlite db to store processed tweets (and corresponding toots ids) sql = sqlite3.connect('tootbot.db') db = sql.cursor() db.execute('''CREATE TABLE IF NOT EXISTS tweets (tweet text, toot text, twitter text, mastodon text, instance text)''') if len(sys.argv)>4: instance = sys.argv[4] else: instance = 'amicale.net' if len(sys.argv)>5: days = int(sys.argv[5]) else: days = 1 twitter = sys.argv[1] mastodon = sys.argv[2] passwd = sys.argv[3] mastodon_api = None d = feedparser.parse('http://twitrss.me/twitter_user_to_rss/?user='+twitter) for t in reversed(d.entries[0:5]): # check if this tweet has been processed db.execute('SELECT * FROM tweets WHERE tweet = ? AND twitter = ? and mastodon = ? and instance = ?',(t.id, twitter, mastodon, instance)) last = db.fetchone() # process only unprocessed tweets less than 1 day old if last is None and (datetime.now()-datetime(t.published_parsed.tm_year, t.published_parsed.tm_mon, t.published_parsed.tm_mday, t.published_parsed.tm_hour, t.published_parsed.tm_min, t.published_parsed.tm_sec) < timedelta(days=days)): if mastodon_api is None: # Create application if it does not exist if not os.path.isfile(instance+'.secret'): if Mastodon.create_app( 'tootbot', api_base_url='https://'+instance, to_file = instance+'.secret' ): print('tootbot app created on instance '+instance) else: print('failed to create app on instance '+instance) sys.exit(1) try: mastodon_api = Mastodon( client_id=instance+'.secret', api_base_url='https://'+instance ) mastodon_api.log_in( username=mastodon, password=passwd, scopes=['read', 'write'], to_file=mastodon+".secret" ) except: print("ERROR: First Login Failed!") sys.exit(1) #h = BeautifulSoup(t.summary_detail.value, "html.parser") c = t.title # if t.author != '(%s)' % twitter: # c = ("RT %s\n" % t.author[1:-1]) + c toot_media = [] # get the pictures... for p in re.finditer(r"https://pbs.twimg.com/[^ \xa0\"]*", t.summary): media = requests.get(p.group(0)) media_posted = mastodon_api.media_post(media.content, mime_type=media.headers.get('content-type')) toot_media.append(media_posted['id']) # replace t.co link by original URL m = re.search(r"http[^ \xa0]*", c) if m != None: l = m.group(0) r = requests.get(l, allow_redirects=False) if r.status_code in {301,302}: c = c.replace(l,r.headers.get('Location')) # remove pic.twitter.com links m = re.search(r"pic.twitter.com[^ \xa0]*", c) if m != None: l = m.group(0) c = c.replace(l,' ') # remove ellipsis c = c.replace('\xa0…',' ') if toot_media is not None: toot = mastodon_api.status_post(c, in_reply_to_id=None, media_ids=toot_media, sensitive=False, visibility='public', spoiler_text=None) if "id" in toot: db.execute("INSERT INTO tweets VALUES ( ? , ? , ? , ? , ? )", (t.id, toot["id"], twitter, mastodon, instance)) sql.commit() break