Skip to content

Instantly share code, notes, and snippets.

@FdelMazo
Last active March 5, 2020 03:54
Show Gist options
  • Save FdelMazo/7eb48d9dc33fd76da40d2efa6667f2ef to your computer and use it in GitHub Desktop.
Save FdelMazo/7eb48d9dc33fd76da40d2efa6667f2ef to your computer and use it in GitHub Desktop.
MoviesReDownloader: Fetches for movies in bad quality in your hard drive and downloads them in good quality.

Quick Usage

  • Throw downloader.py and rarbgapi.py to the working folder and just write python downloader.py
    • Fetches for 'bad' movies (as in, not in good quality) and writes them on a list on 'fetch.txt'
    • Windows: Downloads them through magnet links in good quality.
    • Linux: Copies the magnet links to 'magnets.txt'

Complete Features and options

  • -h, --help show this help message and exit
  • --fetch-only Just fecth the bad movies. Generates fetch.txt
  • --download-only Just downloads the movies listed in fetch.txt
  • --dry-run Only show what would be done without modifying files
import os
import re
import webbrowser
from rarbgapi import RarbgAPI
import argparse
#DOWNLOADING
GOOD_CODEC = "RARBG"
GOOD_SPECS = [GOOD_CODEC, "1080p", "AAC", "H264"] #Specs to be downloaded
BAD_SPECS = ["Ganool", "YIFY", "YTS", "Ozlem", "fgt", "MKVCage", "usabit", "vppv", "etrg", "sujaidr"] #specs to be fecthed
# BAD_SPECS += ["xvid", "dvdrip"] #specs to be fecthed
"""********** FETCH ************"""
def fetch_movies():
print("Fetching movies")
movie_list = set()
for directory in os.listdir():
if not os.path.isdir(directory): continue
if any(bad_spec.lower() in directory.lower() for bad_spec in BAD_SPECS):
movie_list.add(directory)
for archivo in os.listdir(directory):
if any(bad_spec.lower() in archivo.lower() for bad_spec in BAD_SPECS):
movie_list.add(directory)
for movie in movie_list:
print("\t* {}".format(movie))
print("Movies fetched")
return movie_list
def list_to_txt(movie_list, output):
archivo = open(output, 'a')
print("List passed to {}.".format(output))
for linea in movie_list:
archivo.write(linea+'\n')
archivo.close()
"""********** MAIN ************"""
def main():
parser = argparse.ArgumentParser()
group = parser.add_mutually_exclusive_group()
group.add_argument('--fetch-only',help='Just fecth the bad movies. Generates fetch.txt', action='store_true')
group.add_argument('--download-only',help='Just downloads the movies listed in fetch.txt', action='store_true')
parser.add_argument('--dry-run', help='Only show what would be done without modifying files', action='append_const', const=("dry_run",True), dest='flags')
args = parser.parse_args()
flags = dict(args.flags) if args.flags else {}
if not args.download_only:
movie_list = fetch_movies()
if movie_list: list_to_txt(movie_list, 'fetch.txt')
else: print("There are no bad peliculas")
if not args.fetch_only:
peliculas = open('fetch.txt').read().splitlines()
client = RarbgAPI()
not_downloaded = []
for pelicula in peliculas:
print(pelicula)
search_pelicula = re.split('[.]| ',pelicula)
for i, tag in enumerate(search_pelicula):
if "720p" in tag or "1080p" in tag or "480p" in tag:
search_pelicula = search_pelicula[:i]
break
search_pelicula = ' '.join(search_pelicula)
print("\t* Searching for: {}".format(search_pelicula+ " " + ' '.join(GOOD_SPECS)))
torrents = client.search(search_string = search_pelicula+" "+' '.join(GOOD_SPECS), limit=25, category = 44)
if not torrents:
not_downloaded.append(search_pelicula)
print("\t* Not downloaded: {}".format(search_pelicula))
continue
print("\t* Downloaded: {}".format(torrents[0].filename))
if not flags.get('dry_run'):
if "Windows" in os.name: os.startfile(torrents[0].download)
else:
with open('magnets.txt', 'a') as archivo: archivo.write(torrents[0].download+'\n')
if not_downloaded: list_to_txt(not_downloaded, "not_downloaded.txt")
main()
# https://github.com/verybada/rarbgapi
import time
import logging
import requests
import threading
class TokenExpireException(Exception):
pass
# pylint: disable=too-many-instance-attributes,too-few-public-methods
class Torrent(object):
'''
brief
{
"filename":"Off.Piste.2016.iNTERNAL.BDRip.x264-LiBRARiANS",
"category":"Movies/x264",
"download":"magnet:..."
}
extended
{
"title":"Off.Piste.2016.iNTERNAL.BDRip.x264-LiBRARiANS",
"category":"Movies/x264",
"download":"magnet:...",
"seeders":12,
"leechers":6,
"size":504519520,
"pubdate":"2017-05-21 02:13:49 +0000",
"episode_info":{
"imdb":"tt4443856",
"tvrage":null,
"tvdb":null,
"themoviedb":"430293"
},
"ranked":1,
"info_page":"https://torrentapi.org/...."
}
'''
def __init__(self, mapping):
self._raw = mapping
self.is_extended = 'title' in self._raw
self.category = self._raw['category']
self.download = self._raw['download']
self.filename = self._raw.get('filename') or self._raw.get('title')
self.size = self._raw.get('size')
self.pubdate = self._raw.get('pubdate')
self.page = self._raw.get('info_page')
def __str__(self):
return "{} {}".format(self.filename)
def json_hook(dct):
error_code = dct.get('error_code')
if error_code in [2, 4]:
raise TokenExpireException('Token expired')
if 'download' in dct:
return Torrent(dct)
return dct
class _RarbgAPIv2(object):
'''
API reference
https://torrentapi.org/apidocs_v2.txt
'''
ENDPOINT = 'http://torrentapi.org/pubapi_v2.php'
def __init__(self):
super(_RarbgAPIv2, self).__init__()
self._endpoint = self.ENDPOINT
# pylint: disable=no-self-use
def _requests(self, method, url, params=None):
sess = requests.Session()
req = requests.Request(method, url, params=params)
preq = req.prepare()
resp = sess.send(preq)
resp.raise_for_status()
return resp
def _get_token(self):
'''
{"token":"xxxxx"}
'''
params = {
'get_token': 'get_token'
}
return self._requests('GET', self._endpoint, params)
def _query(self, mode, token=None, **kwargs):
params = {
'mode': mode,
'token': token
}
for key, value in kwargs.items():
if key not in ['search_string', 'sort', 'limit',
'category', 'format_']:
raise ValueError('unsupported parameter %s' % key)
if value is None:
continue
params[key] = value
return self._requests('GET', self._endpoint, params)
def request(func):
# pylint: disable=protected-access
def wrapper(self, *args, **kwargs):
max_retries = retries = self._options['retries']
while retries > 0:
try:
backoff = 2**(max_retries - retries)
if not self._bucket.acquire(1, timeout=2):
raise ValueError('accquire token timeout')
if not self._token:
raise TokenExpireException('Empty token')
resp = func(self, token=self._token, *args, **kwargs)
json_ = resp.json(object_hook=json_hook)
if 'torrent_results' not in json_:
print("\t* No torrents found for {}".format(kwargs['search_string']))
return
return json_['torrent_results']
except TokenExpireException:
resp = self._get_token()
content = resp.json()
self._token = content['token']
self._log.debug('token=%s', self._token)
except Exception as exp: # pylint: disable=broad-except
self._log.exception('Unexpected exceptin %s', exp)
retries -= 1
if not retries:
raise
else:
retries -= 1
finally:
time.sleep(backoff)
assert 0, 'Unexpected response'
return wrapper
class RarbgAPI(_RarbgAPIv2):
CATEGORY_MOVIE_XVID = 16
CATEGORY_MOVIE_XVID_720P = 48
CATEGORY_MOVIE_H264 = 17
CATEGORY_MOVIE_H264_1080P = 44
CATEGORY_MOVIE_H264_720P = 45
CATEGORY_MOVIE_H264_3D = 47
CATEGORY_MOVIE_FULL_BD = 42
CATEGORY_MOVIE_BD_REMUX = 46
CATEGORY_TV_EPISODES = 18
CATEGORY_TV_EPISODES_HD = 41
CATEGORY_TV_EPISODES_UHD = 49
CATEGORY_MUSIC_MP3 = 23
CATEGORY_MUSIC_FLAC = 25
CATEGORY_GAMES_PC_ISO = 27
CATEGORY_GAMES_PC_RIP = 28
CATEGORY_GAMES_PS3 = 40
CATEGORY_GAMES_XBOX = 32
CATEGORY_SOFTWARE = 33
CATEGORY_EBOOK = 35
def __init__(self, **options):
super(RarbgAPI, self).__init__()
self._token = None
self._bucket = LeakyBucket(0.5)
self._log = logging.getLogger(__name__)
default_options = {
'retries': 5,
}
if options:
default_options.update(options)
self._options = default_options
@request
# pylint: disable=too-many-arguments
def list(self, search_string=None, sort=None, limit=None,
category=None, format_=None, **kwargs):
return self._query('list', search_string=search_string, sort=sort,
limit=limit, category=category, format_=format_,
**kwargs)
@request
# pylint: disable=too-many-arguments
def search(self, search_string=None, sort=None, limit=None,
category=None, format_=None, **kwargs):
return self._query('search', search_string=search_string,
sort=sort, limit=limit, category=category,
format_=format_, **kwargs)
class LeakyBucket(object): # pylint: disable=too-few-public-methods
def __init__(self, rate):
self._rate = rate
self._lock = threading.Lock()
self._last_time = int(time.time())
self._token = 0
def acquire(self, token, timeout=None):
delay = 1
while True:
with self._lock: # pylint: disable=not-context-manager
now = int(time.time())
diff = now - self._last_time
if diff:
self._token += self._rate * diff
self._last_time = now
if token <= self._token:
self._token -= token
return True
if isinstance(timeout, (int, float)):
if timeout <= 0:
return False
timeout -= delay
time.sleep(delay)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment