# Author: Nic Wolfe <nic@wolfeden.ca> # URL: http://code.google.com/p/sickbeard/ # # This file is part of SickRage. # # SickRage is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # SickRage is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with SickRage. If not, see <http://www.gnu.org/licenses/>. from __future__ import with_statement import os import ctypes import random import re import socket import stat import tempfile import time import traceback import urllib import urllib2 import hashlib import httplib import urlparse import uuid import base64 import zipfile import datetime import errno import ast import operator from contextlib import closing import sickbeard import adba import requests import certifi import xmltodict import subprocess from sickbeard.exceptions import MultipleShowObjectsException, ex from sickbeard import logger, classes from sickbeard.common import USER_AGENT, cpu_presets, mediaExtensions, subtitleExtensions from sickbeard import db from sickbeard import encodingKludge as ek from sickbeard import notifiers from sickbeard import clients from sickbeard.subtitles import isValidLanguage from lib.cachecontrol import CacheControl, caches from itertools import izip, cycle import shutil import lib.shutil_custom shutil.copyfile = lib.shutil_custom.copyfile_custom urllib._urlopener = classes.SickBeardURLopener() def fixGlob(path): path = re.sub(r'\[', '[[]', path) return re.sub(r'(?<!\[)\]', '[]]', path) def indentXML(elem, level=0): ''' Does our pretty printing, makes Matt very happy ''' i = "\n" + level * " " if len(elem): if not elem.text or not elem.text.strip(): elem.text = i + " " if not elem.tail or not elem.tail.strip(): elem.tail = i for elem in elem: indentXML(elem, level + 1) if not elem.tail or not elem.tail.strip(): elem.tail = i else: # Strip out the newlines from text if elem.text: elem.text = elem.text.replace('\n', ' ') if level and (not elem.tail or not elem.tail.strip()): elem.tail = i def remove_extension(name): """ Remove download or media extension from name (if any) """ if name and "." in name: base_name, sep, extension = name.rpartition('.') # @UnusedVariable if base_name and extension.lower() in ['nzb', 'torrent'] + mediaExtensions: name = base_name return name def remove_non_release_groups(name): """ Remove non release groups from name """ if not name: return name # Do not remove all [....] suffixes, or it will break anime releases ## Need to verify this is true now # Check your database for funky release_names and add them here, to improve failed handling, archiving, and history. # select release_name from tv_episodes WHERE LENGTH(release_name); # [eSc], [SSG], [GWC] are valid release groups for non-anime removeWordsList = {'\[rartv\]$': 'searchre', '\[rarbg\]$': 'searchre', '\[eztv\]$': 'searchre', '\[ettv\]$': 'searchre', '\[vtv\]$': 'searchre', '\[GloDLS\]$': 'searchre', '\[silv4\]$': 'searchre', '\[Seedbox\]$': 'searchre', '\[AndroidTwoU\]$': 'searchre', '\.RiPSaLoT$': 'searchre', '-NZBGEEK$': 'searchre', '-RP$': 'searchre', '-20-40$': 'searchre', '\[NO-RAR\] - \[ www\.torrentday\.com \]$': 'searchre', '^\[ www\.TorrentDay\.com \] - ': 'searchre', '^\[ www\.Cpasbien\.pw \] ': 'searchre', } _name = name for remove_string, remove_type in removeWordsList.iteritems(): if remove_type == 'search': _name = _name.replace(remove_string, '') elif remove_type == 'searchre': _name = re.sub(r'(?i)' + remove_string, '', _name) return _name.strip('.- ') def replaceExtension(filename, newExt): ''' >>> replaceExtension('foo.avi', 'mkv') 'foo.mkv' >>> replaceExtension('.vimrc', 'arglebargle') '.vimrc' >>> replaceExtension('a.b.c', 'd') 'a.b.d' >>> replaceExtension('', 'a') '' >>> replaceExtension('foo.bar', '') 'foo.' ''' sepFile = filename.rpartition(".") if sepFile[0] == "": return filename else: return sepFile[0] + "." + newExt def notTorNZBFile(filename): return not (filename.endswith(".torrent") or filename.endswith(".nzb")) def isSyncFile(filename): extension = filename.rpartition(".")[2].lower() #if extension == '!sync' or extension == 'lftp-pget-status' or extension == 'part' or extension == 'bts': syncfiles = sickbeard.SYNC_FILES if extension in syncfiles.split(","): return True else: return False def isMediaFile(filename): # ignore samples if re.search('(^|[\W_])(sample\d*)[\W_]', filename, re.I): return False # ignore MAC OS's retarded "resource fork" files if filename.startswith('._'): return False sepFile = filename.rpartition(".") if re.search('extras?$', sepFile[0], re.I): return False if sepFile[2].lower() in mediaExtensions: return True else: return False def isRarFile(filename): archive_regex = '(?P<file>^(?P<base>(?:(?!\.part\d+\.rar$).)*)\.(?:(?:part0*1\.)?rar)$)' if re.search(archive_regex, filename): return True return False def isBeingWritten(filepath): # Return True if file was modified within 60 seconds. it might still be being written to. ctime = max(ek.ek(os.path.getctime, filepath), ek.ek(os.path.getmtime, filepath)) if ctime > time.time() - 60: return True return False def sanitizeFileName(name): ''' >>> sanitizeFileName('a/b/c') 'a-b-c' >>> sanitizeFileName('abc') 'abc' >>> sanitizeFileName('a"b') 'ab' >>> sanitizeFileName('.a.b..') 'a.b' ''' # remove bad chars from the filename name = re.sub(r'[\\/\*]', '-', name) name = re.sub(r'[:"<>|?]', '', name) name = re.sub(ur'\u2122', '', name) # Trade Mark Sign # remove leading/trailing periods and spaces name = name.strip(' .') return name def _remove_file_failed(file): try: ek.ek(os.remove, file) except: pass def findCertainShow(showList, indexerid): results = [] if not isinstance(indexerid, list): indexerid = [indexerid] if showList and len(indexerid): results = filter(lambda x: int(x.indexerid) in indexerid, showList) if len(results) == 1: return results[0] elif len(results) > 1: raise MultipleShowObjectsException() def makeDir(path): if not ek.ek(os.path.isdir, path): try: ek.ek(os.makedirs, path) # do the library update for synoindex notifiers.synoindex_notifier.addFolder(path) except OSError: return False return True def searchDBForShow(regShowName, log=False): showNames = [re.sub('[. -]', ' ', regShowName)] yearRegex = "([^()]+?)\s*(\()?(\d{4})(?(2)\))$" myDB = db.DBConnection() for showName in showNames: sqlResults = myDB.select("SELECT * FROM tv_shows WHERE show_name LIKE ?", [showName]) if len(sqlResults) == 1: return int(sqlResults[0]["indexer_id"]) else: # if we didn't get exactly one result then try again with the year stripped off if possible match = re.match(yearRegex, showName) if match and match.group(1): if log: logger.log(u"Unable to match original name but trying to manually strip and specify show year", logger.DEBUG) sqlResults = myDB.select( "SELECT * FROM tv_shows WHERE (show_name LIKE ?) AND startyear = ?", [match.group(1) + '%', match.group(3)]) if len(sqlResults) == 0: if log: logger.log(u"Unable to match a record in the DB for " + showName, logger.DEBUG) continue elif len(sqlResults) > 1: if log: logger.log(u"Multiple results for " + showName + " in the DB, unable to match show name", logger.DEBUG) continue else: return int(sqlResults[0]["indexer_id"]) def searchIndexerForShowID(regShowName, indexer=None, indexer_id=None, ui=None): showNames = [re.sub('[. -]', ' ', regShowName)] # Query Indexers for each search term and build the list of results for i in sickbeard.indexerApi().indexers if not indexer else int(indexer or []): # Query Indexers for each search term and build the list of results lINDEXER_API_PARMS = sickbeard.indexerApi(i).api_params.copy() if ui is not None: lINDEXER_API_PARMS['custom_ui'] = ui t = sickbeard.indexerApi(i).indexer(**lINDEXER_API_PARMS) for name in showNames: logger.log(u"Trying to find " + name + " on " + sickbeard.indexerApi(i).name, logger.DEBUG) try: search = t[indexer_id] if indexer_id else t[name] except: continue try: seriesname = search[0]['seriesname'] except: seriesname = None try: series_id = search[0]['id'] except: series_id = None if not (seriesname and series_id): continue ShowObj = findCertainShow(sickbeard.showList, int(series_id)) #Check if we can find the show in our list (if not, it's not the right show) if (indexer_id is None) and (ShowObj is not None) and (ShowObj.indexerid == int(series_id)): return (seriesname, i, int(series_id)) elif (indexer_id is not None) and (int(indexer_id) == int(series_id)): return (seriesname, i, int(indexer_id)) if indexer: break return (None, None, None) def sizeof_fmt(num): ''' >>> sizeof_fmt(2) '2.0 bytes' >>> sizeof_fmt(1024) '1.0 KB' >>> sizeof_fmt(2048) '2.0 KB' >>> sizeof_fmt(2**20) '1.0 MB' >>> sizeof_fmt(1234567) '1.2 MB' ''' for x in ['bytes', 'KB', 'MB', 'GB', 'TB']: if num < 1024.0: return "%3.1f %s" % (num, x) num /= 1024.0 def listMediaFiles(path): if not dir or not ek.ek(os.path.isdir, path): return [] files = [] for curFile in ek.ek(os.listdir, path): fullCurFile = ek.ek(os.path.join, path, curFile) # if it's a folder do it recursively if ek.ek(os.path.isdir, fullCurFile) and not curFile.startswith('.') and not curFile == 'Extras': files += listMediaFiles(fullCurFile) elif isMediaFile(curFile): files.append(fullCurFile) return files def copyFile(srcFile, destFile): ek.ek(shutil.copyfile, srcFile, destFile) try: ek.ek(shutil.copymode, srcFile, destFile) except OSError: pass def moveFile(srcFile, destFile): try: ek.ek(shutil.move, srcFile, destFile) fixSetGroupID(destFile) except OSError: copyFile(srcFile, destFile) ek.ek(os.unlink, srcFile) def link(src, dst): if os.name == 'nt': import ctypes if ctypes.windll.kernel32.CreateHardLinkW(unicode(dst), unicode(src), 0) == 0: raise ctypes.WinError() else: os.link(src, dst) def hardlinkFile(srcFile, destFile): try: ek.ek(link, srcFile, destFile) fixSetGroupID(destFile) except Exception, e: logger.log(u"Failed to create hardlink of " + srcFile + " at " + destFile + ": " + ex(e) + ". Copying instead", logger.ERROR) copyFile(srcFile, destFile) def symlink(src, dst): if os.name == 'nt': import ctypes if ctypes.windll.kernel32.CreateSymbolicLinkW(unicode(dst), unicode(src), 1 if os.path.isdir(src) else 0) in [0, 1280]: raise ctypes.WinError() else: os.symlink(src, dst) def moveAndSymlinkFile(srcFile, destFile): try: ek.ek(shutil.move, srcFile, destFile) fixSetGroupID(destFile) ek.ek(symlink, destFile, srcFile) except: logger.log(u"Failed to create symlink of " + srcFile + " at " + destFile + ". Copying instead", logger.ERROR) copyFile(srcFile, destFile) def make_dirs(path): """ Creates any folders that are missing and assigns them the permissions of their parents """ logger.log(u"Checking if the path " + path + " already exists", logger.DEBUG) if not ek.ek(os.path.isdir, path): # Windows, create all missing folders if os.name == 'nt' or os.name == 'ce': try: logger.log(u"Folder " + path + " didn't exist, creating it", logger.DEBUG) ek.ek(os.makedirs, path) except (OSError, IOError), e: logger.log(u"Failed creating " + path + " : " + ex(e), logger.ERROR) return False # not Windows, create all missing folders and set permissions else: sofar = '' folder_list = path.split(os.path.sep) # look through each subfolder and make sure they all exist for cur_folder in folder_list: sofar += cur_folder + os.path.sep # if it exists then just keep walking down the line if ek.ek(os.path.isdir, sofar): continue try: logger.log(u"Folder " + sofar + " didn't exist, creating it", logger.DEBUG) ek.ek(os.mkdir, sofar) # use normpath to remove end separator, otherwise checks permissions against itself chmodAsParent(ek.ek(os.path.normpath, sofar)) # do the library update for synoindex notifiers.synoindex_notifier.addFolder(sofar) except (OSError, IOError), e: logger.log(u"Failed creating " + sofar + " : " + ex(e), logger.ERROR) return False return True def rename_ep_file(cur_path, new_path, old_path_length=0): """ Creates all folders needed to move a file to its new location, renames it, then cleans up any folders left that are now empty. cur_path: The absolute path to the file you want to move/rename new_path: The absolute path to the destination for the file WITHOUT THE EXTENSION old_path_length: The length of media file path (old name) WITHOUT THE EXTENSION """ new_dest_dir, new_dest_name = os.path.split(new_path) # @UnusedVariable if old_path_length == 0 or old_path_length > len(cur_path): # approach from the right cur_file_name, cur_file_ext = os.path.splitext(cur_path) # @UnusedVariable else: # approach from the left cur_file_ext = cur_path[old_path_length:] cur_file_name = cur_path[:old_path_length] if cur_file_ext[1:] in subtitleExtensions: # Extract subtitle language from filename sublang = os.path.splitext(cur_file_name)[1][1:] # Check if the language extracted from filename is a valid language if isValidLanguage(sublang): cur_file_ext = '.' + sublang + cur_file_ext # put the extension on the incoming file new_path += cur_file_ext make_dirs(os.path.dirname(new_path)) # move the file try: logger.log(u"Renaming file from " + cur_path + " to " + new_path) ek.ek(shutil.move, cur_path, new_path) except (OSError, IOError), e: logger.log(u"Failed renaming " + cur_path + " to " + new_path + ": " + ex(e), logger.ERROR) return False # clean up any old folders that are empty delete_empty_folders(ek.ek(os.path.dirname, cur_path)) return True def delete_empty_folders(check_empty_dir, keep_dir=None): """ Walks backwards up the path and deletes any empty folders found. check_empty_dir: The path to clean (absolute path to a folder) keep_dir: Clean until this path is reached """ # treat check_empty_dir as empty when it only contains these items ignore_items = [] logger.log(u"Trying to clean any empty folders under " + check_empty_dir) # as long as the folder exists and doesn't contain any files, delete it while ek.ek(os.path.isdir, check_empty_dir) and check_empty_dir != keep_dir: check_files = ek.ek(os.listdir, check_empty_dir) if not check_files or (len(check_files) <= len(ignore_items) and all( [check_file in ignore_items for check_file in check_files])): # directory is empty or contains only ignore_items try: logger.log(u"Deleting empty folder: " + check_empty_dir) # need shutil.rmtree when ignore_items is really implemented ek.ek(os.rmdir, check_empty_dir) # do the library update for synoindex notifiers.synoindex_notifier.deleteFolder(check_empty_dir) except OSError, e: logger.log(u"Unable to delete " + check_empty_dir + ": " + repr(e) + " / " + str(e), logger.WARNING) break check_empty_dir = ek.ek(os.path.dirname, check_empty_dir) else: break def fileBitFilter(mode): for bit in [stat.S_IXUSR, stat.S_IXGRP, stat.S_IXOTH, stat.S_ISUID, stat.S_ISGID]: if mode & bit: mode -= bit return mode def chmodAsParent(childPath): if os.name == 'nt' or os.name == 'ce': return parentPath = ek.ek(os.path.dirname, childPath) if not parentPath: logger.log(u"No parent path provided in " + childPath + ", unable to get permissions from it", logger.DEBUG) return parentPathStat = ek.ek(os.stat, parentPath) parentMode = stat.S_IMODE(parentPathStat[stat.ST_MODE]) childPathStat = ek.ek(os.stat, childPath) childPath_mode = stat.S_IMODE(childPathStat[stat.ST_MODE]) if ek.ek(os.path.isfile, childPath): childMode = fileBitFilter(parentMode) else: childMode = parentMode if childPath_mode == childMode: return childPath_owner = childPathStat.st_uid user_id = os.geteuid() # @UndefinedVariable - only available on UNIX if user_id != 0 and user_id != childPath_owner: logger.log(u"Not running as root or owner of " + childPath + ", not trying to set permissions", logger.DEBUG) return try: ek.ek(os.chmod, childPath, childMode) logger.log(u"Setting permissions for %s to %o as parent directory has %o" % (childPath, childMode, parentMode), logger.DEBUG) except OSError: logger.log(u"Failed to set permission for %s to %o" % (childPath, childMode), logger.ERROR) def fixSetGroupID(childPath): if os.name == 'nt' or os.name == 'ce': return parentPath = ek.ek(os.path.dirname, childPath) parentStat = ek.ek(os.stat, parentPath) parentMode = stat.S_IMODE(parentStat[stat.ST_MODE]) if parentMode & stat.S_ISGID: parentGID = parentStat[stat.ST_GID] childStat = ek.ek(os.stat, childPath) childGID = childStat[stat.ST_GID] if childGID == parentGID: return childPath_owner = childStat.st_uid user_id = os.geteuid() # @UndefinedVariable - only available on UNIX if user_id != 0 and user_id != childPath_owner: logger.log(u"Not running as root or owner of " + childPath + ", not trying to set the set-group-ID", logger.DEBUG) return try: ek.ek(os.chown, childPath, -1, parentGID) # @UndefinedVariable - only available on UNIX logger.log(u"Respecting the set-group-ID bit on the parent directory for %s" % (childPath), logger.DEBUG) except OSError: logger.log( u"Failed to respect the set-group-ID bit on the parent directory for %s (setting group ID %i)" % ( childPath, parentGID), logger.ERROR) def is_anime_in_show_list(): for show in sickbeard.showList: if show.is_anime: return True return False def update_anime_support(): sickbeard.ANIMESUPPORT = is_anime_in_show_list() def get_absolute_number_from_season_and_episode(show, season, episode): absolute_number = None if season and episode: myDB = db.DBConnection() sql = "SELECT * FROM tv_episodes WHERE showid = ? and season = ? and episode = ?" sqlResults = myDB.select(sql, [show.indexerid, season, episode]) if len(sqlResults) == 1: absolute_number = int(sqlResults[0]["absolute_number"]) logger.log( "Found absolute_number:" + str(absolute_number) + " by " + str(season) + "x" + str(episode), logger.DEBUG) else: logger.log( "No entries for absolute number in show: " + show.name + " found using " + str(season) + "x" + str( episode), logger.DEBUG) return absolute_number def get_all_episodes_from_absolute_number(show, absolute_numbers, indexer_id=None): episodes = [] season = None if len(absolute_numbers): if not show and indexer_id: show = findCertainShow(sickbeard.showList, indexer_id) for absolute_number in absolute_numbers if show else []: ep = show.getEpisode(None, None, absolute_number=absolute_number) if ep: episodes.append(ep.episode) season = ep.season # this will always take the last found seson so eps that cross the season border are not handeled well return (season, episodes) def sanitizeSceneName(name, anime=False): """ Takes a show name and returns the "scenified" version of it. anime: Some show have a ' in their name(Kuroko's Basketball) and is needed for search. Returns: A string containing the scene version of the show name given. """ if not name: return '' bad_chars = u',:()!?\u2019' if not anime: bad_chars += u"'" # strip out any bad chars for x in bad_chars: name = name.replace(x, "") # tidy up stuff that doesn't belong in scene names name = name.replace("- ", ".").replace(" ", ".").replace("&", "and").replace('/', '.') name = re.sub("\.\.*", ".", name) if name.endswith('.'): name = name[:-1] return name _binOps = { ast.Add: operator.add, ast.Sub: operator.sub, ast.Mult: operator.mul, ast.Div: operator.div, ast.Mod: operator.mod } def arithmeticEval(s): """ A safe eval supporting basic arithmetic operations. :param s: expression to evaluate :return: value """ node = ast.parse(s, mode='eval') def _eval(node): if isinstance(node, ast.Expression): return _eval(node.body) elif isinstance(node, ast.Str): return node.s elif isinstance(node, ast.Num): return node.n elif isinstance(node, ast.BinOp): return _binOps[type(node.op)](_eval(node.left), _eval(node.right)) else: raise Exception('Unsupported type {}'.format(node)) return _eval(node.body) def create_https_certificates(ssl_cert, ssl_key): """ Create self-signed HTTPS certificares and store in paths 'ssl_cert' and 'ssl_key' """ try: from OpenSSL import crypto # @UnresolvedImport from lib.certgen import createKeyPair, createCertRequest, createCertificate, TYPE_RSA, \ serial # @UnresolvedImport except Exception, e: logger.log(u"pyopenssl module missing, please install for https access", logger.WARNING) return False # Create the CA Certificate cakey = createKeyPair(TYPE_RSA, 1024) careq = createCertRequest(cakey, CN='Certificate Authority') cacert = createCertificate(careq, (careq, cakey), serial, (0, 60 * 60 * 24 * 365 * 10)) # ten years cname = 'SickRage' pkey = createKeyPair(TYPE_RSA, 1024) req = createCertRequest(pkey, CN=cname) cert = createCertificate(req, (cacert, cakey), serial, (0, 60 * 60 * 24 * 365 * 10)) # ten years # Save the key and certificate to disk try: open(ssl_key, 'w').write(crypto.dump_privatekey(crypto.FILETYPE_PEM, pkey)) open(ssl_cert, 'w').write(crypto.dump_certificate(crypto.FILETYPE_PEM, cert)) except: logger.log(u"Error creating SSL key and certificate", logger.ERROR) return False return True def backupVersionedFile(old_file, version): numTries = 0 new_file = old_file + '.' + 'v' + str(version) while not ek.ek(os.path.isfile, new_file): if not ek.ek(os.path.isfile, old_file): logger.log(u"Not creating backup, " + old_file + " doesn't exist", logger.DEBUG) break try: logger.log(u"Trying to back up " + old_file + " to " + new_file, logger.DEBUG) shutil.copy(old_file, new_file) logger.log(u"Backup done", logger.DEBUG) break except Exception, e: logger.log(u"Error while trying to back up " + old_file + " to " + new_file + " : " + ex(e), logger.WARNING) numTries += 1 time.sleep(1) logger.log(u"Trying again.", logger.DEBUG) if numTries >= 10: logger.log(u"Unable to back up " + old_file + " to " + new_file + " please do it manually.", logger.ERROR) return False return True def restoreVersionedFile(backup_file, version): numTries = 0 new_file, backup_version = os.path.splitext(backup_file) restore_file = new_file + '.' + 'v' + str(version) if not ek.ek(os.path.isfile, new_file): logger.log(u"Not restoring, " + new_file + " doesn't exist", logger.DEBUG) return False try: logger.log( u"Trying to backup " + new_file + " to " + new_file + "." + "r" + str(version) + " before restoring backup", logger.DEBUG) shutil.move(new_file, new_file + '.' + 'r' + str(version)) except Exception, e: logger.log( u"Error while trying to backup DB file " + restore_file + " before proceeding with restore: " + ex(e), logger.WARNING) return False while not ek.ek(os.path.isfile, new_file): if not ek.ek(os.path.isfile, restore_file): logger.log(u"Not restoring, " + restore_file + " doesn't exist", logger.DEBUG) break try: logger.log(u"Trying to restore " + restore_file + " to " + new_file, logger.DEBUG) shutil.copy(restore_file, new_file) logger.log(u"Restore done", logger.DEBUG) break except Exception, e: logger.log(u"Error while trying to restore " + restore_file + ": " + ex(e), logger.WARNING) numTries += 1 time.sleep(1) logger.log(u"Trying again.", logger.DEBUG) if numTries >= 10: logger.log(u"Unable to restore " + restore_file + " to " + new_file + " please do it manually.", logger.ERROR) return False return True # try to convert to int, if it fails the default will be returned def tryInt(s, s_default=0): try: return int(s) except: return s_default # generates a md5 hash of a file def md5_for_file(filename, block_size=2 ** 16): try: with open(filename, 'rb') as f: md5 = hashlib.md5() while True: data = f.read(block_size) if not data: break md5.update(data) f.close() return md5.hexdigest() except Exception: return None def get_lan_ip(): try:return [ip for ip in socket.gethostbyname_ex(socket.gethostname())[2] if not ip.startswith("127.")][0] except:return socket.gethostname() def check_url(url): """ Check if a URL exists without downloading the whole file. We only check the URL header. """ # see also http://stackoverflow.com/questions/2924422 # http://stackoverflow.com/questions/1140661 good_codes = [httplib.OK, httplib.FOUND, httplib.MOVED_PERMANENTLY] host, path = urlparse.urlparse(url)[1:3] # elems [1] and [2] try: conn = httplib.HTTPConnection(host) conn.request('HEAD', path) return conn.getresponse().status in good_codes except StandardError: return None def anon_url(*url): """ Return a URL string consisting of the Anonymous redirect URL and an arbitrary number of values appended. """ return '' if None in url else '%s%s' % (sickbeard.ANON_REDIRECT, ''.join(str(s) for s in url)) """ Encryption ========== By Pedro Jose Pereira Vieito <pvieito@gmail.com> (@pvieito) * If encryption_version==0 then return data without encryption * The keys should be unique for each device To add a new encryption_version: 1) Code your new encryption_version 2) Update the last encryption_version available in webserve.py 3) Remember to maintain old encryption versions and key generators for retrocompatibility """ # Key Generators unique_key1 = hex(uuid.getnode() ** 2) # Used in encryption v1 # Encryption Functions def encrypt(data, encryption_version=0, decrypt=False): # Version 1: Simple XOR encryption (this is not very secure, but works) if encryption_version == 1: if decrypt: return ''.join(chr(ord(x) ^ ord(y)) for (x, y) in izip(base64.decodestring(data), cycle(unique_key1))) else: return base64.encodestring( ''.join(chr(ord(x) ^ ord(y)) for (x, y) in izip(data, cycle(unique_key1)))).strip() # Version 2: Simple XOR encryption (this is not very secure, but works) elif encryption_version == 2: if decrypt: return ''.join(chr(ord(x) ^ ord(y)) for (x, y) in izip(base64.decodestring(data), cycle(sickbeard.ENCRYPTION_SECRET))) else: return base64.encodestring( ''.join(chr(ord(x) ^ ord(y)) for (x, y) in izip(data, cycle(sickbeard.ENCRYPTION_SECRET)))).strip() # Version 0: Plain text else: return data def decrypt(data, encryption_version=0): return encrypt(data, encryption_version, decrypt=True) def full_sanitizeSceneName(name): return re.sub('[. -]', ' ', sanitizeSceneName(name)).lower().lstrip() def _check_against_names(nameInQuestion, show, season=-1): showNames = [] if season in [-1, 1]: showNames = [show.name] showNames.extend(sickbeard.scene_exceptions.get_scene_exceptions(show.indexerid, season=season)) for showName in showNames: nameFromList = full_sanitizeSceneName(showName) if nameFromList == nameInQuestion: return True return False def get_show(name, tryIndexers=False, trySceneExceptions=False): if not sickbeard.showList: return showObj = None fromCache = False if not name: return showObj try: # check cache for show cache = sickbeard.name_cache.retrieveNameFromCache(name) if cache: fromCache = True showObj = findCertainShow(sickbeard.showList, int(cache)) #try indexers if not showObj and tryIndexers: showObj = findCertainShow(sickbeard.showList, searchIndexerForShowID(full_sanitizeSceneName(name), ui=classes.ShowListUI)[2]) #try scene exceptions if not showObj and trySceneExceptions: ShowID = sickbeard.scene_exceptions.get_scene_exception_by_name(name)[0] if ShowID: showObj = findCertainShow(sickbeard.showList, int(ShowID)) # add show to cache if showObj and not fromCache: sickbeard.name_cache.addNameToCache(name, showObj.indexerid) except Exception as e: logger.log(u"Error when attempting to find show: " + name + " in SickRage: " + str(e), logger.DEBUG) return showObj def is_hidden_folder(folder): """ Returns True if folder is hidden. On Linux based systems hidden folders start with . (dot) folder: Full path of folder to check """ def is_hidden(filepath): name = os.path.basename(os.path.abspath(filepath)) return name.startswith('.') or has_hidden_attribute(filepath) def has_hidden_attribute(filepath): try: attrs = ctypes.windll.kernel32.GetFileAttributesW(unicode(filepath)) assert attrs != -1 result = bool(attrs & 2) except (AttributeError, AssertionError): result = False return result if ek.ek(os.path.isdir, folder): if is_hidden(folder): return True return False def real_path(path): """ Returns: the canonicalized absolute pathname. The resulting path will have no symbolic link, '/./' or '/../' components. """ return ek.ek(os.path.normpath, ek.ek(os.path.normcase, ek.ek(os.path.realpath, path))) def validateShow(show, season=None, episode=None): indexer_lang = show.lang try: lINDEXER_API_PARMS = sickbeard.indexerApi(show.indexer).api_params.copy() if indexer_lang and not indexer_lang == sickbeard.INDEXER_DEFAULT_LANGUAGE: lINDEXER_API_PARMS['language'] = indexer_lang if show.dvdorder != 0: lINDEXER_API_PARMS['dvdorder'] = True t = sickbeard.indexerApi(show.indexer).indexer(**lINDEXER_API_PARMS) if season is None and episode is None: return t return t[show.indexerid][season][episode] except (sickbeard.indexer_episodenotfound, sickbeard.indexer_seasonnotfound): pass def set_up_anidb_connection(): if not sickbeard.USE_ANIDB: logger.log(u"Usage of anidb disabled. Skiping", logger.DEBUG) return False if not sickbeard.ANIDB_USERNAME and not sickbeard.ANIDB_PASSWORD: logger.log(u"anidb username and/or password are not set. Aborting anidb lookup.", logger.DEBUG) return False if not sickbeard.ADBA_CONNECTION: anidb_logger = lambda x: logger.log("ANIDB: " + str(x), logger.DEBUG) try: sickbeard.ADBA_CONNECTION = adba.Connection(keepAlive=True, log=anidb_logger) except Exception as e: logger.log(u"anidb exception msg: " + str(e)) return False try: if not sickbeard.ADBA_CONNECTION.authed(): sickbeard.ADBA_CONNECTION.auth(sickbeard.ANIDB_USERNAME, sickbeard.ANIDB_PASSWORD) else: return True except Exception as e: logger.log(u"anidb exception msg: " + str(e)) return False return sickbeard.ADBA_CONNECTION.authed() def makeZip(fileList, archive): """ 'fileList' is a list of file names - full path each name 'archive' is the file name for the archive with a full path """ try: a = zipfile.ZipFile(archive, 'w', zipfile.ZIP_DEFLATED) for f in fileList: a.write(f) a.close() return True except Exception as e: logger.log(u"Zip creation error: " + str(e), logger.ERROR) return False def extractZip(archive, targetDir): """ 'fileList' is a list of file names - full path each name 'archive' is the file name for the archive with a full path """ try: if not os.path.exists(targetDir): os.mkdir(targetDir) zip_file = zipfile.ZipFile(archive, 'r') for member in zip_file.namelist(): filename = os.path.basename(member) # skip directories if not filename: continue # copy file (taken from zipfile's extract) source = zip_file.open(member) target = file(os.path.join(targetDir, filename), "wb") shutil.copyfileobj(source, target) source.close() target.close() zip_file.close() return True except Exception as e: logger.log(u"Zip extraction error: " + str(e), logger.ERROR) return False def backupConfigZip(fileList, archive, arcname = None): try: a = zipfile.ZipFile(archive, 'w', zipfile.ZIP_DEFLATED) for f in fileList: a.write(f, os.path.relpath(f, arcname)) a.close() return True except Exception as e: logger.log(u"Zip creation error: " + str(e), logger.ERROR) return False def restoreConfigZip(archive, targetDir): import ntpath try: if not os.path.exists(targetDir): os.mkdir(targetDir) else: def path_leaf(path): head, tail = ntpath.split(path) return tail or ntpath.basename(head) bakFilename = '{0}-{1}'.format(path_leaf(targetDir), datetime.datetime.strftime(datetime.datetime.now(), '%Y%m%d_%H%M%S')) shutil.move(targetDir, os.path.join(ntpath.dirname(targetDir), bakFilename)) zip_file = zipfile.ZipFile(archive, 'r') for member in zip_file.namelist(): zip_file.extract(member, targetDir) zip_file.close() return True except Exception as e: logger.log(u"Zip extraction error: " + str(e), logger.ERROR) shutil.rmtree(targetDir) return False def mapIndexersToShow(showObj): mapped = {} # init mapped indexers object for indexer in sickbeard.indexerApi().indexers: mapped[indexer] = showObj.indexerid if int(indexer) == int(showObj.indexer) else 0 myDB = db.DBConnection() sqlResults = myDB.select( "SELECT * FROM indexer_mapping WHERE indexer_id = ? AND indexer = ?", [showObj.indexerid, showObj.indexer]) # for each mapped entry for curResult in sqlResults: nlist = [i for i in curResult if i is not None] # Check if its mapped with both tvdb and tvrage. if len(nlist) >= 4: logger.log(u"Found indexer mapping in cache for show: " + showObj.name, logger.DEBUG) mapped[int(curResult['mindexer'])] = int(curResult['mindexer_id']) return mapped else: sql_l = [] for indexer in sickbeard.indexerApi().indexers: if indexer == showObj.indexer: mapped[indexer] = showObj.indexerid continue lINDEXER_API_PARMS = sickbeard.indexerApi(indexer).api_params.copy() lINDEXER_API_PARMS['custom_ui'] = classes.ShowListUI t = sickbeard.indexerApi(indexer).indexer(**lINDEXER_API_PARMS) try: mapped_show = t[showObj.name] except Exception: logger.log(u"Unable to map " + sickbeard.indexerApi(showObj.indexer).name + "->" + sickbeard.indexerApi( indexer).name + " for show: " + showObj.name + ", skipping it", logger.DEBUG) continue if mapped_show and len(mapped_show) == 1: logger.log(u"Mapping " + sickbeard.indexerApi(showObj.indexer).name + "->" + sickbeard.indexerApi( indexer).name + " for show: " + showObj.name, logger.DEBUG) mapped[indexer] = int(mapped_show[0]['id']) logger.log(u"Adding indexer mapping to DB for show: " + showObj.name, logger.DEBUG) sql_l.append([ "INSERT OR IGNORE INTO indexer_mapping (indexer_id, indexer, mindexer_id, mindexer) VALUES (?,?,?,?)", [showObj.indexerid, showObj.indexer, int(mapped_show[0]['id']), indexer]]) if len(sql_l) > 0: myDB = db.DBConnection() myDB.mass_action(sql_l) return mapped def touchFile(fname, atime=None): if None != atime: try: with file(fname, 'a'): os.utime(fname, (atime, atime)) return True except Exception as e: if e.errno == errno.ENOSYS: logger.log(u"File air date stamping not available on your OS", logger.DEBUG) elif e.errno == errno.EACCES: logger.log(u"File air date stamping failed(Permission denied). Check permissions for file: {0}".format(fname), logger.ERROR) else: logger.log(u"File air date stamping failed. The error is: {0} and the message is: {1}.".format(e.errno, e.strerror), logger.ERROR) pass return False def _getTempDir(): import getpass """Returns the [system temp dir]/tvdb_api-u501 (or tvdb_api-myuser) """ if hasattr(os, 'getuid'): uid = "u%d" % (os.getuid()) else: # For Windows try: uid = getpass.getuser() except ImportError: return os.path.join(tempfile.gettempdir(), "sickrage") return os.path.join(tempfile.gettempdir(), "sickrage-%s" % (uid)) def codeDescription(status_code): """ Returns the description of the URL error code """ if status_code in clients.http_error_code: return clients.http_error_code[status_code] else: logger.log(u"Unknown error code. Please submit an issue", logger.WARNING) return 'unknown' def _setUpSession(session, headers): """ Returns a session initialized with default cache and parameter settings """ # request session cache_dir = sickbeard.CACHE_DIR or _getTempDir() session = CacheControl(sess=session, cache=caches.FileCache(os.path.join(cache_dir, 'sessions')), cache_etags=False) # request session clear residual referer if 'Referer' in session.headers and not 'Referer' in headers: session.headers.pop('Referer') # request session headers session.headers.update({'User-Agent': USER_AGENT, 'Accept-Encoding': 'gzip,deflate'}) session.headers.update(headers) # request session ssl verify session.verify = certifi.where() # request session proxies if not 'Referer' in session.headers and sickbeard.PROXY_SETTING: logger.log("Using proxy: " + sickbeard.PROXY_SETTING, logger.DEBUG) scheme, address = urllib2.splittype(sickbeard.PROXY_SETTING) address = sickbeard.PROXY_SETTING if scheme else 'http://' + sickbeard.PROXY_SETTING session.proxies = { "http": address, "https": address, } session.headers.update({'Referer': address}) if 'Content-Type' in session.headers: session.headers.pop('Content-Type') return session def headURL(url, params=None, headers={}, timeout=30, session=None, json=False, proxyGlypeProxySSLwarning=None): """ Checks if URL is valid, without reading it """ session = _setUpSession(session, headers) session.params = params try: resp = session.head(url, timeout=timeout, allow_redirects=True) if not resp.ok: logger.log(u"Requested url " + url + " returned status code is " + str( resp.status_code) + ': ' + codeDescription(resp.status_code), logger.DEBUG) return False if proxyGlypeProxySSLwarning is not None: if re.search('The site you are attempting to browse is on a secure connection', resp.text): resp = session.head(proxyGlypeProxySSLwarning, timeout=timeout, allow_redirects=True) if not resp.ok: logger.log(u"GlypeProxySSLwarning: Requested headURL " + url + " returned status code is " + str( resp.status_code) + ': ' + codeDescription(resp.status_code), logger.DEBUG) return False return resp.status_code == 200 except requests.exceptions.HTTPError, e: logger.log(u"HTTP error in headURL {0}. Error: {1}".format(url,e.errno), logger.WARNING) except requests.exceptions.ConnectionError, e: logger.log(u"Connection error to {0}. Error: {1}".format(url,e.message), logger.WARNING) except requests.exceptions.Timeout, e: logger.log(u"Connection timed out accessing {0}. Error: {1}".format(url,e.message), logger.WARNING) except Exception as e: logger.log(u"Unknown exception in headURL {0}. Error: {1}".format(url,e.message), logger.WARNING) logger.log(traceback.format_exc(), logger.WARNING) return False def getURL(url, post_data=None, params={}, headers={}, timeout=30, session=None, json=False, proxyGlypeProxySSLwarning=None): """ Returns a byte-string retrieved from the url provider. """ session = _setUpSession(session, headers) session.params = params try: # decide if we get or post data to server if post_data: session.headers.update({'Content-Type': 'application/x-www-form-urlencoded'}) resp = session.post(url, data=post_data, timeout=timeout, allow_redirects=True) else: resp = session.get(url, timeout=timeout, allow_redirects=True) if not resp.ok: logger.log(u"Requested url " + url + " returned status code is " + str( resp.status_code) + ': ' + codeDescription(resp.status_code), logger.DEBUG) return if proxyGlypeProxySSLwarning is not None: if re.search('The site you are attempting to browse is on a secure connection', resp.text): resp = session.get(proxyGlypeProxySSLwarning, timeout=timeout, allow_redirects=True) if not resp.ok: logger.log(u"GlypeProxySSLwarning: Requested url " + url + " returned status code is " + str( resp.status_code) + ': ' + codeDescription(resp.status_code), logger.DEBUG) return except requests.exceptions.HTTPError, e: logger.log(u"HTTP error in getURL {0}. Error: {1}".format(url,e.errno), logger.WARNING) return except requests.exceptions.ConnectionError, e: logger.log(u"Connection error to {0}. Error: {1}".format(url,e.message), logger.WARNING) return except requests.exceptions.Timeout, e: logger.log(u"Connection timed out accessing {0}. Error: {1}".format(url,e.message), logger.WARNING) return except Exception as e: logger.log(u"Unknown exception in getURL {0}. Error: {1}".format(url,e.message), logger.WARNING) logger.log(traceback.format_exc(), logger.WARNING) return return resp.content if not json else resp.json() def download_file(url, filename, session=None, headers={}): session = _setUpSession(session, headers) session.stream = True try: with closing(session.get(url, allow_redirects=True)) as resp: if not resp.ok: logger.log(u"Requested url " + url + " returned status code is " + str( resp.status_code) + ': ' + codeDescription(resp.status_code), logger.DEBUG) return False try: with open(filename, 'wb') as fp: for chunk in resp.iter_content(chunk_size=1024): if chunk: fp.write(chunk) fp.flush() chmodAsParent(filename) except: logger.log(u"Problem setting permissions or writing file to: %s" % filename, logger.WARNING) except requests.exceptions.HTTPError, e: _remove_file_failed(filename) logger.log(u"HTTP error " + str(e.errno) + " while loading URL " + url, logger.WARNING) return False except requests.exceptions.ConnectionError, e: _remove_file_failed(filename) logger.log(u"Connection error " + str(e.message) + " while loading URL " + url, logger.WARNING) return False except requests.exceptions.Timeout, e: _remove_file_failed(filename) logger.log(u"Connection timed out " + str(e.message) + " while loading URL " + url, logger.WARNING) return False except EnvironmentError, e: _remove_file_failed(filename) logger.log(u"Unable to save the file: " + ex(e), logger.WARNING) return False except Exception: _remove_file_failed(filename) logger.log(u"Unknown exception while loading URL " + url + ": " + traceback.format_exc(), logger.WARNING) return False return True def get_size(start_path='.'): total_size = 0 for dirpath, dirnames, filenames in ek.ek(os.walk, start_path): for f in filenames: fp = ek.ek(os.path.join, dirpath, f) try: total_size += ek.ek(os.path.getsize, fp) except OSError as e: logger.log('Unable to get size for file {filePath}. Error msg is: {errorMsg}'.format(filePath=fp, errorMsg=str(e)), logger.ERROR) logger.log(traceback.format_exc(), logger.DEBUG) return total_size def generateApiKey(): """ Return a new randomized API_KEY """ try: from hashlib import md5 except ImportError: from md5 import md5 # Create some values to seed md5 t = str(time.time()) r = str(random.random()) # Create the md5 instance and give it the current time m = md5(t) # Update the md5 instance with the random variable m.update(r) # Return a hex digest of the md5, eg 49f68a5c8493ec2c0bf489821c21fc3b logger.log(u"New API generated") return m.hexdigest() def pretty_filesize(file_bytes): file_bytes = float(file_bytes) if file_bytes >= 1099511627776: terabytes = file_bytes / 1099511627776 size = '%.2f TB' % terabytes elif file_bytes >= 1073741824: gigabytes = file_bytes / 1073741824 size = '%.2f GB' % gigabytes elif file_bytes >= 1048576: megabytes = file_bytes / 1048576 size = '%.2f MB' % megabytes elif file_bytes >= 1024: kilobytes = file_bytes / 1024 size = '%.2f KB' % kilobytes else: size = '%.2f b' % file_bytes return size if __name__ == '__main__': import doctest doctest.testmod() def remove_article(text=''): return re.sub(r'(?i)^(?:(?:A(?!\s+to)n?)|The)\s(\w)', r'\1', text) def generateCookieSecret(): return base64.b64encode(uuid.uuid4().bytes + uuid.uuid4().bytes) def verify_freespace(src, dest, oldfile=None): """ Checks if the target system has enough free space to copy or move a file, Returns true if there is, False if there isn't. Also returns True if the OS doesn't support this option """ if not isinstance(oldfile, list): oldfile = [oldfile] logger.log("Trying to determine free space on destination drive", logger.DEBUG) if hasattr(os, 'statvfs'): # POSIX def disk_usage(path): st = os.statvfs(path) free = st.f_bavail * st.f_frsize return free elif os.name == 'nt': # Windows import sys def disk_usage(path): _, total, free = ctypes.c_ulonglong(), ctypes.c_ulonglong(), \ ctypes.c_ulonglong() if sys.version_info >= (3,) or isinstance(path, unicode): fun = ctypes.windll.kernel32.GetDiskFreeSpaceExW else: fun = ctypes.windll.kernel32.GetDiskFreeSpaceExA ret = fun(path, ctypes.byref(_), ctypes.byref(total), ctypes.byref(free)) if ret == 0: logger.log("Unable to determine free space, something went wrong", logger.WARNING) raise ctypes.WinError() return free.value else: logger.log("Unable to determine free space on your OS") return True if not ek.ek(os.path.isfile, src): logger.log("A path to a file is required for the source. " + src + " is not a file.", logger.WARNING) return True try: diskfree = disk_usage(dest) except: logger.log("Unable to determine free space, so I will assume there is enough.", logger.WARNING) return True neededspace = ek.ek(os.path.getsize, src) if oldfile: for file in oldfile: if ek.ek(os.path.isfile, file.location): diskfree += ek.ek(os.path.getsize, file.location) if diskfree > neededspace: return True else: logger.log("Not enough free space: Needed: " + str(neededspace) + " bytes (" + pretty_filesize(neededspace) + "), found: " + str(diskfree) + " bytes (" + pretty_filesize(diskfree) + ")", logger.WARNING) return False # https://gist.github.com/thatalextaylor/7408395 def pretty_time_delta(seconds): sign_string = '-' if seconds < 0 else '' seconds = abs(int(seconds)) days, seconds = divmod(seconds, 86400) hours, seconds = divmod(seconds, 3600) minutes, seconds = divmod(seconds, 60) if days > 0: return '%s%dd%02dh%02dm%02ds' % (sign_string, days, hours, minutes, seconds) elif hours > 0: return '%s%02dh%02dm%02ds' % (sign_string, hours, minutes, seconds) elif minutes > 0: return '%s%02dm%02ds' % (sign_string, minutes, seconds) else: return '%s%02ds' % (sign_string, seconds) def isFileLocked(file, writeLockCheck=False): ''' Checks to see if a file is locked. Performs three checks 1. Checks if the file even exists 2. Attempts to open the file for reading. This will determine if the file has a write lock. Write locks occur when the file is being edited or copied to, e.g. a file copy destination 3. If the readLockCheck parameter is True, attempts to rename the file. If this fails the file is open by some other process for reading. The file can be read, but not written to or deleted. @param file: the file being checked @param writeLockCheck: when true will check if the file is locked for writing (prevents move operations) ''' if not ek.ek(os.path.exists, file): return True try: f = ek.ek(open, file, 'r') f.close() except IOError: return True if(writeLockCheck): lockFile = file + ".lckchk" if ek.ek(os.path.exists, lockFile): ek.ek(os.remove, lockFile) try: ek.ek(os.rename, file, lockFile) time.sleep(1) ek.ek(os.rename, lockFile, file) except (OSError, IOError): return True return False