|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
from __future__ import absolute_import
|
|
|
|
import logging
|
|
|
|
import io
|
|
|
|
import os
|
|
|
|
import rarfile
|
|
|
|
import zipfile
|
|
|
|
|
|
|
|
from requests import Session
|
|
|
|
from guessit import guessit
|
|
|
|
from subliminal_patch.exceptions import ParseResponseError
|
|
|
|
from subliminal_patch.providers import Provider
|
|
|
|
from subliminal.providers import ParserBeautifulSoup
|
|
|
|
from subliminal_patch.subtitle import Subtitle
|
|
|
|
from subliminal.video import Episode
|
|
|
|
from subliminal.subtitle import SUBTITLE_EXTENSIONS, fix_line_ending,guess_matches
|
|
|
|
from subzero.language import Language
|
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
class LegendasdivxSubtitle(Subtitle):
|
|
|
|
"""Legendasdivx Subtitle."""
|
|
|
|
provider_name = 'legendasdivx'
|
|
|
|
|
|
|
|
def __init__(self, language, video, data):
|
|
|
|
super(LegendasdivxSubtitle, self).__init__(language)
|
|
|
|
self.language = language
|
|
|
|
self.page_link = data['link']
|
|
|
|
self.hits=data['hits']
|
|
|
|
self.exact_match=data['exact_match']
|
|
|
|
self.description=data['description'].lower()
|
|
|
|
self.video = video
|
|
|
|
self.videoname =data['videoname']
|
|
|
|
|
|
|
|
@property
|
|
|
|
def id(self):
|
|
|
|
return self.page_link
|
|
|
|
|
|
|
|
@property
|
|
|
|
def release_info(self):
|
|
|
|
return self.description
|
|
|
|
|
|
|
|
def get_matches(self, video):
|
|
|
|
matches = set()
|
|
|
|
|
|
|
|
if self.videoname.lower() in self.description:
|
|
|
|
matches.update(['title'])
|
|
|
|
matches.update(['season'])
|
|
|
|
matches.update(['episode'])
|
|
|
|
|
|
|
|
# episode
|
|
|
|
if video.title and video.title.lower() in self.description:
|
|
|
|
matches.update(['title'])
|
|
|
|
if video.year and '{:04d}'.format(video.year) in self.description:
|
|
|
|
matches.update(['year'])
|
|
|
|
|
|
|
|
if isinstance(video, Episode):
|
|
|
|
# already matched in search query
|
|
|
|
if video.season and 's{:02d}'.format(video.season) in self.description:
|
|
|
|
matches.update(['season'])
|
|
|
|
if video.episode and 'e{:02d}'.format(video.episode) in self.description:
|
|
|
|
matches.update(['episode'])
|
|
|
|
if video.episode and video.season and video.series:
|
|
|
|
if '{}.s{:02d}e{:02d}'.format(video.series.lower(),video.season,video.episode) in self.description:
|
|
|
|
matches.update(['series'])
|
|
|
|
matches.update(['season'])
|
|
|
|
matches.update(['episode'])
|
|
|
|
if '{} s{:02d}e{:02d}'.format(video.series.lower(),video.season,video.episode) in self.description:
|
|
|
|
matches.update(['series'])
|
|
|
|
matches.update(['season'])
|
|
|
|
matches.update(['episode'])
|
|
|
|
|
|
|
|
# release_group
|
|
|
|
if video.release_group and video.release_group.lower() in self.description:
|
|
|
|
matches.update(['release_group'])
|
|
|
|
|
|
|
|
# resolution
|
|
|
|
|
|
|
|
if video.resolution and video.resolution.lower() in self.description:
|
|
|
|
matches.update(['resolution'])
|
|
|
|
|
|
|
|
# format
|
|
|
|
formats = []
|
|
|
|
if video.format:
|
|
|
|
formats = [video.format.lower()]
|
|
|
|
if formats[0] == "web-dl":
|
|
|
|
formats.append("webdl")
|
|
|
|
formats.append("webrip")
|
|
|
|
formats.append("web ")
|
|
|
|
for frmt in formats:
|
|
|
|
if frmt.lower() in self.description:
|
|
|
|
matches.update(['format'])
|
|
|
|
break
|
|
|
|
|
|
|
|
# video_codec
|
|
|
|
if video.video_codec:
|
|
|
|
video_codecs = [video.video_codec.lower()]
|
|
|
|
if video_codecs[0] == "h264":
|
|
|
|
formats.append("x264")
|
|
|
|
elif video_codecs[0] == "h265":
|
|
|
|
formats.append("x265")
|
|
|
|
for vc in formats:
|
|
|
|
if vc.lower() in self.description:
|
|
|
|
matches.update(['video_codec'])
|
|
|
|
break
|
|
|
|
|
|
|
|
matches |= guess_matches(video, guessit(self.description))
|
|
|
|
return matches
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
class LegendasdivxProvider(Provider):
|
|
|
|
"""Legendasdivx Provider."""
|
|
|
|
languages = {Language('por', 'BR')} | {Language('por')}
|
|
|
|
SEARCH_THROTTLE = 8
|
|
|
|
site = 'https://www.legendasdivx.pt'
|
|
|
|
headers = {
|
|
|
|
'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64; rv:72.0) Gecko/20100101 Firefox/72.0',
|
|
|
|
'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8',
|
|
|
|
'Origin': 'https://www.legendasdivx.pt',
|
|
|
|
'Referer': 'https://www.legendasdivx.pt',
|
|
|
|
'Pragma': 'no-cache',
|
|
|
|
'Cache-Control': 'no-cache'
|
|
|
|
}
|
|
|
|
loginpage = site + '/forum/ucp.php?mode=login'
|
|
|
|
searchurl = site + '/modules.php?name=Downloads&file=jz&d_op=search&op=_jz00&query={query}'
|
|
|
|
language_list = list(languages)
|
|
|
|
|
|
|
|
|
|
|
|
def __init__(self, username, password):
|
|
|
|
self.username = username
|
|
|
|
self.password = password
|
|
|
|
|
|
|
|
def initialize(self):
|
|
|
|
self.session = Session()
|
|
|
|
self.login()
|
|
|
|
|
|
|
|
def terminate(self):
|
|
|
|
self.logout()
|
|
|
|
self.session.close()
|
|
|
|
|
|
|
|
def login(self):
|
|
|
|
logger.info('Logging in')
|
|
|
|
self.headers['Referer'] = self.site + '/index.php'
|
|
|
|
self.session.headers.update(self.headers.items())
|
|
|
|
res = self.session.get(self.loginpage)
|
|
|
|
bsoup = ParserBeautifulSoup(res.content, ['lxml'])
|
|
|
|
|
|
|
|
_allinputs = bsoup.findAll('input')
|
|
|
|
fields = {}
|
|
|
|
for field in _allinputs:
|
|
|
|
fields[field.get('name')] = field.get('value')
|
|
|
|
|
|
|
|
fields['username'] = self.username
|
|
|
|
fields['password'] = self.password
|
|
|
|
fields['autologin'] = 'on'
|
|
|
|
fields['viewonline'] = 'on'
|
|
|
|
|
|
|
|
self.headers['Referer'] = self.loginpage
|
|
|
|
self.session.headers.update(self.headers.items())
|
|
|
|
res = self.session.post(self.loginpage, fields)
|
|
|
|
try:
|
|
|
|
logger.debug('Got session id %s' %
|
|
|
|
self.session.cookies.get_dict()['PHPSESSID'])
|
|
|
|
except KeyError as e:
|
|
|
|
logger.error(repr(e))
|
|
|
|
logger.error("Didn't get session id, check your credentials")
|
|
|
|
return False
|
|
|
|
except Exception as e:
|
|
|
|
logger.error(repr(e))
|
|
|
|
logger.error('uncached error #legendasdivx #AA')
|
|
|
|
return False
|
|
|
|
|
|
|
|
return True
|
|
|
|
def logout(self):
|
|
|
|
# need to figure this out
|
|
|
|
return True
|
|
|
|
|
|
|
|
def query(self, video, language):
|
|
|
|
try:
|
|
|
|
logger.debug('Got session id %s' %
|
|
|
|
self.session.cookies.get_dict()['PHPSESSID'])
|
|
|
|
except Exception as e:
|
|
|
|
self.login()
|
|
|
|
return []
|
|
|
|
|
|
|
|
language_ids = '0'
|
|
|
|
if isinstance(language, (tuple, list, set)):
|
|
|
|
if len(language) == 1:
|
|
|
|
language_ids = ','.join(sorted(l.opensubtitles for l in language))
|
|
|
|
if language_ids == 'por':
|
|
|
|
language_ids = '&form_cat=28'
|
|
|
|
else:
|
|
|
|
language_ids = '&form_cat=29'
|
|
|
|
|
|
|
|
querytext = video.name
|
|
|
|
querytext = os.path.basename(querytext)
|
|
|
|
querytext, _ = os.path.splitext(querytext)
|
|
|
|
videoname = querytext
|
|
|
|
querytext = querytext.lower()
|
|
|
|
querytext = querytext.replace(
|
|
|
|
".", "+").replace("[", "").replace("]", "")
|
|
|
|
if language_ids != '0':
|
|
|
|
querytext = querytext + language_ids
|
|
|
|
self.headers['Referer'] = self.site + '/index.php'
|
|
|
|
self.session.headers.update(self.headers.items())
|
|
|
|
res = self.session.get(self.searchurl.format(query=querytext))
|
|
|
|
# form_cat=28 = br
|
|
|
|
# form_cat=29 = pt
|
|
|
|
if "A legenda não foi encontrada" in res.text:
|
|
|
|
logger.warning('%s not found', querytext)
|
|
|
|
return []
|
|
|
|
|
|
|
|
bsoup = ParserBeautifulSoup(res.content, ['html.parser'])
|
|
|
|
_allsubs = bsoup.findAll("div", {"class": "sub_box"})
|
|
|
|
subtitles = []
|
|
|
|
lang = Language.fromopensubtitles("pob")
|
|
|
|
for _subbox in _allsubs:
|
|
|
|
hits=0
|
|
|
|
for th in _subbox.findAll("th", {"class": "color2"}):
|
|
|
|
if th.string == 'Hits:':
|
|
|
|
hits = int(th.parent.find("td").string)
|
|
|
|
if th.string == 'Idioma:':
|
|
|
|
lang = th.parent.find("td").find ("img").get ('src')
|
|
|
|
if 'brazil' in lang:
|
|
|
|
lang = Language.fromopensubtitles('pob')
|
|
|
|
else:
|
|
|
|
lang = Language.fromopensubtitles('por')
|
|
|
|
|
|
|
|
|
|
|
|
description = _subbox.find("td", {"class": "td_desc brd_up"})
|
|
|
|
download = _subbox.find("a", {"class": "sub_download"})
|
|
|
|
try:
|
|
|
|
# sometimes BSoup just doesn't get the link
|
|
|
|
logger.debug(download.get('href'))
|
|
|
|
except Exception as e:
|
|
|
|
logger.warning('skipping subbox on %s' % self.searchurl.format(query=querytext))
|
|
|
|
continue
|
|
|
|
|
|
|
|
exact_match = False
|
|
|
|
if video.name.lower() in description.get_text().lower():
|
|
|
|
exact_match = True
|
|
|
|
data = {'link': self.site + '/modules.php' + download.get('href'),
|
|
|
|
'exact_match': exact_match,
|
|
|
|
'hits': hits,
|
|
|
|
'videoname': videoname,
|
|
|
|
'description': description.get_text() }
|
|
|
|
subtitles.append(
|
|
|
|
LegendasdivxSubtitle(lang, video, data)
|
|
|
|
)
|
|
|
|
|
|
|
|
return subtitles
|
|
|
|
|
|
|
|
def list_subtitles(self, video, languages):
|
|
|
|
return self.query(video, languages)
|
|
|
|
|
|
|
|
def download_subtitle(self, subtitle):
|
|
|
|
res = self.session.get(subtitle.page_link)
|
|
|
|
if res:
|
|
|
|
if res.text == '500':
|
|
|
|
raise ValueError('Error 500 on server')
|
|
|
|
|
|
|
|
archive = self._get_archive(res.content)
|
|
|
|
# extract the subtitle
|
|
|
|
subtitle_content = self._get_subtitle_from_archive(archive)
|
|
|
|
subtitle.content = fix_line_ending(subtitle_content)
|
|
|
|
subtitle.normalize()
|
|
|
|
|
|
|
|
return subtitle
|
|
|
|
raise ValueError('Problems conecting to the server')
|
|
|
|
|
|
|
|
def _get_archive(self, content):
|
|
|
|
# open the archive
|
|
|
|
# stole^H^H^H^H^H inspired from subvix provider
|
|
|
|
archive_stream = io.BytesIO(content)
|
|
|
|
if rarfile.is_rarfile(archive_stream):
|
|
|
|
logger.debug('Identified rar archive')
|
|
|
|
archive = rarfile.RarFile(archive_stream)
|
|
|
|
elif zipfile.is_zipfile(archive_stream):
|
|
|
|
logger.debug('Identified zip archive')
|
|
|
|
archive = zipfile.ZipFile(archive_stream)
|
|
|
|
else:
|
|
|
|
# raise ParseResponseError('Unsupported compressed format')
|
|
|
|
raise Exception('Unsupported compressed format')
|
|
|
|
|
|
|
|
return archive
|
|
|
|
|
|
|
|
def _get_subtitle_from_archive(self, archive):
|
|
|
|
# some files have a non subtitle with .txt extension
|
|
|
|
_tmp = list(SUBTITLE_EXTENSIONS)
|
|
|
|
_tmp.remove('.txt')
|
|
|
|
_subtitle_extensions = tuple(_tmp)
|
|
|
|
|
|
|
|
for name in archive.namelist():
|
|
|
|
# discard hidden files
|
|
|
|
if os.path.split(name)[-1].startswith('.'):
|
|
|
|
continue
|
|
|
|
|
|
|
|
# discard non-subtitle files
|
|
|
|
if not name.lower().endswith(_subtitle_extensions):
|
|
|
|
continue
|
|
|
|
|
|
|
|
logger.debug("returning from archive: %s" % name)
|
|
|
|
return archive.read(name)
|
|
|
|
|
|
|
|
raise ParseResponseError('Can not find the subtitle in the compressed file')
|