# encoding: utf-8 from __future__ import unicode_literals import re from .crunchyroll import CrunchyrollIE from .common import InfoExtractor from ..compat import compat_HTTPError from ..utils import ( ExtractorError, int_or_none, remove_start, xpath_text, ) class SoompiBaseIE(InfoExtractor): def _get_episodes(self, webpage, episode_filter=None): episodes = self._parse_json( self._search_regex( r'VIDEOS\s*=\s*(\[.+?\]);', webpage, 'episodes JSON'), None) return list(filter(episode_filter, episodes)) class SoompiIE(SoompiBaseIE, CrunchyrollIE): IE_NAME = 'soompi' _VALID_URL = r'https?://tv\.soompi\.com/(?:en/)?watch/(?P<id>[0-9]+)' _TESTS = [{ 'url': 'http://tv.soompi.com/en/watch/29235', 'info_dict': { 'id': '29235', 'ext': 'mp4', 'title': 'Episode 1096', 'description': '2015-05-20' }, 'params': { 'skip_download': True, }, }] def _get_episode(self, webpage, video_id): return self._get_episodes(webpage, lambda x: x['id'] == video_id)[0] def _get_subtitles(self, config, video_id): sub_langs = {} for subtitle in config.findall('./{default}preload/subtitles/subtitle'): sub_langs[subtitle.attrib['id']] = subtitle.attrib['title'] subtitles = {} for s in config.findall('./{default}preload/subtitle'): lang_code = sub_langs.get(s.attrib['id']) if not lang_code: continue sub_id = s.get('id') data = xpath_text(s, './data', 'data') iv = xpath_text(s, './iv', 'iv') if not id or not iv or not data: continue subtitle = self._decrypt_subtitles(data, iv, sub_id).decode('utf-8') subtitles[lang_code] = self._extract_subtitles(subtitle) return subtitles def _real_extract(self, url): video_id = self._match_id(url) try: webpage = self._download_webpage( url, video_id, 'Downloading episode page') except ExtractorError as ee: if isinstance(ee.cause, compat_HTTPError) and ee.cause.code == 403: webpage = ee.cause.read() block_message = self._html_search_regex( r'(?s)<div class="block-message">(.+?)</div>', webpage, 'block message', default=None) if block_message: raise ExtractorError(block_message, expected=True) raise formats = [] config = None for format_id in re.findall(r'\?quality=([0-9a-zA-Z]+)', webpage): config = self._download_xml( 'http://tv.soompi.com/en/show/_/%s-config.xml?mode=hls&quality=%s' % (video_id, format_id), video_id, 'Downloading %s XML' % format_id) m3u8_url = xpath_text( config, './{default}preload/stream_info/file', '%s m3u8 URL' % format_id) if not m3u8_url: continue formats.extend(self._extract_m3u8_formats( m3u8_url, video_id, 'mp4', m3u8_id=format_id)) self._sort_formats(formats) episode = self._get_episode(webpage, video_id) title = episode['name'] description = episode.get('description') duration = int_or_none(episode.get('duration')) thumbnails = [{ 'id': thumbnail_id, 'url': thumbnail_url, } for thumbnail_id, thumbnail_url in episode.get('img_url', {}).items()] subtitles = self.extract_subtitles(config, video_id) return { 'id': video_id, 'title': title, 'description': description, 'thumbnails': thumbnails, 'duration': duration, 'formats': formats, 'subtitles': subtitles } class SoompiShowIE(SoompiBaseIE): IE_NAME = 'soompi:show' _VALID_URL = r'https?://tv\.soompi\.com/en/shows/(?P<id>[0-9a-zA-Z\-_]+)' _TESTS = [{ 'url': 'http://tv.soompi.com/en/shows/liar-game', 'info_dict': { 'id': 'liar-game', 'title': 'Liar Game', 'description': 'md5:52c02bce0c1a622a95823591d0589b66', }, 'playlist_count': 14, }] def _real_extract(self, url): show_id = self._match_id(url) webpage = self._download_webpage( url, show_id, 'Downloading show page') title = remove_start(self._og_search_title(webpage), 'SoompiTV | ') description = self._og_search_description(webpage) entries = [ self.url_result('http://tv.soompi.com/en/watch/%s' % episode['id'], 'Soompi') for episode in self._get_episodes(webpage)] return self.playlist_result(entries, show_id, title, description)