1
1
mirror of https://github.com/ytdl-org/youtube-dl synced 2024-11-25 04:46:50 +01:00
youtube-dl/youtube_dl/extractor/leeco.py

369 lines
13 KiB
Python
Raw Normal View History

# coding: utf-8
from __future__ import unicode_literals
2015-02-26 01:30:18 +01:00
import datetime
2016-03-02 19:42:19 +01:00
import hashlib
2015-02-24 17:45:11 +01:00
import re
import time
from .common import InfoExtractor
2015-02-26 01:30:18 +01:00
from ..compat import (
compat_b64decode,
compat_ord,
2016-01-31 00:13:21 +01:00
compat_str,
compat_urllib_parse_urlencode,
2015-02-26 01:30:18 +01:00
)
from ..utils import (
determine_ext,
2016-03-02 19:42:19 +01:00
encode_data_uri,
2015-02-26 01:30:18 +01:00
ExtractorError,
2016-03-02 19:42:19 +01:00
int_or_none,
orderedSet,
2015-02-26 01:30:18 +01:00
parse_iso8601,
2015-12-20 12:09:45 +01:00
str_or_none,
url_basename,
2016-06-26 09:17:26 +02:00
urshift,
2015-02-26 01:30:18 +01:00
)
class LeIE(InfoExtractor):
2015-07-10 18:51:26 +02:00
IE_DESC = '乐视网'
_VALID_URL = r'https?://(?:www\.le\.com/ptv/vplay|(?:sports\.le|(?:www\.)?lesports)\.com/(?:match|video))/(?P<id>\d+)\.html'
_GEO_COUNTRIES = ['CN']
_URL_TEMPLATE = 'http://www.le.com/ptv/vplay/%s.html'
_TESTS = [{
2016-03-02 19:38:41 +01:00
'url': 'http://www.le.com/ptv/vplay/22005890.html',
'md5': 'edadcfe5406976f42f9f266057ee5e40',
'info_dict': {
'id': '22005890',
'ext': 'mp4',
'title': '第87届奥斯卡颁奖礼完美落幕 《鸟人》成最大赢家',
2015-02-24 19:10:55 +01:00
'description': 'md5:a9cb175fd753e2962176b7beca21a47c',
},
'params': {
'hls_prefer_native': True,
},
}, {
2016-03-02 19:38:41 +01:00
'url': 'http://www.le.com/ptv/vplay/1415246.html',
'info_dict': {
2015-02-24 19:10:55 +01:00
'id': '1415246',
'ext': 'mp4',
2015-02-24 19:10:55 +01:00
'title': '美人天下01',
'description': 'md5:28942e650e82ed4fcc8e4de919ee854d',
2015-02-24 19:10:55 +01:00
},
'params': {
'hls_prefer_native': True,
},
}, {
'note': 'This video is available only in Mainland China, thus a proxy is needed',
2016-03-02 19:38:41 +01:00
'url': 'http://www.le.com/ptv/vplay/1118082.html',
'md5': '2424c74948a62e5f31988438979c5ad1',
'info_dict': {
'id': '1118082',
'ext': 'mp4',
'title': '与龙共舞 完整版',
'description': 'md5:7506a5eeb1722bb9d4068f85024e3986',
},
'params': {
'hls_prefer_native': True,
},
}, {
'url': 'http://sports.le.com/video/25737697.html',
'only_matching': True,
}, {
'url': 'http://www.lesports.com/match/1023203003.html',
'only_matching': True,
}, {
'url': 'http://sports.le.com/match/1023203003.html',
'only_matching': True,
}]
# ror() and calc_time_key() are reversed from a embedded swf file in LetvPlayer.swf
def ror(self, param1, param2):
_loc3_ = 0
while _loc3_ < param2:
2016-06-26 09:17:26 +02:00
param1 = urshift(param1, 1) + ((param1 & 1) << 31)
_loc3_ += 1
return param1
2015-02-26 01:30:18 +01:00
def calc_time_key(self, param1):
_loc2_ = 185025305
return self.ror(param1, _loc2_ % 17) ^ _loc2_
2016-07-08 16:35:20 +02:00
# see M3U8Encryption class in KLetvPlayer.swf
@staticmethod
def decrypt_m3u8(encrypted_data):
if encrypted_data[:5].decode('utf-8').lower() != 'vc_01':
return encrypted_data
encrypted_data = encrypted_data[5:]
2016-02-26 17:57:35 +01:00
_loc4_ = bytearray(2 * len(encrypted_data))
for idx, val in enumerate(encrypted_data):
b = compat_ord(val)
_loc4_[2 * idx] = b // 16
_loc4_[2 * idx + 1] = b % 16
idx = len(_loc4_) - 11
_loc4_ = _loc4_[idx:] + _loc4_[:idx]
2016-02-26 17:57:35 +01:00
_loc7_ = bytearray(len(encrypted_data))
for i in range(len(encrypted_data)):
_loc7_[i] = _loc4_[2 * i] * 16 + _loc4_[2 * i + 1]
return bytes(_loc7_)
2016-07-08 16:35:20 +02:00
def _check_errors(self, play_json):
# Check for errors
playstatus = play_json['msgs']['playstatus']
if playstatus['status'] == 0:
flag = playstatus['flag']
if flag == 1:
self.raise_geo_restricted()
else:
raise ExtractorError('Generic error. flag = %d' % flag, expected=True)
2016-07-08 16:35:20 +02:00
def _real_extract(self, url):
media_id = self._match_id(url)
page = self._download_webpage(url, media_id)
play_json_flash = self._download_json(
'http://player-pc.le.com/mms/out/video/playJson',
2016-07-08 16:35:20 +02:00
media_id, 'Downloading flash playJson data', query={
'id': media_id,
'platid': 1,
'splatid': 101,
'format': 1,
'source': 1000,
2016-07-08 16:35:20 +02:00
'tkey': self.calc_time_key(int(time.time())),
'domain': 'www.le.com',
'region': 'cn',
2016-07-08 16:35:20 +02:00
},
headers=self.geo_verification_headers())
self._check_errors(play_json_flash)
2016-07-08 16:35:20 +02:00
def get_flash_urls(media_url, format_id):
nodes_data = self._download_json(
media_url, media_id,
'Download JSON metadata for format %s' % format_id,
query={
'm3v': 1,
'format': 1,
'expect': 3,
'tss': 'ios',
})
2016-07-08 16:35:20 +02:00
req = self._request_webpage(
nodes_data['nodelist'][0]['location'], media_id,
note='Downloading m3u8 information for format %s' % format_id)
m3u8_data = self.decrypt_m3u8(req.read())
return {
'hls': encode_data_uri(m3u8_data, 'application/vnd.apple.mpegurl'),
}
extracted_formats = []
formats = []
playurl = play_json_flash['msgs']['playurl']
play_domain = playurl['domain'][0]
for format_id, format_data in playurl.get('dispatch', []).items():
if format_id in extracted_formats:
continue
extracted_formats.append(format_id)
media_url = play_domain + format_data[0]
for protocol, format_url in get_flash_urls(media_url, format_id).items():
f = {
'url': format_url,
'ext': determine_ext(format_data[1]),
'format_id': '%s-%s' % (protocol, format_id),
'protocol': 'm3u8_native' if protocol == 'hls' else 'http',
'quality': int_or_none(format_id),
}
if format_id[-1:] == 'p':
f['height'] = int_or_none(format_id[:-1])
formats.append(f)
2016-07-08 16:35:20 +02:00
self._sort_formats(formats, ('height', 'quality', 'format_id'))
publish_time = parse_iso8601(self._html_search_regex(
r'发布时间&nbsp;([^<>]+) ', page, 'publish time', default=None),
delimiter=' ', timezone=datetime.timedelta(hours=8))
2015-02-24 19:10:55 +01:00
description = self._html_search_meta('description', page, fatal=False)
return {
'id': media_id,
2016-07-08 16:35:20 +02:00
'formats': formats,
'title': playurl['title'],
'thumbnail': playurl['pic'],
2015-02-24 19:10:55 +01:00
'description': description,
'timestamp': publish_time,
}
2015-02-24 17:45:11 +01:00
class LePlaylistIE(InfoExtractor):
_VALID_URL = r'https?://[a-z]+\.le\.com/(?!video)[a-z]+/(?P<id>[a-z0-9_]+)'
2015-02-24 17:45:11 +01:00
_TESTS = [{
2016-03-02 19:38:41 +01:00
'url': 'http://www.le.com/tv/46177.html',
2015-02-24 17:45:11 +01:00
'info_dict': {
'id': '46177',
'title': '美人天下',
'description': 'md5:395666ff41b44080396e59570dbac01c'
},
'playlist_count': 35
}, {
2016-03-02 19:38:41 +01:00
'url': 'http://tv.le.com/izt/wuzetian/index.html',
2015-02-24 17:45:11 +01:00
'info_dict': {
'id': 'wuzetian',
'title': '武媚娘传奇',
'description': 'md5:e12499475ab3d50219e5bba00b3cb248'
},
2015-02-24 19:10:55 +01:00
# This playlist contains some extra videos other than the drama itself
'playlist_mincount': 96
2015-02-24 17:45:11 +01:00
}, {
2016-03-02 19:38:41 +01:00
'url': 'http://tv.le.com/pzt/lswjzzjc/index.shtml',
# This series is moved to http://www.le.com/tv/10005297.html
'only_matching': True,
}, {
'url': 'http://www.le.com/comic/92063.html',
'only_matching': True,
}, {
'url': 'http://list.le.com/listn/c1009_sc532002_d2_p1_o1.html',
'only_matching': True,
2015-02-24 17:45:11 +01:00
}]
2015-12-20 12:09:45 +01:00
@classmethod
def suitable(cls, url):
return False if LeIE.suitable(url) else super(LePlaylistIE, cls).suitable(url)
def _real_extract(self, url):
playlist_id = self._match_id(url)
page = self._download_webpage(url, playlist_id)
# Currently old domain names are still used in playlists
media_ids = orderedSet(re.findall(
r'<a[^>]+href="http://www\.letv\.com/ptv/vplay/(\d+)\.html', page))
entries = [self.url_result(LeIE._URL_TEMPLATE % media_id, ie='Le')
for media_id in media_ids]
title = self._html_search_meta('keywords', page,
fatal=False).split('')[0]
description = self._html_search_meta('description', page, fatal=False)
return self.playlist_result(entries, playlist_id, playlist_title=title,
playlist_description=description)
2015-12-20 12:09:45 +01:00
class LetvCloudIE(InfoExtractor):
# Most of *.letv.com is changed to *.le.com on 2016/01/02
# but yuntv.letv.com is kept, so also keep the extractor name
2015-12-20 12:09:45 +01:00
IE_DESC = '乐视云'
_VALID_URL = r'https?://yuntv\.letv\.com/bcloud.html\?.+'
2015-12-20 12:09:45 +01:00
_TESTS = [{
'url': 'http://yuntv.letv.com/bcloud.html?uu=p7jnfw5hw9&vu=467623dedf',
'md5': '26450599afd64c513bc77030ad15db44',
'info_dict': {
'id': 'p7jnfw5hw9_467623dedf',
'ext': 'mp4',
'title': 'Video p7jnfw5hw9_467623dedf',
2015-12-20 12:09:45 +01:00
},
}, {
'url': 'http://yuntv.letv.com/bcloud.html?uu=p7jnfw5hw9&vu=ec93197892&pu=2c7cd40209&auto_play=1&gpcflag=1&width=640&height=360',
2016-01-31 00:13:21 +01:00
'md5': 'e03d9cc8d9c13191e1caf277e42dbd31',
2015-12-20 12:09:45 +01:00
'info_dict': {
'id': 'p7jnfw5hw9_ec93197892',
'ext': 'mp4',
'title': 'Video p7jnfw5hw9_ec93197892',
2015-12-20 12:09:45 +01:00
},
}, {
'url': 'http://yuntv.letv.com/bcloud.html?uu=p7jnfw5hw9&vu=187060b6fd',
2016-01-31 00:13:21 +01:00
'md5': 'cb988699a776b22d4a41b9d43acfb3ac',
2015-12-20 12:09:45 +01:00
'info_dict': {
'id': 'p7jnfw5hw9_187060b6fd',
'ext': 'mp4',
'title': 'Video p7jnfw5hw9_187060b6fd',
2015-12-20 12:09:45 +01:00
},
}]
2016-01-31 00:13:21 +01:00
@staticmethod
def sign_data(obj):
if obj['cf'] == 'flash':
salt = '2f9d6924b33a165a6d8b5d3d42f4f987'
items = ['cf', 'format', 'ran', 'uu', 'ver', 'vu']
elif obj['cf'] == 'html5':
salt = 'fbeh5player12c43eccf2bec3300344'
items = ['cf', 'ran', 'uu', 'bver', 'vu']
input_data = ''.join([item + obj[item] for item in items]) + salt
obj['sign'] = hashlib.md5(input_data.encode('utf-8')).hexdigest()
def _get_formats(self, cf, uu, vu, media_id):
def get_play_json(cf, timestamp):
data = {
'cf': cf,
'ver': '2.2',
'bver': 'firefox44.0',
'format': 'json',
'uu': uu,
'vu': vu,
'ran': compat_str(timestamp),
}
self.sign_data(data)
return self._download_json(
'http://api.letvcloud.com/gpc.php?' + compat_urllib_parse_urlencode(data),
2016-01-31 00:13:21 +01:00
media_id, 'Downloading playJson data for type %s' % cf)
play_json = get_play_json(cf, time.time())
# The server time may be different from local time
if play_json.get('code') == 10071:
play_json = get_play_json(cf, play_json['timestamp'])
2015-12-20 12:09:45 +01:00
if not play_json.get('data'):
if play_json.get('message'):
raise ExtractorError('Letv cloud said: %s' % play_json['message'], expected=True)
elif play_json.get('code'):
raise ExtractorError('Letv cloud returned error %d' % play_json['code'], expected=True)
else:
raise ExtractorError('Letv cloud returned an unknwon error')
def b64decode(s):
return compat_b64decode(s).decode('utf-8')
2016-01-19 20:17:35 +01:00
formats = []
for media in play_json['data']['video_info']['media'].values():
play_url = media['play_url']
url = b64decode(play_url['main_url'])
decoded_url = b64decode(url_basename(url))
2016-01-19 20:17:35 +01:00
formats.append({
'url': url,
'ext': determine_ext(decoded_url),
2016-03-04 06:38:45 +01:00
'format_id': str_or_none(play_url.get('vtype')),
2016-01-19 20:17:35 +01:00
'format_note': str_or_none(play_url.get('definition')),
'width': int_or_none(play_url.get('vwidth')),
'height': int_or_none(play_url.get('vheight')),
})
2016-01-31 00:13:21 +01:00
return formats
def _real_extract(self, url):
uu_mobj = re.search(r'uu=([\w]+)', url)
vu_mobj = re.search(r'vu=([\w]+)', url)
2016-01-31 00:13:21 +01:00
if not uu_mobj or not vu_mobj:
raise ExtractorError('Invalid URL: %s' % url, expected=True)
uu = uu_mobj.group(1)
vu = vu_mobj.group(1)
media_id = uu + '_' + vu
formats = self._get_formats('flash', uu, vu, media_id) + self._get_formats('html5', uu, vu, media_id)
2015-12-20 12:09:45 +01:00
self._sort_formats(formats)
return {
'id': media_id,
'title': 'Video %s' % media_id,
2015-12-20 12:09:45 +01:00
'formats': formats,
}