1
1
mirror of https://github.com/ytdl-org/youtube-dl synced 2024-12-21 17:57:55 +01:00
youtube-dl/youtube_dl/extractor/weibo.py

141 lines
4.4 KiB
Python
Raw Normal View History

# coding: utf-8
from __future__ import unicode_literals
from .common import InfoExtractor
import json
import random
2017-12-26 09:18:35 +01:00
import re
2017-12-26 09:28:47 +01:00
from ..compat import (
2018-01-09 11:12:55 +01:00
compat_parse_qs,
compat_str,
2017-12-26 09:28:47 +01:00
)
2017-12-26 09:18:35 +01:00
from ..utils import (
js_to_json,
2017-12-29 18:11:30 +01:00
strip_jsonp,
urlencode_postdata,
2017-12-26 09:18:35 +01:00
)
2017-12-26 09:38:51 +01:00
class WeiboIE(InfoExtractor):
_VALID_URL = r'https?://weibo\.com/[0-9]+/(?P<id>[a-zA-Z0-9]+)'
_TEST = {
2017-12-26 09:38:51 +01:00
'url': 'https://weibo.com/6275294458/Fp6RGfbff?type=comment',
'info_dict': {
'id': 'Fp6RGfbff',
'ext': 'mp4',
'title': 'You should have servants to massage you,... 来自Hosico_猫 - 微博',
}
}
def _real_extract(self, url):
video_id = self._match_id(url)
# to get Referer url for genvisitor
2018-01-09 11:12:55 +01:00
webpage, urlh = self._download_webpage_handle(url, video_id)
visitor_url = urlh.geturl()
2018-01-09 11:12:55 +01:00
if 'passport.weibo.com' in visitor_url:
# first visit
visitor_data = self._download_json(
'https://passport.weibo.com/visitor/genvisitor', video_id,
note='Generating first-visit data',
transform_source=strip_jsonp,
headers={'Referer': visitor_url},
data=urlencode_postdata({
'cb': 'gen_callback',
'fp': json.dumps({
'os': '2',
'browser': 'Gecko57,0,0,0',
'fonts': 'undefined',
'screenInfo': '1440*900*24',
'plugins': '',
}),
}))
tid = visitor_data['data']['tid']
cnfd = '%03d' % visitor_data['data']['confidence']
self._download_webpage(
'https://passport.weibo.com/visitor/visitor', video_id,
note='Running first-visit callback',
query={
'a': 'incarnate',
't': tid,
'w': 2,
'c': cnfd,
'cb': 'cross_domain',
'from': 'weibo',
'_rand': random.random(),
})
webpage = self._download_webpage(
url, video_id, note='Revisiting webpage')
title = self._html_search_regex(
r'<title>(.+?)</title>', webpage, 'title')
video_formats = compat_parse_qs(self._search_regex(
r'video-sources=\\\"(.+?)\"', webpage, 'video_sources'))
2018-01-09 11:12:55 +01:00
formats = []
supported_resolutions = (480, 720)
for res in supported_resolutions:
vid_urls = video_formats.get(compat_str(res))
if not vid_urls or not isinstance(vid_urls, list):
continue
2018-01-09 11:12:55 +01:00
vid_url = vid_urls[0]
formats.append({
'url': vid_url,
'height': res,
})
2018-01-09 11:12:55 +01:00
self._sort_formats(formats)
2017-12-26 09:38:51 +01:00
2018-01-09 11:12:55 +01:00
uploader = self._og_search_property(
'nick-name', webpage, 'uploader', default=None)
2017-12-11 08:56:54 +01:00
return {
2017-12-26 09:38:51 +01:00
'id': video_id,
'title': title,
'uploader': uploader,
'formats': formats
}
2017-12-26 09:18:35 +01:00
class WeiboMobileIE(InfoExtractor):
_VALID_URL = r'https?://m\.weibo\.cn/status/(?P<id>[0-9]+)(\?.+)?'
2017-12-26 09:18:35 +01:00
_TEST = {
2017-12-26 09:38:51 +01:00
'url': 'https://m.weibo.cn/status/4189191225395228?wm=3333_2001&sourcetype=weixin&featurecode=newtitle&from=singlemessage&isappinstalled=0',
'info_dict': {
'id': '4189191225395228',
'ext': 'mp4',
'title': '午睡当然是要甜甜蜜蜜的啦',
'uploader': '柴犬柴犬'
}
}
2017-12-26 09:18:35 +01:00
def _real_extract(self, url):
video_id = self._match_id(url)
# to get Referer url for genvisitor
2018-01-09 11:12:55 +01:00
webpage = self._download_webpage(url, video_id, note='visit the page')
weibo_info = self._parse_json(self._search_regex(
r'var\s+\$render_data\s*=\s*\[({.*})\]\[0\]\s*\|\|\s*{};',
webpage, 'js_code', flags=re.DOTALL),
video_id, transform_source=js_to_json)
status_data = weibo_info.get('status', {})
page_info = status_data.get('page_info')
title = status_data['status_title']
uploader = status_data.get('user', {}).get('screen_name')
2017-12-26 09:18:35 +01:00
return {
2017-12-26 09:38:51 +01:00
'id': video_id,
'title': title,
'uploader': uploader,
'url': page_info['media_info']['stream_url']
2017-12-26 09:38:51 +01:00
}