mirror of
https://github.com/ytdl-org/youtube-dl.git
synced 2024-11-21 22:51:24 +01:00
395 lines
13 KiB
Python
395 lines
13 KiB
Python
# coding: utf-8
|
||
from __future__ import unicode_literals
|
||
|
||
import hashlib
|
||
import itertools
|
||
import re
|
||
import time
|
||
|
||
from .common import InfoExtractor
|
||
from ..compat import (
|
||
compat_str,
|
||
compat_urllib_parse_urlencode,
|
||
)
|
||
from ..utils import (
|
||
clean_html,
|
||
decode_packed_codes,
|
||
get_element_by_id,
|
||
get_element_by_attribute,
|
||
ExtractorError,
|
||
ohdave_rsa_encrypt,
|
||
remove_start,
|
||
)
|
||
|
||
|
||
def md5_text(text):
|
||
return hashlib.md5(text.encode('utf-8')).hexdigest()
|
||
|
||
|
||
class IqiyiSDK(object):
|
||
def __init__(self, target, ip, timestamp):
|
||
self.target = target
|
||
self.ip = ip
|
||
self.timestamp = timestamp
|
||
|
||
@staticmethod
|
||
def split_sum(data):
|
||
return compat_str(sum(map(lambda p: int(p, 16), list(data))))
|
||
|
||
@staticmethod
|
||
def digit_sum(num):
|
||
if isinstance(num, int):
|
||
num = compat_str(num)
|
||
return compat_str(sum(map(int, num)))
|
||
|
||
def even_odd(self):
|
||
even = self.digit_sum(compat_str(self.timestamp)[::2])
|
||
odd = self.digit_sum(compat_str(self.timestamp)[1::2])
|
||
return even, odd
|
||
|
||
def preprocess(self, chunksize):
|
||
self.target = md5_text(self.target)
|
||
chunks = []
|
||
for i in range(32 // chunksize):
|
||
chunks.append(self.target[chunksize * i:chunksize * (i + 1)])
|
||
if 32 % chunksize:
|
||
chunks.append(self.target[32 - 32 % chunksize:])
|
||
return chunks, list(map(int, self.ip.split('.')))
|
||
|
||
def mod(self, modulus):
|
||
chunks, ip = self.preprocess(32)
|
||
self.target = chunks[0] + ''.join(map(lambda p: compat_str(p % modulus), ip))
|
||
|
||
def split(self, chunksize):
|
||
modulus_map = {
|
||
4: 256,
|
||
5: 10,
|
||
8: 100,
|
||
}
|
||
|
||
chunks, ip = self.preprocess(chunksize)
|
||
ret = ''
|
||
for i in range(len(chunks)):
|
||
ip_part = compat_str(ip[i] % modulus_map[chunksize]) if i < 4 else ''
|
||
if chunksize == 8:
|
||
ret += ip_part + chunks[i]
|
||
else:
|
||
ret += chunks[i] + ip_part
|
||
self.target = ret
|
||
|
||
def handle_input16(self):
|
||
self.target = md5_text(self.target)
|
||
self.target = self.split_sum(self.target[:16]) + self.target + self.split_sum(self.target[16:])
|
||
|
||
def handle_input8(self):
|
||
self.target = md5_text(self.target)
|
||
ret = ''
|
||
for i in range(4):
|
||
part = self.target[8 * i:8 * (i + 1)]
|
||
ret += self.split_sum(part) + part
|
||
self.target = ret
|
||
|
||
def handleSum(self):
|
||
self.target = md5_text(self.target)
|
||
self.target = self.split_sum(self.target) + self.target
|
||
|
||
def date(self, scheme):
|
||
self.target = md5_text(self.target)
|
||
d = time.localtime(self.timestamp)
|
||
strings = {
|
||
'y': compat_str(d.tm_year),
|
||
'm': '%02d' % d.tm_mon,
|
||
'd': '%02d' % d.tm_mday,
|
||
}
|
||
self.target += ''.join(map(lambda c: strings[c], list(scheme)))
|
||
|
||
def split_time_even_odd(self):
|
||
even, odd = self.even_odd()
|
||
self.target = odd + md5_text(self.target) + even
|
||
|
||
def split_time_odd_even(self):
|
||
even, odd = self.even_odd()
|
||
self.target = even + md5_text(self.target) + odd
|
||
|
||
def split_ip_time_sum(self):
|
||
chunks, ip = self.preprocess(32)
|
||
self.target = compat_str(sum(ip)) + chunks[0] + self.digit_sum(self.timestamp)
|
||
|
||
def split_time_ip_sum(self):
|
||
chunks, ip = self.preprocess(32)
|
||
self.target = self.digit_sum(self.timestamp) + chunks[0] + compat_str(sum(ip))
|
||
|
||
|
||
class IqiyiSDKInterpreter(object):
|
||
def __init__(self, sdk_code):
|
||
self.sdk_code = sdk_code
|
||
|
||
def run(self, target, ip, timestamp):
|
||
self.sdk_code = decode_packed_codes(self.sdk_code)
|
||
|
||
functions = re.findall(r'input=([a-zA-Z0-9]+)\(input', self.sdk_code)
|
||
|
||
sdk = IqiyiSDK(target, ip, timestamp)
|
||
|
||
other_functions = {
|
||
'handleSum': sdk.handleSum,
|
||
'handleInput8': sdk.handle_input8,
|
||
'handleInput16': sdk.handle_input16,
|
||
'splitTimeEvenOdd': sdk.split_time_even_odd,
|
||
'splitTimeOddEven': sdk.split_time_odd_even,
|
||
'splitIpTimeSum': sdk.split_ip_time_sum,
|
||
'splitTimeIpSum': sdk.split_time_ip_sum,
|
||
}
|
||
for function in functions:
|
||
if re.match(r'mod\d+', function):
|
||
sdk.mod(int(function[3:]))
|
||
elif re.match(r'date[ymd]{3}', function):
|
||
sdk.date(function[4:])
|
||
elif re.match(r'split\d+', function):
|
||
sdk.split(int(function[5:]))
|
||
elif function in other_functions:
|
||
other_functions[function]()
|
||
else:
|
||
raise ExtractorError('Unknown function %s' % function)
|
||
|
||
return sdk.target
|
||
|
||
|
||
class IqiyiIE(InfoExtractor):
|
||
IE_NAME = 'iqiyi'
|
||
IE_DESC = '爱奇艺'
|
||
|
||
_VALID_URL = r'https?://(?:(?:[^.]+\.)?iqiyi\.com|www\.pps\.tv)/.+\.html'
|
||
|
||
_NETRC_MACHINE = 'iqiyi'
|
||
|
||
_TESTS = [{
|
||
'url': 'http://www.iqiyi.com/v_19rrojlavg.html',
|
||
# MD5 checksum differs on my machine and Travis CI
|
||
'info_dict': {
|
||
'id': '9c1fb1b99d192b21c559e5a1a2cb3c73',
|
||
'ext': 'mp4',
|
||
'title': '美国德州空中惊现奇异云团 酷似UFO',
|
||
}
|
||
}, {
|
||
'url': 'http://www.iqiyi.com/v_19rrhnnclk.html',
|
||
'md5': 'b7dc800a4004b1b57749d9abae0472da',
|
||
'info_dict': {
|
||
'id': 'e3f585b550a280af23c98b6cb2be19fb',
|
||
'ext': 'mp4',
|
||
# This can be either Simplified Chinese or Traditional Chinese
|
||
'title': r're:^(?:名侦探柯南 国语版:第752集 迫近灰原秘密的黑影 下篇|名偵探柯南 國語版:第752集 迫近灰原秘密的黑影 下篇)$',
|
||
},
|
||
'skip': 'Geo-restricted to China',
|
||
}, {
|
||
'url': 'http://www.iqiyi.com/w_19rt6o8t9p.html',
|
||
'only_matching': True,
|
||
}, {
|
||
'url': 'http://www.iqiyi.com/a_19rrhbc6kt.html',
|
||
'only_matching': True,
|
||
}, {
|
||
'url': 'http://yule.iqiyi.com/pcb.html',
|
||
'info_dict': {
|
||
'id': '4a0af228fddb55ec96398a364248ed7f',
|
||
'ext': 'mp4',
|
||
'title': '第2017-04-21期 女艺人频遭极端粉丝骚扰',
|
||
},
|
||
}, {
|
||
# VIP-only video. The first 2 parts (6 minutes) are available without login
|
||
# MD5 sums omitted as values are different on Travis CI and my machine
|
||
'url': 'http://www.iqiyi.com/v_19rrny4w8w.html',
|
||
'info_dict': {
|
||
'id': 'f3cf468b39dddb30d676f89a91200dc1',
|
||
'ext': 'mp4',
|
||
'title': '泰坦尼克号',
|
||
},
|
||
'skip': 'Geo-restricted to China',
|
||
}, {
|
||
'url': 'http://www.iqiyi.com/a_19rrhb8ce1.html',
|
||
'info_dict': {
|
||
'id': '202918101',
|
||
'title': '灌篮高手 国语版',
|
||
},
|
||
'playlist_count': 101,
|
||
}, {
|
||
'url': 'http://www.pps.tv/w_19rrbav0ph.html',
|
||
'only_matching': True,
|
||
}]
|
||
|
||
_FORMATS_MAP = {
|
||
'96': 1, # 216p, 240p
|
||
'1': 2, # 336p, 360p
|
||
'2': 3, # 480p, 504p
|
||
'21': 4, # 504p
|
||
'4': 5, # 720p
|
||
'17': 5, # 720p
|
||
'5': 6, # 1072p, 1080p
|
||
'18': 7, # 1080p
|
||
}
|
||
|
||
def _real_initialize(self):
|
||
self._login()
|
||
|
||
@staticmethod
|
||
def _rsa_fun(data):
|
||
# public key extracted from http://static.iqiyi.com/js/qiyiV2/20160129180840/jobs/i18n/i18nIndex.js
|
||
N = 0xab86b6371b5318aaa1d3c9e612a9f1264f372323c8c0f19875b5fc3b3fd3afcc1e5bec527aa94bfa85bffc157e4245aebda05389a5357b75115ac94f074aefcd
|
||
e = 65537
|
||
|
||
return ohdave_rsa_encrypt(data, e, N)
|
||
|
||
def _login(self):
|
||
username, password = self._get_login_info()
|
||
|
||
# No authentication to be performed
|
||
if not username:
|
||
return True
|
||
|
||
data = self._download_json(
|
||
'http://kylin.iqiyi.com/get_token', None,
|
||
note='Get token for logging', errnote='Unable to get token for logging')
|
||
sdk = data['sdk']
|
||
timestamp = int(time.time())
|
||
target = '/apis/reglogin/login.action?lang=zh_TW&area_code=null&email=%s&passwd=%s&agenttype=1&from=undefined&keeplogin=0&piccode=&fromurl=&_pos=1' % (
|
||
username, self._rsa_fun(password.encode('utf-8')))
|
||
|
||
interp = IqiyiSDKInterpreter(sdk)
|
||
sign = interp.run(target, data['ip'], timestamp)
|
||
|
||
validation_params = {
|
||
'target': target,
|
||
'server': 'BEA3AA1908656AABCCFF76582C4C6660',
|
||
'token': data['token'],
|
||
'bird_src': 'f8d91d57af224da7893dd397d52d811a',
|
||
'sign': sign,
|
||
'bird_t': timestamp,
|
||
}
|
||
validation_result = self._download_json(
|
||
'http://kylin.iqiyi.com/validate?' + compat_urllib_parse_urlencode(validation_params), None,
|
||
note='Validate credentials', errnote='Unable to validate credentials')
|
||
|
||
MSG_MAP = {
|
||
'P00107': 'please login via the web interface and enter the CAPTCHA code',
|
||
'P00117': 'bad username or password',
|
||
}
|
||
|
||
code = validation_result['code']
|
||
if code != 'A00000':
|
||
msg = MSG_MAP.get(code)
|
||
if not msg:
|
||
msg = 'error %s' % code
|
||
if validation_result.get('msg'):
|
||
msg += ': ' + validation_result['msg']
|
||
self._downloader.report_warning('unable to log in: ' + msg)
|
||
return False
|
||
|
||
return True
|
||
|
||
def get_raw_data(self, tvid, video_id):
|
||
tm = int(time.time() * 1000)
|
||
|
||
key = 'd5fb4bd9d50c4be6948c97edd7254b0e'
|
||
sc = md5_text(compat_str(tm) + key + tvid)
|
||
params = {
|
||
'tvid': tvid,
|
||
'vid': video_id,
|
||
'src': '76f90cbd92f94a2e925d83e8ccd22cb7',
|
||
'sc': sc,
|
||
't': tm,
|
||
}
|
||
|
||
return self._download_json(
|
||
'http://cache.m.iqiyi.com/jp/tmts/%s/%s/' % (tvid, video_id),
|
||
video_id, transform_source=lambda s: remove_start(s, 'var tvInfoJs='),
|
||
query=params, headers=self.geo_verification_headers())
|
||
|
||
def _extract_playlist(self, webpage):
|
||
PAGE_SIZE = 50
|
||
|
||
links = re.findall(
|
||
r'<a[^>]+class="site-piclist_pic_link"[^>]+href="(http://www\.iqiyi\.com/.+\.html)"',
|
||
webpage)
|
||
if not links:
|
||
return
|
||
|
||
album_id = self._search_regex(
|
||
r'albumId\s*:\s*(\d+),', webpage, 'album ID')
|
||
album_title = self._search_regex(
|
||
r'data-share-title="([^"]+)"', webpage, 'album title', fatal=False)
|
||
|
||
entries = list(map(self.url_result, links))
|
||
|
||
# Start from 2 because links in the first page are already on webpage
|
||
for page_num in itertools.count(2):
|
||
pagelist_page = self._download_webpage(
|
||
'http://cache.video.qiyi.com/jp/avlist/%s/%d/%d/' % (album_id, page_num, PAGE_SIZE),
|
||
album_id,
|
||
note='Download playlist page %d' % page_num,
|
||
errnote='Failed to download playlist page %d' % page_num)
|
||
pagelist = self._parse_json(
|
||
remove_start(pagelist_page, 'var tvInfoJs='), album_id)
|
||
vlist = pagelist['data']['vlist']
|
||
for item in vlist:
|
||
entries.append(self.url_result(item['vurl']))
|
||
if len(vlist) < PAGE_SIZE:
|
||
break
|
||
|
||
return self.playlist_result(entries, album_id, album_title)
|
||
|
||
def _real_extract(self, url):
|
||
webpage = self._download_webpage(
|
||
url, 'temp_id', note='download video page')
|
||
|
||
# There's no simple way to determine whether an URL is a playlist or not
|
||
# Sometimes there are playlist links in individual videos, so treat it
|
||
# as a single video first
|
||
tvid = self._search_regex(
|
||
r'data-(?:player|shareplattrigger)-tvid\s*=\s*[\'"](\d+)', webpage, 'tvid', default=None)
|
||
if tvid is None:
|
||
playlist_result = self._extract_playlist(webpage)
|
||
if playlist_result:
|
||
return playlist_result
|
||
raise ExtractorError('Can\'t find any video')
|
||
|
||
video_id = self._search_regex(
|
||
r'data-(?:player|shareplattrigger)-videoid\s*=\s*[\'"]([a-f\d]+)', webpage, 'video_id')
|
||
|
||
formats = []
|
||
for _ in range(5):
|
||
raw_data = self.get_raw_data(tvid, video_id)
|
||
|
||
if raw_data['code'] != 'A00000':
|
||
if raw_data['code'] == 'A00111':
|
||
self.raise_geo_restricted()
|
||
raise ExtractorError('Unable to load data. Error code: ' + raw_data['code'])
|
||
|
||
data = raw_data['data']
|
||
|
||
for stream in data['vidl']:
|
||
if 'm3utx' not in stream:
|
||
continue
|
||
vd = compat_str(stream['vd'])
|
||
formats.append({
|
||
'url': stream['m3utx'],
|
||
'format_id': vd,
|
||
'ext': 'mp4',
|
||
'preference': self._FORMATS_MAP.get(vd, -1),
|
||
'protocol': 'm3u8_native',
|
||
})
|
||
|
||
if formats:
|
||
break
|
||
|
||
self._sleep(5, video_id)
|
||
|
||
self._sort_formats(formats)
|
||
title = (get_element_by_id('widget-videotitle', webpage)
|
||
or clean_html(get_element_by_attribute('class', 'mod-play-tit', webpage))
|
||
or self._html_search_regex(r'<span[^>]+data-videochanged-title="word"[^>]*>([^<]+)</span>', webpage, 'title'))
|
||
|
||
return {
|
||
'id': video_id,
|
||
'title': title,
|
||
'formats': formats,
|
||
}
|