mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-11-30 03:01:25 +01:00
Compare commits
No commits in common. "37ff05333317b624d1085b6471253862d04aed95" and "69926d80e88f991ab0a45f26a6de1ff8d4d27240" have entirely different histories.
37ff053333
...
69926d80e8
|
@ -4,12 +4,9 @@ from .common import InfoExtractor
|
|||
from ..utils import (
|
||||
ExtractorError,
|
||||
clean_html,
|
||||
filter_dict,
|
||||
get_element_by_id,
|
||||
int_or_none,
|
||||
join_nonempty,
|
||||
js_to_json,
|
||||
qualities,
|
||||
url_or_none,
|
||||
urljoin,
|
||||
)
|
||||
|
@ -47,22 +44,27 @@ class KukuluLiveIE(InfoExtractor):
|
|||
'only_matching': True,
|
||||
}]
|
||||
|
||||
def _get_quality_meta(self, video_id, desc, code, force_h264=None):
|
||||
desc += ' (force_h264)' if force_h264 else ''
|
||||
def _get_quality_meta(self, video_id, desc, code, force_h264=''):
|
||||
description = desc if force_h264 == '' else f'{desc} (force_h264)'
|
||||
qs = self._download_webpage(
|
||||
'https://live.erinn.biz/live.player.fplayer.php', video_id,
|
||||
f'Downloading {desc} quality metadata', f'Unable to download {desc} quality metadata',
|
||||
query=filter_dict({
|
||||
query={
|
||||
'hash': video_id,
|
||||
'action': f'get{code}liveByAjax',
|
||||
'force_h264': force_h264,
|
||||
}))
|
||||
},
|
||||
note=f'Downloading {description} quality metadata',
|
||||
errnote=f'Unable to download {description} quality metadata')
|
||||
return urllib.parse.parse_qs(qs)
|
||||
|
||||
def _add_quality_formats(self, formats, quality_meta):
|
||||
vcodec = traverse_obj(quality_meta, ('vcodec', 0, {str}))
|
||||
quality = traverse_obj(quality_meta, ('now_quality', 0, {str}))
|
||||
quality_priority = qualities(('low', 'h264', 'high'))(quality)
|
||||
vcodec = traverse_obj(quality_meta, ('vcodec', 0))
|
||||
quality = traverse_obj(quality_meta, ('now_quality', 0))
|
||||
quality_priority = {
|
||||
'high': 3,
|
||||
'h264': 2,
|
||||
'low': 1,
|
||||
}.get(quality, 0)
|
||||
if traverse_obj(quality_meta, ('hlsaddr', 0, {url_or_none})):
|
||||
formats.append({
|
||||
'format_id': quality,
|
||||
|
@ -73,7 +75,7 @@ class KukuluLiveIE(InfoExtractor):
|
|||
})
|
||||
if traverse_obj(quality_meta, ('hlsaddr_audioonly', 0, {url_or_none})):
|
||||
formats.append({
|
||||
'format_id': join_nonempty(quality, 'audioonly'),
|
||||
'format_id': f'{quality}-audioonly',
|
||||
'url': quality_meta['hlsaddr_audioonly'][0],
|
||||
'ext': 'm4a',
|
||||
'vcodec': 'none',
|
||||
|
@ -84,22 +86,25 @@ class KukuluLiveIE(InfoExtractor):
|
|||
video_id = self._match_id(url)
|
||||
html = self._download_webpage(url, video_id)
|
||||
|
||||
if '>タイムシフトが見つかりませんでした。<' in html:
|
||||
raise ExtractorError('This stream has expired', expected=True)
|
||||
|
||||
title = clean_html(
|
||||
get_element_by_id('livetitle', html.replace('<SPAN', '<span').replace('SPAN>', 'span>')))
|
||||
title = clean_html(get_element_by_id('livetitle', html.replace('<SPAN', '<span').replace('SPAN>', 'span>')))
|
||||
description = self._html_search_meta('Description', html)
|
||||
thumbnail = self._html_search_meta(['og:image', 'twitter:image'], html)
|
||||
|
||||
if self._search_regex(r'(var\s+timeshift\s*=\s*false)', html, 'is livestream', default=False):
|
||||
is_live_stream = 'var timeshift = false;' in html
|
||||
is_vod = 'var timeshift = true;' in html
|
||||
|
||||
if is_live_stream:
|
||||
qualities = [
|
||||
('high', 'Z'),
|
||||
('low', 'ForceLow'),
|
||||
]
|
||||
formats = []
|
||||
for (desc, code) in [('high', 'Z'), ('low', 'ForceLow')]:
|
||||
for (desc, code) in qualities:
|
||||
quality_meta = self._get_quality_meta(video_id, desc, code)
|
||||
self._add_quality_formats(formats, quality_meta)
|
||||
if desc == 'high' and traverse_obj(quality_meta, ('vcodec', 0)) == 'HEVC':
|
||||
self._add_quality_formats(
|
||||
formats, self._get_quality_meta(video_id, desc, code, force_h264='1'))
|
||||
if desc == 'high' and quality_meta.get('vcodec')[0] == 'HEVC':
|
||||
h264_meta = self._get_quality_meta(video_id, desc, code, force_h264='1')
|
||||
self._add_quality_formats(formats, h264_meta)
|
||||
|
||||
return {
|
||||
'id': video_id,
|
||||
|
@ -110,31 +115,42 @@ class KukuluLiveIE(InfoExtractor):
|
|||
'formats': formats,
|
||||
}
|
||||
|
||||
# VOD extraction
|
||||
if is_vod:
|
||||
player_html = self._download_webpage(
|
||||
'https://live.erinn.biz/live.timeshift.fplayer.php', video_id,
|
||||
'Downloading player html', 'Unable to download player html', query={'hash': video_id})
|
||||
query={'hash': video_id},
|
||||
note='Downloading player html',
|
||||
errnote='Unable to download player html')
|
||||
|
||||
sources = traverse_obj(self._search_json(
|
||||
sources_json = self._search_json(
|
||||
r'var\s+fplayer_source\s*=', player_html, 'stream data', video_id,
|
||||
contains_pattern=r'\[(?s:.+)\]', transform_source=js_to_json), lambda _, v: v['file'])
|
||||
contains_pattern=r'\[(?s:.+)\]', transform_source=js_to_json)
|
||||
|
||||
def entries(segments, playlist=True):
|
||||
for i, segment in enumerate(segments, 1):
|
||||
yield {
|
||||
'id': f'{video_id}_{i}' if playlist else video_id,
|
||||
'title': f'{title} (Part {i})' if playlist else title,
|
||||
def parse_segment(segment, segment_id, segment_title):
|
||||
path = segment.get('file')
|
||||
if not path:
|
||||
return None
|
||||
formats = [{
|
||||
'url': urljoin('https://live.erinn.biz', path),
|
||||
'ext': 'mp4',
|
||||
'protocol': 'm3u8_native',
|
||||
}]
|
||||
return {
|
||||
'id': segment_id,
|
||||
'title': segment_title,
|
||||
'description': description,
|
||||
'timestamp': traverse_obj(segment, ('time_start', {int_or_none})),
|
||||
'thumbnail': thumbnail,
|
||||
'formats': [{
|
||||
'url': urljoin('https://live.erinn.biz', segment['file']),
|
||||
'ext': 'mp4',
|
||||
'protocol': 'm3u8_native',
|
||||
}],
|
||||
'formats': formats,
|
||||
}
|
||||
|
||||
if len(sources) == 1:
|
||||
return next(entries(sources, playlist=False))
|
||||
if len(sources_json) == 1:
|
||||
return parse_segment(sources_json[0], video_id, title)
|
||||
|
||||
return self.playlist_result(entries(sources), video_id, title, description, multi_video=True)
|
||||
entries = []
|
||||
for i, segment in enumerate(sources_json):
|
||||
if entry := parse_segment(segment, f'{video_id}_{i}', f'{title} (Part {i + 1})'):
|
||||
entries.append(entry)
|
||||
return self.playlist_result(entries, video_id, title, description, multi_video=True)
|
||||
|
||||
raise ExtractorError('Could not detect media type')
|
||||
|
|
Loading…
Reference in New Issue
Block a user