Compare commits

..

No commits in common. "37ff05333317b624d1085b6471253862d04aed95" and "69926d80e88f991ab0a45f26a6de1ff8d4d27240" have entirely different histories.

View File

@ -4,12 +4,9 @@ from .common import InfoExtractor
from ..utils import ( from ..utils import (
ExtractorError, ExtractorError,
clean_html, clean_html,
filter_dict,
get_element_by_id, get_element_by_id,
int_or_none, int_or_none,
join_nonempty,
js_to_json, js_to_json,
qualities,
url_or_none, url_or_none,
urljoin, urljoin,
) )
@ -47,22 +44,27 @@ class KukuluLiveIE(InfoExtractor):
'only_matching': True, 'only_matching': True,
}] }]
def _get_quality_meta(self, video_id, desc, code, force_h264=None): def _get_quality_meta(self, video_id, desc, code, force_h264=''):
desc += ' (force_h264)' if force_h264 else '' description = desc if force_h264 == '' else f'{desc} (force_h264)'
qs = self._download_webpage( qs = self._download_webpage(
'https://live.erinn.biz/live.player.fplayer.php', video_id, 'https://live.erinn.biz/live.player.fplayer.php', video_id,
f'Downloading {desc} quality metadata', f'Unable to download {desc} quality metadata', query={
query=filter_dict({
'hash': video_id, 'hash': video_id,
'action': f'get{code}liveByAjax', 'action': f'get{code}liveByAjax',
'force_h264': force_h264, 'force_h264': force_h264,
})) },
note=f'Downloading {description} quality metadata',
errnote=f'Unable to download {description} quality metadata')
return urllib.parse.parse_qs(qs) return urllib.parse.parse_qs(qs)
def _add_quality_formats(self, formats, quality_meta): def _add_quality_formats(self, formats, quality_meta):
vcodec = traverse_obj(quality_meta, ('vcodec', 0, {str})) vcodec = traverse_obj(quality_meta, ('vcodec', 0))
quality = traverse_obj(quality_meta, ('now_quality', 0, {str})) quality = traverse_obj(quality_meta, ('now_quality', 0))
quality_priority = qualities(('low', 'h264', 'high'))(quality) quality_priority = {
'high': 3,
'h264': 2,
'low': 1,
}.get(quality, 0)
if traverse_obj(quality_meta, ('hlsaddr', 0, {url_or_none})): if traverse_obj(quality_meta, ('hlsaddr', 0, {url_or_none})):
formats.append({ formats.append({
'format_id': quality, 'format_id': quality,
@ -73,7 +75,7 @@ class KukuluLiveIE(InfoExtractor):
}) })
if traverse_obj(quality_meta, ('hlsaddr_audioonly', 0, {url_or_none})): if traverse_obj(quality_meta, ('hlsaddr_audioonly', 0, {url_or_none})):
formats.append({ formats.append({
'format_id': join_nonempty(quality, 'audioonly'), 'format_id': f'{quality}-audioonly',
'url': quality_meta['hlsaddr_audioonly'][0], 'url': quality_meta['hlsaddr_audioonly'][0],
'ext': 'm4a', 'ext': 'm4a',
'vcodec': 'none', 'vcodec': 'none',
@ -84,22 +86,25 @@ class KukuluLiveIE(InfoExtractor):
video_id = self._match_id(url) video_id = self._match_id(url)
html = self._download_webpage(url, video_id) html = self._download_webpage(url, video_id)
if '>タイムシフトが見つかりませんでした。<' in html: title = clean_html(get_element_by_id('livetitle', html.replace('<SPAN', '<span').replace('SPAN>', 'span>')))
raise ExtractorError('This stream has expired', expected=True)
title = clean_html(
get_element_by_id('livetitle', html.replace('<SPAN', '<span').replace('SPAN>', 'span>')))
description = self._html_search_meta('Description', html) description = self._html_search_meta('Description', html)
thumbnail = self._html_search_meta(['og:image', 'twitter:image'], html) thumbnail = self._html_search_meta(['og:image', 'twitter:image'], html)
if self._search_regex(r'(var\s+timeshift\s*=\s*false)', html, 'is livestream', default=False): is_live_stream = 'var timeshift = false;' in html
is_vod = 'var timeshift = true;' in html
if is_live_stream:
qualities = [
('high', 'Z'),
('low', 'ForceLow'),
]
formats = [] formats = []
for (desc, code) in [('high', 'Z'), ('low', 'ForceLow')]: for (desc, code) in qualities:
quality_meta = self._get_quality_meta(video_id, desc, code) quality_meta = self._get_quality_meta(video_id, desc, code)
self._add_quality_formats(formats, quality_meta) self._add_quality_formats(formats, quality_meta)
if desc == 'high' and traverse_obj(quality_meta, ('vcodec', 0)) == 'HEVC': if desc == 'high' and quality_meta.get('vcodec')[0] == 'HEVC':
self._add_quality_formats( h264_meta = self._get_quality_meta(video_id, desc, code, force_h264='1')
formats, self._get_quality_meta(video_id, desc, code, force_h264='1')) self._add_quality_formats(formats, h264_meta)
return { return {
'id': video_id, 'id': video_id,
@ -110,31 +115,42 @@ class KukuluLiveIE(InfoExtractor):
'formats': formats, 'formats': formats,
} }
# VOD extraction if is_vod:
player_html = self._download_webpage( player_html = self._download_webpage(
'https://live.erinn.biz/live.timeshift.fplayer.php', video_id, 'https://live.erinn.biz/live.timeshift.fplayer.php', video_id,
'Downloading player html', 'Unable to download player html', query={'hash': video_id}) query={'hash': video_id},
note='Downloading player html',
errnote='Unable to download player html')
sources = traverse_obj(self._search_json( sources_json = self._search_json(
r'var\s+fplayer_source\s*=', player_html, 'stream data', video_id, r'var\s+fplayer_source\s*=', player_html, 'stream data', video_id,
contains_pattern=r'\[(?s:.+)\]', transform_source=js_to_json), lambda _, v: v['file']) contains_pattern=r'\[(?s:.+)\]', transform_source=js_to_json)
def entries(segments, playlist=True): def parse_segment(segment, segment_id, segment_title):
for i, segment in enumerate(segments, 1): path = segment.get('file')
yield { if not path:
'id': f'{video_id}_{i}' if playlist else video_id, return None
'title': f'{title} (Part {i})' if playlist else title, formats = [{
'url': urljoin('https://live.erinn.biz', path),
'ext': 'mp4',
'protocol': 'm3u8_native',
}]
return {
'id': segment_id,
'title': segment_title,
'description': description, 'description': description,
'timestamp': traverse_obj(segment, ('time_start', {int_or_none})), 'timestamp': traverse_obj(segment, ('time_start', {int_or_none})),
'thumbnail': thumbnail, 'thumbnail': thumbnail,
'formats': [{ 'formats': formats,
'url': urljoin('https://live.erinn.biz', segment['file']),
'ext': 'mp4',
'protocol': 'm3u8_native',
}],
} }
if len(sources) == 1: if len(sources_json) == 1:
return next(entries(sources, playlist=False)) return parse_segment(sources_json[0], video_id, title)
return self.playlist_result(entries(sources), video_id, title, description, multi_video=True) entries = []
for i, segment in enumerate(sources_json):
if entry := parse_segment(segment, f'{video_id}_{i}', f'{title} (Part {i + 1})'):
entries.append(entry)
return self.playlist_result(entries, video_id, title, description, multi_video=True)
raise ExtractorError('Could not detect media type')