mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-11-27 01:31:25 +01:00
Compare commits
21 Commits
d080c7eae5
...
86932beabd
Author | SHA1 | Date | |
---|---|---|---|
|
86932beabd | ||
|
b83ca24eb7 | ||
|
240a7d43c8 | ||
|
f13df591d4 | ||
|
357b639718 | ||
|
698680fc3c | ||
|
74a969dc74 | ||
|
017bbfe491 | ||
|
b36acfe977 | ||
|
44d1b925ae | ||
|
e2aafefd23 | ||
|
46e494d8d2 | ||
|
1a735dc0d8 | ||
|
9aee1ace28 | ||
|
0ace89202f | ||
|
21e451ea5a | ||
|
b9c7f716ba | ||
|
1c34d99ab1 | ||
|
b1892db015 | ||
|
bd16713820 | ||
|
308e713d9e |
3
.github/workflows/build.yml
vendored
3
.github/workflows/build.yml
vendored
|
@ -504,7 +504,8 @@ jobs:
|
||||||
- windows32
|
- windows32
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/download-artifact@v4
|
- name: Download artifacts
|
||||||
|
uses: actions/download-artifact@v4
|
||||||
with:
|
with:
|
||||||
path: artifact
|
path: artifact
|
||||||
pattern: build-bin-*
|
pattern: build-bin-*
|
||||||
|
|
17
.github/workflows/release-master.yml
vendored
17
.github/workflows/release-master.yml
vendored
|
@ -28,3 +28,20 @@ jobs:
|
||||||
actions: write # For cleaning up cache
|
actions: write # For cleaning up cache
|
||||||
id-token: write # mandatory for trusted publishing
|
id-token: write # mandatory for trusted publishing
|
||||||
secrets: inherit
|
secrets: inherit
|
||||||
|
|
||||||
|
publish_pypi:
|
||||||
|
needs: [release]
|
||||||
|
if: vars.MASTER_PYPI_PROJECT != ''
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
permissions:
|
||||||
|
id-token: write # mandatory for trusted publishing
|
||||||
|
steps:
|
||||||
|
- name: Download artifacts
|
||||||
|
uses: actions/download-artifact@v4
|
||||||
|
with:
|
||||||
|
path: dist
|
||||||
|
name: build-pypi
|
||||||
|
- name: Publish to PyPI
|
||||||
|
uses: pypa/gh-action-pypi-publish@release/v1
|
||||||
|
with:
|
||||||
|
verbose: true
|
||||||
|
|
17
.github/workflows/release-nightly.yml
vendored
17
.github/workflows/release-nightly.yml
vendored
|
@ -41,3 +41,20 @@ jobs:
|
||||||
actions: write # For cleaning up cache
|
actions: write # For cleaning up cache
|
||||||
id-token: write # mandatory for trusted publishing
|
id-token: write # mandatory for trusted publishing
|
||||||
secrets: inherit
|
secrets: inherit
|
||||||
|
|
||||||
|
publish_pypi:
|
||||||
|
needs: [release]
|
||||||
|
if: vars.NIGHTLY_PYPI_PROJECT != ''
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
permissions:
|
||||||
|
id-token: write # mandatory for trusted publishing
|
||||||
|
steps:
|
||||||
|
- name: Download artifacts
|
||||||
|
uses: actions/download-artifact@v4
|
||||||
|
with:
|
||||||
|
path: dist
|
||||||
|
name: build-pypi
|
||||||
|
- name: Publish to PyPI
|
||||||
|
uses: pypa/gh-action-pypi-publish@release/v1
|
||||||
|
with:
|
||||||
|
verbose: true
|
||||||
|
|
19
.github/workflows/release.yml
vendored
19
.github/workflows/release.yml
vendored
|
@ -2,10 +2,6 @@ name: Release
|
||||||
on:
|
on:
|
||||||
workflow_call:
|
workflow_call:
|
||||||
inputs:
|
inputs:
|
||||||
prerelease:
|
|
||||||
required: false
|
|
||||||
default: true
|
|
||||||
type: boolean
|
|
||||||
source:
|
source:
|
||||||
required: false
|
required: false
|
||||||
default: ''
|
default: ''
|
||||||
|
@ -18,6 +14,10 @@ on:
|
||||||
required: false
|
required: false
|
||||||
default: ''
|
default: ''
|
||||||
type: string
|
type: string
|
||||||
|
prerelease:
|
||||||
|
required: false
|
||||||
|
default: true
|
||||||
|
type: boolean
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
inputs:
|
inputs:
|
||||||
source:
|
source:
|
||||||
|
@ -278,11 +278,20 @@ jobs:
|
||||||
make clean-cache
|
make clean-cache
|
||||||
python -m build --no-isolation .
|
python -m build --no-isolation .
|
||||||
|
|
||||||
|
- name: Upload artifacts
|
||||||
|
if: github.event_name != 'workflow_dispatch'
|
||||||
|
uses: actions/upload-artifact@v4
|
||||||
|
with:
|
||||||
|
name: build-pypi
|
||||||
|
path: |
|
||||||
|
dist/*
|
||||||
|
compression-level: 0
|
||||||
|
|
||||||
- name: Publish to PyPI
|
- name: Publish to PyPI
|
||||||
|
if: github.event_name == 'workflow_dispatch'
|
||||||
uses: pypa/gh-action-pypi-publish@release/v1
|
uses: pypa/gh-action-pypi-publish@release/v1
|
||||||
with:
|
with:
|
||||||
verbose: true
|
verbose: true
|
||||||
attestations: false # Currently doesn't work w/ reusable workflows (breaks nightly)
|
|
||||||
|
|
||||||
publish:
|
publish:
|
||||||
needs: [prepare, build]
|
needs: [prepare, build]
|
||||||
|
|
|
@ -52,7 +52,7 @@ default = [
|
||||||
"pycryptodomex",
|
"pycryptodomex",
|
||||||
"requests>=2.32.2,<3",
|
"requests>=2.32.2,<3",
|
||||||
"urllib3>=1.26.17,<3",
|
"urllib3>=1.26.17,<3",
|
||||||
"websockets>=13.0",
|
"websockets>=13.0,<14",
|
||||||
]
|
]
|
||||||
curl-cffi = [
|
curl-cffi = [
|
||||||
"curl-cffi==0.5.10; os_name=='nt' and implementation_name=='cpython'",
|
"curl-cffi==0.5.10; os_name=='nt' and implementation_name=='cpython'",
|
||||||
|
|
|
@ -24,7 +24,7 @@ try:
|
||||||
from Crypto.Cipher import AES, PKCS1_OAEP, Blowfish, PKCS1_v1_5 # noqa: F401
|
from Crypto.Cipher import AES, PKCS1_OAEP, Blowfish, PKCS1_v1_5 # noqa: F401
|
||||||
from Crypto.Hash import CMAC, SHA1 # noqa: F401
|
from Crypto.Hash import CMAC, SHA1 # noqa: F401
|
||||||
from Crypto.PublicKey import RSA # noqa: F401
|
from Crypto.PublicKey import RSA # noqa: F401
|
||||||
except ImportError:
|
except (ImportError, OSError):
|
||||||
__version__ = f'broken {__version__}'.strip()
|
__version__ = f'broken {__version__}'.strip()
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -4,76 +4,100 @@ import urllib.parse
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
clean_html,
|
ExtractorError,
|
||||||
join_nonempty,
|
int_or_none,
|
||||||
strip_or_none,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
class FptplayIE(InfoExtractor):
|
class FptplayIE(InfoExtractor):
|
||||||
_VALID_URL = r'https?://fptplay\.vn/xem-video/[^/]+\-(?P<id>\w+)(?:/tap-(?P<episode>\d+)?/?(?:[?#]|$)|)'
|
_VALID_URL = r'https?://fptplay\.vn/xem-video/[^/]+\-(?P<id>[a-f0-9]+)'
|
||||||
_GEO_COUNTRIES = ['VN']
|
_GEO_COUNTRIES = ['VN']
|
||||||
IE_NAME = 'fptplay'
|
IE_NAME = 'fptplay'
|
||||||
IE_DESC = 'fptplay.vn'
|
IE_DESC = 'fptplay.vn'
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
'url': 'https://fptplay.vn/xem-video/nhan-duyen-dai-nhan-xin-dung-buoc-621a123016f369ebbde55945',
|
'url': 'https://fptplay.vn/xem-video/jumanji-tro-choi-ky-ao-615c9b232089bd0509bfbf42',
|
||||||
'md5': 'ca0ee9bc63446c0c3e9a90186f7d6b33',
|
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': '621a123016f369ebbde55945',
|
'id': '615c9b232089bd0509bfbf42',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'title': 'Nhân Duyên Đại Nhân Xin Dừng Bước - Tập 1A',
|
'title': 'Jumanji: Welcome To The Jungle',
|
||||||
'description': 'md5:23cf7d1ce0ade8e21e76ae482e6a8c6c',
|
'description': 'Phim theo chân một nhóm bốn học sinh phổ thông bị phạt dọn dẹp tầng hầm trường học. Tại đó, họ phát hiện ra trò chơi cổ mang tên Jumanji.',
|
||||||
|
'thumbnail': 'https://images.fptplay.net/media/OTT/VOD/2023/03/13/jumanji-tro-choi-ky-ao-fpt-play-1678685776013_Background_1920x1080_over.jpg',
|
||||||
|
'release_year': '2017',
|
||||||
},
|
},
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://fptplay.vn/xem-video/ma-toi-la-dai-gia-61f3aa8a6b3b1d2e73c60eb5/tap-3',
|
'url': 'https://fptplay.vn/xem-video/sang-nhu-trang-trong-may-6156d8292089bd2184e26238',
|
||||||
'md5': 'b35be968c909b3e4e1e20ca45dd261b1',
|
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': '61f3aa8a6b3b1d2e73c60eb5',
|
'id': '346034',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'title': 'Má Tôi Là Đại Gia - Tập 3',
|
'title': 'Bright As The Moon',
|
||||||
'description': 'md5:ff8ba62fb6e98ef8875c42edff641d1c',
|
'description': '',
|
||||||
|
'release_year': '2021',
|
||||||
|
'season_number': '1',
|
||||||
|
'episode': 'Tập 1',
|
||||||
|
'episode_number': '1',
|
||||||
|
'duration': '2665',
|
||||||
},
|
},
|
||||||
}, {
|
|
||||||
'url': 'https://fptplay.vn/xem-video/lap-toi-do-giam-under-the-skin-6222d9684ec7230fa6e627a2/tap-4',
|
|
||||||
'md5': 'bcb06c55ec14786d7d4eda07fa1ccbb9',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '6222d9684ec7230fa6e627a2',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': 'Lạp Tội Đồ Giám - Tập 2B',
|
|
||||||
'description': 'md5:e5a47e9d35fbf7e9479ca8a77204908b',
|
|
||||||
},
|
|
||||||
}, {
|
|
||||||
'url': 'https://fptplay.vn/xem-video/nha-co-chuyen-hi-alls-well-ends-well-1997-6218995f6af792ee370459f0',
|
|
||||||
'only_matching': True,
|
|
||||||
}]
|
}]
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
video_id, slug_episode = self._match_valid_url(url).group('id', 'episode')
|
contentId = self._match_id(url)
|
||||||
webpage = self._download_webpage(url, video_id=video_id, fatal=False) or ''
|
# Need valid cookie with Bearer token, else it won't work
|
||||||
title = self._search_regex(
|
token = self._get_cookies(url).get('token')
|
||||||
r'(?s)<h4\s+class="mb-1 text-2xl text-white"[^>]*>(.+)</h4>', webpage, 'title', fatal=False)
|
res = self._download_json(self.get_api_with_st_token(contentId), contentId, expected_status=406)
|
||||||
real_episode = slug_episode if not title else self._search_regex(
|
if res['result']['episode_type'] == 0:
|
||||||
r'<p.+title="(?P<episode>[^">]+)"\s+class="epi-title active"', webpage, 'episode', fatal=False)
|
# movie or single video
|
||||||
title = strip_or_none(title) or self._html_search_meta(('og:title', 'twitter:title'), webpage)
|
manifest = self._download_json(self.get_api_with_st_token(contentId, 0), contentId, headers={'authorization': f'Bearer {token.value}'}, expected_status=406)
|
||||||
|
if manifest.get('msg') != 'success':
|
||||||
|
raise ExtractorError(f" - Got an error, response: {manifest.get('msg')}", expected=True)
|
||||||
|
formats, subtitles = self._extract_m3u8_formats_and_subtitles(manifest['data']['url'], contentId)
|
||||||
|
return {
|
||||||
|
'id': contentId,
|
||||||
|
'title': res['result']['title_origin'] if res['result']['title_origin'] else res['result']['title_vie'],
|
||||||
|
'description': res['result']['description'],
|
||||||
|
'thumbnail': res['result']['thumb'],
|
||||||
|
'release_year': int_or_none(res['result']['movie_release_date']),
|
||||||
|
'duration': int_or_none(res['result']['duration']),
|
||||||
|
'formats': formats,
|
||||||
|
'subtitles': subtitles,
|
||||||
|
}
|
||||||
|
else:
|
||||||
|
# playlist
|
||||||
|
entries = []
|
||||||
|
for episode in res['result']['episodes']:
|
||||||
|
if episode['is_trailer'] == 1:
|
||||||
|
continue
|
||||||
|
manifest = self._download_json(self.get_api_with_st_token(contentId, episode['_id']), episode['_id'], headers={'authorization': f'Bearer {token.value}'}, expected_status=406)
|
||||||
|
if manifest.get('msg') != 'success':
|
||||||
|
raise ExtractorError(f" - Got an error, response: {manifest.get('msg')}", expected=True)
|
||||||
|
formats, subtitles = self._extract_m3u8_formats_and_subtitles(manifest['data']['url'], episode['_id'])
|
||||||
|
entry = {
|
||||||
|
'id': episode['ref_episode_id'],
|
||||||
|
'title': res['result']['title_origin'] if res['result']['title_origin'] else res['result']['title_vie'],
|
||||||
|
'description': episode['description'],
|
||||||
|
'thumbnail': episode['thumb'],
|
||||||
|
'release_year': int_or_none(res['result']['movie_release_date']),
|
||||||
|
'season_number': 1, # Assuming season 1 for simplicity
|
||||||
|
'episode': episode['title'],
|
||||||
|
'episode_number': episode['_id'] + 1,
|
||||||
|
'duration': int_or_none(episode['duration']),
|
||||||
|
'formats': formats,
|
||||||
|
'subtitles': subtitles,
|
||||||
|
}
|
||||||
|
entries.append(entry)
|
||||||
|
return {
|
||||||
|
'_type': 'playlist',
|
||||||
|
'id': contentId,
|
||||||
|
'title': res['result']['title_origin'] if res['result']['title_origin'] else res['result']['title_vie'],
|
||||||
|
'entries': entries,
|
||||||
|
}
|
||||||
|
|
||||||
info = self._download_json(
|
def get_api_with_st_token(self, video_id, episode=None):
|
||||||
self.get_api_with_st_token(video_id, int(slug_episode) - 1 if slug_episode else 0), video_id)
|
if episode is not None:
|
||||||
formats, subtitles = self._extract_m3u8_formats_and_subtitles(info['data']['url'], video_id, 'mp4')
|
path = f'/api/v7.1_w/stream/vod/{video_id}/{0 if episode is None else episode}/adaptive_bitrate'
|
||||||
return {
|
else:
|
||||||
'id': video_id,
|
path = f'/api/v7.1_w/vod/detail/{video_id}'
|
||||||
'title': join_nonempty(title, real_episode, delim=' - '),
|
|
||||||
'description': (
|
|
||||||
clean_html(self._search_regex(r'<p\s+class="overflow-hidden"[^>]*>(.+)</p>', webpage, 'description'))
|
|
||||||
or self._html_search_meta(('og:description', 'twitter:description'), webpage)),
|
|
||||||
'formats': formats,
|
|
||||||
'subtitles': subtitles,
|
|
||||||
}
|
|
||||||
|
|
||||||
def get_api_with_st_token(self, video_id, episode):
|
|
||||||
path = f'/api/v6.2_w/stream/vod/{video_id}/{episode}/auto_vip'
|
|
||||||
timestamp = int(time.time()) + 10800
|
timestamp = int(time.time()) + 10800
|
||||||
|
t = hashlib.md5(f'6ea6d2a4e2d3a4bd5e275401aa086d{timestamp}{path}'.encode()).hexdigest().upper()
|
||||||
t = hashlib.md5(f'WEBv6Dkdsad90dasdjlALDDDS{timestamp}{path}'.encode()).hexdigest().upper()
|
|
||||||
r = 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/'
|
r = 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/'
|
||||||
n = [int(f'0x{t[2 * o: 2 * o + 2]}', 16) for o in range(len(t) // 2)]
|
n = [int(f'0x{t[2 * o: 2 * o + 2]}', 16) for o in range(len(t) // 2)]
|
||||||
|
|
||||||
|
@ -89,7 +113,7 @@ class FptplayIE(InfoExtractor):
|
||||||
i[n] = e[c]
|
i[n] = e[c]
|
||||||
n += 1
|
n += 1
|
||||||
c += 1
|
c += 1
|
||||||
if 3 == n:
|
if n == 3:
|
||||||
a[0] = (252 & i[0]) >> 2
|
a[0] = (252 & i[0]) >> 2
|
||||||
a[1] = ((3 & i[0]) << 4) + ((240 & i[1]) >> 4)
|
a[1] = ((3 & i[0]) << 4) + ((240 & i[1]) >> 4)
|
||||||
a[2] = ((15 & i[1]) << 2) + ((192 & i[2]) >> 6)
|
a[2] = ((15 & i[1]) << 2) + ((192 & i[2]) >> 6)
|
||||||
|
@ -100,18 +124,16 @@ class FptplayIE(InfoExtractor):
|
||||||
if n:
|
if n:
|
||||||
for o in range(n, 3):
|
for o in range(n, 3):
|
||||||
i[o] = 0
|
i[o] = 0
|
||||||
|
|
||||||
for o in range(n + 1):
|
for o in range(n + 1):
|
||||||
a[0] = (252 & i[0]) >> 2
|
a[0] = (252 & i[0]) >> 2
|
||||||
a[1] = ((3 & i[0]) << 4) + ((240 & i[1]) >> 4)
|
a[1] = ((3 & i[0]) << 4) + ((240 & i[1]) >> 4)
|
||||||
a[2] = ((15 & i[1]) << 2) + ((192 & i[2]) >> 6)
|
a[2] = ((15 & i[1]) << 2) + ((192 & i[2]) >> 6)
|
||||||
a[3] = (63 & i[2])
|
a[3] = (63 & i[2])
|
||||||
t += r[a[o]]
|
t += r[a[o]]
|
||||||
n += 1
|
n += 1
|
||||||
while n < 3:
|
while n < 3:
|
||||||
t += ''
|
t += ''
|
||||||
n += 1
|
n += 1
|
||||||
return t
|
return t
|
||||||
|
|
||||||
st_token = convert(n).replace('+', '-').replace('/', '_').replace('=', '')
|
st_token = convert(n).replace('+', '-').replace('/', '_').replace('=', '')
|
||||||
return f'https://api.fptplay.net{path}?{urllib.parse.urlencode({"st": st_token, "e": timestamp})}'
|
return f"https://api.fptplay.net{path}?{urllib.parse.urlencode({'st': st_token, 'e': timestamp})}"
|
||||||
|
|
Loading…
Reference in New Issue
Block a user