mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-11-27 01:31:25 +01:00
Compare commits
7 Commits
b206bc3b5c
...
5d6d87ce82
Author | SHA1 | Date | |
---|---|---|---|
|
5d6d87ce82 | ||
|
b83ca24eb7 | ||
|
240a7d43c8 | ||
|
f13df591d4 | ||
|
440238ebc9 | ||
|
5ea1d902d3 | ||
|
47fb5ba647 |
3
.github/workflows/build.yml
vendored
3
.github/workflows/build.yml
vendored
|
@ -504,7 +504,8 @@ jobs:
|
||||||
- windows32
|
- windows32
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/download-artifact@v4
|
- name: Download artifacts
|
||||||
|
uses: actions/download-artifact@v4
|
||||||
with:
|
with:
|
||||||
path: artifact
|
path: artifact
|
||||||
pattern: build-bin-*
|
pattern: build-bin-*
|
||||||
|
|
17
.github/workflows/release-master.yml
vendored
17
.github/workflows/release-master.yml
vendored
|
@ -28,3 +28,20 @@ jobs:
|
||||||
actions: write # For cleaning up cache
|
actions: write # For cleaning up cache
|
||||||
id-token: write # mandatory for trusted publishing
|
id-token: write # mandatory for trusted publishing
|
||||||
secrets: inherit
|
secrets: inherit
|
||||||
|
|
||||||
|
publish_pypi:
|
||||||
|
needs: [release]
|
||||||
|
if: vars.MASTER_PYPI_PROJECT != ''
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
permissions:
|
||||||
|
id-token: write # mandatory for trusted publishing
|
||||||
|
steps:
|
||||||
|
- name: Download artifacts
|
||||||
|
uses: actions/download-artifact@v4
|
||||||
|
with:
|
||||||
|
path: dist
|
||||||
|
name: build-pypi
|
||||||
|
- name: Publish to PyPI
|
||||||
|
uses: pypa/gh-action-pypi-publish@release/v1
|
||||||
|
with:
|
||||||
|
verbose: true
|
||||||
|
|
17
.github/workflows/release-nightly.yml
vendored
17
.github/workflows/release-nightly.yml
vendored
|
@ -41,3 +41,20 @@ jobs:
|
||||||
actions: write # For cleaning up cache
|
actions: write # For cleaning up cache
|
||||||
id-token: write # mandatory for trusted publishing
|
id-token: write # mandatory for trusted publishing
|
||||||
secrets: inherit
|
secrets: inherit
|
||||||
|
|
||||||
|
publish_pypi:
|
||||||
|
needs: [release]
|
||||||
|
if: vars.NIGHTLY_PYPI_PROJECT != ''
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
permissions:
|
||||||
|
id-token: write # mandatory for trusted publishing
|
||||||
|
steps:
|
||||||
|
- name: Download artifacts
|
||||||
|
uses: actions/download-artifact@v4
|
||||||
|
with:
|
||||||
|
path: dist
|
||||||
|
name: build-pypi
|
||||||
|
- name: Publish to PyPI
|
||||||
|
uses: pypa/gh-action-pypi-publish@release/v1
|
||||||
|
with:
|
||||||
|
verbose: true
|
||||||
|
|
19
.github/workflows/release.yml
vendored
19
.github/workflows/release.yml
vendored
|
@ -2,10 +2,6 @@ name: Release
|
||||||
on:
|
on:
|
||||||
workflow_call:
|
workflow_call:
|
||||||
inputs:
|
inputs:
|
||||||
prerelease:
|
|
||||||
required: false
|
|
||||||
default: true
|
|
||||||
type: boolean
|
|
||||||
source:
|
source:
|
||||||
required: false
|
required: false
|
||||||
default: ''
|
default: ''
|
||||||
|
@ -18,6 +14,10 @@ on:
|
||||||
required: false
|
required: false
|
||||||
default: ''
|
default: ''
|
||||||
type: string
|
type: string
|
||||||
|
prerelease:
|
||||||
|
required: false
|
||||||
|
default: true
|
||||||
|
type: boolean
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
inputs:
|
inputs:
|
||||||
source:
|
source:
|
||||||
|
@ -278,11 +278,20 @@ jobs:
|
||||||
make clean-cache
|
make clean-cache
|
||||||
python -m build --no-isolation .
|
python -m build --no-isolation .
|
||||||
|
|
||||||
|
- name: Upload artifacts
|
||||||
|
if: github.event_name != 'workflow_dispatch'
|
||||||
|
uses: actions/upload-artifact@v4
|
||||||
|
with:
|
||||||
|
name: build-pypi
|
||||||
|
path: |
|
||||||
|
dist/*
|
||||||
|
compression-level: 0
|
||||||
|
|
||||||
- name: Publish to PyPI
|
- name: Publish to PyPI
|
||||||
|
if: github.event_name == 'workflow_dispatch'
|
||||||
uses: pypa/gh-action-pypi-publish@release/v1
|
uses: pypa/gh-action-pypi-publish@release/v1
|
||||||
with:
|
with:
|
||||||
verbose: true
|
verbose: true
|
||||||
attestations: false # Currently doesn't work w/ reusable workflows (breaks nightly)
|
|
||||||
|
|
||||||
publish:
|
publish:
|
||||||
needs: [prepare, build]
|
needs: [prepare, build]
|
||||||
|
|
|
@ -52,7 +52,7 @@ default = [
|
||||||
"pycryptodomex",
|
"pycryptodomex",
|
||||||
"requests>=2.32.2,<3",
|
"requests>=2.32.2,<3",
|
||||||
"urllib3>=1.26.17,<3",
|
"urllib3>=1.26.17,<3",
|
||||||
"websockets>=13.0",
|
"websockets>=13.0,<14",
|
||||||
]
|
]
|
||||||
curl-cffi = [
|
curl-cffi = [
|
||||||
"curl-cffi==0.5.10; os_name=='nt' and implementation_name=='cpython'",
|
"curl-cffi==0.5.10; os_name=='nt' and implementation_name=='cpython'",
|
||||||
|
|
|
@ -24,7 +24,7 @@ try:
|
||||||
from Crypto.Cipher import AES, PKCS1_OAEP, Blowfish, PKCS1_v1_5 # noqa: F401
|
from Crypto.Cipher import AES, PKCS1_OAEP, Blowfish, PKCS1_v1_5 # noqa: F401
|
||||||
from Crypto.Hash import CMAC, SHA1 # noqa: F401
|
from Crypto.Hash import CMAC, SHA1 # noqa: F401
|
||||||
from Crypto.PublicKey import RSA # noqa: F401
|
from Crypto.PublicKey import RSA # noqa: F401
|
||||||
except ImportError:
|
except (ImportError, OSError):
|
||||||
__version__ = f'broken {__version__}'.strip()
|
__version__ = f'broken {__version__}'.strip()
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -78,7 +78,7 @@ class LBRYBaseIE(InfoExtractor):
|
||||||
|
|
||||||
return info
|
return info
|
||||||
|
|
||||||
def _fetch_page(self, display_id, url, params, page):
|
def _fetch_page(self, display_id, url, params, metapage_nr, page):
|
||||||
page += 1
|
page += 1
|
||||||
page_params = {
|
page_params = {
|
||||||
'no_totals': True,
|
'no_totals': True,
|
||||||
|
@ -86,8 +86,12 @@ class LBRYBaseIE(InfoExtractor):
|
||||||
'page_size': self._PAGE_SIZE,
|
'page_size': self._PAGE_SIZE,
|
||||||
**params,
|
**params,
|
||||||
}
|
}
|
||||||
|
if metapage_nr == 0:
|
||||||
|
resource = f'page {page}'
|
||||||
|
else:
|
||||||
|
resource = f'page {metapage_nr+1}_{page}'
|
||||||
result = self._call_api_proxy(
|
result = self._call_api_proxy(
|
||||||
'claim_search', display_id, page_params, f'page {page}')
|
'claim_search', display_id, page_params, resource)
|
||||||
for item in traverse_obj(result, ('items', lambda _, v: v['name'] and v['claim_id'])):
|
for item in traverse_obj(result, ('items', lambda _, v: v['name'] and v['claim_id'])):
|
||||||
yield {
|
yield {
|
||||||
**self._parse_stream(item, url),
|
**self._parse_stream(item, url),
|
||||||
|
@ -96,6 +100,32 @@ class LBRYBaseIE(InfoExtractor):
|
||||||
'url': self._permanent_url(url, item['name'], item['claim_id']),
|
'url': self._permanent_url(url, item['name'], item['claim_id']),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
def _metapage_entries(self, display_id, url, params):
|
||||||
|
if 'release_time' in params:
|
||||||
|
raise ExtractorError('release_time isn\'t allowed because _metapage_entires needs to specify it.')
|
||||||
|
if not ('order_by' in params and params['order_by'] == ['release_time']):
|
||||||
|
raise ExtractorError('videos must be sorted by release_time for _metapage_entries to work.')
|
||||||
|
|
||||||
|
last_metapage = []
|
||||||
|
metapage = OnDemandPagedList(
|
||||||
|
functools.partial(self._fetch_page, display_id, url, params, 0),
|
||||||
|
self._PAGE_SIZE).getslice()
|
||||||
|
|
||||||
|
metapage_nr = 1
|
||||||
|
while len(metapage) > 0:
|
||||||
|
yield from metapage
|
||||||
|
|
||||||
|
next_metapage_params = {
|
||||||
|
**params,
|
||||||
|
'release_time': '<={}'.format(metapage[-1]['release_timestamp']),
|
||||||
|
}
|
||||||
|
last_metapage = metapage
|
||||||
|
metapage = OnDemandPagedList(
|
||||||
|
functools.partial(self._fetch_page, display_id, url, next_metapage_params, metapage_nr),
|
||||||
|
self._PAGE_SIZE).getslice()
|
||||||
|
metapage = [x for x in metapage if x not in last_metapage]
|
||||||
|
metapage_nr += 1
|
||||||
|
|
||||||
def _playlist_entries(self, url, display_id, claim_param, metadata):
|
def _playlist_entries(self, url, display_id, claim_param, metadata):
|
||||||
qs = parse_qs(url)
|
qs = parse_qs(url)
|
||||||
content = qs.get('content', [None])[0]
|
content = qs.get('content', [None])[0]
|
||||||
|
@ -123,8 +153,12 @@ class LBRYBaseIE(InfoExtractor):
|
||||||
languages.append('none')
|
languages.append('none')
|
||||||
params['any_languages'] = languages
|
params['any_languages'] = languages
|
||||||
|
|
||||||
|
if qs.get('order', ['new'])[0] == 'new':
|
||||||
|
entries = self._metapage_entries(display_id, url, params)
|
||||||
|
else:
|
||||||
|
self.report_warning('Extraction is limited to 1000 Videos when not sorting by newest.')
|
||||||
entries = OnDemandPagedList(
|
entries = OnDemandPagedList(
|
||||||
functools.partial(self._fetch_page, display_id, url, params),
|
functools.partial(self._fetch_page, display_id, url, params, 0),
|
||||||
self._PAGE_SIZE)
|
self._PAGE_SIZE)
|
||||||
|
|
||||||
return self.playlist_result(
|
return self.playlist_result(
|
||||||
|
|
Loading…
Reference in New Issue
Block a user