mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-11-24 00:01:54 +01:00
Compare commits
15 Commits
b5a1447ea7
...
d4395538fd
Author | SHA1 | Date | |
---|---|---|---|
|
d4395538fd | ||
|
f919729538 | ||
|
7ea2787920 | ||
|
f7257588bd | ||
|
da252d9d32 | ||
|
e079ffbda6 | ||
|
2009cb27e1 | ||
|
f351440f1d | ||
|
f9d98509a8 | ||
|
37cd7660ea | ||
|
d867f99622 | ||
|
10fc719bc7 | ||
|
eb15fd5a32 | ||
|
7cecd299e4 | ||
|
52c0ffe40a |
12
CONTRIBUTORS
12
CONTRIBUTORS
|
@ -695,3 +695,15 @@ KBelmin
|
||||||
kesor
|
kesor
|
||||||
MellowKyler
|
MellowKyler
|
||||||
Wesley107772
|
Wesley107772
|
||||||
|
a13ssandr0
|
||||||
|
ChocoLZS
|
||||||
|
doe1080
|
||||||
|
hugovdev
|
||||||
|
jshumphrey
|
||||||
|
julionc
|
||||||
|
manavchaudhary1
|
||||||
|
powergold1
|
||||||
|
Sakura286
|
||||||
|
SamDecrock
|
||||||
|
stratus-ss
|
||||||
|
subrat-lima
|
||||||
|
|
58
Changelog.md
58
Changelog.md
|
@ -4,6 +4,64 @@
|
||||||
# To create a release, dispatch the https://github.com/yt-dlp/yt-dlp/actions/workflows/release.yml workflow on master
|
# To create a release, dispatch the https://github.com/yt-dlp/yt-dlp/actions/workflows/release.yml workflow on master
|
||||||
-->
|
-->
|
||||||
|
|
||||||
|
### 2024.11.18
|
||||||
|
|
||||||
|
#### Important changes
|
||||||
|
- **Login with OAuth is no longer supported for YouTube**
|
||||||
|
Due to a change made by the site, yt-dlp is longer able to support OAuth login for YouTube. [Read more](https://github.com/yt-dlp/yt-dlp/issues/11462#issuecomment-2471703090)
|
||||||
|
|
||||||
|
#### Core changes
|
||||||
|
- [Catch broken Cryptodome installations](https://github.com/yt-dlp/yt-dlp/commit/b83ca24eb72e1e558b0185bd73975586c0bc0546) ([#11486](https://github.com/yt-dlp/yt-dlp/issues/11486)) by [seproDev](https://github.com/seproDev)
|
||||||
|
- **utils**
|
||||||
|
- [Fix `join_nonempty`, add `**kwargs` to `unpack`](https://github.com/yt-dlp/yt-dlp/commit/39d79c9b9cf23411d935910685c40aa1a2fdb409) ([#11559](https://github.com/yt-dlp/yt-dlp/issues/11559)) by [Grub4K](https://github.com/Grub4K)
|
||||||
|
- `subs_list_to_dict`: [Add `lang` default parameter](https://github.com/yt-dlp/yt-dlp/commit/c014fbcddcb4c8f79d914ac5bb526758b540ea33) ([#11508](https://github.com/yt-dlp/yt-dlp/issues/11508)) by [Grub4K](https://github.com/Grub4K)
|
||||||
|
|
||||||
|
#### Extractor changes
|
||||||
|
- [Allow `ext` override for thumbnails](https://github.com/yt-dlp/yt-dlp/commit/eb64ae7d5def6df2aba74fb703e7f168fb299865) ([#11545](https://github.com/yt-dlp/yt-dlp/issues/11545)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **adobepass**: [Fix provider requests](https://github.com/yt-dlp/yt-dlp/commit/85fdc66b6e01d19a94b4f39b58e3c0cf23600902) ([#11472](https://github.com/yt-dlp/yt-dlp/issues/11472)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **archive.org**: [Fix comments extraction](https://github.com/yt-dlp/yt-dlp/commit/f2a4983df7a64c4e93b56f79dbd16a781bd90206) ([#11527](https://github.com/yt-dlp/yt-dlp/issues/11527)) by [jshumphrey](https://github.com/jshumphrey)
|
||||||
|
- **bandlab**: [Add extractors](https://github.com/yt-dlp/yt-dlp/commit/6365e92589e4bc17b8fffb0125a716d144ad2137) ([#11535](https://github.com/yt-dlp/yt-dlp/issues/11535)) by [seproDev](https://github.com/seproDev)
|
||||||
|
- **chaturbate**
|
||||||
|
- [Extract from API and support impersonation](https://github.com/yt-dlp/yt-dlp/commit/720b3dc453c342bc2e8df7dbc0acaab4479de46c) ([#11555](https://github.com/yt-dlp/yt-dlp/issues/11555)) by [powergold1](https://github.com/powergold1) (With fixes in [7cecd29](https://github.com/yt-dlp/yt-dlp/commit/7cecd299e4a5ef1f0f044b2fedc26f17e41f15e3) by [seproDev](https://github.com/seproDev))
|
||||||
|
- [Support alternate domains](https://github.com/yt-dlp/yt-dlp/commit/a9f85670d03ab993dc589f21a9ffffcad61392d5) ([#10595](https://github.com/yt-dlp/yt-dlp/issues/10595)) by [manavchaudhary1](https://github.com/manavchaudhary1)
|
||||||
|
- **cloudflarestream**: [Avoid extraction via videodelivery.net](https://github.com/yt-dlp/yt-dlp/commit/2db8c2e7d57a1784b06057c48e3e91023720d195) ([#11478](https://github.com/yt-dlp/yt-dlp/issues/11478)) by [hugovdev](https://github.com/hugovdev)
|
||||||
|
- **ctvnews**
|
||||||
|
- [Fix extractor](https://github.com/yt-dlp/yt-dlp/commit/f351440f1dc5b3dfbfc5737b037a869d946056fe) ([#11534](https://github.com/yt-dlp/yt-dlp/issues/11534)) by [bashonly](https://github.com/bashonly), [jshumphrey](https://github.com/jshumphrey)
|
||||||
|
- [Fix playlist ID extraction](https://github.com/yt-dlp/yt-dlp/commit/f9d98509a898737c12977b2e2117277bada2c196) ([#8892](https://github.com/yt-dlp/yt-dlp/issues/8892)) by [qbnu](https://github.com/qbnu)
|
||||||
|
- **digitalconcerthall**: [Support login with access/refresh tokens](https://github.com/yt-dlp/yt-dlp/commit/f7257588bdff5f0b0452635a66b253a783c97357) ([#11571](https://github.com/yt-dlp/yt-dlp/issues/11571)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **facebook**: [Fix formats extraction](https://github.com/yt-dlp/yt-dlp/commit/bacc31b05a04181b63100c481565256b14813a5e) ([#11513](https://github.com/yt-dlp/yt-dlp/issues/11513)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **gamedevtv**: [Add extractor](https://github.com/yt-dlp/yt-dlp/commit/be3579aaf0c3b71a0a3195e1955415d5e4d6b3d8) ([#11368](https://github.com/yt-dlp/yt-dlp/issues/11368)) by [bashonly](https://github.com/bashonly), [stratus-ss](https://github.com/stratus-ss)
|
||||||
|
- **goplay**: [Fix extractor](https://github.com/yt-dlp/yt-dlp/commit/6b43a8d84b881d769b480ba6e20ec691e9d1b92d) ([#11466](https://github.com/yt-dlp/yt-dlp/issues/11466)) by [bashonly](https://github.com/bashonly), [SamDecrock](https://github.com/SamDecrock)
|
||||||
|
- **kenh14**: [Add extractor](https://github.com/yt-dlp/yt-dlp/commit/eb15fd5a32d8b35ef515f7a3d1158c03025648ff) ([#3996](https://github.com/yt-dlp/yt-dlp/issues/3996)) by [krichbanana](https://github.com/krichbanana), [pzhlkj6612](https://github.com/pzhlkj6612)
|
||||||
|
- **litv**: [Fix extractor](https://github.com/yt-dlp/yt-dlp/commit/e079ffbda66de150c0a9ebef05e89f61bb4d5f76) ([#11071](https://github.com/yt-dlp/yt-dlp/issues/11071)) by [jiru](https://github.com/jiru)
|
||||||
|
- **mixchmovie**: [Add extractor](https://github.com/yt-dlp/yt-dlp/commit/0ec9bfed4d4a52bfb4f8733da1acf0aeeae21e6b) ([#10897](https://github.com/yt-dlp/yt-dlp/issues/10897)) by [Sakura286](https://github.com/Sakura286)
|
||||||
|
- **patreon**: [Fix comments extraction](https://github.com/yt-dlp/yt-dlp/commit/1d253b0a27110d174c40faf8fb1c999d099e0cde) ([#11530](https://github.com/yt-dlp/yt-dlp/issues/11530)) by [bashonly](https://github.com/bashonly), [jshumphrey](https://github.com/jshumphrey)
|
||||||
|
- **pialive**: [Add extractor](https://github.com/yt-dlp/yt-dlp/commit/d867f99622ef7fba690b08da56c39d739b822bb7) ([#10811](https://github.com/yt-dlp/yt-dlp/issues/10811)) by [ChocoLZS](https://github.com/ChocoLZS)
|
||||||
|
- **radioradicale**: [Add extractor](https://github.com/yt-dlp/yt-dlp/commit/70c55cb08f780eab687e881ef42bb5c6007d290b) ([#5607](https://github.com/yt-dlp/yt-dlp/issues/5607)) by [a13ssandr0](https://github.com/a13ssandr0), [pzhlkj6612](https://github.com/pzhlkj6612)
|
||||||
|
- **reddit**: [Improve error handling](https://github.com/yt-dlp/yt-dlp/commit/7ea2787920cccc6b8ea30791993d114fbd564434) ([#11573](https://github.com/yt-dlp/yt-dlp/issues/11573)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **redgifsuser**: [Fix extraction](https://github.com/yt-dlp/yt-dlp/commit/d215fba7edb69d4fa665f43663756fd260b1489f) ([#11531](https://github.com/yt-dlp/yt-dlp/issues/11531)) by [jshumphrey](https://github.com/jshumphrey)
|
||||||
|
- **rutube**: [Rework extractors](https://github.com/yt-dlp/yt-dlp/commit/e398217aae19bb25f91797bfbe8a3243698d7f45) ([#11480](https://github.com/yt-dlp/yt-dlp/issues/11480)) by [seproDev](https://github.com/seproDev)
|
||||||
|
- **sonylivseries**: [Add `sort_order` extractor-arg](https://github.com/yt-dlp/yt-dlp/commit/2009cb27e17014787bf63eaa2ada51293d54f22a) ([#11569](https://github.com/yt-dlp/yt-dlp/issues/11569)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **soop**: [Fix thumbnail extraction](https://github.com/yt-dlp/yt-dlp/commit/c699bafc5038b59c9afe8c2e69175fb66424c832) ([#11545](https://github.com/yt-dlp/yt-dlp/issues/11545)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **spankbang**: [Support browser impersonation](https://github.com/yt-dlp/yt-dlp/commit/8388ec256f7753b02488788e3cfa771f6e1db247) ([#11542](https://github.com/yt-dlp/yt-dlp/issues/11542)) by [jshumphrey](https://github.com/jshumphrey)
|
||||||
|
- **spreaker**
|
||||||
|
- [Support episode pages and access keys](https://github.com/yt-dlp/yt-dlp/commit/c39016f66df76d14284c705736ca73db8055d8de) ([#11489](https://github.com/yt-dlp/yt-dlp/issues/11489)) by [julionc](https://github.com/julionc)
|
||||||
|
- [Support podcast and feed pages](https://github.com/yt-dlp/yt-dlp/commit/c6737310619022248f5d0fd13872073cac168453) ([#10968](https://github.com/yt-dlp/yt-dlp/issues/10968)) by [subrat-lima](https://github.com/subrat-lima)
|
||||||
|
- **youtube**
|
||||||
|
- [Player client maintenance](https://github.com/yt-dlp/yt-dlp/commit/637d62a3a9fc723d68632c1af25c30acdadeeb85) ([#11528](https://github.com/yt-dlp/yt-dlp/issues/11528)) by [bashonly](https://github.com/bashonly), [seproDev](https://github.com/seproDev)
|
||||||
|
- [Remove broken OAuth support](https://github.com/yt-dlp/yt-dlp/commit/52c0ffe40ad6e8404d93296f575007b05b04c686) ([#11558](https://github.com/yt-dlp/yt-dlp/issues/11558)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- tab: [Fix podcasts tab extraction](https://github.com/yt-dlp/yt-dlp/commit/37cd7660eaff397c551ee18d80507702342b0c2b) ([#11567](https://github.com/yt-dlp/yt-dlp/issues/11567)) by [seproDev](https://github.com/seproDev)
|
||||||
|
|
||||||
|
#### Misc. changes
|
||||||
|
- **build**
|
||||||
|
- [Bump PyInstaller version pin to `>=6.11.1`](https://github.com/yt-dlp/yt-dlp/commit/f9c8deb4e5887ff5150e911ac0452e645f988044) ([#11507](https://github.com/yt-dlp/yt-dlp/issues/11507)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- [Enable attestations for trusted publishing](https://github.com/yt-dlp/yt-dlp/commit/f13df591d4d7ca8e2f31b35c9c91e69ba9e9b013) ([#11420](https://github.com/yt-dlp/yt-dlp/issues/11420)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- [Pin `websockets` version to >=13.0,<14](https://github.com/yt-dlp/yt-dlp/commit/240a7d43c8a67ffb86d44dc276805aa43c358dcc) ([#11488](https://github.com/yt-dlp/yt-dlp/issues/11488)) by [bashonly](https://github.com/bashonly)
|
||||||
|
- **cleanup**
|
||||||
|
- [Deprecate more compat functions](https://github.com/yt-dlp/yt-dlp/commit/f95a92b3d0169a784ee15a138fbe09d82b2754a1) ([#11439](https://github.com/yt-dlp/yt-dlp/issues/11439)) by [seproDev](https://github.com/seproDev)
|
||||||
|
- [Remove dead extractors](https://github.com/yt-dlp/yt-dlp/commit/10fc719bc7f1eef469389c5219102266ef411f29) ([#11566](https://github.com/yt-dlp/yt-dlp/issues/11566)) by [doe1080](https://github.com/doe1080)
|
||||||
|
- Miscellaneous: [da252d9](https://github.com/yt-dlp/yt-dlp/commit/da252d9d322af3e2178ac5eae324809502a0a862) by [bashonly](https://github.com/bashonly), [Grub4K](https://github.com/Grub4K), [seproDev](https://github.com/seproDev)
|
||||||
|
|
||||||
### 2024.11.04
|
### 2024.11.04
|
||||||
|
|
||||||
#### Important changes
|
#### Important changes
|
||||||
|
|
|
@ -342,8 +342,9 @@ If you fork the project on GitHub, you can run your fork's [build workflow](.git
|
||||||
extractor plugins; postprocessor plugins can
|
extractor plugins; postprocessor plugins can
|
||||||
only be loaded from the default plugin
|
only be loaded from the default plugin
|
||||||
directories
|
directories
|
||||||
--flat-playlist Do not extract the videos of a playlist,
|
--flat-playlist Do not extract a playlist's URL result
|
||||||
only list them
|
entries; some entry metadata may be missing
|
||||||
|
and downloading may be bypassed
|
||||||
--no-flat-playlist Fully extract the videos of a playlist
|
--no-flat-playlist Fully extract the videos of a playlist
|
||||||
(default)
|
(default)
|
||||||
--live-from-start Download livestreams from the start.
|
--live-from-start Download livestreams from the start.
|
||||||
|
@ -1866,8 +1867,8 @@ The following extractors use this feature:
|
||||||
#### bilibili
|
#### bilibili
|
||||||
* `prefer_multi_flv`: Prefer extracting flv formats over mp4 for older videos that still provide legacy formats
|
* `prefer_multi_flv`: Prefer extracting flv formats over mp4 for older videos that still provide legacy formats
|
||||||
|
|
||||||
#### digitalconcerthall
|
#### sonylivseries
|
||||||
* `prefer_combined_hls`: Prefer extracting combined/pre-merged video and audio HLS formats. This will exclude 4K/HEVC video and lossless/FLAC audio formats, which are only available as split video/audio HLS formats
|
* `sort_order`: Episode sort order for series extraction - one of `asc` (ascending, oldest first) or `desc` (descending, newest first). Default is `asc`
|
||||||
|
|
||||||
**Note**: These options may be changed/removed in the future without concern for backward compatibility
|
**Note**: These options may be changed/removed in the future without concern for backward compatibility
|
||||||
|
|
||||||
|
|
|
@ -234,5 +234,10 @@
|
||||||
"when": "57212a5f97ce367590aaa5c3e9a135eead8f81f7",
|
"when": "57212a5f97ce367590aaa5c3e9a135eead8f81f7",
|
||||||
"short": "[ie/vimeo] Fix API retries (#11351)",
|
"short": "[ie/vimeo] Fix API retries (#11351)",
|
||||||
"authors": ["bashonly"]
|
"authors": ["bashonly"]
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"action": "add",
|
||||||
|
"when": "52c0ffe40ad6e8404d93296f575007b05b04c686",
|
||||||
|
"short": "[priority] **Login with OAuth is no longer supported for YouTube**\nDue to a change made by the site, yt-dlp is longer able to support OAuth login for YouTube. [Read more](https://github.com/yt-dlp/yt-dlp/issues/11462#issuecomment-2471703090)"
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
|
|
|
@ -129,6 +129,8 @@
|
||||||
- **Bandcamp:album**
|
- **Bandcamp:album**
|
||||||
- **Bandcamp:user**
|
- **Bandcamp:user**
|
||||||
- **Bandcamp:weekly**
|
- **Bandcamp:weekly**
|
||||||
|
- **Bandlab**
|
||||||
|
- **BandlabPlaylist**
|
||||||
- **BannedVideo**
|
- **BannedVideo**
|
||||||
- **bbc**: [*bbc*](## "netrc machine") BBC
|
- **bbc**: [*bbc*](## "netrc machine") BBC
|
||||||
- **bbc.co.uk**: [*bbc*](## "netrc machine") BBC iPlayer
|
- **bbc.co.uk**: [*bbc*](## "netrc machine") BBC iPlayer
|
||||||
|
@ -484,6 +486,7 @@
|
||||||
- **Gab**
|
- **Gab**
|
||||||
- **GabTV**
|
- **GabTV**
|
||||||
- **Gaia**: [*gaia*](## "netrc machine")
|
- **Gaia**: [*gaia*](## "netrc machine")
|
||||||
|
- **GameDevTVDashboard**: [*gamedevtv*](## "netrc machine")
|
||||||
- **GameJolt**
|
- **GameJolt**
|
||||||
- **GameJoltCommunity**
|
- **GameJoltCommunity**
|
||||||
- **GameJoltGame**
|
- **GameJoltGame**
|
||||||
|
@ -651,6 +654,8 @@
|
||||||
- **Karaoketv**
|
- **Karaoketv**
|
||||||
- **Katsomo**: (**Currently broken**)
|
- **Katsomo**: (**Currently broken**)
|
||||||
- **KelbyOne**: (**Currently broken**)
|
- **KelbyOne**: (**Currently broken**)
|
||||||
|
- **Kenh14Playlist**
|
||||||
|
- **Kenh14Video**
|
||||||
- **Ketnet**
|
- **Ketnet**
|
||||||
- **khanacademy**
|
- **khanacademy**
|
||||||
- **khanacademy:unit**
|
- **khanacademy:unit**
|
||||||
|
@ -784,10 +789,6 @@
|
||||||
- **MicrosoftLearnSession**
|
- **MicrosoftLearnSession**
|
||||||
- **MicrosoftMedius**
|
- **MicrosoftMedius**
|
||||||
- **microsoftstream**: Microsoft Stream
|
- **microsoftstream**: Microsoft Stream
|
||||||
- **mildom**: Record ongoing live by specific user in Mildom
|
|
||||||
- **mildom:clip**: Clip in Mildom
|
|
||||||
- **mildom:user:vod**: Download all VODs from specific user in Mildom
|
|
||||||
- **mildom:vod**: VOD in Mildom
|
|
||||||
- **minds**
|
- **minds**
|
||||||
- **minds:channel**
|
- **minds:channel**
|
||||||
- **minds:group**
|
- **minds:group**
|
||||||
|
@ -798,6 +799,7 @@
|
||||||
- **MiTele**: mitele.es
|
- **MiTele**: mitele.es
|
||||||
- **mixch**
|
- **mixch**
|
||||||
- **mixch:archive**
|
- **mixch:archive**
|
||||||
|
- **mixch:movie**
|
||||||
- **mixcloud**
|
- **mixcloud**
|
||||||
- **mixcloud:playlist**
|
- **mixcloud:playlist**
|
||||||
- **mixcloud:user**
|
- **mixcloud:user**
|
||||||
|
@ -1060,8 +1062,8 @@
|
||||||
- **PhilharmonieDeParis**: Philharmonie de Paris
|
- **PhilharmonieDeParis**: Philharmonie de Paris
|
||||||
- **phoenix.de**
|
- **phoenix.de**
|
||||||
- **Photobucket**
|
- **Photobucket**
|
||||||
|
- **PiaLive**
|
||||||
- **Piapro**: [*piapro*](## "netrc machine")
|
- **Piapro**: [*piapro*](## "netrc machine")
|
||||||
- **PIAULIZAPortal**: ulizaportal.jp - PIA LIVE STREAM
|
|
||||||
- **Picarto**
|
- **Picarto**
|
||||||
- **PicartoVod**
|
- **PicartoVod**
|
||||||
- **Piksel**
|
- **Piksel**
|
||||||
|
@ -1088,8 +1090,6 @@
|
||||||
- **PodbayFMChannel**
|
- **PodbayFMChannel**
|
||||||
- **Podchaser**
|
- **Podchaser**
|
||||||
- **podomatic**: (**Currently broken**)
|
- **podomatic**: (**Currently broken**)
|
||||||
- **Pokemon**
|
|
||||||
- **PokemonWatch**
|
|
||||||
- **PokerGo**: [*pokergo*](## "netrc machine")
|
- **PokerGo**: [*pokergo*](## "netrc machine")
|
||||||
- **PokerGoCollection**: [*pokergo*](## "netrc machine")
|
- **PokerGoCollection**: [*pokergo*](## "netrc machine")
|
||||||
- **PolsatGo**
|
- **PolsatGo**
|
||||||
|
@ -1160,6 +1160,7 @@
|
||||||
- **RadioJavan**: (**Currently broken**)
|
- **RadioJavan**: (**Currently broken**)
|
||||||
- **radiokapital**
|
- **radiokapital**
|
||||||
- **radiokapital:show**
|
- **radiokapital:show**
|
||||||
|
- **RadioRadicale**
|
||||||
- **RadioZetPodcast**
|
- **RadioZetPodcast**
|
||||||
- **radlive**
|
- **radlive**
|
||||||
- **radlive:channel**
|
- **radlive:channel**
|
||||||
|
@ -1367,9 +1368,7 @@
|
||||||
- **spotify**: Spotify episodes (**Currently broken**)
|
- **spotify**: Spotify episodes (**Currently broken**)
|
||||||
- **spotify:show**: Spotify shows (**Currently broken**)
|
- **spotify:show**: Spotify shows (**Currently broken**)
|
||||||
- **Spreaker**
|
- **Spreaker**
|
||||||
- **SpreakerPage**
|
|
||||||
- **SpreakerShow**
|
- **SpreakerShow**
|
||||||
- **SpreakerShowPage**
|
|
||||||
- **SpringboardPlatform**
|
- **SpringboardPlatform**
|
||||||
- **Sprout**
|
- **Sprout**
|
||||||
- **SproutVideo**
|
- **SproutVideo**
|
||||||
|
@ -1570,6 +1569,8 @@
|
||||||
- **UFCTV**: [*ufctv*](## "netrc machine")
|
- **UFCTV**: [*ufctv*](## "netrc machine")
|
||||||
- **ukcolumn**: (**Currently broken**)
|
- **ukcolumn**: (**Currently broken**)
|
||||||
- **UKTVPlay**
|
- **UKTVPlay**
|
||||||
|
- **UlizaPlayer**
|
||||||
|
- **UlizaPortal**: ulizaportal.jp
|
||||||
- **umg:de**: Universal Music Deutschland (**Currently broken**)
|
- **umg:de**: Universal Music Deutschland (**Currently broken**)
|
||||||
- **Unistra**
|
- **Unistra**
|
||||||
- **Unity**: (**Currently broken**)
|
- **Unity**: (**Currently broken**)
|
||||||
|
@ -1587,8 +1588,6 @@
|
||||||
- **Varzesh3**: (**Currently broken**)
|
- **Varzesh3**: (**Currently broken**)
|
||||||
- **Vbox7**
|
- **Vbox7**
|
||||||
- **Veo**
|
- **Veo**
|
||||||
- **Veoh**
|
|
||||||
- **veoh:user**
|
|
||||||
- **Vesti**: Вести.Ru (**Currently broken**)
|
- **Vesti**: Вести.Ru (**Currently broken**)
|
||||||
- **Vevo**
|
- **Vevo**
|
||||||
- **VevoPlaylist**
|
- **VevoPlaylist**
|
||||||
|
|
|
@ -946,6 +946,10 @@ from .kaltura import KalturaIE
|
||||||
from .kankanews import KankaNewsIE
|
from .kankanews import KankaNewsIE
|
||||||
from .karaoketv import KaraoketvIE
|
from .karaoketv import KaraoketvIE
|
||||||
from .kelbyone import KelbyOneIE
|
from .kelbyone import KelbyOneIE
|
||||||
|
from .kenh14 import (
|
||||||
|
Kenh14PlaylistIE,
|
||||||
|
Kenh14VideoIE,
|
||||||
|
)
|
||||||
from .khanacademy import (
|
from .khanacademy import (
|
||||||
KhanAcademyIE,
|
KhanAcademyIE,
|
||||||
KhanAcademyUnitIE,
|
KhanAcademyUnitIE,
|
||||||
|
@ -1135,12 +1139,6 @@ from .microsoftembed import (
|
||||||
MicrosoftMediusIE,
|
MicrosoftMediusIE,
|
||||||
)
|
)
|
||||||
from .microsoftstream import MicrosoftStreamIE
|
from .microsoftstream import MicrosoftStreamIE
|
||||||
from .mildom import (
|
|
||||||
MildomClipIE,
|
|
||||||
MildomIE,
|
|
||||||
MildomUserVodIE,
|
|
||||||
MildomVodIE,
|
|
||||||
)
|
|
||||||
from .minds import (
|
from .minds import (
|
||||||
MindsChannelIE,
|
MindsChannelIE,
|
||||||
MindsGroupIE,
|
MindsGroupIE,
|
||||||
|
@ -1522,8 +1520,8 @@ from .pgatour import PGATourIE
|
||||||
from .philharmoniedeparis import PhilharmonieDeParisIE
|
from .philharmoniedeparis import PhilharmonieDeParisIE
|
||||||
from .phoenix import PhoenixIE
|
from .phoenix import PhoenixIE
|
||||||
from .photobucket import PhotobucketIE
|
from .photobucket import PhotobucketIE
|
||||||
|
from .pialive import PiaLiveIE
|
||||||
from .piapro import PiaproIE
|
from .piapro import PiaproIE
|
||||||
from .piaulizaportal import PIAULIZAPortalIE
|
|
||||||
from .picarto import (
|
from .picarto import (
|
||||||
PicartoIE,
|
PicartoIE,
|
||||||
PicartoVodIE,
|
PicartoVodIE,
|
||||||
|
@ -1560,10 +1558,6 @@ from .podbayfm import (
|
||||||
)
|
)
|
||||||
from .podchaser import PodchaserIE
|
from .podchaser import PodchaserIE
|
||||||
from .podomatic import PodomaticIE
|
from .podomatic import PodomaticIE
|
||||||
from .pokemon import (
|
|
||||||
PokemonIE,
|
|
||||||
PokemonWatchIE,
|
|
||||||
)
|
|
||||||
from .pokergo import (
|
from .pokergo import (
|
||||||
PokerGoCollectionIE,
|
PokerGoCollectionIE,
|
||||||
PokerGoIE,
|
PokerGoIE,
|
||||||
|
@ -2257,6 +2251,10 @@ from .ufctv import (
|
||||||
)
|
)
|
||||||
from .ukcolumn import UkColumnIE
|
from .ukcolumn import UkColumnIE
|
||||||
from .uktvplay import UKTVPlayIE
|
from .uktvplay import UKTVPlayIE
|
||||||
|
from .uliza import (
|
||||||
|
UlizaPlayerIE,
|
||||||
|
UlizaPortalIE,
|
||||||
|
)
|
||||||
from .umg import UMGDeIE
|
from .umg import UMGDeIE
|
||||||
from .unistra import UnistraIE
|
from .unistra import UnistraIE
|
||||||
from .unity import UnityIE
|
from .unity import UnityIE
|
||||||
|
@ -2285,10 +2283,6 @@ from .utreon import UtreonIE
|
||||||
from .varzesh3 import Varzesh3IE
|
from .varzesh3 import Varzesh3IE
|
||||||
from .vbox7 import Vbox7IE
|
from .vbox7 import Vbox7IE
|
||||||
from .veo import VeoIE
|
from .veo import VeoIE
|
||||||
from .veoh import (
|
|
||||||
VeohIE,
|
|
||||||
VeohUserIE,
|
|
||||||
)
|
|
||||||
from .vesti import VestiIE
|
from .vesti import VestiIE
|
||||||
from .vevo import (
|
from .vevo import (
|
||||||
VevoIE,
|
VevoIE,
|
||||||
|
|
|
@ -1,4 +1,3 @@
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
ExtractorError,
|
ExtractorError,
|
||||||
|
|
|
@ -79,7 +79,7 @@ class ChaturbateIE(InfoExtractor):
|
||||||
'formats': self._extract_m3u8_formats(m3u8_url, video_id, ext='mp4', live=True),
|
'formats': self._extract_m3u8_formats(m3u8_url, video_id, ext='mp4', live=True),
|
||||||
}
|
}
|
||||||
|
|
||||||
def _extract_from_webpage(self, video_id, tld):
|
def _extract_from_html(self, video_id, tld):
|
||||||
webpage = self._download_webpage(
|
webpage = self._download_webpage(
|
||||||
f'https://chaturbate.{tld}/{video_id}/', video_id,
|
f'https://chaturbate.{tld}/{video_id}/', video_id,
|
||||||
headers=self.geo_verification_headers(), impersonate=True)
|
headers=self.geo_verification_headers(), impersonate=True)
|
||||||
|
@ -151,4 +151,4 @@ class ChaturbateIE(InfoExtractor):
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
video_id, tld = self._match_valid_url(url).group('id', 'tld')
|
video_id, tld = self._match_valid_url(url).group('id', 'tld')
|
||||||
return self._extract_from_api(video_id, tld) or self._extract_from_webpage(video_id, tld)
|
return self._extract_from_api(video_id, tld) or self._extract_from_html(video_id, tld)
|
||||||
|
|
|
@ -3767,7 +3767,7 @@ class InfoExtractor:
|
||||||
""" Merge subtitle dictionaries, language by language. """
|
""" Merge subtitle dictionaries, language by language. """
|
||||||
if target is None:
|
if target is None:
|
||||||
target = {}
|
target = {}
|
||||||
for d in dicts:
|
for d in filter(None, dicts):
|
||||||
for lang, subs in d.items():
|
for lang, subs in d.items():
|
||||||
target[lang] = cls._merge_subtitle_items(target.get(lang, []), subs)
|
target[lang] = cls._merge_subtitle_items(target.get(lang, []), subs)
|
||||||
return target
|
return target
|
||||||
|
|
|
@ -1,14 +1,27 @@
|
||||||
|
import json
|
||||||
import re
|
import re
|
||||||
|
import urllib.parse
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..utils import orderedSet
|
from .ninecninemedia import NineCNineMediaIE
|
||||||
|
from ..utils import extract_attributes, orderedSet
|
||||||
|
from ..utils.traversal import find_element, traverse_obj
|
||||||
|
|
||||||
|
|
||||||
class CTVNewsIE(InfoExtractor):
|
class CTVNewsIE(InfoExtractor):
|
||||||
_VALID_URL = r'https?://(?:.+?\.)?ctvnews\.ca/(?:video\?(?:clip|playlist|bin)Id=|.*?)(?P<id>[0-9.]+)'
|
_BASE_REGEX = r'https?://(?:[^.]+\.)?ctvnews\.ca/'
|
||||||
|
_VIDEO_ID_RE = r'(?P<id>\d{5,})'
|
||||||
|
_PLAYLIST_ID_RE = r'(?P<id>\d\.\d{5,})'
|
||||||
|
_VALID_URL = [
|
||||||
|
rf'{_BASE_REGEX}video/c{_VIDEO_ID_RE}',
|
||||||
|
rf'{_BASE_REGEX}video(?:-gallery)?/?\?clipId={_VIDEO_ID_RE}',
|
||||||
|
rf'{_BASE_REGEX}video/?\?(?:playlist|bin)Id={_PLAYLIST_ID_RE}',
|
||||||
|
rf'{_BASE_REGEX}(?!video/)[^?#]*?{_PLAYLIST_ID_RE}/?(?:$|[?#])',
|
||||||
|
rf'{_BASE_REGEX}(?!video/)[^?#]+\?binId={_PLAYLIST_ID_RE}',
|
||||||
|
]
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
'url': 'http://www.ctvnews.ca/video?clipId=901995',
|
'url': 'http://www.ctvnews.ca/video?clipId=901995',
|
||||||
'md5': '9b8624ba66351a23e0b6e1391971f9af',
|
'md5': 'b608f466c7fa24b9666c6439d766ab7e',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': '901995',
|
'id': '901995',
|
||||||
'ext': 'flv',
|
'ext': 'flv',
|
||||||
|
@ -16,6 +29,33 @@ class CTVNewsIE(InfoExtractor):
|
||||||
'description': 'md5:958dd3b4f5bbbf0ed4d045c790d89285',
|
'description': 'md5:958dd3b4f5bbbf0ed4d045c790d89285',
|
||||||
'timestamp': 1467286284,
|
'timestamp': 1467286284,
|
||||||
'upload_date': '20160630',
|
'upload_date': '20160630',
|
||||||
|
'categories': [],
|
||||||
|
'season_number': 0,
|
||||||
|
'season': 'Season 0',
|
||||||
|
'tags': [],
|
||||||
|
'series': 'CTV News National | Archive | Stories 2',
|
||||||
|
'season_id': '57981',
|
||||||
|
'thumbnail': r're:https?://.*\.jpg$',
|
||||||
|
'duration': 764.631,
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://barrie.ctvnews.ca/video/c3030933-here_s-what_s-making-news-for-nov--15?binId=1272429',
|
||||||
|
'md5': '8b8c2b33c5c1803e3c26bc74ff8694d5',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '3030933',
|
||||||
|
'ext': 'flv',
|
||||||
|
'title': 'Here’s what’s making news for Nov. 15',
|
||||||
|
'description': 'Here are the top stories we’re working on for CTV News at 11 for Nov. 15',
|
||||||
|
'thumbnail': 'http://images2.9c9media.com/image_asset/2021_2_22_a602e68e-1514-410e-a67a-e1f7cccbacab_png_2000x1125.jpg',
|
||||||
|
'season_id': '58104',
|
||||||
|
'season_number': 0,
|
||||||
|
'tags': [],
|
||||||
|
'season': 'Season 0',
|
||||||
|
'categories': [],
|
||||||
|
'series': 'CTV News Barrie',
|
||||||
|
'upload_date': '20241116',
|
||||||
|
'duration': 42.943,
|
||||||
|
'timestamp': 1731722452,
|
||||||
},
|
},
|
||||||
}, {
|
}, {
|
||||||
'url': 'http://www.ctvnews.ca/video?playlistId=1.2966224',
|
'url': 'http://www.ctvnews.ca/video?playlistId=1.2966224',
|
||||||
|
@ -31,6 +71,72 @@ class CTVNewsIE(InfoExtractor):
|
||||||
'id': '1.2876780',
|
'id': '1.2876780',
|
||||||
},
|
},
|
||||||
'playlist_mincount': 100,
|
'playlist_mincount': 100,
|
||||||
|
}, {
|
||||||
|
'url': 'https://www.ctvnews.ca/it-s-been-23-years-since-toronto-called-in-the-army-after-a-major-snowstorm-1.5736957',
|
||||||
|
'info_dict':
|
||||||
|
{
|
||||||
|
'id': '1.5736957',
|
||||||
|
},
|
||||||
|
'playlist_mincount': 6,
|
||||||
|
}, {
|
||||||
|
'url': 'https://www.ctvnews.ca/business/respondents-to-bank-of-canada-questionnaire-largely-oppose-creating-a-digital-loonie-1.6665797',
|
||||||
|
'md5': '24bc4b88cdc17d8c3fc01dfc228ab72c',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '2695026',
|
||||||
|
'ext': 'flv',
|
||||||
|
'season_id': '89852',
|
||||||
|
'series': 'From CTV News Channel',
|
||||||
|
'description': 'md5:796a985a23cacc7e1e2fafefd94afd0a',
|
||||||
|
'season': '2023',
|
||||||
|
'title': 'Bank of Canada asks public about digital currency',
|
||||||
|
'categories': [],
|
||||||
|
'tags': [],
|
||||||
|
'upload_date': '20230526',
|
||||||
|
'season_number': 2023,
|
||||||
|
'thumbnail': 'http://images2.9c9media.com/image_asset/2019_3_28_35f5afc3-10f6-4d92-b194-8b9a86f55c6a_png_1920x1080.jpg',
|
||||||
|
'timestamp': 1685105157,
|
||||||
|
'duration': 253.553,
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://stox.ctvnews.ca/video-gallery?clipId=582589',
|
||||||
|
'md5': '135cc592df607d29dddc931f1b756ae2',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '582589',
|
||||||
|
'ext': 'flv',
|
||||||
|
'categories': [],
|
||||||
|
'timestamp': 1427906183,
|
||||||
|
'season_number': 0,
|
||||||
|
'duration': 125.559,
|
||||||
|
'thumbnail': 'http://images2.9c9media.com/image_asset/2019_3_28_35f5afc3-10f6-4d92-b194-8b9a86f55c6a_png_1920x1080.jpg',
|
||||||
|
'series': 'CTV News Stox',
|
||||||
|
'description': 'CTV original footage of the rise and fall of the Berlin Wall.',
|
||||||
|
'title': 'Berlin Wall',
|
||||||
|
'season_id': '63817',
|
||||||
|
'season': 'Season 0',
|
||||||
|
'tags': [],
|
||||||
|
'upload_date': '20150401',
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://ottawa.ctvnews.ca/features/regional-contact/regional-contact-archive?binId=1.1164587#3023759',
|
||||||
|
'md5': 'a14c0603557decc6531260791c23cc5e',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '3023759',
|
||||||
|
'ext': 'flv',
|
||||||
|
'season_number': 2024,
|
||||||
|
'timestamp': 1731798000,
|
||||||
|
'season': '2024',
|
||||||
|
'episode': 'Episode 125',
|
||||||
|
'description': 'CTV News Ottawa at Six',
|
||||||
|
'duration': 2712.076,
|
||||||
|
'episode_number': 125,
|
||||||
|
'upload_date': '20241116',
|
||||||
|
'title': 'CTV News Ottawa at Six for Saturday, November 16, 2024',
|
||||||
|
'thumbnail': 'http://images2.9c9media.com/image_asset/2019_3_28_35f5afc3-10f6-4d92-b194-8b9a86f55c6a_png_1920x1080.jpg',
|
||||||
|
'categories': [],
|
||||||
|
'tags': [],
|
||||||
|
'series': 'CTV News Ottawa at Six',
|
||||||
|
'season_id': '92667',
|
||||||
|
},
|
||||||
}, {
|
}, {
|
||||||
'url': 'http://www.ctvnews.ca/1.810401',
|
'url': 'http://www.ctvnews.ca/1.810401',
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
|
@ -42,29 +148,35 @@ class CTVNewsIE(InfoExtractor):
|
||||||
'only_matching': True,
|
'only_matching': True,
|
||||||
}]
|
}]
|
||||||
|
|
||||||
|
def _ninecninemedia_url_result(self, clip_id):
|
||||||
|
return self.url_result(f'9c9media:ctvnews_web:{clip_id}', NineCNineMediaIE, clip_id)
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
page_id = self._match_id(url)
|
page_id = self._match_id(url)
|
||||||
|
|
||||||
def ninecninemedia_url_result(clip_id):
|
if mobj := re.fullmatch(self._VIDEO_ID_RE, urllib.parse.urlparse(url).fragment):
|
||||||
return {
|
page_id = mobj.group('id')
|
||||||
'_type': 'url_transparent',
|
|
||||||
'id': clip_id,
|
|
||||||
'url': f'9c9media:ctvnews_web:{clip_id}',
|
|
||||||
'ie_key': 'NineCNineMedia',
|
|
||||||
}
|
|
||||||
|
|
||||||
if page_id.isdigit():
|
if re.fullmatch(self._VIDEO_ID_RE, page_id):
|
||||||
return ninecninemedia_url_result(page_id)
|
return self._ninecninemedia_url_result(page_id)
|
||||||
else:
|
|
||||||
webpage = self._download_webpage(f'http://www.ctvnews.ca/{page_id}', page_id, query={
|
webpage = self._download_webpage(f'https://www.ctvnews.ca/{page_id}', page_id, query={
|
||||||
'ot': 'example.AjaxPageLayout.ot',
|
'ot': 'example.AjaxPageLayout.ot',
|
||||||
'maxItemsPerPage': 1000000,
|
'maxItemsPerPage': 1000000,
|
||||||
})
|
})
|
||||||
entries = [ninecninemedia_url_result(clip_id) for clip_id in orderedSet(
|
entries = [self._ninecninemedia_url_result(clip_id)
|
||||||
re.findall(r'clip\.id\s*=\s*(\d+);', webpage))]
|
for clip_id in orderedSet(re.findall(r'clip\.id\s*=\s*(\d+);', webpage))]
|
||||||
if not entries:
|
if not entries:
|
||||||
webpage = self._download_webpage(url, page_id)
|
webpage = self._download_webpage(url, page_id)
|
||||||
if 'getAuthStates("' in webpage:
|
if 'getAuthStates("' in webpage:
|
||||||
entries = [ninecninemedia_url_result(clip_id) for clip_id in
|
entries = [self._ninecninemedia_url_result(clip_id) for clip_id in
|
||||||
self._search_regex(r'getAuthStates\("([\d+,]+)"', webpage, 'clip ids').split(',')]
|
self._search_regex(r'getAuthStates\("([\d+,]+)"', webpage, 'clip ids').split(',')]
|
||||||
return self.playlist_result(entries, page_id)
|
else:
|
||||||
|
entries = [
|
||||||
|
self._ninecninemedia_url_result(clip_id) for clip_id in
|
||||||
|
traverse_obj(webpage, (
|
||||||
|
{find_element(tag='jasper-player-container', html=True)},
|
||||||
|
{extract_attributes}, 'axis-ids', {json.loads}, ..., 'axisId', {str}))
|
||||||
|
]
|
||||||
|
|
||||||
|
return self.playlist_result(entries, page_id)
|
||||||
|
|
|
@ -1,7 +1,10 @@
|
||||||
|
import time
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..networking.exceptions import HTTPError
|
from ..networking.exceptions import HTTPError
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
ExtractorError,
|
ExtractorError,
|
||||||
|
jwt_decode_hs256,
|
||||||
parse_codecs,
|
parse_codecs,
|
||||||
try_get,
|
try_get,
|
||||||
url_or_none,
|
url_or_none,
|
||||||
|
@ -13,9 +16,6 @@ from ..utils.traversal import traverse_obj
|
||||||
class DigitalConcertHallIE(InfoExtractor):
|
class DigitalConcertHallIE(InfoExtractor):
|
||||||
IE_DESC = 'DigitalConcertHall extractor'
|
IE_DESC = 'DigitalConcertHall extractor'
|
||||||
_VALID_URL = r'https?://(?:www\.)?digitalconcerthall\.com/(?P<language>[a-z]+)/(?P<type>film|concert|work)/(?P<id>[0-9]+)-?(?P<part>[0-9]+)?'
|
_VALID_URL = r'https?://(?:www\.)?digitalconcerthall\.com/(?P<language>[a-z]+)/(?P<type>film|concert|work)/(?P<id>[0-9]+)-?(?P<part>[0-9]+)?'
|
||||||
_OAUTH_URL = 'https://api.digitalconcerthall.com/v2/oauth2/token'
|
|
||||||
_USER_AGENT = 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/17.5 Safari/605.1.15'
|
|
||||||
_ACCESS_TOKEN = None
|
|
||||||
_NETRC_MACHINE = 'digitalconcerthall'
|
_NETRC_MACHINE = 'digitalconcerthall'
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
'note': 'Playlist with only one video',
|
'note': 'Playlist with only one video',
|
||||||
|
@ -69,59 +69,157 @@ class DigitalConcertHallIE(InfoExtractor):
|
||||||
'params': {'skip_download': 'm3u8'},
|
'params': {'skip_download': 'm3u8'},
|
||||||
'playlist_count': 1,
|
'playlist_count': 1,
|
||||||
}]
|
}]
|
||||||
|
_LOGIN_HINT = ('Use --username token --password ACCESS_TOKEN where ACCESS_TOKEN '
|
||||||
|
'is the "access_token_production" from your browser local storage')
|
||||||
|
_REFRESH_HINT = 'or else use a "refresh_token" with --username refresh --password REFRESH_TOKEN'
|
||||||
|
_OAUTH_URL = 'https://api.digitalconcerthall.com/v2/oauth2/token'
|
||||||
|
_CLIENT_ID = 'dch.webapp'
|
||||||
|
_CLIENT_SECRET = '2ySLN+2Fwb'
|
||||||
|
_USER_AGENT = 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/17.5 Safari/605.1.15'
|
||||||
|
_OAUTH_HEADERS = {
|
||||||
|
'Accept': 'application/json',
|
||||||
|
'Content-Type': 'application/x-www-form-urlencoded;charset=UTF-8',
|
||||||
|
'Origin': 'https://www.digitalconcerthall.com',
|
||||||
|
'Referer': 'https://www.digitalconcerthall.com/',
|
||||||
|
'User-Agent': _USER_AGENT,
|
||||||
|
}
|
||||||
|
_access_token = None
|
||||||
|
_access_token_expiry = 0
|
||||||
|
_refresh_token = None
|
||||||
|
|
||||||
def _perform_login(self, username, password):
|
@property
|
||||||
login_token = self._download_json(
|
def _access_token_is_expired(self):
|
||||||
self._OAUTH_URL,
|
return self._access_token_expiry - 30 <= int(time.time())
|
||||||
None, 'Obtaining token', errnote='Unable to obtain token', data=urlencode_postdata({
|
|
||||||
|
def _set_access_token(self, value):
|
||||||
|
self._access_token = value
|
||||||
|
self._access_token_expiry = traverse_obj(value, ({jwt_decode_hs256}, 'exp', {int})) or 0
|
||||||
|
|
||||||
|
def _cache_tokens(self, /):
|
||||||
|
self.cache.store(self._NETRC_MACHINE, 'tokens', {
|
||||||
|
'access_token': self._access_token,
|
||||||
|
'refresh_token': self._refresh_token,
|
||||||
|
})
|
||||||
|
|
||||||
|
def _fetch_new_tokens(self, invalidate=False):
|
||||||
|
if invalidate:
|
||||||
|
self.report_warning('Access token has been invalidated')
|
||||||
|
self._set_access_token(None)
|
||||||
|
|
||||||
|
if not self._access_token_is_expired:
|
||||||
|
return
|
||||||
|
|
||||||
|
if not self._refresh_token:
|
||||||
|
self._set_access_token(None)
|
||||||
|
self._cache_tokens()
|
||||||
|
raise ExtractorError(
|
||||||
|
'Access token has expired or been invalidated. '
|
||||||
|
'Get a new "access_token_production" value from your browser '
|
||||||
|
f'and try again, {self._REFRESH_HINT}', expected=True)
|
||||||
|
|
||||||
|
# If we only have a refresh token, we need a temporary "initial token" for the refresh flow
|
||||||
|
bearer_token = self._access_token or self._download_json(
|
||||||
|
self._OAUTH_URL, None, 'Obtaining initial token', 'Unable to obtain initial token',
|
||||||
|
data=urlencode_postdata({
|
||||||
'affiliate': 'none',
|
'affiliate': 'none',
|
||||||
'grant_type': 'device',
|
'grant_type': 'device',
|
||||||
'device_vendor': 'unknown',
|
'device_vendor': 'unknown',
|
||||||
# device_model 'Safari' gets split streams of 4K/HEVC video and lossless/FLAC audio
|
# device_model 'Safari' gets split streams of 4K/HEVC video and lossless/FLAC audio,
|
||||||
'device_model': 'unknown' if self._configuration_arg('prefer_combined_hls') else 'Safari',
|
# but this is no longer effective since actual login is not possible anymore
|
||||||
'app_id': 'dch.webapp',
|
'device_model': 'unknown',
|
||||||
|
'app_id': self._CLIENT_ID,
|
||||||
'app_distributor': 'berlinphil',
|
'app_distributor': 'berlinphil',
|
||||||
'app_version': '1.84.0',
|
'app_version': '1.95.0',
|
||||||
'client_secret': '2ySLN+2Fwb',
|
'client_secret': self._CLIENT_SECRET,
|
||||||
}), headers={
|
}), headers=self._OAUTH_HEADERS)['access_token']
|
||||||
'Accept': 'application/json',
|
|
||||||
'Content-Type': 'application/x-www-form-urlencoded;charset=UTF-8',
|
|
||||||
'User-Agent': self._USER_AGENT,
|
|
||||||
})['access_token']
|
|
||||||
try:
|
try:
|
||||||
login_response = self._download_json(
|
response = self._download_json(
|
||||||
self._OAUTH_URL,
|
self._OAUTH_URL, None, 'Refreshing token', 'Unable to refresh token',
|
||||||
None, note='Logging in', errnote='Unable to login', data=urlencode_postdata({
|
data=urlencode_postdata({
|
||||||
'grant_type': 'password',
|
'grant_type': 'refresh_token',
|
||||||
'username': username,
|
'refresh_token': self._refresh_token,
|
||||||
'password': password,
|
'client_id': self._CLIENT_ID,
|
||||||
|
'client_secret': self._CLIENT_SECRET,
|
||||||
}), headers={
|
}), headers={
|
||||||
'Accept': 'application/json',
|
**self._OAUTH_HEADERS,
|
||||||
'Content-Type': 'application/x-www-form-urlencoded;charset=UTF-8',
|
'Authorization': f'Bearer {bearer_token}',
|
||||||
'Referer': 'https://www.digitalconcerthall.com',
|
|
||||||
'Authorization': f'Bearer {login_token}',
|
|
||||||
'User-Agent': self._USER_AGENT,
|
|
||||||
})
|
})
|
||||||
except ExtractorError as error:
|
except ExtractorError as e:
|
||||||
if isinstance(error.cause, HTTPError) and error.cause.status == 401:
|
if isinstance(e.cause, HTTPError) and e.cause.status == 401:
|
||||||
raise ExtractorError('Invalid username or password', expected=True)
|
self._set_access_token(None)
|
||||||
|
self._refresh_token = None
|
||||||
|
self._cache_tokens()
|
||||||
|
raise ExtractorError('Your tokens have been invalidated', expected=True)
|
||||||
raise
|
raise
|
||||||
self._ACCESS_TOKEN = login_response['access_token']
|
|
||||||
|
self._set_access_token(response['access_token'])
|
||||||
|
if refresh_token := traverse_obj(response, ('refresh_token', {str})):
|
||||||
|
self.write_debug('New refresh token granted')
|
||||||
|
self._refresh_token = refresh_token
|
||||||
|
self._cache_tokens()
|
||||||
|
|
||||||
|
def _perform_login(self, username, password):
|
||||||
|
self.report_login()
|
||||||
|
|
||||||
|
if username == 'refresh':
|
||||||
|
self._refresh_token = password
|
||||||
|
self._fetch_new_tokens()
|
||||||
|
|
||||||
|
if username == 'token':
|
||||||
|
if not traverse_obj(password, {jwt_decode_hs256}):
|
||||||
|
raise ExtractorError(
|
||||||
|
f'The access token passed to yt-dlp is not valid. {self._LOGIN_HINT}', expected=True)
|
||||||
|
self._set_access_token(password)
|
||||||
|
self._cache_tokens()
|
||||||
|
|
||||||
|
if username in ('refresh', 'token'):
|
||||||
|
if self.get_param('cachedir') is not False:
|
||||||
|
token_type = 'access' if username == 'token' else 'refresh'
|
||||||
|
self.to_screen(f'Your {token_type} token has been cached to disk. To use the cached '
|
||||||
|
'token next time, pass --username cache along with any password')
|
||||||
|
return
|
||||||
|
|
||||||
|
if username != 'cache':
|
||||||
|
raise ExtractorError(
|
||||||
|
'Login with username and password is no longer supported '
|
||||||
|
f'for this site. {self._LOGIN_HINT}, {self._REFRESH_HINT}', expected=True)
|
||||||
|
|
||||||
|
# Try cached access_token
|
||||||
|
cached_tokens = self.cache.load(self._NETRC_MACHINE, 'tokens', default={})
|
||||||
|
self._set_access_token(cached_tokens.get('access_token'))
|
||||||
|
self._refresh_token = cached_tokens.get('refresh_token')
|
||||||
|
if not self._access_token_is_expired:
|
||||||
|
return
|
||||||
|
|
||||||
|
# Try cached refresh_token
|
||||||
|
self._fetch_new_tokens(invalidate=True)
|
||||||
|
|
||||||
def _real_initialize(self):
|
def _real_initialize(self):
|
||||||
if not self._ACCESS_TOKEN:
|
if not self._access_token:
|
||||||
self.raise_login_required(method='password')
|
self.raise_login_required(
|
||||||
|
'All content on this site is only available for registered users. '
|
||||||
|
f'{self._LOGIN_HINT}, {self._REFRESH_HINT}', method=None)
|
||||||
|
|
||||||
def _entries(self, items, language, type_, **kwargs):
|
def _entries(self, items, language, type_, **kwargs):
|
||||||
for item in items:
|
for item in items:
|
||||||
video_id = item['id']
|
video_id = item['id']
|
||||||
stream_info = self._download_json(
|
|
||||||
self._proto_relative_url(item['_links']['streams']['href']), video_id, headers={
|
for should_retry in (True, False):
|
||||||
'Accept': 'application/json',
|
self._fetch_new_tokens(invalidate=not should_retry)
|
||||||
'Authorization': f'Bearer {self._ACCESS_TOKEN}',
|
try:
|
||||||
'Accept-Language': language,
|
stream_info = self._download_json(
|
||||||
'User-Agent': self._USER_AGENT,
|
self._proto_relative_url(item['_links']['streams']['href']), video_id, headers={
|
||||||
})
|
'Accept': 'application/json',
|
||||||
|
'Authorization': f'Bearer {self._access_token}',
|
||||||
|
'Accept-Language': language,
|
||||||
|
'User-Agent': self._USER_AGENT,
|
||||||
|
})
|
||||||
|
break
|
||||||
|
except ExtractorError as error:
|
||||||
|
if should_retry and isinstance(error.cause, HTTPError) and error.cause.status == 401:
|
||||||
|
continue
|
||||||
|
raise
|
||||||
|
|
||||||
formats = []
|
formats = []
|
||||||
for m3u8_url in traverse_obj(stream_info, ('channel', ..., 'stream', ..., 'url', {url_or_none})):
|
for m3u8_url in traverse_obj(stream_info, ('channel', ..., 'stream', ..., 'url', {url_or_none})):
|
||||||
|
@ -157,7 +255,6 @@ class DigitalConcertHallIE(InfoExtractor):
|
||||||
'Accept': 'application/json',
|
'Accept': 'application/json',
|
||||||
'Accept-Language': language,
|
'Accept-Language': language,
|
||||||
'User-Agent': self._USER_AGENT,
|
'User-Agent': self._USER_AGENT,
|
||||||
'Authorization': f'Bearer {self._ACCESS_TOKEN}',
|
|
||||||
})
|
})
|
||||||
videos = [vid_info] if type_ == 'film' else traverse_obj(vid_info, ('_embedded', ..., ...))
|
videos = [vid_info] if type_ == 'film' else traverse_obj(vid_info, ('_embedded', ..., ...))
|
||||||
|
|
||||||
|
|
|
@ -569,7 +569,7 @@ class FacebookIE(InfoExtractor):
|
||||||
if dash_manifest:
|
if dash_manifest:
|
||||||
formats.extend(self._parse_mpd_formats(
|
formats.extend(self._parse_mpd_formats(
|
||||||
compat_etree_fromstring(urllib.parse.unquote_plus(dash_manifest)),
|
compat_etree_fromstring(urllib.parse.unquote_plus(dash_manifest)),
|
||||||
mpd_url=url_or_none(video.get('dash_manifest_url')) or mpd_url))
|
mpd_url=url_or_none(vid_data.get('dash_manifest_url')) or mpd_url))
|
||||||
|
|
||||||
def process_formats(info):
|
def process_formats(info):
|
||||||
# Downloads with browser's User-Agent are rate limited. Working around
|
# Downloads with browser's User-Agent are rate limited. Working around
|
||||||
|
|
160
yt_dlp/extractor/kenh14.py
Normal file
160
yt_dlp/extractor/kenh14.py
Normal file
|
@ -0,0 +1,160 @@
|
||||||
|
from .common import InfoExtractor
|
||||||
|
from ..utils import (
|
||||||
|
clean_html,
|
||||||
|
extract_attributes,
|
||||||
|
get_element_by_class,
|
||||||
|
get_element_html_by_attribute,
|
||||||
|
get_elements_html_by_class,
|
||||||
|
int_or_none,
|
||||||
|
parse_duration,
|
||||||
|
parse_iso8601,
|
||||||
|
remove_start,
|
||||||
|
strip_or_none,
|
||||||
|
unescapeHTML,
|
||||||
|
update_url,
|
||||||
|
url_or_none,
|
||||||
|
)
|
||||||
|
from ..utils.traversal import traverse_obj
|
||||||
|
|
||||||
|
|
||||||
|
class Kenh14VideoIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https?://video\.kenh14\.vn/(?:video/)?[\w-]+-(?P<id>[0-9]+)\.chn'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://video.kenh14.vn/video/mo-hop-iphone-14-pro-max-nguon-unbox-therapy-316173.chn',
|
||||||
|
'md5': '1ed67f9c3a1e74acf15db69590cf6210',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '316173',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'Video mở hộp iPhone 14 Pro Max (Nguồn: Unbox Therapy)',
|
||||||
|
'description': 'Video mở hộp iPhone 14 Pro MaxVideo mở hộp iPhone 14 Pro Max (Nguồn: Unbox Therapy)',
|
||||||
|
'thumbnail': r're:^https?://videothumbs\.mediacdn\.vn/.*\.jpg$',
|
||||||
|
'tags': [],
|
||||||
|
'uploader': 'Unbox Therapy',
|
||||||
|
'upload_date': '20220517',
|
||||||
|
'view_count': int,
|
||||||
|
'duration': 722.86,
|
||||||
|
'timestamp': 1652764468,
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://video.kenh14.vn/video-316174.chn',
|
||||||
|
'md5': '2b41877d2afaf4a3f487ceda8e5c7cbd',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '316174',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'Khoảnh khắc VĐV nằm gục khóc sau chiến thắng: 7 năm trời Việt Nam mới có HCV kiếm chém nữ, chỉ có 8 tháng để khổ luyện trước khi lên sàn đấu',
|
||||||
|
'description': 'md5:de86aa22e143e2b277bce8ec9c6f17dc',
|
||||||
|
'thumbnail': r're:^https?://videothumbs\.mediacdn\.vn/.*\.jpg$',
|
||||||
|
'tags': [],
|
||||||
|
'upload_date': '20220517',
|
||||||
|
'view_count': int,
|
||||||
|
'duration': 70.04,
|
||||||
|
'timestamp': 1652766021,
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://video.kenh14.vn/0-344740.chn',
|
||||||
|
'md5': 'b843495d5e728142c8870c09b46df2a9',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '344740',
|
||||||
|
'ext': 'mov',
|
||||||
|
'title': 'Kỳ Duyên đầy căng thẳng trong buổi ra quân đi Miss Universe, nghi thức tuyên thuệ lần đầu xuất hiện gây nhiều tranh cãi',
|
||||||
|
'description': 'md5:2a2dbb4a7397169fb21ee68f09160497',
|
||||||
|
'thumbnail': r're:^https?://kenh14cdn\.com/.*\.jpg$',
|
||||||
|
'tags': ['kỳ duyên', 'Kỳ Duyên tuyên thuệ', 'miss universe'],
|
||||||
|
'uploader': 'Quang Vũ',
|
||||||
|
'upload_date': '20241024',
|
||||||
|
'view_count': int,
|
||||||
|
'duration': 198.88,
|
||||||
|
'timestamp': 1729741590,
|
||||||
|
},
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
video_id = self._match_id(url)
|
||||||
|
webpage = self._download_webpage(url, video_id)
|
||||||
|
|
||||||
|
attrs = extract_attributes(get_element_html_by_attribute('type', 'VideoStream', webpage) or '')
|
||||||
|
direct_url = attrs['data-vid']
|
||||||
|
|
||||||
|
metadata = self._download_json(
|
||||||
|
'https://api.kinghub.vn/video/api/v1/detailVideoByGet?FileName={}'.format(
|
||||||
|
remove_start(direct_url, 'kenh14cdn.com/')), video_id, fatal=False)
|
||||||
|
|
||||||
|
formats = [{'url': f'https://{direct_url}', 'format_id': 'http', 'quality': 1}]
|
||||||
|
subtitles = {}
|
||||||
|
video_data = self._download_json(
|
||||||
|
f'https://{direct_url}.json', video_id, note='Downloading video data', fatal=False)
|
||||||
|
if hls_url := traverse_obj(video_data, ('hls', {url_or_none})):
|
||||||
|
fmts, subs = self._extract_m3u8_formats_and_subtitles(
|
||||||
|
hls_url, video_id, m3u8_id='hls', fatal=False)
|
||||||
|
formats.extend(fmts)
|
||||||
|
self._merge_subtitles(subs, target=subtitles)
|
||||||
|
if dash_url := traverse_obj(video_data, ('mpd', {url_or_none})):
|
||||||
|
fmts, subs = self._extract_mpd_formats_and_subtitles(
|
||||||
|
dash_url, video_id, mpd_id='dash', fatal=False)
|
||||||
|
formats.extend(fmts)
|
||||||
|
self._merge_subtitles(subs, target=subtitles)
|
||||||
|
|
||||||
|
return {
|
||||||
|
**traverse_obj(metadata, {
|
||||||
|
'duration': ('duration', {parse_duration}),
|
||||||
|
'uploader': ('author', {strip_or_none}),
|
||||||
|
'timestamp': ('uploadtime', {parse_iso8601(delimiter=' ')}),
|
||||||
|
'view_count': ('views', {int_or_none}),
|
||||||
|
}),
|
||||||
|
'id': video_id,
|
||||||
|
'title': (
|
||||||
|
traverse_obj(metadata, ('title', {strip_or_none}))
|
||||||
|
or clean_html(self._og_search_title(webpage))
|
||||||
|
or clean_html(get_element_by_class('vdbw-title', webpage))),
|
||||||
|
'formats': formats,
|
||||||
|
'subtitles': subtitles,
|
||||||
|
'description': (
|
||||||
|
clean_html(self._og_search_description(webpage))
|
||||||
|
or clean_html(get_element_by_class('vdbw-sapo', webpage))),
|
||||||
|
'thumbnail': (self._og_search_thumbnail(webpage) or attrs.get('data-thumb')),
|
||||||
|
'tags': traverse_obj(self._html_search_meta('keywords', webpage), (
|
||||||
|
{lambda x: x.split(';')}, ..., filter)),
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class Kenh14PlaylistIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https?://video\.kenh14\.vn/playlist/[\w-]+-(?P<id>[0-9]+)\.chn'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://video.kenh14.vn/playlist/tran-tinh-naked-love-mua-2-71.chn',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '71',
|
||||||
|
'title': 'Trần Tình (Naked love) mùa 2',
|
||||||
|
'description': 'md5:e9522339304956dea931722dd72eddb2',
|
||||||
|
'thumbnail': r're:^https?://kenh14cdn\.com/.*\.png$',
|
||||||
|
},
|
||||||
|
'playlist_count': 9,
|
||||||
|
}, {
|
||||||
|
'url': 'https://video.kenh14.vn/playlist/0-72.chn',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '72',
|
||||||
|
'title': 'Lau Lại Đầu Từ',
|
||||||
|
'description': 'Cùng xem xưa và nay có gì khác biệt nhé!',
|
||||||
|
'thumbnail': r're:^https?://kenh14cdn\.com/.*\.png$',
|
||||||
|
},
|
||||||
|
'playlist_count': 6,
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
playlist_id = self._match_id(url)
|
||||||
|
webpage = self._download_webpage(url, playlist_id)
|
||||||
|
|
||||||
|
category_detail = get_element_by_class('category-detail', webpage) or ''
|
||||||
|
embed_info = traverse_obj(
|
||||||
|
self._yield_json_ld(webpage, playlist_id),
|
||||||
|
(lambda _, v: v['name'] and v['alternateName'], any)) or {}
|
||||||
|
|
||||||
|
return self.playlist_from_matches(
|
||||||
|
get_elements_html_by_class('video-item', webpage), playlist_id,
|
||||||
|
(clean_html(get_element_by_class('name', category_detail)) or unescapeHTML(embed_info.get('name'))),
|
||||||
|
getter=lambda x: 'https://video.kenh14.vn/video/video-{}.chn'.format(extract_attributes(x)['data-id']),
|
||||||
|
ie=Kenh14VideoIE, playlist_description=(
|
||||||
|
clean_html(get_element_by_class('description', category_detail))
|
||||||
|
or unescapeHTML(embed_info.get('alternateName'))),
|
||||||
|
thumbnail=traverse_obj(
|
||||||
|
self._og_search_thumbnail(webpage),
|
||||||
|
({url_or_none}, {update_url(query=None)})))
|
|
@ -1,30 +1,32 @@
|
||||||
import json
|
import json
|
||||||
|
import uuid
|
||||||
|
|
||||||
from .common import InfoExtractor
|
from .common import InfoExtractor
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
ExtractorError,
|
ExtractorError,
|
||||||
int_or_none,
|
int_or_none,
|
||||||
|
join_nonempty,
|
||||||
smuggle_url,
|
smuggle_url,
|
||||||
traverse_obj,
|
traverse_obj,
|
||||||
try_call,
|
try_call,
|
||||||
unsmuggle_url,
|
unsmuggle_url,
|
||||||
|
urljoin,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
class LiTVIE(InfoExtractor):
|
class LiTVIE(InfoExtractor):
|
||||||
_VALID_URL = r'https?://(?:www\.)?litv\.tv/(?:vod|promo)/[^/]+/(?:content\.do)?\?.*?\b(?:content_)?id=(?P<id>[^&]+)'
|
_VALID_URL = r'https?://(?:www\.)?litv\.tv/(?:[^/?#]+/watch/|vod/[^/?#]+/content\.do\?content_id=)(?P<id>[\w-]+)'
|
||||||
|
_URL_TEMPLATE = 'https://www.litv.tv/%s/watch/%s'
|
||||||
_URL_TEMPLATE = 'https://www.litv.tv/vod/%s/content.do?content_id=%s'
|
_GEO_COUNTRIES = ['TW']
|
||||||
|
|
||||||
_TESTS = [{
|
_TESTS = [{
|
||||||
'url': 'https://www.litv.tv/vod/drama/content.do?brc_id=root&id=VOD00041610&isUHEnabled=true&autoPlay=1',
|
'url': 'https://www.litv.tv/drama/watch/VOD00041610',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'VOD00041606',
|
'id': 'VOD00041606',
|
||||||
'title': '花千骨',
|
'title': '花千骨',
|
||||||
},
|
},
|
||||||
'playlist_count': 51, # 50 episodes + 1 trailer
|
'playlist_count': 51, # 50 episodes + 1 trailer
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://www.litv.tv/vod/drama/content.do?brc_id=root&id=VOD00041610&isUHEnabled=true&autoPlay=1',
|
'url': 'https://www.litv.tv/drama/watch/VOD00041610',
|
||||||
'md5': 'b90ff1e9f1d8f5cfcd0a44c3e2b34c7a',
|
'md5': 'b90ff1e9f1d8f5cfcd0a44c3e2b34c7a',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'VOD00041610',
|
'id': 'VOD00041610',
|
||||||
|
@ -32,16 +34,15 @@ class LiTVIE(InfoExtractor):
|
||||||
'title': '花千骨第1集',
|
'title': '花千骨第1集',
|
||||||
'thumbnail': r're:https?://.*\.jpg$',
|
'thumbnail': r're:https?://.*\.jpg$',
|
||||||
'description': '《花千骨》陸劇線上看。十六年前,平靜的村莊內,一名女嬰隨異相出生,途徑此地的蜀山掌門清虛道長算出此女命運非同一般,她體內散發的異香易招惹妖魔。一念慈悲下,他在村莊周邊設下結界阻擋妖魔入侵,讓其年滿十六後去蜀山,並賜名花千骨。',
|
'description': '《花千骨》陸劇線上看。十六年前,平靜的村莊內,一名女嬰隨異相出生,途徑此地的蜀山掌門清虛道長算出此女命運非同一般,她體內散發的異香易招惹妖魔。一念慈悲下,他在村莊周邊設下結界阻擋妖魔入侵,讓其年滿十六後去蜀山,並賜名花千骨。',
|
||||||
'categories': ['奇幻', '愛情', '中國', '仙俠'],
|
'categories': ['奇幻', '愛情', '仙俠', '古裝'],
|
||||||
'episode': 'Episode 1',
|
'episode': 'Episode 1',
|
||||||
'episode_number': 1,
|
'episode_number': 1,
|
||||||
},
|
},
|
||||||
'params': {
|
'params': {
|
||||||
'noplaylist': True,
|
'noplaylist': True,
|
||||||
},
|
},
|
||||||
'skip': 'Georestricted to Taiwan',
|
|
||||||
}, {
|
}, {
|
||||||
'url': 'https://www.litv.tv/promo/miyuezhuan/?content_id=VOD00044841&',
|
'url': 'https://www.litv.tv/drama/watch/VOD00044841',
|
||||||
'md5': '88322ea132f848d6e3e18b32a832b918',
|
'md5': '88322ea132f848d6e3e18b32a832b918',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'VOD00044841',
|
'id': 'VOD00044841',
|
||||||
|
@ -55,94 +56,62 @@ class LiTVIE(InfoExtractor):
|
||||||
def _extract_playlist(self, playlist_data, content_type):
|
def _extract_playlist(self, playlist_data, content_type):
|
||||||
all_episodes = [
|
all_episodes = [
|
||||||
self.url_result(smuggle_url(
|
self.url_result(smuggle_url(
|
||||||
self._URL_TEMPLATE % (content_type, episode['contentId']),
|
self._URL_TEMPLATE % (content_type, episode['content_id']),
|
||||||
{'force_noplaylist': True})) # To prevent infinite recursion
|
{'force_noplaylist': True})) # To prevent infinite recursion
|
||||||
for episode in traverse_obj(playlist_data, ('seasons', ..., 'episode', lambda _, v: v['contentId']))]
|
for episode in traverse_obj(playlist_data, ('seasons', ..., 'episodes', lambda _, v: v['content_id']))]
|
||||||
|
|
||||||
return self.playlist_result(all_episodes, playlist_data['contentId'], playlist_data.get('title'))
|
return self.playlist_result(all_episodes, playlist_data['content_id'], playlist_data.get('title'))
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
url, smuggled_data = unsmuggle_url(url, {})
|
url, smuggled_data = unsmuggle_url(url, {})
|
||||||
|
|
||||||
video_id = self._match_id(url)
|
video_id = self._match_id(url)
|
||||||
|
|
||||||
webpage = self._download_webpage(url, video_id)
|
webpage = self._download_webpage(url, video_id)
|
||||||
|
vod_data = self._search_nextjs_data(webpage, video_id)['props']['pageProps']
|
||||||
|
|
||||||
if self._search_regex(
|
program_info = traverse_obj(vod_data, ('programInformation', {dict})) or {}
|
||||||
r'(?i)<meta\s[^>]*http-equiv="refresh"\s[^>]*content="[0-9]+;\s*url=https://www\.litv\.tv/"',
|
playlist_data = traverse_obj(vod_data, ('seriesTree'))
|
||||||
webpage, 'meta refresh redirect', default=False, group=0):
|
if playlist_data and self._yes_playlist(program_info.get('series_id'), video_id, smuggled_data):
|
||||||
raise ExtractorError('No such content found', expected=True)
|
return self._extract_playlist(playlist_data, program_info.get('content_type'))
|
||||||
|
|
||||||
program_info = self._parse_json(self._search_regex(
|
asset_id = traverse_obj(program_info, ('assets', 0, 'asset_id', {str}))
|
||||||
r'var\s+programInfo\s*=\s*([^;]+)', webpage, 'VOD data', default='{}'),
|
if asset_id: # This is a VOD
|
||||||
video_id)
|
media_type = 'vod'
|
||||||
|
else: # This is a live stream
|
||||||
|
asset_id = program_info['content_id']
|
||||||
|
media_type = program_info['content_type']
|
||||||
|
puid = try_call(lambda: self._get_cookies('https://www.litv.tv/')['PUID'].value)
|
||||||
|
if puid:
|
||||||
|
endpoint = 'get-urls'
|
||||||
|
else:
|
||||||
|
puid = str(uuid.uuid4())
|
||||||
|
endpoint = 'get-urls-no-auth'
|
||||||
|
video_data = self._download_json(
|
||||||
|
f'https://www.litv.tv/api/{endpoint}', video_id,
|
||||||
|
data=json.dumps({'AssetId': asset_id, 'MediaType': media_type, 'puid': puid}).encode(),
|
||||||
|
headers={'Content-Type': 'application/json'})
|
||||||
|
|
||||||
# In browsers `getProgramInfo` request is always issued. Usually this
|
if error := traverse_obj(video_data, ('error', {dict})):
|
||||||
# endpoint gives the same result as the data embedded in the webpage.
|
error_msg = traverse_obj(error, ('message', {str}))
|
||||||
# If, for some reason, there are no embedded data, we do an extra request.
|
if error_msg and 'OutsideRegionError' in error_msg:
|
||||||
if 'assetId' not in program_info:
|
|
||||||
program_info = self._download_json(
|
|
||||||
'https://www.litv.tv/vod/ajax/getProgramInfo', video_id,
|
|
||||||
query={'contentId': video_id},
|
|
||||||
headers={'Accept': 'application/json'})
|
|
||||||
|
|
||||||
series_id = program_info['seriesId']
|
|
||||||
if self._yes_playlist(series_id, video_id, smuggled_data):
|
|
||||||
playlist_data = self._download_json(
|
|
||||||
'https://www.litv.tv/vod/ajax/getSeriesTree', video_id,
|
|
||||||
query={'seriesId': series_id}, headers={'Accept': 'application/json'})
|
|
||||||
return self._extract_playlist(playlist_data, program_info['contentType'])
|
|
||||||
|
|
||||||
video_data = self._parse_json(self._search_regex(
|
|
||||||
r'uiHlsUrl\s*=\s*testBackendData\(([^;]+)\);',
|
|
||||||
webpage, 'video data', default='{}'), video_id)
|
|
||||||
if not video_data:
|
|
||||||
payload = {'assetId': program_info['assetId']}
|
|
||||||
puid = try_call(lambda: self._get_cookies('https://www.litv.tv/')['PUID'].value)
|
|
||||||
if puid:
|
|
||||||
payload.update({
|
|
||||||
'type': 'auth',
|
|
||||||
'puid': puid,
|
|
||||||
})
|
|
||||||
endpoint = 'getUrl'
|
|
||||||
else:
|
|
||||||
payload.update({
|
|
||||||
'watchDevices': program_info['watchDevices'],
|
|
||||||
'contentType': program_info['contentType'],
|
|
||||||
})
|
|
||||||
endpoint = 'getMainUrlNoAuth'
|
|
||||||
video_data = self._download_json(
|
|
||||||
f'https://www.litv.tv/vod/ajax/{endpoint}', video_id,
|
|
||||||
data=json.dumps(payload).encode(),
|
|
||||||
headers={'Content-Type': 'application/json'})
|
|
||||||
|
|
||||||
if not video_data.get('fullpath'):
|
|
||||||
error_msg = video_data.get('errorMessage')
|
|
||||||
if error_msg == 'vod.error.outsideregionerror':
|
|
||||||
self.raise_geo_restricted('This video is available in Taiwan only')
|
self.raise_geo_restricted('This video is available in Taiwan only')
|
||||||
if error_msg:
|
elif error_msg:
|
||||||
raise ExtractorError(f'{self.IE_NAME} said: {error_msg}', expected=True)
|
raise ExtractorError(f'{self.IE_NAME} said: {error_msg}', expected=True)
|
||||||
raise ExtractorError(f'Unexpected result from {self.IE_NAME}')
|
raise ExtractorError(f'Unexpected error from {self.IE_NAME}')
|
||||||
|
|
||||||
formats = self._extract_m3u8_formats(
|
formats = self._extract_m3u8_formats(
|
||||||
video_data['fullpath'], video_id, ext='mp4',
|
video_data['result']['AssetURLs'][0], video_id, ext='mp4', m3u8_id='hls')
|
||||||
entry_protocol='m3u8_native', m3u8_id='hls')
|
|
||||||
for a_format in formats:
|
for a_format in formats:
|
||||||
# LiTV HLS segments doesn't like compressions
|
# LiTV HLS segments doesn't like compressions
|
||||||
a_format.setdefault('http_headers', {})['Accept-Encoding'] = 'identity'
|
a_format.setdefault('http_headers', {})['Accept-Encoding'] = 'identity'
|
||||||
|
|
||||||
title = program_info['title'] + program_info.get('secondaryMark', '')
|
|
||||||
description = program_info.get('description')
|
|
||||||
thumbnail = program_info.get('imageFile')
|
|
||||||
categories = [item['name'] for item in program_info.get('category', [])]
|
|
||||||
episode = int_or_none(program_info.get('episode'))
|
|
||||||
|
|
||||||
return {
|
return {
|
||||||
'id': video_id,
|
'id': video_id,
|
||||||
'formats': formats,
|
'formats': formats,
|
||||||
'title': title,
|
'title': join_nonempty('title', 'secondary_mark', delim='', from_dict=program_info),
|
||||||
'description': description,
|
**traverse_obj(program_info, {
|
||||||
'thumbnail': thumbnail,
|
'description': ('description', {str}),
|
||||||
'categories': categories,
|
'thumbnail': ('picture', {urljoin('https://p-cdnstatic.svc.litv.tv/')}),
|
||||||
'episode_number': episode,
|
'categories': ('genres', ..., 'name', {str}),
|
||||||
|
'episode_number': ('episode', {int_or_none}),
|
||||||
|
}),
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,291 +0,0 @@
|
||||||
import functools
|
|
||||||
import json
|
|
||||||
import uuid
|
|
||||||
|
|
||||||
from .common import InfoExtractor
|
|
||||||
from ..utils import (
|
|
||||||
ExtractorError,
|
|
||||||
OnDemandPagedList,
|
|
||||||
determine_ext,
|
|
||||||
dict_get,
|
|
||||||
float_or_none,
|
|
||||||
traverse_obj,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class MildomBaseIE(InfoExtractor):
|
|
||||||
_GUEST_ID = None
|
|
||||||
|
|
||||||
def _call_api(self, url, video_id, query=None, note='Downloading JSON metadata', body=None):
|
|
||||||
if not self._GUEST_ID:
|
|
||||||
self._GUEST_ID = f'pc-gp-{uuid.uuid4()}'
|
|
||||||
|
|
||||||
content = self._download_json(
|
|
||||||
url, video_id, note=note, data=json.dumps(body).encode() if body else None,
|
|
||||||
headers={'Content-Type': 'application/json'} if body else {},
|
|
||||||
query={
|
|
||||||
'__guest_id': self._GUEST_ID,
|
|
||||||
'__platform': 'web',
|
|
||||||
**(query or {}),
|
|
||||||
})
|
|
||||||
|
|
||||||
if content['code'] != 0:
|
|
||||||
raise ExtractorError(
|
|
||||||
f'Mildom says: {content["message"]} (code {content["code"]})',
|
|
||||||
expected=True)
|
|
||||||
return content['body']
|
|
||||||
|
|
||||||
|
|
||||||
class MildomIE(MildomBaseIE):
|
|
||||||
IE_NAME = 'mildom'
|
|
||||||
IE_DESC = 'Record ongoing live by specific user in Mildom'
|
|
||||||
_VALID_URL = r'https?://(?:(?:www|m)\.)mildom\.com/(?P<id>\d+)'
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
video_id = self._match_id(url)
|
|
||||||
webpage = self._download_webpage(f'https://www.mildom.com/{video_id}', video_id)
|
|
||||||
|
|
||||||
enterstudio = self._call_api(
|
|
||||||
'https://cloudac.mildom.com/nonolive/gappserv/live/enterstudio', video_id,
|
|
||||||
note='Downloading live metadata', query={'user_id': video_id})
|
|
||||||
result_video_id = enterstudio.get('log_id', video_id)
|
|
||||||
|
|
||||||
servers = self._call_api(
|
|
||||||
'https://cloudac.mildom.com/nonolive/gappserv/live/liveserver', result_video_id,
|
|
||||||
note='Downloading live server list', query={
|
|
||||||
'user_id': video_id,
|
|
||||||
'live_server_type': 'hls',
|
|
||||||
})
|
|
||||||
|
|
||||||
playback_token = self._call_api(
|
|
||||||
'https://cloudac.mildom.com/nonolive/gappserv/live/token', result_video_id,
|
|
||||||
note='Obtaining live playback token', body={'host_id': video_id, 'type': 'hls'})
|
|
||||||
playback_token = traverse_obj(playback_token, ('data', ..., 'token'), get_all=False)
|
|
||||||
if not playback_token:
|
|
||||||
raise ExtractorError('Failed to obtain live playback token')
|
|
||||||
|
|
||||||
formats = self._extract_m3u8_formats(
|
|
||||||
f'{servers["stream_server"]}/{video_id}_master.m3u8?{playback_token}',
|
|
||||||
result_video_id, 'mp4', headers={
|
|
||||||
'Referer': 'https://www.mildom.com/',
|
|
||||||
'Origin': 'https://www.mildom.com',
|
|
||||||
})
|
|
||||||
|
|
||||||
for fmt in formats:
|
|
||||||
fmt.setdefault('http_headers', {})['Referer'] = 'https://www.mildom.com/'
|
|
||||||
|
|
||||||
return {
|
|
||||||
'id': result_video_id,
|
|
||||||
'title': self._html_search_meta('twitter:description', webpage, default=None) or traverse_obj(enterstudio, 'anchor_intro'),
|
|
||||||
'description': traverse_obj(enterstudio, 'intro', 'live_intro', expected_type=str),
|
|
||||||
'timestamp': float_or_none(enterstudio.get('live_start_ms'), scale=1000),
|
|
||||||
'uploader': self._html_search_meta('twitter:title', webpage, default=None) or traverse_obj(enterstudio, 'loginname'),
|
|
||||||
'uploader_id': video_id,
|
|
||||||
'formats': formats,
|
|
||||||
'is_live': True,
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
class MildomVodIE(MildomBaseIE):
|
|
||||||
IE_NAME = 'mildom:vod'
|
|
||||||
IE_DESC = 'VOD in Mildom'
|
|
||||||
_VALID_URL = r'https?://(?:(?:www|m)\.)mildom\.com/playback/(?P<user_id>\d+)/(?P<id>(?P=user_id)-[a-zA-Z0-9]+-?[0-9]*)'
|
|
||||||
_TESTS = [{
|
|
||||||
'url': 'https://www.mildom.com/playback/10882672/10882672-1597662269',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '10882672-1597662269',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': '始めてのミルダム配信じゃぃ!',
|
|
||||||
'thumbnail': r're:^https?://.*\.(png|jpg)$',
|
|
||||||
'upload_date': '20200817',
|
|
||||||
'duration': 4138.37,
|
|
||||||
'description': 'ゲームをしたくて!',
|
|
||||||
'timestamp': 1597662269.0,
|
|
||||||
'uploader_id': '10882672',
|
|
||||||
'uploader': 'kson組長(けいそん)',
|
|
||||||
},
|
|
||||||
}, {
|
|
||||||
'url': 'https://www.mildom.com/playback/10882672/10882672-1597758589870-477',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '10882672-1597758589870-477',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': '【kson】感染メイズ!麻酔銃で無双する',
|
|
||||||
'thumbnail': r're:^https?://.*\.(png|jpg)$',
|
|
||||||
'timestamp': 1597759093.0,
|
|
||||||
'uploader': 'kson組長(けいそん)',
|
|
||||||
'duration': 4302.58,
|
|
||||||
'uploader_id': '10882672',
|
|
||||||
'description': 'このステージ絶対乗り越えたい',
|
|
||||||
'upload_date': '20200818',
|
|
||||||
},
|
|
||||||
}, {
|
|
||||||
'url': 'https://www.mildom.com/playback/10882672/10882672-buha9td2lrn97fk2jme0',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '10882672-buha9td2lrn97fk2jme0',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': '【kson組長】CART RACER!!!',
|
|
||||||
'thumbnail': r're:^https?://.*\.(png|jpg)$',
|
|
||||||
'uploader_id': '10882672',
|
|
||||||
'uploader': 'kson組長(けいそん)',
|
|
||||||
'upload_date': '20201104',
|
|
||||||
'timestamp': 1604494797.0,
|
|
||||||
'duration': 4657.25,
|
|
||||||
'description': 'WTF',
|
|
||||||
},
|
|
||||||
}]
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
user_id, video_id = self._match_valid_url(url).group('user_id', 'id')
|
|
||||||
webpage = self._download_webpage(f'https://www.mildom.com/playback/{user_id}/{video_id}', video_id)
|
|
||||||
|
|
||||||
autoplay = self._call_api(
|
|
||||||
'https://cloudac.mildom.com/nonolive/videocontent/playback/getPlaybackDetail', video_id,
|
|
||||||
note='Downloading playback metadata', query={
|
|
||||||
'v_id': video_id,
|
|
||||||
})['playback']
|
|
||||||
|
|
||||||
formats = [{
|
|
||||||
'url': autoplay['audio_url'],
|
|
||||||
'format_id': 'audio',
|
|
||||||
'protocol': 'm3u8_native',
|
|
||||||
'vcodec': 'none',
|
|
||||||
'acodec': 'aac',
|
|
||||||
'ext': 'm4a',
|
|
||||||
}]
|
|
||||||
for fmt in autoplay['video_link']:
|
|
||||||
formats.append({
|
|
||||||
'format_id': 'video-{}'.format(fmt['name']),
|
|
||||||
'url': fmt['url'],
|
|
||||||
'protocol': 'm3u8_native',
|
|
||||||
'width': fmt['level'] * autoplay['video_width'] // autoplay['video_height'],
|
|
||||||
'height': fmt['level'],
|
|
||||||
'vcodec': 'h264',
|
|
||||||
'acodec': 'aac',
|
|
||||||
'ext': 'mp4',
|
|
||||||
})
|
|
||||||
|
|
||||||
return {
|
|
||||||
'id': video_id,
|
|
||||||
'title': self._html_search_meta(('og:description', 'description'), webpage, default=None) or autoplay.get('title'),
|
|
||||||
'description': traverse_obj(autoplay, 'video_intro'),
|
|
||||||
'timestamp': float_or_none(autoplay.get('publish_time'), scale=1000),
|
|
||||||
'duration': float_or_none(autoplay.get('video_length'), scale=1000),
|
|
||||||
'thumbnail': dict_get(autoplay, ('upload_pic', 'video_pic')),
|
|
||||||
'uploader': traverse_obj(autoplay, ('author_info', 'login_name')),
|
|
||||||
'uploader_id': user_id,
|
|
||||||
'formats': formats,
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
class MildomClipIE(MildomBaseIE):
|
|
||||||
IE_NAME = 'mildom:clip'
|
|
||||||
IE_DESC = 'Clip in Mildom'
|
|
||||||
_VALID_URL = r'https?://(?:(?:www|m)\.)mildom\.com/clip/(?P<id>(?P<user_id>\d+)-[a-zA-Z0-9]+)'
|
|
||||||
_TESTS = [{
|
|
||||||
'url': 'https://www.mildom.com/clip/10042245-63921673e7b147ebb0806d42b5ba5ce9',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '10042245-63921673e7b147ebb0806d42b5ba5ce9',
|
|
||||||
'title': '全然違ったよ',
|
|
||||||
'timestamp': 1619181890,
|
|
||||||
'duration': 59,
|
|
||||||
'thumbnail': r're:https?://.+',
|
|
||||||
'uploader': 'ざきんぽ',
|
|
||||||
'uploader_id': '10042245',
|
|
||||||
},
|
|
||||||
}, {
|
|
||||||
'url': 'https://www.mildom.com/clip/10111524-ebf4036e5aa8411c99fb3a1ae0902864',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '10111524-ebf4036e5aa8411c99fb3a1ae0902864',
|
|
||||||
'title': 'かっこいい',
|
|
||||||
'timestamp': 1621094003,
|
|
||||||
'duration': 59,
|
|
||||||
'thumbnail': r're:https?://.+',
|
|
||||||
'uploader': '(ルーキー',
|
|
||||||
'uploader_id': '10111524',
|
|
||||||
},
|
|
||||||
}, {
|
|
||||||
'url': 'https://www.mildom.com/clip/10660174-2c539e6e277c4aaeb4b1fbe8d22cb902',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '10660174-2c539e6e277c4aaeb4b1fbe8d22cb902',
|
|
||||||
'title': 'あ',
|
|
||||||
'timestamp': 1614769431,
|
|
||||||
'duration': 31,
|
|
||||||
'thumbnail': r're:https?://.+',
|
|
||||||
'uploader': 'ドルゴルスレンギーン=ダグワドルジ',
|
|
||||||
'uploader_id': '10660174',
|
|
||||||
},
|
|
||||||
}]
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
user_id, video_id = self._match_valid_url(url).group('user_id', 'id')
|
|
||||||
webpage = self._download_webpage(f'https://www.mildom.com/clip/{video_id}', video_id)
|
|
||||||
|
|
||||||
clip_detail = self._call_api(
|
|
||||||
'https://cloudac-cf-jp.mildom.com/nonolive/videocontent/clip/detail', video_id,
|
|
||||||
note='Downloading playback metadata', query={
|
|
||||||
'clip_id': video_id,
|
|
||||||
})
|
|
||||||
|
|
||||||
return {
|
|
||||||
'id': video_id,
|
|
||||||
'title': self._html_search_meta(
|
|
||||||
('og:description', 'description'), webpage, default=None) or clip_detail.get('title'),
|
|
||||||
'timestamp': float_or_none(clip_detail.get('create_time')),
|
|
||||||
'duration': float_or_none(clip_detail.get('length')),
|
|
||||||
'thumbnail': clip_detail.get('cover'),
|
|
||||||
'uploader': traverse_obj(clip_detail, ('user_info', 'loginname')),
|
|
||||||
'uploader_id': user_id,
|
|
||||||
|
|
||||||
'url': clip_detail['url'],
|
|
||||||
'ext': determine_ext(clip_detail.get('url'), 'mp4'),
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
class MildomUserVodIE(MildomBaseIE):
|
|
||||||
IE_NAME = 'mildom:user:vod'
|
|
||||||
IE_DESC = 'Download all VODs from specific user in Mildom'
|
|
||||||
_VALID_URL = r'https?://(?:(?:www|m)\.)mildom\.com/profile/(?P<id>\d+)'
|
|
||||||
_TESTS = [{
|
|
||||||
'url': 'https://www.mildom.com/profile/10093333',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '10093333',
|
|
||||||
'title': 'Uploads from ねこばたけ',
|
|
||||||
},
|
|
||||||
'playlist_mincount': 732,
|
|
||||||
}, {
|
|
||||||
'url': 'https://www.mildom.com/profile/10882672',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '10882672',
|
|
||||||
'title': 'Uploads from kson組長(けいそん)',
|
|
||||||
},
|
|
||||||
'playlist_mincount': 201,
|
|
||||||
}]
|
|
||||||
|
|
||||||
def _fetch_page(self, user_id, page):
|
|
||||||
page += 1
|
|
||||||
reply = self._call_api(
|
|
||||||
'https://cloudac.mildom.com/nonolive/videocontent/profile/playbackList',
|
|
||||||
user_id, note=f'Downloading page {page}', query={
|
|
||||||
'user_id': user_id,
|
|
||||||
'page': page,
|
|
||||||
'limit': '30',
|
|
||||||
})
|
|
||||||
if not reply:
|
|
||||||
return
|
|
||||||
for x in reply:
|
|
||||||
v_id = x.get('v_id')
|
|
||||||
if not v_id:
|
|
||||||
continue
|
|
||||||
yield self.url_result(f'https://www.mildom.com/playback/{user_id}/{v_id}')
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
user_id = self._match_id(url)
|
|
||||||
self.to_screen(f'This will download all VODs belonging to user. To download ongoing live video, use "https://www.mildom.com/{user_id}" instead')
|
|
||||||
|
|
||||||
profile = self._call_api(
|
|
||||||
'https://cloudac.mildom.com/nonolive/gappserv/user/profileV2', user_id,
|
|
||||||
query={'user_id': user_id}, note='Downloading user profile')['user_info']
|
|
||||||
|
|
||||||
return self.playlist_result(
|
|
||||||
OnDemandPagedList(functools.partial(self._fetch_page, user_id), 30),
|
|
||||||
user_id, f'Uploads from {profile["loginname"]}')
|
|
122
yt_dlp/extractor/pialive.py
Normal file
122
yt_dlp/extractor/pialive.py
Normal file
|
@ -0,0 +1,122 @@
|
||||||
|
from .common import InfoExtractor
|
||||||
|
from ..utils import (
|
||||||
|
ExtractorError,
|
||||||
|
clean_html,
|
||||||
|
extract_attributes,
|
||||||
|
get_element_by_class,
|
||||||
|
get_element_html_by_class,
|
||||||
|
multipart_encode,
|
||||||
|
str_or_none,
|
||||||
|
unified_timestamp,
|
||||||
|
url_or_none,
|
||||||
|
)
|
||||||
|
from ..utils.traversal import traverse_obj
|
||||||
|
|
||||||
|
|
||||||
|
class PiaLiveIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https?://player\.pia-live\.jp/stream/(?P<id>[\w-]+)'
|
||||||
|
_PLAYER_ROOT_URL = 'https://player.pia-live.jp/'
|
||||||
|
_PIA_LIVE_API_URL = 'https://api.pia-live.jp'
|
||||||
|
_API_KEY = 'kfds)FKFps-dms9e'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://player.pia-live.jp/stream/4JagFBEIM14s_hK9aXHKf3k3F3bY5eoHFQxu68TC6krUDqGOwN4d61dCWQYOd6CTxl4hjya9dsfEZGsM4uGOUdax60lEI4twsXGXf7crmz8Gk__GhupTrWxA7RFRVt76',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '88f3109a-f503-4d0f-a9f7-9f39ac745d84',
|
||||||
|
'display_id': '2431867_001',
|
||||||
|
'title': 'こながめでたい日2024の視聴ページ | PIA LIVE STREAM(ぴあライブストリーム)',
|
||||||
|
'live_status': 'was_live',
|
||||||
|
'comment_count': int,
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'getcomments': True,
|
||||||
|
'skip_download': True,
|
||||||
|
'ignore_no_formats_error': True,
|
||||||
|
},
|
||||||
|
'skip': 'The video is no longer available',
|
||||||
|
}, {
|
||||||
|
'url': 'https://player.pia-live.jp/stream/4JagFBEIM14s_hK9aXHKf3k3F3bY5eoHFQxu68TC6krJdu0GVBVbVy01IwpJ6J3qBEm3d9TCTt1d0eWpsZGj7DrOjVOmS7GAWGwyscMgiThopJvzgWC4H5b-7XQjAfRZ',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '9ce8b8ba-f6d1-4d1f-83a0-18c3148ded93',
|
||||||
|
'display_id': '2431867_002',
|
||||||
|
'title': 'こながめでたい日2024の視聴ページ | PIA LIVE STREAM(ぴあライブストリーム)',
|
||||||
|
'live_status': 'was_live',
|
||||||
|
'comment_count': int,
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'getcomments': True,
|
||||||
|
'skip_download': True,
|
||||||
|
'ignore_no_formats_error': True,
|
||||||
|
},
|
||||||
|
'skip': 'The video is no longer available',
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _extract_var(self, variable, html):
|
||||||
|
return self._search_regex(
|
||||||
|
rf'(?:var|const|let)\s+{variable}\s*=\s*(["\'])(?P<value>(?:(?!\1).)+)\1',
|
||||||
|
html, f'variable {variable}', group='value')
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
video_key = self._match_id(url)
|
||||||
|
webpage = self._download_webpage(url, video_key)
|
||||||
|
|
||||||
|
program_code = self._extract_var('programCode', webpage)
|
||||||
|
article_code = self._extract_var('articleCode', webpage)
|
||||||
|
title = self._html_extract_title(webpage)
|
||||||
|
|
||||||
|
if get_element_html_by_class('play-end', webpage):
|
||||||
|
raise ExtractorError('The video is no longer available', expected=True, video_id=program_code)
|
||||||
|
|
||||||
|
if start_info := clean_html(get_element_by_class('play-waiting__date', webpage)):
|
||||||
|
date, time = self._search_regex(
|
||||||
|
r'(?P<date>\d{4}/\d{1,2}/\d{1,2})\([月火水木金土日]\)(?P<time>\d{2}:\d{2})',
|
||||||
|
start_info, 'start_info', fatal=False, group=('date', 'time'))
|
||||||
|
if date and time:
|
||||||
|
release_timestamp_str = f'{date} {time} +09:00'
|
||||||
|
release_timestamp = unified_timestamp(release_timestamp_str)
|
||||||
|
self.raise_no_formats(f'The video will be available after {release_timestamp_str}', expected=True)
|
||||||
|
return {
|
||||||
|
'id': program_code,
|
||||||
|
'title': title,
|
||||||
|
'live_status': 'is_upcoming',
|
||||||
|
'release_timestamp': release_timestamp,
|
||||||
|
}
|
||||||
|
|
||||||
|
payload, content_type = multipart_encode({
|
||||||
|
'play_url': video_key,
|
||||||
|
'api_key': self._API_KEY,
|
||||||
|
})
|
||||||
|
api_data_and_headers = {
|
||||||
|
'data': payload,
|
||||||
|
'headers': {'Content-Type': content_type, 'Referer': self._PLAYER_ROOT_URL},
|
||||||
|
}
|
||||||
|
|
||||||
|
player_tag_list = self._download_json(
|
||||||
|
f'{self._PIA_LIVE_API_URL}/perf/player-tag-list/{program_code}', program_code,
|
||||||
|
'Fetching player tag list', 'Unable to fetch player tag list', **api_data_and_headers)
|
||||||
|
|
||||||
|
return self.url_result(
|
||||||
|
extract_attributes(player_tag_list['data']['movie_one_tag'])['src'],
|
||||||
|
url_transparent=True, title=title, display_id=program_code,
|
||||||
|
__post_extractor=self.extract_comments(program_code, article_code, api_data_and_headers))
|
||||||
|
|
||||||
|
def _get_comments(self, program_code, article_code, api_data_and_headers):
|
||||||
|
chat_room_url = traverse_obj(self._download_json(
|
||||||
|
f'{self._PIA_LIVE_API_URL}/perf/chat-tag-list/{program_code}/{article_code}', program_code,
|
||||||
|
'Fetching chat info', 'Unable to fetch chat info', fatal=False, **api_data_and_headers),
|
||||||
|
('data', 'chat_one_tag', {extract_attributes}, 'src', {url_or_none}))
|
||||||
|
if not chat_room_url:
|
||||||
|
return
|
||||||
|
comment_page = self._download_webpage(
|
||||||
|
chat_room_url, program_code, 'Fetching comment page', 'Unable to fetch comment page',
|
||||||
|
fatal=False, headers={'Referer': self._PLAYER_ROOT_URL})
|
||||||
|
if not comment_page:
|
||||||
|
return
|
||||||
|
yield from traverse_obj(self._search_json(
|
||||||
|
r'var\s+_history\s*=', comment_page, 'comment list',
|
||||||
|
program_code, contains_pattern=r'\[(?s:.+)\]', fatal=False), (..., {
|
||||||
|
'timestamp': (0, {int}),
|
||||||
|
'author_is_uploader': (1, {lambda x: x == 2}),
|
||||||
|
'author': (2, {str}),
|
||||||
|
'text': (3, {str}),
|
||||||
|
'id': (4, {str_or_none}),
|
||||||
|
}))
|
|
@ -1,70 +0,0 @@
|
||||||
from .common import InfoExtractor
|
|
||||||
from ..utils import (
|
|
||||||
ExtractorError,
|
|
||||||
int_or_none,
|
|
||||||
parse_qs,
|
|
||||||
time_seconds,
|
|
||||||
traverse_obj,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class PIAULIZAPortalIE(InfoExtractor):
|
|
||||||
IE_DESC = 'ulizaportal.jp - PIA LIVE STREAM'
|
|
||||||
_VALID_URL = r'https?://(?:www\.)?ulizaportal\.jp/pages/(?P<id>[\da-f]{8}-(?:[\da-f]{4}-){3}[\da-f]{12})'
|
|
||||||
_TESTS = [{
|
|
||||||
'url': 'https://ulizaportal.jp/pages/005f18b7-e810-5618-cb82-0987c5755d44',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '005f18b7-e810-5618-cb82-0987c5755d44',
|
|
||||||
'title': 'プレゼンテーションプレイヤーのサンプル',
|
|
||||||
'live_status': 'not_live',
|
|
||||||
},
|
|
||||||
'params': {
|
|
||||||
'skip_download': True,
|
|
||||||
'ignore_no_formats_error': True,
|
|
||||||
},
|
|
||||||
}, {
|
|
||||||
'url': 'https://ulizaportal.jp/pages/005e1b23-fe93-5780-19a0-98e917cc4b7d?expires=4102412400&signature=f422a993b683e1068f946caf406d211c17d1ef17da8bef3df4a519502155aa91&version=1',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '005e1b23-fe93-5780-19a0-98e917cc4b7d',
|
|
||||||
'title': '【確認用】視聴サンプルページ(ULIZA)',
|
|
||||||
'live_status': 'not_live',
|
|
||||||
},
|
|
||||||
'params': {
|
|
||||||
'skip_download': True,
|
|
||||||
'ignore_no_formats_error': True,
|
|
||||||
},
|
|
||||||
}]
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
video_id = self._match_id(url)
|
|
||||||
|
|
||||||
expires = int_or_none(traverse_obj(parse_qs(url), ('expires', 0)))
|
|
||||||
if expires and expires <= time_seconds():
|
|
||||||
raise ExtractorError('The link is expired.', video_id=video_id, expected=True)
|
|
||||||
|
|
||||||
webpage = self._download_webpage(url, video_id)
|
|
||||||
|
|
||||||
player_data = self._download_webpage(
|
|
||||||
self._search_regex(
|
|
||||||
r'<script [^>]*\bsrc="(https://player-api\.p\.uliza\.jp/v1/players/[^"]+)"',
|
|
||||||
webpage, 'player data url'),
|
|
||||||
video_id, headers={'Referer': 'https://ulizaportal.jp/'},
|
|
||||||
note='Fetching player data', errnote='Unable to fetch player data')
|
|
||||||
|
|
||||||
formats = self._extract_m3u8_formats(
|
|
||||||
self._search_regex(
|
|
||||||
r'["\'](https://vms-api\.p\.uliza\.jp/v1/prog-index\.m3u8[^"\']+)', player_data,
|
|
||||||
'm3u8 url', default=None),
|
|
||||||
video_id, fatal=False)
|
|
||||||
m3u8_type = self._search_regex(
|
|
||||||
r'/hls/(dvr|video)/', traverse_obj(formats, (0, 'url')), 'm3u8 type', default=None)
|
|
||||||
|
|
||||||
return {
|
|
||||||
'id': video_id,
|
|
||||||
'title': self._html_extract_title(webpage),
|
|
||||||
'formats': formats,
|
|
||||||
'live_status': {
|
|
||||||
'video': 'is_live',
|
|
||||||
'dvr': 'was_live', # short-term archives
|
|
||||||
}.get(m3u8_type, 'not_live'), # VOD or long-term archives
|
|
||||||
}
|
|
|
@ -1,136 +0,0 @@
|
||||||
from .common import InfoExtractor
|
|
||||||
from ..utils import (
|
|
||||||
ExtractorError,
|
|
||||||
extract_attributes,
|
|
||||||
int_or_none,
|
|
||||||
js_to_json,
|
|
||||||
merge_dicts,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class PokemonIE(InfoExtractor):
|
|
||||||
_VALID_URL = r'https?://(?:www\.)?pokemon\.com/[a-z]{2}(?:.*?play=(?P<id>[a-z0-9]{32})|/(?:[^/]+/)+(?P<display_id>[^/?#&]+))'
|
|
||||||
_TESTS = [{
|
|
||||||
'url': 'https://www.pokemon.com/us/pokemon-episodes/20_30-the-ol-raise-and-switch/',
|
|
||||||
'md5': '2fe8eaec69768b25ef898cda9c43062e',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'afe22e30f01c41f49d4f1d9eab5cd9a4',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': 'The Ol’ Raise and Switch!',
|
|
||||||
'description': 'md5:7db77f7107f98ba88401d3adc80ff7af',
|
|
||||||
},
|
|
||||||
'add_id': ['LimelightMedia'],
|
|
||||||
}, {
|
|
||||||
# no data-video-title
|
|
||||||
'url': 'https://www.pokemon.com/fr/episodes-pokemon/films-pokemon/pokemon-lascension-de-darkrai-2008',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'dfbaf830d7e54e179837c50c0c6cc0e1',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': "Pokémon : L'ascension de Darkrai",
|
|
||||||
'description': 'md5:d1dbc9e206070c3e14a06ff557659fb5',
|
|
||||||
},
|
|
||||||
'add_id': ['LimelightMedia'],
|
|
||||||
'params': {
|
|
||||||
'skip_download': True,
|
|
||||||
},
|
|
||||||
}, {
|
|
||||||
'url': 'http://www.pokemon.com/uk/pokemon-episodes/?play=2e8b5c761f1d4a9286165d7748c1ece2',
|
|
||||||
'only_matching': True,
|
|
||||||
}, {
|
|
||||||
'url': 'http://www.pokemon.com/fr/episodes-pokemon/18_09-un-hiver-inattendu/',
|
|
||||||
'only_matching': True,
|
|
||||||
}, {
|
|
||||||
'url': 'http://www.pokemon.com/de/pokemon-folgen/01_20-bye-bye-smettbo/',
|
|
||||||
'only_matching': True,
|
|
||||||
}]
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
video_id, display_id = self._match_valid_url(url).groups()
|
|
||||||
webpage = self._download_webpage(url, video_id or display_id)
|
|
||||||
video_data = extract_attributes(self._search_regex(
|
|
||||||
r'(<[^>]+data-video-id="{}"[^>]*>)'.format(video_id if video_id else '[a-z0-9]{32}'),
|
|
||||||
webpage, 'video data element'))
|
|
||||||
video_id = video_data['data-video-id']
|
|
||||||
title = video_data.get('data-video-title') or self._html_search_meta(
|
|
||||||
'pkm-title', webpage, ' title', default=None) or self._search_regex(
|
|
||||||
r'<h1[^>]+\bclass=["\']us-title[^>]+>([^<]+)', webpage, 'title')
|
|
||||||
return {
|
|
||||||
'_type': 'url_transparent',
|
|
||||||
'id': video_id,
|
|
||||||
'url': f'limelight:media:{video_id}',
|
|
||||||
'title': title,
|
|
||||||
'description': video_data.get('data-video-summary'),
|
|
||||||
'thumbnail': video_data.get('data-video-poster'),
|
|
||||||
'series': 'Pokémon',
|
|
||||||
'season_number': int_or_none(video_data.get('data-video-season')),
|
|
||||||
'episode': title,
|
|
||||||
'episode_number': int_or_none(video_data.get('data-video-episode')),
|
|
||||||
'ie_key': 'LimelightMedia',
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
class PokemonWatchIE(InfoExtractor):
|
|
||||||
_VALID_URL = r'https?://watch\.pokemon\.com/[a-z]{2}-[a-z]{2}/(?:#/)?player(?:\.html)?\?id=(?P<id>[a-z0-9]{32})'
|
|
||||||
_API_URL = 'https://www.pokemon.com/api/pokemontv/v2/channels/{0:}'
|
|
||||||
_TESTS = [{
|
|
||||||
'url': 'https://watch.pokemon.com/en-us/player.html?id=8309a40969894a8e8d5bc1311e9c5667',
|
|
||||||
'md5': '62833938a31e61ab49ada92f524c42ff',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '8309a40969894a8e8d5bc1311e9c5667',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': 'Lillier and the Staff!',
|
|
||||||
'description': 'md5:338841b8c21b283d24bdc9b568849f04',
|
|
||||||
},
|
|
||||||
}, {
|
|
||||||
'url': 'https://watch.pokemon.com/en-us/#/player?id=3fe7752ba09141f0b0f7756d1981c6b2',
|
|
||||||
'only_matching': True,
|
|
||||||
}, {
|
|
||||||
'url': 'https://watch.pokemon.com/de-de/player.html?id=b3c402e111a4459eb47e12160ab0ba07',
|
|
||||||
'only_matching': True,
|
|
||||||
}]
|
|
||||||
|
|
||||||
def _extract_media(self, channel_array, video_id):
|
|
||||||
for channel in channel_array:
|
|
||||||
for media in channel.get('media'):
|
|
||||||
if media.get('id') == video_id:
|
|
||||||
return media
|
|
||||||
return None
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
video_id = self._match_id(url)
|
|
||||||
|
|
||||||
info = {
|
|
||||||
'_type': 'url',
|
|
||||||
'id': video_id,
|
|
||||||
'url': f'limelight:media:{video_id}',
|
|
||||||
'ie_key': 'LimelightMedia',
|
|
||||||
}
|
|
||||||
|
|
||||||
# API call can be avoided entirely if we are listing formats
|
|
||||||
if self.get_param('listformats', False):
|
|
||||||
return info
|
|
||||||
|
|
||||||
webpage = self._download_webpage(url, video_id)
|
|
||||||
build_vars = self._parse_json(self._search_regex(
|
|
||||||
r'(?s)buildVars\s*=\s*({.*?})', webpage, 'build vars'),
|
|
||||||
video_id, transform_source=js_to_json)
|
|
||||||
region = build_vars.get('region')
|
|
||||||
channel_array = self._download_json(self._API_URL.format(region), video_id)
|
|
||||||
video_data = self._extract_media(channel_array, video_id)
|
|
||||||
|
|
||||||
if video_data is None:
|
|
||||||
raise ExtractorError(
|
|
||||||
f'Video {video_id} does not exist', expected=True)
|
|
||||||
|
|
||||||
info['_type'] = 'url_transparent'
|
|
||||||
images = video_data.get('images')
|
|
||||||
|
|
||||||
return merge_dicts(info, {
|
|
||||||
'title': video_data.get('title'),
|
|
||||||
'description': video_data.get('description'),
|
|
||||||
'thumbnail': images.get('medium') or images.get('small'),
|
|
||||||
'series': 'Pokémon',
|
|
||||||
'season_number': int_or_none(video_data.get('season')),
|
|
||||||
'episode': video_data.get('title'),
|
|
||||||
'episode_number': int_or_none(video_data.get('episode')),
|
|
||||||
})
|
|
|
@ -259,6 +259,8 @@ class RedditIE(InfoExtractor):
|
||||||
f'https://www.reddit.com/{slug}/.json', video_id, expected_status=403)
|
f'https://www.reddit.com/{slug}/.json', video_id, expected_status=403)
|
||||||
except ExtractorError as e:
|
except ExtractorError as e:
|
||||||
if isinstance(e.cause, json.JSONDecodeError):
|
if isinstance(e.cause, json.JSONDecodeError):
|
||||||
|
if self._get_cookies('https://www.reddit.com/').get('reddit_session'):
|
||||||
|
raise ExtractorError('Your IP address is unable to access the Reddit API', expected=True)
|
||||||
self.raise_login_required('Account authentication is required')
|
self.raise_login_required('Account authentication is required')
|
||||||
raise
|
raise
|
||||||
|
|
||||||
|
|
|
@ -13,7 +13,10 @@ from ..utils import (
|
||||||
unified_timestamp,
|
unified_timestamp,
|
||||||
url_or_none,
|
url_or_none,
|
||||||
)
|
)
|
||||||
from ..utils.traversal import traverse_obj
|
from ..utils.traversal import (
|
||||||
|
subs_list_to_dict,
|
||||||
|
traverse_obj,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
class RutubeBaseIE(InfoExtractor):
|
class RutubeBaseIE(InfoExtractor):
|
||||||
|
@ -92,11 +95,11 @@ class RutubeBaseIE(InfoExtractor):
|
||||||
hls_url, video_id, 'mp4', fatal=False, m3u8_id='hls')
|
hls_url, video_id, 'mp4', fatal=False, m3u8_id='hls')
|
||||||
formats.extend(fmts)
|
formats.extend(fmts)
|
||||||
self._merge_subtitles(subs, target=subtitles)
|
self._merge_subtitles(subs, target=subtitles)
|
||||||
for caption in traverse_obj(options, ('captions', lambda _, v: url_or_none(v['file']))):
|
self._merge_subtitles(traverse_obj(options, ('captions', ..., {
|
||||||
subtitles.setdefault(caption.get('code') or 'ru', []).append({
|
'id': 'code',
|
||||||
'url': caption['file'],
|
'url': 'file',
|
||||||
'name': caption.get('langTitle'),
|
'name': ('langTitle', {str}),
|
||||||
})
|
}, all, {subs_list_to_dict(lang='ru')})), target=subtitles)
|
||||||
return formats, subtitles
|
return formats, subtitles
|
||||||
|
|
||||||
def _download_and_extract_formats_and_subtitles(self, video_id, query=None):
|
def _download_and_extract_formats_and_subtitles(self, video_id, query=None):
|
||||||
|
|
|
@ -199,8 +199,9 @@ class SonyLIVSeriesIE(InfoExtractor):
|
||||||
},
|
},
|
||||||
}]
|
}]
|
||||||
_API_BASE = 'https://apiv2.sonyliv.com/AGL'
|
_API_BASE = 'https://apiv2.sonyliv.com/AGL'
|
||||||
|
_SORT_ORDERS = ('asc', 'desc')
|
||||||
|
|
||||||
def _entries(self, show_id):
|
def _entries(self, show_id, sort_order):
|
||||||
headers = {
|
headers = {
|
||||||
'Accept': 'application/json, text/plain, */*',
|
'Accept': 'application/json, text/plain, */*',
|
||||||
'Referer': 'https://www.sonyliv.com',
|
'Referer': 'https://www.sonyliv.com',
|
||||||
|
@ -215,6 +216,9 @@ class SonyLIVSeriesIE(InfoExtractor):
|
||||||
'from': '0',
|
'from': '0',
|
||||||
'to': '49',
|
'to': '49',
|
||||||
}), ('resultObj', 'containers', 0, 'containers', lambda _, v: int_or_none(v['id'])))
|
}), ('resultObj', 'containers', 0, 'containers', lambda _, v: int_or_none(v['id'])))
|
||||||
|
|
||||||
|
if sort_order == 'desc':
|
||||||
|
seasons = reversed(seasons)
|
||||||
for season in seasons:
|
for season in seasons:
|
||||||
season_id = str(season['id'])
|
season_id = str(season['id'])
|
||||||
note = traverse_obj(season, ('metadata', 'title', {str})) or 'season'
|
note = traverse_obj(season, ('metadata', 'title', {str})) or 'season'
|
||||||
|
@ -226,7 +230,7 @@ class SonyLIVSeriesIE(InfoExtractor):
|
||||||
'from': str(cursor),
|
'from': str(cursor),
|
||||||
'to': str(cursor + 99),
|
'to': str(cursor + 99),
|
||||||
'orderBy': 'episodeNumber',
|
'orderBy': 'episodeNumber',
|
||||||
'sortOrder': 'asc',
|
'sortOrder': sort_order,
|
||||||
}), ('resultObj', 'containers', 0, 'containers', lambda _, v: int_or_none(v['id'])))
|
}), ('resultObj', 'containers', 0, 'containers', lambda _, v: int_or_none(v['id'])))
|
||||||
if not episodes:
|
if not episodes:
|
||||||
break
|
break
|
||||||
|
@ -237,4 +241,10 @@ class SonyLIVSeriesIE(InfoExtractor):
|
||||||
|
|
||||||
def _real_extract(self, url):
|
def _real_extract(self, url):
|
||||||
show_id = self._match_id(url)
|
show_id = self._match_id(url)
|
||||||
return self.playlist_result(self._entries(show_id), playlist_id=show_id)
|
|
||||||
|
sort_order = self._configuration_arg('sort_order', [self._SORT_ORDERS[0]])[0]
|
||||||
|
if sort_order not in self._SORT_ORDERS:
|
||||||
|
raise ValueError(
|
||||||
|
f'Invalid sort order "{sort_order}". Allowed values are: {", ".join(self._SORT_ORDERS)}')
|
||||||
|
|
||||||
|
return self.playlist_result(self._entries(show_id, sort_order), playlist_id=show_id)
|
||||||
|
|
|
@ -241,7 +241,7 @@ class SoundcloudBaseIE(InfoExtractor):
|
||||||
format_urls.add(format_url)
|
format_urls.add(format_url)
|
||||||
formats.append({
|
formats.append({
|
||||||
'format_id': 'download',
|
'format_id': 'download',
|
||||||
'ext': urlhandle_detect_ext(urlh) or 'mp3',
|
'ext': urlhandle_detect_ext(urlh, default='mp3'),
|
||||||
'filesize': int_or_none(urlh.headers.get('Content-Length')),
|
'filesize': int_or_none(urlh.headers.get('Content-Length')),
|
||||||
'url': format_url,
|
'url': format_url,
|
||||||
'quality': 10,
|
'quality': 10,
|
||||||
|
|
113
yt_dlp/extractor/uliza.py
Normal file
113
yt_dlp/extractor/uliza.py
Normal file
|
@ -0,0 +1,113 @@
|
||||||
|
from .common import InfoExtractor
|
||||||
|
from ..utils import (
|
||||||
|
ExtractorError,
|
||||||
|
int_or_none,
|
||||||
|
make_archive_id,
|
||||||
|
parse_qs,
|
||||||
|
time_seconds,
|
||||||
|
)
|
||||||
|
from ..utils.traversal import traverse_obj
|
||||||
|
|
||||||
|
|
||||||
|
class UlizaPlayerIE(InfoExtractor):
|
||||||
|
_VALID_URL = r'https://player-api\.p\.uliza\.jp/v1/players/[^?#]+\?(?:[^#]*&)?name=(?P<id>[^#&]+)'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://player-api.p.uliza.jp/v1/players/timeshift-disabled/pia/admin?type=normal&playerobjectname=ulizaPlayer&name=livestream01_dvr&repeatable=true',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '88f3109a-f503-4d0f-a9f7-9f39ac745d84',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': '88f3109a-f503-4d0f-a9f7-9f39ac745d84',
|
||||||
|
'live_status': 'was_live',
|
||||||
|
'_old_archive_ids': ['piaulizaportal 88f3109a-f503-4d0f-a9f7-9f39ac745d84'],
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://player-api.p.uliza.jp/v1/players/uliza_jp_gallery_normal/promotion/admin?type=presentation&name=cookings&targetid=player1',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'ae350126-5e22-4a7f-a8ac-8d0fd448b800',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': 'ae350126-5e22-4a7f-a8ac-8d0fd448b800',
|
||||||
|
'live_status': 'not_live',
|
||||||
|
'_old_archive_ids': ['piaulizaportal ae350126-5e22-4a7f-a8ac-8d0fd448b800'],
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://player-api.p.uliza.jp/v1/players/default-player/pia/admin?type=normal&name=pia_movie_uliza_fix&targetid=ulizahtml5&repeatable=true',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '0644ecc8-e354-41b4-b957-3b08a2d63df1',
|
||||||
|
'ext': 'mp4',
|
||||||
|
'title': '0644ecc8-e354-41b4-b957-3b08a2d63df1',
|
||||||
|
'live_status': 'not_live',
|
||||||
|
'_old_archive_ids': ['piaulizaportal 0644ecc8-e354-41b4-b957-3b08a2d63df1'],
|
||||||
|
},
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
display_id = self._match_id(url)
|
||||||
|
player_data = self._download_webpage(
|
||||||
|
url, display_id, headers={'Referer': 'https://player-api.p.uliza.jp/'},
|
||||||
|
note='Fetching player data', errnote='Unable to fetch player data')
|
||||||
|
|
||||||
|
m3u8_url = self._search_regex(
|
||||||
|
r'["\'](https://vms-api\.p\.uliza\.jp/v1/prog-index\.m3u8[^"\']+)', player_data, 'm3u8 url')
|
||||||
|
video_id = parse_qs(m3u8_url).get('ss', [display_id])[0]
|
||||||
|
|
||||||
|
formats = self._extract_m3u8_formats(m3u8_url, video_id)
|
||||||
|
m3u8_type = self._search_regex(
|
||||||
|
r'/hls/(dvr|video)/', traverse_obj(formats, (0, 'url')), 'm3u8 type', default=None)
|
||||||
|
return {
|
||||||
|
'id': video_id,
|
||||||
|
'title': video_id,
|
||||||
|
'formats': formats,
|
||||||
|
'live_status': {
|
||||||
|
'video': 'is_live',
|
||||||
|
'dvr': 'was_live', # short-term archives
|
||||||
|
}.get(m3u8_type, 'not_live'), # VOD or long-term archives
|
||||||
|
'_old_archive_ids': [make_archive_id('PIAULIZAPortal', video_id)],
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class UlizaPortalIE(InfoExtractor):
|
||||||
|
IE_DESC = 'ulizaportal.jp'
|
||||||
|
_VALID_URL = r'https?://(?:www\.)?ulizaportal\.jp/pages/(?P<id>[\da-f]{8}-(?:[\da-f]{4}-){3}[\da-f]{12})'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'https://ulizaportal.jp/pages/005f18b7-e810-5618-cb82-0987c5755d44',
|
||||||
|
'info_dict': {
|
||||||
|
'id': 'ae350126-5e22-4a7f-a8ac-8d0fd448b800',
|
||||||
|
'display_id': '005f18b7-e810-5618-cb82-0987c5755d44',
|
||||||
|
'title': 'プレゼンテーションプレイヤーのサンプル',
|
||||||
|
'live_status': 'not_live',
|
||||||
|
'_old_archive_ids': ['piaulizaportal ae350126-5e22-4a7f-a8ac-8d0fd448b800'],
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': True,
|
||||||
|
'ignore_no_formats_error': True,
|
||||||
|
},
|
||||||
|
}, {
|
||||||
|
'url': 'https://ulizaportal.jp/pages/005e1b23-fe93-5780-19a0-98e917cc4b7d?expires=4102412400&signature=f422a993b683e1068f946caf406d211c17d1ef17da8bef3df4a519502155aa91&version=1',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '0644ecc8-e354-41b4-b957-3b08a2d63df1',
|
||||||
|
'display_id': '005e1b23-fe93-5780-19a0-98e917cc4b7d',
|
||||||
|
'title': '【確認用】視聴サンプルページ(ULIZA)',
|
||||||
|
'live_status': 'not_live',
|
||||||
|
'_old_archive_ids': ['piaulizaportal 0644ecc8-e354-41b4-b957-3b08a2d63df1'],
|
||||||
|
},
|
||||||
|
'params': {
|
||||||
|
'skip_download': True,
|
||||||
|
'ignore_no_formats_error': True,
|
||||||
|
},
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
video_id = self._match_id(url)
|
||||||
|
|
||||||
|
expires = int_or_none(traverse_obj(parse_qs(url), ('expires', 0)))
|
||||||
|
if expires and expires <= time_seconds():
|
||||||
|
raise ExtractorError('The link is expired', video_id=video_id, expected=True)
|
||||||
|
|
||||||
|
webpage = self._download_webpage(url, video_id)
|
||||||
|
|
||||||
|
player_data_url = self._search_regex(
|
||||||
|
r'<script [^>]*\bsrc="(https://player-api\.p\.uliza\.jp/v1/players/[^"]+)"',
|
||||||
|
webpage, 'player data url')
|
||||||
|
return self.url_result(
|
||||||
|
player_data_url, UlizaPlayerIE, url_transparent=True,
|
||||||
|
display_id=video_id, video_title=self._html_extract_title(webpage))
|
|
@ -1,189 +0,0 @@
|
||||||
import functools
|
|
||||||
import json
|
|
||||||
|
|
||||||
from .common import InfoExtractor
|
|
||||||
from ..utils import (
|
|
||||||
ExtractorError,
|
|
||||||
OnDemandPagedList,
|
|
||||||
int_or_none,
|
|
||||||
parse_duration,
|
|
||||||
qualities,
|
|
||||||
remove_start,
|
|
||||||
strip_or_none,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class VeohIE(InfoExtractor):
|
|
||||||
_VALID_URL = r'https?://(?:www\.)?veoh\.com/(?:watch|videos|embed|iphone/#_Watch)/(?P<id>(?:v|e|yapi-)[\da-zA-Z]+)'
|
|
||||||
|
|
||||||
_TESTS = [{
|
|
||||||
'url': 'http://www.veoh.com/watch/v56314296nk7Zdmz3',
|
|
||||||
'md5': '620e68e6a3cff80086df3348426c9ca3',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'v56314296nk7Zdmz3',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': 'Straight Backs Are Stronger',
|
|
||||||
'description': 'md5:203f976279939a6dc664d4001e13f5f4',
|
|
||||||
'thumbnail': 're:https://fcache\\.veoh\\.com/file/f/th56314296\\.jpg(\\?.*)?',
|
|
||||||
'uploader': 'LUMOback',
|
|
||||||
'duration': 46,
|
|
||||||
'view_count': int,
|
|
||||||
'average_rating': int,
|
|
||||||
'comment_count': int,
|
|
||||||
'age_limit': 0,
|
|
||||||
'categories': ['technology_and_gaming'],
|
|
||||||
'tags': ['posture', 'posture', 'sensor', 'back', 'pain', 'wearable', 'tech', 'lumo'],
|
|
||||||
},
|
|
||||||
}, {
|
|
||||||
'url': 'http://www.veoh.com/embed/v56314296nk7Zdmz3',
|
|
||||||
'only_matching': True,
|
|
||||||
}, {
|
|
||||||
'url': 'http://www.veoh.com/watch/v27701988pbTc4wzN?h1=Chile+workers+cover+up+to+avoid+skin+damage',
|
|
||||||
'md5': '4a6ff84b87d536a6a71e6aa6c0ad07fa',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '27701988',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': 'Chile workers cover up to avoid skin damage',
|
|
||||||
'description': 'md5:2bd151625a60a32822873efc246ba20d',
|
|
||||||
'uploader': 'afp-news',
|
|
||||||
'duration': 123,
|
|
||||||
},
|
|
||||||
'skip': 'This video has been deleted.',
|
|
||||||
}, {
|
|
||||||
'url': 'http://www.veoh.com/watch/v69525809F6Nc4frX',
|
|
||||||
'md5': '4fde7b9e33577bab2f2f8f260e30e979',
|
|
||||||
'note': 'Embedded ooyala video',
|
|
||||||
'info_dict': {
|
|
||||||
'id': '69525809',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': 'Doctors Alter Plan For Preteen\'s Weight Loss Surgery',
|
|
||||||
'description': 'md5:f5a11c51f8fb51d2315bca0937526891',
|
|
||||||
'uploader': 'newsy-videos',
|
|
||||||
},
|
|
||||||
'skip': 'This video has been deleted.',
|
|
||||||
}, {
|
|
||||||
'url': 'http://www.veoh.com/watch/e152215AJxZktGS',
|
|
||||||
'only_matching': True,
|
|
||||||
}, {
|
|
||||||
'url': 'https://www.veoh.com/videos/v16374379WA437rMH',
|
|
||||||
'md5': 'cceb73f3909063d64f4b93d4defca1b3',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'v16374379WA437rMH',
|
|
||||||
'ext': 'mp4',
|
|
||||||
'title': 'Phantasmagoria 2, pt. 1-3',
|
|
||||||
'description': 'Phantasmagoria: a Puzzle of Flesh',
|
|
||||||
'thumbnail': 're:https://fcache\\.veoh\\.com/file/f/th16374379\\.jpg(\\?.*)?',
|
|
||||||
'uploader': 'davidspackage',
|
|
||||||
'duration': 968,
|
|
||||||
'view_count': int,
|
|
||||||
'average_rating': int,
|
|
||||||
'comment_count': int,
|
|
||||||
'age_limit': 18,
|
|
||||||
'categories': ['technology_and_gaming', 'gaming'],
|
|
||||||
'tags': ['puzzle', 'of', 'flesh'],
|
|
||||||
},
|
|
||||||
}]
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
video_id = self._match_id(url)
|
|
||||||
metadata = self._download_json(
|
|
||||||
'https://www.veoh.com/watch/getVideo/' + video_id,
|
|
||||||
video_id)
|
|
||||||
video = metadata['video']
|
|
||||||
title = video['title']
|
|
||||||
|
|
||||||
thumbnail_url = None
|
|
||||||
q = qualities(['Regular', 'HQ'])
|
|
||||||
formats = []
|
|
||||||
for f_id, f_url in video.get('src', {}).items():
|
|
||||||
if not f_url:
|
|
||||||
continue
|
|
||||||
if f_id == 'poster':
|
|
||||||
thumbnail_url = f_url
|
|
||||||
else:
|
|
||||||
formats.append({
|
|
||||||
'format_id': f_id,
|
|
||||||
'quality': q(f_id),
|
|
||||||
'url': f_url,
|
|
||||||
})
|
|
||||||
|
|
||||||
categories = metadata.get('categoryPath')
|
|
||||||
if not categories:
|
|
||||||
category = remove_start(strip_or_none(video.get('category')), 'category_')
|
|
||||||
categories = [category] if category else None
|
|
||||||
tags = video.get('tags')
|
|
||||||
|
|
||||||
return {
|
|
||||||
'id': video_id,
|
|
||||||
'title': title,
|
|
||||||
'description': video.get('description'),
|
|
||||||
'thumbnail': thumbnail_url,
|
|
||||||
'uploader': video.get('author', {}).get('nickname'),
|
|
||||||
'duration': int_or_none(video.get('lengthBySec')) or parse_duration(video.get('length')),
|
|
||||||
'view_count': int_or_none(video.get('views')),
|
|
||||||
'formats': formats,
|
|
||||||
'average_rating': int_or_none(video.get('rating')),
|
|
||||||
'comment_count': int_or_none(video.get('numOfComments')),
|
|
||||||
'age_limit': 18 if video.get('contentRatingId') == 2 else 0,
|
|
||||||
'categories': categories,
|
|
||||||
'tags': tags.split(', ') if tags else None,
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
class VeohUserIE(VeohIE): # XXX: Do not subclass from concrete IE
|
|
||||||
_VALID_URL = r'https?://(?:www\.)?veoh\.com/users/(?P<id>[\w-]+)'
|
|
||||||
IE_NAME = 'veoh:user'
|
|
||||||
|
|
||||||
_TESTS = [
|
|
||||||
{
|
|
||||||
'url': 'https://www.veoh.com/users/valentinazoe',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'valentinazoe',
|
|
||||||
'title': 'valentinazoe (Uploads)',
|
|
||||||
},
|
|
||||||
'playlist_mincount': 75,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
'url': 'https://www.veoh.com/users/PiensaLibre',
|
|
||||||
'info_dict': {
|
|
||||||
'id': 'PiensaLibre',
|
|
||||||
'title': 'PiensaLibre (Uploads)',
|
|
||||||
},
|
|
||||||
'playlist_mincount': 2,
|
|
||||||
}]
|
|
||||||
|
|
||||||
_PAGE_SIZE = 16
|
|
||||||
|
|
||||||
def _fetch_page(self, uploader, page):
|
|
||||||
response = self._download_json(
|
|
||||||
'https://www.veoh.com/users/published/videos', uploader,
|
|
||||||
note=f'Downloading videos page {page + 1}',
|
|
||||||
headers={
|
|
||||||
'x-csrf-token': self._TOKEN,
|
|
||||||
'content-type': 'application/json;charset=UTF-8',
|
|
||||||
},
|
|
||||||
data=json.dumps({
|
|
||||||
'username': uploader,
|
|
||||||
'maxResults': self._PAGE_SIZE,
|
|
||||||
'page': page + 1,
|
|
||||||
'requestName': 'userPage',
|
|
||||||
}).encode())
|
|
||||||
if not response.get('success'):
|
|
||||||
raise ExtractorError(response['message'])
|
|
||||||
|
|
||||||
for video in response['videos']:
|
|
||||||
yield self.url_result(f'https://www.veoh.com/watch/{video["permalinkId"]}', VeohIE,
|
|
||||||
video['permalinkId'], video.get('title'))
|
|
||||||
|
|
||||||
def _real_initialize(self):
|
|
||||||
webpage = self._download_webpage(
|
|
||||||
'https://www.veoh.com', None, note='Downloading authorization token')
|
|
||||||
self._TOKEN = self._search_regex(
|
|
||||||
r'csrfToken:\s*(["\'])(?P<token>[0-9a-zA-Z]{40})\1', webpage,
|
|
||||||
'request token', group='token')
|
|
||||||
|
|
||||||
def _real_extract(self, url):
|
|
||||||
uploader = self._match_id(url)
|
|
||||||
return self.playlist_result(OnDemandPagedList(
|
|
||||||
functools.partial(self._fetch_page, uploader),
|
|
||||||
self._PAGE_SIZE), uploader, f'{uploader} (Uploads)')
|
|
|
@ -22,7 +22,7 @@ import urllib.parse
|
||||||
from .common import InfoExtractor, SearchInfoExtractor
|
from .common import InfoExtractor, SearchInfoExtractor
|
||||||
from .openload import PhantomJSwrapper
|
from .openload import PhantomJSwrapper
|
||||||
from ..jsinterp import JSInterpreter
|
from ..jsinterp import JSInterpreter
|
||||||
from ..networking.exceptions import HTTPError, TransportError, network_exceptions
|
from ..networking.exceptions import HTTPError, network_exceptions
|
||||||
from ..utils import (
|
from ..utils import (
|
||||||
NO_DEFAULT,
|
NO_DEFAULT,
|
||||||
ExtractorError,
|
ExtractorError,
|
||||||
|
@ -50,12 +50,12 @@ from ..utils import (
|
||||||
parse_iso8601,
|
parse_iso8601,
|
||||||
parse_qs,
|
parse_qs,
|
||||||
qualities,
|
qualities,
|
||||||
|
remove_end,
|
||||||
remove_start,
|
remove_start,
|
||||||
smuggle_url,
|
smuggle_url,
|
||||||
str_or_none,
|
str_or_none,
|
||||||
str_to_int,
|
str_to_int,
|
||||||
strftime_or_none,
|
strftime_or_none,
|
||||||
time_seconds,
|
|
||||||
traverse_obj,
|
traverse_obj,
|
||||||
try_call,
|
try_call,
|
||||||
try_get,
|
try_get,
|
||||||
|
@ -573,208 +573,18 @@ class YoutubeBaseInfoExtractor(InfoExtractor):
|
||||||
self._check_login_required()
|
self._check_login_required()
|
||||||
|
|
||||||
def _perform_login(self, username, password):
|
def _perform_login(self, username, password):
|
||||||
auth_type, _, user = (username or '').partition('+')
|
if username.startswith('oauth'):
|
||||||
|
|
||||||
if auth_type != 'oauth':
|
|
||||||
raise ExtractorError(self._youtube_login_hint, expected=True)
|
|
||||||
|
|
||||||
self._initialize_oauth(user, password)
|
|
||||||
|
|
||||||
'''
|
|
||||||
OAuth 2.0 Device Authorization Grant flow, used by the YouTube TV client (youtube.com/tv).
|
|
||||||
|
|
||||||
For more information regarding OAuth 2.0 and the Device Authorization Grant flow in general, see:
|
|
||||||
- https://developers.google.com/identity/protocols/oauth2/limited-input-device
|
|
||||||
- https://accounts.google.com/.well-known/openid-configuration
|
|
||||||
- https://www.rfc-editor.org/rfc/rfc8628
|
|
||||||
- https://www.rfc-editor.org/rfc/rfc6749
|
|
||||||
|
|
||||||
Note: The official client appears to use a proxied version of the oauth2 endpoints on youtube.com/o/oauth2,
|
|
||||||
which applies some modifications to the response (such as returning errors as 200 OK).
|
|
||||||
Since the client works with the standard API, we will use that as it is well-documented.
|
|
||||||
'''
|
|
||||||
|
|
||||||
_OAUTH_PROFILE = None
|
|
||||||
_OAUTH_ACCESS_TOKEN_CACHE = {}
|
|
||||||
_OAUTH_DISPLAY_ID = 'oauth'
|
|
||||||
|
|
||||||
# YouTube TV (TVHTML5) client. You can find these at youtube.com/tv
|
|
||||||
_OAUTH_CLIENT_ID = '861556708454-d6dlm3lh05idd8npek18k6be8ba3oc68.apps.googleusercontent.com'
|
|
||||||
_OAUTH_CLIENT_SECRET = 'SboVhoG9s0rNafixCSGGKXAT'
|
|
||||||
_OAUTH_SCOPE = 'http://gdata.youtube.com https://www.googleapis.com/auth/youtube-paid-content'
|
|
||||||
|
|
||||||
# From https://accounts.google.com/.well-known/openid-configuration
|
|
||||||
# Technically, these should be fetched dynamically and not hard-coded.
|
|
||||||
# However, as these endpoints rarely change, we can risk saving an extra request for every invocation.
|
|
||||||
_OAUTH_DEVICE_AUTHORIZATION_ENDPOINT = 'https://oauth2.googleapis.com/device/code'
|
|
||||||
_OAUTH_TOKEN_ENDPOINT = 'https://oauth2.googleapis.com/token'
|
|
||||||
|
|
||||||
@property
|
|
||||||
def _oauth_cache_key(self):
|
|
||||||
return f'oauth_refresh_token_{self._OAUTH_PROFILE}'
|
|
||||||
|
|
||||||
def _read_oauth_error_response(self, response):
|
|
||||||
return traverse_obj(
|
|
||||||
self._webpage_read_content(response, self._OAUTH_TOKEN_ENDPOINT, self._OAUTH_DISPLAY_ID, fatal=False),
|
|
||||||
({json.loads}, 'error', {str}))
|
|
||||||
|
|
||||||
def _set_oauth_info(self, token_response):
|
|
||||||
YoutubeBaseInfoExtractor._OAUTH_ACCESS_TOKEN_CACHE.setdefault(self._OAUTH_PROFILE, {}).update({
|
|
||||||
'access_token': token_response['access_token'],
|
|
||||||
'token_type': token_response['token_type'],
|
|
||||||
'expiry': time_seconds(
|
|
||||||
seconds=traverse_obj(token_response, ('expires_in', {float_or_none}), default=300) - 10),
|
|
||||||
})
|
|
||||||
refresh_token = traverse_obj(token_response, ('refresh_token', {str}))
|
|
||||||
if refresh_token:
|
|
||||||
self.cache.store(self._NETRC_MACHINE, self._oauth_cache_key, refresh_token)
|
|
||||||
YoutubeBaseInfoExtractor._OAUTH_ACCESS_TOKEN_CACHE[self._OAUTH_PROFILE]['refresh_token'] = refresh_token
|
|
||||||
|
|
||||||
def _initialize_oauth(self, user, refresh_token):
|
|
||||||
self._OAUTH_PROFILE = user or 'default'
|
|
||||||
|
|
||||||
if self._OAUTH_PROFILE in YoutubeBaseInfoExtractor._OAUTH_ACCESS_TOKEN_CACHE:
|
|
||||||
self.write_debug(f'{self._OAUTH_DISPLAY_ID}: Using cached access token for profile "{self._OAUTH_PROFILE}"')
|
|
||||||
return
|
|
||||||
|
|
||||||
YoutubeBaseInfoExtractor._OAUTH_ACCESS_TOKEN_CACHE[self._OAUTH_PROFILE] = {}
|
|
||||||
|
|
||||||
if refresh_token:
|
|
||||||
msg = f'{self._OAUTH_DISPLAY_ID}: Using password input as refresh token'
|
|
||||||
if self.get_param('cachedir') is not False:
|
|
||||||
msg += ' and caching token to disk; you should supply an empty password next time'
|
|
||||||
self.to_screen(msg)
|
|
||||||
self.cache.store(self._NETRC_MACHINE, self._oauth_cache_key, refresh_token)
|
|
||||||
else:
|
|
||||||
refresh_token = self.cache.load(self._NETRC_MACHINE, self._oauth_cache_key)
|
|
||||||
|
|
||||||
if refresh_token:
|
|
||||||
YoutubeBaseInfoExtractor._OAUTH_ACCESS_TOKEN_CACHE[self._OAUTH_PROFILE]['refresh_token'] = refresh_token
|
|
||||||
try:
|
|
||||||
token_response = self._refresh_token(refresh_token)
|
|
||||||
except ExtractorError as e:
|
|
||||||
error_msg = str(e.orig_msg).replace('Failed to refresh access token: ', '')
|
|
||||||
self.report_warning(f'{self._OAUTH_DISPLAY_ID}: Failed to refresh access token: {error_msg}')
|
|
||||||
token_response = self._oauth_authorize
|
|
||||||
else:
|
|
||||||
token_response = self._oauth_authorize
|
|
||||||
|
|
||||||
self._set_oauth_info(token_response)
|
|
||||||
self.write_debug(f'{self._OAUTH_DISPLAY_ID}: Logged in using profile "{self._OAUTH_PROFILE}"')
|
|
||||||
|
|
||||||
def _refresh_token(self, refresh_token):
|
|
||||||
try:
|
|
||||||
token_response = self._download_json(
|
|
||||||
self._OAUTH_TOKEN_ENDPOINT,
|
|
||||||
video_id=self._OAUTH_DISPLAY_ID,
|
|
||||||
note='Refreshing access token',
|
|
||||||
data=json.dumps({
|
|
||||||
'client_id': self._OAUTH_CLIENT_ID,
|
|
||||||
'client_secret': self._OAUTH_CLIENT_SECRET,
|
|
||||||
'refresh_token': refresh_token,
|
|
||||||
'grant_type': 'refresh_token',
|
|
||||||
}).encode(),
|
|
||||||
headers={'Content-Type': 'application/json'})
|
|
||||||
except ExtractorError as e:
|
|
||||||
if isinstance(e.cause, HTTPError):
|
|
||||||
error = self._read_oauth_error_response(e.cause.response)
|
|
||||||
if error == 'invalid_grant':
|
|
||||||
# RFC6749 § 5.2
|
|
||||||
raise ExtractorError(
|
|
||||||
'Failed to refresh access token: Refresh token is invalid, revoked, or expired (invalid_grant)',
|
|
||||||
expected=True, video_id=self._OAUTH_DISPLAY_ID)
|
|
||||||
raise ExtractorError(
|
|
||||||
f'Failed to refresh access token: Authorization server returned error {error}',
|
|
||||||
video_id=self._OAUTH_DISPLAY_ID)
|
|
||||||
raise
|
|
||||||
return token_response
|
|
||||||
|
|
||||||
@property
|
|
||||||
def _oauth_authorize(self):
|
|
||||||
code_response = self._download_json(
|
|
||||||
self._OAUTH_DEVICE_AUTHORIZATION_ENDPOINT,
|
|
||||||
video_id=self._OAUTH_DISPLAY_ID,
|
|
||||||
note='Initializing authorization flow',
|
|
||||||
data=json.dumps({
|
|
||||||
'client_id': self._OAUTH_CLIENT_ID,
|
|
||||||
'scope': self._OAUTH_SCOPE,
|
|
||||||
}).encode(),
|
|
||||||
headers={'Content-Type': 'application/json'})
|
|
||||||
|
|
||||||
verification_url = traverse_obj(code_response, ('verification_url', {str}))
|
|
||||||
user_code = traverse_obj(code_response, ('user_code', {str}))
|
|
||||||
if not verification_url or not user_code:
|
|
||||||
raise ExtractorError(
|
raise ExtractorError(
|
||||||
'Authorization server did not provide verification_url or user_code', video_id=self._OAUTH_DISPLAY_ID)
|
f'Login with OAuth is no longer supported. {self._youtube_login_hint}', expected=True)
|
||||||
|
|
||||||
# note: The whitespace is intentional
|
self.report_warning(
|
||||||
self.to_screen(
|
f'Login with password is not supported for YouTube. {self._youtube_login_hint}')
|
||||||
f'{self._OAUTH_DISPLAY_ID}: To give yt-dlp access to your account, '
|
|
||||||
f'go to {verification_url} and enter code {user_code}')
|
|
||||||
|
|
||||||
# RFC8628 § 3.5: default poll interval is 5 seconds if not provided
|
|
||||||
poll_interval = traverse_obj(code_response, ('interval', {int}), default=5)
|
|
||||||
|
|
||||||
for retry in self.RetryManager():
|
|
||||||
while True:
|
|
||||||
try:
|
|
||||||
token_response = self._download_json(
|
|
||||||
self._OAUTH_TOKEN_ENDPOINT,
|
|
||||||
video_id=self._OAUTH_DISPLAY_ID,
|
|
||||||
note=False,
|
|
||||||
errnote='Failed to request access token',
|
|
||||||
data=json.dumps({
|
|
||||||
'client_id': self._OAUTH_CLIENT_ID,
|
|
||||||
'client_secret': self._OAUTH_CLIENT_SECRET,
|
|
||||||
'device_code': code_response['device_code'],
|
|
||||||
'grant_type': 'urn:ietf:params:oauth:grant-type:device_code',
|
|
||||||
}).encode(),
|
|
||||||
headers={'Content-Type': 'application/json'})
|
|
||||||
except ExtractorError as e:
|
|
||||||
if isinstance(e.cause, TransportError):
|
|
||||||
retry.error = e
|
|
||||||
break
|
|
||||||
elif isinstance(e.cause, HTTPError):
|
|
||||||
error = self._read_oauth_error_response(e.cause.response)
|
|
||||||
if not error:
|
|
||||||
retry.error = e
|
|
||||||
break
|
|
||||||
|
|
||||||
if error == 'authorization_pending':
|
|
||||||
time.sleep(poll_interval)
|
|
||||||
continue
|
|
||||||
elif error == 'expired_token':
|
|
||||||
raise ExtractorError(
|
|
||||||
'Authorization timed out', expected=True, video_id=self._OAUTH_DISPLAY_ID)
|
|
||||||
elif error == 'access_denied':
|
|
||||||
raise ExtractorError(
|
|
||||||
'You denied access to an account', expected=True, video_id=self._OAUTH_DISPLAY_ID)
|
|
||||||
elif error == 'slow_down':
|
|
||||||
# RFC8628 § 3.5: add 5 seconds to the poll interval
|
|
||||||
poll_interval += 5
|
|
||||||
time.sleep(poll_interval)
|
|
||||||
continue
|
|
||||||
else:
|
|
||||||
raise ExtractorError(
|
|
||||||
f'Authorization server returned an error when fetching access token: {error}',
|
|
||||||
video_id=self._OAUTH_DISPLAY_ID)
|
|
||||||
raise
|
|
||||||
|
|
||||||
return token_response
|
|
||||||
|
|
||||||
def _update_oauth(self):
|
|
||||||
token = YoutubeBaseInfoExtractor._OAUTH_ACCESS_TOKEN_CACHE.get(self._OAUTH_PROFILE)
|
|
||||||
if token is None or token['expiry'] > time.time():
|
|
||||||
return
|
|
||||||
|
|
||||||
self._set_oauth_info(self._refresh_token(token['refresh_token']))
|
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def _youtube_login_hint(self):
|
def _youtube_login_hint(self):
|
||||||
return ('Use --username=oauth[+PROFILE] --password="" to log in using oauth, '
|
return (f'{self._login_hint(method="cookies")}. Also see '
|
||||||
f'or else u{self._login_hint(method="cookies")[1:]}. '
|
'https://github.com/yt-dlp/yt-dlp/wiki/Extractors#exporting-youtube-cookies '
|
||||||
'See https://github.com/yt-dlp/yt-dlp/wiki/Extractors#logging-in-with-oauth for more on how to use oauth. '
|
'for tips on effectively exporting YouTube cookies')
|
||||||
'See https://github.com/yt-dlp/yt-dlp/wiki/Extractors#exporting-youtube-cookies for help with cookies')
|
|
||||||
|
|
||||||
def _check_login_required(self):
|
def _check_login_required(self):
|
||||||
if self._LOGIN_REQUIRED and not self.is_authenticated:
|
if self._LOGIN_REQUIRED and not self.is_authenticated:
|
||||||
|
@ -924,7 +734,7 @@ class YoutubeBaseInfoExtractor(InfoExtractor):
|
||||||
|
|
||||||
@functools.cached_property
|
@functools.cached_property
|
||||||
def is_authenticated(self):
|
def is_authenticated(self):
|
||||||
return self._OAUTH_PROFILE or bool(self._generate_sapisidhash_header())
|
return bool(self._generate_sapisidhash_header())
|
||||||
|
|
||||||
def extract_ytcfg(self, video_id, webpage):
|
def extract_ytcfg(self, video_id, webpage):
|
||||||
if not webpage:
|
if not webpage:
|
||||||
|
@ -934,16 +744,6 @@ class YoutubeBaseInfoExtractor(InfoExtractor):
|
||||||
r'ytcfg\.set\s*\(\s*({.+?})\s*\)\s*;', webpage, 'ytcfg',
|
r'ytcfg\.set\s*\(\s*({.+?})\s*\)\s*;', webpage, 'ytcfg',
|
||||||
default='{}'), video_id, fatal=False) or {}
|
default='{}'), video_id, fatal=False) or {}
|
||||||
|
|
||||||
def _generate_oauth_headers(self):
|
|
||||||
self._update_oauth()
|
|
||||||
oauth_token = YoutubeBaseInfoExtractor._OAUTH_ACCESS_TOKEN_CACHE.get(self._OAUTH_PROFILE)
|
|
||||||
if not oauth_token:
|
|
||||||
return {}
|
|
||||||
|
|
||||||
return {
|
|
||||||
'Authorization': f'{oauth_token["token_type"]} {oauth_token["access_token"]}',
|
|
||||||
}
|
|
||||||
|
|
||||||
def _generate_cookie_auth_headers(self, *, ytcfg=None, account_syncid=None, session_index=None, origin=None, **kwargs):
|
def _generate_cookie_auth_headers(self, *, ytcfg=None, account_syncid=None, session_index=None, origin=None, **kwargs):
|
||||||
headers = {}
|
headers = {}
|
||||||
account_syncid = account_syncid or self._extract_account_syncid(ytcfg)
|
account_syncid = account_syncid or self._extract_account_syncid(ytcfg)
|
||||||
|
@ -973,14 +773,10 @@ class YoutubeBaseInfoExtractor(InfoExtractor):
|
||||||
'Origin': origin,
|
'Origin': origin,
|
||||||
'X-Goog-Visitor-Id': visitor_data or self._extract_visitor_data(ytcfg),
|
'X-Goog-Visitor-Id': visitor_data or self._extract_visitor_data(ytcfg),
|
||||||
'User-Agent': self._ytcfg_get_safe(ytcfg, lambda x: x['INNERTUBE_CONTEXT']['client']['userAgent'], default_client=default_client),
|
'User-Agent': self._ytcfg_get_safe(ytcfg, lambda x: x['INNERTUBE_CONTEXT']['client']['userAgent'], default_client=default_client),
|
||||||
**self._generate_oauth_headers(),
|
|
||||||
**self._generate_cookie_auth_headers(ytcfg=ytcfg, account_syncid=account_syncid, session_index=session_index, origin=origin),
|
**self._generate_cookie_auth_headers(ytcfg=ytcfg, account_syncid=account_syncid, session_index=session_index, origin=origin),
|
||||||
}
|
}
|
||||||
return filter_dict(headers)
|
return filter_dict(headers)
|
||||||
|
|
||||||
def _generate_webpage_headers(self):
|
|
||||||
return self._generate_oauth_headers()
|
|
||||||
|
|
||||||
def _download_ytcfg(self, client, video_id):
|
def _download_ytcfg(self, client, video_id):
|
||||||
url = {
|
url = {
|
||||||
'web': 'https://www.youtube.com',
|
'web': 'https://www.youtube.com',
|
||||||
|
@ -990,8 +786,7 @@ class YoutubeBaseInfoExtractor(InfoExtractor):
|
||||||
if not url:
|
if not url:
|
||||||
return {}
|
return {}
|
||||||
webpage = self._download_webpage(
|
webpage = self._download_webpage(
|
||||||
url, video_id, fatal=False, note=f'Downloading {client.replace("_", " ").strip()} client config',
|
url, video_id, fatal=False, note=f'Downloading {client.replace("_", " ").strip()} client config')
|
||||||
headers=self._generate_webpage_headers())
|
|
||||||
return self.extract_ytcfg(video_id, webpage) or {}
|
return self.extract_ytcfg(video_id, webpage) or {}
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
|
@ -3256,8 +3051,7 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
||||||
code = self._download_webpage(
|
code = self._download_webpage(
|
||||||
player_url, video_id, fatal=fatal,
|
player_url, video_id, fatal=fatal,
|
||||||
note='Downloading player ' + player_id,
|
note='Downloading player ' + player_id,
|
||||||
errnote=f'Download of {player_url} failed',
|
errnote=f'Download of {player_url} failed')
|
||||||
headers=self._generate_webpage_headers())
|
|
||||||
if code:
|
if code:
|
||||||
self._code_cache[player_id] = code
|
self._code_cache[player_id] = code
|
||||||
return self._code_cache.get(player_id)
|
return self._code_cache.get(player_id)
|
||||||
|
@ -3540,8 +3334,7 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
||||||
|
|
||||||
self._download_webpage(
|
self._download_webpage(
|
||||||
url, video_id, f'Marking {label}watched',
|
url, video_id, f'Marking {label}watched',
|
||||||
'Unable to mark watched', fatal=False,
|
'Unable to mark watched', fatal=False)
|
||||||
headers=self._generate_webpage_headers())
|
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def _extract_from_webpage(cls, url, webpage):
|
def _extract_from_webpage(cls, url, webpage):
|
||||||
|
@ -4523,7 +4316,7 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
||||||
if pp:
|
if pp:
|
||||||
query['pp'] = pp
|
query['pp'] = pp
|
||||||
webpage = self._download_webpage(
|
webpage = self._download_webpage(
|
||||||
webpage_url, video_id, fatal=False, query=query, headers=self._generate_webpage_headers())
|
webpage_url, video_id, fatal=False, query=query)
|
||||||
|
|
||||||
master_ytcfg = self.extract_ytcfg(video_id, webpage) or self._get_default_ytcfg()
|
master_ytcfg = self.extract_ytcfg(video_id, webpage) or self._get_default_ytcfg()
|
||||||
|
|
||||||
|
@ -4666,6 +4459,9 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
||||||
self.raise_geo_restricted(subreason, countries, metadata_available=True)
|
self.raise_geo_restricted(subreason, countries, metadata_available=True)
|
||||||
reason += f'. {subreason}'
|
reason += f'. {subreason}'
|
||||||
if reason:
|
if reason:
|
||||||
|
if 'sign in' in reason.lower():
|
||||||
|
reason = remove_end(reason, 'This helps protect our community. Learn more')
|
||||||
|
reason = f'{remove_end(reason.strip(), ".")}. {self._youtube_login_hint}'
|
||||||
self.raise_no_formats(reason, expected=True)
|
self.raise_no_formats(reason, expected=True)
|
||||||
|
|
||||||
keywords = get_first(video_details, 'keywords', expected_type=list) or []
|
keywords = get_first(video_details, 'keywords', expected_type=list) or []
|
||||||
|
@ -5291,7 +5087,7 @@ class YoutubeTabBaseInfoExtractor(YoutubeBaseInfoExtractor):
|
||||||
def _rich_entries(self, rich_grid_renderer):
|
def _rich_entries(self, rich_grid_renderer):
|
||||||
renderer = traverse_obj(
|
renderer = traverse_obj(
|
||||||
rich_grid_renderer,
|
rich_grid_renderer,
|
||||||
('content', ('videoRenderer', 'reelItemRenderer', 'playlistRenderer', 'shortsLockupViewModel'), any)) or {}
|
('content', ('videoRenderer', 'reelItemRenderer', 'playlistRenderer', 'shortsLockupViewModel', 'lockupViewModel'), any)) or {}
|
||||||
video_id = renderer.get('videoId')
|
video_id = renderer.get('videoId')
|
||||||
if video_id:
|
if video_id:
|
||||||
yield self._extract_video(renderer)
|
yield self._extract_video(renderer)
|
||||||
|
@ -5318,6 +5114,18 @@ class YoutubeTabBaseInfoExtractor(YoutubeBaseInfoExtractor):
|
||||||
})),
|
})),
|
||||||
thumbnails=self._extract_thumbnails(renderer, 'thumbnail', final_key='sources'))
|
thumbnails=self._extract_thumbnails(renderer, 'thumbnail', final_key='sources'))
|
||||||
return
|
return
|
||||||
|
# lockupViewModel extraction
|
||||||
|
content_id = renderer.get('contentId')
|
||||||
|
if content_id and renderer.get('contentType') == 'LOCKUP_CONTENT_TYPE_PODCAST':
|
||||||
|
yield self.url_result(
|
||||||
|
f'https://www.youtube.com/playlist?list={content_id}',
|
||||||
|
ie=YoutubeTabIE, video_id=content_id,
|
||||||
|
**traverse_obj(renderer, {
|
||||||
|
'title': ('metadata', 'lockupMetadataViewModel', 'title', 'content', {str}),
|
||||||
|
}),
|
||||||
|
thumbnails=self._extract_thumbnails(renderer, (
|
||||||
|
'contentImage', 'collectionThumbnailViewModel', 'primaryThumbnail', 'thumbnailViewModel', 'image'), final_key='sources'))
|
||||||
|
return
|
||||||
|
|
||||||
def _video_entry(self, video_renderer):
|
def _video_entry(self, video_renderer):
|
||||||
video_id = video_renderer.get('videoId')
|
video_id = video_renderer.get('videoId')
|
||||||
|
@ -5811,7 +5619,7 @@ class YoutubeTabBaseInfoExtractor(YoutubeBaseInfoExtractor):
|
||||||
webpage, data = None, None
|
webpage, data = None, None
|
||||||
for retry in self.RetryManager(fatal=fatal):
|
for retry in self.RetryManager(fatal=fatal):
|
||||||
try:
|
try:
|
||||||
webpage = self._download_webpage(url, item_id, note='Downloading webpage', headers=self._generate_webpage_headers())
|
webpage = self._download_webpage(url, item_id, note='Downloading webpage')
|
||||||
data = self.extract_yt_initial_data(item_id, webpage or '', fatal=fatal) or {}
|
data = self.extract_yt_initial_data(item_id, webpage or '', fatal=fatal) or {}
|
||||||
except ExtractorError as e:
|
except ExtractorError as e:
|
||||||
if isinstance(e.cause, network_exceptions):
|
if isinstance(e.cause, network_exceptions):
|
||||||
|
@ -6910,22 +6718,22 @@ class YoutubeTabIE(YoutubeTabBaseInfoExtractor):
|
||||||
},
|
},
|
||||||
'playlist_count': 0,
|
'playlist_count': 0,
|
||||||
}, {
|
}, {
|
||||||
# Podcasts tab, with rich entry playlistRenderers
|
# Podcasts tab, with rich entry lockupViewModel
|
||||||
'url': 'https://www.youtube.com/@99percentinvisiblepodcast/podcasts',
|
'url': 'https://www.youtube.com/@99percentinvisiblepodcast/podcasts',
|
||||||
'info_dict': {
|
'info_dict': {
|
||||||
'id': 'UCVMF2HD4ZgC0QHpU9Yq5Xrw',
|
'id': 'UCVMF2HD4ZgC0QHpU9Yq5Xrw',
|
||||||
'channel_id': 'UCVMF2HD4ZgC0QHpU9Yq5Xrw',
|
'channel_id': 'UCVMF2HD4ZgC0QHpU9Yq5Xrw',
|
||||||
'uploader_url': 'https://www.youtube.com/@99percentinvisiblepodcast',
|
'uploader_url': 'https://www.youtube.com/@99percentinvisiblepodcast',
|
||||||
'description': 'md5:3a0ed38f1ad42a68ef0428c04a15695c',
|
'description': 'md5:3a0ed38f1ad42a68ef0428c04a15695c',
|
||||||
'title': '99 Percent Invisible - Podcasts',
|
'title': '99% Invisible - Podcasts',
|
||||||
'uploader': '99 Percent Invisible',
|
'uploader': '99% Invisible',
|
||||||
'channel_follower_count': int,
|
'channel_follower_count': int,
|
||||||
'channel_url': 'https://www.youtube.com/channel/UCVMF2HD4ZgC0QHpU9Yq5Xrw',
|
'channel_url': 'https://www.youtube.com/channel/UCVMF2HD4ZgC0QHpU9Yq5Xrw',
|
||||||
'tags': [],
|
'tags': [],
|
||||||
'channel': '99 Percent Invisible',
|
'channel': '99% Invisible',
|
||||||
'uploader_id': '@99percentinvisiblepodcast',
|
'uploader_id': '@99percentinvisiblepodcast',
|
||||||
},
|
},
|
||||||
'playlist_count': 0,
|
'playlist_count': 5,
|
||||||
}, {
|
}, {
|
||||||
# Releases tab, with rich entry playlistRenderers (same as Podcasts tab)
|
# Releases tab, with rich entry playlistRenderers (same as Podcasts tab)
|
||||||
'url': 'https://www.youtube.com/@AHimitsu/releases',
|
'url': 'https://www.youtube.com/@AHimitsu/releases',
|
||||||
|
|
|
@ -419,7 +419,9 @@ def create_parser():
|
||||||
general.add_option(
|
general.add_option(
|
||||||
'--flat-playlist',
|
'--flat-playlist',
|
||||||
action='store_const', dest='extract_flat', const='in_playlist', default=False,
|
action='store_const', dest='extract_flat', const='in_playlist', default=False,
|
||||||
help='Do not extract the videos of a playlist, only list them')
|
help=(
|
||||||
|
'Do not extract a playlist\'s URL result entries; '
|
||||||
|
'some entry metadata may be missing and downloading may be bypassed'))
|
||||||
general.add_option(
|
general.add_option(
|
||||||
'--no-flat-playlist',
|
'--no-flat-playlist',
|
||||||
action='store_false', dest='extract_flat',
|
action='store_false', dest='extract_flat',
|
||||||
|
|
|
@ -1,8 +1,8 @@
|
||||||
# Autogenerated by devscripts/update-version.py
|
# Autogenerated by devscripts/update-version.py
|
||||||
|
|
||||||
__version__ = '2024.11.04'
|
__version__ = '2024.11.18'
|
||||||
|
|
||||||
RELEASE_GIT_HEAD = '197d0b03b6a3c8fe4fa5ace630eeffec629bf72c'
|
RELEASE_GIT_HEAD = '7ea2787920cccc6b8ea30791993d114fbd564434'
|
||||||
|
|
||||||
VARIANT = None
|
VARIANT = None
|
||||||
|
|
||||||
|
@ -12,4 +12,4 @@ CHANNEL = 'stable'
|
||||||
|
|
||||||
ORIGIN = 'yt-dlp/yt-dlp'
|
ORIGIN = 'yt-dlp/yt-dlp'
|
||||||
|
|
||||||
_pkg_version = '2024.11.04'
|
_pkg_version = '2024.11.18'
|
||||||
|
|
Loading…
Reference in New Issue
Block a user