attempt to call API for version-3 player to get more fields

This commit is contained in:
Stefan Lobbenmeier 2024-03-29 19:31:13 -03:00
parent 3905f64920
commit 6988130d0c

View File

@ -1,6 +1,7 @@
import re import re
from .common import InfoExtractor from .common import InfoExtractor
from ..compat import compat_str
from ..utils import ( from ..utils import (
NO_DEFAULT, NO_DEFAULT,
ExtractorError, ExtractorError,
@ -13,6 +14,7 @@ from ..utils import (
parse_codecs, parse_codecs,
qualities, qualities,
traverse_obj, traverse_obj,
try_call,
try_get, try_get,
unified_timestamp, unified_timestamp,
update_url_query, update_url_query,
@ -28,11 +30,11 @@ class ZDFBaseIE(InfoExtractor):
def _call_api(self, url, video_id, item, api_token=None, referrer=None): def _call_api(self, url, video_id, item, api_token=None, referrer=None):
headers = {} headers = {}
if api_token: if api_token:
headers['Api-Auth'] = f'Bearer {api_token}' headers['Api-Auth'] = 'Bearer %s' % api_token
if referrer: if referrer:
headers['Referer'] = referrer headers['Referer'] = referrer
return self._download_json( return self._download_json(
url, video_id, f'Downloading JSON {item}', headers=headers) url, video_id, 'Downloading JSON %s' % item, headers=headers)
@staticmethod @staticmethod
def _extract_subtitles(src): def _extract_subtitles(src):
@ -72,7 +74,7 @@ class ZDFBaseIE(InfoExtractor):
f.update({ f.update({
'url': format_url, 'url': format_url,
'format_id': join_nonempty('http', meta.get('type'), meta.get('quality')), 'format_id': join_nonempty('http', meta.get('type'), meta.get('quality')),
'tbr': int_or_none(self._search_regex(r'_(\d+)k_', format_url, 'tbr', default=None)), 'tbr': int_or_none(self._search_regex(r'_(\d+)k_', format_url, 'tbr', default=None))
}) })
new_formats = [f] new_formats = [f]
formats.extend(merge_dicts(f, { formats.extend(merge_dicts(f, {
@ -235,7 +237,7 @@ class ZDFIE(ZDFBaseIE):
'timestamp': 1641355200, 'timestamp': 1641355200,
'upload_date': '20220105', 'upload_date': '20220105',
}, },
'skip': 'No longer available "Diese Seite wurde leider nicht gefunden"', 'skip': 'No longer available "Diese Seite wurde leider nicht gefunden"'
}, { }, {
'url': 'https://www.zdf.de/serien/soko-stuttgart/das-geld-anderer-leute-100.html', 'url': 'https://www.zdf.de/serien/soko-stuttgart/das-geld-anderer-leute-100.html',
'info_dict': { 'info_dict': {
@ -269,7 +271,7 @@ class ZDFIE(ZDFBaseIE):
t = content['mainVideoContent']['http://zdf.de/rels/target'] t = content['mainVideoContent']['http://zdf.de/rels/target']
ptmd_path = traverse_obj(t, ( ptmd_path = traverse_obj(t, (
(('streams', 'default'), None), (('streams', 'default'), None),
('http://zdf.de/rels/streams/ptmd', 'http://zdf.de/rels/streams/ptmd-template'), ('http://zdf.de/rels/streams/ptmd', 'http://zdf.de/rels/streams/ptmd-template')
), get_all=False) ), get_all=False)
if not ptmd_path: if not ptmd_path:
raise ExtractorError('Could not extract ptmd_path') raise ExtractorError('Could not extract ptmd_path')
@ -302,7 +304,7 @@ class ZDFIE(ZDFBaseIE):
chapters = [{ chapters = [{
'start_time': chap.get('anchorOffset'), 'start_time': chap.get('anchorOffset'),
'end_time': next_chap.get('anchorOffset'), 'end_time': next_chap.get('anchorOffset'),
'title': chap.get('anchorLabel'), 'title': chap.get('anchorLabel')
} for chap, next_chap in zip(chapter_marks, chapter_marks[1:])] } for chap, next_chap in zip(chapter_marks, chapter_marks[1:])]
return merge_dicts(info, { return merge_dicts(info, {
@ -312,16 +314,23 @@ class ZDFIE(ZDFBaseIE):
'timestamp': unified_timestamp(content.get('editorialDate')), 'timestamp': unified_timestamp(content.get('editorialDate')),
'thumbnails': thumbnails, 'thumbnails': thumbnails,
'chapters': chapters or None, 'chapters': chapters or None,
'season': 'test'
}) })
def _extract_regular(self, url, player, video_id): def _extract_regular(self, url, player, video_id):
content = self._call_api( player_content_v2 = player['content']
player['content'], video_id, 'content', player['apiToken'], url) player_content_v3 = update_url_query(player_content_v2, {'profile': 'player-3'})
return self._extract_entry(player['content'], player, content, video_id)
content = try_call(
lambda: self._call_api(player_content_v3, video_id, 'content', player['apiToken'], url),
lambda: self._call_api(player_content_v2, video_id, 'content', player['apiToken'], url),
)
return self._extract_entry(player_content_v2, player, content, video_id)
def _extract_mobile(self, video_id): def _extract_mobile(self, video_id):
video = self._download_json( video = self._download_json(
f'https://zdf-cdn.live.cellular.de/mediathekV2/document/{video_id}', 'https://zdf-cdn.live.cellular.de/mediathekV2/document/%s' % video_id,
video_id) video_id)
formats = [] formats = []
@ -340,7 +349,7 @@ class ZDFIE(ZDFBaseIE):
if isinstance(teaser_bild, dict): if isinstance(teaser_bild, dict):
for thumbnail_key, thumbnail in teaser_bild.items(): for thumbnail_key, thumbnail in teaser_bild.items():
thumbnail_url = try_get( thumbnail_url = try_get(
thumbnail, lambda x: x['url'], str) thumbnail, lambda x: x['url'], compat_str)
if thumbnail_url: if thumbnail_url:
thumbnails.append({ thumbnails.append({
'url': thumbnail_url, 'url': thumbnail_url,
@ -355,7 +364,7 @@ class ZDFIE(ZDFBaseIE):
'description': document.get('beschreibung'), 'description': document.get('beschreibung'),
'duration': int_or_none(document.get('length')), 'duration': int_or_none(document.get('length')),
'timestamp': unified_timestamp(document.get('date')) or unified_timestamp( 'timestamp': unified_timestamp(document.get('date')) or unified_timestamp(
try_get(video, lambda x: x['meta']['editorialDate'], str)), try_get(video, lambda x: x['meta']['editorialDate'], compat_str)),
'thumbnails': thumbnails, 'thumbnails': thumbnails,
'subtitles': self._extract_subtitles(document), 'subtitles': self._extract_subtitles(document),
'formats': formats, 'formats': formats,
@ -404,10 +413,10 @@ class ZDFChannelIE(ZDFBaseIE):
@classmethod @classmethod
def suitable(cls, url): def suitable(cls, url):
return False if ZDFIE.suitable(url) else super().suitable(url) return False if ZDFIE.suitable(url) else super(ZDFChannelIE, cls).suitable(url)
def _og_search_title(self, webpage, fatal=False): def _og_search_title(self, webpage, fatal=False):
title = super()._og_search_title(webpage, fatal=fatal) title = super(ZDFChannelIE, self)._og_search_title(webpage, fatal=fatal)
return re.split(r'\s+[-|]\s+ZDF(?:mediathek)?$', title or '')[0] or None return re.split(r'\s+[-|]\s+ZDF(?:mediathek)?$', title or '')[0] or None
def _real_extract(self, url): def _real_extract(self, url):
@ -416,7 +425,7 @@ class ZDFChannelIE(ZDFBaseIE):
webpage = self._download_webpage(url, channel_id) webpage = self._download_webpage(url, channel_id)
matches = re.finditer( matches = re.finditer(
rf'''<div\b[^>]*?\sdata-plusbar-id\s*=\s*(["'])(?P<p_id>[\w-]+)\1[^>]*?\sdata-plusbar-url=\1(?P<url>{ZDFIE._VALID_URL})\1''', r'''<div\b[^>]*?\sdata-plusbar-id\s*=\s*(["'])(?P<p_id>[\w-]+)\1[^>]*?\sdata-plusbar-url=\1(?P<url>%s)\1''' % ZDFIE._VALID_URL,
webpage) webpage)
if self._downloader.params.get('noplaylist', False): if self._downloader.params.get('noplaylist', False):
@ -427,11 +436,11 @@ class ZDFChannelIE(ZDFBaseIE):
if entry: if entry:
return entry return entry
else: else:
self.to_screen(f'Downloading playlist {channel_id} - add --no-playlist to download just the main video') self.to_screen('Downloading playlist %s - add --no-playlist to download just the main video' % (channel_id, ))
def check_video(m): def check_video(m):
v_ref = self._search_regex( v_ref = self._search_regex(
r'''(<a\b[^>]*?\shref\s*=[^>]+?\sdata-target-id\s*=\s*(["']){}\2[^>]*>)'''.format(m.group('p_id')), r'''(<a\b[^>]*?\shref\s*=[^>]+?\sdata-target-id\s*=\s*(["'])%s\2[^>]*>)''' % (m.group('p_id'), ),
webpage, 'check id', default='') webpage, 'check id', default='')
v_ref = extract_attributes(v_ref) v_ref = extract_attributes(v_ref)
return v_ref.get('data-target-video-type') != 'novideo' return v_ref.get('data-target-video-type') != 'novideo'