mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2026-04-02 07:02:36 +00:00
[devscripts] Handle ejs updates for requirements files (#16374)
Authored by: bashonly, Grub4K Co-authored-by: Simon Sawicki <contact@grub4k.dev>
This commit is contained in:
parent
cdc465a346
commit
fcccbc6849
@ -22,8 +22,11 @@ from devscripts.utils import run_process
|
|||||||
REQUIREMENTS_PATH = pathlib.Path(__file__).parent.parent / 'bundle/requirements'
|
REQUIREMENTS_PATH = pathlib.Path(__file__).parent.parent / 'bundle/requirements'
|
||||||
INPUT_TMPL = 'requirements-{}.in'
|
INPUT_TMPL = 'requirements-{}.in'
|
||||||
OUTPUT_TMPL = 'requirements-{}.txt'
|
OUTPUT_TMPL = 'requirements-{}.txt'
|
||||||
COOLDOWN_DATE = (dt.datetime.today() - dt.timedelta(days=5)).strftime('%Y-%m-%d')
|
|
||||||
CUSTOM_COMPILE_COMMAND = 'python -m devscripts.update_bundle_requirements'
|
CUSTOM_COMPILE_COMMAND = 'python -m devscripts.update_bundle_requirements'
|
||||||
|
COOLDOWN_DATE = (dt.date.today() - dt.timedelta(days=7)).isoformat()
|
||||||
|
FUTURE_DATE = (dt.date.today() + dt.timedelta(days=1)).isoformat()
|
||||||
|
|
||||||
|
COOLDOWN_EXCEPTIONS = ('protobug', 'yt-dlp-ejs')
|
||||||
|
|
||||||
LINUX_GNU_PYTHON_VERSION = '3.13'
|
LINUX_GNU_PYTHON_VERSION = '3.13'
|
||||||
LINUX_MUSL_PYTHON_VERISON = '3.14'
|
LINUX_MUSL_PYTHON_VERISON = '3.14'
|
||||||
@ -152,14 +155,20 @@ def write_requirements_input(filepath: pathlib.Path, *args: str) -> None:
|
|||||||
def run_pip_compile(python_platform: str, python_version: str, requirements_input_path: pathlib.Path, *args: str) -> str:
|
def run_pip_compile(python_platform: str, python_version: str, requirements_input_path: pathlib.Path, *args: str) -> str:
|
||||||
return run_process(
|
return run_process(
|
||||||
'uv', 'pip', 'compile',
|
'uv', 'pip', 'compile',
|
||||||
|
'--no-config',
|
||||||
|
'--quiet',
|
||||||
|
'--no-progress',
|
||||||
|
'--color=never',
|
||||||
'--upgrade',
|
'--upgrade',
|
||||||
f'--exclude-newer={COOLDOWN_DATE}',
|
f'--exclude-newer={COOLDOWN_DATE}',
|
||||||
|
*(f'--exclude-newer-package={package}={FUTURE_DATE}' for package in COOLDOWN_EXCEPTIONS),
|
||||||
f'--python-platform={python_platform}',
|
f'--python-platform={python_platform}',
|
||||||
f'--python-version={python_version}',
|
f'--python-version={python_version}',
|
||||||
'--generate-hashes',
|
'--generate-hashes',
|
||||||
'--no-strip-markers',
|
'--no-strip-markers',
|
||||||
f'--custom-compile-command={CUSTOM_COMPILE_COMMAND}',
|
f'--custom-compile-command={CUSTOM_COMPILE_COMMAND}',
|
||||||
str(requirements_input_path),
|
str(requirements_input_path),
|
||||||
|
'--format=requirements.txt',
|
||||||
*args)
|
*args)
|
||||||
|
|
||||||
|
|
||||||
@ -174,7 +183,7 @@ def main():
|
|||||||
base_requirements_path.write_text(f'pyinstaller=={pyinstaller_version}\n')
|
base_requirements_path.write_text(f'pyinstaller=={pyinstaller_version}\n')
|
||||||
pyinstaller_builds_deps = run_pip_compile(
|
pyinstaller_builds_deps = run_pip_compile(
|
||||||
target.platform, target.version, base_requirements_path,
|
target.platform, target.version, base_requirements_path,
|
||||||
'--color=never', '--no-emit-package=pyinstaller').stdout
|
'--no-emit-package=pyinstaller').stdout
|
||||||
requirements_path = REQUIREMENTS_PATH / OUTPUT_TMPL.format(target_suffix)
|
requirements_path = REQUIREMENTS_PATH / OUTPUT_TMPL.format(target_suffix)
|
||||||
requirements_path.write_text(PYINSTALLER_BUILDS_TMPL.format(
|
requirements_path.write_text(PYINSTALLER_BUILDS_TMPL.format(
|
||||||
pyinstaller_builds_deps, asset_info['browser_download_url'], asset_info['digest']))
|
pyinstaller_builds_deps, asset_info['browser_download_url'], asset_info['digest']))
|
||||||
|
|||||||
@ -1,6 +1,7 @@
|
|||||||
#!/usr/bin/env python3
|
#!/usr/bin/env python3
|
||||||
from __future__ import annotations
|
from __future__ import annotations
|
||||||
|
|
||||||
|
import collections.abc
|
||||||
import contextlib
|
import contextlib
|
||||||
import io
|
import io
|
||||||
import json
|
import json
|
||||||
@ -18,7 +19,9 @@ HASHES = {{
|
|||||||
{hash_mapping}
|
{hash_mapping}
|
||||||
}}
|
}}
|
||||||
'''
|
'''
|
||||||
PREFIX = ' "yt-dlp-ejs=='
|
PACKAGE_NAME = 'yt-dlp-ejs'
|
||||||
|
PREFIX = f' "{PACKAGE_NAME}=='
|
||||||
|
PYPI_ARTIFACT_NAME = PACKAGE_NAME.replace('-', '_')
|
||||||
BASE_PATH = pathlib.Path(__file__).parent.parent
|
BASE_PATH = pathlib.Path(__file__).parent.parent
|
||||||
PYPROJECT_PATH = BASE_PATH / 'pyproject.toml'
|
PYPROJECT_PATH = BASE_PATH / 'pyproject.toml'
|
||||||
PACKAGE_PATH = BASE_PATH / 'yt_dlp/extractor/youtube/jsc/_builtin/vendor'
|
PACKAGE_PATH = BASE_PATH / 'yt_dlp/extractor/youtube/jsc/_builtin/vendor'
|
||||||
@ -32,6 +35,58 @@ ASSETS = {
|
|||||||
'yt.solver.core.js': True,
|
'yt.solver.core.js': True,
|
||||||
}
|
}
|
||||||
MAKEFILE_PATH = BASE_PATH / 'Makefile'
|
MAKEFILE_PATH = BASE_PATH / 'Makefile'
|
||||||
|
REQUIREMENTS_PATH = BASE_PATH / 'bundle/requirements'
|
||||||
|
|
||||||
|
|
||||||
|
def requirements_needs_update(
|
||||||
|
lines: collections.abc.Iterable[str],
|
||||||
|
package: str,
|
||||||
|
version: str,
|
||||||
|
):
|
||||||
|
identifier = f'{package}=='
|
||||||
|
for line in lines:
|
||||||
|
if line.startswith(identifier):
|
||||||
|
return not line.removeprefix(identifier).startswith(version)
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
def requirements_update(
|
||||||
|
lines: collections.abc.Iterable[str],
|
||||||
|
package: str,
|
||||||
|
new_version: str,
|
||||||
|
new_hashes: list[str],
|
||||||
|
):
|
||||||
|
first_comment = True
|
||||||
|
current = []
|
||||||
|
for line in lines:
|
||||||
|
if not line.endswith('\n'):
|
||||||
|
line += '\n'
|
||||||
|
|
||||||
|
if first_comment:
|
||||||
|
comment_line = line.strip()
|
||||||
|
if comment_line.startswith('#'):
|
||||||
|
yield line
|
||||||
|
continue
|
||||||
|
|
||||||
|
first_comment = False
|
||||||
|
yield '# It was later updated using devscripts/update_ejs.py\n'
|
||||||
|
|
||||||
|
current.append(line)
|
||||||
|
if line.endswith('\\\n'):
|
||||||
|
# continue logical line
|
||||||
|
continue
|
||||||
|
|
||||||
|
if not current[0].startswith(f'{package}=='):
|
||||||
|
yield from current
|
||||||
|
|
||||||
|
else:
|
||||||
|
yield f'{package}=={new_version} \\\n'
|
||||||
|
for digest in new_hashes[:-1]:
|
||||||
|
yield f' --hash={digest} \\\n'
|
||||||
|
yield f' --hash={new_hashes[-1]}\n'
|
||||||
|
|
||||||
|
current.clear()
|
||||||
|
|
||||||
|
|
||||||
def request(url: str):
|
def request(url: str):
|
||||||
@ -93,7 +148,7 @@ def main():
|
|||||||
current_version, _, _ = line.removeprefix(PREFIX).partition('"')
|
current_version, _, _ = line.removeprefix(PREFIX).partition('"')
|
||||||
|
|
||||||
if not current_version:
|
if not current_version:
|
||||||
print('yt-dlp-ejs dependency line could not be found')
|
print(f'{PACKAGE_NAME} dependency line could not be found')
|
||||||
return
|
return
|
||||||
|
|
||||||
makefile_info = makefile_variables(keys_only=True)
|
makefile_info = makefile_variables(keys_only=True)
|
||||||
@ -110,27 +165,36 @@ def main():
|
|||||||
|
|
||||||
version = info['tag_name']
|
version = info['tag_name']
|
||||||
if version == current_version:
|
if version == current_version:
|
||||||
print(f'yt-dlp-ejs is up to date! ({version})')
|
print(f'{PACKAGE_NAME} is up to date! ({version})')
|
||||||
return
|
return
|
||||||
|
|
||||||
print(f'Updating yt-dlp-ejs from {current_version} to {version}')
|
print(f'Updating {PACKAGE_NAME} from {current_version} to {version}')
|
||||||
hashes = []
|
hashes = []
|
||||||
|
requirements_hashes = []
|
||||||
wheel_info = {}
|
wheel_info = {}
|
||||||
for asset in info['assets']:
|
for asset in info['assets']:
|
||||||
name = asset['name']
|
name = asset['name']
|
||||||
is_wheel = name.startswith('yt_dlp_ejs-') and name.endswith('.whl')
|
digest = asset['digest']
|
||||||
|
|
||||||
|
# Is it the source distribution? If so, we only need its hash for the requirements files
|
||||||
|
if name == f'{PYPI_ARTIFACT_NAME}-{version}.tar.gz':
|
||||||
|
requirements_hashes.append(digest)
|
||||||
|
continue
|
||||||
|
|
||||||
|
is_wheel = name.startswith(f'{PYPI_ARTIFACT_NAME}-') and name.endswith('.whl')
|
||||||
if not is_wheel and name not in ASSETS:
|
if not is_wheel and name not in ASSETS:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
with request(asset['browser_download_url']) as resp:
|
with request(asset['browser_download_url']) as resp:
|
||||||
data = resp.read()
|
data = resp.read()
|
||||||
|
|
||||||
# verify digest from github
|
# verify digest from github
|
||||||
digest = asset['digest']
|
|
||||||
algo, _, expected = digest.partition(':')
|
algo, _, expected = digest.partition(':')
|
||||||
hexdigest = hashlib.new(algo, data).hexdigest()
|
hexdigest = hashlib.new(algo, data).hexdigest()
|
||||||
assert hexdigest == expected, f'downloaded attest mismatch ({hexdigest!r} != {expected!r})'
|
assert hexdigest == expected, f'downloaded attest mismatch ({hexdigest!r} != {expected!r})'
|
||||||
|
|
||||||
if is_wheel:
|
if is_wheel:
|
||||||
|
requirements_hashes.append(digest)
|
||||||
wheel_info = makefile_variables(version, name, digest, data)
|
wheel_info = makefile_variables(version, name, digest, data)
|
||||||
continue
|
continue
|
||||||
|
|
||||||
@ -161,6 +225,12 @@ def main():
|
|||||||
makefile = makefile.replace(f'{key} = {makefile_info[key]}', f'{key} = {wheel_info[key]}')
|
makefile = makefile.replace(f'{key} = {makefile_info[key]}', f'{key} = {wheel_info[key]}')
|
||||||
MAKEFILE_PATH.write_text(makefile)
|
MAKEFILE_PATH.write_text(makefile)
|
||||||
|
|
||||||
|
for req in REQUIREMENTS_PATH.glob('requirements-*.txt'):
|
||||||
|
lines = req.read_text().splitlines(True)
|
||||||
|
if requirements_needs_update(lines, PACKAGE_NAME, version):
|
||||||
|
with req.open(mode='w') as f:
|
||||||
|
f.writelines(requirements_update(lines, PACKAGE_NAME, version, requirements_hashes))
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user