summaryrefslogtreecommitdiffstats
path: root/bin/update/create_partial_update.py
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-15 09:29:03 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-15 09:29:03 +0000
commit5a7157d319477830426797532e02ac39d3b859f4 (patch)
tree3773f5ce209bee14a5643e98672e0f3828c71434 /bin/update/create_partial_update.py
parentReleasing progress-linux version 4:24.2.0-3~progress7.99u1. (diff)
downloadlibreoffice-5a7157d319477830426797532e02ac39d3b859f4.tar.xz
libreoffice-5a7157d319477830426797532e02ac39d3b859f4.zip
Merging upstream version 4:24.2.1.
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'bin/update/create_partial_update.py')
-rwxr-xr-xbin/update/create_partial_update.py185
1 files changed, 46 insertions, 139 deletions
diff --git a/bin/update/create_partial_update.py b/bin/update/create_partial_update.py
index 2730c4765f..8c49bd9159 100755
--- a/bin/update/create_partial_update.py
+++ b/bin/update/create_partial_update.py
@@ -1,90 +1,14 @@
#!/usr/bin/env python3
+import glob
import json
import os
+import re
import subprocess
import sys
-import requests
-
-from path import UpdaterPath, mkdir_p, convert_to_unix, convert_to_native
+from path import UpdaterPath, convert_to_native
from signing import sign_mar_file
-from tools import get_file_info, get_hash
-from uncompress_mar import extract_mar
-
-BUF_SIZE = 1024
-current_dir_path = os.path.dirname(os.path.realpath(convert_to_unix(__file__)))
-
-
-class InvalidFileException(Exception):
-
- def __init__(self, *args, **kwargs):
- super().__init__(self, *args, **kwargs)
-
-
-def download_file(filepath, url, hash_string):
- with open(filepath, "wb") as f:
- response = requests.get(url, stream=True)
-
- if not response.ok:
- return
-
- for block in response.iter_content(1024):
- f.write(block)
-
- file_hash = get_hash(filepath)
-
- if file_hash != hash_string:
- raise InvalidFileException(
- "file hash does not match for file %s: Expected %s, Got: %s" % (url, hash_string, file_hash))
-
-
-def handle_language(lang_entries, filedir):
- langs = {}
- for lang, data in lang_entries.items():
- lang_dir = os.path.join(filedir, lang)
- lang_file = os.path.join(lang_dir, "lang.mar")
- mkdir_p(lang_dir)
- download_file(lang_file, data["url"], data["hash"])
- dir_path = os.path.join(lang_dir, "lang")
- mkdir_p(dir_path)
- extract_mar(lang_file, dir_path)
- langs[lang] = dir_path
-
- return langs
-
-
-def download_mar_for_update_channel_and_platform(server_url, channel, platform, temp_dir):
- base_url = server_url + "update/partial-targets/1/"
- url = base_url + platform + "/" + channel
- r = requests.get(url)
- if r.status_code != 200:
- print(r.content)
- raise Exception("download failed")
-
- update_info = json.loads(r.content.decode("utf-8"))
- update_files = update_info['updates']
- downloaded_updates = {}
- for update_file in update_files:
- build = update_file["build"]
- filedir = os.path.join(temp_dir, build)
-
- mkdir_p(filedir)
-
- filepath = filedir + "/complete.mar"
- url = update_file["update"]["url"]
- expected_hash = update_file["update"]["hash"]
- download_file(filepath, url, expected_hash)
-
- dir_path = os.path.join(filedir, "complete")
- mkdir_p(dir_path)
- extract_mar(filepath, dir_path)
-
- downloaded_updates[build] = {"complete": dir_path}
-
- langs = handle_language(update_file["languages"], filedir)
- downloaded_updates[build]["languages"] = langs
-
- return downloaded_updates
+from tools import get_file_info, uncompress_file_to_dir
def generate_file_name(old_build_id, mar_name_prefix):
@@ -92,16 +16,6 @@ def generate_file_name(old_build_id, mar_name_prefix):
return name
-def generate_lang_file_name(old_build_id, mar_name_prefix, lang):
- name = "%s_%s_from_%s_partial.mar" % (mar_name_prefix, lang, old_build_id)
- return name
-
-
-def add_single_dir(path):
- dir_name = [os.path.join(path, name) for name in os.listdir(path) if os.path.isdir(os.path.join(path, name))]
- return dir_name[0]
-
-
def main():
workdir = sys.argv[1]
@@ -109,60 +23,53 @@ def main():
updater_path.ensure_dir_exist()
mar_name_prefix = sys.argv[2]
- server_url = sys.argv[3]
- channel = sys.argv[4]
- certificate_path = sys.argv[5]
- certificate_name = sys.argv[6]
- base_url = sys.argv[7]
- platform = sys.argv[8]
- build_id = sys.argv[9]
-
- current_build_path = updater_path.get_current_build_dir()
- mar_dir = updater_path.get_mar_dir()
- temp_dir = updater_path.get_previous_build_dir()
- update_dir = updater_path.get_update_dir()
-
- current_build_path = add_single_dir(current_build_path)
- if sys.platform == "cygwin":
- current_build_path = add_single_dir(current_build_path)
-
- updates = download_mar_for_update_channel_and_platform(server_url, channel, platform, temp_dir)
-
- data = {"partials": []}
+ channel = sys.argv[3]
+ certificate_path = sys.argv[4]
+ certificate_name = sys.argv[5]
+ base_url = sys.argv[6]
+ product_name = sys.argv[7]
+ version = sys.argv[8]
+ old_msi = sys.argv[9]
+
+ old_uncompress_dir = uncompress_file_to_dir(old_msi, updater_path.get_previous_build_dir())
+ versionini = os.path.join(old_uncompress_dir, 'program', 'version.ini') #TODO: Linux, macOS
+ old_build_id = None
+ with open(versionini) as f:
+ for l in f:
+ m = re.fullmatch('buildid=(.*)', l.rstrip())
+ if m:
+ old_build_id = m.group(1)
+ break
+ if old_build_id is None:
+ raise Exception(f'Cannot find buildid in {versionini}')
+
+ new_msi_file_glob = os.path.join(updater_path.get_workdir(), "installation", product_name, "msi", "install", "*", f'{product_name}_*.msi')
+ new_msi_files = glob.glob(new_msi_file_glob)
+ if len(new_msi_files) != 1:
+ raise Exception(f'`{new_msi_file_glob}` does not match exactly one file')
+ new_msi_file = new_msi_files[0]
+ new_uncompress_dir = uncompress_file_to_dir(new_msi_file, updater_path.get_current_build_dir())
- for build, update in updates.items():
- file_name = generate_file_name(build, mar_name_prefix)
- mar_file = os.path.join(update_dir, file_name)
- subprocess.call([os.path.join(current_dir_path, 'make_incremental_update.sh'), convert_to_native(mar_file),
- convert_to_native(update["complete"]), convert_to_native(current_build_path)])
- sign_mar_file(update_dir, certificate_path, certificate_name, mar_file, mar_name_prefix)
-
- partial_info = {"file": get_file_info(mar_file, base_url), "from": build, "to": build_id,
- "languages": {}}
-
- # on Windows we don't use language packs
- if sys.platform != "cygwin":
- for lang, lang_info in update["languages"].items():
- lang_name = generate_lang_file_name(build, mar_name_prefix, lang)
-
- # write the file into the final directory
- lang_mar_file = os.path.join(update_dir, lang_name)
+ update_dir = updater_path.get_update_dir()
- # the directory of the old language file is of the form
- # workdir/mar/language/en-US/LibreOffice_<version>_<os>_archive_langpack_<lang>/
- language_dir = add_single_dir(os.path.join(mar_dir, "language", lang))
- subprocess.call(
- [os.path.join(current_dir_path, 'make_incremental_update.sh'), convert_to_native(lang_mar_file),
- convert_to_native(lang_info), convert_to_native(language_dir)])
- sign_mar_file(update_dir, certificate_path, certificate_name, lang_mar_file, mar_name_prefix)
+ file_name = generate_file_name(old_build_id, mar_name_prefix)
+ mar_file = os.path.join(update_dir, file_name)
- # add the partial language info
- partial_info["languages"][lang] = get_file_info(lang_mar_file, base_url)
+ os.putenv('MOZ_PRODUCT_VERSION', version)
+ os.putenv('MAR_CHANNEL_ID', 'LOOnlineUpdater')
+ subprocess.call([os.path.join(workdir, 'UnpackedTarball/onlineupdate/tools/update-packaging/make_incremental_update.sh'), convert_to_native(mar_file),
+ convert_to_native(old_uncompress_dir), convert_to_native(new_uncompress_dir)])
- data["partials"].append(partial_info)
+ sign_mar_file(update_dir, certificate_path, certificate_name, mar_file, mar_name_prefix)
- with open(os.path.join(update_dir, "partial_update_info.json"), "w") as f:
- json.dump(data, f)
+ data = {
+ 'from': old_build_id,
+ 'see also': '',
+ 'update': get_file_info(mar_file, base_url),
+ 'languages': {}
+ };
+ with open(os.path.join(update_dir, channel), "w") as f:
+ json.dump(data, f, indent=4)
if __name__ == '__main__':