diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-19 00:47:55 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-19 00:47:55 +0000 |
commit | 26a029d407be480d791972afb5975cf62c9360a6 (patch) | |
tree | f435a8308119effd964b339f76abb83a57c29483 /third_party/rust/encoding_rs/generate-encoding-data.py | |
parent | Initial commit. (diff) | |
download | firefox-26a029d407be480d791972afb5975cf62c9360a6.tar.xz firefox-26a029d407be480d791972afb5975cf62c9360a6.zip |
Adding upstream version 124.0.1.upstream/124.0.1
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'third_party/rust/encoding_rs/generate-encoding-data.py')
-rw-r--r-- | third_party/rust/encoding_rs/generate-encoding-data.py | 2008 |
1 files changed, 2008 insertions, 0 deletions
diff --git a/third_party/rust/encoding_rs/generate-encoding-data.py b/third_party/rust/encoding_rs/generate-encoding-data.py new file mode 100644 index 0000000000..99cec1adc5 --- /dev/null +++ b/third_party/rust/encoding_rs/generate-encoding-data.py @@ -0,0 +1,2008 @@ +#!/usr/bin/python + +# Copyright Mozilla Foundation. See the COPYRIGHT +# file at the top-level directory of this distribution. +# +# Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or +# https://www.apache.org/licenses/LICENSE-2.0> or the MIT license +# <LICENSE-MIT or https://opensource.org/licenses/MIT>, at your +# option. This file may not be copied, modified, or distributed +# except according to those terms. + +import json +import subprocess +import sys +import os.path + +if (not os.path.isfile("../encoding/encodings.json")) or (not os.path.isfile("../encoding/indexes.json")): + sys.stderr.write("This script needs a clone of https://github.com/whatwg/encoding/ (preferably at revision f381389) next to the encoding_rs directory.\n"); + sys.exit(-1) + +if not os.path.isfile("../encoding_c/src/lib.rs"): + sys.stderr.write("This script also writes the generated parts of the encoding_c crate and needs a clone of https://github.com/hsivonen/encoding_c next to the encoding_rs directory.\n"); + sys.exit(-1) + +if not os.path.isfile("../codepage/src/lib.rs"): + sys.stderr.write("This script also writes the generated parts of the codepage crate and needs a clone of https://github.com/hsivonen/codepage next to the encoding_rs directory.\n"); + sys.exit(-1) + +def cmp_from_end(one, other): + c = cmp(len(one), len(other)) + if c != 0: + return c + i = len(one) - 1 + while i >= 0: + c = cmp(one[i], other[i]) + if c != 0: + return c + i -= 1 + return 0 + + +class Label: + def __init__(self, label, preferred): + self.label = label + self.preferred = preferred + def __cmp__(self, other): + return cmp_from_end(self.label, other.label) + +class CodePage: + def __init__(self, code_page, preferred): + self.code_page = code_page + self.preferred = preferred + def __cmp__(self, other): + return self.code_page, other.code_page + +def static_u16_table(name, data): + data_file.write('''pub static %s: [u16; %d] = [ + ''' % (name, len(data))) + + for i in xrange(len(data)): + data_file.write('0x%04X,\n' % data[i]) + + data_file.write(''']; + + ''') + +def static_u16_table_from_indexable(name, data, item, feature): + data_file.write('''#[cfg(all( + feature = "less-slow-%s", + not(feature = "fast-%s") +))] +static %s: [u16; %d] = [ + ''' % (feature, feature, name, len(data))) + + for i in xrange(len(data)): + data_file.write('0x%04X,\n' % data[i][item]) + + data_file.write(''']; + + ''') + +def static_u8_pair_table_from_indexable(name, data, item, feature): + data_file.write('''#[cfg(all( + feature = "less-slow-%s", + not(feature = "fast-%s") +))] +static %s: [[u8; 2]; %d] = [ + ''' % (feature, feature, name, len(data))) + + for i in xrange(len(data)): + data_file.write('[0x%02X, 0x%02X],\n' % data[i][item]) + + data_file.write(''']; + + ''') + +def static_u8_pair_table(name, data, feature): + data_file.write('''#[cfg(feature = "%s")] +static %s: [[u8; 2]; %d] = [ + ''' % (feature, name, len(data))) + + for i in xrange(len(data)): + pair = data[i] + if not pair: + pair = (0, 0) + data_file.write('[0x%02X, 0x%02X],\n' % pair) + + data_file.write(''']; + + ''') + +preferred = [] + +dom = [] + +labels = [] + +data = json.load(open("../encoding/encodings.json", "r")) + +indexes = json.load(open("../encoding/indexes.json", "r")) + +single_byte = [] + +multi_byte = [] + +def to_camel_name(name): + if name == u"iso-8859-8-i": + return u"Iso8I" + if name.startswith(u"iso-8859-"): + return name.replace(u"iso-8859-", u"Iso") + return name.title().replace(u"X-", u"").replace(u"-", u"").replace(u"_", u"") + +def to_constant_name(name): + return name.replace(u"-", u"_").upper() + +def to_snake_name(name): + return name.replace(u"-", u"_").lower() + +def to_dom_name(name): + return name + +# Guestimate based on +# https://w3techs.com/technologies/overview/character_encoding/all +# whose methodology is known to be bogus, but the results are credible for +# this purpose. UTF-16LE lifted up due to prevalence on Windows and +# "ANSI codepages" prioritized. +encodings_by_code_page_frequency = [ + "UTF-8", + "UTF-16LE", + "windows-1252", + "windows-1251", + "GBK", + "Shift_JIS", + "EUC-KR", + "windows-1250", + "windows-1256", + "windows-1254", + "Big5", + "windows-874", + "windows-1255", + "windows-1253", + "windows-1257", + "windows-1258", + "EUC-JP", + "ISO-8859-2", + "ISO-8859-15", + "ISO-8859-7", + "KOI8-R", + "gb18030", + "ISO-8859-5", + "ISO-8859-8-I", + "ISO-8859-4", + "ISO-8859-6", + "ISO-2022-JP", + "KOI8-U", + "ISO-8859-13", + "ISO-8859-3", + "UTF-16BE", + "IBM866", + "ISO-8859-10", + "ISO-8859-8", + "macintosh", + "x-mac-cyrillic", + "ISO-8859-14", + "ISO-8859-16", +] + +encodings_by_code_page = { + 932: "Shift_JIS", + 936: "GBK", + 949: "EUC-KR", + 950: "Big5", + 866: "IBM866", + 874: "windows-874", + 1200: "UTF-16LE", + 1201: "UTF-16BE", + 1250: "windows-1250", + 1251: "windows-1251", + 1252: "windows-1252", + 1253: "windows-1253", + 1254: "windows-1254", + 1255: "windows-1255", + 1256: "windows-1256", + 1257: "windows-1257", + 1258: "windows-1258", + 10000: "macintosh", + 10017: "x-mac-cyrillic", + 20866: "KOI8-R", + 20932: "EUC-JP", + 21866: "KOI8-U", + 28592: "ISO-8859-2", + 28593: "ISO-8859-3", + 28594: "ISO-8859-4", + 28595: "ISO-8859-5", + 28596: "ISO-8859-6", + 28597: "ISO-8859-7", + 28598: "ISO-8859-8", + 28600: "ISO-8859-10", + 28603: "ISO-8859-13", + 28604: "ISO-8859-14", + 28605: "ISO-8859-15", + 28606: "ISO-8859-16", + 38598: "ISO-8859-8-I", + 50221: "ISO-2022-JP", + 54936: "gb18030", + 65001: "UTF-8", +} + +code_pages_by_encoding = {} + +for code_page, encoding in encodings_by_code_page.iteritems(): + code_pages_by_encoding[encoding] = code_page + +encoding_by_alias_code_page = { + 951: "Big5", + 10007: "x-mac-cyrillic", + 20936: "GBK", + 20949: "EUC-KR", + 21010: "UTF-16LE", # Undocumented; needed by calamine for Excel compat + 28591: "windows-1252", + 28599: "windows-1254", + 28601: "windows-874", + 50220: "ISO-2022-JP", + 50222: "ISO-2022-JP", + 50225: "replacement", # ISO-2022-KR + 50227: "replacement", # ISO-2022-CN + 51949: "EUC-JP", + 51936: "GBK", + 51949: "EUC-KR", + 52936: "replacement", # HZ +} + +code_pages = [] + +for name in encodings_by_code_page_frequency: + code_pages.append(code_pages_by_encoding[name]) + +encodings_by_code_page.update(encoding_by_alias_code_page) + +temp_keys = encodings_by_code_page.keys() +temp_keys.sort() +for code_page in temp_keys: + if not code_page in code_pages: + code_pages.append(code_page) + +# The position in the index (0 is the first index entry, +# i.e. byte value 0x80) that starts the longest run of +# consecutive code points. Must not be in the first +# quadrant. If the character to be encoded is not in this +# run, the part of the index after the run is searched +# forward. Then the part of the index from 32 to the start +# of the run. The first quadrant is searched last. +# +# If there is no obviously most useful longest run, +# the index here is just used to affect the search order. +start_of_longest_run_in_single_byte = { + "IBM866": 96, # 0 would be longest, but we don't want to start in the first quadrant + "windows-874": 33, + "windows-1250": 92, + "windows-1251": 64, + "windows-1252": 32, + "windows-1253": 83, + "windows-1254": 95, + "windows-1255": 96, + "windows-1256": 65, + "windows-1257": 95, # not actually longest + "windows-1258": 95, # not actually longest + "macintosh": 106, # useless + "x-mac-cyrillic": 96, + "KOI8-R": 64, # not actually longest + "KOI8-U": 64, # not actually longest + "ISO-8859-2": 95, # not actually longest + "ISO-8859-3": 95, # not actually longest + "ISO-8859-4": 95, # not actually longest + "ISO-8859-5": 46, + "ISO-8859-6": 65, + "ISO-8859-7": 83, + "ISO-8859-8": 96, + "ISO-8859-10": 90, # not actually longest + "ISO-8859-13": 95, # not actually longest + "ISO-8859-14": 95, + "ISO-8859-15": 63, + "ISO-8859-16": 95, # not actually longest +} + +# + +for group in data: + if group["heading"] == "Legacy single-byte encodings": + single_byte = group["encodings"] + else: + multi_byte.extend(group["encodings"]) + for encoding in group["encodings"]: + preferred.append(encoding["name"]) + for label in encoding["labels"]: + labels.append(Label(label, encoding["name"])) + +for name in preferred: + dom.append(to_dom_name(name)) + +preferred.sort() +labels.sort() +dom.sort(cmp=cmp_from_end) + +longest_label_length = 0 +longest_name_length = 0 +longest_label = None +longest_name = None + +for name in preferred: + if len(name) > longest_name_length: + longest_name_length = len(name) + longest_name = name + +for label in labels: + if len(label.label) > longest_label_length: + longest_label_length = len(label.label) + longest_label = label.label + +def longest_run_for_single_byte(name): + if name == u"ISO-8859-8-I": + name = u"ISO-8859-8" + index = indexes[name.lower()] + run_byte_offset = start_of_longest_run_in_single_byte[name] + run_bmp_offset = index[run_byte_offset] + previous_code_point = run_bmp_offset + run_length = 1 + while True: + i = run_byte_offset + run_length + if i == len(index): + break + code_point = index[i] + if previous_code_point + 1 != code_point: + break + previous_code_point = code_point + run_length += 1 + return (run_bmp_offset, run_byte_offset, run_length) + +def is_single_byte(name): + for encoding in single_byte: + if name == encoding["name"]: + return True + return False + +def read_non_generated(path): + partially_generated_file = open(path, "r") + full = partially_generated_file.read() + partially_generated_file.close() + + generated_begin = "// BEGIN GENERATED CODE. PLEASE DO NOT EDIT." + generated_end = "// END GENERATED CODE" + + generated_begin_index = full.find(generated_begin) + if generated_begin_index < 0: + sys.stderr.write("Can't find generated code start marker in %s. Exiting.\n" % path) + sys.exit(-1) + generated_end_index = full.find(generated_end) + if generated_end_index < 0: + sys.stderr.write("Can't find generated code end marker in %s. Exiting.\n" % path) + sys.exit(-1) + + return (full[0:generated_begin_index + len(generated_begin)], + full[generated_end_index:]) + +(lib_rs_begin, lib_rs_end) = read_non_generated("src/lib.rs") + +label_file = open("src/lib.rs", "w") + +label_file.write(lib_rs_begin) +label_file.write(""" +// Instead, please regenerate using generate-encoding-data.py + +const LONGEST_LABEL_LENGTH: usize = %d; // %s + +""" % (longest_label_length, longest_label)) + +for name in preferred: + variant = None + if is_single_byte(name): + (run_bmp_offset, run_byte_offset, run_length) = longest_run_for_single_byte(name) + variant = "SingleByte(&data::SINGLE_BYTE_DATA.%s, 0x%04X, %d, %d)" % (to_snake_name(u"iso-8859-8" if name == u"ISO-8859-8-I" else name), run_bmp_offset, run_byte_offset, run_length) + else: + variant = to_camel_name(name) + + docfile = open("doc/%s.txt" % name, "r") + doctext = docfile.read() + docfile.close() + + label_file.write('''/// The initializer for the [%s](static.%s.html) encoding. +/// +/// For use only for taking the address of this form when +/// Rust prohibits the use of the non-`_INIT` form directly, +/// such as in initializers of other `static`s. If in doubt, +/// use the corresponding non-`_INIT` reference-typed `static`. +/// +/// This part of the public API will go away if Rust changes +/// to make the referent of `pub const FOO: &'static Encoding` +/// unique cross-crate or if Rust starts allowing static arrays +/// to be initialized with `pub static FOO: &'static Encoding` +/// items. +pub static %s_INIT: Encoding = Encoding { + name: "%s", + variant: VariantEncoding::%s, +}; + +/// The %s encoding. +/// +%s/// +/// This will change from `static` to `const` if Rust changes +/// to make the referent of `pub const FOO: &'static Encoding` +/// unique cross-crate, so don't take the address of this +/// `static`. +pub static %s: &'static Encoding = &%s_INIT; + +''' % (to_dom_name(name), to_constant_name(name), to_constant_name(name), to_dom_name(name), variant, to_dom_name(name), doctext, to_constant_name(name), to_constant_name(name))) + +label_file.write("""static LABELS_SORTED: [&'static str; %d] = [ +""" % len(labels)) + +for label in labels: + label_file.write('''"%s",\n''' % label.label) + +label_file.write("""]; + +static ENCODINGS_IN_LABEL_SORT: [&'static Encoding; %d] = [ +""" % len(labels)) + +for label in labels: + label_file.write('''&%s_INIT,\n''' % to_constant_name(label.preferred)) + +label_file.write(''']; + +''') +label_file.write(lib_rs_end) +label_file.close() + +label_test_file = open("src/test_labels_names.rs", "w") +label_test_file.write('''// Any copyright to the test code below this comment is dedicated to the +// Public Domain. http://creativecommons.org/publicdomain/zero/1.0/ + +// THIS IS A GENERATED FILE. PLEASE DO NOT EDIT. +// Instead, please regenerate using generate-encoding-data.py + +use super::*; + +#[test] +fn test_all_labels() { +''') + +for label in labels: + label_test_file.write('''assert_eq!(Encoding::for_label(b"%s"), Some(%s));\n''' % (label.label, to_constant_name(label.preferred))) + +label_test_file.write('''} +''') +label_test_file.close() + +def null_to_zero(code_point): + if not code_point: + code_point = 0 + return code_point + +(data_rs_begin, data_rs_end) = read_non_generated("src/data.rs") + +data_file = open("src/data.rs", "w") +data_file.write(data_rs_begin) +data_file.write(''' +// Instead, please regenerate using generate-encoding-data.py + +#[repr(align(64))] // Align to cache lines +pub struct SingleByteData { +''') + +# Single-byte + +for encoding in single_byte: + name = encoding["name"] + if name == u"ISO-8859-8-I": + continue + + data_file.write(''' pub %s: [u16; 128], +''' % to_snake_name(name)) + +data_file.write('''} + +pub static SINGLE_BYTE_DATA: SingleByteData = SingleByteData { +''') + +for encoding in single_byte: + name = encoding["name"] + if name == u"ISO-8859-8-I": + continue + + data_file.write(''' %s: [ +''' % to_snake_name(name)) + + for code_point in indexes[name.lower()]: + data_file.write('0x%04X,\n' % null_to_zero(code_point)) + + data_file.write('''], +''') + +data_file.write('''}; + +''') + +# Big5 + +index = indexes["big5"] + +astralness = [] +low_bits = [] + +for code_point in index[942:19782]: + if code_point: + astralness.append(1 if code_point > 0xFFFF else 0) + low_bits.append(code_point & 0xFFFF) + else: + astralness.append(0) + low_bits.append(0) + +# pad length to multiple of 32 +for j in xrange(32 - (len(astralness) % 32)): + astralness.append(0) + +data_file.write('''#[cfg_attr(feature = "cargo-clippy", allow(unreadable_literal))] +static BIG5_ASTRALNESS: [u32; %d] = [ +''' % (len(astralness) / 32)) + +i = 0 +while i < len(astralness): + accu = 0 + for j in xrange(32): + accu |= astralness[i + j] << j + data_file.write('0x%08X,\n' % accu) + i += 32 + +data_file.write(''']; + +''') + +static_u16_table("BIG5_LOW_BITS", low_bits) + +# Encoder table for Level 1 Hanzi +# Note: If we were OK with doubling this table, we +# could use a directly-indexable table instead... +level1_hanzi_index = index[5495:10896] +level1_hanzi_pairs = [] +for i in xrange(len(level1_hanzi_index)): + hanzi_lead = (i / 157) + 0xA4 + hanzi_trail = (i % 157) + hanzi_trail += 0x40 if hanzi_trail < 0x3F else 0x62 + level1_hanzi_pairs.append((level1_hanzi_index[i], (hanzi_lead, hanzi_trail))) +level1_hanzi_pairs.append((0x4E5A, (0xC8, 0x7B))) +level1_hanzi_pairs.append((0x5202, (0xC8, 0x7D))) +level1_hanzi_pairs.append((0x9FB0, (0xC8, 0xA1))) +level1_hanzi_pairs.append((0x5188, (0xC8, 0xA2))) +level1_hanzi_pairs.append((0x9FB1, (0xC8, 0xA3))) +level1_hanzi_pairs.sort(key=lambda x: x[0]) + +static_u16_table_from_indexable("BIG5_LEVEL1_HANZI_CODE_POINTS", level1_hanzi_pairs, 0, "big5-hanzi-encode") +static_u8_pair_table_from_indexable("BIG5_LEVEL1_HANZI_BYTES", level1_hanzi_pairs, 1, "big5-hanzi-encode") + +# Fast Unified Ideograph encode +big5_unified_ideograph_bytes = [None] * (0x9FCC - 0x4E00) +for row in xrange(0x7E - 0x20): + for column in xrange(157): + pointer = 5024 + column + (row * 157) + code_point = index[pointer] + if code_point and code_point >= 0x4E00 and code_point <= 0x9FCB: + unified_offset = code_point - 0x4E00 + unified_lead = 0xA1 + row + unified_trail = (0x40 if column < 0x3F else 0x62) + column + if code_point == 0x5341 or code_point == 0x5345 or not big5_unified_ideograph_bytes[unified_offset]: + big5_unified_ideograph_bytes[unified_offset] = (unified_lead, unified_trail) + +static_u8_pair_table("BIG5_UNIFIED_IDEOGRAPH_BYTES", big5_unified_ideograph_bytes, "fast-big5-hanzi-encode") + +# JIS0208 + +index = indexes["jis0208"] + +# JIS 0208 Level 1 Kanji +static_u16_table("JIS0208_LEVEL1_KANJI", index[1410:4375]) + +# JIS 0208 Level 2 Kanji and Additional Kanji +static_u16_table("JIS0208_LEVEL2_AND_ADDITIONAL_KANJI", index[4418:7808]) + +# IBM Kanji +static_u16_table("IBM_KANJI", index[8272:8632]) + +# Check that the other instance is the same +if index[8272:8632] != index[10744:11104]: + raise Error() + +# JIS 0208 symbols (all non-Kanji, non-range items) +symbol_index = [] +symbol_triples = [] +pointers_to_scan = [ + (0, 188), + (658, 691), + (1159, 1221), +] +in_run = False +run_start_pointer = 0 +run_start_array_index = 0 +for (start, end) in pointers_to_scan: + for i in range(start, end): + code_point = index[i] + if in_run: + if code_point: + symbol_index.append(code_point) + else: + symbol_triples.append(run_start_pointer) + symbol_triples.append(i - run_start_pointer) + symbol_triples.append(run_start_array_index) + in_run = False + else: + if code_point: + in_run = True + run_start_pointer = i + run_start_array_index = len(symbol_index) + symbol_index.append(code_point) + if in_run: + symbol_triples.append(run_start_pointer) + symbol_triples.append(end - run_start_pointer) + symbol_triples.append(run_start_array_index) + in_run = False +if in_run: + raise Error() + +# Now add manually the two overlapping slices of +# index from the NEC/IBM extensions. +run_start_array_index = len(symbol_index) +symbol_index.extend(index[10736:10744]) +# Later +symbol_triples.append(10736) +symbol_triples.append(8) +symbol_triples.append(run_start_array_index) +# Earlier +symbol_triples.append(8644) +symbol_triples.append(4) +symbol_triples.append(run_start_array_index) + +static_u16_table("JIS0208_SYMBOLS", symbol_index) +static_u16_table("JIS0208_SYMBOL_TRIPLES", symbol_triples) + +# Write down the magic numbers needed when preferring the earlier case +data_file.write('''const IBM_SYMBOL_START: usize = %d;''' % (run_start_array_index + 1)) +data_file.write('''const IBM_SYMBOL_END: usize = %d;''' % (run_start_array_index + 4)) +data_file.write('''const IBM_SYMBOL_POINTER_START: usize = %d;''' % 8645) + +# JIS 0208 ranges (excluding kana) +range_triples = [] +pointers_to_scan = [ + (188, 281), + (470, 657), + (1128, 1159), + (8634, 8644), + (10716, 10736), +] +in_run = False +run_start_pointer = 0 +run_start_code_point = 0 +previous_code_point = 0 +for (start, end) in pointers_to_scan: + for i in range(start, end): + code_point = index[i] + if in_run: + if code_point: + if previous_code_point + 1 != code_point: + range_triples.append(run_start_pointer) + range_triples.append(i - run_start_pointer) + range_triples.append(run_start_code_point) + run_start_pointer = i + run_start_code_point = code_point + previous_code_point = code_point + else: + range_triples.append(run_start_pointer) + range_triples.append(i - run_start_pointer) + range_triples.append(run_start_code_point) + run_start_pointer = 0 + run_start_code_point = 0 + previous_code_point = 0 + in_run = False + else: + if code_point: + in_run = True + run_start_pointer = i + run_start_code_point = code_point + previous_code_point = code_point + if in_run: + range_triples.append(run_start_pointer) + range_triples.append(end - run_start_pointer) + range_triples.append(run_start_code_point) + run_start_pointer = 0 + run_start_code_point = 0 + previous_code_point = 0 + in_run = False +if in_run: + raise Error() + +static_u16_table("JIS0208_RANGE_TRIPLES", range_triples) + +# Encoder table for Level 1 Kanji +# Note: If we were OK with 30 KB more footprint, we +# could use a directly-indexable table instead... +level1_kanji_index = index[1410:4375] +level1_kanji_pairs = [] +for i in xrange(len(level1_kanji_index)): + pointer = 1410 + i + (lead, trail) = divmod(pointer, 188) + lead += 0x81 if lead < 0x1F else 0xC1 + trail += 0x40 if trail < 0x3F else 0x41 + level1_kanji_pairs.append((level1_kanji_index[i], (lead, trail))) +level1_kanji_pairs.sort(key=lambda x: x[0]) + +static_u16_table_from_indexable("JIS0208_LEVEL1_KANJI_CODE_POINTS", level1_kanji_pairs, 0, "kanji-encode") +static_u8_pair_table_from_indexable("JIS0208_LEVEL1_KANJI_SHIFT_JIS_BYTES", level1_kanji_pairs, 1, "kanji-encode") + +# Fast encoder table for Kanji +kanji_bytes = [None] * (0x9FA1 - 0x4E00) +for pointer in xrange(len(index)): + code_point = index[pointer] + if code_point and code_point >= 0x4E00 and code_point <= 0x9FA0: + (lead, trail) = divmod(pointer, 188) + lead += 0x81 if lead < 0x1F else 0xC1 + trail += 0x40 if trail < 0x3F else 0x41 + # unset the high bit of lead if IBM Kanji + if pointer >= 8272: + lead = lead & 0x7F + kanji_bytes[code_point - 0x4E00] = (lead, trail) + +static_u8_pair_table("JIS0208_KANJI_BYTES", kanji_bytes, "fast-kanji-encode") + +# ISO-2022-JP half-width katakana + +# index is still jis0208 +half_width_index = indexes["iso-2022-jp-katakana"] + +data_file.write('''pub static ISO_2022_JP_HALF_WIDTH_TRAIL: [u8; %d] = [ +''' % len(half_width_index)) + +for i in xrange(len(half_width_index)): + code_point = half_width_index[i] + pointer = index.index(code_point) + trail = pointer % 94 + 0x21 + data_file.write('0x%02X,\n' % trail) + +data_file.write(''']; + +''') + +# EUC-KR + +index = indexes["euc-kr"] + +# Unicode 1.1 Hangul above the old KS X 1001 block +# Compressed form takes 35% of uncompressed form +pointers = [] +offsets = [] +previous_code_point = 0 +for row in xrange(0x20): + for column in xrange(190): + i = column + (row * 190) + # Skip the gaps + if (column >= 0x1A and column < 0x20) or (column >= 0x3A and column < 0x40): + continue + code_point = index[i] + if previous_code_point > code_point: + raise Error() + if code_point - previous_code_point != 1: + adjustment = 0 + if column >= 0x40: + adjustment = 12 + elif column >= 0x20: + adjustment = 6 + pointers.append(column - adjustment + (row * (190 - 12))) + offsets.append(code_point) + previous_code_point = code_point + +static_u16_table("CP949_TOP_HANGUL_POINTERS", pointers) +static_u16_table("CP949_TOP_HANGUL_OFFSETS", offsets) + +# Unicode 1.1 Hangul to the left of the old KS X 1001 block +pointers = [] +offsets = [] +previous_code_point = 0 +for row in xrange(0x46 - 0x20): + for column in xrange(190 - 94): + i = 6080 + column + (row * 190) + # Skip the gaps + if (column >= 0x1A and column < 0x20) or (column >= 0x3A and column < 0x40): + continue + if i > 13127: + # Exclude unassigned on partial last row + break + code_point = index[i] + if previous_code_point > code_point: + raise Error() + if code_point - previous_code_point != 1: + adjustment = 0 + if column >= 0x40: + adjustment = 12 + elif column >= 0x20: + adjustment = 6 + pointers.append(column - adjustment + (row * (190 - 94 - 12))) + offsets.append(code_point) + previous_code_point = code_point + +static_u16_table("CP949_LEFT_HANGUL_POINTERS", pointers) +static_u16_table("CP949_LEFT_HANGUL_OFFSETS", offsets) + +# KS X 1001 Hangul +hangul_index = [] +previous_code_point = 0 +for row in xrange(0x48 - 0x2F): + for column in xrange(94): + code_point = index[9026 + column + (row * 190)] + if previous_code_point >= code_point: + raise Error() + hangul_index.append(code_point) + previous_code_point = code_point + +static_u16_table("KSX1001_HANGUL", hangul_index) + +# KS X 1001 Hanja +hanja_index = [] +for row in xrange(0x7D - 0x49): + for column in xrange(94): + hanja_index.append(index[13966 + column + (row * 190)]) + +static_u16_table("KSX1001_HANJA", hanja_index) + +# KS X 1001 symbols +symbol_index = [] +for i in range(6176, 6270): + symbol_index.append(index[i]) +for i in range(6366, 6437): + symbol_index.append(index[i]) + +static_u16_table("KSX1001_SYMBOLS", symbol_index) + +# KS X 1001 Uppercase Latin +subindex = [] +for i in range(7506, 7521): + subindex.append(null_to_zero(index[i])) + +static_u16_table("KSX1001_UPPERCASE", subindex) + +# KS X 1001 Lowercase Latin +subindex = [] +for i in range(7696, 7712): + subindex.append(index[i]) + +static_u16_table("KSX1001_LOWERCASE", subindex) + +# KS X 1001 Box drawing +subindex = [] +for i in range(7126, 7194): + subindex.append(index[i]) + +static_u16_table("KSX1001_BOX", subindex) + +# KS X 1001 other +pointers = [] +offsets = [] +previous_code_point = 0 +for row in xrange(10): + for column in xrange(94): + i = 6556 + column + (row * 190) + code_point = index[i] + # Exclude ranges that were processed as lookup tables + # or that contain unmapped cells by filling them with + # ASCII. Upon encode, ASCII code points will + # never appear as the search key. + if (i >= 6946 and i <= 6950): + code_point = i - 6946 + elif (i >= 6961 and i <= 6967): + code_point = i - 6961 + elif (i >= 6992 and i <= 6999): + code_point = i - 6992 + elif (i >= 7024 and i <= 7029): + code_point = i - 7024 + elif (i >= 7126 and i <= 7219): + code_point = i - 7126 + elif (i >= 7395 and i <= 7409): + code_point = i - 7395 + elif (i >= 7506 and i <= 7521): + code_point = i - 7506 + elif (i >= 7696 and i <= 7711): + code_point = i - 7696 + elif (i >= 7969 and i <= 7979): + code_point = i - 7969 + elif (i >= 8162 and i <= 8169): + code_point = i - 8162 + elif (i >= 8299 and i <= 8313): + code_point = i - 8299 + elif (i >= 8347 and i <= 8359): + code_point = i - 8347 + if code_point - previous_code_point != 1: + pointers.append(column + (row * 94)) + offsets.append(code_point) + previous_code_point = code_point + +static_u16_table("KSX1001_OTHER_POINTERS", pointers) +# Omit the last offset, because the end of the last line +# is unmapped, so we don't want to look at it. +static_u16_table("KSX1001_OTHER_UNSORTED_OFFSETS", offsets[:-1]) + +# Fast Hangul and Hanja encode +hangul_bytes = [None] * (0xD7A4 - 0xAC00) +hanja_unified_bytes = [None] * (0x9F9D - 0x4E00) +hanja_compatibility_bytes = [None] * (0xFA0C - 0xF900) +for row in xrange(0x7D): + for column in xrange(190): + pointer = column + (row * 190) + code_point = index[pointer] + if code_point: + lead = 0x81 + row + trail = 0x41 + column + if code_point >= 0xAC00 and code_point < 0xD7A4: + hangul_bytes[code_point - 0xAC00] = (lead, trail) + elif code_point >= 0x4E00 and code_point < 0x9F9D: + hanja_unified_bytes[code_point - 0x4E00] = (lead, trail) + elif code_point >= 0xF900 and code_point < 0xFA0C: + hanja_compatibility_bytes[code_point - 0xF900] = (lead, trail) + +static_u8_pair_table("CP949_HANGUL_BYTES", hangul_bytes, "fast-hangul-encode") +static_u8_pair_table("KSX1001_UNIFIED_HANJA_BYTES", hanja_unified_bytes, "fast-hanja-encode") +static_u8_pair_table("KSX1001_COMPATIBILITY_HANJA_BYTES", hanja_compatibility_bytes, "fast-hanja-encode") + +# JIS 0212 + +index = indexes["jis0212"] + +# JIS 0212 Kanji +static_u16_table("JIS0212_KANJI", index[1410:7211]) + +# JIS 0212 accented (all non-Kanji, non-range items) +symbol_index = [] +symbol_triples = [] +pointers_to_scan = [ + (0, 596), + (608, 644), + (656, 1409), +] +in_run = False +run_start_pointer = 0 +run_start_array_index = 0 +for (start, end) in pointers_to_scan: + for i in range(start, end): + code_point = index[i] + if in_run: + if code_point: + symbol_index.append(code_point) + elif index[i + 1]: + symbol_index.append(0) + else: + symbol_triples.append(run_start_pointer) + symbol_triples.append(i - run_start_pointer) + symbol_triples.append(run_start_array_index) + in_run = False + else: + if code_point: + in_run = True + run_start_pointer = i + run_start_array_index = len(symbol_index) + symbol_index.append(code_point) + if in_run: + symbol_triples.append(run_start_pointer) + symbol_triples.append(end - run_start_pointer) + symbol_triples.append(run_start_array_index) + in_run = False +if in_run: + raise Error() + +static_u16_table("JIS0212_ACCENTED", symbol_index) +static_u16_table("JIS0212_ACCENTED_TRIPLES", symbol_triples) + +# gb18030 + +index = indexes["gb18030"] + +# Unicode 1.1 ideographs above the old GB2312 block +# Compressed form takes 63% of uncompressed form +pointers = [] +offsets = [] +previous_code_point = 0 +for i in xrange(6080): + code_point = index[i] + if previous_code_point > code_point: + raise Error() + if code_point - previous_code_point != 1: + pointers.append(i) + offsets.append(code_point) + previous_code_point = code_point + +static_u16_table("GBK_TOP_IDEOGRAPH_POINTERS", pointers) +static_u16_table("GBK_TOP_IDEOGRAPH_OFFSETS", offsets) + +# Unicode 1.1 ideographs to the left of the old GB2312 block +# Compressed form takes 40% of uncompressed form +pointers = [] +offsets = [] +previous_code_point = 0 +for row in xrange(0x7D - 0x29): + for column in xrange(190 - 94): + i = 7790 + column + (row * 190) + if i > 23650: + # Exclude compatibility ideographs at the end + break + code_point = index[i] + if previous_code_point > code_point: + raise Error() + if code_point - previous_code_point != 1: + pointers.append(column + (row * (190 - 94))) + offsets.append(code_point) + previous_code_point = code_point + +static_u16_table("GBK_LEFT_IDEOGRAPH_POINTERS", pointers) +static_u16_table("GBK_LEFT_IDEOGRAPH_OFFSETS", offsets) + +# GBK other (excl. Ext A, Compat & PUA at the bottom) +pointers = [] +offsets = [] +previous_code_point = 0 +for row in xrange(0x29 - 0x20): + for column in xrange(190 - 94): + i = 6080 + column + (row * 190) + code_point = index[i] + if code_point - previous_code_point != 1: + pointers.append(column + (row * (190 - 94))) + offsets.append(code_point) + previous_code_point = code_point + +pointers.append((190 - 94) * (0x29 - 0x20)) +static_u16_table("GBK_OTHER_POINTERS", pointers) +static_u16_table("GBK_OTHER_UNSORTED_OFFSETS", offsets) + +# GBK bottom: Compatibility ideagraphs, Ext A and PUA +bottom_index = [] +# 5 compat following Unified Ideographs +for i in range(23651, 23656): + bottom_index.append(index[i]) +# Last row +for i in range(23750, 23846): + bottom_index.append(index[i]) + +static_u16_table("GBK_BOTTOM", bottom_index) + +# GB2312 Hanzi +# (and the 5 PUA code points in between Level 1 and Level 2) +hanzi_index = [] +for row in xrange(0x77 - 0x2F): + for column in xrange(94): + hanzi_index.append(index[9026 + column + (row * 190)]) + +static_u16_table("GB2312_HANZI", hanzi_index) + +# GB2312 symbols +symbol_index = [] +for i in xrange(94): + symbol_index.append(index[6176 + i]) + +static_u16_table("GB2312_SYMBOLS", symbol_index) + +# GB2312 symbols on Greek row (incl. PUA) +symbol_index = [] +for i in xrange(22): + symbol_index.append(index[7189 + i]) + +static_u16_table("GB2312_SYMBOLS_AFTER_GREEK", symbol_index) + +# GB2312 Pinyin +pinyin_index = [] +for i in xrange(32): + pinyin_index.append(index[7506 + i]) + +static_u16_table("GB2312_PINYIN", pinyin_index) + +# GB2312 other (excl. bottom PUA) +pointers = [] +offsets = [] +previous_code_point = 0 +for row in xrange(14): + for column in xrange(94): + i = 6366 + column + (row * 190) + code_point = index[i] + # Exclude the two ranges that were processed as + # lookup tables above by filling them with + # ASCII. Upon encode, ASCII code points will + # never appear as the search key. + if (i >= 7189 and i < 7189 + 22): + code_point = i - 7189 + elif (i >= 7506 and i < 7506 + 32): + code_point = i - 7506 + if code_point - previous_code_point != 1: + pointers.append(column + (row * 94)) + offsets.append(code_point) + previous_code_point = code_point + +pointers.append(14 * 94) +static_u16_table("GB2312_OTHER_POINTERS", pointers) +static_u16_table("GB2312_OTHER_UNSORTED_OFFSETS", offsets) + +# Non-gbk code points +pointers = [] +offsets = [] +for pair in indexes["gb18030-ranges"]: + if pair[1] == 0x10000: + break # the last entry doesn't fit in u16 + pointers.append(pair[0]) + offsets.append(pair[1]) + +static_u16_table("GB18030_RANGE_POINTERS", pointers) +static_u16_table("GB18030_RANGE_OFFSETS", offsets) + +# Encoder table for Level 1 Hanzi +# The units here really fit into 12 bits, but since we're +# looking for speed here, let's use 16 bits per unit. +# Once we use 16 bits per unit, we might as well precompute +# the output bytes. +level1_hanzi_index = hanzi_index[:(94 * (0xD8 - 0xB0) - 5)] +level1_hanzi_pairs = [] +for i in xrange(len(level1_hanzi_index)): + hanzi_lead = (i / 94) + 0xB0 + hanzi_trail = (i % 94) + 0xA1 + level1_hanzi_pairs.append((level1_hanzi_index[i], (hanzi_lead, hanzi_trail))) +level1_hanzi_pairs.sort(key=lambda x: x[0]) + +static_u16_table_from_indexable("GB2312_LEVEL1_HANZI_CODE_POINTS", level1_hanzi_pairs, 0, "gb-hanzi-encode") +static_u8_pair_table_from_indexable("GB2312_LEVEL1_HANZI_BYTES", level1_hanzi_pairs, 1, "gb-hanzi-encode") + +# Fast Hanzi encoder table +hanzi_bytes = [None] * (0x9FA7 - 0x4E00) +for row in xrange(126): + for column in xrange(190): + pointer = column + (row * 190) + code_point = index[pointer] + if code_point and code_point >= 0x4E00 and code_point <= 0x9FA6: + hanzi_lead = 0x81 + row + hanzi_trail = column + (0x40 if column < 0x3F else 0x41) + hanzi_bytes[code_point - 0x4E00] = (hanzi_lead, hanzi_trail) + +static_u8_pair_table("GBK_HANZI_BYTES", hanzi_bytes, "fast-gb-hanzi-encode") + +data_file.write(data_rs_end) + +data_file.close() + +# Variant + +variant_file = open("src/variant.rs", "w") +variant_file.write('''// Copyright Mozilla Foundation. See the COPYRIGHT +// file at the top-level directory of this distribution. +// +// Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or +// https://www.apache.org/licenses/LICENSE-2.0> or the MIT license +// <LICENSE-MIT or https://opensource.org/licenses/MIT>, at your +// option. This file may not be copied, modified, or distributed +// except according to those terms. + +// THIS IS A GENERATED FILE. PLEASE DO NOT EDIT. +// Instead, please regenerate using generate-encoding-data.py + +//! This module provides enums that wrap the various decoders and encoders. +//! The purpose is to make `Decoder` and `Encoder` `Sized` by writing the +//! dispatch explicitly for a finite set of specialized decoders and encoders. +//! Unfortunately, this means the compiler doesn't generate the dispatch code +//! and it has to be written here instead. +//! +//! The purpose of making `Decoder` and `Encoder` `Sized` is to allow stack +//! allocation in Rust code, including the convenience methods on `Encoding`. + +''') + +encoding_variants = [u"single-byte",] +for encoding in multi_byte: + if encoding["name"] in [u"UTF-16LE", u"UTF-16BE"]: + continue + else: + encoding_variants.append(encoding["name"]) +encoding_variants.append(u"UTF-16") + +decoder_variants = [] +for variant in encoding_variants: + if variant == u"GBK": + continue + decoder_variants.append(variant) + +encoder_variants = [] +for variant in encoding_variants: + if variant in [u"replacement", u"GBK", u"UTF-16"]: + continue + encoder_variants.append(variant) + +for variant in decoder_variants: + variant_file.write("use %s::*;\n" % to_snake_name(variant)) + +variant_file.write('''use super::*; + +pub enum VariantDecoder { +''') + +for variant in decoder_variants: + variant_file.write(" %s(%sDecoder),\n" % (to_camel_name(variant), to_camel_name(variant))) + +variant_file.write('''} + +impl VariantDecoder { +''') + +def write_variant_method(name, mut, arg_list, ret, variants, excludes, kind): + variant_file.write('''pub fn %s(&''' % name) + if mut: + variant_file.write('''mut ''') + variant_file.write('''self''') + for arg in arg_list: + variant_file.write(''', %s: %s''' % (arg[0], arg[1])) + variant_file.write(''')''') + if ret: + variant_file.write(''' -> %s''' % ret) + variant_file.write(''' {\nmatch *self {\n''') + for variant in variants: + variant_file.write('''Variant%s::%s(ref ''' % (kind, to_camel_name(variant))) + if mut: + variant_file.write('''mut ''') + if variant in excludes: + variant_file.write('''v) => (),''') + continue + variant_file.write('''v) => v.%s(''' % name) + first = True + for arg in arg_list: + if not first: + variant_file.write(''', ''') + first = False + variant_file.write(arg[0]) + variant_file.write('''),\n''') + variant_file.write('''}\n}\n\n''') + +write_variant_method("max_utf16_buffer_length", False, [("byte_length", "usize")], "Option<usize>", decoder_variants, [], "Decoder") + +write_variant_method("max_utf8_buffer_length_without_replacement", False, [("byte_length", "usize")], "Option<usize>", decoder_variants, [], "Decoder") + +write_variant_method("max_utf8_buffer_length", False, [("byte_length", "usize")], "Option<usize>", decoder_variants, [], "Decoder") + +write_variant_method("decode_to_utf16_raw", True, [("src", "&[u8]"), + ("dst", "&mut [u16]"), + ("last", "bool")], "(DecoderResult, usize, usize)", decoder_variants, [], "Decoder") + +write_variant_method("decode_to_utf8_raw", True, [("src", "&[u8]"), + ("dst", "&mut [u8]"), + ("last", "bool")], "(DecoderResult, usize, usize)", decoder_variants, [], "Decoder") + +variant_file.write(''' + + pub fn latin1_byte_compatible_up_to(&self, buffer: &[u8]) -> Option<usize> { + match *self { + VariantDecoder::SingleByte(ref v) => { + return Some(v.latin1_byte_compatible_up_to(buffer)); + } + VariantDecoder::Utf8(ref v) => { + if !v.in_neutral_state() { + return None; + } + } + VariantDecoder::Gb18030(ref v) => { + if !v.in_neutral_state() { + return None; + } + } + VariantDecoder::Big5(ref v) => { + if !v.in_neutral_state() { + return None; + } + } + VariantDecoder::EucJp(ref v) => { + if !v.in_neutral_state() { + return None; + } + } + VariantDecoder::Iso2022Jp(ref v) => { + if v.in_neutral_state() { + return Some(Encoding::iso_2022_jp_ascii_valid_up_to(buffer)); + } + return None; + } + VariantDecoder::ShiftJis(ref v) => { + if !v.in_neutral_state() { + return None; + } + } + VariantDecoder::EucKr(ref v) => { + if !v.in_neutral_state() { + return None; + } + } + VariantDecoder::UserDefined(_) => {} + VariantDecoder::Replacement(_) | VariantDecoder::Utf16(_) => { + return None; + } + }; + Some(Encoding::ascii_valid_up_to(buffer)) + } +} + +pub enum VariantEncoder { +''') + +for variant in encoder_variants: + variant_file.write(" %s(%sEncoder),\n" % (to_camel_name(variant), to_camel_name(variant))) + +variant_file.write('''} + +impl VariantEncoder { + pub fn has_pending_state(&self) -> bool { + match *self { + VariantEncoder::Iso2022Jp(ref v) => { + v.has_pending_state() + } + _ => false, + } + } +''') + +write_variant_method("max_buffer_length_from_utf16_without_replacement", False, [("u16_length", "usize")], "Option<usize>", encoder_variants, [], "Encoder") + +write_variant_method("max_buffer_length_from_utf8_without_replacement", False, [("byte_length", "usize")], "Option<usize>", encoder_variants, [], "Encoder") + +write_variant_method("encode_from_utf16_raw", True, [("src", "&[u16]"), + ("dst", "&mut [u8]"), + ("last", "bool")], "(EncoderResult, usize, usize)", encoder_variants, [], "Encoder") + +write_variant_method("encode_from_utf8_raw", True, [("src", "&str"), + ("dst", "&mut [u8]"), + ("last", "bool")], "(EncoderResult, usize, usize)", encoder_variants, [], "Encoder") + + +variant_file.write('''} + +pub enum VariantEncoding { + SingleByte(&'static [u16; 128], u16, u8, u8),''') + +for encoding in multi_byte: + variant_file.write("%s,\n" % to_camel_name(encoding["name"])) + +variant_file.write('''} + +impl VariantEncoding { + pub fn new_variant_decoder(&self) -> VariantDecoder { + match *self { + VariantEncoding::SingleByte(table, _, _, _) => SingleByteDecoder::new(table), + VariantEncoding::Utf8 => Utf8Decoder::new(), + VariantEncoding::Gbk | VariantEncoding::Gb18030 => Gb18030Decoder::new(), + VariantEncoding::Big5 => Big5Decoder::new(), + VariantEncoding::EucJp => EucJpDecoder::new(), + VariantEncoding::Iso2022Jp => Iso2022JpDecoder::new(), + VariantEncoding::ShiftJis => ShiftJisDecoder::new(), + VariantEncoding::EucKr => EucKrDecoder::new(), + VariantEncoding::Replacement => ReplacementDecoder::new(), + VariantEncoding::UserDefined => UserDefinedDecoder::new(), + VariantEncoding::Utf16Be => Utf16Decoder::new(true), + VariantEncoding::Utf16Le => Utf16Decoder::new(false), + } + } + + pub fn new_encoder(&self, encoding: &'static Encoding) -> Encoder { + match *self { + VariantEncoding::SingleByte(table, run_bmp_offset, run_byte_offset, run_length) => SingleByteEncoder::new(encoding, table, run_bmp_offset, run_byte_offset, run_length), + VariantEncoding::Utf8 => Utf8Encoder::new(encoding), + VariantEncoding::Gbk => Gb18030Encoder::new(encoding, false), + VariantEncoding::Gb18030 => Gb18030Encoder::new(encoding, true), + VariantEncoding::Big5 => Big5Encoder::new(encoding), + VariantEncoding::EucJp => EucJpEncoder::new(encoding), + VariantEncoding::Iso2022Jp => Iso2022JpEncoder::new(encoding), + VariantEncoding::ShiftJis => ShiftJisEncoder::new(encoding), + VariantEncoding::EucKr => EucKrEncoder::new(encoding), + VariantEncoding::UserDefined => UserDefinedEncoder::new(encoding), + VariantEncoding::Utf16Be | VariantEncoding::Replacement | + VariantEncoding::Utf16Le => unreachable!(), + } + } + + pub fn is_single_byte(&self) -> bool { + match *self { + VariantEncoding::SingleByte(_, _, _, _) | VariantEncoding::UserDefined => true, + _ => false, + } + } +} +''') + +variant_file.close() + +(ffi_rs_begin, ffi_rs_end) = read_non_generated("../encoding_c/src/lib.rs") + +ffi_file = open("../encoding_c/src/lib.rs", "w") + +ffi_file.write(ffi_rs_begin) +ffi_file.write(""" +// Instead, please regenerate using generate-encoding-data.py + +/// The minimum length of buffers that may be passed to `encoding_name()`. +pub const ENCODING_NAME_MAX_LENGTH: usize = %d; // %s + +""" % (longest_name_length, longest_name)) + +for name in preferred: + ffi_file.write('''/// The %s encoding. +#[no_mangle] +pub static %s_ENCODING: ConstEncoding = ConstEncoding(&%s_INIT); + +''' % (to_dom_name(name), to_constant_name(name), to_constant_name(name))) + +ffi_file.write(ffi_rs_end) +ffi_file.close() + +(single_byte_rs_begin, single_byte_rs_end) = read_non_generated("src/single_byte.rs") + +single_byte_file = open("src/single_byte.rs", "w") + +single_byte_file.write(single_byte_rs_begin) +single_byte_file.write(""" +// Instead, please regenerate using generate-encoding-data.py + + #[test] + fn test_single_byte_decode() {""") + +idx = 0 # for Miri, return after 2nd test +for name in preferred: + if name == u"ISO-8859-8-I": + continue; + if is_single_byte(name): + single_byte_file.write(""" + decode_single_byte(%s, &data::SINGLE_BYTE_DATA.%s);""" % (to_constant_name(name), to_snake_name(name))) + idx += 1 + if idx == 2: + single_byte_file.write(""" + if cfg!(miri) { + // Miri is too slow + return; + }""") + +single_byte_file.write(""" + } + + #[test] + fn test_single_byte_encode() {""") + + +idx = 0 # for Miri, return after 2nd test +for name in preferred: + if name == u"ISO-8859-8-I": + continue; + if is_single_byte(name): + single_byte_file.write(""" + encode_single_byte(%s, &data::SINGLE_BYTE_DATA.%s);""" % (to_constant_name(name), to_snake_name(name))) + idx += 1 + if idx == 2: + single_byte_file.write(""" + if cfg!(miri) { + // Miri is too slow + return; + }""") + + +single_byte_file.write(""" + } +""") + +single_byte_file.write(single_byte_rs_end) +single_byte_file.close() + +static_file = open("../encoding_c/include/encoding_rs_statics.h", "w") + +static_file.write("""// Copyright Mozilla Foundation. See the COPYRIGHT +// file at the top-level directory of this distribution. +// +// Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or +// https://www.apache.org/licenses/LICENSE-2.0> or the MIT license +// <LICENSE-MIT or https://opensource.org/licenses/MIT>, at your +// option. This file may not be copied, modified, or distributed +// except according to those terms. + +// THIS IS A GENERATED FILE. PLEASE DO NOT EDIT. +// Instead, please regenerate using generate-encoding-data.py + +// This file is not meant to be included directly. Instead, encoding_rs.h +// includes this file. + +#ifndef encoding_rs_statics_h_ +#define encoding_rs_statics_h_ + +#ifndef ENCODING_RS_ENCODING +#define ENCODING_RS_ENCODING Encoding +#ifndef __cplusplus +typedef struct Encoding_ Encoding; +#endif +#endif + +#ifndef ENCODING_RS_NOT_NULL_CONST_ENCODING_PTR +#define ENCODING_RS_NOT_NULL_CONST_ENCODING_PTR const ENCODING_RS_ENCODING* +#endif + +#ifndef ENCODING_RS_ENCODER +#define ENCODING_RS_ENCODER Encoder +#ifndef __cplusplus +typedef struct Encoder_ Encoder; +#endif +#endif + +#ifndef ENCODING_RS_DECODER +#define ENCODING_RS_DECODER Decoder +#ifndef __cplusplus +typedef struct Decoder_ Decoder; +#endif +#endif + +#define INPUT_EMPTY 0 + +#define OUTPUT_FULL 0xFFFFFFFF + +// %s +#define ENCODING_NAME_MAX_LENGTH %d + +""" % (longest_name, longest_name_length)) + +for name in preferred: + static_file.write('''/// The %s encoding. +extern ENCODING_RS_NOT_NULL_CONST_ENCODING_PTR const %s_ENCODING; + +''' % (to_dom_name(name), to_constant_name(name))) + +static_file.write("""#endif // encoding_rs_statics_h_ +""") +static_file.close() + +(utf_8_rs_begin, utf_8_rs_end) = read_non_generated("src/utf_8.rs") + +utf_8_file = open("src/utf_8.rs", "w") + +utf_8_file.write(utf_8_rs_begin) +utf_8_file.write(""" +// Instead, please regenerate using generate-encoding-data.py + +pub static UTF8_DATA: Utf8Data = Utf8Data { + table: [ +""") + +for i in range(256): + combined = (1 << 2) # invalid lead + if i < 0x80 or i > 0xBF: + combined |= (1 << 3) # normal trail + if i < 0xA0 or i > 0xBF: + combined |= (1 << 4) # three-byte special lower bound + if i < 0x80 or i > 0x9F: + combined |= (1 << 5) # three-byte special upper bound + if i < 0x90 or i > 0xBF: + combined |= (1 << 6) # four-byte special lower bound + if i < 0x80 or i > 0x8F: + combined |= (1 << 7) # four-byte special upper bound + utf_8_file.write("%d," % combined) + +for i in range(128, 256): + lane = (1 << 2) # invalid lead + if i >= 0xC2 and i <= 0xDF: + lane = (1 << 3) # normal trail + elif i == 0xE0: + lane = (1 << 4) # three-byte special lower bound + elif i >= 0xE1 and i <= 0xEC: + lane = (1 << 3) # normal trail + elif i == 0xED: + lane = (1 << 5) # three-byte special upper bound + elif i >= 0xEE and i <= 0xEF: + lane = (1 << 3) # normal trail + elif i == 0xF0: + lane = (1 << 6) # four-byte special lower bound + elif i >= 0xF1 and i <= 0xF3: + lane = (1 << 3) # normal trail + elif i == 0xF4: + lane = (1 << 7) # four-byte special upper bound + utf_8_file.write("%d," % lane) + +utf_8_file.write(""" + ], +}; + +""") + +utf_8_file.write(utf_8_rs_end) +utf_8_file.close() + +# Unit tests + +TEST_HEADER = '''Any copyright to the test code below this comment is dedicated to the +Public Domain. http://creativecommons.org/publicdomain/zero/1.0/ + +This is a generated file. Please do not edit. +Instead, please regenerate using generate-encoding-data.py +''' + +index = indexes["jis0208"] + +jis0208_in_file = open("src/test_data/jis0208_in.txt", "w") +jis0208_in_file.write(TEST_HEADER) +for pointer in range(0, 94 * 94): + (lead, trail) = divmod(pointer, 94) + lead += 0xA1 + trail += 0xA1 + jis0208_in_file.write("%s%s\n" % (chr(lead), chr(trail))) +jis0208_in_file.close() + +jis0208_in_ref_file = open("src/test_data/jis0208_in_ref.txt", "w") +jis0208_in_ref_file.write(TEST_HEADER) +for pointer in range(0, 94 * 94): + code_point = index[pointer] + if code_point: + jis0208_in_ref_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) + else: + jis0208_in_ref_file.write(u"\uFFFD\n".encode("utf-8")) +jis0208_in_ref_file.close() + +jis0208_out_file = open("src/test_data/jis0208_out.txt", "w") +jis0208_out_ref_file = open("src/test_data/jis0208_out_ref.txt", "w") +jis0208_out_file.write(TEST_HEADER) +jis0208_out_ref_file.write(TEST_HEADER) +for pointer in range(0, 94 * 94): + code_point = index[pointer] + if code_point: + revised_pointer = pointer + if revised_pointer == 8644 or (revised_pointer >= 1207 and revised_pointer < 1220): + revised_pointer = index.index(code_point) + (lead, trail) = divmod(revised_pointer, 94) + lead += 0xA1 + trail += 0xA1 + jis0208_out_ref_file.write("%s%s\n" % (chr(lead), chr(trail))) + jis0208_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) +jis0208_out_file.close() +jis0208_out_ref_file.close() + +shift_jis_in_file = open("src/test_data/shift_jis_in.txt", "w") +shift_jis_in_file.write(TEST_HEADER) +for pointer in range(0, len(index)): + (lead, trail) = divmod(pointer, 188) + lead += 0x81 if lead < 0x1F else 0xC1 + trail += 0x40 if trail < 0x3F else 0x41 + shift_jis_in_file.write("%s%s\n" % (chr(lead), chr(trail))) +shift_jis_in_file.close() + +shift_jis_in_ref_file = open("src/test_data/shift_jis_in_ref.txt", "w") +shift_jis_in_ref_file.write(TEST_HEADER) +for pointer in range(0, len(index)): + code_point = 0xE000 - 8836 + pointer if pointer >= 8836 and pointer <= 10715 else index[pointer] + if code_point: + shift_jis_in_ref_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) + else: + trail = pointer % 188 + trail += 0x40 if trail < 0x3F else 0x41 + if trail < 0x80: + shift_jis_in_ref_file.write((u"\uFFFD%s\n" % unichr(trail)).encode("utf-8")) + else: + shift_jis_in_ref_file.write(u"\uFFFD\n".encode("utf-8")) +shift_jis_in_ref_file.close() + +shift_jis_out_file = open("src/test_data/shift_jis_out.txt", "w") +shift_jis_out_ref_file = open("src/test_data/shift_jis_out_ref.txt", "w") +shift_jis_out_file.write(TEST_HEADER) +shift_jis_out_ref_file.write(TEST_HEADER) +for pointer in range(0, 8272): + code_point = index[pointer] + if code_point: + revised_pointer = pointer + if revised_pointer >= 1207 and revised_pointer < 1220: + revised_pointer = index.index(code_point) + (lead, trail) = divmod(revised_pointer, 188) + lead += 0x81 if lead < 0x1F else 0xC1 + trail += 0x40 if trail < 0x3F else 0x41 + shift_jis_out_ref_file.write("%s%s\n" % (chr(lead), chr(trail))) + shift_jis_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) +for pointer in range(8836, len(index)): + code_point = index[pointer] + if code_point: + revised_pointer = index.index(code_point) + if revised_pointer >= 8272 and revised_pointer < 8836: + revised_pointer = pointer + (lead, trail) = divmod(revised_pointer, 188) + lead += 0x81 if lead < 0x1F else 0xC1 + trail += 0x40 if trail < 0x3F else 0x41 + shift_jis_out_ref_file.write("%s%s\n" % (chr(lead), chr(trail))) + shift_jis_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) +shift_jis_out_file.close() +shift_jis_out_ref_file.close() + +iso_2022_jp_in_file = open("src/test_data/iso_2022_jp_in.txt", "w") +iso_2022_jp_in_file.write(TEST_HEADER) +for pointer in range(0, 94 * 94): + (lead, trail) = divmod(pointer, 94) + lead += 0x21 + trail += 0x21 + iso_2022_jp_in_file.write("\x1B$B%s%s\x1B(B\n" % (chr(lead), chr(trail))) +iso_2022_jp_in_file.close() + +iso_2022_jp_in_ref_file = open("src/test_data/iso_2022_jp_in_ref.txt", "w") +iso_2022_jp_in_ref_file.write(TEST_HEADER) +for pointer in range(0, 94 * 94): + code_point = index[pointer] + if code_point: + iso_2022_jp_in_ref_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) + else: + iso_2022_jp_in_ref_file.write(u"\uFFFD\n".encode("utf-8")) +iso_2022_jp_in_ref_file.close() + +iso_2022_jp_out_file = open("src/test_data/iso_2022_jp_out.txt", "w") +iso_2022_jp_out_ref_file = open("src/test_data/iso_2022_jp_out_ref.txt", "w") +iso_2022_jp_out_file.write(TEST_HEADER) +iso_2022_jp_out_ref_file.write(TEST_HEADER) +for pointer in range(0, 94 * 94): + code_point = index[pointer] + if code_point: + revised_pointer = pointer + if revised_pointer == 8644 or (revised_pointer >= 1207 and revised_pointer < 1220): + revised_pointer = index.index(code_point) + (lead, trail) = divmod(revised_pointer, 94) + lead += 0x21 + trail += 0x21 + iso_2022_jp_out_ref_file.write("\x1B$B%s%s\x1B(B\n" % (chr(lead), chr(trail))) + iso_2022_jp_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) +for i in xrange(len(half_width_index)): + code_point = i + 0xFF61 + normalized_code_point = half_width_index[i] + pointer = index.index(normalized_code_point) + (lead, trail) = divmod(pointer, 94) + lead += 0x21 + trail += 0x21 + iso_2022_jp_out_ref_file.write("\x1B$B%s%s\x1B(B\n" % (chr(lead), chr(trail))) + iso_2022_jp_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) +iso_2022_jp_out_file.close() +iso_2022_jp_out_ref_file.close() + +index = indexes["euc-kr"] + +euc_kr_in_file = open("src/test_data/euc_kr_in.txt", "w") +euc_kr_in_file.write(TEST_HEADER) +for pointer in range(0, len(index)): + (lead, trail) = divmod(pointer, 190) + lead += 0x81 + trail += 0x41 + euc_kr_in_file.write("%s%s\n" % (chr(lead), chr(trail))) +euc_kr_in_file.close() + +euc_kr_in_ref_file = open("src/test_data/euc_kr_in_ref.txt", "w") +euc_kr_in_ref_file.write(TEST_HEADER) +for pointer in range(0, len(index)): + code_point = index[pointer] + if code_point: + euc_kr_in_ref_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) + else: + trail = pointer % 190 + trail += 0x41 + if trail < 0x80: + euc_kr_in_ref_file.write((u"\uFFFD%s\n" % unichr(trail)).encode("utf-8")) + else: + euc_kr_in_ref_file.write(u"\uFFFD\n".encode("utf-8")) +euc_kr_in_ref_file.close() + +euc_kr_out_file = open("src/test_data/euc_kr_out.txt", "w") +euc_kr_out_ref_file = open("src/test_data/euc_kr_out_ref.txt", "w") +euc_kr_out_file.write(TEST_HEADER) +euc_kr_out_ref_file.write(TEST_HEADER) +for pointer in range(0, len(index)): + code_point = index[pointer] + if code_point: + (lead, trail) = divmod(pointer, 190) + lead += 0x81 + trail += 0x41 + euc_kr_out_ref_file.write("%s%s\n" % (chr(lead), chr(trail))) + euc_kr_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) +euc_kr_out_file.close() +euc_kr_out_ref_file.close() + +index = indexes["gb18030"] + +gb18030_in_file = open("src/test_data/gb18030_in.txt", "w") +gb18030_in_file.write(TEST_HEADER) +for pointer in range(0, len(index)): + (lead, trail) = divmod(pointer, 190) + lead += 0x81 + trail += 0x40 if trail < 0x3F else 0x41 + gb18030_in_file.write("%s%s\n" % (chr(lead), chr(trail))) +gb18030_in_file.close() + +gb18030_in_ref_file = open("src/test_data/gb18030_in_ref.txt", "w") +gb18030_in_ref_file.write(TEST_HEADER) +for pointer in range(0, len(index)): + code_point = index[pointer] + if code_point: + gb18030_in_ref_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) + else: + trail = pointer % 190 + trail += 0x40 if trail < 0x3F else 0x41 + if trail < 0x80: + gb18030_in_ref_file.write((u"\uFFFD%s\n" % unichr(trail)).encode("utf-8")) + else: + gb18030_in_ref_file.write(u"\uFFFD\n".encode("utf-8")) +gb18030_in_ref_file.close() + +gb18030_out_file = open("src/test_data/gb18030_out.txt", "w") +gb18030_out_ref_file = open("src/test_data/gb18030_out_ref.txt", "w") +gb18030_out_file.write(TEST_HEADER) +gb18030_out_ref_file.write(TEST_HEADER) +for pointer in range(0, len(index)): + if pointer == 6555: + continue + code_point = index[pointer] + if code_point: + (lead, trail) = divmod(pointer, 190) + lead += 0x81 + trail += 0x40 if trail < 0x3F else 0x41 + gb18030_out_ref_file.write("%s%s\n" % (chr(lead), chr(trail))) + gb18030_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) +gb18030_out_file.close() +gb18030_out_ref_file.close() + +index = indexes["big5"] + +big5_in_file = open("src/test_data/big5_in.txt", "w") +big5_in_file.write(TEST_HEADER) +for pointer in range(0, len(index)): + (lead, trail) = divmod(pointer, 157) + lead += 0x81 + trail += 0x40 if trail < 0x3F else 0x62 + big5_in_file.write("%s%s\n" % (chr(lead), chr(trail))) +big5_in_file.close() + +big5_two_characters = { + 1133: u"\u00CA\u0304", + 1135: u"\u00CA\u030C", + 1164: u"\u00EA\u0304", + 1166: u"\u00EA\u030C", +} + +big5_in_ref_file = open("src/test_data/big5_in_ref.txt", "w") +big5_in_ref_file.write(TEST_HEADER) +for pointer in range(0, len(index)): + if pointer in big5_two_characters.keys(): + big5_in_ref_file.write((u"%s\n" % big5_two_characters[pointer]).encode("utf-8")) + continue + code_point = index[pointer] + if code_point: + big5_in_ref_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) + else: + trail = pointer % 157 + trail += 0x40 if trail < 0x3F else 0x62 + if trail < 0x80: + big5_in_ref_file.write((u"\uFFFD%s\n" % unichr(trail)).encode("utf-8")) + else: + big5_in_ref_file.write(u"\uFFFD\n".encode("utf-8")) +big5_in_ref_file.close() + +prefer_last = [ + 0x2550, + 0x255E, + 0x2561, + 0x256A, + 0x5341, + 0x5345, +] + +pointer_for_prefer_last = [] + +for code_point in prefer_last: + # Python lists don't have .rindex() :-( + for i in xrange(len(index) - 1, -1, -1): + candidate = index[i] + if candidate == code_point: + pointer_for_prefer_last.append(i) + break + +big5_out_file = open("src/test_data/big5_out.txt", "w") +big5_out_ref_file = open("src/test_data/big5_out_ref.txt", "w") +big5_out_file.write(TEST_HEADER) +big5_out_ref_file.write(TEST_HEADER) +for pointer in range(((0xA1 - 0x81) * 157), len(index)): + code_point = index[pointer] + if code_point: + if code_point in prefer_last: + if pointer != pointer_for_prefer_last[prefer_last.index(code_point)]: + continue + else: + if pointer != index.index(code_point): + continue + (lead, trail) = divmod(pointer, 157) + lead += 0x81 + trail += 0x40 if trail < 0x3F else 0x62 + big5_out_ref_file.write("%s%s\n" % (chr(lead), chr(trail))) + big5_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) +big5_out_file.close() +big5_out_ref_file.close() + +index = indexes["jis0212"] + +jis0212_in_file = open("src/test_data/jis0212_in.txt", "w") +jis0212_in_file.write(TEST_HEADER) +for pointer in range(0, len(index)): + (lead, trail) = divmod(pointer, 94) + lead += 0xA1 + trail += 0xA1 + jis0212_in_file.write("\x8F%s%s\n" % (chr(lead), chr(trail))) +jis0212_in_file.close() + +jis0212_in_ref_file = open("src/test_data/jis0212_in_ref.txt", "w") +jis0212_in_ref_file.write(TEST_HEADER) +for pointer in range(0, len(index)): + code_point = index[pointer] + if code_point: + jis0212_in_ref_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) + else: + jis0212_in_ref_file.write(u"\uFFFD\n".encode("utf-8")) +jis0212_in_ref_file.close() + +(codepage_begin, codepage_end) = read_non_generated("../codepage/src/lib.rs") + +codepage_file = open("../codepage/src/lib.rs", "w") + +codepage_file.write(codepage_begin) +codepage_file.write(""" +// Instead, please regenerate using generate-encoding-data.py + +/// Supported code page numbers in estimated order of usage frequency +static CODE_PAGES: [u16; %d] = [ +""" % len(code_pages)) + +for code_page in code_pages: + codepage_file.write(" %d,\n" % code_page) + +codepage_file.write("""]; + +/// Encodings corresponding to the code page numbers in the same order +static ENCODINGS: [&'static Encoding; %d] = [ +""" % len(code_pages)) + +for code_page in code_pages: + name = encodings_by_code_page[code_page] + codepage_file.write(" &%s_INIT,\n" % to_constant_name(name)) + +codepage_file.write("""]; + +""") + +codepage_file.write(codepage_end) +codepage_file.close() + +(codepage_test_begin, codepage_test_end) = read_non_generated("../codepage/src/tests.rs") + +codepage_test_file = open("../codepage/src/tests.rs", "w") + +codepage_test_file.write(codepage_test_begin) +codepage_test_file.write(""" +// Instead, please regenerate using generate-encoding-data.py + +#[test] +fn test_to_encoding() { + assert_eq!(to_encoding(0), None); + +""") + +for code_page in code_pages: + codepage_test_file.write(" assert_eq!(to_encoding(%d), Some(%s));\n" % (code_page, to_constant_name(encodings_by_code_page[code_page]))) + +codepage_test_file.write("""} + +#[test] +fn test_from_encoding() { +""") + +for name in preferred: + if code_pages_by_encoding.has_key(name): + codepage_test_file.write(" assert_eq!(from_encoding(%s), Some(%d));\n" % (to_constant_name(name), code_pages_by_encoding[name])) + else: + codepage_test_file.write(" assert_eq!(from_encoding(%s), None);\n" % to_constant_name(name)) + +codepage_test_file.write("""} +""") + +codepage_test_file.write(codepage_test_end) +codepage_test_file.close() + +subprocess.call(["cargo", "fmt"]) |