#!/usr/bin/python # Copyright 2013-2016 Mozilla Foundation. See the COPYRIGHT # file at the top-level directory of this distribution. # # Licensed under the Apache License, Version 2.0 or the MIT license # , at your # option. This file may not be copied, modified, or distributed # except according to those terms. import json import subprocess import sys import os.path if (not os.path.isfile("../encoding/encodings.json")) or (not os.path.isfile("../encoding/indexes.json")): sys.stderr.write("This script needs a clone of https://github.com/whatwg/encoding/ (preferably at revision f381389) next to the encoding_rs directory.\n"); sys.exit(-1) if not os.path.isfile("../encoding_c/src/lib.rs"): sys.stderr.write("This script also writes the generated parts of the encoding_c crate and needs a clone of https://github.com/hsivonen/encoding_c next to the encoding_rs directory.\n"); sys.exit(-1) if not os.path.isfile("../codepage/src/lib.rs"): sys.stderr.write("This script also writes the generated parts of the codepage crate and needs a clone of https://github.com/hsivonen/codepage next to the encoding_rs directory.\n"); sys.exit(-1) def cmp_from_end(one, other): c = cmp(len(one), len(other)) if c != 0: return c i = len(one) - 1 while i >= 0: c = cmp(one[i], other[i]) if c != 0: return c i -= 1 return 0 class Label: def __init__(self, label, preferred): self.label = label self.preferred = preferred def __cmp__(self, other): return cmp_from_end(self.label, other.label) class CodePage: def __init__(self, code_page, preferred): self.code_page = code_page self.preferred = preferred def __cmp__(self, other): return self.code_page, other.code_page def static_u16_table(name, data): data_file.write('''pub static %s: [u16; %d] = [ ''' % (name, len(data))) for i in xrange(len(data)): data_file.write('0x%04X,\n' % data[i]) data_file.write(''']; ''') def static_u16_table_from_indexable(name, data, item, feature): data_file.write('''#[cfg(all( feature = "less-slow-%s", not(feature = "fast-%s") ))] static %s: [u16; %d] = [ ''' % (feature, feature, name, len(data))) for i in xrange(len(data)): data_file.write('0x%04X,\n' % data[i][item]) data_file.write(''']; ''') def static_u8_pair_table_from_indexable(name, data, item, feature): data_file.write('''#[cfg(all( feature = "less-slow-%s", not(feature = "fast-%s") ))] static %s: [[u8; 2]; %d] = [ ''' % (feature, feature, name, len(data))) for i in xrange(len(data)): data_file.write('[0x%02X, 0x%02X],\n' % data[i][item]) data_file.write(''']; ''') def static_u8_pair_table(name, data, feature): data_file.write('''#[cfg(feature = "%s")] static %s: [[u8; 2]; %d] = [ ''' % (feature, name, len(data))) for i in xrange(len(data)): pair = data[i] if not pair: pair = (0, 0) data_file.write('[0x%02X, 0x%02X],\n' % pair) data_file.write(''']; ''') preferred = [] dom = [] labels = [] data = json.load(open("../encoding/encodings.json", "r")) indexes = json.load(open("../encoding/indexes.json", "r")) single_byte = [] multi_byte = [] def to_camel_name(name): if name == u"iso-8859-8-i": return u"Iso8I" if name.startswith(u"iso-8859-"): return name.replace(u"iso-8859-", u"Iso") return name.title().replace(u"X-", u"").replace(u"-", u"").replace(u"_", u"") def to_constant_name(name): return name.replace(u"-", u"_").upper() def to_snake_name(name): return name.replace(u"-", u"_").lower() def to_dom_name(name): return name # Guestimate based on # https://w3techs.com/technologies/overview/character_encoding/all # whose methodology is known to be bogus, but the results are credible for # this purpose. UTF-16LE lifted up due to prevalence on Windows and # "ANSI codepages" prioritized. encodings_by_code_page_frequency = [ "UTF-8", "UTF-16LE", "windows-1252", "windows-1251", "GBK", "Shift_JIS", "EUC-KR", "windows-1250", "windows-1256", "windows-1254", "Big5", "windows-874", "windows-1255", "windows-1253", "windows-1257", "windows-1258", "EUC-JP", "ISO-8859-2", "ISO-8859-15", "ISO-8859-7", "KOI8-R", "gb18030", "ISO-8859-5", "ISO-8859-8-I", "ISO-8859-4", "ISO-8859-6", "ISO-2022-JP", "KOI8-U", "ISO-8859-13", "ISO-8859-3", "UTF-16BE", "IBM866", "ISO-8859-10", "ISO-8859-8", "macintosh", "x-mac-cyrillic", "ISO-8859-14", "ISO-8859-16", ] encodings_by_code_page = { 932: "Shift_JIS", 936: "GBK", 949: "EUC-KR", 950: "Big5", 866: "IBM866", 874: "windows-874", 1200: "UTF-16LE", 1201: "UTF-16BE", 1250: "windows-1250", 1251: "windows-1251", 1252: "windows-1252", 1253: "windows-1253", 1254: "windows-1254", 1255: "windows-1255", 1256: "windows-1256", 1257: "windows-1257", 1258: "windows-1258", 10000: "macintosh", 10017: "x-mac-cyrillic", 20866: "KOI8-R", 20932: "EUC-JP", 21866: "KOI8-U", 28592: "ISO-8859-2", 28593: "ISO-8859-3", 28594: "ISO-8859-4", 28595: "ISO-8859-5", 28596: "ISO-8859-6", 28597: "ISO-8859-7", 28598: "ISO-8859-8", 28600: "ISO-8859-10", 28603: "ISO-8859-13", 28604: "ISO-8859-14", 28605: "ISO-8859-15", 28606: "ISO-8859-16", 38598: "ISO-8859-8-I", 50221: "ISO-2022-JP", 54936: "gb18030", 65001: "UTF-8", } code_pages_by_encoding = {} for code_page, encoding in encodings_by_code_page.iteritems(): code_pages_by_encoding[encoding] = code_page encoding_by_alias_code_page = { 951: "Big5", 10007: "x-mac-cyrillic", 20936: "GBK", 20949: "EUC-KR", 21010: "UTF-16LE", # Undocumented; needed by calamine for Excel compat 28591: "windows-1252", 28599: "windows-1254", 28601: "windows-874", 50220: "ISO-2022-JP", 50222: "ISO-2022-JP", 50225: "replacement", # ISO-2022-KR 50227: "replacement", # ISO-2022-CN 51949: "EUC-JP", 51936: "GBK", 51949: "EUC-KR", 52936: "replacement", # HZ } code_pages = [] for name in encodings_by_code_page_frequency: code_pages.append(code_pages_by_encoding[name]) encodings_by_code_page.update(encoding_by_alias_code_page) temp_keys = encodings_by_code_page.keys() temp_keys.sort() for code_page in temp_keys: if not code_page in code_pages: code_pages.append(code_page) # The position in the index (0 is the first index entry, # i.e. byte value 0x80) that starts the longest run of # consecutive code points. Must not be in the first # quadrant. If the character to be encoded is not in this # run, the part of the index after the run is searched # forward. Then the part of the index from 32 to the start # of the run. The first quadrant is searched last. # # If there is no obviously most useful longest run, # the index here is just used to affect the search order. start_of_longest_run_in_single_byte = { "IBM866": 96, # 0 would be longest, but we don't want to start in the first quadrant "windows-874": 33, "windows-1250": 92, "windows-1251": 64, "windows-1252": 32, "windows-1253": 83, "windows-1254": 95, "windows-1255": 96, "windows-1256": 65, "windows-1257": 95, # not actually longest "windows-1258": 95, # not actually longest "macintosh": 106, # useless "x-mac-cyrillic": 96, "KOI8-R": 64, # not actually longest "KOI8-U": 64, # not actually longest "ISO-8859-2": 95, # not actually longest "ISO-8859-3": 95, # not actually longest "ISO-8859-4": 95, # not actually longest "ISO-8859-5": 46, "ISO-8859-6": 65, "ISO-8859-7": 83, "ISO-8859-8": 96, "ISO-8859-10": 90, # not actually longest "ISO-8859-13": 95, # not actually longest "ISO-8859-14": 95, "ISO-8859-15": 63, "ISO-8859-16": 95, # not actually longest } # for group in data: if group["heading"] == "Legacy single-byte encodings": single_byte = group["encodings"] else: multi_byte.extend(group["encodings"]) for encoding in group["encodings"]: preferred.append(encoding["name"]) for label in encoding["labels"]: labels.append(Label(label, encoding["name"])) for name in preferred: dom.append(to_dom_name(name)) preferred.sort() labels.sort() dom.sort(cmp=cmp_from_end) longest_label_length = 0 longest_name_length = 0 longest_label = None longest_name = None for name in preferred: if len(name) > longest_name_length: longest_name_length = len(name) longest_name = name for label in labels: if len(label.label) > longest_label_length: longest_label_length = len(label.label) longest_label = label.label def longest_run_for_single_byte(name): if name == u"ISO-8859-8-I": name = u"ISO-8859-8" index = indexes[name.lower()] run_byte_offset = start_of_longest_run_in_single_byte[name] run_bmp_offset = index[run_byte_offset] previous_code_point = run_bmp_offset run_length = 1 while True: i = run_byte_offset + run_length if i == len(index): break code_point = index[i] if previous_code_point + 1 != code_point: break previous_code_point = code_point run_length += 1 return (run_bmp_offset, run_byte_offset, run_length) def is_single_byte(name): for encoding in single_byte: if name == encoding["name"]: return True return False def read_non_generated(path): partially_generated_file = open(path, "r") full = partially_generated_file.read() partially_generated_file.close() generated_begin = "// BEGIN GENERATED CODE. PLEASE DO NOT EDIT." generated_end = "// END GENERATED CODE" generated_begin_index = full.find(generated_begin) if generated_begin_index < 0: sys.stderr.write("Can't find generated code start marker in %s. Exiting.\n" % path) sys.exit(-1) generated_end_index = full.find(generated_end) if generated_end_index < 0: sys.stderr.write("Can't find generated code end marker in %s. Exiting.\n" % path) sys.exit(-1) return (full[0:generated_begin_index + len(generated_begin)], full[generated_end_index:]) (lib_rs_begin, lib_rs_end) = read_non_generated("src/lib.rs") label_file = open("src/lib.rs", "w") label_file.write(lib_rs_begin) label_file.write(""" // Instead, please regenerate using generate-encoding-data.py const LONGEST_LABEL_LENGTH: usize = %d; // %s """ % (longest_label_length, longest_label)) for name in preferred: variant = None if is_single_byte(name): (run_bmp_offset, run_byte_offset, run_length) = longest_run_for_single_byte(name) variant = "SingleByte(&data::SINGLE_BYTE_DATA.%s, 0x%04X, %d, %d)" % (to_snake_name(u"iso-8859-8" if name == u"ISO-8859-8-I" else name), run_bmp_offset, run_byte_offset, run_length) else: variant = to_camel_name(name) docfile = open("doc/%s.txt" % name, "r") doctext = docfile.read() docfile.close() label_file.write('''/// The initializer for the [%s](static.%s.html) encoding. /// /// For use only for taking the address of this form when /// Rust prohibits the use of the non-`_INIT` form directly, /// such as in initializers of other `static`s. If in doubt, /// use the corresponding non-`_INIT` reference-typed `static`. /// /// This part of the public API will go away if Rust changes /// to make the referent of `pub const FOO: &'static Encoding` /// unique cross-crate or if Rust starts allowing static arrays /// to be initialized with `pub static FOO: &'static Encoding` /// items. pub static %s_INIT: Encoding = Encoding { name: "%s", variant: VariantEncoding::%s, }; /// The %s encoding. /// %s/// /// This will change from `static` to `const` if Rust changes /// to make the referent of `pub const FOO: &'static Encoding` /// unique cross-crate, so don't take the address of this /// `static`. pub static %s: &'static Encoding = &%s_INIT; ''' % (to_dom_name(name), to_constant_name(name), to_constant_name(name), to_dom_name(name), variant, to_dom_name(name), doctext, to_constant_name(name), to_constant_name(name))) label_file.write("""static LABELS_SORTED: [&'static str; %d] = [ """ % len(labels)) for label in labels: label_file.write('''"%s",\n''' % label.label) label_file.write("""]; static ENCODINGS_IN_LABEL_SORT: [&'static Encoding; %d] = [ """ % len(labels)) for label in labels: label_file.write('''&%s_INIT,\n''' % to_constant_name(label.preferred)) label_file.write(''']; ''') label_file.write(lib_rs_end) label_file.close() label_test_file = open("src/test_labels_names.rs", "w") label_test_file.write('''// Any copyright to the test code below this comment is dedicated to the // Public Domain. http://creativecommons.org/publicdomain/zero/1.0/ // THIS IS A GENERATED FILE. PLEASE DO NOT EDIT. // Instead, please regenerate using generate-encoding-data.py use super::*; #[test] fn test_all_labels() { ''') for label in labels: label_test_file.write('''assert_eq!(Encoding::for_label(b"%s"), Some(%s));\n''' % (label.label, to_constant_name(label.preferred))) label_test_file.write('''} ''') label_test_file.close() def null_to_zero(code_point): if not code_point: code_point = 0 return code_point (data_rs_begin, data_rs_end) = read_non_generated("src/data.rs") data_file = open("src/data.rs", "w") data_file.write(data_rs_begin) data_file.write(''' // Instead, please regenerate using generate-encoding-data.py #[repr(align(64))] // Align to cache lines pub struct SingleByteData { ''') # Single-byte for encoding in single_byte: name = encoding["name"] if name == u"ISO-8859-8-I": continue data_file.write(''' pub %s: [u16; 128], ''' % to_snake_name(name)) data_file.write('''} pub static SINGLE_BYTE_DATA: SingleByteData = SingleByteData { ''') for encoding in single_byte: name = encoding["name"] if name == u"ISO-8859-8-I": continue data_file.write(''' %s: [ ''' % to_snake_name(name)) for code_point in indexes[name.lower()]: data_file.write('0x%04X,\n' % null_to_zero(code_point)) data_file.write('''], ''') data_file.write('''}; ''') # Big5 index = indexes["big5"] astralness = [] low_bits = [] for code_point in index[942:19782]: if code_point: astralness.append(1 if code_point > 0xFFFF else 0) low_bits.append(code_point & 0xFFFF) else: astralness.append(0) low_bits.append(0) # pad length to multiple of 32 for j in xrange(32 - (len(astralness) % 32)): astralness.append(0) data_file.write('''#[cfg_attr(feature = "cargo-clippy", allow(unreadable_literal))] static BIG5_ASTRALNESS: [u32; %d] = [ ''' % (len(astralness) / 32)) i = 0 while i < len(astralness): accu = 0 for j in xrange(32): accu |= astralness[i + j] << j data_file.write('0x%08X,\n' % accu) i += 32 data_file.write(''']; ''') static_u16_table("BIG5_LOW_BITS", low_bits) # Encoder table for Level 1 Hanzi # Note: If we were OK with doubling this table, we # could use a directly-indexable table instead... level1_hanzi_index = index[5495:10896] level1_hanzi_pairs = [] for i in xrange(len(level1_hanzi_index)): hanzi_lead = (i / 157) + 0xA4 hanzi_trail = (i % 157) hanzi_trail += 0x40 if hanzi_trail < 0x3F else 0x62 level1_hanzi_pairs.append((level1_hanzi_index[i], (hanzi_lead, hanzi_trail))) level1_hanzi_pairs.append((0x4E5A, (0xC8, 0x7B))) level1_hanzi_pairs.append((0x5202, (0xC8, 0x7D))) level1_hanzi_pairs.append((0x9FB0, (0xC8, 0xA1))) level1_hanzi_pairs.append((0x5188, (0xC8, 0xA2))) level1_hanzi_pairs.append((0x9FB1, (0xC8, 0xA3))) level1_hanzi_pairs.sort(key=lambda x: x[0]) static_u16_table_from_indexable("BIG5_LEVEL1_HANZI_CODE_POINTS", level1_hanzi_pairs, 0, "big5-hanzi-encode") static_u8_pair_table_from_indexable("BIG5_LEVEL1_HANZI_BYTES", level1_hanzi_pairs, 1, "big5-hanzi-encode") # Fast Unified Ideograph encode big5_unified_ideograph_bytes = [None] * (0x9FCC - 0x4E00) for row in xrange(0x7E - 0x20): for column in xrange(157): pointer = 5024 + column + (row * 157) code_point = index[pointer] if code_point and code_point >= 0x4E00 and code_point <= 0x9FCB: unified_offset = code_point - 0x4E00 unified_lead = 0xA1 + row unified_trail = (0x40 if column < 0x3F else 0x62) + column if code_point == 0x5341 or code_point == 0x5345 or not big5_unified_ideograph_bytes[unified_offset]: big5_unified_ideograph_bytes[unified_offset] = (unified_lead, unified_trail) static_u8_pair_table("BIG5_UNIFIED_IDEOGRAPH_BYTES", big5_unified_ideograph_bytes, "fast-big5-hanzi-encode") # JIS0208 index = indexes["jis0208"] # JIS 0208 Level 1 Kanji static_u16_table("JIS0208_LEVEL1_KANJI", index[1410:4375]) # JIS 0208 Level 2 Kanji and Additional Kanji static_u16_table("JIS0208_LEVEL2_AND_ADDITIONAL_KANJI", index[4418:7808]) # IBM Kanji static_u16_table("IBM_KANJI", index[8272:8632]) # Check that the other instance is the same if index[8272:8632] != index[10744:11104]: raise Error() # JIS 0208 symbols (all non-Kanji, non-range items) symbol_index = [] symbol_triples = [] pointers_to_scan = [ (0, 188), (658, 691), (1159, 1221), ] in_run = False run_start_pointer = 0 run_start_array_index = 0 for (start, end) in pointers_to_scan: for i in range(start, end): code_point = index[i] if in_run: if code_point: symbol_index.append(code_point) else: symbol_triples.append(run_start_pointer) symbol_triples.append(i - run_start_pointer) symbol_triples.append(run_start_array_index) in_run = False else: if code_point: in_run = True run_start_pointer = i run_start_array_index = len(symbol_index) symbol_index.append(code_point) if in_run: symbol_triples.append(run_start_pointer) symbol_triples.append(end - run_start_pointer) symbol_triples.append(run_start_array_index) in_run = False if in_run: raise Error() # Now add manually the two overlapping slices of # index from the NEC/IBM extensions. run_start_array_index = len(symbol_index) symbol_index.extend(index[10736:10744]) # Later symbol_triples.append(10736) symbol_triples.append(8) symbol_triples.append(run_start_array_index) # Earlier symbol_triples.append(8644) symbol_triples.append(4) symbol_triples.append(run_start_array_index) static_u16_table("JIS0208_SYMBOLS", symbol_index) static_u16_table("JIS0208_SYMBOL_TRIPLES", symbol_triples) # Write down the magic numbers needed when preferring the earlier case data_file.write('''const IBM_SYMBOL_START: usize = %d;''' % (run_start_array_index + 1)) data_file.write('''const IBM_SYMBOL_END: usize = %d;''' % (run_start_array_index + 4)) data_file.write('''const IBM_SYMBOL_POINTER_START: usize = %d;''' % 8645) # JIS 0208 ranges (excluding kana) range_triples = [] pointers_to_scan = [ (188, 281), (470, 657), (1128, 1159), (8634, 8644), (10716, 10736), ] in_run = False run_start_pointer = 0 run_start_code_point = 0 previous_code_point = 0 for (start, end) in pointers_to_scan: for i in range(start, end): code_point = index[i] if in_run: if code_point: if previous_code_point + 1 != code_point: range_triples.append(run_start_pointer) range_triples.append(i - run_start_pointer) range_triples.append(run_start_code_point) run_start_pointer = i run_start_code_point = code_point previous_code_point = code_point else: range_triples.append(run_start_pointer) range_triples.append(i - run_start_pointer) range_triples.append(run_start_code_point) run_start_pointer = 0 run_start_code_point = 0 previous_code_point = 0 in_run = False else: if code_point: in_run = True run_start_pointer = i run_start_code_point = code_point previous_code_point = code_point if in_run: range_triples.append(run_start_pointer) range_triples.append(end - run_start_pointer) range_triples.append(run_start_code_point) run_start_pointer = 0 run_start_code_point = 0 previous_code_point = 0 in_run = False if in_run: raise Error() static_u16_table("JIS0208_RANGE_TRIPLES", range_triples) # Encoder table for Level 1 Kanji # Note: If we were OK with 30 KB more footprint, we # could use a directly-indexable table instead... level1_kanji_index = index[1410:4375] level1_kanji_pairs = [] for i in xrange(len(level1_kanji_index)): pointer = 1410 + i (lead, trail) = divmod(pointer, 188) lead += 0x81 if lead < 0x1F else 0xC1 trail += 0x40 if trail < 0x3F else 0x41 level1_kanji_pairs.append((level1_kanji_index[i], (lead, trail))) level1_kanji_pairs.sort(key=lambda x: x[0]) static_u16_table_from_indexable("JIS0208_LEVEL1_KANJI_CODE_POINTS", level1_kanji_pairs, 0, "kanji-encode") static_u8_pair_table_from_indexable("JIS0208_LEVEL1_KANJI_SHIFT_JIS_BYTES", level1_kanji_pairs, 1, "kanji-encode") # Fast encoder table for Kanji kanji_bytes = [None] * (0x9FA1 - 0x4E00) for pointer in xrange(len(index)): code_point = index[pointer] if code_point and code_point >= 0x4E00 and code_point <= 0x9FA0: (lead, trail) = divmod(pointer, 188) lead += 0x81 if lead < 0x1F else 0xC1 trail += 0x40 if trail < 0x3F else 0x41 # unset the high bit of lead if IBM Kanji if pointer >= 8272: lead = lead & 0x7F kanji_bytes[code_point - 0x4E00] = (lead, trail) static_u8_pair_table("JIS0208_KANJI_BYTES", kanji_bytes, "fast-kanji-encode") # ISO-2022-JP half-width katakana # index is still jis0208 half_width_index = indexes["iso-2022-jp-katakana"] data_file.write('''pub static ISO_2022_JP_HALF_WIDTH_TRAIL: [u8; %d] = [ ''' % len(half_width_index)) for i in xrange(len(half_width_index)): code_point = half_width_index[i] pointer = index.index(code_point) trail = pointer % 94 + 0x21 data_file.write('0x%02X,\n' % trail) data_file.write(''']; ''') # EUC-KR index = indexes["euc-kr"] # Unicode 1.1 Hangul above the old KS X 1001 block # Compressed form takes 35% of uncompressed form pointers = [] offsets = [] previous_code_point = 0 for row in xrange(0x20): for column in xrange(190): i = column + (row * 190) # Skip the gaps if (column >= 0x1A and column < 0x20) or (column >= 0x3A and column < 0x40): continue code_point = index[i] if previous_code_point > code_point: raise Error() if code_point - previous_code_point != 1: adjustment = 0 if column >= 0x40: adjustment = 12 elif column >= 0x20: adjustment = 6 pointers.append(column - adjustment + (row * (190 - 12))) offsets.append(code_point) previous_code_point = code_point static_u16_table("CP949_TOP_HANGUL_POINTERS", pointers) static_u16_table("CP949_TOP_HANGUL_OFFSETS", offsets) # Unicode 1.1 Hangul to the left of the old KS X 1001 block pointers = [] offsets = [] previous_code_point = 0 for row in xrange(0x46 - 0x20): for column in xrange(190 - 94): i = 6080 + column + (row * 190) # Skip the gaps if (column >= 0x1A and column < 0x20) or (column >= 0x3A and column < 0x40): continue if i > 13127: # Exclude unassigned on partial last row break code_point = index[i] if previous_code_point > code_point: raise Error() if code_point - previous_code_point != 1: adjustment = 0 if column >= 0x40: adjustment = 12 elif column >= 0x20: adjustment = 6 pointers.append(column - adjustment + (row * (190 - 94 - 12))) offsets.append(code_point) previous_code_point = code_point static_u16_table("CP949_LEFT_HANGUL_POINTERS", pointers) static_u16_table("CP949_LEFT_HANGUL_OFFSETS", offsets) # KS X 1001 Hangul hangul_index = [] previous_code_point = 0 for row in xrange(0x48 - 0x2F): for column in xrange(94): code_point = index[9026 + column + (row * 190)] if previous_code_point >= code_point: raise Error() hangul_index.append(code_point) previous_code_point = code_point static_u16_table("KSX1001_HANGUL", hangul_index) # KS X 1001 Hanja hanja_index = [] for row in xrange(0x7D - 0x49): for column in xrange(94): hanja_index.append(index[13966 + column + (row * 190)]) static_u16_table("KSX1001_HANJA", hanja_index) # KS X 1001 symbols symbol_index = [] for i in range(6176, 6270): symbol_index.append(index[i]) for i in range(6366, 6437): symbol_index.append(index[i]) static_u16_table("KSX1001_SYMBOLS", symbol_index) # KS X 1001 Uppercase Latin subindex = [] for i in range(7506, 7521): subindex.append(null_to_zero(index[i])) static_u16_table("KSX1001_UPPERCASE", subindex) # KS X 1001 Lowercase Latin subindex = [] for i in range(7696, 7712): subindex.append(index[i]) static_u16_table("KSX1001_LOWERCASE", subindex) # KS X 1001 Box drawing subindex = [] for i in range(7126, 7194): subindex.append(index[i]) static_u16_table("KSX1001_BOX", subindex) # KS X 1001 other pointers = [] offsets = [] previous_code_point = 0 for row in xrange(10): for column in xrange(94): i = 6556 + column + (row * 190) code_point = index[i] # Exclude ranges that were processed as lookup tables # or that contain unmapped cells by filling them with # ASCII. Upon encode, ASCII code points will # never appear as the search key. if (i >= 6946 and i <= 6950): code_point = i - 6946 elif (i >= 6961 and i <= 6967): code_point = i - 6961 elif (i >= 6992 and i <= 6999): code_point = i - 6992 elif (i >= 7024 and i <= 7029): code_point = i - 7024 elif (i >= 7126 and i <= 7219): code_point = i - 7126 elif (i >= 7395 and i <= 7409): code_point = i - 7395 elif (i >= 7506 and i <= 7521): code_point = i - 7506 elif (i >= 7696 and i <= 7711): code_point = i - 7696 elif (i >= 7969 and i <= 7979): code_point = i - 7969 elif (i >= 8162 and i <= 8169): code_point = i - 8162 elif (i >= 8299 and i <= 8313): code_point = i - 8299 elif (i >= 8347 and i <= 8359): code_point = i - 8347 if code_point - previous_code_point != 1: pointers.append(column + (row * 94)) offsets.append(code_point) previous_code_point = code_point static_u16_table("KSX1001_OTHER_POINTERS", pointers) # Omit the last offset, because the end of the last line # is unmapped, so we don't want to look at it. static_u16_table("KSX1001_OTHER_UNSORTED_OFFSETS", offsets[:-1]) # Fast Hangul and Hanja encode hangul_bytes = [None] * (0xD7A4 - 0xAC00) hanja_unified_bytes = [None] * (0x9F9D - 0x4E00) hanja_compatibility_bytes = [None] * (0xFA0C - 0xF900) for row in xrange(0x7D): for column in xrange(190): pointer = column + (row * 190) code_point = index[pointer] if code_point: lead = 0x81 + row trail = 0x41 + column if code_point >= 0xAC00 and code_point < 0xD7A4: hangul_bytes[code_point - 0xAC00] = (lead, trail) elif code_point >= 0x4E00 and code_point < 0x9F9D: hanja_unified_bytes[code_point - 0x4E00] = (lead, trail) elif code_point >= 0xF900 and code_point < 0xFA0C: hanja_compatibility_bytes[code_point - 0xF900] = (lead, trail) static_u8_pair_table("CP949_HANGUL_BYTES", hangul_bytes, "fast-hangul-encode") static_u8_pair_table("KSX1001_UNIFIED_HANJA_BYTES", hanja_unified_bytes, "fast-hanja-encode") static_u8_pair_table("KSX1001_COMPATIBILITY_HANJA_BYTES", hanja_compatibility_bytes, "fast-hanja-encode") # JIS 0212 index = indexes["jis0212"] # JIS 0212 Kanji static_u16_table("JIS0212_KANJI", index[1410:7211]) # JIS 0212 accented (all non-Kanji, non-range items) symbol_index = [] symbol_triples = [] pointers_to_scan = [ (0, 596), (608, 644), (656, 1409), ] in_run = False run_start_pointer = 0 run_start_array_index = 0 for (start, end) in pointers_to_scan: for i in range(start, end): code_point = index[i] if in_run: if code_point: symbol_index.append(code_point) elif index[i + 1]: symbol_index.append(0) else: symbol_triples.append(run_start_pointer) symbol_triples.append(i - run_start_pointer) symbol_triples.append(run_start_array_index) in_run = False else: if code_point: in_run = True run_start_pointer = i run_start_array_index = len(symbol_index) symbol_index.append(code_point) if in_run: symbol_triples.append(run_start_pointer) symbol_triples.append(end - run_start_pointer) symbol_triples.append(run_start_array_index) in_run = False if in_run: raise Error() static_u16_table("JIS0212_ACCENTED", symbol_index) static_u16_table("JIS0212_ACCENTED_TRIPLES", symbol_triples) # gb18030 index = indexes["gb18030"] # Unicode 1.1 ideographs above the old GB2312 block # Compressed form takes 63% of uncompressed form pointers = [] offsets = [] previous_code_point = 0 for i in xrange(6080): code_point = index[i] if previous_code_point > code_point: raise Error() if code_point - previous_code_point != 1: pointers.append(i) offsets.append(code_point) previous_code_point = code_point static_u16_table("GBK_TOP_IDEOGRAPH_POINTERS", pointers) static_u16_table("GBK_TOP_IDEOGRAPH_OFFSETS", offsets) # Unicode 1.1 ideographs to the left of the old GB2312 block # Compressed form takes 40% of uncompressed form pointers = [] offsets = [] previous_code_point = 0 for row in xrange(0x7D - 0x29): for column in xrange(190 - 94): i = 7790 + column + (row * 190) if i > 23650: # Exclude compatibility ideographs at the end break code_point = index[i] if previous_code_point > code_point: raise Error() if code_point - previous_code_point != 1: pointers.append(column + (row * (190 - 94))) offsets.append(code_point) previous_code_point = code_point static_u16_table("GBK_LEFT_IDEOGRAPH_POINTERS", pointers) static_u16_table("GBK_LEFT_IDEOGRAPH_OFFSETS", offsets) # GBK other (excl. Ext A, Compat & PUA at the bottom) pointers = [] offsets = [] previous_code_point = 0 for row in xrange(0x29 - 0x20): for column in xrange(190 - 94): i = 6080 + column + (row * 190) code_point = index[i] if code_point - previous_code_point != 1: pointers.append(column + (row * (190 - 94))) offsets.append(code_point) previous_code_point = code_point pointers.append((190 - 94) * (0x29 - 0x20)) static_u16_table("GBK_OTHER_POINTERS", pointers) static_u16_table("GBK_OTHER_UNSORTED_OFFSETS", offsets) # GBK bottom: Compatibility ideagraphs, Ext A and PUA bottom_index = [] # 5 compat following Unified Ideographs for i in range(23651, 23656): bottom_index.append(index[i]) # Last row for i in range(23750, 23846): bottom_index.append(index[i]) static_u16_table("GBK_BOTTOM", bottom_index) # GB2312 Hanzi # (and the 5 PUA code points in between Level 1 and Level 2) hanzi_index = [] for row in xrange(0x77 - 0x2F): for column in xrange(94): hanzi_index.append(index[9026 + column + (row * 190)]) static_u16_table("GB2312_HANZI", hanzi_index) # GB2312 symbols symbol_index = [] for i in xrange(94): symbol_index.append(index[6176 + i]) static_u16_table("GB2312_SYMBOLS", symbol_index) # GB2312 symbols on Greek row (incl. PUA) symbol_index = [] for i in xrange(22): symbol_index.append(index[7189 + i]) static_u16_table("GB2312_SYMBOLS_AFTER_GREEK", symbol_index) # GB2312 Pinyin pinyin_index = [] for i in xrange(32): pinyin_index.append(index[7506 + i]) static_u16_table("GB2312_PINYIN", pinyin_index) # GB2312 other (excl. bottom PUA) pointers = [] offsets = [] previous_code_point = 0 for row in xrange(14): for column in xrange(94): i = 6366 + column + (row * 190) code_point = index[i] # Exclude the two ranges that were processed as # lookup tables above by filling them with # ASCII. Upon encode, ASCII code points will # never appear as the search key. if (i >= 7189 and i < 7189 + 22): code_point = i - 7189 elif (i >= 7506 and i < 7506 + 32): code_point = i - 7506 if code_point - previous_code_point != 1: pointers.append(column + (row * 94)) offsets.append(code_point) previous_code_point = code_point pointers.append(14 * 94) static_u16_table("GB2312_OTHER_POINTERS", pointers) static_u16_table("GB2312_OTHER_UNSORTED_OFFSETS", offsets) # Non-gbk code points pointers = [] offsets = [] for pair in indexes["gb18030-ranges"]: if pair[1] == 0x10000: break # the last entry doesn't fit in u16 pointers.append(pair[0]) offsets.append(pair[1]) static_u16_table("GB18030_RANGE_POINTERS", pointers) static_u16_table("GB18030_RANGE_OFFSETS", offsets) # Encoder table for Level 1 Hanzi # The units here really fit into 12 bits, but since we're # looking for speed here, let's use 16 bits per unit. # Once we use 16 bits per unit, we might as well precompute # the output bytes. level1_hanzi_index = hanzi_index[:(94 * (0xD8 - 0xB0) - 5)] level1_hanzi_pairs = [] for i in xrange(len(level1_hanzi_index)): hanzi_lead = (i / 94) + 0xB0 hanzi_trail = (i % 94) + 0xA1 level1_hanzi_pairs.append((level1_hanzi_index[i], (hanzi_lead, hanzi_trail))) level1_hanzi_pairs.sort(key=lambda x: x[0]) static_u16_table_from_indexable("GB2312_LEVEL1_HANZI_CODE_POINTS", level1_hanzi_pairs, 0, "gb-hanzi-encode") static_u8_pair_table_from_indexable("GB2312_LEVEL1_HANZI_BYTES", level1_hanzi_pairs, 1, "gb-hanzi-encode") # Fast Hanzi encoder table hanzi_bytes = [None] * (0x9FA7 - 0x4E00) for row in xrange(126): for column in xrange(190): pointer = column + (row * 190) code_point = index[pointer] if code_point and code_point >= 0x4E00 and code_point <= 0x9FA6: hanzi_lead = 0x81 + row hanzi_trail = column + (0x40 if column < 0x3F else 0x41) hanzi_bytes[code_point - 0x4E00] = (hanzi_lead, hanzi_trail) static_u8_pair_table("GBK_HANZI_BYTES", hanzi_bytes, "fast-gb-hanzi-encode") data_file.write(data_rs_end) data_file.close() # Variant variant_file = open("src/variant.rs", "w") variant_file.write('''// Copyright 2015-2016 Mozilla Foundation. See the COPYRIGHT // file at the top-level directory of this distribution. // // Licensed under the Apache License, Version 2.0 or the MIT license // , at your // option. This file may not be copied, modified, or distributed // except according to those terms. // THIS IS A GENERATED FILE. PLEASE DO NOT EDIT. // Instead, please regenerate using generate-encoding-data.py //! This module provides enums that wrap the various decoders and encoders. //! The purpose is to make `Decoder` and `Encoder` `Sized` by writing the //! dispatch explicitly for a finite set of specialized decoders and encoders. //! Unfortunately, this means the compiler doesn't generate the dispatch code //! and it has to be written here instead. //! //! The purpose of making `Decoder` and `Encoder` `Sized` is to allow stack //! allocation in Rust code, including the convenience methods on `Encoding`. ''') encoding_variants = [u"single-byte",] for encoding in multi_byte: if encoding["name"] in [u"UTF-16LE", u"UTF-16BE"]: continue else: encoding_variants.append(encoding["name"]) encoding_variants.append(u"UTF-16") decoder_variants = [] for variant in encoding_variants: if variant == u"GBK": continue decoder_variants.append(variant) encoder_variants = [] for variant in encoding_variants: if variant in [u"replacement", u"GBK", u"UTF-16"]: continue encoder_variants.append(variant) for variant in decoder_variants: variant_file.write("use %s::*;\n" % to_snake_name(variant)) variant_file.write('''use super::*; pub enum VariantDecoder { ''') for variant in decoder_variants: variant_file.write(" %s(%sDecoder),\n" % (to_camel_name(variant), to_camel_name(variant))) variant_file.write('''} impl VariantDecoder { ''') def write_variant_method(name, mut, arg_list, ret, variants, excludes, kind): variant_file.write('''pub fn %s(&''' % name) if mut: variant_file.write('''mut ''') variant_file.write('''self''') for arg in arg_list: variant_file.write(''', %s: %s''' % (arg[0], arg[1])) variant_file.write(''')''') if ret: variant_file.write(''' -> %s''' % ret) variant_file.write(''' {\nmatch *self {\n''') for variant in variants: variant_file.write('''Variant%s::%s(ref ''' % (kind, to_camel_name(variant))) if mut: variant_file.write('''mut ''') if variant in excludes: variant_file.write('''v) => (),''') continue variant_file.write('''v) => v.%s(''' % name) first = True for arg in arg_list: if not first: variant_file.write(''', ''') first = False variant_file.write(arg[0]) variant_file.write('''),\n''') variant_file.write('''}\n}\n\n''') write_variant_method("max_utf16_buffer_length", False, [("byte_length", "usize")], "Option", decoder_variants, [], "Decoder") write_variant_method("max_utf8_buffer_length_without_replacement", False, [("byte_length", "usize")], "Option", decoder_variants, [], "Decoder") write_variant_method("max_utf8_buffer_length", False, [("byte_length", "usize")], "Option", decoder_variants, [], "Decoder") write_variant_method("decode_to_utf16_raw", True, [("src", "&[u8]"), ("dst", "&mut [u16]"), ("last", "bool")], "(DecoderResult, usize, usize)", decoder_variants, [], "Decoder") write_variant_method("decode_to_utf8_raw", True, [("src", "&[u8]"), ("dst", "&mut [u8]"), ("last", "bool")], "(DecoderResult, usize, usize)", decoder_variants, [], "Decoder") variant_file.write(''' pub fn latin1_byte_compatible_up_to(&self, buffer: &[u8]) -> Option { match *self { VariantDecoder::SingleByte(ref v) => { return Some(v.latin1_byte_compatible_up_to(buffer)); } VariantDecoder::Utf8(ref v) => { if !v.in_neutral_state() { return None; } } VariantDecoder::Gb18030(ref v) => { if !v.in_neutral_state() { return None; } } VariantDecoder::Big5(ref v) => { if !v.in_neutral_state() { return None; } } VariantDecoder::EucJp(ref v) => { if !v.in_neutral_state() { return None; } } VariantDecoder::Iso2022Jp(ref v) => { if v.in_neutral_state() { return Some(Encoding::iso_2022_jp_ascii_valid_up_to(buffer)); } return None; } VariantDecoder::ShiftJis(ref v) => { if !v.in_neutral_state() { return None; } } VariantDecoder::EucKr(ref v) => { if !v.in_neutral_state() { return None; } } VariantDecoder::UserDefined(_) => {} VariantDecoder::Replacement(_) | VariantDecoder::Utf16(_) => { return None; } }; Some(Encoding::ascii_valid_up_to(buffer)) } } pub enum VariantEncoder { ''') for variant in encoder_variants: variant_file.write(" %s(%sEncoder),\n" % (to_camel_name(variant), to_camel_name(variant))) variant_file.write('''} impl VariantEncoder { pub fn has_pending_state(&self) -> bool { match *self { VariantEncoder::Iso2022Jp(ref v) => { v.has_pending_state() } _ => false, } } ''') write_variant_method("max_buffer_length_from_utf16_without_replacement", False, [("u16_length", "usize")], "Option", encoder_variants, [], "Encoder") write_variant_method("max_buffer_length_from_utf8_without_replacement", False, [("byte_length", "usize")], "Option", encoder_variants, [], "Encoder") write_variant_method("encode_from_utf16_raw", True, [("src", "&[u16]"), ("dst", "&mut [u8]"), ("last", "bool")], "(EncoderResult, usize, usize)", encoder_variants, [], "Encoder") write_variant_method("encode_from_utf8_raw", True, [("src", "&str"), ("dst", "&mut [u8]"), ("last", "bool")], "(EncoderResult, usize, usize)", encoder_variants, [], "Encoder") variant_file.write('''} pub enum VariantEncoding { SingleByte(&'static [u16; 128], u16, u8, u8),''') for encoding in multi_byte: variant_file.write("%s,\n" % to_camel_name(encoding["name"])) variant_file.write('''} impl VariantEncoding { pub fn new_variant_decoder(&self) -> VariantDecoder { match *self { VariantEncoding::SingleByte(table, _, _, _) => SingleByteDecoder::new(table), VariantEncoding::Utf8 => Utf8Decoder::new(), VariantEncoding::Gbk | VariantEncoding::Gb18030 => Gb18030Decoder::new(), VariantEncoding::Big5 => Big5Decoder::new(), VariantEncoding::EucJp => EucJpDecoder::new(), VariantEncoding::Iso2022Jp => Iso2022JpDecoder::new(), VariantEncoding::ShiftJis => ShiftJisDecoder::new(), VariantEncoding::EucKr => EucKrDecoder::new(), VariantEncoding::Replacement => ReplacementDecoder::new(), VariantEncoding::UserDefined => UserDefinedDecoder::new(), VariantEncoding::Utf16Be => Utf16Decoder::new(true), VariantEncoding::Utf16Le => Utf16Decoder::new(false), } } pub fn new_encoder(&self, encoding: &'static Encoding) -> Encoder { match *self { VariantEncoding::SingleByte(table, run_bmp_offset, run_byte_offset, run_length) => SingleByteEncoder::new(encoding, table, run_bmp_offset, run_byte_offset, run_length), VariantEncoding::Utf8 => Utf8Encoder::new(encoding), VariantEncoding::Gbk => Gb18030Encoder::new(encoding, false), VariantEncoding::Gb18030 => Gb18030Encoder::new(encoding, true), VariantEncoding::Big5 => Big5Encoder::new(encoding), VariantEncoding::EucJp => EucJpEncoder::new(encoding), VariantEncoding::Iso2022Jp => Iso2022JpEncoder::new(encoding), VariantEncoding::ShiftJis => ShiftJisEncoder::new(encoding), VariantEncoding::EucKr => EucKrEncoder::new(encoding), VariantEncoding::UserDefined => UserDefinedEncoder::new(encoding), VariantEncoding::Utf16Be | VariantEncoding::Replacement | VariantEncoding::Utf16Le => unreachable!(), } } pub fn is_single_byte(&self) -> bool { match *self { VariantEncoding::SingleByte(_, _, _, _) | VariantEncoding::UserDefined => true, _ => false, } } } ''') variant_file.close() (ffi_rs_begin, ffi_rs_end) = read_non_generated("../encoding_c/src/lib.rs") ffi_file = open("../encoding_c/src/lib.rs", "w") ffi_file.write(ffi_rs_begin) ffi_file.write(""" // Instead, please regenerate using generate-encoding-data.py /// The minimum length of buffers that may be passed to `encoding_name()`. pub const ENCODING_NAME_MAX_LENGTH: usize = %d; // %s """ % (longest_name_length, longest_name)) for name in preferred: ffi_file.write('''/// The %s encoding. #[no_mangle] pub static %s_ENCODING: ConstEncoding = ConstEncoding(&%s_INIT); ''' % (to_dom_name(name), to_constant_name(name), to_constant_name(name))) ffi_file.write(ffi_rs_end) ffi_file.close() (single_byte_rs_begin, single_byte_rs_end) = read_non_generated("src/single_byte.rs") single_byte_file = open("src/single_byte.rs", "w") single_byte_file.write(single_byte_rs_begin) single_byte_file.write(""" // Instead, please regenerate using generate-encoding-data.py #[test] fn test_single_byte_decode() {""") for name in preferred: if name == u"ISO-8859-8-I": continue; if is_single_byte(name): single_byte_file.write(""" decode_single_byte(%s, &data::SINGLE_BYTE_DATA.%s);""" % (to_constant_name(name), to_snake_name(name))) single_byte_file.write(""" } #[test] fn test_single_byte_encode() {""") for name in preferred: if name == u"ISO-8859-8-I": continue; if is_single_byte(name): single_byte_file.write(""" encode_single_byte(%s, &data::SINGLE_BYTE_DATA.%s);""" % (to_constant_name(name), to_snake_name(name))) single_byte_file.write(""" } """) single_byte_file.write(single_byte_rs_end) single_byte_file.close() static_file = open("../encoding_c/include/encoding_rs_statics.h", "w") static_file.write("""// Copyright 2016 Mozilla Foundation. See the COPYRIGHT // file at the top-level directory of this distribution. // // Licensed under the Apache License, Version 2.0 or the MIT license // , at your // option. This file may not be copied, modified, or distributed // except according to those terms. // THIS IS A GENERATED FILE. PLEASE DO NOT EDIT. // Instead, please regenerate using generate-encoding-data.py // This file is not meant to be included directly. Instead, encoding_rs.h // includes this file. #ifndef encoding_rs_statics_h_ #define encoding_rs_statics_h_ #ifndef ENCODING_RS_ENCODING #define ENCODING_RS_ENCODING Encoding #ifndef __cplusplus typedef struct Encoding_ Encoding; #endif #endif #ifndef ENCODING_RS_NOT_NULL_CONST_ENCODING_PTR #define ENCODING_RS_NOT_NULL_CONST_ENCODING_PTR const ENCODING_RS_ENCODING* #endif #ifndef ENCODING_RS_ENCODER #define ENCODING_RS_ENCODER Encoder #ifndef __cplusplus typedef struct Encoder_ Encoder; #endif #endif #ifndef ENCODING_RS_DECODER #define ENCODING_RS_DECODER Decoder #ifndef __cplusplus typedef struct Decoder_ Decoder; #endif #endif #define INPUT_EMPTY 0 #define OUTPUT_FULL 0xFFFFFFFF // %s #define ENCODING_NAME_MAX_LENGTH %d """ % (longest_name, longest_name_length)) for name in preferred: static_file.write('''/// The %s encoding. extern ENCODING_RS_NOT_NULL_CONST_ENCODING_PTR const %s_ENCODING; ''' % (to_dom_name(name), to_constant_name(name))) static_file.write("""#endif // encoding_rs_statics_h_ """) static_file.close() (utf_8_rs_begin, utf_8_rs_end) = read_non_generated("src/utf_8.rs") utf_8_file = open("src/utf_8.rs", "w") utf_8_file.write(utf_8_rs_begin) utf_8_file.write(""" // Instead, please regenerate using generate-encoding-data.py pub static UTF8_DATA: Utf8Data = Utf8Data { table: [ """) for i in range(256): combined = (1 << 2) # invalid lead if i < 0x80 or i > 0xBF: combined |= (1 << 3) # normal trail if i < 0xA0 or i > 0xBF: combined |= (1 << 4) # three-byte special lower bound if i < 0x80 or i > 0x9F: combined |= (1 << 5) # three-byte special upper bound if i < 0x90 or i > 0xBF: combined |= (1 << 6) # four-byte special lower bound if i < 0x80 or i > 0x8F: combined |= (1 << 7) # four-byte special upper bound utf_8_file.write("%d," % combined) for i in range(128, 256): lane = (1 << 2) # invalid lead if i >= 0xC2 and i <= 0xDF: lane = (1 << 3) # normal trail elif i == 0xE0: lane = (1 << 4) # three-byte special lower bound elif i >= 0xE1 and i <= 0xEC: lane = (1 << 3) # normal trail elif i == 0xED: lane = (1 << 5) # three-byte special upper bound elif i >= 0xEE and i <= 0xEF: lane = (1 << 3) # normal trail elif i == 0xF0: lane = (1 << 6) # four-byte special lower bound elif i >= 0xF1 and i <= 0xF3: lane = (1 << 3) # normal trail elif i == 0xF4: lane = (1 << 7) # four-byte special upper bound utf_8_file.write("%d," % lane) utf_8_file.write(""" ], }; """) utf_8_file.write(utf_8_rs_end) utf_8_file.close() # Unit tests TEST_HEADER = '''Any copyright to the test code below this comment is dedicated to the Public Domain. http://creativecommons.org/publicdomain/zero/1.0/ This is a generated file. Please do not edit. Instead, please regenerate using generate-encoding-data.py ''' index = indexes["jis0208"] jis0208_in_file = open("src/test_data/jis0208_in.txt", "w") jis0208_in_file.write(TEST_HEADER) for pointer in range(0, 94 * 94): (lead, trail) = divmod(pointer, 94) lead += 0xA1 trail += 0xA1 jis0208_in_file.write("%s%s\n" % (chr(lead), chr(trail))) jis0208_in_file.close() jis0208_in_ref_file = open("src/test_data/jis0208_in_ref.txt", "w") jis0208_in_ref_file.write(TEST_HEADER) for pointer in range(0, 94 * 94): code_point = index[pointer] if code_point: jis0208_in_ref_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) else: jis0208_in_ref_file.write(u"\uFFFD\n".encode("utf-8")) jis0208_in_ref_file.close() jis0208_out_file = open("src/test_data/jis0208_out.txt", "w") jis0208_out_ref_file = open("src/test_data/jis0208_out_ref.txt", "w") jis0208_out_file.write(TEST_HEADER) jis0208_out_ref_file.write(TEST_HEADER) for pointer in range(0, 94 * 94): code_point = index[pointer] if code_point: revised_pointer = pointer if revised_pointer == 8644 or (revised_pointer >= 1207 and revised_pointer < 1220): revised_pointer = index.index(code_point) (lead, trail) = divmod(revised_pointer, 94) lead += 0xA1 trail += 0xA1 jis0208_out_ref_file.write("%s%s\n" % (chr(lead), chr(trail))) jis0208_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) jis0208_out_file.close() jis0208_out_ref_file.close() shift_jis_in_file = open("src/test_data/shift_jis_in.txt", "w") shift_jis_in_file.write(TEST_HEADER) for pointer in range(0, len(index)): (lead, trail) = divmod(pointer, 188) lead += 0x81 if lead < 0x1F else 0xC1 trail += 0x40 if trail < 0x3F else 0x41 shift_jis_in_file.write("%s%s\n" % (chr(lead), chr(trail))) shift_jis_in_file.close() shift_jis_in_ref_file = open("src/test_data/shift_jis_in_ref.txt", "w") shift_jis_in_ref_file.write(TEST_HEADER) for pointer in range(0, len(index)): code_point = 0xE000 - 8836 + pointer if pointer >= 8836 and pointer <= 10715 else index[pointer] if code_point: shift_jis_in_ref_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) else: trail = pointer % 188 trail += 0x40 if trail < 0x3F else 0x41 if trail < 0x80: shift_jis_in_ref_file.write((u"\uFFFD%s\n" % unichr(trail)).encode("utf-8")) else: shift_jis_in_ref_file.write(u"\uFFFD\n".encode("utf-8")) shift_jis_in_ref_file.close() shift_jis_out_file = open("src/test_data/shift_jis_out.txt", "w") shift_jis_out_ref_file = open("src/test_data/shift_jis_out_ref.txt", "w") shift_jis_out_file.write(TEST_HEADER) shift_jis_out_ref_file.write(TEST_HEADER) for pointer in range(0, 8272): code_point = index[pointer] if code_point: revised_pointer = pointer if revised_pointer >= 1207 and revised_pointer < 1220: revised_pointer = index.index(code_point) (lead, trail) = divmod(revised_pointer, 188) lead += 0x81 if lead < 0x1F else 0xC1 trail += 0x40 if trail < 0x3F else 0x41 shift_jis_out_ref_file.write("%s%s\n" % (chr(lead), chr(trail))) shift_jis_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) for pointer in range(8836, len(index)): code_point = index[pointer] if code_point: revised_pointer = index.index(code_point) if revised_pointer >= 8272 and revised_pointer < 8836: revised_pointer = pointer (lead, trail) = divmod(revised_pointer, 188) lead += 0x81 if lead < 0x1F else 0xC1 trail += 0x40 if trail < 0x3F else 0x41 shift_jis_out_ref_file.write("%s%s\n" % (chr(lead), chr(trail))) shift_jis_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) shift_jis_out_file.close() shift_jis_out_ref_file.close() iso_2022_jp_in_file = open("src/test_data/iso_2022_jp_in.txt", "w") iso_2022_jp_in_file.write(TEST_HEADER) for pointer in range(0, 94 * 94): (lead, trail) = divmod(pointer, 94) lead += 0x21 trail += 0x21 iso_2022_jp_in_file.write("\x1B$B%s%s\x1B(B\n" % (chr(lead), chr(trail))) iso_2022_jp_in_file.close() iso_2022_jp_in_ref_file = open("src/test_data/iso_2022_jp_in_ref.txt", "w") iso_2022_jp_in_ref_file.write(TEST_HEADER) for pointer in range(0, 94 * 94): code_point = index[pointer] if code_point: iso_2022_jp_in_ref_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) else: iso_2022_jp_in_ref_file.write(u"\uFFFD\n".encode("utf-8")) iso_2022_jp_in_ref_file.close() iso_2022_jp_out_file = open("src/test_data/iso_2022_jp_out.txt", "w") iso_2022_jp_out_ref_file = open("src/test_data/iso_2022_jp_out_ref.txt", "w") iso_2022_jp_out_file.write(TEST_HEADER) iso_2022_jp_out_ref_file.write(TEST_HEADER) for pointer in range(0, 94 * 94): code_point = index[pointer] if code_point: revised_pointer = pointer if revised_pointer == 8644 or (revised_pointer >= 1207 and revised_pointer < 1220): revised_pointer = index.index(code_point) (lead, trail) = divmod(revised_pointer, 94) lead += 0x21 trail += 0x21 iso_2022_jp_out_ref_file.write("\x1B$B%s%s\x1B(B\n" % (chr(lead), chr(trail))) iso_2022_jp_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) for i in xrange(len(half_width_index)): code_point = i + 0xFF61 normalized_code_point = half_width_index[i] pointer = index.index(normalized_code_point) (lead, trail) = divmod(pointer, 94) lead += 0x21 trail += 0x21 iso_2022_jp_out_ref_file.write("\x1B$B%s%s\x1B(B\n" % (chr(lead), chr(trail))) iso_2022_jp_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) iso_2022_jp_out_file.close() iso_2022_jp_out_ref_file.close() index = indexes["euc-kr"] euc_kr_in_file = open("src/test_data/euc_kr_in.txt", "w") euc_kr_in_file.write(TEST_HEADER) for pointer in range(0, len(index)): (lead, trail) = divmod(pointer, 190) lead += 0x81 trail += 0x41 euc_kr_in_file.write("%s%s\n" % (chr(lead), chr(trail))) euc_kr_in_file.close() euc_kr_in_ref_file = open("src/test_data/euc_kr_in_ref.txt", "w") euc_kr_in_ref_file.write(TEST_HEADER) for pointer in range(0, len(index)): code_point = index[pointer] if code_point: euc_kr_in_ref_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) else: trail = pointer % 190 trail += 0x41 if trail < 0x80: euc_kr_in_ref_file.write((u"\uFFFD%s\n" % unichr(trail)).encode("utf-8")) else: euc_kr_in_ref_file.write(u"\uFFFD\n".encode("utf-8")) euc_kr_in_ref_file.close() euc_kr_out_file = open("src/test_data/euc_kr_out.txt", "w") euc_kr_out_ref_file = open("src/test_data/euc_kr_out_ref.txt", "w") euc_kr_out_file.write(TEST_HEADER) euc_kr_out_ref_file.write(TEST_HEADER) for pointer in range(0, len(index)): code_point = index[pointer] if code_point: (lead, trail) = divmod(pointer, 190) lead += 0x81 trail += 0x41 euc_kr_out_ref_file.write("%s%s\n" % (chr(lead), chr(trail))) euc_kr_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) euc_kr_out_file.close() euc_kr_out_ref_file.close() index = indexes["gb18030"] gb18030_in_file = open("src/test_data/gb18030_in.txt", "w") gb18030_in_file.write(TEST_HEADER) for pointer in range(0, len(index)): (lead, trail) = divmod(pointer, 190) lead += 0x81 trail += 0x40 if trail < 0x3F else 0x41 gb18030_in_file.write("%s%s\n" % (chr(lead), chr(trail))) gb18030_in_file.close() gb18030_in_ref_file = open("src/test_data/gb18030_in_ref.txt", "w") gb18030_in_ref_file.write(TEST_HEADER) for pointer in range(0, len(index)): code_point = index[pointer] if code_point: gb18030_in_ref_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) else: trail = pointer % 190 trail += 0x40 if trail < 0x3F else 0x41 if trail < 0x80: gb18030_in_ref_file.write((u"\uFFFD%s\n" % unichr(trail)).encode("utf-8")) else: gb18030_in_ref_file.write(u"\uFFFD\n".encode("utf-8")) gb18030_in_ref_file.close() gb18030_out_file = open("src/test_data/gb18030_out.txt", "w") gb18030_out_ref_file = open("src/test_data/gb18030_out_ref.txt", "w") gb18030_out_file.write(TEST_HEADER) gb18030_out_ref_file.write(TEST_HEADER) for pointer in range(0, len(index)): if pointer == 6555: continue code_point = index[pointer] if code_point: (lead, trail) = divmod(pointer, 190) lead += 0x81 trail += 0x40 if trail < 0x3F else 0x41 gb18030_out_ref_file.write("%s%s\n" % (chr(lead), chr(trail))) gb18030_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) gb18030_out_file.close() gb18030_out_ref_file.close() index = indexes["big5"] big5_in_file = open("src/test_data/big5_in.txt", "w") big5_in_file.write(TEST_HEADER) for pointer in range(0, len(index)): (lead, trail) = divmod(pointer, 157) lead += 0x81 trail += 0x40 if trail < 0x3F else 0x62 big5_in_file.write("%s%s\n" % (chr(lead), chr(trail))) big5_in_file.close() big5_two_characters = { 1133: u"\u00CA\u0304", 1135: u"\u00CA\u030C", 1164: u"\u00EA\u0304", 1166: u"\u00EA\u030C", } big5_in_ref_file = open("src/test_data/big5_in_ref.txt", "w") big5_in_ref_file.write(TEST_HEADER) for pointer in range(0, len(index)): if pointer in big5_two_characters.keys(): big5_in_ref_file.write((u"%s\n" % big5_two_characters[pointer]).encode("utf-8")) continue code_point = index[pointer] if code_point: big5_in_ref_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) else: trail = pointer % 157 trail += 0x40 if trail < 0x3F else 0x62 if trail < 0x80: big5_in_ref_file.write((u"\uFFFD%s\n" % unichr(trail)).encode("utf-8")) else: big5_in_ref_file.write(u"\uFFFD\n".encode("utf-8")) big5_in_ref_file.close() prefer_last = [ 0x2550, 0x255E, 0x2561, 0x256A, 0x5341, 0x5345, ] pointer_for_prefer_last = [] for code_point in prefer_last: # Python lists don't have .rindex() :-( for i in xrange(len(index) - 1, -1, -1): candidate = index[i] if candidate == code_point: pointer_for_prefer_last.append(i) break big5_out_file = open("src/test_data/big5_out.txt", "w") big5_out_ref_file = open("src/test_data/big5_out_ref.txt", "w") big5_out_file.write(TEST_HEADER) big5_out_ref_file.write(TEST_HEADER) for pointer in range(((0xA1 - 0x81) * 157), len(index)): code_point = index[pointer] if code_point: if code_point in prefer_last: if pointer != pointer_for_prefer_last[prefer_last.index(code_point)]: continue else: if pointer != index.index(code_point): continue (lead, trail) = divmod(pointer, 157) lead += 0x81 trail += 0x40 if trail < 0x3F else 0x62 big5_out_ref_file.write("%s%s\n" % (chr(lead), chr(trail))) big5_out_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) big5_out_file.close() big5_out_ref_file.close() index = indexes["jis0212"] jis0212_in_file = open("src/test_data/jis0212_in.txt", "w") jis0212_in_file.write(TEST_HEADER) for pointer in range(0, len(index)): (lead, trail) = divmod(pointer, 94) lead += 0xA1 trail += 0xA1 jis0212_in_file.write("\x8F%s%s\n" % (chr(lead), chr(trail))) jis0212_in_file.close() jis0212_in_ref_file = open("src/test_data/jis0212_in_ref.txt", "w") jis0212_in_ref_file.write(TEST_HEADER) for pointer in range(0, len(index)): code_point = index[pointer] if code_point: jis0212_in_ref_file.write((u"%s\n" % unichr(code_point)).encode("utf-8")) else: jis0212_in_ref_file.write(u"\uFFFD\n".encode("utf-8")) jis0212_in_ref_file.close() (codepage_begin, codepage_end) = read_non_generated("../codepage/src/lib.rs") codepage_file = open("../codepage/src/lib.rs", "w") codepage_file.write(codepage_begin) codepage_file.write(""" // Instead, please regenerate using generate-encoding-data.py /// Supported code page numbers in estimated order of usage frequency static CODE_PAGES: [u16; %d] = [ """ % len(code_pages)) for code_page in code_pages: codepage_file.write(" %d,\n" % code_page) codepage_file.write("""]; /// Encodings corresponding to the code page numbers in the same order static ENCODINGS: [&'static Encoding; %d] = [ """ % len(code_pages)) for code_page in code_pages: name = encodings_by_code_page[code_page] codepage_file.write(" &%s_INIT,\n" % to_constant_name(name)) codepage_file.write("""]; """) codepage_file.write(codepage_end) codepage_file.close() (codepage_test_begin, codepage_test_end) = read_non_generated("../codepage/src/tests.rs") codepage_test_file = open("../codepage/src/tests.rs", "w") codepage_test_file.write(codepage_test_begin) codepage_test_file.write(""" // Instead, please regenerate using generate-encoding-data.py #[test] fn test_to_encoding() { assert_eq!(to_encoding(0), None); """) for code_page in code_pages: codepage_test_file.write(" assert_eq!(to_encoding(%d), Some(%s));\n" % (code_page, to_constant_name(encodings_by_code_page[code_page]))) codepage_test_file.write("""} #[test] fn test_from_encoding() { """) for name in preferred: if code_pages_by_encoding.has_key(name): codepage_test_file.write(" assert_eq!(from_encoding(%s), Some(%d));\n" % (to_constant_name(name), code_pages_by_encoding[name])) else: codepage_test_file.write(" assert_eq!(from_encoding(%s), None);\n" % to_constant_name(name)) codepage_test_file.write("""} """) codepage_test_file.write(codepage_test_end) codepage_test_file.close() subprocess.call(["cargo", "fmt"])