to use the bisect built-in module we need to have two separate tables, one with the ranges themselves (which we pass to bisect to get an index) and the other containing the script name for each range. Also, allow the buildUCD.py script to load data files from a local directory, e.g. to allow downstream maintainers to rebuild the generated modules from local files instead of downloading from Unicode website.
193 lines
5.9 KiB
Python
Executable File
193 lines
5.9 KiB
Python
Executable File
#!/usr/bin/env python
|
|
"""
|
|
Tools to parse data files from the Unicode Character Database.
|
|
"""
|
|
|
|
from __future__ import print_function, absolute_import, division
|
|
from __future__ import unicode_literals
|
|
from fontTools.misc.py23 import *
|
|
|
|
try:
|
|
from urllib.request import urlopen
|
|
except ImportError:
|
|
from urllib2 import urlopen
|
|
from contextlib import closing, contextmanager
|
|
import re
|
|
from codecs import iterdecode
|
|
import logging
|
|
import os
|
|
from os.path import abspath, dirname, join as pjoin, pardir, sep
|
|
|
|
|
|
UNIDATA_URL = "https://unicode.org/Public/UNIDATA/"
|
|
|
|
# by default save output files to ../Lib/fontTools/unicodedata/
|
|
UNIDATA_PATH = pjoin(abspath(dirname(__file__)), pardir,
|
|
"Lib", "fontTools", "unicodedata") + sep
|
|
|
|
SRC_ENCODING = "# -*- coding: utf-8 -*-\n"
|
|
|
|
NOTICE = "# NOTE: This file was auto-generated with MetaTools/buildUCD.py.\n"
|
|
|
|
MAX_UNICODE = 0x10FFFF
|
|
|
|
log = logging.getLogger()
|
|
|
|
|
|
@contextmanager
|
|
def open_unidata_file(filename):
|
|
"""Open a text file from https://unicode.org/Public/UNIDATA/"""
|
|
url = UNIDATA_URL + filename
|
|
with closing(urlopen(url)) as response:
|
|
yield iterdecode(response, encoding="utf-8")
|
|
|
|
|
|
def parse_unidata_header(infile):
|
|
"""Read the top header of data files, until the first line
|
|
that does not start with '#'.
|
|
"""
|
|
header = []
|
|
line = next(infile)
|
|
while line.startswith("#"):
|
|
header.append(line)
|
|
line = next(infile)
|
|
return "".join(header)
|
|
|
|
|
|
def parse_range_properties(infile, default="Unknown"):
|
|
"""Parse a Unicode data file containing a column with one character or
|
|
a range of characters, and another column containing a property value
|
|
separated by a semicolon. Comments after '#' are ignored.
|
|
|
|
If the ranges defined in the data file are not continuous, assign the
|
|
'default' property to the unassigned codepoints.
|
|
|
|
Return a list of (start, end, property_name) tuples.
|
|
"""
|
|
ranges = []
|
|
line_regex = re.compile(
|
|
r"^"
|
|
r"([0-9A-F]{4,6})" # first character code
|
|
r"(?:\.\.([0-9A-F]{4,6}))?" # optional second character code
|
|
r"\s*;\s*"
|
|
r"([^#]+)") # everything up to the potential comment
|
|
for line in infile:
|
|
match = line_regex.match(line)
|
|
if not match:
|
|
continue
|
|
|
|
first, last, data = match.groups()
|
|
if last is None:
|
|
last = first
|
|
|
|
first = int(first, 16)
|
|
last = int(last, 16)
|
|
data = data.rstrip()
|
|
|
|
ranges.append((first, last, data))
|
|
|
|
ranges.sort()
|
|
|
|
# fill the gaps between explicitly defined ranges
|
|
last_start, last_end = -1, -1
|
|
full_ranges = []
|
|
for start, end, name in ranges:
|
|
assert last_end < start
|
|
assert start <= end
|
|
if start - last_end > 1:
|
|
full_ranges.append((last_end+1, start-1, default))
|
|
full_ranges.append((start, end, name))
|
|
last_start, last_end = start, end
|
|
if last_end != MAX_UNICODE:
|
|
full_ranges.append((last_end+1, MAX_UNICODE, default))
|
|
|
|
# reduce total number of ranges by combining continuous ones
|
|
last_start, last_end, last_name = full_ranges.pop(0)
|
|
merged_ranges = []
|
|
for start, end, name in full_ranges:
|
|
if name == last_name:
|
|
continue
|
|
else:
|
|
merged_ranges.append((last_start, start-1, last_name))
|
|
last_start, line_end, last_name = start, end, name
|
|
merged_ranges.append((last_start, MAX_UNICODE, last_name))
|
|
|
|
# make sure that the ranges cover the full unicode repertoire
|
|
assert merged_ranges[0][0] == 0
|
|
for (cs, ce, cn), (ns, ne, nn) in zip(merged_ranges, merged_ranges[1:]):
|
|
assert ce+1 == ns
|
|
assert merged_ranges[-1][1] == MAX_UNICODE
|
|
|
|
return merged_ranges
|
|
|
|
|
|
def build_scripts(local_ucd=None, output_path=None):
|
|
"""Fetch "Scripts.txt" data file from Unicode official website, parse
|
|
the script ranges and write them as a list of Python tuples to
|
|
'fontTools.unicodedata.scripts'.
|
|
|
|
To load "Scripts.txt" from a local directory, you can use the
|
|
'local_ucd' argument.
|
|
"""
|
|
if not output_path:
|
|
output_path = UNIDATA_PATH + "scripts.py"
|
|
|
|
filename = "Scripts.txt"
|
|
if local_ucd:
|
|
log.info("loading %r from local directory %r", filename, local_ucd)
|
|
cm = open(pjoin(local_ucd, filename), "r", encoding="utf-8")
|
|
else:
|
|
log.info("downloading %r from %r", filename, UNIDATA_URL)
|
|
cm = open_unidata_file(filename)
|
|
|
|
with cm as f:
|
|
header = parse_unidata_header(f)
|
|
ranges = parse_range_properties(f)
|
|
|
|
max_name_length = max(len(n) for _, _, n in ranges)
|
|
|
|
with open(output_path, "w", encoding="utf-8") as f:
|
|
f.write(SRC_ENCODING)
|
|
f.write("#\n")
|
|
f.write(NOTICE)
|
|
f.write("# Source: {}{}\n".format(UNIDATA_URL, filename))
|
|
f.write("#\n")
|
|
f.write(header+"\n\n")
|
|
|
|
f.write("SCRIPT_RANGES = [\n")
|
|
for first, last, script_name in ranges:
|
|
f.write(" 0x{:0>4X}, # .. 0x{:0>4X} ; {}\n".format(
|
|
first, last, tostr(script_name)))
|
|
f.write("]\n")
|
|
|
|
f.write("\n")
|
|
f.write("SCRIPT_NAMES = [\n")
|
|
for first, last, script_name in ranges:
|
|
script_name = "'{}',".format(script_name)
|
|
f.write(" {} # {:0>4X}..{:0>4X}\n".format(
|
|
script_name.ljust(max_name_length+3), first, last))
|
|
f.write("]\n")
|
|
|
|
log.info("saved new file: %r", os.path.normpath(output_path))
|
|
|
|
|
|
def main():
|
|
import argparse
|
|
|
|
parser = argparse.ArgumentParser(
|
|
description="Generate fontTools.unicodedata from UCD data files")
|
|
parser.add_argument(
|
|
'--ucd-path', help="Path to local folder containing UCD data files")
|
|
parser.add_argument('-q', '--quiet', action="store_true")
|
|
options = parser.parse_args()
|
|
|
|
level = "WARNING" if options.quiet else "INFO"
|
|
logging.basicConfig(level=level, format="%(message)s")
|
|
|
|
build_scripts(local_ucd=options.ucd_path)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
import sys
|
|
sys.exit(main())
|