# Copyright 2013 Google, Inc. All Rights Reserved. # # Google Author(s): Behdad Esfahbod from __future__ import print_function, division, absolute_import from fontTools.misc.py23 import * from fontTools import ttLib from fontTools.ttLib.tables import otTables from fontTools.misc import psCharStrings from fontTools.pens.basePen import NullPen from fontTools.misc.loggingTools import Timer import sys import struct import array import logging from collections import Counter from types import MethodType __usage__ = "pyftsubset font-file [glyph...] [--option=value]..." __doc__="""\ pyftsubset -- OpenType font subsetter and optimizer pyftsubset is an OpenType font subsetter and optimizer, based on fontTools. It accepts any TT- or CFF-flavored OpenType (.otf or .ttf) or WOFF (.woff) font file. The subsetted glyph set is based on the specified glyphs or characters, and specified OpenType layout features. The tool also performs some size-reducing optimizations, aimed for using subset fonts as webfonts. Individual optimizations can be enabled or disabled, and are enabled by default when they are safe. Usage: """+__usage__+""" At least one glyph or one of --gids, --gids-file, --glyphs, --glyphs-file, --text, --text-file, --unicodes, or --unicodes-file, must be specified. Arguments: font-file The input font file. glyph Specify one or more glyph identifiers to include in the subset. Must be PS glyph names, or the special string '*' to keep the entire glyph set. Initial glyph set specification: These options populate the initial glyph set. Same option can appear multiple times, and the results are accummulated. --gids=[,...] Specify comma/whitespace-separated list of glyph IDs or ranges as decimal numbers. For example, --gids=10-12,14 adds glyphs with numbers 10, 11, 12, and 14. --gids-file= Like --gids but reads from a file. Anything after a '#' on any line is ignored as comments. --glyphs=[,...] Specify comma/whitespace-separated PS glyph names to add to the subset. Note that only PS glyph names are accepted, not gidNNN, U+XXXX, etc that are accepted on the command line. The special string '*' will keep the entire glyph set. --glyphs-file= Like --glyphs but reads from a file. Anything after a '#' on any line is ignored as comments. --text= Specify characters to include in the subset, as UTF-8 string. --text-file= Like --text but reads from a file. Newline character are not added to the subset. --unicodes=[,...] Specify comma/whitespace-separated list of Unicode codepoints or ranges as hex numbers, optionally prefixed with 'U+', 'u', etc. For example, --unicodes=41-5a,61-7a adds ASCII letters, so does the more verbose --unicodes=U+0041-005A,U+0061-007A. The special strings '*' will choose all Unicode characters mapped by the font. --unicodes-file= Like --unicodes, but reads from a file. Anything after a '#' on any line in the file is ignored as comments. --ignore-missing-glyphs Do not fail if some requested glyphs or gids are not available in the font. --no-ignore-missing-glyphs Stop and fail if some requested glyphs or gids are not available in the font. [default] --ignore-missing-unicodes [default] Do not fail if some requested Unicode characters (including those indirectly specified using --text or --text-file) are not available in the font. --no-ignore-missing-unicodes Stop and fail if some requested Unicode characters are not available in the font. Note the default discrepancy between ignoring missing glyphs versus unicodes. This is for historical reasons and in the future --no-ignore-missing-unicodes might become default. Other options: For the other options listed below, to see the current value of the option, pass a value of '?' to it, with or without a '='. Examples: $ pyftsubset --glyph-names? Current setting for 'glyph-names' is: False $ ./pyftsubset --name-IDs=? Current setting for 'name-IDs' is: [1, 2] $ ./pyftsubset --hinting? --no-hinting --hinting? Current setting for 'hinting' is: True Current setting for 'hinting' is: False Output options: --output-file= The output font file. If not specified, the subsetted font will be saved in as font-file.subset. --flavor= Specify flavor of output font file. May be 'woff' or 'woff2'. Note that WOFF2 requires the Brotli Python extension, available at https://github.com/google/brotli --with-zopfli Use the Google Zopfli algorithm to compress WOFF. The output is 3-8 % smaller than pure zlib, but the compression speed is much slower. The Zopfli Python bindings are available at: https://pypi.python.org/pypi/zopfli Glyph set expansion: These options control how additional glyphs are added to the subset. --notdef-glyph Add the '.notdef' glyph to the subset (ie, keep it). [default] --no-notdef-glyph Drop the '.notdef' glyph unless specified in the glyph set. This saves a few bytes, but is not possible for Postscript-flavored fonts, as those require '.notdef'. For TrueType-flavored fonts, this works fine as long as no unsupported glyphs are requested from the font. --notdef-outline Keep the outline of '.notdef' glyph. The '.notdef' glyph outline is used when glyphs not supported by the font are to be shown. It is not needed otherwise. --no-notdef-outline When including a '.notdef' glyph, remove its outline. This saves a few bytes. [default] --recommended-glyphs Add glyphs 0, 1, 2, and 3 to the subset, as recommended for TrueType-flavored fonts: '.notdef', 'NULL' or '.null', 'CR', 'space'. Some legacy software might require this, but no modern system does. --no-recommended-glyphs Do not add glyphs 0, 1, 2, and 3 to the subset, unless specified in glyph set. [default] --layout-features[+|-]=[,...] Specify (=), add to (+=) or exclude from (-=) the comma-separated set of OpenType layout feature tags that will be preserved. Glyph variants used by the preserved features are added to the specified subset glyph set. By default, 'calt', 'ccmp', 'clig', 'curs', 'dnom', 'frac', 'kern', 'liga', 'locl', 'mark', 'mkmk', 'numr', 'rclt', 'rlig', 'rvrn', and all features required for script shaping are preserved. To see the full list, try '--layout-features=?'. Use '*' to keep all features. Multiple --layout-features options can be provided if necessary. Examples: --layout-features+=onum,pnum,ss01 * Keep the default set of features and 'onum', 'pnum', 'ss01'. --layout-features-='mark','mkmk' * Keep the default set of features but drop 'mark' and 'mkmk'. --layout-features='kern' * Only keep the 'kern' feature, drop all others. --layout-features='' * Drop all features. --layout-features='*' * Keep all features. --layout-features+=aalt --layout-features-=vrt2 * Keep default set of features plus 'aalt', but drop 'vrt2'. Hinting options: --hinting Keep hinting [default] --no-hinting Drop glyph-specific hinting and font-wide hinting tables, as well as remove hinting-related bits and pieces from other tables (eg. GPOS). See --hinting-tables for list of tables that are dropped by default. Instructions and hints are stripped from 'glyf' and 'CFF ' tables respectively. This produces (sometimes up to 30%) smaller fonts that are suitable for extremely high-resolution systems, like high-end mobile devices and retina displays. Optimization options: --desubroutinize Remove CFF use of subroutinizes. Subroutinization is a way to make CFF fonts smaller. For small subsets however, desubroutinizing might make the font smaller. It has even been reported that desubroutinized CFF fonts compress better (produce smaller output) WOFF and WOFF2 fonts. Also see note under --no-hinting. --no-desubroutinize [default] Leave CFF subroutinizes as is, only throw away unused subroutinizes. Font table options: --drop-tables[+|-]=[,
...] Specify (=), add to (+=) or exclude from (-=) the comma-separated set of tables that will be be dropped. By default, the following tables are dropped: 'BASE', 'JSTF', 'DSIG', 'EBDT', 'EBLC', 'EBSC', 'SVG ', 'PCLT', 'LTSH' and Graphite tables: 'Feat', 'Glat', 'Gloc', 'Silf', 'Sill' and color tables: 'CBLC', 'CBDT', 'sbix'. The tool will attempt to subset the remaining tables. Examples: --drop-tables-='SVG ' * Drop the default set of tables but keep 'SVG '. --drop-tables+=GSUB * Drop the default set of tables and 'GSUB'. --drop-tables=DSIG * Only drop the 'DSIG' table, keep all others. --drop-tables= * Keep all tables. --no-subset-tables+=
[,
...] Add to the set of tables that will not be subsetted. By default, the following tables are included in this list, as they do not need subsetting (ignore the fact that 'loca' is listed here): 'gasp', 'head', 'hhea', 'maxp', 'vhea', 'OS/2', 'loca', 'name', 'cvt ', 'fpgm', 'prep', 'VMDX', 'DSIG', 'CPAL', 'MVAR', 'STAT'. By default, tables that the tool does not know how to subset and are not specified here will be dropped from the font, unless --passthrough-tables option is passed. Example: --no-subset-tables+=FFTM * Keep 'FFTM' table in the font by preventing subsetting. --passthrough-tables Do not drop tables that the tool does not know how to subset. --no-passthrough-tables Tables that the tool does not know how to subset and are not specified in --no-subset-tables will be dropped from the font. [default] --hinting-tables[-]=
[,
...] Specify (=), add to (+=) or exclude from (-=) the list of font-wide hinting tables that will be dropped if --no-hinting is specified, Examples: --hinting-tables-='VDMX' * Drop font-wide hinting tables except 'VDMX'. --hinting-tables='' * Keep all font-wide hinting tables (but strip hints from glyphs). --legacy-kern Keep TrueType 'kern' table even when OpenType 'GPOS' is available. --no-legacy-kern Drop TrueType 'kern' table if OpenType 'GPOS' is available. [default] Font naming options: These options control what is retained in the 'name' table. For numerical codes, see: http://www.microsoft.com/typography/otspec/name.htm --name-IDs[+|-]=[,...] Specify (=), add to (+=) or exclude from (-=) the set of 'name' table entry nameIDs that will be preserved. By default only nameID 1 (Family) and nameID 2 (Style) are preserved. Use '*' to keep all entries. Examples: --name-IDs+=0,4,6 * Also keep Copyright, Full name and PostScript name entry. --name-IDs='' * Drop all 'name' table entries. --name-IDs='*' * keep all 'name' table entries --name-legacy Keep legacy (non-Unicode) 'name' table entries (0.x, 1.x etc.). XXX Note: This might be needed for some fonts that have no Unicode name entires for English. See: https://github.com/behdad/fonttools/issues/146 --no-name-legacy Drop legacy (non-Unicode) 'name' table entries [default] --name-languages[+|-]=[,] Specify (=), add to (+=) or exclude from (-=) the set of 'name' table langIDs that will be preserved. By default only records with langID 0x0409 (English) are preserved. Use '*' to keep all langIDs. --obfuscate-names Make the font unusable as a system font by replacing name IDs 1, 2, 3, 4, and 6 with dummy strings (it is still fully functional as webfont). Glyph naming and encoding options: --glyph-names Keep PS glyph names in TT-flavored fonts. In general glyph names are not needed for correct use of the font. However, some PDF generators and PDF viewers might rely on glyph names to extract Unicode text from PDF documents. --no-glyph-names Drop PS glyph names in TT-flavored fonts, by using 'post' table version 3.0. [default] --legacy-cmap Keep the legacy 'cmap' subtables (0.x, 1.x, 4.x etc.). --no-legacy-cmap Drop the legacy 'cmap' subtables. [default] --symbol-cmap Keep the 3.0 symbol 'cmap'. --no-symbol-cmap Drop the 3.0 symbol 'cmap'. [default] Other font-specific options: --recalc-bounds Recalculate font bounding boxes. --no-recalc-bounds Keep original font bounding boxes. This is faster and still safe for all practical purposes. [default] --recalc-timestamp Set font 'modified' timestamp to current time. --no-recalc-timestamp Do not modify font 'modified' timestamp. [default] --canonical-order Order tables as recommended in the OpenType standard. This is not required by the standard, nor by any known implementation. --no-canonical-order Keep original order of font tables. This is faster. [default] --prune-unicode-ranges Update the 'OS/2 ulUnicodeRange*' bits after subsetting. The Unicode ranges defined in the OpenType specification v1.7 are intersected with the Unicode codepoints specified in the font's Unicode 'cmap' subtables: when no overlap is found, the bit will be switched off. However, it will *not* be switched on if an intersection is found. [default] --no-prune-unicode-ranges Don't change the 'OS/2 ulUnicodeRange*' bits. --recalc-average-width Update the 'OS/2 xAvgCharWidth' field after subsetting. --no-recalc-average-width Don't change the 'OS/2 xAvgCharWidth' field. [default] Application options: --verbose Display verbose information of the subsetting process. --timing Display detailed timing information of the subsetting process. --xml Display the TTX XML representation of subsetted font. Example: Produce a subset containing the characters ' !"#$%' without performing size-reducing optimizations: $ pyftsubset font.ttf --unicodes="U+0020-0025" \\ --layout-features='*' --glyph-names --symbol-cmap --legacy-cmap \\ --notdef-glyph --notdef-outline --recommended-glyphs \\ --name-IDs='*' --name-legacy --name-languages='*' """ log = logging.getLogger("fontTools.subset") def _log_glyphs(self, glyphs, font=None): self.info("Glyph names: %s", sorted(glyphs)) if font: reverseGlyphMap = font.getReverseGlyphMap() self.info("Glyph IDs: %s", sorted(reverseGlyphMap[g] for g in glyphs)) # bind "glyphs" function to 'log' object log.glyphs = MethodType(_log_glyphs, log) # I use a different timing channel so I can configure it separately from the # main module's logger timer = Timer(logger=logging.getLogger("fontTools.subset.timer")) def _add_method(*clazzes): """Returns a decorator function that adds a new method to one or more classes.""" def wrapper(method): done = [] for clazz in clazzes: if clazz in done: continue # Support multiple names of a clazz done.append(clazz) assert clazz.__name__ != 'DefaultTable', \ 'Oops, table class not found.' assert not hasattr(clazz, method.__name__), \ "Oops, class '%s' has method '%s'." % (clazz.__name__, method.__name__) setattr(clazz, method.__name__, method) return None return wrapper def _uniq_sort(l): return sorted(set(l)) def _set_update(s, *others): # Jython's set.update only takes one other argument. # Emulate real set.update... for other in others: s.update(other) def _dict_subset(d, glyphs): return {g:d[g] for g in glyphs} @_add_method(otTables.Coverage) def intersect(self, glyphs): """Returns ascending list of matching coverage values.""" return [i for i,g in enumerate(self.glyphs) if g in glyphs] @_add_method(otTables.Coverage) def intersect_glyphs(self, glyphs): """Returns set of intersecting glyphs.""" return set(g for g in self.glyphs if g in glyphs) @_add_method(otTables.Coverage) def subset(self, glyphs): """Returns ascending list of remaining coverage values.""" indices = self.intersect(glyphs) self.glyphs = [g for g in self.glyphs if g in glyphs] return indices @_add_method(otTables.Coverage) def remap(self, coverage_map): """Remaps coverage.""" self.glyphs = [self.glyphs[i] for i in coverage_map] @_add_method(otTables.ClassDef) def intersect(self, glyphs): """Returns ascending list of matching class values.""" return _uniq_sort( ([0] if any(g not in self.classDefs for g in glyphs) else []) + [v for g,v in self.classDefs.items() if g in glyphs]) @_add_method(otTables.ClassDef) def intersect_class(self, glyphs, klass): """Returns set of glyphs matching class.""" if klass == 0: return set(g for g in glyphs if g not in self.classDefs) return set(g for g,v in self.classDefs.items() if v == klass and g in glyphs) @_add_method(otTables.ClassDef) def subset(self, glyphs, remap=False): """Returns ascending list of remaining classes.""" self.classDefs = {g:v for g,v in self.classDefs.items() if g in glyphs} # Note: while class 0 has the special meaning of "not matched", # if no glyph will ever /not match/, we can optimize class 0 out too. indices = _uniq_sort( ([0] if any(g not in self.classDefs for g in glyphs) else []) + list(self.classDefs.values())) if remap: self.remap(indices) return indices @_add_method(otTables.ClassDef) def remap(self, class_map): """Remaps classes.""" self.classDefs = {g:class_map.index(v) for g,v in self.classDefs.items()} @_add_method(otTables.SingleSubst) def closure_glyphs(self, s, cur_glyphs): s.glyphs.update(v for g,v in self.mapping.items() if g in cur_glyphs) @_add_method(otTables.SingleSubst) def subset_glyphs(self, s): self.mapping = {g:v for g,v in self.mapping.items() if g in s.glyphs and v in s.glyphs} return bool(self.mapping) @_add_method(otTables.MultipleSubst) def closure_glyphs(self, s, cur_glyphs): for glyph, subst in self.mapping.items(): if glyph in cur_glyphs: _set_update(s.glyphs, subst) @_add_method(otTables.MultipleSubst) def subset_glyphs(self, s): self.mapping = {g:v for g,v in self.mapping.items() if g in s.glyphs and all(sub in s.glyphs for sub in v)} return bool(self.mapping) @_add_method(otTables.AlternateSubst) def closure_glyphs(self, s, cur_glyphs): _set_update(s.glyphs, *(vlist for g,vlist in self.alternates.items() if g in cur_glyphs)) @_add_method(otTables.AlternateSubst) def subset_glyphs(self, s): self.alternates = {g:vlist for g,vlist in self.alternates.items() if g in s.glyphs and all(v in s.glyphs for v in vlist)} return bool(self.alternates) @_add_method(otTables.LigatureSubst) def closure_glyphs(self, s, cur_glyphs): _set_update(s.glyphs, *([seq.LigGlyph for seq in seqs if all(c in s.glyphs for c in seq.Component)] for g,seqs in self.ligatures.items() if g in cur_glyphs)) @_add_method(otTables.LigatureSubst) def subset_glyphs(self, s): self.ligatures = {g:v for g,v in self.ligatures.items() if g in s.glyphs} self.ligatures = {g:[seq for seq in seqs if seq.LigGlyph in s.glyphs and all(c in s.glyphs for c in seq.Component)] for g,seqs in self.ligatures.items()} self.ligatures = {g:v for g,v in self.ligatures.items() if v} return bool(self.ligatures) @_add_method(otTables.ReverseChainSingleSubst) def closure_glyphs(self, s, cur_glyphs): if self.Format == 1: indices = self.Coverage.intersect(cur_glyphs) if(not indices or not all(c.intersect(s.glyphs) for c in self.LookAheadCoverage + self.BacktrackCoverage)): return s.glyphs.update(self.Substitute[i] for i in indices) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.ReverseChainSingleSubst) def subset_glyphs(self, s): if self.Format == 1: indices = self.Coverage.subset(s.glyphs) self.Substitute = [self.Substitute[i] for i in indices] # Now drop rules generating glyphs we don't want indices = [i for i,sub in enumerate(self.Substitute) if sub in s.glyphs] self.Substitute = [self.Substitute[i] for i in indices] self.Coverage.remap(indices) self.GlyphCount = len(self.Substitute) return bool(self.GlyphCount and all(c.subset(s.glyphs) for c in self.LookAheadCoverage+self.BacktrackCoverage)) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.SinglePos) def subset_glyphs(self, s): if self.Format == 1: return len(self.Coverage.subset(s.glyphs)) elif self.Format == 2: indices = self.Coverage.subset(s.glyphs) values = self.Value count = len(values) self.Value = [values[i] for i in indices if i < count] self.ValueCount = len(self.Value) return bool(self.ValueCount) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.SinglePos) def prune_post_subset(self, options): if not options.hinting: # Drop device tables self.ValueFormat &= ~0x00F0 return True @_add_method(otTables.PairPos) def subset_glyphs(self, s): if self.Format == 1: indices = self.Coverage.subset(s.glyphs) pairs = self.PairSet count = len(pairs) self.PairSet = [pairs[i] for i in indices if i < count] for p in self.PairSet: p.PairValueRecord = [r for r in p.PairValueRecord if r.SecondGlyph in s.glyphs] p.PairValueCount = len(p.PairValueRecord) # Remove empty pairsets indices = [i for i,p in enumerate(self.PairSet) if p.PairValueCount] self.Coverage.remap(indices) self.PairSet = [self.PairSet[i] for i in indices] self.PairSetCount = len(self.PairSet) return bool(self.PairSetCount) elif self.Format == 2: class1_map = [c for c in self.ClassDef1.subset(s.glyphs, remap=True) if c < self.Class1Count] class2_map = [c for c in self.ClassDef2.subset(s.glyphs, remap=True) if c < self.Class2Count] self.Class1Record = [self.Class1Record[i] for i in class1_map] for c in self.Class1Record: c.Class2Record = [c.Class2Record[i] for i in class2_map] self.Class1Count = len(class1_map) self.Class2Count = len(class2_map) return bool(self.Class1Count and self.Class2Count and self.Coverage.subset(s.glyphs)) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.PairPos) def prune_post_subset(self, options): if not options.hinting: # Drop device tables self.ValueFormat1 &= ~0x00F0 self.ValueFormat2 &= ~0x00F0 return True @_add_method(otTables.CursivePos) def subset_glyphs(self, s): if self.Format == 1: indices = self.Coverage.subset(s.glyphs) records = self.EntryExitRecord count = len(records) self.EntryExitRecord = [records[i] for i in indices if i < count] self.EntryExitCount = len(self.EntryExitRecord) return bool(self.EntryExitCount) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.Anchor) def prune_hints(self): # Drop device tables / contour anchor point self.ensureDecompiled() self.Format = 1 @_add_method(otTables.CursivePos) def prune_post_subset(self, options): if not options.hinting: for rec in self.EntryExitRecord: if rec.EntryAnchor: rec.EntryAnchor.prune_hints() if rec.ExitAnchor: rec.ExitAnchor.prune_hints() return True @_add_method(otTables.MarkBasePos) def subset_glyphs(self, s): if self.Format == 1: mark_indices = self.MarkCoverage.subset(s.glyphs) self.MarkArray.MarkRecord = [self.MarkArray.MarkRecord[i] for i in mark_indices] self.MarkArray.MarkCount = len(self.MarkArray.MarkRecord) base_indices = self.BaseCoverage.subset(s.glyphs) self.BaseArray.BaseRecord = [self.BaseArray.BaseRecord[i] for i in base_indices] self.BaseArray.BaseCount = len(self.BaseArray.BaseRecord) # Prune empty classes class_indices = _uniq_sort(v.Class for v in self.MarkArray.MarkRecord) self.ClassCount = len(class_indices) for m in self.MarkArray.MarkRecord: m.Class = class_indices.index(m.Class) for b in self.BaseArray.BaseRecord: b.BaseAnchor = [b.BaseAnchor[i] for i in class_indices] return bool(self.ClassCount and self.MarkArray.MarkCount and self.BaseArray.BaseCount) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.MarkBasePos) def prune_post_subset(self, options): if not options.hinting: for m in self.MarkArray.MarkRecord: if m.MarkAnchor: m.MarkAnchor.prune_hints() for b in self.BaseArray.BaseRecord: for a in b.BaseAnchor: if a: a.prune_hints() return True @_add_method(otTables.MarkLigPos) def subset_glyphs(self, s): if self.Format == 1: mark_indices = self.MarkCoverage.subset(s.glyphs) self.MarkArray.MarkRecord = [self.MarkArray.MarkRecord[i] for i in mark_indices] self.MarkArray.MarkCount = len(self.MarkArray.MarkRecord) ligature_indices = self.LigatureCoverage.subset(s.glyphs) self.LigatureArray.LigatureAttach = [self.LigatureArray.LigatureAttach[i] for i in ligature_indices] self.LigatureArray.LigatureCount = len(self.LigatureArray.LigatureAttach) # Prune empty classes class_indices = _uniq_sort(v.Class for v in self.MarkArray.MarkRecord) self.ClassCount = len(class_indices) for m in self.MarkArray.MarkRecord: m.Class = class_indices.index(m.Class) for l in self.LigatureArray.LigatureAttach: for c in l.ComponentRecord: c.LigatureAnchor = [c.LigatureAnchor[i] for i in class_indices] return bool(self.ClassCount and self.MarkArray.MarkCount and self.LigatureArray.LigatureCount) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.MarkLigPos) def prune_post_subset(self, options): if not options.hinting: for m in self.MarkArray.MarkRecord: if m.MarkAnchor: m.MarkAnchor.prune_hints() for l in self.LigatureArray.LigatureAttach: for c in l.ComponentRecord: for a in c.LigatureAnchor: if a: a.prune_hints() return True @_add_method(otTables.MarkMarkPos) def subset_glyphs(self, s): if self.Format == 1: mark1_indices = self.Mark1Coverage.subset(s.glyphs) self.Mark1Array.MarkRecord = [self.Mark1Array.MarkRecord[i] for i in mark1_indices] self.Mark1Array.MarkCount = len(self.Mark1Array.MarkRecord) mark2_indices = self.Mark2Coverage.subset(s.glyphs) self.Mark2Array.Mark2Record = [self.Mark2Array.Mark2Record[i] for i in mark2_indices] self.Mark2Array.MarkCount = len(self.Mark2Array.Mark2Record) # Prune empty classes class_indices = _uniq_sort(v.Class for v in self.Mark1Array.MarkRecord) self.ClassCount = len(class_indices) for m in self.Mark1Array.MarkRecord: m.Class = class_indices.index(m.Class) for b in self.Mark2Array.Mark2Record: b.Mark2Anchor = [b.Mark2Anchor[i] for i in class_indices] return bool(self.ClassCount and self.Mark1Array.MarkCount and self.Mark2Array.MarkCount) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.MarkMarkPos) def prune_post_subset(self, options): if not options.hinting: # Drop device tables or contour anchor point for m in self.Mark1Array.MarkRecord: if m.MarkAnchor: m.MarkAnchor.prune_hints() for b in self.Mark2Array.Mark2Record: for m in b.Mark2Anchor: if m: m.prune_hints() return True @_add_method(otTables.SingleSubst, otTables.MultipleSubst, otTables.AlternateSubst, otTables.LigatureSubst, otTables.ReverseChainSingleSubst, otTables.SinglePos, otTables.PairPos, otTables.CursivePos, otTables.MarkBasePos, otTables.MarkLigPos, otTables.MarkMarkPos) def subset_lookups(self, lookup_indices): pass @_add_method(otTables.SingleSubst, otTables.MultipleSubst, otTables.AlternateSubst, otTables.LigatureSubst, otTables.ReverseChainSingleSubst, otTables.SinglePos, otTables.PairPos, otTables.CursivePos, otTables.MarkBasePos, otTables.MarkLigPos, otTables.MarkMarkPos) def collect_lookups(self): return [] @_add_method(otTables.SingleSubst, otTables.MultipleSubst, otTables.AlternateSubst, otTables.LigatureSubst, otTables.ReverseChainSingleSubst, otTables.ContextSubst, otTables.ChainContextSubst, otTables.ContextPos, otTables.ChainContextPos) def prune_post_subset(self, options): return True @_add_method(otTables.SingleSubst, otTables.AlternateSubst, otTables.ReverseChainSingleSubst) def may_have_non_1to1(self): return False @_add_method(otTables.MultipleSubst, otTables.LigatureSubst, otTables.ContextSubst, otTables.ChainContextSubst) def may_have_non_1to1(self): return True @_add_method(otTables.ContextSubst, otTables.ChainContextSubst, otTables.ContextPos, otTables.ChainContextPos) def __subset_classify_context(self): class ContextHelper(object): def __init__(self, klass, Format): if klass.__name__.endswith('Subst'): Typ = 'Sub' Type = 'Subst' else: Typ = 'Pos' Type = 'Pos' if klass.__name__.startswith('Chain'): Chain = 'Chain' InputIdx = 1 DataLen = 3 else: Chain = '' InputIdx = 0 DataLen = 1 ChainTyp = Chain+Typ self.Typ = Typ self.Type = Type self.Chain = Chain self.ChainTyp = ChainTyp self.InputIdx = InputIdx self.DataLen = DataLen self.LookupRecord = Type+'LookupRecord' if Format == 1: Coverage = lambda r: r.Coverage ChainCoverage = lambda r: r.Coverage ContextData = lambda r:(None,) ChainContextData = lambda r:(None, None, None) SetContextData = None SetChainContextData = None RuleData = lambda r:(r.Input,) ChainRuleData = lambda r:(r.Backtrack, r.Input, r.LookAhead) def SetRuleData(r, d): (r.Input,) = d (r.GlyphCount,) = (len(x)+1 for x in d) def ChainSetRuleData(r, d): (r.Backtrack, r.Input, r.LookAhead) = d (r.BacktrackGlyphCount,r.InputGlyphCount,r.LookAheadGlyphCount,) = (len(d[0]),len(d[1])+1,len(d[2])) elif Format == 2: Coverage = lambda r: r.Coverage ChainCoverage = lambda r: r.Coverage ContextData = lambda r:(r.ClassDef,) ChainContextData = lambda r:(r.BacktrackClassDef, r.InputClassDef, r.LookAheadClassDef) def SetContextData(r, d): (r.ClassDef,) = d def SetChainContextData(r, d): (r.BacktrackClassDef, r.InputClassDef, r.LookAheadClassDef) = d RuleData = lambda r:(r.Class,) ChainRuleData = lambda r:(r.Backtrack, r.Input, r.LookAhead) def SetRuleData(r, d): (r.Class,) = d (r.GlyphCount,) = (len(x)+1 for x in d) def ChainSetRuleData(r, d): (r.Backtrack, r.Input, r.LookAhead) = d (r.BacktrackGlyphCount,r.InputGlyphCount,r.LookAheadGlyphCount,) = (len(d[0]),len(d[1])+1,len(d[2])) elif Format == 3: Coverage = lambda r: r.Coverage[0] ChainCoverage = lambda r: r.InputCoverage[0] ContextData = None ChainContextData = None SetContextData = None SetChainContextData = None RuleData = lambda r: r.Coverage ChainRuleData = lambda r:(r.BacktrackCoverage + r.InputCoverage + r.LookAheadCoverage) def SetRuleData(r, d): (r.Coverage,) = d (r.GlyphCount,) = (len(x) for x in d) def ChainSetRuleData(r, d): (r.BacktrackCoverage, r.InputCoverage, r.LookAheadCoverage) = d (r.BacktrackGlyphCount,r.InputGlyphCount,r.LookAheadGlyphCount,) = (len(x) for x in d) else: assert 0, "unknown format: %s" % Format if Chain: self.Coverage = ChainCoverage self.ContextData = ChainContextData self.SetContextData = SetChainContextData self.RuleData = ChainRuleData self.SetRuleData = ChainSetRuleData else: self.Coverage = Coverage self.ContextData = ContextData self.SetContextData = SetContextData self.RuleData = RuleData self.SetRuleData = SetRuleData if Format == 1: self.Rule = ChainTyp+'Rule' self.RuleCount = ChainTyp+'RuleCount' self.RuleSet = ChainTyp+'RuleSet' self.RuleSetCount = ChainTyp+'RuleSetCount' self.Intersect = lambda glyphs, c, r: [r] if r in glyphs else [] elif Format == 2: self.Rule = ChainTyp+'ClassRule' self.RuleCount = ChainTyp+'ClassRuleCount' self.RuleSet = ChainTyp+'ClassSet' self.RuleSetCount = ChainTyp+'ClassSetCount' self.Intersect = lambda glyphs, c, r: (c.intersect_class(glyphs, r) if c else (set(glyphs) if r == 0 else set())) self.ClassDef = 'InputClassDef' if Chain else 'ClassDef' self.ClassDefIndex = 1 if Chain else 0 self.Input = 'Input' if Chain else 'Class' if self.Format not in [1, 2, 3]: return None # Don't shoot the messenger; let it go if not hasattr(self.__class__, "__ContextHelpers"): self.__class__.__ContextHelpers = {} if self.Format not in self.__class__.__ContextHelpers: helper = ContextHelper(self.__class__, self.Format) self.__class__.__ContextHelpers[self.Format] = helper return self.__class__.__ContextHelpers[self.Format] @_add_method(otTables.ContextSubst, otTables.ChainContextSubst) def closure_glyphs(self, s, cur_glyphs): c = self.__subset_classify_context() indices = c.Coverage(self).intersect(cur_glyphs) if not indices: return [] cur_glyphs = c.Coverage(self).intersect_glyphs(cur_glyphs) if self.Format == 1: ContextData = c.ContextData(self) rss = getattr(self, c.RuleSet) rssCount = getattr(self, c.RuleSetCount) for i in indices: if i >= rssCount or not rss[i]: continue for r in getattr(rss[i], c.Rule): if not r: continue if not all(all(c.Intersect(s.glyphs, cd, k) for k in klist) for cd,klist in zip(ContextData, c.RuleData(r))): continue chaos = set() for ll in getattr(r, c.LookupRecord): if not ll: continue seqi = ll.SequenceIndex if seqi in chaos: # TODO Can we improve this? pos_glyphs = None else: if seqi == 0: pos_glyphs = frozenset([c.Coverage(self).glyphs[i]]) else: pos_glyphs = frozenset([r.Input[seqi - 1]]) lookup = s.table.LookupList.Lookup[ll.LookupListIndex] chaos.add(seqi) if lookup.may_have_non_1to1(): chaos.update(range(seqi, len(r.Input)+2)) lookup.closure_glyphs(s, cur_glyphs=pos_glyphs) elif self.Format == 2: ClassDef = getattr(self, c.ClassDef) indices = ClassDef.intersect(cur_glyphs) ContextData = c.ContextData(self) rss = getattr(self, c.RuleSet) rssCount = getattr(self, c.RuleSetCount) for i in indices: if i >= rssCount or not rss[i]: continue for r in getattr(rss[i], c.Rule): if not r: continue if not all(all(c.Intersect(s.glyphs, cd, k) for k in klist) for cd,klist in zip(ContextData, c.RuleData(r))): continue chaos = set() for ll in getattr(r, c.LookupRecord): if not ll: continue seqi = ll.SequenceIndex if seqi in chaos: # TODO Can we improve this? pos_glyphs = None else: if seqi == 0: pos_glyphs = frozenset(ClassDef.intersect_class(cur_glyphs, i)) else: pos_glyphs = frozenset(ClassDef.intersect_class(s.glyphs, getattr(r, c.Input)[seqi - 1])) lookup = s.table.LookupList.Lookup[ll.LookupListIndex] chaos.add(seqi) if lookup.may_have_non_1to1(): chaos.update(range(seqi, len(getattr(r, c.Input))+2)) lookup.closure_glyphs(s, cur_glyphs=pos_glyphs) elif self.Format == 3: if not all(x.intersect(s.glyphs) for x in c.RuleData(self)): return [] r = self chaos = set() for ll in getattr(r, c.LookupRecord): if not ll: continue seqi = ll.SequenceIndex if seqi in chaos: # TODO Can we improve this? pos_glyphs = None else: if seqi == 0: pos_glyphs = frozenset(cur_glyphs) else: pos_glyphs = frozenset(r.InputCoverage[seqi].intersect_glyphs(s.glyphs)) lookup = s.table.LookupList.Lookup[ll.LookupListIndex] chaos.add(seqi) if lookup.may_have_non_1to1(): chaos.update(range(seqi, len(r.InputCoverage)+1)) lookup.closure_glyphs(s, cur_glyphs=pos_glyphs) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.ContextSubst, otTables.ContextPos, otTables.ChainContextSubst, otTables.ChainContextPos) def subset_glyphs(self, s): c = self.__subset_classify_context() if self.Format == 1: indices = self.Coverage.subset(s.glyphs) rss = getattr(self, c.RuleSet) rssCount = getattr(self, c.RuleSetCount) rss = [rss[i] for i in indices if i < rssCount] for rs in rss: if not rs: continue ss = getattr(rs, c.Rule) ss = [r for r in ss if r and all(all(g in s.glyphs for g in glist) for glist in c.RuleData(r))] setattr(rs, c.Rule, ss) setattr(rs, c.RuleCount, len(ss)) # Prune empty rulesets indices = [i for i,rs in enumerate(rss) if rs and getattr(rs, c.Rule)] self.Coverage.remap(indices) rss = [rss[i] for i in indices] setattr(self, c.RuleSet, rss) setattr(self, c.RuleSetCount, len(rss)) return bool(rss) elif self.Format == 2: if not self.Coverage.subset(s.glyphs): return False ContextData = c.ContextData(self) klass_maps = [x.subset(s.glyphs, remap=True) if x else None for x in ContextData] # Keep rulesets for class numbers that survived. indices = klass_maps[c.ClassDefIndex] rss = getattr(self, c.RuleSet) rssCount = getattr(self, c.RuleSetCount) rss = [rss[i] for i in indices if i < rssCount] del rssCount # Delete, but not renumber, unreachable rulesets. indices = getattr(self, c.ClassDef).intersect(self.Coverage.glyphs) rss = [rss if i in indices else None for i,rss in enumerate(rss)] for rs in rss: if not rs: continue ss = getattr(rs, c.Rule) ss = [r for r in ss if r and all(all(k in klass_map for k in klist) for klass_map,klist in zip(klass_maps, c.RuleData(r)))] setattr(rs, c.Rule, ss) setattr(rs, c.RuleCount, len(ss)) # Remap rule classes for r in ss: c.SetRuleData(r, [[klass_map.index(k) for k in klist] for klass_map,klist in zip(klass_maps, c.RuleData(r))]) # Prune empty rulesets rss = [rs if rs and getattr(rs, c.Rule) else None for rs in rss] while rss and rss[-1] is None: del rss[-1] setattr(self, c.RuleSet, rss) setattr(self, c.RuleSetCount, len(rss)) # TODO: We can do a second round of remapping class values based # on classes that are actually used in at least one rule. Right # now we subset classes to c.glyphs only. Or better, rewrite # the above to do that. return bool(rss) elif self.Format == 3: return all(x.subset(s.glyphs) for x in c.RuleData(self)) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.ContextSubst, otTables.ChainContextSubst, otTables.ContextPos, otTables.ChainContextPos) def subset_lookups(self, lookup_indices): c = self.__subset_classify_context() if self.Format in [1, 2]: for rs in getattr(self, c.RuleSet): if not rs: continue for r in getattr(rs, c.Rule): if not r: continue setattr(r, c.LookupRecord, [ll for ll in getattr(r, c.LookupRecord) if ll and ll.LookupListIndex in lookup_indices]) for ll in getattr(r, c.LookupRecord): if not ll: continue ll.LookupListIndex = lookup_indices.index(ll.LookupListIndex) elif self.Format == 3: setattr(self, c.LookupRecord, [ll for ll in getattr(self, c.LookupRecord) if ll and ll.LookupListIndex in lookup_indices]) for ll in getattr(self, c.LookupRecord): if not ll: continue ll.LookupListIndex = lookup_indices.index(ll.LookupListIndex) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.ContextSubst, otTables.ChainContextSubst, otTables.ContextPos, otTables.ChainContextPos) def collect_lookups(self): c = self.__subset_classify_context() if self.Format in [1, 2]: return [ll.LookupListIndex for rs in getattr(self, c.RuleSet) if rs for r in getattr(rs, c.Rule) if r for ll in getattr(r, c.LookupRecord) if ll] elif self.Format == 3: return [ll.LookupListIndex for ll in getattr(self, c.LookupRecord) if ll] else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.ExtensionSubst) def closure_glyphs(self, s, cur_glyphs): if self.Format == 1: self.ExtSubTable.closure_glyphs(s, cur_glyphs) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.ExtensionSubst) def may_have_non_1to1(self): if self.Format == 1: return self.ExtSubTable.may_have_non_1to1() else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.ExtensionSubst, otTables.ExtensionPos) def subset_glyphs(self, s): if self.Format == 1: return self.ExtSubTable.subset_glyphs(s) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.ExtensionSubst, otTables.ExtensionPos) def prune_post_subset(self, options): if self.Format == 1: return self.ExtSubTable.prune_post_subset(options) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.ExtensionSubst, otTables.ExtensionPos) def subset_lookups(self, lookup_indices): if self.Format == 1: return self.ExtSubTable.subset_lookups(lookup_indices) else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.ExtensionSubst, otTables.ExtensionPos) def collect_lookups(self): if self.Format == 1: return self.ExtSubTable.collect_lookups() else: assert 0, "unknown format: %s" % self.Format @_add_method(otTables.Lookup) def closure_glyphs(self, s, cur_glyphs=None): if cur_glyphs is None: cur_glyphs = frozenset(s.glyphs) # Memoize if (id(self), cur_glyphs) in s._doneLookups: return s._doneLookups.add((id(self), cur_glyphs)) if self in s._activeLookups: raise Exception("Circular loop in lookup recursion") s._activeLookups.append(self) for st in self.SubTable: if not st: continue st.closure_glyphs(s, cur_glyphs) assert(s._activeLookups[-1] == self) del s._activeLookups[-1] @_add_method(otTables.Lookup) def subset_glyphs(self, s): self.SubTable = [st for st in self.SubTable if st and st.subset_glyphs(s)] self.SubTableCount = len(self.SubTable) return bool(self.SubTableCount) @_add_method(otTables.Lookup) def prune_post_subset(self, options): ret = False for st in self.SubTable: if not st: continue if st.prune_post_subset(options): ret = True return ret @_add_method(otTables.Lookup) def subset_lookups(self, lookup_indices): for s in self.SubTable: s.subset_lookups(lookup_indices) @_add_method(otTables.Lookup) def collect_lookups(self): return sum((st.collect_lookups() for st in self.SubTable if st), []) @_add_method(otTables.Lookup) def may_have_non_1to1(self): return any(st.may_have_non_1to1() for st in self.SubTable if st) @_add_method(otTables.LookupList) def subset_glyphs(self, s): """Returns the indices of nonempty lookups.""" return [i for i,l in enumerate(self.Lookup) if l and l.subset_glyphs(s)] @_add_method(otTables.LookupList) def prune_post_subset(self, options): ret = False for l in self.Lookup: if not l: continue if l.prune_post_subset(options): ret = True return ret @_add_method(otTables.LookupList) def subset_lookups(self, lookup_indices): self.ensureDecompiled() self.Lookup = [self.Lookup[i] for i in lookup_indices if i < self.LookupCount] self.LookupCount = len(self.Lookup) for l in self.Lookup: l.subset_lookups(lookup_indices) @_add_method(otTables.LookupList) def neuter_lookups(self, lookup_indices): """Sets lookups not in lookup_indices to None.""" self.ensureDecompiled() self.Lookup = [l if i in lookup_indices else None for i,l in enumerate(self.Lookup)] @_add_method(otTables.LookupList) def closure_lookups(self, lookup_indices): """Returns sorted index of all lookups reachable from lookup_indices.""" lookup_indices = _uniq_sort(lookup_indices) recurse = lookup_indices while True: recurse_lookups = sum((self.Lookup[i].collect_lookups() for i in recurse if i < self.LookupCount), []) recurse_lookups = [l for l in recurse_lookups if l not in lookup_indices and l < self.LookupCount] if not recurse_lookups: return _uniq_sort(lookup_indices) recurse_lookups = _uniq_sort(recurse_lookups) lookup_indices.extend(recurse_lookups) recurse = recurse_lookups @_add_method(otTables.Feature) def subset_lookups(self, lookup_indices): """"Returns True if feature is non-empty afterwards.""" self.LookupListIndex = [l for l in self.LookupListIndex if l in lookup_indices] # Now map them. self.LookupListIndex = [lookup_indices.index(l) for l in self.LookupListIndex] self.LookupCount = len(self.LookupListIndex) return self.LookupCount or self.FeatureParams @_add_method(otTables.FeatureList) def subset_lookups(self, lookup_indices): """Returns the indices of nonempty features.""" # Note: Never ever drop feature 'pref', even if it's empty. # HarfBuzz chooses shaper for Khmer based on presence of this # feature. See thread at: # http://lists.freedesktop.org/archives/harfbuzz/2012-November/002660.html return [i for i,f in enumerate(self.FeatureRecord) if (f.Feature.subset_lookups(lookup_indices) or f.FeatureTag == 'pref')] @_add_method(otTables.FeatureList) def collect_lookups(self, feature_indices): return sum((self.FeatureRecord[i].Feature.LookupListIndex for i in feature_indices if i < self.FeatureCount), []) @_add_method(otTables.FeatureList) def subset_features(self, feature_indices): self.ensureDecompiled() self.FeatureRecord = [self.FeatureRecord[i] for i in feature_indices] self.FeatureCount = len(self.FeatureRecord) return bool(self.FeatureCount) @_add_method(otTables.FeatureTableSubstitution) def subset_lookups(self, lookup_indices): """Returns the indices of nonempty features.""" return [r.FeatureIndex for r in self.SubstitutionRecord if r.Feature.subset_lookups(lookup_indices)] @_add_method(otTables.FeatureVariations) def subset_lookups(self, lookup_indices): """Returns the indices of nonempty features.""" return sum((f.FeatureTableSubstitution.subset_lookups(lookup_indices) for f in self.FeatureVariationRecord), []) @_add_method(otTables.FeatureVariations) def collect_lookups(self, feature_indices): return sum((r.Feature.LookupListIndex for vr in self.FeatureVariationRecord for r in vr.FeatureTableSubstitution.SubstitutionRecord if r.FeatureIndex in feature_indices), []) @_add_method(otTables.FeatureTableSubstitution) def subset_features(self, feature_indices): self.ensureDecompiled() self.SubstitutionRecord = [r for r in self.SubstitutionRecord if r.FeatureIndex in feature_indices] self.SubstitutionCount = len(self.SubstitutionRecord) return bool(self.SubstitutionCount) @_add_method(otTables.FeatureVariations) def subset_features(self, feature_indices): self.ensureDecompiled() self.FeaturVariationRecord = [r for r in self.FeatureVariationRecord if r.FeatureTableSubstitution.subset_features(feature_indices)] self.FeatureVariationCount = len(self.FeatureVariationRecord) return bool(self.FeatureVariationCount) @_add_method(otTables.DefaultLangSys, otTables.LangSys) def subset_features(self, feature_indices): if self.ReqFeatureIndex in feature_indices: self.ReqFeatureIndex = feature_indices.index(self.ReqFeatureIndex) else: self.ReqFeatureIndex = 65535 self.FeatureIndex = [f for f in self.FeatureIndex if f in feature_indices] # Now map them. self.FeatureIndex = [feature_indices.index(f) for f in self.FeatureIndex if f in feature_indices] self.FeatureCount = len(self.FeatureIndex) return bool(self.FeatureCount or self.ReqFeatureIndex != 65535) @_add_method(otTables.DefaultLangSys, otTables.LangSys) def collect_features(self): feature_indices = self.FeatureIndex[:] if self.ReqFeatureIndex != 65535: feature_indices.append(self.ReqFeatureIndex) return _uniq_sort(feature_indices) @_add_method(otTables.Script) def subset_features(self, feature_indices, keepEmptyDefaultLangSys=False): if(self.DefaultLangSys and not self.DefaultLangSys.subset_features(feature_indices) and not keepEmptyDefaultLangSys): self.DefaultLangSys = None self.LangSysRecord = [l for l in self.LangSysRecord if l.LangSys.subset_features(feature_indices)] self.LangSysCount = len(self.LangSysRecord) return bool(self.LangSysCount or self.DefaultLangSys) @_add_method(otTables.Script) def collect_features(self): feature_indices = [l.LangSys.collect_features() for l in self.LangSysRecord] if self.DefaultLangSys: feature_indices.append(self.DefaultLangSys.collect_features()) return _uniq_sort(sum(feature_indices, [])) @_add_method(otTables.ScriptList) def subset_features(self, feature_indices, retain_empty): # https://bugzilla.mozilla.org/show_bug.cgi?id=1331737#c32 self.ScriptRecord = [s for s in self.ScriptRecord if s.Script.subset_features(feature_indices, s.ScriptTag=='DFLT') or retain_empty] self.ScriptCount = len(self.ScriptRecord) return bool(self.ScriptCount) @_add_method(otTables.ScriptList) def collect_features(self): return _uniq_sort(sum((s.Script.collect_features() for s in self.ScriptRecord), [])) # CBLC will inherit it @_add_method(ttLib.getTableClass('EBLC')) def subset_glyphs(self, s): for strike in self.strikes: for indexSubTable in strike.indexSubTables: indexSubTable.names = [n for n in indexSubTable.names if n in s.glyphs] strike.indexSubTables = [i for i in strike.indexSubTables if i.names] self.strikes = [s for s in self.strikes if s.indexSubTables] return True # CBDC will inherit it @_add_method(ttLib.getTableClass('EBDT')) def subset_glyphs(self, s): self.strikeData = [{g: strike[g] for g in s.glyphs if g in strike} for strike in self.strikeData] return True @_add_method(ttLib.getTableClass('GSUB')) def closure_glyphs(self, s): s.table = self.table if self.table.ScriptList: feature_indices = self.table.ScriptList.collect_features() else: feature_indices = [] if self.table.FeatureList: lookup_indices = self.table.FeatureList.collect_lookups(feature_indices) else: lookup_indices = [] if getattr(self.table, 'FeatureVariations', None): lookup_indices += self.table.FeatureVariations.collect_lookups(feature_indices) lookup_indices = _uniq_sort(lookup_indices) if self.table.LookupList: while True: orig_glyphs = frozenset(s.glyphs) s._activeLookups = [] s._doneLookups = set() for i in lookup_indices: if i >= self.table.LookupList.LookupCount: continue if not self.table.LookupList.Lookup[i]: continue self.table.LookupList.Lookup[i].closure_glyphs(s) del s._activeLookups, s._doneLookups if orig_glyphs == s.glyphs: break del s.table @_add_method(ttLib.getTableClass('GSUB'), ttLib.getTableClass('GPOS')) def subset_glyphs(self, s): s.glyphs = s.glyphs_gsubed if self.table.LookupList: lookup_indices = self.table.LookupList.subset_glyphs(s) else: lookup_indices = [] self.subset_lookups(lookup_indices) return True @_add_method(ttLib.getTableClass('GSUB'), ttLib.getTableClass('GPOS')) def retain_empty_scripts(self): # https://github.com/behdad/fonttools/issues/518 # https://bugzilla.mozilla.org/show_bug.cgi?id=1080739#c15 return self.__class__ == ttLib.getTableClass('GSUB') @_add_method(ttLib.getTableClass('GSUB'), ttLib.getTableClass('GPOS')) def subset_lookups(self, lookup_indices): """Retains specified lookups, then removes empty features, language systems, and scripts.""" if self.table.LookupList: self.table.LookupList.subset_lookups(lookup_indices) if self.table.FeatureList: feature_indices = self.table.FeatureList.subset_lookups(lookup_indices) else: feature_indices = [] if getattr(self.table, 'FeatureVariations', None): feature_indices += self.table.FeatureVariations.subset_lookups(lookup_indices) feature_indices = _uniq_sort(feature_indices) if self.table.FeatureList: self.table.FeatureList.subset_features(feature_indices) if getattr(self.table, 'FeatureVariations', None): self.table.FeatureVariations.subset_features(feature_indices) if self.table.ScriptList: self.table.ScriptList.subset_features(feature_indices, self.retain_empty_scripts()) @_add_method(ttLib.getTableClass('GSUB'), ttLib.getTableClass('GPOS')) def neuter_lookups(self, lookup_indices): """Sets lookups not in lookup_indices to None.""" if self.table.LookupList: self.table.LookupList.neuter_lookups(lookup_indices) @_add_method(ttLib.getTableClass('GSUB'), ttLib.getTableClass('GPOS')) def prune_lookups(self, remap=True): """Remove (default) or neuter unreferenced lookups""" if self.table.ScriptList: feature_indices = self.table.ScriptList.collect_features() else: feature_indices = [] if self.table.FeatureList: lookup_indices = self.table.FeatureList.collect_lookups(feature_indices) else: lookup_indices = [] if getattr(self.table, 'FeatureVariations', None): lookup_indices += self.table.FeatureVariations.collect_lookups(feature_indices) lookup_indices = _uniq_sort(lookup_indices) if self.table.LookupList: lookup_indices = self.table.LookupList.closure_lookups(lookup_indices) else: lookup_indices = [] if remap: self.subset_lookups(lookup_indices) else: self.neuter_lookups(lookup_indices) @_add_method(ttLib.getTableClass('GSUB'), ttLib.getTableClass('GPOS')) def subset_feature_tags(self, feature_tags): if self.table.FeatureList: feature_indices = \ [i for i,f in enumerate(self.table.FeatureList.FeatureRecord) if f.FeatureTag in feature_tags] self.table.FeatureList.subset_features(feature_indices) if getattr(self.table, 'FeatureVariations', None): self.table.FeatureVariations.subset_features(feature_indices) else: feature_indices = [] if self.table.ScriptList: self.table.ScriptList.subset_features(feature_indices, self.retain_empty_scripts()) @_add_method(ttLib.getTableClass('GSUB'), ttLib.getTableClass('GPOS')) def prune_features(self): """Remove unreferenced features""" if self.table.ScriptList: feature_indices = self.table.ScriptList.collect_features() else: feature_indices = [] if self.table.FeatureList: self.table.FeatureList.subset_features(feature_indices) if getattr(self.table, 'FeatureVariations', None): self.table.FeatureVariations.subset_features(feature_indices) if self.table.ScriptList: self.table.ScriptList.subset_features(feature_indices, self.retain_empty_scripts()) @_add_method(ttLib.getTableClass('GSUB'), ttLib.getTableClass('GPOS')) def prune_pre_subset(self, font, options): # Drop undesired features if '*' not in options.layout_features: self.subset_feature_tags(options.layout_features) # Neuter unreferenced lookups self.prune_lookups(remap=False) return True @_add_method(ttLib.getTableClass('GSUB'), ttLib.getTableClass('GPOS')) def remove_redundant_langsys(self): table = self.table if not table.ScriptList or not table.FeatureList: return features = table.FeatureList.FeatureRecord for s in table.ScriptList.ScriptRecord: d = s.Script.DefaultLangSys if not d: continue for lr in s.Script.LangSysRecord[:]: l = lr.LangSys # Compare d and l if len(d.FeatureIndex) != len(l.FeatureIndex): continue if (d.ReqFeatureIndex == 65535) != (l.ReqFeatureIndex == 65535): continue if d.ReqFeatureIndex != 65535: if features[d.ReqFeatureIndex] != features[l.ReqFeatureIndex]: continue for i in range(len(d.FeatureIndex)): if features[d.FeatureIndex[i]] != features[l.FeatureIndex[i]]: break else: # LangSys and default are equal; delete LangSys s.Script.LangSysRecord.remove(lr) @_add_method(ttLib.getTableClass('GSUB'), ttLib.getTableClass('GPOS')) def prune_post_subset(self, options): table = self.table self.prune_lookups() # XXX Is this actually needed?! if table.LookupList: table.LookupList.prune_post_subset(options) # XXX Next two lines disabled because OTS is stupid and # doesn't like NULL offsets here. #if not table.LookupList.Lookup: # table.LookupList = None if not table.LookupList: table.FeatureList = None if table.FeatureList: self.remove_redundant_langsys() # Remove unreferenced features self.prune_features() # XXX Next two lines disabled because OTS is stupid and # doesn't like NULL offsets here. #if table.FeatureList and not table.FeatureList.FeatureRecord: # table.FeatureList = None # Never drop scripts themselves as them just being available # holds semantic significance. # XXX Next two lines disabled because OTS is stupid and # doesn't like NULL offsets here. #if table.ScriptList and not table.ScriptList.ScriptRecord: # table.ScriptList = None if not table.FeatureList and hasattr(table, 'FeatureVariations'): table.FeatureVariations = None if hasattr(table, 'FeatureVariations') and not table.FeatureVariations: if table.Version == 0x00010001: table.Version = 0x00010000 return True @_add_method(ttLib.getTableClass('GDEF')) def subset_glyphs(self, s): glyphs = s.glyphs_gsubed table = self.table if table.LigCaretList: indices = table.LigCaretList.Coverage.subset(glyphs) table.LigCaretList.LigGlyph = [table.LigCaretList.LigGlyph[i] for i in indices] table.LigCaretList.LigGlyphCount = len(table.LigCaretList.LigGlyph) if table.MarkAttachClassDef: table.MarkAttachClassDef.classDefs = \ {g:v for g,v in table.MarkAttachClassDef.classDefs.items() if g in glyphs} if table.GlyphClassDef: table.GlyphClassDef.classDefs = \ {g:v for g,v in table.GlyphClassDef.classDefs.items() if g in glyphs} if table.AttachList: indices = table.AttachList.Coverage.subset(glyphs) GlyphCount = table.AttachList.GlyphCount table.AttachList.AttachPoint = [table.AttachList.AttachPoint[i] for i in indices if i < GlyphCount] table.AttachList.GlyphCount = len(table.AttachList.AttachPoint) if hasattr(table, "MarkGlyphSetsDef") and table.MarkGlyphSetsDef: for coverage in table.MarkGlyphSetsDef.Coverage: coverage.subset(glyphs) # TODO: The following is disabled. If enabling, we need to go fixup all # lookups that use MarkFilteringSet and map their set. # indices = table.MarkGlyphSetsDef.Coverage = \ # [c for c in table.MarkGlyphSetsDef.Coverage if c.glyphs] return True @_add_method(ttLib.getTableClass('GDEF')) def prune_post_subset(self, options): table = self.table # XXX check these against OTS if table.LigCaretList and not table.LigCaretList.LigGlyphCount: table.LigCaretList = None if table.MarkAttachClassDef and not table.MarkAttachClassDef.classDefs: table.MarkAttachClassDef = None if table.GlyphClassDef and not table.GlyphClassDef.classDefs: table.GlyphClassDef = None if table.AttachList and not table.AttachList.GlyphCount: table.AttachList = None if hasattr(table, "VarStore"): if table.VarStore.VarDataCount == 0: if table.Version == 0x00010003: table.Version = 0x00010002 if (not hasattr(table, "MarkGlyphSetsDef") or not table.MarkGlyphSetsDef or not table.MarkGlyphSetsDef.Coverage): table.MarkGlyphSetsDef = None if table.Version == 0x00010002: table.Version = 0x00010000 return bool(table.LigCaretList or table.MarkAttachClassDef or table.GlyphClassDef or table.AttachList or (table.Version >= 0x00010002 and table.MarkGlyphSetsDef)) @_add_method(ttLib.getTableClass('kern')) def prune_pre_subset(self, font, options): # Prune unknown kern table types self.kernTables = [t for t in self.kernTables if hasattr(t, 'kernTable')] return bool(self.kernTables) @_add_method(ttLib.getTableClass('kern')) def subset_glyphs(self, s): glyphs = s.glyphs_gsubed for t in self.kernTables: t.kernTable = {(a,b):v for (a,b),v in t.kernTable.items() if a in glyphs and b in glyphs} self.kernTables = [t for t in self.kernTables if t.kernTable] return bool(self.kernTables) @_add_method(ttLib.getTableClass('vmtx')) def subset_glyphs(self, s): self.metrics = _dict_subset(self.metrics, s.glyphs) return bool(self.metrics) @_add_method(ttLib.getTableClass('hmtx')) def subset_glyphs(self, s): self.metrics = _dict_subset(self.metrics, s.glyphs) return True # Required table @_add_method(ttLib.getTableClass('hdmx')) def subset_glyphs(self, s): self.hdmx = {sz:_dict_subset(l, s.glyphs) for sz,l in self.hdmx.items()} return bool(self.hdmx) @_add_method(ttLib.getTableClass('ankr')) def subset_glyphs(self, s): table = self.table.AnchorPoints assert table.Format == 0, "unknown 'ankr' format %s" % table.Format table.Anchors = {glyph: table.Anchors[glyph] for glyph in s.glyphs if glyph in table.Anchors} return len(table.Anchors) > 0 @_add_method(ttLib.getTableClass('bsln')) def closure_glyphs(self, s): table = self.table.Baseline if table.Format in (2, 3): s.glyphs.add(table.StandardGlyph) @_add_method(ttLib.getTableClass('bsln')) def subset_glyphs(self, s): table = self.table.Baseline if table.Format in (1, 3): baselines = {glyph: table.BaselineValues.get(glyph, table.DefaultBaseline) for glyph in s.glyphs} if len(baselines) > 0: mostCommon, _cnt = Counter(baselines.values()).most_common(1)[0] table.DefaultBaseline = mostCommon baselines = {glyph: b for glyph, b in baselines.items() if b != mostCommon} if len(baselines) > 0: table.BaselineValues = baselines else: table.Format = {1: 0, 3: 2}[table.Format] del table.BaselineValues return True @_add_method(ttLib.getTableClass('lcar')) def subset_glyphs(self, s): table = self.table.LigatureCarets if table.Format in (0, 1): table.Carets = {glyph: table.Carets[glyph] for glyph in s.glyphs if glyph in table.Carets} return len(table.Carets) > 0 else: assert False, "unknown 'lcar' format %s" % table.Format @_add_method(ttLib.getTableClass('gvar')) def prune_pre_subset(self, font, options): if options.notdef_glyph and not options.notdef_outline: self.variations[font.glyphOrder[0]] = [] return True @_add_method(ttLib.getTableClass('gvar')) def subset_glyphs(self, s): self.variations = _dict_subset(self.variations, s.glyphs) self.glyphCount = len(self.variations) return bool(self.variations) @_add_method(ttLib.getTableClass('VORG')) def subset_glyphs(self, s): self.VOriginRecords = {g:v for g,v in self.VOriginRecords.items() if g in s.glyphs} self.numVertOriginYMetrics = len(self.VOriginRecords) return True # Never drop; has default metrics @_add_method(ttLib.getTableClass('opbd')) def subset_glyphs(self, s): table = self.table.OpticalBounds if table.Format == 0: table.OpticalBoundsDeltas = {glyph: table.OpticalBoundsDeltas[glyph] for glyph in s.glyphs if glyph in table.OpticalBoundsDeltas} return len(table.OpticalBoundsDeltas) > 0 elif table.Format == 1: table.OpticalBoundsPoints = {glyph: table.OpticalBoundsPoints[glyph] for glyph in s.glyphs if glyph in table.OpticalBoundsPoints} return len(table.OpticalBoundsPoints) > 0 else: assert False, "unknown 'opbd' format %s" % table.Format @_add_method(ttLib.getTableClass('post')) def prune_pre_subset(self, font, options): if not options.glyph_names: self.formatType = 3.0 return True # Required table @_add_method(ttLib.getTableClass('post')) def subset_glyphs(self, s): self.extraNames = [] # This seems to do it return True # Required table @_add_method(ttLib.getTableClass('prop')) def subset_glyphs(self, s): prop = self.table.GlyphProperties if prop.Format == 0: return prop.DefaultProperties != 0 elif prop.Format == 1: prop.Properties = {g: prop.Properties.get(g, prop.DefaultProperties) for g in s.glyphs} mostCommon, _cnt = Counter(prop.Properties.values()).most_common(1)[0] prop.DefaultProperties = mostCommon prop.Properties = {g: prop for g, prop in prop.Properties.items() if prop != mostCommon} if len(prop.Properties) == 0: del prop.Properties prop.Format = 0 return prop.DefaultProperties != 0 return True else: assert False, "unknown 'prop' format %s" % prop.Format @_add_method(ttLib.getTableClass('COLR')) def closure_glyphs(self, s): decompose = s.glyphs while decompose: layers = set() for g in decompose: for l in self.ColorLayers.get(g, []): layers.add(l.name) layers -= s.glyphs s.glyphs.update(layers) decompose = layers @_add_method(ttLib.getTableClass('COLR')) def subset_glyphs(self, s): self.ColorLayers = {g: self.ColorLayers[g] for g in s.glyphs if g in self.ColorLayers} return bool(self.ColorLayers) # TODO: prune unused palettes @_add_method(ttLib.getTableClass('CPAL')) def prune_post_subset(self, options): return True @_add_method(otTables.MathGlyphConstruction) def closure_glyphs(self, glyphs): variants = set() for v in self.MathGlyphVariantRecord: variants.add(v.VariantGlyph) if self.GlyphAssembly: for p in self.GlyphAssembly.PartRecords: variants.add(p.glyph) return variants @_add_method(otTables.MathVariants) def closure_glyphs(self, s): glyphs = frozenset(s.glyphs) variants = set() if self.VertGlyphCoverage: indices = self.VertGlyphCoverage.intersect(glyphs) for i in indices: variants.update(self.VertGlyphConstruction[i].closure_glyphs(glyphs)) if self.HorizGlyphCoverage: indices = self.HorizGlyphCoverage.intersect(glyphs) for i in indices: variants.update(self.HorizGlyphConstruction[i].closure_glyphs(glyphs)) s.glyphs.update(variants) @_add_method(ttLib.getTableClass('MATH')) def closure_glyphs(self, s): self.table.MathVariants.closure_glyphs(s) @_add_method(otTables.MathItalicsCorrectionInfo) def subset_glyphs(self, s): indices = self.Coverage.subset(s.glyphs) self.ItalicsCorrection = [self.ItalicsCorrection[i] for i in indices] self.ItalicsCorrectionCount = len(self.ItalicsCorrection) return bool(self.ItalicsCorrectionCount) @_add_method(otTables.MathTopAccentAttachment) def subset_glyphs(self, s): indices = self.TopAccentCoverage.subset(s.glyphs) self.TopAccentAttachment = [self.TopAccentAttachment[i] for i in indices] self.TopAccentAttachmentCount = len(self.TopAccentAttachment) return bool(self.TopAccentAttachmentCount) @_add_method(otTables.MathKernInfo) def subset_glyphs(self, s): indices = self.MathKernCoverage.subset(s.glyphs) self.MathKernInfoRecords = [self.MathKernInfoRecords[i] for i in indices] self.MathKernCount = len(self.MathKernInfoRecords) return bool(self.MathKernCount) @_add_method(otTables.MathGlyphInfo) def subset_glyphs(self, s): if self.MathItalicsCorrectionInfo: self.MathItalicsCorrectionInfo.subset_glyphs(s) if self.MathTopAccentAttachment: self.MathTopAccentAttachment.subset_glyphs(s) if self.MathKernInfo: self.MathKernInfo.subset_glyphs(s) if self.ExtendedShapeCoverage: self.ExtendedShapeCoverage.subset(s.glyphs) return True @_add_method(otTables.MathVariants) def subset_glyphs(self, s): if self.VertGlyphCoverage: indices = self.VertGlyphCoverage.subset(s.glyphs) self.VertGlyphConstruction = [self.VertGlyphConstruction[i] for i in indices] self.VertGlyphCount = len(self.VertGlyphConstruction) if self.HorizGlyphCoverage: indices = self.HorizGlyphCoverage.subset(s.glyphs) self.HorizGlyphConstruction = [self.HorizGlyphConstruction[i] for i in indices] self.HorizGlyphCount = len(self.HorizGlyphConstruction) return True @_add_method(ttLib.getTableClass('MATH')) def subset_glyphs(self, s): s.glyphs = s.glyphs_mathed self.table.MathGlyphInfo.subset_glyphs(s) self.table.MathVariants.subset_glyphs(s) return True @_add_method(ttLib.getTableModule('glyf').Glyph) def remapComponentsFast(self, indices): if not self.data or struct.unpack(">h", self.data[:2])[0] >= 0: return # Not composite data = array.array("B", self.data) i = 10 more = 1 while more: flags =(data[i] << 8) | data[i+1] glyphID =(data[i+2] << 8) | data[i+3] # Remap glyphID = indices.index(glyphID) data[i+2] = glyphID >> 8 data[i+3] = glyphID & 0xFF i += 4 flags = int(flags) if flags & 0x0001: i += 4 # ARG_1_AND_2_ARE_WORDS else: i += 2 if flags & 0x0008: i += 2 # WE_HAVE_A_SCALE elif flags & 0x0040: i += 4 # WE_HAVE_AN_X_AND_Y_SCALE elif flags & 0x0080: i += 8 # WE_HAVE_A_TWO_BY_TWO more = flags & 0x0020 # MORE_COMPONENTS self.data = data.tostring() @_add_method(ttLib.getTableClass('glyf')) def closure_glyphs(self, s): glyphSet = self.glyphs decompose = s.glyphs while decompose: components = set() for g in decompose: if g not in glyphSet: continue gl = glyphSet[g] for c in gl.getComponentNames(self): components.add(c) components -= s.glyphs s.glyphs.update(components) decompose = components @_add_method(ttLib.getTableClass('glyf')) def prune_pre_subset(self, font, options): if options.notdef_glyph and not options.notdef_outline: g = self[self.glyphOrder[0]] # Yay, easy! g.__dict__.clear() g.data = "" return True @_add_method(ttLib.getTableClass('glyf')) def subset_glyphs(self, s): self.glyphs = _dict_subset(self.glyphs, s.glyphs) indices = [i for i,g in enumerate(self.glyphOrder) if g in s.glyphs] for v in self.glyphs.values(): if hasattr(v, "data"): v.remapComponentsFast(indices) else: pass # No need self.glyphOrder = [g for g in self.glyphOrder if g in s.glyphs] # Don't drop empty 'glyf' tables, otherwise 'loca' doesn't get subset. return True @_add_method(ttLib.getTableClass('glyf')) def prune_post_subset(self, options): remove_hinting = not options.hinting for v in self.glyphs.values(): v.trim(remove_hinting=remove_hinting) return True class _ClosureGlyphsT2Decompiler(psCharStrings.SimpleT2Decompiler): def __init__(self, components, localSubrs, globalSubrs): psCharStrings.SimpleT2Decompiler.__init__(self, localSubrs, globalSubrs) self.components = components def op_endchar(self, index): args = self.popall() if len(args) >= 4: from fontTools.encodings.StandardEncoding import StandardEncoding # endchar can do seac accent bulding; The T2 spec says it's deprecated, # but recent software that shall remain nameless does output it. adx, ady, bchar, achar = args[-4:] baseGlyph = StandardEncoding[bchar] accentGlyph = StandardEncoding[achar] self.components.add(baseGlyph) self.components.add(accentGlyph) @_add_method(ttLib.getTableClass('CFF ')) def closure_glyphs(self, s): cff = self.cff assert len(cff) == 1 font = cff[cff.keys()[0]] glyphSet = font.CharStrings decompose = s.glyphs while decompose: components = set() for g in decompose: if g not in glyphSet: continue gl = glyphSet[g] subrs = getattr(gl.private, "Subrs", []) decompiler = _ClosureGlyphsT2Decompiler(components, subrs, gl.globalSubrs) decompiler.execute(gl) components -= s.glyphs s.glyphs.update(components) decompose = components @_add_method(ttLib.getTableClass('CFF ')) def prune_pre_subset(self, font, options): cff = self.cff # CFF table must have one font only cff.fontNames = cff.fontNames[:1] if options.notdef_glyph and not options.notdef_outline: for fontname in cff.keys(): font = cff[fontname] c, fdSelectIndex = font.CharStrings.getItemAndSelector('.notdef') if hasattr(font, 'FDArray') and font.FDArray is not None: private = font.FDArray[fdSelectIndex].Private else: private = font.Private dfltWdX = private.defaultWidthX nmnlWdX = private.nominalWidthX pen = NullPen() c.draw(pen) # this will set the charstring's width if c.width != dfltWdX: c.program = [c.width - nmnlWdX, 'endchar'] else: c.program = ['endchar'] # Clear useless Encoding for fontname in cff.keys(): font = cff[fontname] # https://github.com/behdad/fonttools/issues/620 font.Encoding = "StandardEncoding" return True # bool(cff.fontNames) @_add_method(ttLib.getTableClass('CFF ')) def subset_glyphs(self, s): cff = self.cff for fontname in cff.keys(): font = cff[fontname] cs = font.CharStrings # Load all glyphs for g in font.charset: if g not in s.glyphs: continue c, _ = cs.getItemAndSelector(g) if cs.charStringsAreIndexed: indices = [i for i,g in enumerate(font.charset) if g in s.glyphs] csi = cs.charStringsIndex csi.items = [csi.items[i] for i in indices] del csi.file, csi.offsets if hasattr(font, "FDSelect"): sel = font.FDSelect # XXX We want to set sel.format to None, such that the # most compact format is selected. However, OTS was # broken and couldn't parse a FDSelect format 0 that # happened before CharStrings. As such, always force # format 3 until we fix cffLib to always generate # FDSelect after CharStrings. # https://github.com/khaledhosny/ots/pull/31 #sel.format = None sel.format = 3 sel.gidArray = [sel.gidArray[i] for i in indices] cs.charStrings = {g:indices.index(v) for g,v in cs.charStrings.items() if g in s.glyphs} else: cs.charStrings = {g:v for g,v in cs.charStrings.items() if g in s.glyphs} font.charset = [g for g in font.charset if g in s.glyphs] font.numGlyphs = len(font.charset) return True # any(cff[fontname].numGlyphs for fontname in cff.keys()) @_add_method(psCharStrings.T2CharString) def subset_subroutines(self, subrs, gsubrs): p = self.program assert len(p) for i in range(1, len(p)): if p[i] == 'callsubr': assert isinstance(p[i-1], int) p[i-1] = subrs._used.index(p[i-1] + subrs._old_bias) - subrs._new_bias elif p[i] == 'callgsubr': assert isinstance(p[i-1], int) p[i-1] = gsubrs._used.index(p[i-1] + gsubrs._old_bias) - gsubrs._new_bias @_add_method(psCharStrings.T2CharString) def drop_hints(self): hints = self._hints if hints.deletions: p = self.program for idx in reversed(hints.deletions): del p[idx-2:idx] if hints.has_hint: assert not hints.deletions or hints.last_hint <= hints.deletions[0] self.program = self.program[hints.last_hint:] if hasattr(self, 'width'): # Insert width back if needed if self.width != self.private.defaultWidthX: self.program.insert(0, self.width - self.private.nominalWidthX) if hints.has_hintmask: i = 0 p = self.program while i < len(p): if p[i] in ['hintmask', 'cntrmask']: assert i + 1 <= len(p) del p[i:i+2] continue i += 1 assert len(self.program) del self._hints class _MarkingT2Decompiler(psCharStrings.SimpleT2Decompiler): def __init__(self, localSubrs, globalSubrs): psCharStrings.SimpleT2Decompiler.__init__(self, localSubrs, globalSubrs) for subrs in [localSubrs, globalSubrs]: if subrs and not hasattr(subrs, "_used"): subrs._used = set() def op_callsubr(self, index): self.localSubrs._used.add(self.operandStack[-1]+self.localBias) psCharStrings.SimpleT2Decompiler.op_callsubr(self, index) def op_callgsubr(self, index): self.globalSubrs._used.add(self.operandStack[-1]+self.globalBias) psCharStrings.SimpleT2Decompiler.op_callgsubr(self, index) class _DehintingT2Decompiler(psCharStrings.T2WidthExtractor): class Hints(object): def __init__(self): # Whether calling this charstring produces any hint stems # Note that if a charstring starts with hintmask, it will # have has_hint set to True, because it *might* produce an # implicit vstem if called under certain conditions. self.has_hint = False # Index to start at to drop all hints self.last_hint = 0 # Index up to which we know more hints are possible. # Only relevant if status is 0 or 1. self.last_checked = 0 # The status means: # 0: after dropping hints, this charstring is empty # 1: after dropping hints, there may be more hints # continuing after this # 2: no more hints possible after this charstring self.status = 0 # Has hintmask instructions; not recursive self.has_hintmask = False # List of indices of calls to empty subroutines to remove. self.deletions = [] pass def __init__(self, css, localSubrs, globalSubrs, nominalWidthX, defaultWidthX): self._css = css psCharStrings.T2WidthExtractor.__init__( self, localSubrs, globalSubrs, nominalWidthX, defaultWidthX) def execute(self, charString): old_hints = charString._hints if hasattr(charString, '_hints') else None charString._hints = self.Hints() psCharStrings.T2WidthExtractor.execute(self, charString) hints = charString._hints if hints.has_hint or hints.has_hintmask: self._css.add(charString) if hints.status != 2: # Check from last_check, make sure we didn't have any operators. for i in range(hints.last_checked, len(charString.program) - 1): if isinstance(charString.program[i], str): hints.status = 2 break else: hints.status = 1 # There's *something* here hints.last_checked = len(charString.program) if old_hints: assert hints.__dict__ == old_hints.__dict__ def op_callsubr(self, index): subr = self.localSubrs[self.operandStack[-1]+self.localBias] psCharStrings.T2WidthExtractor.op_callsubr(self, index) self.processSubr(index, subr) def op_callgsubr(self, index): subr = self.globalSubrs[self.operandStack[-1]+self.globalBias] psCharStrings.T2WidthExtractor.op_callgsubr(self, index) self.processSubr(index, subr) def op_hstem(self, index): psCharStrings.T2WidthExtractor.op_hstem(self, index) self.processHint(index) def op_vstem(self, index): psCharStrings.T2WidthExtractor.op_vstem(self, index) self.processHint(index) def op_hstemhm(self, index): psCharStrings.T2WidthExtractor.op_hstemhm(self, index) self.processHint(index) def op_vstemhm(self, index): psCharStrings.T2WidthExtractor.op_vstemhm(self, index) self.processHint(index) def op_hintmask(self, index): rv = psCharStrings.T2WidthExtractor.op_hintmask(self, index) self.processHintmask(index) return rv def op_cntrmask(self, index): rv = psCharStrings.T2WidthExtractor.op_cntrmask(self, index) self.processHintmask(index) return rv def processHintmask(self, index): cs = self.callingStack[-1] hints = cs._hints hints.has_hintmask = True if hints.status != 2: # Check from last_check, see if we may be an implicit vstem for i in range(hints.last_checked, index - 1): if isinstance(cs.program[i], str): hints.status = 2 break else: # We are an implicit vstem hints.has_hint = True hints.last_hint = index + 1 hints.status = 0 hints.last_checked = index + 1 def processHint(self, index): cs = self.callingStack[-1] hints = cs._hints hints.has_hint = True hints.last_hint = index hints.last_checked = index def processSubr(self, index, subr): cs = self.callingStack[-1] hints = cs._hints subr_hints = subr._hints # Check from last_check, make sure we didn't have # any operators. if hints.status != 2: for i in range(hints.last_checked, index - 1): if isinstance(cs.program[i], str): hints.status = 2 break hints.last_checked = index if hints.status != 2: if subr_hints.has_hint: hints.has_hint = True # Decide where to chop off from if subr_hints.status == 0: hints.last_hint = index else: hints.last_hint = index - 2 # Leave the subr call in elif subr_hints.status == 0: hints.deletions.append(index) hints.status = max(hints.status, subr_hints.status) class _DesubroutinizingT2Decompiler(psCharStrings.SimpleT2Decompiler): def __init__(self, localSubrs, globalSubrs): psCharStrings.SimpleT2Decompiler.__init__(self, localSubrs, globalSubrs) def execute(self, charString): # Note: Currently we recompute _desubroutinized each time. # This is more robust in some cases, but in other places we assume # that each subroutine always expands to the same code, so # maybe it doesn't matter. To speed up we can just not # recompute _desubroutinized if it's there. For now I just # double-check that it desubroutinized to the same thing. old_desubroutinized = charString._desubroutinized if hasattr(charString, '_desubroutinized') else None charString._patches = [] psCharStrings.SimpleT2Decompiler.execute(self, charString) desubroutinized = charString.program[:] for idx,expansion in reversed (charString._patches): assert idx >= 2 assert desubroutinized[idx - 1] in ['callsubr', 'callgsubr'], desubroutinized[idx - 1] assert type(desubroutinized[idx - 2]) == int if expansion[-1] == 'return': expansion = expansion[:-1] desubroutinized[idx-2:idx] = expansion if 'endchar' in desubroutinized: # Cut off after first endchar desubroutinized = desubroutinized[:desubroutinized.index('endchar') + 1] else: if not len(desubroutinized) or desubroutinized[-1] != 'return': desubroutinized.append('return') charString._desubroutinized = desubroutinized del charString._patches if old_desubroutinized: assert desubroutinized == old_desubroutinized def op_callsubr(self, index): subr = self.localSubrs[self.operandStack[-1]+self.localBias] psCharStrings.SimpleT2Decompiler.op_callsubr(self, index) self.processSubr(index, subr) def op_callgsubr(self, index): subr = self.globalSubrs[self.operandStack[-1]+self.globalBias] psCharStrings.SimpleT2Decompiler.op_callgsubr(self, index) self.processSubr(index, subr) def processSubr(self, index, subr): cs = self.callingStack[-1] cs._patches.append((index, subr._desubroutinized)) @_add_method(ttLib.getTableClass('CFF ')) def prune_post_subset(self, options): cff = self.cff for fontname in cff.keys(): font = cff[fontname] cs = font.CharStrings # Drop unused FontDictionaries if hasattr(font, "FDSelect"): sel = font.FDSelect indices = _uniq_sort(sel.gidArray) sel.gidArray = [indices.index (ss) for ss in sel.gidArray] arr = font.FDArray arr.items = [arr[i] for i in indices] del arr.file, arr.offsets # Desubroutinize if asked for if options.desubroutinize: for g in font.charset: c, _ = cs.getItemAndSelector(g) c.decompile() subrs = getattr(c.private, "Subrs", []) decompiler = _DesubroutinizingT2Decompiler(subrs, c.globalSubrs) decompiler.execute(c) c.program = c._desubroutinized # Drop hints if not needed if not options.hinting: # This can be tricky, but doesn't have to. What we do is: # # - Run all used glyph charstrings and recurse into subroutines, # - For each charstring (including subroutines), if it has any # of the hint stem operators, we mark it as such. # Upon returning, for each charstring we note all the # subroutine calls it makes that (recursively) contain a stem, # - Dropping hinting then consists of the following two ops: # * Drop the piece of the program in each charstring before the # last call to a stem op or a stem-calling subroutine, # * Drop all hintmask operations. # - It's trickier... A hintmask right after hints and a few numbers # will act as an implicit vstemhm. As such, we track whether # we have seen any non-hint operators so far and do the right # thing, recursively... Good luck understanding that :( css = set() for g in font.charset: c, _ = cs.getItemAndSelector(g) c.decompile() subrs = getattr(c.private, "Subrs", []) decompiler = _DehintingT2Decompiler(css, subrs, c.globalSubrs, c.private.nominalWidthX, c.private.defaultWidthX) decompiler.execute(c) c.width = decompiler.width for charstring in css: charstring.drop_hints() del css # Drop font-wide hinting values all_privs = [] if hasattr(font, 'FDSelect'): all_privs.extend(fd.Private for fd in font.FDArray) else: all_privs.append(font.Private) for priv in all_privs: for k in ['BlueValues', 'OtherBlues', 'FamilyBlues', 'FamilyOtherBlues', 'BlueScale', 'BlueShift', 'BlueFuzz', 'StemSnapH', 'StemSnapV', 'StdHW', 'StdVW']: if hasattr(priv, k): setattr(priv, k, None) # Renumber subroutines to remove unused ones # Mark all used subroutines for g in font.charset: c, _ = cs.getItemAndSelector(g) subrs = getattr(c.private, "Subrs", []) decompiler = _MarkingT2Decompiler(subrs, c.globalSubrs) decompiler.execute(c) all_subrs = [font.GlobalSubrs] if hasattr(font, 'FDSelect'): all_subrs.extend(fd.Private.Subrs for fd in font.FDArray if hasattr(fd.Private, 'Subrs') and fd.Private.Subrs) elif hasattr(font.Private, 'Subrs') and font.Private.Subrs: all_subrs.append(font.Private.Subrs) subrs = set(subrs) # Remove duplicates # Prepare for subrs in all_subrs: if not hasattr(subrs, '_used'): subrs._used = set() subrs._used = _uniq_sort(subrs._used) subrs._old_bias = psCharStrings.calcSubrBias(subrs) subrs._new_bias = psCharStrings.calcSubrBias(subrs._used) # Renumber glyph charstrings for g in font.charset: c, _ = cs.getItemAndSelector(g) subrs = getattr(c.private, "Subrs", []) c.subset_subroutines (subrs, font.GlobalSubrs) # Renumber subroutines themselves for subrs in all_subrs: if subrs == font.GlobalSubrs: if not hasattr(font, 'FDSelect') and hasattr(font.Private, 'Subrs'): local_subrs = font.Private.Subrs else: local_subrs = [] else: local_subrs = subrs subrs.items = [subrs.items[i] for i in subrs._used] if hasattr(subrs, 'file'): del subrs.file if hasattr(subrs, 'offsets'): del subrs.offsets for subr in subrs.items: subr.subset_subroutines (local_subrs, font.GlobalSubrs) # Delete local SubrsIndex if empty if hasattr(font, 'FDSelect'): for fd in font.FDArray: _delete_empty_subrs(fd.Private) else: _delete_empty_subrs(font.Private) # Cleanup for subrs in all_subrs: del subrs._used, subrs._old_bias, subrs._new_bias return True def _delete_empty_subrs(private_dict): if hasattr(private_dict, 'Subrs') and not private_dict.Subrs: if 'Subrs' in private_dict.rawDict: del private_dict.rawDict['Subrs'] del private_dict.Subrs @_add_method(ttLib.getTableClass('cmap')) def closure_glyphs(self, s): tables = [t for t in self.tables if t.isUnicode()] # Close glyphs for table in tables: if table.format == 14: for cmap in table.uvsDict.values(): glyphs = {g for u,g in cmap if u in s.unicodes_requested} if None in glyphs: glyphs.remove(None) s.glyphs.update(glyphs) else: cmap = table.cmap intersection = s.unicodes_requested.intersection(cmap.keys()) s.glyphs.update(cmap[u] for u in intersection) # Calculate unicodes_missing s.unicodes_missing = s.unicodes_requested.copy() for table in tables: s.unicodes_missing.difference_update(table.cmap) @_add_method(ttLib.getTableClass('cmap')) def prune_pre_subset(self, font, options): if not options.legacy_cmap: # Drop non-Unicode / non-Symbol cmaps self.tables = [t for t in self.tables if t.isUnicode() or t.isSymbol()] if not options.symbol_cmap: self.tables = [t for t in self.tables if not t.isSymbol()] # TODO(behdad) Only keep one subtable? # For now, drop format=0 which can't be subset_glyphs easily? self.tables = [t for t in self.tables if t.format != 0] self.numSubTables = len(self.tables) return True # Required table @_add_method(ttLib.getTableClass('cmap')) def subset_glyphs(self, s): s.glyphs = None # We use s.glyphs_requested and s.unicodes_requested only for t in self.tables: if t.format == 14: # TODO(behdad) We drop all the default-UVS mappings # for glyphs_requested. So it's the caller's responsibility to make # sure those are included. t.uvsDict = {v:[(u,g) for u,g in l if g in s.glyphs_requested or u in s.unicodes_requested] for v,l in t.uvsDict.items()} t.uvsDict = {v:l for v,l in t.uvsDict.items() if l} elif t.isUnicode(): t.cmap = {u:g for u,g in t.cmap.items() if g in s.glyphs_requested or u in s.unicodes_requested} else: t.cmap = {u:g for u,g in t.cmap.items() if g in s.glyphs_requested} self.tables = [t for t in self.tables if (t.cmap if t.format != 14 else t.uvsDict)] self.numSubTables = len(self.tables) # TODO(behdad) Convert formats when needed. # In particular, if we have a format=12 without non-BMP # characters, either drop format=12 one or convert it # to format=4 if there's not one. return True # Required table @_add_method(ttLib.getTableClass('DSIG')) def prune_pre_subset(self, font, options): # Drop all signatures since they will be invalid self.usNumSigs = 0 self.signatureRecords = [] return True @_add_method(ttLib.getTableClass('maxp')) def prune_pre_subset(self, font, options): if not options.hinting: if self.tableVersion == 0x00010000: self.maxZones = 1 self.maxTwilightPoints = 0 self.maxStorage = 0 self.maxFunctionDefs = 0 self.maxInstructionDefs = 0 self.maxStackElements = 0 self.maxSizeOfInstructions = 0 return True @_add_method(ttLib.getTableClass('name')) def prune_pre_subset(self, font, options): nameIDs = set(options.name_IDs) fvar = font.get('fvar') if fvar: nameIDs.update([axis.axisNameID for axis in fvar.axes]) nameIDs.update([inst.subfamilyNameID for inst in fvar.instances]) nameIDs.update([inst.postscriptNameID for inst in fvar.instances if inst.postscriptNameID != 0xFFFF]) if '*' not in options.name_IDs: self.names = [n for n in self.names if n.nameID in nameIDs] if not options.name_legacy: # TODO(behdad) Sometimes (eg Apple Color Emoji) there's only a macroman # entry for Latin and no Unicode names. self.names = [n for n in self.names if n.isUnicode()] # TODO(behdad) Option to keep only one platform's if '*' not in options.name_languages: # TODO(behdad) This is Windows-platform specific! self.names = [n for n in self.names if n.langID in options.name_languages] if options.obfuscate_names: namerecs = [] for n in self.names: if n.nameID in [1, 4]: n.string = ".\x7f".encode('utf_16_be') if n.isUnicode() else ".\x7f" elif n.nameID in [2, 6]: n.string = "\x7f".encode('utf_16_be') if n.isUnicode() else "\x7f" elif n.nameID == 3: n.string = "" elif n.nameID in [16, 17, 18]: continue namerecs.append(n) self.names = namerecs return True # Required table # TODO(behdad) OS/2 ulCodePageRange? # TODO(behdad) Drop AAT tables. # TODO(behdad) Drop unneeded GSUB/GPOS Script/LangSys entries. # TODO(behdad) Drop empty GSUB/GPOS, and GDEF if no GSUB/GPOS left # TODO(behdad) Drop GDEF subitems if unused by lookups # TODO(behdad) Avoid recursing too much (in GSUB/GPOS and in CFF) # TODO(behdad) Text direction considerations. # TODO(behdad) Text script / language considerations. # TODO(behdad) Optionally drop 'kern' table if GPOS available # TODO(behdad) Implement --unicode='*' to choose all cmap'ed # TODO(behdad) Drop old-spec Indic scripts class Options(object): class OptionError(Exception): pass class UnknownOptionError(OptionError): pass # spaces in tag names (e.g. "SVG ", "cvt ") are stripped by the argument parser _drop_tables_default = ['BASE', 'JSTF', 'DSIG', 'EBDT', 'EBLC', 'EBSC', 'SVG', 'PCLT', 'LTSH'] _drop_tables_default += ['Feat', 'Glat', 'Gloc', 'Silf', 'Sill'] # Graphite _drop_tables_default += ['sbix'] # Color _no_subset_tables_default = ['avar', 'fvar', 'gasp', 'head', 'hhea', 'maxp', 'vhea', 'OS/2', 'loca', 'name', 'cvt', 'fpgm', 'prep', 'VDMX', 'DSIG', 'CPAL', 'MVAR', 'STAT'] _hinting_tables_default = ['cvar', 'cvt', 'fpgm', 'prep', 'hdmx', 'VDMX'] # Based on HarfBuzz shapers _layout_features_groups = { # Default shaper 'common': ['rvrn', 'ccmp', 'liga', 'locl', 'mark', 'mkmk', 'rlig'], 'fractions': ['frac', 'numr', 'dnom'], 'horizontal': ['calt', 'clig', 'curs', 'kern', 'rclt'], 'vertical': ['valt', 'vert', 'vkrn', 'vpal', 'vrt2'], 'ltr': ['ltra', 'ltrm'], 'rtl': ['rtla', 'rtlm'], # Complex shapers 'arabic': ['init', 'medi', 'fina', 'isol', 'med2', 'fin2', 'fin3', 'cswh', 'mset', 'stch'], 'hangul': ['ljmo', 'vjmo', 'tjmo'], 'tibetan': ['abvs', 'blws', 'abvm', 'blwm'], 'indic': ['nukt', 'akhn', 'rphf', 'rkrf', 'pref', 'blwf', 'half', 'abvf', 'pstf', 'cfar', 'vatu', 'cjct', 'init', 'pres', 'abvs', 'blws', 'psts', 'haln', 'dist', 'abvm', 'blwm'], } _layout_features_default = _uniq_sort(sum( iter(_layout_features_groups.values()), [])) def __init__(self, **kwargs): self.drop_tables = self._drop_tables_default[:] self.no_subset_tables = self._no_subset_tables_default[:] self.passthrough_tables = False # keep/drop tables we can't subset self.hinting_tables = self._hinting_tables_default[:] self.legacy_kern = False # drop 'kern' table if GPOS available self.layout_features = self._layout_features_default[:] self.ignore_missing_glyphs = False self.ignore_missing_unicodes = True self.hinting = True self.glyph_names = False self.legacy_cmap = False self.symbol_cmap = False self.name_IDs = [1, 2, 3, 4, 6] # https://github.com/fonttools/fonttools/issues/1170#issuecomment-364631225 self.name_legacy = False self.name_languages = [0x0409] # English self.obfuscate_names = False # to make webfont unusable as a system font self.notdef_glyph = True # gid0 for TrueType / .notdef for CFF self.notdef_outline = False # No need for notdef to have an outline really self.recommended_glyphs = False # gid1, gid2, gid3 for TrueType self.recalc_bounds = False # Recalculate font bounding boxes self.recalc_timestamp = False # Recalculate font modified timestamp self.prune_unicode_ranges = True # Clear unused 'ulUnicodeRange' bits self.recalc_average_width = False # update 'xAvgCharWidth' self.canonical_order = None # Order tables as recommended self.flavor = None # May be 'woff' or 'woff2' self.with_zopfli = False # use zopfli instead of zlib for WOFF 1.0 self.desubroutinize = False # Desubroutinize CFF CharStrings self.verbose = False self.timing = False self.xml = False self.set(**kwargs) def set(self, **kwargs): for k,v in kwargs.items(): if not hasattr(self, k): raise self.UnknownOptionError("Unknown option '%s'" % k) setattr(self, k, v) def parse_opts(self, argv, ignore_unknown=[]): posargs = [] passthru_options = [] for a in argv: orig_a = a if not a.startswith('--'): posargs.append(a) continue a = a[2:] i = a.find('=') op = '=' if i == -1: if a.startswith("no-"): k = a[3:] if k == "canonical-order": # reorderTables=None is faster than False (the latter # still reorders to "keep" the original table order) v = None else: v = False else: k = a v = True if k.endswith("?"): k = k[:-1] v = '?' else: k = a[:i] if k[-1] in "-+": op = k[-1]+'=' # Op is '-=' or '+=' now. k = k[:-1] v = a[i+1:] ok = k k = k.replace('-', '_') if not hasattr(self, k): if ignore_unknown is True or ok in ignore_unknown: passthru_options.append(orig_a) continue else: raise self.UnknownOptionError("Unknown option '%s'" % a) ov = getattr(self, k) if v == '?': print("Current setting for '%s' is: %s" % (ok, ov)) continue if isinstance(ov, bool): v = bool(v) elif isinstance(ov, int): v = int(v) elif isinstance(ov, str): v = str(v) # redundant elif isinstance(ov, list): if isinstance(v, bool): raise self.OptionError("Option '%s' requires values to be specified using '='" % a) vv = v.replace(',', ' ').split() if vv == ['']: vv = [] vv = [int(x, 0) if len(x) and x[0] in "0123456789" else x for x in vv] if op == '=': v = vv elif op == '+=': v = ov v.extend(vv) elif op == '-=': v = ov for x in vv: if x in v: v.remove(x) else: assert False setattr(self, k, v) return posargs + passthru_options class Subsetter(object): class SubsettingError(Exception): pass class MissingGlyphsSubsettingError(SubsettingError): pass class MissingUnicodesSubsettingError(SubsettingError): pass def __init__(self, options=None): if not options: options = Options() self.options = options self.unicodes_requested = set() self.glyph_names_requested = set() self.glyph_ids_requested = set() def populate(self, glyphs=[], gids=[], unicodes=[], text=""): self.unicodes_requested.update(unicodes) if isinstance(text, bytes): text = text.decode("utf_8") text_utf32 = text.encode("utf-32-be") nchars = len(text_utf32)//4 for u in struct.unpack('>%dL' % nchars, text_utf32): self.unicodes_requested.add(u) self.glyph_names_requested.update(glyphs) self.glyph_ids_requested.update(gids) def _prune_pre_subset(self, font): for tag in self._sort_tables(font): if (tag.strip() in self.options.drop_tables or (tag.strip() in self.options.hinting_tables and not self.options.hinting) or (tag == 'kern' and (not self.options.legacy_kern and 'GPOS' in font))): log.info("%s dropped", tag) del font[tag] continue clazz = ttLib.getTableClass(tag) if hasattr(clazz, 'prune_pre_subset'): with timer("load '%s'" % tag): table = font[tag] with timer("prune '%s'" % tag): retain = table.prune_pre_subset(font, self.options) if not retain: log.info("%s pruned to empty; dropped", tag) del font[tag] continue else: log.info("%s pruned", tag) def _closure_glyphs(self, font): realGlyphs = set(font.getGlyphOrder()) glyph_order = font.getGlyphOrder() self.glyphs_requested = set() self.glyphs_requested.update(self.glyph_names_requested) self.glyphs_requested.update(glyph_order[i] for i in self.glyph_ids_requested if i < len(glyph_order)) self.glyphs_missing = set() self.glyphs_missing.update(self.glyphs_requested.difference(realGlyphs)) self.glyphs_missing.update(i for i in self.glyph_ids_requested if i >= len(glyph_order)) if self.glyphs_missing: log.info("Missing requested glyphs: %s", self.glyphs_missing) if not self.options.ignore_missing_glyphs: raise self.MissingGlyphsSubsettingError(self.glyphs_missing) self.glyphs = self.glyphs_requested.copy() self.unicodes_missing = set() if 'cmap' in font: with timer("close glyph list over 'cmap'"): font['cmap'].closure_glyphs(self) self.glyphs.intersection_update(realGlyphs) self.glyphs_cmaped = frozenset(self.glyphs) if self.unicodes_missing: missing = ["U+%04X" % u for u in self.unicodes_missing] log.info("Missing glyphs for requested Unicodes: %s", missing) if not self.options.ignore_missing_unicodes: raise self.MissingUnicodesSubsettingError(missing) del missing if self.options.notdef_glyph: if 'glyf' in font: self.glyphs.add(font.getGlyphName(0)) log.info("Added gid0 to subset") else: self.glyphs.add('.notdef') log.info("Added .notdef to subset") if self.options.recommended_glyphs: if 'glyf' in font: for i in range(min(4, len(font.getGlyphOrder()))): self.glyphs.add(font.getGlyphName(i)) log.info("Added first four glyphs to subset") if 'GSUB' in font: with timer("close glyph list over 'GSUB'"): log.info("Closing glyph list over 'GSUB': %d glyphs before", len(self.glyphs)) log.glyphs(self.glyphs, font=font) font['GSUB'].closure_glyphs(self) self.glyphs.intersection_update(realGlyphs) log.info("Closed glyph list over 'GSUB': %d glyphs after", len(self.glyphs)) log.glyphs(self.glyphs, font=font) self.glyphs_gsubed = frozenset(self.glyphs) if 'MATH' in font: with timer("close glyph list over 'MATH'"): log.info("Closing glyph list over 'MATH': %d glyphs before", len(self.glyphs)) log.glyphs(self.glyphs, font=font) font['MATH'].closure_glyphs(self) self.glyphs.intersection_update(realGlyphs) log.info("Closed glyph list over 'MATH': %d glyphs after", len(self.glyphs)) log.glyphs(self.glyphs, font=font) self.glyphs_mathed = frozenset(self.glyphs) for table in ('COLR', 'bsln'): if table in font: with timer("close glyph list over '%s'" % table): log.info("Closing glyph list over '%s': %d glyphs before", table, len(self.glyphs)) log.glyphs(self.glyphs, font=font) font[table].closure_glyphs(self) self.glyphs.intersection_update(realGlyphs) log.info("Closed glyph list over '%s': %d glyphs after", table, len(self.glyphs)) log.glyphs(self.glyphs, font=font) if 'glyf' in font: with timer("close glyph list over 'glyf'"): log.info("Closing glyph list over 'glyf': %d glyphs before", len(self.glyphs)) log.glyphs(self.glyphs, font=font) font['glyf'].closure_glyphs(self) self.glyphs.intersection_update(realGlyphs) log.info("Closed glyph list over 'glyf': %d glyphs after", len(self.glyphs)) log.glyphs(self.glyphs, font=font) self.glyphs_glyfed = frozenset(self.glyphs) if 'CFF ' in font: with timer("close glyph list over 'CFF '"): log.info("Closing glyph list over 'CFF ': %d glyphs before", len(self.glyphs)) log.glyphs(self.glyphs, font=font) font['CFF '].closure_glyphs(self) self.glyphs.intersection_update(realGlyphs) log.info("Closed glyph list over 'CFF ': %d glyphs after", len(self.glyphs)) log.glyphs(self.glyphs, font=font) self.glyphs_cffed = frozenset(self.glyphs) self.glyphs_all = frozenset(self.glyphs) log.info("Retaining %d glyphs", len(self.glyphs_all)) del self.glyphs def _subset_glyphs(self, font): for tag in self._sort_tables(font): clazz = ttLib.getTableClass(tag) if tag.strip() in self.options.no_subset_tables: log.info("%s subsetting not needed", tag) elif hasattr(clazz, 'subset_glyphs'): with timer("subset '%s'" % tag): table = font[tag] self.glyphs = self.glyphs_all retain = table.subset_glyphs(self) del self.glyphs if not retain: log.info("%s subsetted to empty; dropped", tag) del font[tag] else: log.info("%s subsetted", tag) elif self.options.passthrough_tables: log.info("%s NOT subset; don't know how to subset", tag) else: log.info("%s NOT subset; don't know how to subset; dropped", tag) del font[tag] with timer("subset GlyphOrder"): glyphOrder = font.getGlyphOrder() glyphOrder = [g for g in glyphOrder if g in self.glyphs_all] font.setGlyphOrder(glyphOrder) font._buildReverseGlyphOrderDict() def _prune_post_subset(self, font): for tag in font.keys(): if tag == 'GlyphOrder': continue if tag == 'OS/2' and self.options.prune_unicode_ranges: old_uniranges = font[tag].getUnicodeRanges() new_uniranges = font[tag].recalcUnicodeRanges(font, pruneOnly=True) if old_uniranges != new_uniranges: log.info("%s Unicode ranges pruned: %s", tag, sorted(new_uniranges)) if self.options.recalc_average_width: widths = [m[0] for m in font["hmtx"].metrics.values() if m[0] > 0] avg_width = round(sum(widths) / len(widths)) if avg_width != font[tag].xAvgCharWidth: font[tag].xAvgCharWidth = avg_width log.info("%s xAvgCharWidth updated: %d", tag, avg_width) clazz = ttLib.getTableClass(tag) if hasattr(clazz, 'prune_post_subset'): with timer("prune '%s'" % tag): table = font[tag] retain = table.prune_post_subset(self.options) if not retain: log.info("%s pruned to empty; dropped", tag) del font[tag] else: log.info("%s pruned", tag) def _sort_tables(self, font): tagOrder = ['fvar', 'avar', 'gvar', 'name', 'glyf'] tagOrder = {t: i + 1 for i, t in enumerate(tagOrder)} tags = sorted(font.keys(), key=lambda tag: tagOrder.get(tag, 0)) return [t for t in tags if t != 'GlyphOrder'] def subset(self, font): self._prune_pre_subset(font) self._closure_glyphs(font) self._subset_glyphs(font) self._prune_post_subset(font) @timer("load font") def load_font(fontFile, options, allowVID=False, checkChecksums=False, dontLoadGlyphNames=False, lazy=True): font = ttLib.TTFont(fontFile, allowVID=allowVID, checkChecksums=checkChecksums, recalcBBoxes=options.recalc_bounds, recalcTimestamp=options.recalc_timestamp, lazy=lazy) # Hack: # # If we don't need glyph names, change 'post' class to not try to # load them. It avoid lots of headache with broken fonts as well # as loading time. # # Ideally ttLib should provide a way to ask it to skip loading # glyph names. But it currently doesn't provide such a thing. # if dontLoadGlyphNames: post = ttLib.getTableClass('post') saved = post.decode_format_2_0 post.decode_format_2_0 = post.decode_format_3_0 f = font['post'] if f.formatType == 2.0: f.formatType = 3.0 post.decode_format_2_0 = saved return font @timer("compile and save font") def save_font(font, outfile, options): if options.flavor and not hasattr(font, 'flavor'): raise Exception("fonttools version does not support flavors.") if options.with_zopfli and options.flavor == "woff": from fontTools.ttLib import sfnt sfnt.USE_ZOPFLI = True font.flavor = options.flavor font.save(outfile, reorderTables=options.canonical_order) def parse_unicodes(s): import re s = re.sub (r"0[xX]", " ", s) s = re.sub (r"[<+>,;&#\\xXuU\n ]", " ", s) l = [] for item in s.split(): fields = item.split('-') if len(fields) == 1: l.append(int(item, 16)) else: start,end = fields l.extend(range(int(start, 16), int(end, 16)+1)) return l def parse_gids(s): l = [] for item in s.replace(',', ' ').split(): fields = item.split('-') if len(fields) == 1: l.append(int(fields[0])) else: l.extend(range(int(fields[0]), int(fields[1])+1)) return l def parse_glyphs(s): return s.replace(',', ' ').split() def usage(): print("usage:", __usage__, file=sys.stderr) print("Try pyftsubset --help for more information.\n", file=sys.stderr) @timer("make one with everything (TOTAL TIME)") def main(args=None): from os.path import splitext from fontTools import configLogger if args is None: args = sys.argv[1:] if '--help' in args: print(__doc__) return 0 options = Options() try: args = options.parse_opts(args, ignore_unknown=['gids', 'gids-file', 'glyphs', 'glyphs-file', 'text', 'text-file', 'unicodes', 'unicodes-file', 'output-file']) except options.OptionError as e: usage() print("ERROR:", e, file=sys.stderr) return 2 if len(args) < 2: usage() return 1 configLogger(level=logging.INFO if options.verbose else logging.WARNING) if options.timing: timer.logger.setLevel(logging.DEBUG) else: timer.logger.disabled = True fontfile = args[0] args = args[1:] subsetter = Subsetter(options=options) basename, extension = splitext(fontfile) outfile = basename + '.subset' + extension glyphs = [] gids = [] unicodes = [] wildcard_glyphs = False wildcard_unicodes = False text = "" for g in args: if g == '*': wildcard_glyphs = True continue if g.startswith('--output-file='): outfile = g[14:] continue if g.startswith('--text='): text += g[7:] continue if g.startswith('--text-file='): text += open(g[12:], encoding='utf-8').read().replace('\n', '') continue if g.startswith('--unicodes='): if g[11:] == '*': wildcard_unicodes = True else: unicodes.extend(parse_unicodes(g[11:])) continue if g.startswith('--unicodes-file='): for line in open(g[16:]).readlines(): unicodes.extend(parse_unicodes(line.split('#')[0])) continue if g.startswith('--gids='): gids.extend(parse_gids(g[7:])) continue if g.startswith('--gids-file='): for line in open(g[12:]).readlines(): gids.extend(parse_gids(line.split('#')[0])) continue if g.startswith('--glyphs='): if g[9:] == '*': wildcard_glyphs = True else: glyphs.extend(parse_glyphs(g[9:])) continue if g.startswith('--glyphs-file='): for line in open(g[14:]).readlines(): glyphs.extend(parse_glyphs(line.split('#')[0])) continue glyphs.append(g) dontLoadGlyphNames = not options.glyph_names and not glyphs font = load_font(fontfile, options, dontLoadGlyphNames=dontLoadGlyphNames) with timer("compile glyph list"): if wildcard_glyphs: glyphs.extend(font.getGlyphOrder()) if wildcard_unicodes: for t in font['cmap'].tables: if t.isUnicode(): unicodes.extend(t.cmap.keys()) assert '' not in glyphs log.info("Text: '%s'" % text) log.info("Unicodes: %s", unicodes) log.info("Glyphs: %s", glyphs) log.info("Gids: %s", gids) subsetter.populate(glyphs=glyphs, gids=gids, unicodes=unicodes, text=text) subsetter.subset(font) save_font(font, outfile, options) if options.verbose: import os log.info("Input font:% 7d bytes: %s" % (os.path.getsize(fontfile), fontfile)) log.info("Subset font:% 7d bytes: %s" % (os.path.getsize(outfile), outfile)) if options.xml: font.saveXML(sys.stdout) font.close() __all__ = [ 'Options', 'Subsetter', 'load_font', 'save_font', 'parse_gids', 'parse_glyphs', 'parse_unicodes', 'main' ] if __name__ == '__main__': sys.exit(main())