forked from TrueCloudLab/lego
1712 lines
48 KiB
Go
1712 lines
48 KiB
Go
// Copyright 2013 The Go Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style
|
|
// license that can be found in the LICENSE file.
|
|
|
|
// +build ignore
|
|
|
|
// Language tag table generator.
|
|
// Data read from the web.
|
|
|
|
package main
|
|
|
|
import (
|
|
"bufio"
|
|
"flag"
|
|
"fmt"
|
|
"io"
|
|
"io/ioutil"
|
|
"log"
|
|
"math"
|
|
"reflect"
|
|
"regexp"
|
|
"sort"
|
|
"strconv"
|
|
"strings"
|
|
|
|
"golang.org/x/text/internal/gen"
|
|
"golang.org/x/text/internal/tag"
|
|
"golang.org/x/text/unicode/cldr"
|
|
)
|
|
|
|
var (
|
|
test = flag.Bool("test",
|
|
false,
|
|
"test existing tables; can be used to compare web data with package data.")
|
|
outputFile = flag.String("output",
|
|
"tables.go",
|
|
"output file for generated tables")
|
|
)
|
|
|
|
var comment = []string{
|
|
`
|
|
lang holds an alphabetically sorted list of ISO-639 language identifiers.
|
|
All entries are 4 bytes. The index of the identifier (divided by 4) is the language tag.
|
|
For 2-byte language identifiers, the two successive bytes have the following meaning:
|
|
- if the first letter of the 2- and 3-letter ISO codes are the same:
|
|
the second and third letter of the 3-letter ISO code.
|
|
- otherwise: a 0 and a by 2 bits right-shifted index into altLangISO3.
|
|
For 3-byte language identifiers the 4th byte is 0.`,
|
|
`
|
|
langNoIndex is a bit vector of all 3-letter language codes that are not used as an index
|
|
in lookup tables. The language ids for these language codes are derived directly
|
|
from the letters and are not consecutive.`,
|
|
`
|
|
altLangISO3 holds an alphabetically sorted list of 3-letter language code alternatives
|
|
to 2-letter language codes that cannot be derived using the method described above.
|
|
Each 3-letter code is followed by its 1-byte langID.`,
|
|
`
|
|
altLangIndex is used to convert indexes in altLangISO3 to langIDs.`,
|
|
`
|
|
langAliasMap maps langIDs to their suggested replacements.`,
|
|
`
|
|
script is an alphabetically sorted list of ISO 15924 codes. The index
|
|
of the script in the string, divided by 4, is the internal scriptID.`,
|
|
`
|
|
isoRegionOffset needs to be added to the index of regionISO to obtain the regionID
|
|
for 2-letter ISO codes. (The first isoRegionOffset regionIDs are reserved for
|
|
the UN.M49 codes used for groups.)`,
|
|
`
|
|
regionISO holds a list of alphabetically sorted 2-letter ISO region codes.
|
|
Each 2-letter codes is followed by two bytes with the following meaning:
|
|
- [A-Z}{2}: the first letter of the 2-letter code plus these two
|
|
letters form the 3-letter ISO code.
|
|
- 0, n: index into altRegionISO3.`,
|
|
`
|
|
regionTypes defines the status of a region for various standards.`,
|
|
`
|
|
m49 maps regionIDs to UN.M49 codes. The first isoRegionOffset entries are
|
|
codes indicating collections of regions.`,
|
|
`
|
|
m49Index gives indexes into fromM49 based on the three most significant bits
|
|
of a 10-bit UN.M49 code. To search an UN.M49 code in fromM49, search in
|
|
fromM49[m49Index[msb39(code)]:m49Index[msb3(code)+1]]
|
|
for an entry where the first 7 bits match the 7 lsb of the UN.M49 code.
|
|
The region code is stored in the 9 lsb of the indexed value.`,
|
|
`
|
|
fromM49 contains entries to map UN.M49 codes to regions. See m49Index for details.`,
|
|
`
|
|
altRegionISO3 holds a list of 3-letter region codes that cannot be
|
|
mapped to 2-letter codes using the default algorithm. This is a short list.`,
|
|
`
|
|
altRegionIDs holds a list of regionIDs the positions of which match those
|
|
of the 3-letter ISO codes in altRegionISO3.`,
|
|
`
|
|
variantNumSpecialized is the number of specialized variants in variants.`,
|
|
`
|
|
suppressScript is an index from langID to the dominant script for that language,
|
|
if it exists. If a script is given, it should be suppressed from the language tag.`,
|
|
`
|
|
likelyLang is a lookup table, indexed by langID, for the most likely
|
|
scripts and regions given incomplete information. If more entries exist for a
|
|
given language, region and script are the index and size respectively
|
|
of the list in likelyLangList.`,
|
|
`
|
|
likelyLangList holds lists info associated with likelyLang.`,
|
|
`
|
|
likelyRegion is a lookup table, indexed by regionID, for the most likely
|
|
languages and scripts given incomplete information. If more entries exist
|
|
for a given regionID, lang and script are the index and size respectively
|
|
of the list in likelyRegionList.
|
|
TODO: exclude containers and user-definable regions from the list.`,
|
|
`
|
|
likelyRegionList holds lists info associated with likelyRegion.`,
|
|
`
|
|
likelyScript is a lookup table, indexed by scriptID, for the most likely
|
|
languages and regions given a script.`,
|
|
`
|
|
matchLang holds pairs of langIDs of base languages that are typically
|
|
mutually intelligible. Each pair is associated with a confidence and
|
|
whether the intelligibility goes one or both ways.`,
|
|
`
|
|
matchScript holds pairs of scriptIDs where readers of one script
|
|
can typically also read the other. Each is associated with a confidence.`,
|
|
`
|
|
nRegionGroups is the number of region groups.`,
|
|
`
|
|
regionInclusion maps region identifiers to sets of regions in regionInclusionBits,
|
|
where each set holds all groupings that are directly connected in a region
|
|
containment graph.`,
|
|
`
|
|
regionInclusionBits is an array of bit vectors where every vector represents
|
|
a set of region groupings. These sets are used to compute the distance
|
|
between two regions for the purpose of language matching.`,
|
|
`
|
|
regionInclusionNext marks, for each entry in regionInclusionBits, the set of
|
|
all groups that are reachable from the groups set in the respective entry.`,
|
|
}
|
|
|
|
// TODO: consider changing some of these structures to tries. This can reduce
|
|
// memory, but may increase the need for memory allocations. This could be
|
|
// mitigated if we can piggyback on language tags for common cases.
|
|
|
|
func failOnError(e error) {
|
|
if e != nil {
|
|
log.Panic(e)
|
|
}
|
|
}
|
|
|
|
type setType int
|
|
|
|
const (
|
|
Indexed setType = 1 + iota // all elements must be of same size
|
|
Linear
|
|
)
|
|
|
|
type stringSet struct {
|
|
s []string
|
|
sorted, frozen bool
|
|
|
|
// We often need to update values after the creation of an index is completed.
|
|
// We include a convenience map for keeping track of this.
|
|
update map[string]string
|
|
typ setType // used for checking.
|
|
}
|
|
|
|
func (ss *stringSet) clone() stringSet {
|
|
c := *ss
|
|
c.s = append([]string(nil), c.s...)
|
|
return c
|
|
}
|
|
|
|
func (ss *stringSet) setType(t setType) {
|
|
if ss.typ != t && ss.typ != 0 {
|
|
log.Panicf("type %d cannot be assigned as it was already %d", t, ss.typ)
|
|
}
|
|
}
|
|
|
|
// parse parses a whitespace-separated string and initializes ss with its
|
|
// components.
|
|
func (ss *stringSet) parse(s string) {
|
|
scan := bufio.NewScanner(strings.NewReader(s))
|
|
scan.Split(bufio.ScanWords)
|
|
for scan.Scan() {
|
|
ss.add(scan.Text())
|
|
}
|
|
}
|
|
|
|
func (ss *stringSet) assertChangeable() {
|
|
if ss.frozen {
|
|
log.Panic("attempt to modify a frozen stringSet")
|
|
}
|
|
}
|
|
|
|
func (ss *stringSet) add(s string) {
|
|
ss.assertChangeable()
|
|
ss.s = append(ss.s, s)
|
|
ss.sorted = ss.frozen
|
|
}
|
|
|
|
func (ss *stringSet) freeze() {
|
|
ss.compact()
|
|
ss.frozen = true
|
|
}
|
|
|
|
func (ss *stringSet) compact() {
|
|
if ss.sorted {
|
|
return
|
|
}
|
|
a := ss.s
|
|
sort.Strings(a)
|
|
k := 0
|
|
for i := 1; i < len(a); i++ {
|
|
if a[k] != a[i] {
|
|
a[k+1] = a[i]
|
|
k++
|
|
}
|
|
}
|
|
ss.s = a[:k+1]
|
|
ss.sorted = ss.frozen
|
|
}
|
|
|
|
type funcSorter struct {
|
|
fn func(a, b string) bool
|
|
sort.StringSlice
|
|
}
|
|
|
|
func (s funcSorter) Less(i, j int) bool {
|
|
return s.fn(s.StringSlice[i], s.StringSlice[j])
|
|
}
|
|
|
|
func (ss *stringSet) sortFunc(f func(a, b string) bool) {
|
|
ss.compact()
|
|
sort.Sort(funcSorter{f, sort.StringSlice(ss.s)})
|
|
}
|
|
|
|
func (ss *stringSet) remove(s string) {
|
|
ss.assertChangeable()
|
|
if i, ok := ss.find(s); ok {
|
|
copy(ss.s[i:], ss.s[i+1:])
|
|
ss.s = ss.s[:len(ss.s)-1]
|
|
}
|
|
}
|
|
|
|
func (ss *stringSet) replace(ol, nu string) {
|
|
ss.s[ss.index(ol)] = nu
|
|
ss.sorted = ss.frozen
|
|
}
|
|
|
|
func (ss *stringSet) index(s string) int {
|
|
ss.setType(Indexed)
|
|
i, ok := ss.find(s)
|
|
if !ok {
|
|
if i < len(ss.s) {
|
|
log.Panicf("find: item %q is not in list. Closest match is %q.", s, ss.s[i])
|
|
}
|
|
log.Panicf("find: item %q is not in list", s)
|
|
|
|
}
|
|
return i
|
|
}
|
|
|
|
func (ss *stringSet) find(s string) (int, bool) {
|
|
ss.compact()
|
|
i := sort.SearchStrings(ss.s, s)
|
|
return i, i != len(ss.s) && ss.s[i] == s
|
|
}
|
|
|
|
func (ss *stringSet) slice() []string {
|
|
ss.compact()
|
|
return ss.s
|
|
}
|
|
|
|
func (ss *stringSet) updateLater(v, key string) {
|
|
if ss.update == nil {
|
|
ss.update = map[string]string{}
|
|
}
|
|
ss.update[v] = key
|
|
}
|
|
|
|
// join joins the string and ensures that all entries are of the same length.
|
|
func (ss *stringSet) join() string {
|
|
ss.setType(Indexed)
|
|
n := len(ss.s[0])
|
|
for _, s := range ss.s {
|
|
if len(s) != n {
|
|
log.Panicf("join: not all entries are of the same length: %q", s)
|
|
}
|
|
}
|
|
ss.s = append(ss.s, strings.Repeat("\xff", n))
|
|
return strings.Join(ss.s, "")
|
|
}
|
|
|
|
// ianaEntry holds information for an entry in the IANA Language Subtag Repository.
|
|
// All types use the same entry.
|
|
// See http://tools.ietf.org/html/bcp47#section-5.1 for a description of the various
|
|
// fields.
|
|
type ianaEntry struct {
|
|
typ string
|
|
description []string
|
|
scope string
|
|
added string
|
|
preferred string
|
|
deprecated string
|
|
suppressScript string
|
|
macro string
|
|
prefix []string
|
|
}
|
|
|
|
type builder struct {
|
|
w *gen.CodeWriter
|
|
hw io.Writer // MultiWriter for w and w.Hash
|
|
data *cldr.CLDR
|
|
supp *cldr.SupplementalData
|
|
|
|
// indices
|
|
locale stringSet // common locales
|
|
lang stringSet // canonical language ids (2 or 3 letter ISO codes) with data
|
|
langNoIndex stringSet // 3-letter ISO codes with no associated data
|
|
script stringSet // 4-letter ISO codes
|
|
region stringSet // 2-letter ISO or 3-digit UN M49 codes
|
|
variant stringSet // 4-8-alphanumeric variant code.
|
|
|
|
// Region codes that are groups with their corresponding group IDs.
|
|
groups map[int]index
|
|
|
|
// langInfo
|
|
registry map[string]*ianaEntry
|
|
}
|
|
|
|
type index uint
|
|
|
|
func newBuilder(w *gen.CodeWriter) *builder {
|
|
r := gen.OpenCLDRCoreZip()
|
|
defer r.Close()
|
|
d := &cldr.Decoder{}
|
|
data, err := d.DecodeZip(r)
|
|
failOnError(err)
|
|
b := builder{
|
|
w: w,
|
|
hw: io.MultiWriter(w, w.Hash),
|
|
data: data,
|
|
supp: data.Supplemental(),
|
|
}
|
|
b.parseRegistry()
|
|
return &b
|
|
}
|
|
|
|
func (b *builder) parseRegistry() {
|
|
r := gen.OpenIANAFile("assignments/language-subtag-registry")
|
|
defer r.Close()
|
|
b.registry = make(map[string]*ianaEntry)
|
|
|
|
scan := bufio.NewScanner(r)
|
|
scan.Split(bufio.ScanWords)
|
|
var record *ianaEntry
|
|
for more := scan.Scan(); more; {
|
|
key := scan.Text()
|
|
more = scan.Scan()
|
|
value := scan.Text()
|
|
switch key {
|
|
case "Type:":
|
|
record = &ianaEntry{typ: value}
|
|
case "Subtag:", "Tag:":
|
|
if s := strings.SplitN(value, "..", 2); len(s) > 1 {
|
|
for a := s[0]; a <= s[1]; a = inc(a) {
|
|
b.addToRegistry(a, record)
|
|
}
|
|
} else {
|
|
b.addToRegistry(value, record)
|
|
}
|
|
case "Suppress-Script:":
|
|
record.suppressScript = value
|
|
case "Added:":
|
|
record.added = value
|
|
case "Deprecated:":
|
|
record.deprecated = value
|
|
case "Macrolanguage:":
|
|
record.macro = value
|
|
case "Preferred-Value:":
|
|
record.preferred = value
|
|
case "Prefix:":
|
|
record.prefix = append(record.prefix, value)
|
|
case "Scope:":
|
|
record.scope = value
|
|
case "Description:":
|
|
buf := []byte(value)
|
|
for more = scan.Scan(); more; more = scan.Scan() {
|
|
b := scan.Bytes()
|
|
if b[0] == '%' || b[len(b)-1] == ':' {
|
|
break
|
|
}
|
|
buf = append(buf, ' ')
|
|
buf = append(buf, b...)
|
|
}
|
|
record.description = append(record.description, string(buf))
|
|
continue
|
|
default:
|
|
continue
|
|
}
|
|
more = scan.Scan()
|
|
}
|
|
if scan.Err() != nil {
|
|
log.Panic(scan.Err())
|
|
}
|
|
}
|
|
|
|
func (b *builder) addToRegistry(key string, entry *ianaEntry) {
|
|
if info, ok := b.registry[key]; ok {
|
|
if info.typ != "language" || entry.typ != "extlang" {
|
|
log.Fatalf("parseRegistry: tag %q already exists", key)
|
|
}
|
|
} else {
|
|
b.registry[key] = entry
|
|
}
|
|
}
|
|
|
|
var commentIndex = make(map[string]string)
|
|
|
|
func init() {
|
|
for _, s := range comment {
|
|
key := strings.TrimSpace(strings.SplitN(s, " ", 2)[0])
|
|
commentIndex[key] = s
|
|
}
|
|
}
|
|
|
|
func (b *builder) comment(name string) {
|
|
if s := commentIndex[name]; len(s) > 0 {
|
|
b.w.WriteComment(s)
|
|
} else {
|
|
fmt.Fprintln(b.w)
|
|
}
|
|
}
|
|
|
|
func (b *builder) pf(f string, x ...interface{}) {
|
|
fmt.Fprintf(b.hw, f, x...)
|
|
fmt.Fprint(b.hw, "\n")
|
|
}
|
|
|
|
func (b *builder) p(x ...interface{}) {
|
|
fmt.Fprintln(b.hw, x...)
|
|
}
|
|
|
|
func (b *builder) addSize(s int) {
|
|
b.w.Size += s
|
|
b.pf("// Size: %d bytes", s)
|
|
}
|
|
|
|
func (b *builder) writeConst(name string, x interface{}) {
|
|
b.comment(name)
|
|
b.w.WriteConst(name, x)
|
|
}
|
|
|
|
// writeConsts computes f(v) for all v in values and writes the results
|
|
// as constants named _v to a single constant block.
|
|
func (b *builder) writeConsts(f func(string) int, values ...string) {
|
|
b.pf("const (")
|
|
for _, v := range values {
|
|
b.pf("\t_%s = %v", v, f(v))
|
|
}
|
|
b.pf(")")
|
|
}
|
|
|
|
// writeType writes the type of the given value, which must be a struct.
|
|
func (b *builder) writeType(value interface{}) {
|
|
b.comment(reflect.TypeOf(value).Name())
|
|
b.w.WriteType(value)
|
|
}
|
|
|
|
func (b *builder) writeSlice(name string, ss interface{}) {
|
|
b.writeSliceAddSize(name, 0, ss)
|
|
}
|
|
|
|
func (b *builder) writeSliceAddSize(name string, extraSize int, ss interface{}) {
|
|
b.comment(name)
|
|
b.w.Size += extraSize
|
|
v := reflect.ValueOf(ss)
|
|
t := v.Type().Elem()
|
|
b.pf("// Size: %d bytes, %d elements", v.Len()*int(t.Size())+extraSize, v.Len())
|
|
|
|
fmt.Fprintf(b.w, "var %s = ", name)
|
|
b.w.WriteArray(ss)
|
|
b.p()
|
|
}
|
|
|
|
type fromTo struct {
|
|
from, to uint16
|
|
}
|
|
|
|
func (b *builder) writeSortedMap(name string, ss *stringSet, index func(s string) uint16) {
|
|
ss.sortFunc(func(a, b string) bool {
|
|
return index(a) < index(b)
|
|
})
|
|
m := []fromTo{}
|
|
for _, s := range ss.s {
|
|
m = append(m, fromTo{index(s), index(ss.update[s])})
|
|
}
|
|
b.writeSlice(name, m)
|
|
}
|
|
|
|
const base = 'z' - 'a' + 1
|
|
|
|
func strToInt(s string) uint {
|
|
v := uint(0)
|
|
for i := 0; i < len(s); i++ {
|
|
v *= base
|
|
v += uint(s[i] - 'a')
|
|
}
|
|
return v
|
|
}
|
|
|
|
// converts the given integer to the original ASCII string passed to strToInt.
|
|
// len(s) must match the number of characters obtained.
|
|
func intToStr(v uint, s []byte) {
|
|
for i := len(s) - 1; i >= 0; i-- {
|
|
s[i] = byte(v%base) + 'a'
|
|
v /= base
|
|
}
|
|
}
|
|
|
|
func (b *builder) writeBitVector(name string, ss []string) {
|
|
vec := make([]uint8, int(math.Ceil(math.Pow(base, float64(len(ss[0])))/8)))
|
|
for _, s := range ss {
|
|
v := strToInt(s)
|
|
vec[v/8] |= 1 << (v % 8)
|
|
}
|
|
b.writeSlice(name, vec)
|
|
}
|
|
|
|
// TODO: convert this type into a list or two-stage trie.
|
|
func (b *builder) writeMapFunc(name string, m map[string]string, f func(string) uint16) {
|
|
b.comment(name)
|
|
v := reflect.ValueOf(m)
|
|
sz := v.Len() * (2 + int(v.Type().Key().Size()))
|
|
for _, k := range m {
|
|
sz += len(k)
|
|
}
|
|
b.addSize(sz)
|
|
keys := []string{}
|
|
b.pf(`var %s = map[string]uint16{`, name)
|
|
for k := range m {
|
|
keys = append(keys, k)
|
|
}
|
|
sort.Strings(keys)
|
|
for _, k := range keys {
|
|
b.pf("\t%q: %v,", k, f(m[k]))
|
|
}
|
|
b.p("}")
|
|
}
|
|
|
|
func (b *builder) writeMap(name string, m interface{}) {
|
|
b.comment(name)
|
|
v := reflect.ValueOf(m)
|
|
sz := v.Len() * (2 + int(v.Type().Key().Size()) + int(v.Type().Elem().Size()))
|
|
b.addSize(sz)
|
|
f := strings.FieldsFunc(fmt.Sprintf("%#v", m), func(r rune) bool {
|
|
return strings.IndexRune("{}, ", r) != -1
|
|
})
|
|
sort.Strings(f[1:])
|
|
b.pf(`var %s = %s{`, name, f[0])
|
|
for _, kv := range f[1:] {
|
|
b.pf("\t%s,", kv)
|
|
}
|
|
b.p("}")
|
|
}
|
|
|
|
func (b *builder) langIndex(s string) uint16 {
|
|
if s == "und" {
|
|
return 0
|
|
}
|
|
if i, ok := b.lang.find(s); ok {
|
|
return uint16(i)
|
|
}
|
|
return uint16(strToInt(s)) + uint16(len(b.lang.s))
|
|
}
|
|
|
|
// inc advances the string to its lexicographical successor.
|
|
func inc(s string) string {
|
|
const maxTagLength = 4
|
|
var buf [maxTagLength]byte
|
|
intToStr(strToInt(strings.ToLower(s))+1, buf[:len(s)])
|
|
for i := 0; i < len(s); i++ {
|
|
if s[i] <= 'Z' {
|
|
buf[i] -= 'a' - 'A'
|
|
}
|
|
}
|
|
return string(buf[:len(s)])
|
|
}
|
|
|
|
func (b *builder) parseIndices() {
|
|
meta := b.supp.Metadata
|
|
|
|
for k, v := range b.registry {
|
|
var ss *stringSet
|
|
switch v.typ {
|
|
case "language":
|
|
if len(k) == 2 || v.suppressScript != "" || v.scope == "special" {
|
|
b.lang.add(k)
|
|
continue
|
|
} else {
|
|
ss = &b.langNoIndex
|
|
}
|
|
case "region":
|
|
ss = &b.region
|
|
case "script":
|
|
ss = &b.script
|
|
case "variant":
|
|
ss = &b.variant
|
|
default:
|
|
continue
|
|
}
|
|
ss.add(k)
|
|
}
|
|
// Include any language for which there is data.
|
|
for _, lang := range b.data.Locales() {
|
|
if x := b.data.RawLDML(lang); false ||
|
|
x.LocaleDisplayNames != nil ||
|
|
x.Characters != nil ||
|
|
x.Delimiters != nil ||
|
|
x.Measurement != nil ||
|
|
x.Dates != nil ||
|
|
x.Numbers != nil ||
|
|
x.Units != nil ||
|
|
x.ListPatterns != nil ||
|
|
x.Collations != nil ||
|
|
x.Segmentations != nil ||
|
|
x.Rbnf != nil ||
|
|
x.Annotations != nil ||
|
|
x.Metadata != nil {
|
|
|
|
from := strings.Split(lang, "_")
|
|
if lang := from[0]; lang != "root" {
|
|
b.lang.add(lang)
|
|
}
|
|
}
|
|
}
|
|
// Include locales for plural rules, which uses a different structure.
|
|
for _, plurals := range b.data.Supplemental().Plurals {
|
|
for _, rules := range plurals.PluralRules {
|
|
for _, lang := range strings.Split(rules.Locales, " ") {
|
|
if lang = strings.Split(lang, "_")[0]; lang != "root" {
|
|
b.lang.add(lang)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
// Include languages in likely subtags.
|
|
for _, m := range b.supp.LikelySubtags.LikelySubtag {
|
|
from := strings.Split(m.From, "_")
|
|
b.lang.add(from[0])
|
|
}
|
|
// Include ISO-639 alpha-3 bibliographic entries.
|
|
for _, a := range meta.Alias.LanguageAlias {
|
|
if a.Reason == "bibliographic" {
|
|
b.langNoIndex.add(a.Type)
|
|
}
|
|
}
|
|
// Include regions in territoryAlias (not all are in the IANA registry!)
|
|
for _, reg := range b.supp.Metadata.Alias.TerritoryAlias {
|
|
if len(reg.Type) == 2 {
|
|
b.region.add(reg.Type)
|
|
}
|
|
}
|
|
|
|
for _, s := range b.lang.s {
|
|
if len(s) == 3 {
|
|
b.langNoIndex.remove(s)
|
|
}
|
|
}
|
|
b.writeConst("numLanguages", len(b.lang.slice())+len(b.langNoIndex.slice()))
|
|
b.writeConst("numScripts", len(b.script.slice()))
|
|
b.writeConst("numRegions", len(b.region.slice()))
|
|
|
|
// Add dummy codes at the start of each list to represent "unspecified".
|
|
b.lang.add("---")
|
|
b.script.add("----")
|
|
b.region.add("---")
|
|
|
|
// common locales
|
|
b.locale.parse(meta.DefaultContent.Locales)
|
|
}
|
|
|
|
// TODO: region inclusion data will probably not be use used in future matchers.
|
|
|
|
func (b *builder) computeRegionGroups() {
|
|
b.groups = make(map[int]index)
|
|
|
|
// Create group indices.
|
|
for i := 1; b.region.s[i][0] < 'A'; i++ { // Base M49 indices on regionID.
|
|
b.groups[i] = index(len(b.groups))
|
|
}
|
|
for _, g := range b.supp.TerritoryContainment.Group {
|
|
// Skip UN and EURO zone as they are flattening the containment
|
|
// relationship.
|
|
if g.Type == "EZ" || g.Type == "UN" {
|
|
continue
|
|
}
|
|
group := b.region.index(g.Type)
|
|
if _, ok := b.groups[group]; !ok {
|
|
b.groups[group] = index(len(b.groups))
|
|
}
|
|
}
|
|
if len(b.groups) > 64 {
|
|
log.Fatalf("only 64 groups supported, found %d", len(b.groups))
|
|
}
|
|
b.writeConst("nRegionGroups", len(b.groups))
|
|
}
|
|
|
|
var langConsts = []string{
|
|
"af", "am", "ar", "az", "bg", "bn", "ca", "cs", "da", "de", "el", "en", "es",
|
|
"et", "fa", "fi", "fil", "fr", "gu", "he", "hi", "hr", "hu", "hy", "id", "is",
|
|
"it", "ja", "ka", "kk", "km", "kn", "ko", "ky", "lo", "lt", "lv", "mk", "ml",
|
|
"mn", "mo", "mr", "ms", "mul", "my", "nb", "ne", "nl", "no", "pa", "pl", "pt",
|
|
"ro", "ru", "sh", "si", "sk", "sl", "sq", "sr", "sv", "sw", "ta", "te", "th",
|
|
"tl", "tn", "tr", "uk", "ur", "uz", "vi", "zh", "zu",
|
|
|
|
// constants for grandfathered tags (if not already defined)
|
|
"jbo", "ami", "bnn", "hak", "tlh", "lb", "nv", "pwn", "tao", "tay", "tsu",
|
|
"nn", "sfb", "vgt", "sgg", "cmn", "nan", "hsn",
|
|
}
|
|
|
|
// writeLanguage generates all tables needed for language canonicalization.
|
|
func (b *builder) writeLanguage() {
|
|
meta := b.supp.Metadata
|
|
|
|
b.writeConst("nonCanonicalUnd", b.lang.index("und"))
|
|
b.writeConsts(func(s string) int { return int(b.langIndex(s)) }, langConsts...)
|
|
b.writeConst("langPrivateStart", b.langIndex("qaa"))
|
|
b.writeConst("langPrivateEnd", b.langIndex("qtz"))
|
|
|
|
// Get language codes that need to be mapped (overlong 3-letter codes,
|
|
// deprecated 2-letter codes, legacy and grandfathered tags.)
|
|
langAliasMap := stringSet{}
|
|
aliasTypeMap := map[string]langAliasType{}
|
|
|
|
// altLangISO3 get the alternative ISO3 names that need to be mapped.
|
|
altLangISO3 := stringSet{}
|
|
// Add dummy start to avoid the use of index 0.
|
|
altLangISO3.add("---")
|
|
altLangISO3.updateLater("---", "aa")
|
|
|
|
lang := b.lang.clone()
|
|
for _, a := range meta.Alias.LanguageAlias {
|
|
if a.Replacement == "" {
|
|
a.Replacement = "und"
|
|
}
|
|
// TODO: support mapping to tags
|
|
repl := strings.SplitN(a.Replacement, "_", 2)[0]
|
|
if a.Reason == "overlong" {
|
|
if len(a.Replacement) == 2 && len(a.Type) == 3 {
|
|
lang.updateLater(a.Replacement, a.Type)
|
|
}
|
|
} else if len(a.Type) <= 3 {
|
|
switch a.Reason {
|
|
case "macrolanguage":
|
|
aliasTypeMap[a.Type] = langMacro
|
|
case "deprecated":
|
|
// handled elsewhere
|
|
continue
|
|
case "bibliographic", "legacy":
|
|
if a.Type == "no" {
|
|
continue
|
|
}
|
|
aliasTypeMap[a.Type] = langLegacy
|
|
default:
|
|
log.Fatalf("new %s alias: %s", a.Reason, a.Type)
|
|
}
|
|
langAliasMap.add(a.Type)
|
|
langAliasMap.updateLater(a.Type, repl)
|
|
}
|
|
}
|
|
// Manually add the mapping of "nb" (Norwegian) to its macro language.
|
|
// This can be removed if CLDR adopts this change.
|
|
langAliasMap.add("nb")
|
|
langAliasMap.updateLater("nb", "no")
|
|
aliasTypeMap["nb"] = langMacro
|
|
|
|
for k, v := range b.registry {
|
|
// Also add deprecated values for 3-letter ISO codes, which CLDR omits.
|
|
if v.typ == "language" && v.deprecated != "" && v.preferred != "" {
|
|
langAliasMap.add(k)
|
|
langAliasMap.updateLater(k, v.preferred)
|
|
aliasTypeMap[k] = langDeprecated
|
|
}
|
|
}
|
|
// Fix CLDR mappings.
|
|
lang.updateLater("tl", "tgl")
|
|
lang.updateLater("sh", "hbs")
|
|
lang.updateLater("mo", "mol")
|
|
lang.updateLater("no", "nor")
|
|
lang.updateLater("tw", "twi")
|
|
lang.updateLater("nb", "nob")
|
|
lang.updateLater("ak", "aka")
|
|
lang.updateLater("bh", "bih")
|
|
|
|
// Ensure that each 2-letter code is matched with a 3-letter code.
|
|
for _, v := range lang.s[1:] {
|
|
s, ok := lang.update[v]
|
|
if !ok {
|
|
if s, ok = lang.update[langAliasMap.update[v]]; !ok {
|
|
continue
|
|
}
|
|
lang.update[v] = s
|
|
}
|
|
if v[0] != s[0] {
|
|
altLangISO3.add(s)
|
|
altLangISO3.updateLater(s, v)
|
|
}
|
|
}
|
|
|
|
// Complete canonicalized language tags.
|
|
lang.freeze()
|
|
for i, v := range lang.s {
|
|
// We can avoid these manual entries by using the IANA registry directly.
|
|
// Seems easier to update the list manually, as changes are rare.
|
|
// The panic in this loop will trigger if we miss an entry.
|
|
add := ""
|
|
if s, ok := lang.update[v]; ok {
|
|
if s[0] == v[0] {
|
|
add = s[1:]
|
|
} else {
|
|
add = string([]byte{0, byte(altLangISO3.index(s))})
|
|
}
|
|
} else if len(v) == 3 {
|
|
add = "\x00"
|
|
} else {
|
|
log.Panicf("no data for long form of %q", v)
|
|
}
|
|
lang.s[i] += add
|
|
}
|
|
b.writeConst("lang", tag.Index(lang.join()))
|
|
|
|
b.writeConst("langNoIndexOffset", len(b.lang.s))
|
|
|
|
// space of all valid 3-letter language identifiers.
|
|
b.writeBitVector("langNoIndex", b.langNoIndex.slice())
|
|
|
|
altLangIndex := []uint16{}
|
|
for i, s := range altLangISO3.slice() {
|
|
altLangISO3.s[i] += string([]byte{byte(len(altLangIndex))})
|
|
if i > 0 {
|
|
idx := b.lang.index(altLangISO3.update[s])
|
|
altLangIndex = append(altLangIndex, uint16(idx))
|
|
}
|
|
}
|
|
b.writeConst("altLangISO3", tag.Index(altLangISO3.join()))
|
|
b.writeSlice("altLangIndex", altLangIndex)
|
|
|
|
b.writeSortedMap("langAliasMap", &langAliasMap, b.langIndex)
|
|
types := make([]langAliasType, len(langAliasMap.s))
|
|
for i, s := range langAliasMap.s {
|
|
types[i] = aliasTypeMap[s]
|
|
}
|
|
b.writeSlice("langAliasTypes", types)
|
|
}
|
|
|
|
var scriptConsts = []string{
|
|
"Latn", "Hani", "Hans", "Hant", "Qaaa", "Qaai", "Qabx", "Zinh", "Zyyy",
|
|
"Zzzz",
|
|
}
|
|
|
|
func (b *builder) writeScript() {
|
|
b.writeConsts(b.script.index, scriptConsts...)
|
|
b.writeConst("script", tag.Index(b.script.join()))
|
|
|
|
supp := make([]uint8, len(b.lang.slice()))
|
|
for i, v := range b.lang.slice()[1:] {
|
|
if sc := b.registry[v].suppressScript; sc != "" {
|
|
supp[i+1] = uint8(b.script.index(sc))
|
|
}
|
|
}
|
|
b.writeSlice("suppressScript", supp)
|
|
|
|
// There is only one deprecated script in CLDR. This value is hard-coded.
|
|
// We check here if the code must be updated.
|
|
for _, a := range b.supp.Metadata.Alias.ScriptAlias {
|
|
if a.Type != "Qaai" {
|
|
log.Panicf("unexpected deprecated stript %q", a.Type)
|
|
}
|
|
}
|
|
}
|
|
|
|
func parseM49(s string) int16 {
|
|
if len(s) == 0 {
|
|
return 0
|
|
}
|
|
v, err := strconv.ParseUint(s, 10, 10)
|
|
failOnError(err)
|
|
return int16(v)
|
|
}
|
|
|
|
var regionConsts = []string{
|
|
"001", "419", "BR", "CA", "ES", "GB", "MD", "PT", "UK", "US",
|
|
"ZZ", "XA", "XC", "XK", // Unofficial tag for Kosovo.
|
|
}
|
|
|
|
func (b *builder) writeRegion() {
|
|
b.writeConsts(b.region.index, regionConsts...)
|
|
|
|
isoOffset := b.region.index("AA")
|
|
m49map := make([]int16, len(b.region.slice()))
|
|
fromM49map := make(map[int16]int)
|
|
altRegionISO3 := ""
|
|
altRegionIDs := []uint16{}
|
|
|
|
b.writeConst("isoRegionOffset", isoOffset)
|
|
|
|
// 2-letter region lookup and mapping to numeric codes.
|
|
regionISO := b.region.clone()
|
|
regionISO.s = regionISO.s[isoOffset:]
|
|
regionISO.sorted = false
|
|
|
|
regionTypes := make([]byte, len(b.region.s))
|
|
|
|
// Is the region valid BCP 47?
|
|
for s, e := range b.registry {
|
|
if len(s) == 2 && s == strings.ToUpper(s) {
|
|
i := b.region.index(s)
|
|
for _, d := range e.description {
|
|
if strings.Contains(d, "Private use") {
|
|
regionTypes[i] = iso3166UserAssigned
|
|
}
|
|
}
|
|
regionTypes[i] |= bcp47Region
|
|
}
|
|
}
|
|
|
|
// Is the region a valid ccTLD?
|
|
r := gen.OpenIANAFile("domains/root/db")
|
|
defer r.Close()
|
|
|
|
buf, err := ioutil.ReadAll(r)
|
|
failOnError(err)
|
|
re := regexp.MustCompile(`"/domains/root/db/([a-z]{2}).html"`)
|
|
for _, m := range re.FindAllSubmatch(buf, -1) {
|
|
i := b.region.index(strings.ToUpper(string(m[1])))
|
|
regionTypes[i] |= ccTLD
|
|
}
|
|
|
|
b.writeSlice("regionTypes", regionTypes)
|
|
|
|
iso3Set := make(map[string]int)
|
|
update := func(iso2, iso3 string) {
|
|
i := regionISO.index(iso2)
|
|
if j, ok := iso3Set[iso3]; !ok && iso3[0] == iso2[0] {
|
|
regionISO.s[i] += iso3[1:]
|
|
iso3Set[iso3] = -1
|
|
} else {
|
|
if ok && j >= 0 {
|
|
regionISO.s[i] += string([]byte{0, byte(j)})
|
|
} else {
|
|
iso3Set[iso3] = len(altRegionISO3)
|
|
regionISO.s[i] += string([]byte{0, byte(len(altRegionISO3))})
|
|
altRegionISO3 += iso3
|
|
altRegionIDs = append(altRegionIDs, uint16(isoOffset+i))
|
|
}
|
|
}
|
|
}
|
|
for _, tc := range b.supp.CodeMappings.TerritoryCodes {
|
|
i := regionISO.index(tc.Type) + isoOffset
|
|
if d := m49map[i]; d != 0 {
|
|
log.Panicf("%s found as a duplicate UN.M49 code of %03d", tc.Numeric, d)
|
|
}
|
|
m49 := parseM49(tc.Numeric)
|
|
m49map[i] = m49
|
|
if r := fromM49map[m49]; r == 0 {
|
|
fromM49map[m49] = i
|
|
} else if r != i {
|
|
dep := b.registry[regionISO.s[r-isoOffset]].deprecated
|
|
if t := b.registry[tc.Type]; t != nil && dep != "" && (t.deprecated == "" || t.deprecated > dep) {
|
|
fromM49map[m49] = i
|
|
}
|
|
}
|
|
}
|
|
for _, ta := range b.supp.Metadata.Alias.TerritoryAlias {
|
|
if len(ta.Type) == 3 && ta.Type[0] <= '9' && len(ta.Replacement) == 2 {
|
|
from := parseM49(ta.Type)
|
|
if r := fromM49map[from]; r == 0 {
|
|
fromM49map[from] = regionISO.index(ta.Replacement) + isoOffset
|
|
}
|
|
}
|
|
}
|
|
for _, tc := range b.supp.CodeMappings.TerritoryCodes {
|
|
if len(tc.Alpha3) == 3 {
|
|
update(tc.Type, tc.Alpha3)
|
|
}
|
|
}
|
|
// This entries are not included in territoryCodes. Mostly 3-letter variants
|
|
// of deleted codes and an entry for QU.
|
|
for _, m := range []struct{ iso2, iso3 string }{
|
|
{"CT", "CTE"},
|
|
{"DY", "DHY"},
|
|
{"HV", "HVO"},
|
|
{"JT", "JTN"},
|
|
{"MI", "MID"},
|
|
{"NH", "NHB"},
|
|
{"NQ", "ATN"},
|
|
{"PC", "PCI"},
|
|
{"PU", "PUS"},
|
|
{"PZ", "PCZ"},
|
|
{"RH", "RHO"},
|
|
{"VD", "VDR"},
|
|
{"WK", "WAK"},
|
|
// These three-letter codes are used for others as well.
|
|
{"FQ", "ATF"},
|
|
} {
|
|
update(m.iso2, m.iso3)
|
|
}
|
|
for i, s := range regionISO.s {
|
|
if len(s) != 4 {
|
|
regionISO.s[i] = s + " "
|
|
}
|
|
}
|
|
b.writeConst("regionISO", tag.Index(regionISO.join()))
|
|
b.writeConst("altRegionISO3", altRegionISO3)
|
|
b.writeSlice("altRegionIDs", altRegionIDs)
|
|
|
|
// Create list of deprecated regions.
|
|
// TODO: consider inserting SF -> FI. Not included by CLDR, but is the only
|
|
// Transitionally-reserved mapping not included.
|
|
regionOldMap := stringSet{}
|
|
// Include regions in territoryAlias (not all are in the IANA registry!)
|
|
for _, reg := range b.supp.Metadata.Alias.TerritoryAlias {
|
|
if len(reg.Type) == 2 && reg.Reason == "deprecated" && len(reg.Replacement) == 2 {
|
|
regionOldMap.add(reg.Type)
|
|
regionOldMap.updateLater(reg.Type, reg.Replacement)
|
|
i, _ := regionISO.find(reg.Type)
|
|
j, _ := regionISO.find(reg.Replacement)
|
|
if k := m49map[i+isoOffset]; k == 0 {
|
|
m49map[i+isoOffset] = m49map[j+isoOffset]
|
|
}
|
|
}
|
|
}
|
|
b.writeSortedMap("regionOldMap", ®ionOldMap, func(s string) uint16 {
|
|
return uint16(b.region.index(s))
|
|
})
|
|
// 3-digit region lookup, groupings.
|
|
for i := 1; i < isoOffset; i++ {
|
|
m := parseM49(b.region.s[i])
|
|
m49map[i] = m
|
|
fromM49map[m] = i
|
|
}
|
|
b.writeSlice("m49", m49map)
|
|
|
|
const (
|
|
searchBits = 7
|
|
regionBits = 9
|
|
)
|
|
if len(m49map) >= 1<<regionBits {
|
|
log.Fatalf("Maximum number of regions exceeded: %d > %d", len(m49map), 1<<regionBits)
|
|
}
|
|
m49Index := [9]int16{}
|
|
fromM49 := []uint16{}
|
|
m49 := []int{}
|
|
for k, _ := range fromM49map {
|
|
m49 = append(m49, int(k))
|
|
}
|
|
sort.Ints(m49)
|
|
for _, k := range m49[1:] {
|
|
val := (k & (1<<searchBits - 1)) << regionBits
|
|
fromM49 = append(fromM49, uint16(val|fromM49map[int16(k)]))
|
|
m49Index[1:][k>>searchBits] = int16(len(fromM49))
|
|
}
|
|
b.writeSlice("m49Index", m49Index)
|
|
b.writeSlice("fromM49", fromM49)
|
|
}
|
|
|
|
const (
|
|
// TODO: put these lists in regionTypes as user data? Could be used for
|
|
// various optimizations and refinements and could be exposed in the API.
|
|
iso3166Except = "AC CP DG EA EU FX IC SU TA UK"
|
|
iso3166Trans = "AN BU CS NT TP YU ZR" // SF is not in our set of Regions.
|
|
// DY and RH are actually not deleted, but indeterminately reserved.
|
|
iso3166DelCLDR = "CT DD DY FQ HV JT MI NH NQ PC PU PZ RH VD WK YD"
|
|
)
|
|
|
|
const (
|
|
iso3166UserAssigned = 1 << iota
|
|
ccTLD
|
|
bcp47Region
|
|
)
|
|
|
|
func find(list []string, s string) int {
|
|
for i, t := range list {
|
|
if t == s {
|
|
return i
|
|
}
|
|
}
|
|
return -1
|
|
}
|
|
|
|
// writeVariants generates per-variant information and creates a map from variant
|
|
// name to index value. We assign index values such that sorting multiple
|
|
// variants by index value will result in the correct order.
|
|
// There are two types of variants: specialized and general. Specialized variants
|
|
// are only applicable to certain language or language-script pairs. Generalized
|
|
// variants apply to any language. Generalized variants always sort after
|
|
// specialized variants. We will therefore always assign a higher index value
|
|
// to a generalized variant than any other variant. Generalized variants are
|
|
// sorted alphabetically among themselves.
|
|
// Specialized variants may also sort after other specialized variants. Such
|
|
// variants will be ordered after any of the variants they may follow.
|
|
// We assume that if a variant x is followed by a variant y, then for any prefix
|
|
// p of x, p-x is a prefix of y. This allows us to order tags based on the
|
|
// maximum of the length of any of its prefixes.
|
|
// TODO: it is possible to define a set of Prefix values on variants such that
|
|
// a total order cannot be defined to the point that this algorithm breaks.
|
|
// In other words, we cannot guarantee the same order of variants for the
|
|
// future using the same algorithm or for non-compliant combinations of
|
|
// variants. For this reason, consider using simple alphabetic sorting
|
|
// of variants and ignore Prefix restrictions altogether.
|
|
func (b *builder) writeVariant() {
|
|
generalized := stringSet{}
|
|
specialized := stringSet{}
|
|
specializedExtend := stringSet{}
|
|
// Collate the variants by type and check assumptions.
|
|
for _, v := range b.variant.slice() {
|
|
e := b.registry[v]
|
|
if len(e.prefix) == 0 {
|
|
generalized.add(v)
|
|
continue
|
|
}
|
|
c := strings.Split(e.prefix[0], "-")
|
|
hasScriptOrRegion := false
|
|
if len(c) > 1 {
|
|
_, hasScriptOrRegion = b.script.find(c[1])
|
|
if !hasScriptOrRegion {
|
|
_, hasScriptOrRegion = b.region.find(c[1])
|
|
|
|
}
|
|
}
|
|
if len(c) == 1 || len(c) == 2 && hasScriptOrRegion {
|
|
// Variant is preceded by a language.
|
|
specialized.add(v)
|
|
continue
|
|
}
|
|
// Variant is preceded by another variant.
|
|
specializedExtend.add(v)
|
|
prefix := c[0] + "-"
|
|
if hasScriptOrRegion {
|
|
prefix += c[1]
|
|
}
|
|
for _, p := range e.prefix {
|
|
// Verify that the prefix minus the last element is a prefix of the
|
|
// predecessor element.
|
|
i := strings.LastIndex(p, "-")
|
|
pred := b.registry[p[i+1:]]
|
|
if find(pred.prefix, p[:i]) < 0 {
|
|
log.Fatalf("prefix %q for variant %q not consistent with predecessor spec", p, v)
|
|
}
|
|
// The sorting used below does not work in the general case. It works
|
|
// if we assume that variants that may be followed by others only have
|
|
// prefixes of the same length. Verify this.
|
|
count := strings.Count(p[:i], "-")
|
|
for _, q := range pred.prefix {
|
|
if c := strings.Count(q, "-"); c != count {
|
|
log.Fatalf("variant %q preceding %q has a prefix %q of size %d; want %d", p[i+1:], v, q, c, count)
|
|
}
|
|
}
|
|
if !strings.HasPrefix(p, prefix) {
|
|
log.Fatalf("prefix %q of variant %q should start with %q", p, v, prefix)
|
|
}
|
|
}
|
|
}
|
|
|
|
// Sort extended variants.
|
|
a := specializedExtend.s
|
|
less := func(v, w string) bool {
|
|
// Sort by the maximum number of elements.
|
|
maxCount := func(s string) (max int) {
|
|
for _, p := range b.registry[s].prefix {
|
|
if c := strings.Count(p, "-"); c > max {
|
|
max = c
|
|
}
|
|
}
|
|
return
|
|
}
|
|
if cv, cw := maxCount(v), maxCount(w); cv != cw {
|
|
return cv < cw
|
|
}
|
|
// Sort by name as tie breaker.
|
|
return v < w
|
|
}
|
|
sort.Sort(funcSorter{less, sort.StringSlice(a)})
|
|
specializedExtend.frozen = true
|
|
|
|
// Create index from variant name to index.
|
|
variantIndex := make(map[string]uint8)
|
|
add := func(s []string) {
|
|
for _, v := range s {
|
|
variantIndex[v] = uint8(len(variantIndex))
|
|
}
|
|
}
|
|
add(specialized.slice())
|
|
add(specializedExtend.s)
|
|
numSpecialized := len(variantIndex)
|
|
add(generalized.slice())
|
|
if n := len(variantIndex); n > 255 {
|
|
log.Fatalf("maximum number of variants exceeded: was %d; want <= 255", n)
|
|
}
|
|
b.writeMap("variantIndex", variantIndex)
|
|
b.writeConst("variantNumSpecialized", numSpecialized)
|
|
}
|
|
|
|
func (b *builder) writeLanguageInfo() {
|
|
}
|
|
|
|
// writeLikelyData writes tables that are used both for finding parent relations and for
|
|
// language matching. Each entry contains additional bits to indicate the status of the
|
|
// data to know when it cannot be used for parent relations.
|
|
func (b *builder) writeLikelyData() {
|
|
const (
|
|
isList = 1 << iota
|
|
scriptInFrom
|
|
regionInFrom
|
|
)
|
|
type ( // generated types
|
|
likelyScriptRegion struct {
|
|
region uint16
|
|
script uint8
|
|
flags uint8
|
|
}
|
|
likelyLangScript struct {
|
|
lang uint16
|
|
script uint8
|
|
flags uint8
|
|
}
|
|
likelyLangRegion struct {
|
|
lang uint16
|
|
region uint16
|
|
}
|
|
// likelyTag is used for getting likely tags for group regions, where
|
|
// the likely region might be a region contained in the group.
|
|
likelyTag struct {
|
|
lang uint16
|
|
region uint16
|
|
script uint8
|
|
}
|
|
)
|
|
var ( // generated variables
|
|
likelyRegionGroup = make([]likelyTag, len(b.groups))
|
|
likelyLang = make([]likelyScriptRegion, len(b.lang.s))
|
|
likelyRegion = make([]likelyLangScript, len(b.region.s))
|
|
likelyScript = make([]likelyLangRegion, len(b.script.s))
|
|
likelyLangList = []likelyScriptRegion{}
|
|
likelyRegionList = []likelyLangScript{}
|
|
)
|
|
type fromTo struct {
|
|
from, to []string
|
|
}
|
|
langToOther := map[int][]fromTo{}
|
|
regionToOther := map[int][]fromTo{}
|
|
for _, m := range b.supp.LikelySubtags.LikelySubtag {
|
|
from := strings.Split(m.From, "_")
|
|
to := strings.Split(m.To, "_")
|
|
if len(to) != 3 {
|
|
log.Fatalf("invalid number of subtags in %q: found %d, want 3", m.To, len(to))
|
|
}
|
|
if len(from) > 3 {
|
|
log.Fatalf("invalid number of subtags: found %d, want 1-3", len(from))
|
|
}
|
|
if from[0] != to[0] && from[0] != "und" {
|
|
log.Fatalf("unexpected language change in expansion: %s -> %s", from, to)
|
|
}
|
|
if len(from) == 3 {
|
|
if from[2] != to[2] {
|
|
log.Fatalf("unexpected region change in expansion: %s -> %s", from, to)
|
|
}
|
|
if from[0] != "und" {
|
|
log.Fatalf("unexpected fully specified from tag: %s -> %s", from, to)
|
|
}
|
|
}
|
|
if len(from) == 1 || from[0] != "und" {
|
|
id := 0
|
|
if from[0] != "und" {
|
|
id = b.lang.index(from[0])
|
|
}
|
|
langToOther[id] = append(langToOther[id], fromTo{from, to})
|
|
} else if len(from) == 2 && len(from[1]) == 4 {
|
|
sid := b.script.index(from[1])
|
|
likelyScript[sid].lang = uint16(b.langIndex(to[0]))
|
|
likelyScript[sid].region = uint16(b.region.index(to[2]))
|
|
} else {
|
|
r := b.region.index(from[len(from)-1])
|
|
if id, ok := b.groups[r]; ok {
|
|
if from[0] != "und" {
|
|
log.Fatalf("region changed unexpectedly: %s -> %s", from, to)
|
|
}
|
|
likelyRegionGroup[id].lang = uint16(b.langIndex(to[0]))
|
|
likelyRegionGroup[id].script = uint8(b.script.index(to[1]))
|
|
likelyRegionGroup[id].region = uint16(b.region.index(to[2]))
|
|
} else {
|
|
regionToOther[r] = append(regionToOther[r], fromTo{from, to})
|
|
}
|
|
}
|
|
}
|
|
b.writeType(likelyLangRegion{})
|
|
b.writeSlice("likelyScript", likelyScript)
|
|
|
|
for id := range b.lang.s {
|
|
list := langToOther[id]
|
|
if len(list) == 1 {
|
|
likelyLang[id].region = uint16(b.region.index(list[0].to[2]))
|
|
likelyLang[id].script = uint8(b.script.index(list[0].to[1]))
|
|
} else if len(list) > 1 {
|
|
likelyLang[id].flags = isList
|
|
likelyLang[id].region = uint16(len(likelyLangList))
|
|
likelyLang[id].script = uint8(len(list))
|
|
for _, x := range list {
|
|
flags := uint8(0)
|
|
if len(x.from) > 1 {
|
|
if x.from[1] == x.to[2] {
|
|
flags = regionInFrom
|
|
} else {
|
|
flags = scriptInFrom
|
|
}
|
|
}
|
|
likelyLangList = append(likelyLangList, likelyScriptRegion{
|
|
region: uint16(b.region.index(x.to[2])),
|
|
script: uint8(b.script.index(x.to[1])),
|
|
flags: flags,
|
|
})
|
|
}
|
|
}
|
|
}
|
|
// TODO: merge suppressScript data with this table.
|
|
b.writeType(likelyScriptRegion{})
|
|
b.writeSlice("likelyLang", likelyLang)
|
|
b.writeSlice("likelyLangList", likelyLangList)
|
|
|
|
for id := range b.region.s {
|
|
list := regionToOther[id]
|
|
if len(list) == 1 {
|
|
likelyRegion[id].lang = uint16(b.langIndex(list[0].to[0]))
|
|
likelyRegion[id].script = uint8(b.script.index(list[0].to[1]))
|
|
if len(list[0].from) > 2 {
|
|
likelyRegion[id].flags = scriptInFrom
|
|
}
|
|
} else if len(list) > 1 {
|
|
likelyRegion[id].flags = isList
|
|
likelyRegion[id].lang = uint16(len(likelyRegionList))
|
|
likelyRegion[id].script = uint8(len(list))
|
|
for i, x := range list {
|
|
if len(x.from) == 2 && i != 0 || i > 0 && len(x.from) != 3 {
|
|
log.Fatalf("unspecified script must be first in list: %v at %d", x.from, i)
|
|
}
|
|
x := likelyLangScript{
|
|
lang: uint16(b.langIndex(x.to[0])),
|
|
script: uint8(b.script.index(x.to[1])),
|
|
}
|
|
if len(list[0].from) > 2 {
|
|
x.flags = scriptInFrom
|
|
}
|
|
likelyRegionList = append(likelyRegionList, x)
|
|
}
|
|
}
|
|
}
|
|
b.writeType(likelyLangScript{})
|
|
b.writeSlice("likelyRegion", likelyRegion)
|
|
b.writeSlice("likelyRegionList", likelyRegionList)
|
|
|
|
b.writeType(likelyTag{})
|
|
b.writeSlice("likelyRegionGroup", likelyRegionGroup)
|
|
}
|
|
|
|
type mutualIntelligibility struct {
|
|
want, have uint16
|
|
distance uint8
|
|
oneway bool
|
|
}
|
|
|
|
type scriptIntelligibility struct {
|
|
wantLang, haveLang uint16
|
|
wantScript, haveScript uint8
|
|
distance uint8
|
|
// Always oneway
|
|
}
|
|
|
|
type regionIntelligibility struct {
|
|
lang uint16 // compact language id
|
|
script uint8 // 0 means any
|
|
group uint8 // 0 means any; if bit 7 is set it means inverse
|
|
distance uint8
|
|
// Always twoway.
|
|
}
|
|
|
|
// writeMatchData writes tables with languages and scripts for which there is
|
|
// mutual intelligibility. The data is based on CLDR's languageMatching data.
|
|
// Note that we use a different algorithm than the one defined by CLDR and that
|
|
// we slightly modify the data. For example, we convert scores to confidence levels.
|
|
// We also drop all region-related data as we use a different algorithm to
|
|
// determine region equivalence.
|
|
func (b *builder) writeMatchData() {
|
|
lm := b.supp.LanguageMatching.LanguageMatches
|
|
cldr.MakeSlice(&lm).SelectAnyOf("type", "written_new")
|
|
|
|
regionHierarchy := map[string][]string{}
|
|
for _, g := range b.supp.TerritoryContainment.Group {
|
|
regions := strings.Split(g.Contains, " ")
|
|
regionHierarchy[g.Type] = append(regionHierarchy[g.Type], regions...)
|
|
}
|
|
regionToGroups := make([]uint8, len(b.region.s))
|
|
|
|
idToIndex := map[string]uint8{}
|
|
for i, mv := range lm[0].MatchVariable {
|
|
if i > 6 {
|
|
log.Fatalf("Too many groups: %d", i)
|
|
}
|
|
idToIndex[mv.Id] = uint8(i + 1)
|
|
// TODO: also handle '-'
|
|
for _, r := range strings.Split(mv.Value, "+") {
|
|
todo := []string{r}
|
|
for k := 0; k < len(todo); k++ {
|
|
r := todo[k]
|
|
regionToGroups[b.region.index(r)] |= 1 << uint8(i)
|
|
todo = append(todo, regionHierarchy[r]...)
|
|
}
|
|
}
|
|
}
|
|
b.writeSlice("regionToGroups", regionToGroups)
|
|
|
|
// maps language id to in- and out-of-group region.
|
|
paradigmLocales := [][3]uint16{}
|
|
locales := strings.Split(lm[0].ParadigmLocales[0].Locales, " ")
|
|
for i := 0; i < len(locales); i += 2 {
|
|
x := [3]uint16{}
|
|
for j := 0; j < 2; j++ {
|
|
pc := strings.SplitN(locales[i+j], "-", 2)
|
|
x[0] = b.langIndex(pc[0])
|
|
if len(pc) == 2 {
|
|
x[1+j] = uint16(b.region.index(pc[1]))
|
|
}
|
|
}
|
|
paradigmLocales = append(paradigmLocales, x)
|
|
}
|
|
b.writeSlice("paradigmLocales", paradigmLocales)
|
|
|
|
b.writeType(mutualIntelligibility{})
|
|
b.writeType(scriptIntelligibility{})
|
|
b.writeType(regionIntelligibility{})
|
|
|
|
matchLang := []mutualIntelligibility{}
|
|
matchScript := []scriptIntelligibility{}
|
|
matchRegion := []regionIntelligibility{}
|
|
// Convert the languageMatch entries in lists keyed by desired language.
|
|
for _, m := range lm[0].LanguageMatch {
|
|
// Different versions of CLDR use different separators.
|
|
desired := strings.Replace(m.Desired, "-", "_", -1)
|
|
supported := strings.Replace(m.Supported, "-", "_", -1)
|
|
d := strings.Split(desired, "_")
|
|
s := strings.Split(supported, "_")
|
|
if len(d) != len(s) {
|
|
log.Fatalf("not supported: desired=%q; supported=%q", desired, supported)
|
|
continue
|
|
}
|
|
distance, _ := strconv.ParseInt(m.Distance, 10, 8)
|
|
switch len(d) {
|
|
case 2:
|
|
if desired == supported && desired == "*_*" {
|
|
continue
|
|
}
|
|
// language-script pair.
|
|
matchScript = append(matchScript, scriptIntelligibility{
|
|
wantLang: uint16(b.langIndex(d[0])),
|
|
haveLang: uint16(b.langIndex(s[0])),
|
|
wantScript: uint8(b.script.index(d[1])),
|
|
haveScript: uint8(b.script.index(s[1])),
|
|
distance: uint8(distance),
|
|
})
|
|
if m.Oneway != "true" {
|
|
matchScript = append(matchScript, scriptIntelligibility{
|
|
wantLang: uint16(b.langIndex(s[0])),
|
|
haveLang: uint16(b.langIndex(d[0])),
|
|
wantScript: uint8(b.script.index(s[1])),
|
|
haveScript: uint8(b.script.index(d[1])),
|
|
distance: uint8(distance),
|
|
})
|
|
}
|
|
case 1:
|
|
if desired == supported && desired == "*" {
|
|
continue
|
|
}
|
|
if distance == 1 {
|
|
// nb == no is already handled by macro mapping. Check there
|
|
// really is only this case.
|
|
if d[0] != "no" || s[0] != "nb" {
|
|
log.Fatalf("unhandled equivalence %s == %s", s[0], d[0])
|
|
}
|
|
continue
|
|
}
|
|
// TODO: consider dropping oneway field and just doubling the entry.
|
|
matchLang = append(matchLang, mutualIntelligibility{
|
|
want: uint16(b.langIndex(d[0])),
|
|
have: uint16(b.langIndex(s[0])),
|
|
distance: uint8(distance),
|
|
oneway: m.Oneway == "true",
|
|
})
|
|
case 3:
|
|
if desired == supported && desired == "*_*_*" {
|
|
continue
|
|
}
|
|
if desired != supported {
|
|
// This is now supported by CLDR, but only one case, which
|
|
// should already be covered by paradigm locales. For instance,
|
|
// test case "und, en, en-GU, en-IN, en-GB ; en-ZA ; en-GB" in
|
|
// testdata/CLDRLocaleMatcherTest.txt tests this.
|
|
if supported != "en_*_GB" {
|
|
log.Fatalf("not supported: desired=%q; supported=%q", desired, supported)
|
|
}
|
|
continue
|
|
}
|
|
ri := regionIntelligibility{
|
|
lang: b.langIndex(d[0]),
|
|
distance: uint8(distance),
|
|
}
|
|
if d[1] != "*" {
|
|
ri.script = uint8(b.script.index(d[1]))
|
|
}
|
|
switch {
|
|
case d[2] == "*":
|
|
ri.group = 0x80 // not contained in anything
|
|
case strings.HasPrefix(d[2], "$!"):
|
|
ri.group = 0x80
|
|
d[2] = "$" + d[2][len("$!"):]
|
|
fallthrough
|
|
case strings.HasPrefix(d[2], "$"):
|
|
ri.group |= idToIndex[d[2]]
|
|
}
|
|
matchRegion = append(matchRegion, ri)
|
|
default:
|
|
log.Fatalf("not supported: desired=%q; supported=%q", desired, supported)
|
|
}
|
|
}
|
|
sort.SliceStable(matchLang, func(i, j int) bool {
|
|
return matchLang[i].distance < matchLang[j].distance
|
|
})
|
|
b.writeSlice("matchLang", matchLang)
|
|
|
|
sort.SliceStable(matchScript, func(i, j int) bool {
|
|
return matchScript[i].distance < matchScript[j].distance
|
|
})
|
|
b.writeSlice("matchScript", matchScript)
|
|
|
|
sort.SliceStable(matchRegion, func(i, j int) bool {
|
|
return matchRegion[i].distance < matchRegion[j].distance
|
|
})
|
|
b.writeSlice("matchRegion", matchRegion)
|
|
}
|
|
|
|
func (b *builder) writeRegionInclusionData() {
|
|
var (
|
|
// mm holds for each group the set of groups with a distance of 1.
|
|
mm = make(map[int][]index)
|
|
|
|
// containment holds for each group the transitive closure of
|
|
// containment of other groups.
|
|
containment = make(map[index][]index)
|
|
)
|
|
for _, g := range b.supp.TerritoryContainment.Group {
|
|
// Skip UN and EURO zone as they are flattening the containment
|
|
// relationship.
|
|
if g.Type == "EZ" || g.Type == "UN" {
|
|
continue
|
|
}
|
|
group := b.region.index(g.Type)
|
|
groupIdx := b.groups[group]
|
|
for _, mem := range strings.Split(g.Contains, " ") {
|
|
r := b.region.index(mem)
|
|
mm[r] = append(mm[r], groupIdx)
|
|
if g, ok := b.groups[r]; ok {
|
|
mm[group] = append(mm[group], g)
|
|
containment[groupIdx] = append(containment[groupIdx], g)
|
|
}
|
|
}
|
|
}
|
|
|
|
regionContainment := make([]uint64, len(b.groups))
|
|
for _, g := range b.groups {
|
|
l := containment[g]
|
|
|
|
// Compute the transitive closure of containment.
|
|
for i := 0; i < len(l); i++ {
|
|
l = append(l, containment[l[i]]...)
|
|
}
|
|
|
|
// Compute the bitmask.
|
|
regionContainment[g] = 1 << g
|
|
for _, v := range l {
|
|
regionContainment[g] |= 1 << v
|
|
}
|
|
}
|
|
b.writeSlice("regionContainment", regionContainment)
|
|
|
|
regionInclusion := make([]uint8, len(b.region.s))
|
|
bvs := make(map[uint64]index)
|
|
// Make the first bitvector positions correspond with the groups.
|
|
for r, i := range b.groups {
|
|
bv := uint64(1 << i)
|
|
for _, g := range mm[r] {
|
|
bv |= 1 << g
|
|
}
|
|
bvs[bv] = i
|
|
regionInclusion[r] = uint8(bvs[bv])
|
|
}
|
|
for r := 1; r < len(b.region.s); r++ {
|
|
if _, ok := b.groups[r]; !ok {
|
|
bv := uint64(0)
|
|
for _, g := range mm[r] {
|
|
bv |= 1 << g
|
|
}
|
|
if bv == 0 {
|
|
// Pick the world for unspecified regions.
|
|
bv = 1 << b.groups[b.region.index("001")]
|
|
}
|
|
if _, ok := bvs[bv]; !ok {
|
|
bvs[bv] = index(len(bvs))
|
|
}
|
|
regionInclusion[r] = uint8(bvs[bv])
|
|
}
|
|
}
|
|
b.writeSlice("regionInclusion", regionInclusion)
|
|
regionInclusionBits := make([]uint64, len(bvs))
|
|
for k, v := range bvs {
|
|
regionInclusionBits[v] = uint64(k)
|
|
}
|
|
// Add bit vectors for increasingly large distances until a fixed point is reached.
|
|
regionInclusionNext := []uint8{}
|
|
for i := 0; i < len(regionInclusionBits); i++ {
|
|
bits := regionInclusionBits[i]
|
|
next := bits
|
|
for i := uint(0); i < uint(len(b.groups)); i++ {
|
|
if bits&(1<<i) != 0 {
|
|
next |= regionInclusionBits[i]
|
|
}
|
|
}
|
|
if _, ok := bvs[next]; !ok {
|
|
bvs[next] = index(len(bvs))
|
|
regionInclusionBits = append(regionInclusionBits, next)
|
|
}
|
|
regionInclusionNext = append(regionInclusionNext, uint8(bvs[next]))
|
|
}
|
|
b.writeSlice("regionInclusionBits", regionInclusionBits)
|
|
b.writeSlice("regionInclusionNext", regionInclusionNext)
|
|
}
|
|
|
|
type parentRel struct {
|
|
lang uint16
|
|
script uint8
|
|
maxScript uint8
|
|
toRegion uint16
|
|
fromRegion []uint16
|
|
}
|
|
|
|
func (b *builder) writeParents() {
|
|
b.writeType(parentRel{})
|
|
|
|
parents := []parentRel{}
|
|
|
|
// Construct parent overrides.
|
|
n := 0
|
|
for _, p := range b.data.Supplemental().ParentLocales.ParentLocale {
|
|
// Skipping non-standard scripts to root is implemented using addTags.
|
|
if p.Parent == "root" {
|
|
continue
|
|
}
|
|
|
|
sub := strings.Split(p.Parent, "_")
|
|
parent := parentRel{lang: b.langIndex(sub[0])}
|
|
if len(sub) == 2 {
|
|
// TODO: check that all undefined scripts are indeed Latn in these
|
|
// cases.
|
|
parent.maxScript = uint8(b.script.index("Latn"))
|
|
parent.toRegion = uint16(b.region.index(sub[1]))
|
|
} else {
|
|
parent.script = uint8(b.script.index(sub[1]))
|
|
parent.maxScript = parent.script
|
|
parent.toRegion = uint16(b.region.index(sub[2]))
|
|
}
|
|
for _, c := range strings.Split(p.Locales, " ") {
|
|
region := b.region.index(c[strings.LastIndex(c, "_")+1:])
|
|
parent.fromRegion = append(parent.fromRegion, uint16(region))
|
|
}
|
|
parents = append(parents, parent)
|
|
n += len(parent.fromRegion)
|
|
}
|
|
b.writeSliceAddSize("parents", n*2, parents)
|
|
}
|
|
|
|
func main() {
|
|
gen.Init()
|
|
|
|
gen.Repackage("gen_common.go", "common.go", "language")
|
|
|
|
w := gen.NewCodeWriter()
|
|
defer w.WriteGoFile("tables.go", "language")
|
|
|
|
fmt.Fprintln(w, `import "golang.org/x/text/internal/tag"`)
|
|
|
|
b := newBuilder(w)
|
|
gen.WriteCLDRVersion(w)
|
|
|
|
b.parseIndices()
|
|
b.writeType(fromTo{})
|
|
b.writeLanguage()
|
|
b.writeScript()
|
|
b.writeRegion()
|
|
b.writeVariant()
|
|
// TODO: b.writeLocale()
|
|
b.computeRegionGroups()
|
|
b.writeLikelyData()
|
|
b.writeMatchData()
|
|
b.writeRegionInclusionData()
|
|
b.writeParents()
|
|
}
|