Module:languages

local require = require

local m_str_utils = require("Module:string utilities") local m_table = require("Module:table") local mw = mw local string = string local table = table local ustring = mw.ustring

local char = string.char local check_object = require("Module:utilities").check_object local concat = table.concat local decode_entities = m_str_utils.decode_entities local decode_uri = m_str_utils.decode_uri local find = string.find local floor = math.floor local gmatch = string.gmatch local gsub = string.gsub local insert = table.insert local ipairs = ipairs local list_to_set = m_table.listToSet local load_data = mw.loadData local match = string.match local next = next local pairs = pairs local remove = table.remove local remove_duplicates = m_table.removeDuplicates local select = select local setmetatable = setmetatable local shallowcopy = m_table.shallowcopy local split = m_str_utils.split local sub = string.sub local type = type local ugsub = ustring.gsub local ulen = m_str_utils.len local ulower = m_str_utils.lower local umatch = ustring.match local uupper = m_str_utils.upper

-- Loaded as needed by findBestScript. local Hans_chars local Hant_chars

local export = {}

--[=[ This module implements fetching of language-specific information and processing text in a given language.

There are two types of languages: full languages and etymology-only languages. The essential difference is that only full languages appear in L2 headings in vocabulary entries, and hence categories like Category:French nouns exist only for full languages. Etymology-only languages have either a full language or another etymology-only language as their parent (in the parent-child inheritance sense), and for etymology-only languages with another etymology-only language as their parent, a full language can always be derived by following the parent links upwards. For example, "Canadian French", code 'fr-CA', is an etymology-only language whose parent is the full language "French", code 'fr'. An example of an etymology-only language with another etymology-only parent is "Northumbrian Old English", code 'ang-nor', which has "Anglian Old English", code 'ang-ang' as its parent; this is an etymology-only language whose parent is "Old English", code "ang", which is a full language. (This is because Northumbrian Old English is considered a variety of Anglian Old English.) Sometimes the parent is the "Undetermined" language, code 'und'; this is the case, for example, for "substrate" languages such as "Pre-Greek", code 'qsb-grc', and "the BMAC substrate", code 'qsb-bma'.

It is important to distinguish language parents from language ancestors. The parent-child relationship is one of containment, i.e. if X is a child of Y, X is considered a variety of Y. On the other hand, the ancestor-descendant relationship is one of descent in time. For example, "Classical Latin", code 'la-cla', and "Late Latin", code 'la-lat', are both etymology-only languages with "Latin", code 'la', as their parents, because both of the former are varieties of Latin. However, Late Latin does *NOT* have Classical Latin as its parent because Late Latin is *not* a variety of Classical Latin; rather, it is a descendant. There is in fact a separate 'ancestors' field that is used to express the ancestor-descendant relationship, and Late Latin's ancestor is given as Classical Latin. It is also important to note that sometimes an etymology-only language is actually the conceptual ancestor of its parent language. This happens, for example, with "Old Italian" (code 'roa-oit'), which is an etymology-only variant of full language "Italian" (code 'it'), and with "Old Latin" (code 'itc-ola'), which is an etymology-only variant of Latin. In both cases, the full language has the etymology-only variant listed as an ancestor. This allows a Latin term to inherit from Old Latin using the inh template (where in this template, "inheritance" refers to ancestral inheritance, i.e. inheritance in time, rather than in the parent-child sense); likewise for Italian and Old Italian.

Full languages come in three subtypes: main namespace. There may also be reconstructed terms for the language, which are placed in the {Reconstruction} namespace and must be prefixed with * to indicate a reconstruction. Most full languages are natural (not constructed) languages, but a few constructed languages (e.g. Esperanto and Volapük,			 among others) are also allowed in the mainspace and considered regular languages. {Reconstruction} namespace. All terms in this language are reconstructed, and must be prefixed with *. Languages such as Proto-Indo-European and Proto-Germanic are in this category. constructed languages (WT:CFI). Its entries must therefore be in						 the Appendix namespace, but they are not reconstructed and therefore should not have * prefixed in links. Most constructed languages are of this subtype.
 * {regular}: This indicates a full language that is attested according to WT:CFI and therefore permitted in the
 * {reconstructed}: This language is not attested according to WT:CFI, and therefore is allowed only in the
 * {appendix-constructed}: This language is attested but does not meet the additional requirements set out for

Both full languages and etymology-only languages have a {Language} object associated with them, which is fetched using the {getByCode} function in Module:languages to convert a language code to a {Language} object. Depending on the options supplied to this function, etymology-only languages may or may not be accepted, and family codes may be accepted (returning a {Family} object as described in Module:families). There are also separate {getByCanonicalName} functions in Module:languages and Module:etymology languages to convert a language's canonical name to a {Language} object (depending on whether the canonical name refers to a full or etymology-only language).

Textual strings belonging to a given language come in several different text variants: t, lang and the like. are stripped to form the entry text (see below), as well as embedded bracketed links that are variously processed further. The display text is generated from the input text by applying language-specific transformations; for most languages, there will be no such transformations. Examples of transformations are bad-character replacements for certain languages (e.g. replacing 'l' or '1' to palochka in certain languages in Cyrillic); and for Thai and Khmer, converting space-separated words to bracketed words and resolving respelling substitutions such as [กรีน/กฺรีน], which indicate how to transliterate given words. from the display text by stripping certain sorts of diacritics on a per-language basis, and sometimes doing other transformations. The concept of entry text only really makes sense for text that does not contain embedded links, meaning that display text containing embedded links will need to have the links individually processed to get per-link entry text in order to generate the resolved display text (see below). two-part links where the first part has entry-text transformations applied, and adding appropriate language-specific fragments) and adding appropriate language and script tagging. This text can be passed directly to MediaWiki for display.  the source translit text may need to be language-specific, e.g Thai and Khmer will need the full unprocessed input  text, whereas other languages may need to work off the display text. [FIXME: It's still unclear to me how embedded  bracketed links are handled in the existing code.] In general, embedded links need to be removed (i.e. converted to  their "bare display" form by taking the right part of two-part links and removing double brackets), but when this  happens is unclear to me [FIXME]. Some languages have a chop-up-and-paste-together scheme that sends parts of the  text through the transliterate mechanism, and for others (those listed in {contiguous_substition} in  Module:languages/data) they receive the full input text, but preprocessed in certain ways. (The wisdom of this is  still unclear to me.) Unlike for all the other text variants except the transcribed text, it is always in the Latin script. "transcription" here means a close approximation to the phonetic form of the language in languages (e.g. Akkadian, Sumerian, Ancient Egyptian, maybe Tibetan) that have a wide difference between the written letters and spoken form. Unlike for all the other text variants other than the transliterated text, it is always in the Latin script. Currently, the transcribed text is always supplied manually be the user; there is no such thing as a {lua|transcribe} method on language objects. sort key is generated from the pagename or a specified sort base by lowercasing, doing language-specific transformations and then uppercasing the result. If the sort base is supplied and is generated from input text, it needs to be converted to display text, have embedded links removed (i.e. resolving them to their right side if they  are two-part links) and have entry text transformations applied. above text variants), but we can skip them for now.
 * 1) The input text is what the user supplies in wikitext, in the parameters to m, l, ux,
 * 1) The display text is the text in the form as it will be displayed to the user. This can include accent marks that
 * 1) The entry text is the text in the form used to generate a link to a Wiktionary entry. This is usually generated
 * 1) The resolved display text is the result of resolving embedded links in the display text (e.g. converting them to
 * 1) The source translit text is the text as supplied to the language-specific {transliterate} method. The form of
 * 1) The transliterated text (or transliteration) is the result of transliterating the source translit text.
 * 1) The transcribed text (or transcription) is the result of transcribing the source translit text, where
 * 1) The sort key is the text used in sort keys for determining the placing of pages in categories they belong to. The
 * 1) There are other text variants that occur in usexes (specifically, there are normalized variants of several of the

The following methods exist on {Language} objects to convert between different text variants: particular, {lua|makeEntryName} is sometimes called on display text (in some paths inside of Module:links) and sometimes called on input text (in other paths inside of Module:links, and usually from other modules). We need to make sure we don't try to convert input text to display text twice, but at the same time we need to support calling it directly on input text since so many modules do this. This means we need to add a parameter indicating whether the passed-in text is input or display text; if that former, we call {lua|makeDisplayText} ourselves.] [FIXME: This needs some rethinking. In particular, it calls {lua|processDisplayText} on its input, which won't work for Thai and Khmer, so we may need language-specific flags indicating whether to pass the input text directly to the language transliterate method. In addition, I'm not sure how embedded links are handled in the existing translit code; a lot of callers remove the links themselves before calling {lua|transliterate}, which I assume is wrong.] ]=]
 * 1) {makeDisplayText}: This converts input text to display text.
 * 2) {lua|makeEntryName}: This converts input or display text to entry text. [FIXME: This needs some rethinking. In
 * 1) {lua|transliterate}: This appears to convert input text with embedded brackets removed into a transliteration.
 * 1) {lua|makeSortKey}: This converts entry text (?) to a sort key. [FIXME: Clarify this.]

local function track(page) require("Module:debug/track")("languages/" .. page) return true end

local function conditionalRequire(modname, useRequire) return (useRequire and require or load_data)(modname) end

local function normalize_code(code, useRequire) return conditionalRequire("Module:languages/data", useRequire).aliases[code] or code end

-- Convert risky characters to HTML entities, which minimizes interference once returned (e.g. for "sms:a", "" etc.). local function escape_risky_characters(text) local encode_entities = require("Module:string/encode entities") -- Spacing characters in isolation generally need to be escaped in order to be properly processed by the MediaWiki software. if umatch(text, "^%s*$") then return encode_entities(text, text) else return encode_entities(text, "!#%&*+/:;<=>?@[\\]_{|}") end end

-- Temporarily convert various formatting characters to PUA to prevent them from being disrupted by the substitution process. local function doTempSubstitutions(text, subbedChars, keepCarets, noTrim) -- Clone so that we don't insert any extra patterns into the table in package.loaded. For some reason, using require seems to keep memory use down; probably because the table is always cloned. local patterns = shallowcopy(require("Module:languages/data/patterns")) if keepCarets then insert(patterns, "((\\+)%^)") insert(patterns, "((%^))") end -- Ensure any whitespace at the beginning and end is temp substituted, to prevent it from being accidentally trimmed. We only want to trim any final spaces added during the substitution process (e.g. by a module), which means we only do this during the first round of temp substitutions. if not noTrim then insert(patterns, "^([\128-\191\244]*(%s+))") insert(patterns, "((%s+)[\128-\191\244]*)$") end -- Pre-substitution, of "" and "", which makes pattern matching more accurate. text = gsub(text, "%f[%[]%[%[", "\1") :gsub("%f[%]]%]%]", "\2") local i, pe = #subbedChars, require("Module:string utilities").pattern_escape for _, pattern in ipairs(patterns) do -- Patterns ending in \0 stand are for things like "" or ""), so the inserted PUA are treated as breaks between terms by modules that scrape info from pages.		local term_divider		pattern = gsub(pattern, "%z$", function(divider) term_divider = divider == "\0" return "" end)		text = gsub(text, pattern, function(...) local m = {...} local m1New = m[1] for k = 2, #m do				local n = i + k - 1 subbedChars[n] = m[k] local byte2 = floor(n / 4096) % 64 + (term_divider and 128 or 136) local byte3 = floor(n / 64) % 64 + 128 local byte4 = n % 64 + 128 m1New = gsub(m1New, pe(m[k]), "\244" .. char(byte2) .. char(byte3) .. char(byte4), 1) end i = i + #m - 1 return m1New end)	end	text = gsub(text, "\1", "%[%[")		:gsub("\2", "%]%]")	return text, subbedChars end

-- Reinsert any formatting that was temporarily substituted. local function undoTempSubstitutions(text, subbedChars) local pe = require("Module:string utilities").pattern_escape for i = 1, #subbedChars do		local byte2 = floor(i / 4096) % 64 + 128 local byte3 = floor(i / 64) % 64 + 128 local byte4 = i % 64 + 128 text = gsub(text, "\244[" .. char(byte2) .. char(byte2+8) .. "]" .. char(byte3) .. char(byte4), pe(subbedChars[i])) end text = gsub(text, "\1", "%[%[") :gsub("\2", "%]%]") return text end

-- Check if the raw text is an unsupported title, and if so return that. Otherwise, remove HTML entities. We do the pre-conversion to avoid loading the unsupported title list unnecessarily. local function checkNoEntities(self, text) local textNoEnc = decode_entities(text) if textNoEnc ~= text and self:loadData("Module:links/data").unsupported_titles[text] then return text else return textNoEnc end end

-- If no script object is provided (or if it's invalid or None), get one. local function checkScript(text, self, sc) if not check_object("script", true, sc) or sc:getCode == "None" then return self:findBestScript(text) else return sc	end end

local function normalize(text, sc) text = sc:fixDiscouragedSequences(text) return sc:toFixedNFD(text) end

-- Split the text into sections, based on the presence of temporarily substituted formatting characters, then iterate over each one to apply substitutions. This avoids putting PUA characters through language-specific modules, which may be unequipped for them. local function iterateSectionSubstitutions(text, subbedChars, keepCarets, self, sc, substitution_data, function_name) local pe = require("Module:string utilities").pattern_escape local fail, cats, sections = nil, {} -- See Module:languages/data. if not find(text, "\244") or self:loadData("Module:languages/data").contiguous_substitution[self._code] then sections = {text} else sections = split(text, "\244[\128-\143][\128-\191]*", true) end for _, section in ipairs(sections) do -- Don't bother processing empty strings or whitespace (which may also not be handled well by dedicated modules). if gsub(section, "%s+", "") ~= "" then local sub, sub_fail, sub_cats = require("Module:languages/doSubstitutions")(section, self, sc, substitution_data, function_name) -- Second round of temporary substitutions, in case any formatting was added by the main substitution process. However, don't do this if the section contains formatting already (as it would have had to have been escaped to reach this stage, and therefore should be given as raw text). if sub and subbedChars then local noSub for _, pattern in ipairs(require("Module:languages/data/patterns")) do if match(section, pattern .. "%z?") then noSub = true end end if not noSub then sub, subbedChars = doTempSubstitutions(sub, subbedChars, keepCarets, true) end end if (not sub) or sub_fail then text = sub fail = sub_fail cats = sub_cats or {} break end text = sub and gsub(text, pe(section), pe(sub), 1) or text if type(sub_cats) == "table" then for _, cat in ipairs(sub_cats) do					insert(cats, cat) end end end end

-- Trim, unless there are only spacing characters, while ignoring any final formatting characters. text = text and text:gsub("^([\128-\191\244]*)%s+(%S)", "%1%2") :gsub("(%S)%s+([\128-\191\244]*)$", "%1%2")

-- Remove duplicate categories. if #cats > 1 then cats = remove_duplicates(cats) end

return text, fail, cats, subbedChars end

-- Process carets (and any escapes). Default to simple removal, if no pattern/replacement is given. local function processCarets(text, pattern, repl) local rep repeat text, rep = gsub(text, "\\\\(\\*^)", "\3%1") until rep == 0 return text:gsub("\\^", "\4") :gsub(pattern or "%^", repl or "") :gsub("\3", "\\") :gsub("\4", "^") end

-- Remove carets if they are used to capitalize parts of transliterations (unless they have been escaped). local function removeCarets(text, sc) if not sc:hasCapitalization and sc:isTransliterated and text:find("^", 1, true) then return processCarets(text) else return text end end

local Language = {}

function Language:loadData(modname) return (self._useRequire and require or mw.loadData)(modname) end

--[==[Returns the language code of the language. Example: lua for French.]==] function Language:getCode return self._code end

--[==[Returns the canonical name of the language. This is the name used to represent that language on Wiktionary, and is guaranteed to be unique to that language alone. Example: lua for French.]==] function Language:getCanonicalName local name = self._name if name == nil then name = self._rawData[1] self._name = name end return name end

--[==[ Return the display form of the language. The display form of a language, family or script is the form it takes when appearing as the  in categories such as   or , and is also the displayed text in {makeCategoryLink} links. For full and etymology-only languages, this is the same as the canonical name, but for families, it reads  (e.g. {"Indo-Iranian languages"}), and for scripts, it reads  (e.g. {"Arabic script"}). ]==] function Language:getDisplayForm local form = self._displayForm if form == nil then form = self:getCanonicalName -- Add article and " substrate" to substrates that lack them. if self:getFamilyCode == "qfa-sub" then if not (match(form, "^[Tt]he ") or match(form, "^[Aa] ")) then form = "a " .. form end if not match(form, "[Ss]ubstrate") then form = form .. " substrate" end end self._displayForm = form end return form end

--[==[Returns the value which should be used in the HTML lang= attribute for tagged text in the language.]==] function Language:getHTMLAttribute(sc, region) local code = self._code if not find(code, "-", 1, true) then return code .. "-" .. sc:getCode .. (region and "-" .. region or "") end local parent = self:getParent region = region or match(code, "%f[%u][%u-]+%f[%U]") if parent then return parent:getHTMLAttribute(sc, region) end -- TODO: ISO family codes can also be used. return "mis-" .. sc:getCode .. (region and "-" .. region or "") end

--[==[Returns a table of the "other names" that the language is known by, excluding the canonical name. The names are not guaranteed to be unique, in that sometimes more than one language is known by the same name. Example: lua for Manx. If  is given and is non-lua, only names explicitly listed in the   field are returned; otherwise, names listed under ,   and   are combined together and returned. For example, for Manx, Manx Gaelic is listed as an alias, while Northern Manx and Southern Manx are listed as varieties. It should be noted that the  field itself is deprecated, and entries listed there should eventually be moved to either   or  .]==] function Language:getOtherNames(onlyOtherNames) if #self._stack == 1 then self:loadInExtraData end return require("Module:language-like").getOtherNames(self, onlyOtherNames) end

--[==[Returns a table of the aliases that the language is known by, excluding the canonical name. Aliases are synonyms for the language in question. The names are not guaranteed to be unique, in that sometimes more than one language is known by the same name. Example: lua for German.]==] function Language:getAliases if #self._stack == 1 then self:loadInExtraData end return self._rawData.aliases or (self._extraData and self._extraData.aliases) or {} end

--[==[ Return a table of the known subvarieties of a given language, excluding subvarieties that have been given explicit etymology-only language codes. The names are not guaranteed to be unique, in that sometimes a given name refers to a subvariety of more than one language. Example: lua for Aymara. Note that the returned value can have nested tables in it, when a subvariety goes by more than one name. Example: lua for Azerbaijani. Here, for example, Afshar, Afshari, Afshar Azerbaijani and Afchar all refer to the same subvariety, whose preferred name is Afshar (the one listed first). To avoid a return value with nested tables in it, specify a non-lua value for the  parameter; in that case, the return value would be lua. ]==] function Language:getVarieties(flatten) if #self._stack == 1 then self:loadInExtraData end return require("Module:language-like").getVarieties(self, flatten) end

--[==[Returns a table of types as a lookup table (with the types as keys).

The possible types are {appendix-constructed}. Note that the types {full} and {etymology-only} also exist for families, so if you want to check specifically for a full language and you have an object that might be a family, you should use and not simply. language or a full language. Note that the types {full} and {etymology-only} also exist for families, so if you want to check specifically for an etymology-only language and you have an					object that might be a family, you should use and not simply. in the main namespace. There may also be reconstructed terms for the language, which are placed in the {Reconstruction} namespace and must be prefixed with * to indicate a reconstruction. Most full languages are natural (not constructed) languages, but a few constructed languages (e.g. Esperanto			 and Volapük, among others) are also allowed in the mainspace and considered regular languages. {Reconstruction} namespace. All terms in this language are reconstructed, and must be prefixed with *. Languages such as Proto-Indo-European and Proto-Germanic are in this category. constructed languages (WT:CFI). Its entries must therefore be in the Appendix namespace, but they are not reconstructed and therefore should not have * prefixed in links. ]==] function Language:getTypes local types = self._types if types == nil then types = {language = true} if self:getFullCode == self._code then types.full = true else types["etymology-only"] = true end for t in gmatch(self._rawData.type, "[^,]+") do			types[t] = true end self._types = types end return types end
 * {language}: This is a language, either full or etymology-only.
 * {full}: This is a "full" (not etymology-only) language, i.e. the union of {regular}, {reconstructed} and
 * {etymology-only}: This is an etymology-only (not full) language, whose parent is another etymology-only
 * {regular}: This indicates a full language that is attested according to WT:CFI and therefore permitted
 * {reconstructed}: This language is not attested according to WT:CFI, and therefore is allowed only in the
 * {appendix-constructed}: This language is attested but does not meet the additional requirements set out for

--[==[Given a list of types as strings, returns true if the language has all of them.]==] function Language:hasType(...) local args, types = {...}, self:getTypes for i = 1, #args do		if not types[args[i]] then return false end end return true end

--[==[Returns a table containing  objects (see Module:wikimedia languages), which represent languages and their codes as they are used in Wikimedia projects for interwiki linking and such. More than one object may be returned, as a single Wiktionary language may correspond to multiple Wikimedia languages. For example, Wiktionary's single code  (Serbo-Croatian) maps to four Wikimedia codes:   (Serbo-Croatian),   (Bosnian),   (Croatian) and   (Serbian). The code for the Wikimedia language is retrieved from the  property in the data modules. If that property is not present, the code of the current language is used. If none of the available codes is actually a valid Wikimedia code, an empty table is returned.]==] function Language:getWikimediaLanguages local wm_langs = self._wikimediaLanguageObjects if wm_langs == nil then local get_wm_lang = require("Module:wikimedia languages").getByCode local codes = self:getWikimediaLanguageCodes wm_langs = {} for i = 1, #codes do			wm_langs[i] = get_wm_lang(codes[i]) end self._wikimediaLanguageObjects = wm_langs end return wm_langs end

function Language:getWikimediaLanguageCodes local wm_langs = self._wikimediaLanguageCodes if wm_langs == nil then wm_langs = self._rawData.wikimedia_codes wm_langs = wm_langs and split(wm_langs, ",", true, true) or {self._code} self._wikimediaLanguageCodes = wm_langs end return wm_langs end

--[==[ Returns the name of the Wikipedia article for the language. `project` specifies the language and project to retrieve the article from, defaulting to {"enwiki"} for the English Wikipedia. Normally if specified it should be the project code for a specific-language Wikipedia e.g. "zhwiki" for the Chinese Wikipedia, but it can be any project, including non-Wikipedia ones. If the project is the English Wikipedia and the property {wikipedia_article} is present in the data module it will be used first. In all other cases, a sitelink will be generated from {:getWikidataItem} (if set). The resulting value (or lack of value) is cached so that subsequent calls are fast. If no value could be determined, and `noCategoryFallback` is {false}, {:getCategoryName} is used as fallback; otherwise, {nil} is returned. Note that if `noCategoryFallback` is {nil} or omitted, it defaults to {false} if the project is the English Wikipedia, otherwise to {true}. In other words, under normal circumstances, if the English Wikipedia article couldn't be retrieved, the return value will fall back to a link to the language's category, but this won't normally happen for any other project. ]==] function Language:getWikipediaArticle(noCategoryFallback, project) project = project or "enwiki" local cached_value if project == "enwiki" then cached_value = self._wikipedia_article if cached_value == nil then cached_value = self._rawData.wikipedia_article end else -- If the project isn't enwiki, default to no category fallback, but this can be overridden by specifying the -- value `false`. if noCategoryFallback == nil then noCategoryFallback = true end if self._non_en_wikipedia_articles == nil then self._non_en_wikipedia_articles = {} end cached_value = self._non_en_wikipedia_articles[project] end if cached_value == nil then -- not false if self:getWikidataItem and mw.wikibase then cached_value = mw.wikibase.sitelink(self:getWikidataItem, project) end if not cached_value then cached_value = false end end -- Now cache the determined value. if project == "enwiki" then self._wikipedia_article = cached_value else self._non_en_wikipedia_articles[project] = cached_value end if not cached_value and not noCategoryFallback then return gsub(self:getCategoryName, "Creole language", "Creole") end return cached_value or nil end

function Language:makeWikipediaLink return "" .. self:getCanonicalName .. "" end

--[==[Returns the Wikidata item id for the language or. This corresponds to the the second field in the data modules.]==] function Language:getWikidataItem local item = self._WikidataItem if item == nil then item = self._rawData[2] -- If the value is nil, it's cached as false. item = item ~= nil and (type(item) == "number" and "Q" .. item or item) or false self._WikidataItem = item end return item or nil end

--[==[Returns a table of  objects for all scripts that the language is written in. See Module:scripts.]==] function Language:getScripts local scripts = self._scriptObjects if scripts == nil then local codes = self:getScriptCodes if codes[1] == "All" then scripts = self:loadData("Module:scripts/data") else local get_script = require("Module:scripts").getByCode scripts = {} for i = 1, #codes do				scripts[i] = get_script(codes[i], nil, nil, self._useRequire) end end self._scriptObjects = scripts end return scripts end

--[==[Returns the table of script codes in the language's data file.]==] function Language:getScriptCodes local scripts = self._scriptCodes if scripts == nil then scripts = self._rawData[4] if scripts then local codes, n = {}, 0 for code in gmatch(scripts, "[^,]+") do				n = n + 1 -- Special handling of "Hants", which represents "Hani", "Hant" and "Hans" collectively. if code == "Hants" then codes[n] = "Hani" codes[n + 1] = "Hant" codes[n + 2] = "Hans" n = n + 2 else codes[n] = code end end scripts = codes else scripts = {"None"} end self._scriptCodes = scripts end return scripts end

--[==[Given some text, this function iterates through the scripts of a given language and tries to find the script that best matches the text. It returns a lua object representing the script. If no match is found at all, it returns the lua script object.]==] function Language:findBestScript(text, forceDetect) local useRequire = self._useRequire if not text or text == "" or text == "-" then return require("Module:scripts").getByCode("None", nil, nil, useRequire) end -- Differs from table returned by getScriptCodes, as Hants is not normalized into its constituents. local codes = self._bestScriptCodes if codes == nil then codes = self._rawData[4] codes = codes and split(codes, ",", true, true) or {"None"} self._bestScriptCodes = codes end local first_sc = codes[1] if first_sc == "All" then return require("Module:scripts").findBestScriptWithoutLang(text) end local get_script = require("Module:scripts").getByCode local codes_len = #codes if not (forceDetect or first_sc == "Hants" or codes_len > 1) then first_sc = get_script(first_sc, nil, nil, useRequire) local charset = first_sc.characters return charset and umatch(text, "[" .. charset .. "]") and first_sc or			get_script("None", nil, nil, useRequire) end -- Remove all formatting characters. text = require("Module:utilities").get_plaintext(text) -- Remove all spaces and any ASCII punctuation. Some non-ASCII punctuation is script-specific, so can't be removed. text = ugsub(text, "[%s!\"#%%&'*,%-./:;?@[\\%]_{}]+", "")	if #text == 0 then		return get_script("None", nil, nil, useRequire)	end	-- Try to match every script against the text,	-- and return the one with the most matching characters.	local bestcount, bestscript, length = 0	for i = 1, codes_len do		local sc = codes[i]		-- Special case for "Hants", which is a special code that represents whichever of "Hant" or "Hans" best matches, or "Hani" if they match equally. This avoids having to list all three. In addition, "Hants" will be treated as the best match if there is at least one matching character, under the assumption that a Han script is desirable in terms that contain a mix of Han and other scripts (not counting those which use Jpan or Kore).		if sc == "Hants" then			local Hani = get_script("Hani", nil, nil, useRequire)			if not Hant_chars then				Hant_chars = self:loadData("Module:zh/data/ts") Hans_chars = self:loadData("Module:zh/data/st") end local t, s, found = 0, 0 -- This is faster than using mw.ustring.gmatch directly. for ch in gmatch(ugsub(text, "[" .. Hani.characters .. "]", "\255%0"), "\255(.[\128-\191]*)") do				found = true if Hant_chars[ch] then t = t + 1 if Hans_chars[ch] then s = s + 1 end elseif Hans_chars[ch] then s = s + 1 else t, s = t + 1, s + 1 end end if found then if t == s then return Hani end return get_script(t > s and "Hant" or "Hans", nil, nil, useRequire) end else sc = get_script(sc, nil, nil, useRequire) if not length then length = ulen(text) end -- Count characters by removing everything in the script's charset and comparing to the original length. local charset = sc.characters local count = charset and length - ulen(ugsub(text, "[" .. charset .. "]+", "")) or 0 if count >= length then return sc			elseif count > bestcount then bestcount = count bestscript = sc			end end end -- Return best matching script, or otherwise None. return bestscript or get_script("None", nil, nil, useRequire) end

--[==[Returns a  object for the language family that the language belongs to. See Module:families.]==] function Language:getFamily local family = self._familyObject if family == nil then family = self:getFamilyCode -- If the value is nil, it's cached as false. family = family and require("Module:families").getByCode(family, self._useRequire) or false self._familyObject = family end return family or nil end

--[==[Returns the family code in the language's data file.]==] function Language:getFamilyCode local family = self._familyCode if family == nil then -- If the value is nil, it's cached as false. family = self._rawData[3] or false self._familyCode = family end return family or nil end

function Language:getFamilyName local family = self._familyName if family == nil then family = self:getFamily -- If the value is nil, it's cached as false. family = family and family:getCanonicalName or false self._familyName = family end return family or nil end

--[==[Check whether the language belongs to `family` (which can be a family code or object). A list of objects can be given in place of `family`; in that case, return true if the language belongs to any of the specified families. Note that some languages (in particular, certain creoles) can have multiple immediate ancestors potentially belonging to different families; in that case, return true if the language belongs to any of the specified families.]==] function Language:inFamily(...) --check_object("family", nil, ...) for _, family in ipairs{...} do		if type(family) == "table" then family = family:getCode end local self_family_code = self:getFamilyCode if not self_family_code then return false elseif self_family_code == family then return true end local self_family = self:getFamily if self_family:inFamily(family) then return true -- If the family isn't a real family (e.g. creoles) check any ancestors. elseif self_family:getFamilyCode == "qfa-not" then local ancestors = self:getAncestors for _, ancestor in ipairs(ancestors) do				if ancestor:inFamily(family) then return true end end end end return false end

function Language:getParent local parent = self._parentObject if parent == nil then parent = self:getParentCode -- If the value is nil, it's cached as false. parent = parent and export.getByCode(parent, nil, true, true, self._useRequire) or false self._parentObject = parent end return parent or nil end

function Language:getParentCode local parent = self._parentCode if parent == nil then -- If the value is nil, it's cached as false. parent = self._rawData[5] or false self._parentCode = parent end return parent or nil end

function Language:getParentName local parent = self._parentName if parent == nil then parent = self:getParent -- If the value is nil, it's cached as false. parent = parent and parent:getCanonicalName or false self._parentName = parent end return parent or nil end

function Language:getParentChain local chain = self._parentChain if chain == nil then chain = {} local parent, n = self:getParent, 0 while parent do			n = n + 1 chain[n] = parent parent = parent:getParent end self._parentChain = chain end return chain end

function Language:hasParent(...) --check_object("language", nil, ...) for _, otherlang in ipairs{...} do		for _, parent in ipairs(self:getParentChain) do			if type(otherlang) == "string" then if otherlang == parent:getCode then return true end else if otherlang:getCode == parent:getCode then return true end end end end return false end

--[==[ If the language is etymology-only, this iterates through parents until a full language or family is found, and the corresponding object is returned. If the language is a full language, then it simply returns itself. ]==] function Language:getFull local full = self._fullObject if full == nil then full = self:getFullCode full = full == self._code and self or			export.getByCode(full, nil, nil, nil, self._useRequire) self._fullObject = full end return full end

--[==[ If the language is an etymology-only language, this iterates through parents until a full language or family is found, and the corresponding code is returned. If the language is a full language, then it simply returns the language code. ]==] function Language:getFullCode return self._fullCode or self._code end

--[==[ If the language is an etymology-only language, this iterates through parents until a full language or family is found, and the corresponding canonical name is returned. If the language is a full language, then it simply returns the canonical name of the language. ]==] function Language:getFullName local full = self._fullName if full == nil then full = self:getFull:getCanonicalName self._fullName = full end return full end

--[==[Returns a table of  objects for all languages that this language is directly descended from. Generally this is only a single language, but creoles, pidgins and mixed languages can have multiple ancestors.]==] function Language:getAncestors if not self._ancestorObjects then self._ancestorObjects = {} local ancestors = shallowcopy(self:getAncestorCodes) if #ancestors > 0 then for _, ancestor in ipairs(ancestors) do				insert(self._ancestorObjects, export.getByCode(ancestor, nil, true, nil, self._useRequire)) end else local fam = self:getFamily local protoLang = fam and fam:getProtoLanguage or nil -- For the cases where the current language is the proto-language -- of its family, or an etymology-only language that is ancestral to that -- proto-language, we need to step up a level higher right from the -- start. if protoLang and (				protoLang:getCode == self._code or				(self:hasType("etymology-only") and protoLang:hasAncestor(self))			) then fam = fam:getFamily protoLang = fam and fam:getProtoLanguage or nil end while not protoLang and not (not fam or fam:getCode == "qfa-not") do				fam = fam:getFamily protoLang = fam and fam:getProtoLanguage or nil end insert(self._ancestorObjects, protoLang) end end return self._ancestorObjects end

do -- Avoid a language being its own ancestor via class inheritance. We only need to check for this if the language has inherited an ancestor table from its parent, because we never want to drop ancestors that have been explicitly set in the data. -- Recursively iterate over ancestors until we either find self or run out. If self is found, return true. local function check_ancestor(self, lang) local codes = lang:getAncestorCodes if not codes then return nil end for i = 1, #codes do			local code = codes[i] if code == self._code then return true end local anc = export.getByCode(code, nil, true, nil, self._useRequire) if check_ancestor(self, anc) then return true end end end

--[==[Returns a table of  codes for all languages that this language is directly descended from. Generally this is only a single language, but creoles, pidgins and mixed languages can have multiple ancestors.]==] function Language:getAncestorCodes if self._ancestorCodes then return self._ancestorCodes end local codes = self._rawData.ancestors if not codes then codes = {} self._ancestorCodes = codes return codes end codes = split(codes, ",", true, true) self._ancestorCodes = codes if (			#codes == 0 or			#self._stack == 1 or			self._stack[#self._stack].ancestors		) then return codes end local i, code = 1 while i <= #codes do			code = codes[i] if check_ancestor(self, self) then remove(codes, i)			else i = i + 1 end end return codes end end

--[==[Given a list of language objects or codes, returns true if at least one of them is an ancestor. This includes any etymology-only children of that ancestor. If the language's ancestor(s) are etymology-only languages, it will also return true for those language parent(s) (e.g. if Vulgar Latin is the ancestor, it will also return true for its parent, Latin). However, a parent is excluded from this if the ancestor is also ancestral to that parent (e.g. if Classical Persian is the ancestor, Persian would return false, because Classical Persian is also ancestral to Persian).]==] function Language:hasAncestor(...) --check_object("language", nil, ...)

local function iterateOverAncestorTree(node, func, parent_check) local ancestors = node:getAncestors local ancestorsParents = {} for _, ancestor in ipairs(ancestors) do			local ret = func(ancestor) or iterateOverAncestorTree(ancestor, func, parent_check) if ret then return ret end end -- Check the parents of any ancestors. We don't do this if checking the parents of the other language, so that we exclude any etymology-only children of those parents that are not directly related (e.g. if the ancestor is Vulgar Latin and we are checking New Latin, we want it to return false because they are on different ancestral branches. As such, if we're already checking the parent of New Latin (Latin) we don't want to compare it to the parent of the ancestor (Latin), as this would be a false positive; it should be one or the other). if not parent_check then return nil end for _, ancestor in ipairs(ancestors) do			local ancestorParents = ancestor:getParentChain for _, ancestorParent in ipairs(ancestorParents) do				if ancestorParent:getCode == self._code or ancestorParent:hasAncestor(ancestor) then break else insert(ancestorsParents, ancestorParent) end end end for _, ancestorParent in ipairs(ancestorsParents) do			local ret = func(ancestorParent) if ret then return ret end end end

local function do_iteration(otherlang, parent_check) -- otherlang can't be self if (type(otherlang) == "string" and otherlang or otherlang:getCode) == self._code then return false end repeat if iterateOverAncestorTree(				self,				function(ancestor)					return ancestor:getCode == (type(otherlang) == "string" and otherlang or otherlang:getCode)				end,				parent_check			) then return true elseif type(otherlang) == "string" then otherlang = export.getByCode(otherlang, nil, true, nil, self._useRequire) end otherlang = otherlang:getParent parent_check = false until not otherlang end

local parent_check = true for _, otherlang in ipairs{...} do		local ret = do_iteration(otherlang, parent_check) if ret then return true end end return false end

function Language:getAncestorChain if not self._ancestorChain then self._ancestorChain = {} local step = self while true do			local ancestors = step:getAncestors step = #ancestors == 1 and ancestors[1] or nil if not step then break end insert(self._ancestorChain, 1, step) end end return self._ancestorChain end

local function fetch_descendants(self, format) local languages = require("Module:languages/code to canonical name") local etymology_languages = require("Module:etymology languages/code to canonical name") local families = require("Module:families/code to canonical name") local descendants = {} local family = self:getFamily -- Iterate over all three datasets. for _, data in ipairs{languages, etymology_languages, families} do		for code in pairs(data) do			local lang = export.getByCode(code, nil, true, true, self._useRequire) -- Test for a descendant. Earlier tests weed out most candidates, while the more intensive tests are only used sparingly. if (				code ~= self._code and -- Not self.				lang:inFamily(family) and -- In the same family.				( family:getProtoLanguageCode == self._code or -- Self is the protolanguage. self:hasDescendant(lang) or -- Full hasDescendant check. (lang:getFullCode == self._code and not self:hasAncestor(lang)) -- Etymology-only child which isn't an ancestor. )			) then if format == "object" then insert(descendants, lang) elseif format == "code" then insert(descendants, code) elseif format == "name" then insert(descendants, lang:getCanonicalName) end end end end return descendants end

function Language:getDescendants if not self._descendantObjects then self._descendantObjects = fetch_descendants(self, "object") end return self._descendantObjects end

function Language:getDescendantCodes if not self._descendantCodes then self._descendantCodes = fetch_descendants(self, "code") end return self._descendantCodes end

function Language:getDescendantNames if not self._descendantNames then self._descendantNames = fetch_descendants(self, "name") end return self._descendantNames end

function Language:hasDescendant(...) for _, lang in ipairs{...} do		if type(lang) == "string" then lang = export.getByCode(lang, nil, true, nil, self._useRequire) end if lang:hasAncestor(self) then return true end end return false end

local function fetch_children(self, format) local m_etym_data = require("Module:etymology languages/data") local self_code = self._code local children = {} for code, data in pairs(m_etym_data) do		local _data = data repeat local parent = _data[5] if parent == self_code then if format == "object" then insert(children, export.getByCode(code, nil, true, nil, self._useRequire)) elseif format == "code" then insert(children, code) elseif format == "name" then insert(children, data[1]) end break end _data = m_etym_data[parent] until not _data end return children end

function Language:getChildren if not self._childObjects then self._childObjects = fetch_children(self, "object") end return self._childObjects end

function Language:getChildrenCodes if not self._childCodes then self._childCodes = fetch_children(self, "code") end return self._childCodes end

function Language:getChildrenNames if not self._childNames then self._childNames = fetch_children(self, "name") end return self._childNames end

function Language:hasChild(...) local lang = ... if not lang then return false elseif type(lang) == "string" then lang = export.getByCode(lang, nil, true, nil, self._useRequire) end if lang:hasParent(self) then return true end return self:hasChild(select(2, ...)) end

--[==[Returns the name of the main category of that language. Example: lua for French, whose category is at Category:French language. Unless optional argument  is given, the language name at the beginning of the returned value will be capitalized. This capitalization is correct for category names, but not if the language name is lowercase and the returned value of this function is used in the middle of a sentence.]==] function Language:getCategoryName(nocap) if not self._categoryName then local name = self:getCanonicalName -- Only add " language" if a full language. if #self._stack == 1 then -- If the name already has "language" in it, don't add it. if not match(name, "[Ll]anguage$") then name = name .. " language" end end self._categoryName = name end if nocap then return self._categoryName else return mw.getContentLanguage:ucfirst(self._categoryName) end end

--[==[Creates a link to the category; the link text is the canonical name.]==] function Language:makeCategoryLink return "" .. self:getDisplayForm .. "" end

function Language:getStandardCharacters(sc) if type(self._rawData.standardChars) ~= "table" then return self._rawData.standardChars else if sc and type(sc) ~= "string" then check_object("script", nil, sc) sc = sc:getCode end if (not sc) or sc == "None" then local scripts = {} for _, script in pairs(self._rawData.standardChars) do				insert(scripts, script) end return concat(scripts) end if self._rawData.standardChars[sc] then return self._rawData.standardChars[sc] .. (self._rawData.standardChars[1] or "") end end end

--[==[Make the entry name (i.e. the correct page name).]==] function Language:makeEntryName(text, sc) if (not text) or text == "" then return text, nil, {} end -- Set `unsupported` as true if certain conditions are met. local unsupported -- Check if there's an unsupported character. \239\191\189 is the replacement character U+FFFD, which can't be typed directly here due to an abuse filter. Unix-style dot-slash notation is also unsupported, as it is used for relative paths in links, as are 3 or more consecutive tildes. -- Note: match is faster with magic characters/charsets; find is faster with plaintext. if (		match(text, "[#<>%[%]_{|}]") or		find(text, "\239\191\189") or		match(text, "%f[^%z/]%.%.?%f[%z/]") or		find(text, "")	) then unsupported = true -- If it looks like an interwiki link. elseif find(text, ":") then local prefix = gsub(text, "^:*(.-):.*", ulower) if (			self:loadData("Module:data/namespaces")[prefix] or			self:loadData("Module:data/interwikis")[prefix]		) then unsupported = true end end

-- Check if the text is a listed unsupported title. local unsupportedTitles = self:loadData("Module:links/data").unsupported_titles if unsupportedTitles[text] then return "Unsupported titles/" .. unsupportedTitles[text], nil, {} end

sc = checkScript(text, self, sc)

local fail, cats text = normalize(text, sc) text, fail, cats = iterateSectionSubstitutions(text, nil, nil, self, sc, self._rawData.entry_name, "makeEntryName")

text = umatch(text, "^[¿¡]?(.-[^%s%p].-)%s*[؟?!;՛՜ ՞ ՟？！︖︕।॥။၊་།]?$") or text

-- Escape unsupported characters so they can be used in titles. ` is used as a delimiter for this, so a raw use of it in an unsupported title is also escaped here to prevent interference; this is only done with unsupported titles, though, so inclusion won't in itself mean a title is treated as unsupported (which is why it's excluded from the earlier test). if unsupported then local unsupported_characters = self:loadData("Module:links/data").unsupported_characters text = text:gsub("[#<>%[%]_`{|}\239]\191?\189?", unsupported_characters) :gsub("%f[^%z/]%.%.?%f[%z/]", function(m)				return gsub(m, "%.", "`period`")			end) :gsub("+", function(m)				return gsub(m, "~", "`tilde`")			end) text = "Unsupported titles/" .. text end

return text, fail, cats end

--[==[Generates alternative forms using a specified method, and returns them as a table. If no method is specified, returns a table containing only the input term.]==] function Language:generateForms(text, sc) if self._rawData.generate_forms then sc = checkScript(text, self, sc) return require("Module:" .. self._rawData.generate_forms).generateForms(text, self._code, sc:getCode) else return {text} end end

--[==[Creates a sort key for the given entry name, following the rules appropriate for the language. This removes diacritical marks from the entry name if they are not considered significant for sorting, and may perform some other changes. Any initial hyphen is also removed, and anything parentheses is removed as well. The  setting for each language in the data modules defines the replacements made by this function, or it gives the name of the module that takes the entry name and returns a sortkey.]==] function Language:makeSortKey(text, sc) if (not text) or text == "" then return text, nil, {} end if match(text, "<[^<>]+>") then track("track HTML tag") end -- Remove directional characters, soft hyphens, strip markers and HTML tags. text = ugsub(text, "[\194\173\226\128\170-\226\128\174\226\129\166-\226\129\169]", "") text = gsub(mw.text.unstrip(text), "<[^<>]+>", "")

text = decode_uri(text, "PATH") text = checkNoEntities(self, text)

-- Remove initial hyphens and * unless the term only consists of spacing + punctuation characters. text = ugsub(text, "^([􀀀-􏿽]*)[-־ـ᠊*]+([􀀀-􏿽]*)(.*[^%s%p].*)", "%1%2%3")

sc = checkScript(text, self, sc)

text = normalize(text, sc) text = removeCarets(text, sc)

-- For languages with dotted dotless i, ensure that "İ" is sorted as "i", and "I" is sorted as "ı". if self:hasDottedDotlessI then text = gsub(text, "I\204\135", "i") -- decomposed "İ" :gsub("I", "ı") text = sc:toFixedNFD(text) end -- Convert to lowercase, make the sortkey, then convert to uppercase. Where the language has dotted dotless i, it is usually not necessary to convert "i" to "İ" and "ı" to "I" first, because "I" will always be interpreted as conventional "I" (not dotless "İ") by any sorting algorithms, which will have been taken into account by the sortkey substitutions themselves. However, if no sortkey substitutions have been specified, then conversion is necessary so as to prevent "i" and "ı" both being sorted as "I". -- An exception is made for scripts that (sometimes) sort by scraping page content, as that means they are sensitive to changes in capitalization (as it changes the target page). local fail, cats if not sc:sortByScraping then text = ulower(text) end

text, fail, cats = iterateSectionSubstitutions(text, nil, nil, self, sc, self._rawData.sort_key, "makeSortKey")

if not sc:sortByScraping then if self:hasDottedDotlessI and not self._rawData.sort_key then text = gsub(gsub(text, "ı", "I"), "i", "İ") text = sc:toFixedNFC(text) end text = uupper(text) end

-- Remove parentheses, as long as they are either preceded or followed by something. text = gsub(text, "(.)[]+", "%1") :gsub("[]+(.)", "%1")

text = escape_risky_characters(text) return text, fail, cats end

--[==[Create the form used as as a basis for display text and transliteration.]==] local function processDisplayText(text, self, sc, keepCarets, keepPrefixes) local subbedChars = {} text, subbedChars = doTempSubstitutions(text, subbedChars, keepCarets)

text = decode_uri(text, "PATH") text = checkNoEntities(self, text)

sc = checkScript(text, self, sc) local fail, cats text = normalize(text, sc) text, fail, cats, subbedChars = iterateSectionSubstitutions(text, subbedChars, keepCarets, self, sc, self._rawData.display_text, "makeDisplayText")

text = removeCarets(text, sc)

-- Remove any interwiki link prefixes (unless they have been escaped or this has been disabled). if find(text, ":") and not keepPrefixes then local rep repeat text, rep = gsub(text, "\\\\(\\*:)", "\3%1") until rep == 0 text = gsub(text, "\\:", "\4") while true do			local prefix = gsub(text, "^(.-):.+", function(m1)				return gsub(m1, "\244[\128-\191]*", "")			end) if not prefix or prefix == text then break end local lower_prefix = ulower(prefix) if not (self:loadData("Module:data/interwikis")[lower_prefix] or prefix == "") then break end text = gsub(text, "^(.-):(.*)", function(m1, m2)				local ret = {}				for subbedChar in gmatch(m1, "\244[\128-\191]*") do					insert(ret, subbedChar)				end				return concat(ret) .. m2			end) end text = gsub(text, "\3", "\\") :gsub("\4", ":") end

return text, fail, cats, subbedChars end

--[==[Make the display text (i.e. what is displayed on the page).]==] function Language:makeDisplayText(text, sc, keepPrefixes) if (not text) or text == "" then return text, nil, {} end local fail, cats, subbedChars text, fail, cats, subbedChars = processDisplayText(text, self, sc, nil, keepPrefixes)

text = escape_risky_characters(text) return undoTempSubstitutions(text, subbedChars), fail, cats end

--[==[Transliterates the text from the given script into the Latin script (see Transliteration and romanization). The language must have the  property for this to work; if it is not present, lua is returned. Returns three values: The  parameter is handled by the transliteration module, and how it is handled is specific to that module. Some transliteration modules may tolerate lua as the script, others require it to be one of the possible scripts that the module can transliterate, and will show an error if it's not one of them. For this reason, the  parameter should always be provided when writing non-language-specific code. The  parameter is used to override the default module that is used to provide the transliteration. This is useful in cases where you need to demonstrate a particular module in use, but there is no default module yet, or you want to demonstrate an alternative version of a transliteration module before making it official. It should not be used in real modules or templates, only for testing. All uses of this parameter are tracked by Tracking/module_override. Known bugs: function Language:transliterate(text, sc, module_override) -- If there is no text, or the language doesn't have transliteration data and there's no override, return nil. if not (self._rawData.translit or module_override) then return nil, false, {} elseif (not text) or text == "" or text == "-" then return text, false, {} end -- If the script is not transliteratable (and no override is given), return nil. sc = checkScript(text, self, sc) if not (sc:isTransliterated or module_override) then return nil, true, {} end
 * 1) The transliteration.
 * 2) A boolean which indicates whether the transliteration failed for an unexpected reason. If lua, then the transliteration either succeeded, or the module is returning nothing in a controlled way (e.g. the input was lua). Generally, this means that no maintenance action is required. If lua, then the transliteration is lua because either the input or output was defective in some way (e.g. Module:ar-translit will not transliterate non-vocalised inputs, and this module will fail partially-completed transliterations in all languages). Note that this value can be manually set by the transliteration module, so make sure to cross-check to ensure it is accurate.
 * 3) A table of categories selected by the transliteration module, which should be in the format expected by lua in Module:utilities.
 * This function assumes {tr(s1) .. tr(s2) == tr(s1 .. s2)}. When this assertion fails, wikitext markups like ''' can cause wrong transliterations.
 * HTML entities like, often used to escape wikitext markups, do not work.]==]

-- Remove any strip markers. text = mw.text.unstrip(text)

-- Get the display text with the keepCarets flag set. local fail, cats, subbedChars text, fail, cats, subbedChars = processDisplayText(text, self, sc, true)

-- Transliterate (using the module override if applicable). text, fail, cats, subbedChars = iterateSectionSubstitutions(text, subbedChars, true, self, sc, module_override or self._rawData.translit, "tr") if not text then return nil, true, cats end -- Incomplete transliterations return nil. local charset = sc.characters if charset and umatch(text, "[" .. charset .. "]") then -- Remove any characters in Latin, which includes Latin characters also included in other scripts (as these are false positives), as well as any PUA substitutions. Anything remaining should only be script code "None" (e.g. numerals). local check_text = ugsub(text, "[" .. require("Module:scripts").getByCode("Latn").characters .. "􀀀-􏿽]+", "") -- Set none_is_last_resort_only flag, so that any non-None chars will cause a script other than "None" to be returned. if require("Module:scripts").findBestScriptWithoutLang(check_text, true):getCode ~= "None" then return nil, true, cats end end

text = escape_risky_characters(text) text = undoTempSubstitutions(text, subbedChars)

-- If the script does not use capitalization, then capitalize any letters of the transliteration which are immediately preceded by a caret (and remove the caret). if text and not sc:hasCapitalization and text:find("^", 1, true) then text = processCarets(text, "%^([\128-\191\244]*%*?)([^\128-\191\244][\128-\191]*)", function(m1, m2)			return m1 .. uupper(m2)		end) end

-- Track module overrides. if module_override ~= nil then track("module_override") end

fail = text == nil and (not not fail) or false

return text, fail, cats end

do local function handle_language_spec(self, spec, sc) local ret = self["_" .. spec] if ret == nil then ret = self._rawData[spec] if type(ret) == "string" then ret = list_to_set(split(ret, ",", true, true)) end self["_" .. spec] = ret end if type(ret) == "table" then ret = ret[sc:getCode] end return not not ret end function Language:overrideManualTranslit(sc) return handle_language_spec(self, "override_translit", sc) end function Language:link_tr(sc) return handle_language_spec(self, "link_tr", sc) end end

--[==[Returns lua if the language has a transliteration module, or lua if it doesn't.]==] function Language:hasTranslit return not not self._rawData.translit end

--[==[Returns lua if the language uses the letters I/ı and İ/i, or lua if it doesn't.]==] function Language:hasDottedDotlessI return not not self._rawData.dotted_dotless_i end

function Language:toJSON(returnTable) local entryNamePatterns = nil local entryNameRemoveDiacritics = nil

if self._rawData.entry_name then entryNameRemoveDiacritics = self._rawData.entry_name.remove_diacritics if self._rawData.entry_name.from then entryNamePatterns = {} for i, from in ipairs(self._rawData.entry_name.from) do				insert(entryNamePatterns, {from = from, to = self._rawData.entry_name.to[i] or ""}) end end end -- mainCode should only end up non-nil if dontCanonicalizeAliases is passed to make_object. local ret = m_table.deepcopy{ ancestors = self:getAncestorCodes, canonicalName = self:getCanonicalName, categoryName = self:getCategoryName("nocap"), code = self._code, mainCode = self._main_code, entryNamePatterns = entryNamePatterns, entryNameRemoveDiacritics = entryNameRemoveDiacritics, family = self:getFamilyCode, otherNames = self:getOtherNames(true), aliases = self:getAliases, varieties = self:getVarieties, scripts = self:getScriptCodes, parent = self._parentCode or nil, full = self._fullCode or nil, type = m_table.keysToList(self:getTypes), wikimediaLanguages = self:getWikimediaLanguageCodes, wikidataItem = self:getWikidataItem, }

if returnTable then return ret else return require("Module:JSON").toJSON(ret) end end

--[==[ This function is not for use in entries or other content pages. Returns a blob of data about the language. The format of this blob is undocumented, and perhaps unstable; it's intended for things like the module's own unit-tests, which are "close friends" with the module and will be kept up-to-date as the format changes. -- Do NOT use these methods! -- All uses should be pre-approved on the talk page! ]==] function Language:getRawData local rawData = {} for _, element in ipairs(self._stack) do		for k, v in pairs(element) do			rawData[k] = v		end end return rawData end

--[==[ This function is not for use in entries or other content pages. Returns a blob of data about the language that contains the "extra data". Much like with getRawData, the format of this blob is undocumented, and perhaps unstable; it's intended for things like the module's own unit-tests, which are "close friends" with the module and will be kept up-to-date as the format changes.]==] function Language:getRawExtraData if #self._stack == 1 then self:loadInExtraData end return self._extraData end

local function getRawExtraLanguageData(self, code) local modulename = export.getExtraDataModuleName(code) return modulename and self:loadData("Module:" .. modulename)[code] or nil end

function Language:loadInExtraData if not self._extraData then -- load extra data from module and assign to _extraData field -- use empty table as a fallback if extra data is nil self._extraData = getRawExtraLanguageData(self, self._code) or {} end end

function export.getDataModuleName(code) local letter = match(code, "^(%l)[%l-]+$") if not letter then return nil elseif find(code, "-", 1, true) then return "languages/data/exceptional" end local code_len = #code return code_len == 2 and "languages/data/2" or code_len == 3 and "languages/data/3/" .. letter or nil end

function export.getExtraDataModuleName(code) local dataModule = export.getDataModuleName(code) return dataModule and dataModule .. "/extra" or nil end

do local key_types = { [2] = "unique", aliases = "unique", otherNames = "unique", type = "append", varieties = "unique" }	function export.makeObject(code, data, useRequire, dontCanonicalizeAliases) if not data then return nil end

-- Convert any aliases. local input_code = code code = normalize_code(code, useRequire) input_code = dontCanonicalizeAliases and input_code or code

if find(data.type, "family") and not data[5] then return require("Module:families").makeObject(code, data, useRequire) end local parent if data[5] then parent = export.getByCode(data[5], nil, true, true, useRequire) else parent = Language end parent.__index = parent

local lang = { _code = input_code, _useRequire = useRequire or nil }		-- This can only happen if dontCanonicalizeAliases is passed to make_object. if code ~= input_code then lang._main_code = code end

-- Full language. if not parent._stack then -- Create stack, accessed with rawData metamethod. local stack = parent._rawData and {parent._rawData, data} or {data} lang._stack = stack lang._rawData = setmetatable({}, {				__index = function(t, k)					local key_type = key_types[k]					-- Data that isn't inherited from the parent.					if key_type == "unique" then						return stack[#stack][k]					-- Data that is appended by each generation.					elseif key_type == "append" then						local parts = {}						for i = 1, #stack do							insert(parts, stack[i][k])						end						if type(parts[1]) == "string" then							return concat(parts, ","), true						end					-- Otherwise, iterate down the stack, looking for a match.					else						local i = #stack						while not stack[i][k] and i > 1 do							i = i - 1						end						return stack[i][k]					end				end,				-- Retain immutability (as writing to rawData will break functionality).				__newindex = function					error("not allowed to edit rawData")				end			}) -- Full code is the parent code. lang._fullCode = parent._code or code -- Etymology-only. else -- Copy over rawData and stack to the new object, and add new layer to stack. lang._rawData = parent._rawData lang._stack = parent._stack insert(lang._stack, data) -- Copy full code. lang._fullCode = parent._fullCode end

return setmetatable(lang, parent) end end

--[==[Finds the language whose code matches the one provided. If it exists, it returns a  object representing the language. Otherwise, it returns lua, unless  is given, in which case an error is generated. If  is lua, a generic error message mentioning the bad code is generated; otherwise   should be a string or number specifying the parameter that the code came from, and this parameter will be mentioned in the error message along with the bad code. If  is specified, etymology-only language codes are allowed and looked up along with normal language codes. If  is specified, language family codes are allowed and looked up along with normal language codes.]==] function export.getByCode(code, paramForError, allowEtymLang, allowFamily, useRequire) if type(code) ~= "string" then local typ if not code then typ = "nil" elseif check_object("language", true, code) then typ = "a language object" elseif check_object("family", true, code) then typ = "a family object" else typ = "a " .. type(code) end error("The function getByCode expects a string as its first argument, but received " .. typ .. ".") end local modulename = export.getDataModuleName(code) local m_data = conditionalRequire("Module:languages/data", useRequire) if m_data.aliases[code] or m_data.track[code] then track(code) end local norm_code = normalize_code(code, useRequire) local data = modulename and conditionalRequire("Module:" .. modulename, useRequire)[norm_code] or		(allowEtymLang and require("Module:etymology languages/track-bad-etym-code")(norm_code) and conditionalRequire("Module:etymology languages/data", useRequire)[norm_code]) or		(allowFamily and conditionalRequire("Module:families/data", useRequire)[norm_code]) or		(allowEtymLang and allowFamily and require("Module:families/track-bad-etym-code")(norm_code) and conditionalRequire("Module:families/data/etymology", useRequire)[norm_code]) local retval = code and data and export.makeObject(code, data, useRequire)

if not retval and paramForError then require("Module:languages/errorGetBy").code(code, paramForError, allowEtymLang, allowFamily) end

return retval end

--[==[Finds the language whose canonical name (the name used to represent that language on Wiktionary) or other name matches the one provided. If it exists, it returns a  object representing the language. Otherwise, it returns lua, unless  is given, in which case an error is generated. If  is specified, etymology-only language codes are allowed and looked up along with normal language codes. If  is specified, language family codes are allowed and looked up along with normal language codes. The canonical name of languages should always be unique (it is an error for two languages on Wiktionary to share the same canonical name), so this is guaranteed to give at most one result. This function is powered by Module:languages/canonical names, which contains a pre-generated mapping of full-language canonical names to codes. It is generated by going through the Category:Language data modules for full languages. When  is specified for the above function, Module:etymology languages/canonical names may also be used, and when   is specified for the above function, Module:families/canonical names may also be used.]==] function export.getByCanonicalName(name, errorIfInvalid, allowEtymLang, allowFamily, useRequire) local byName = conditionalRequire("Module:languages/canonical names", useRequire) local code = byName and byName[name]

if not code and allowEtymLang then byName = conditionalRequire("Module:etymology languages/canonical names", useRequire) code = byName and byName[name] or			byName[gsub(name, " [Ss]ubstrate$", "")] or			byName[gsub(name, "^a ", "")] or			byName[gsub(name, "^a ", "") :gsub(" [Ss]ubstrate$", "")] or -- For etymology families like "ira-pro". -- FIXME: This is not ideal, as it allows " languages" to be appended to any etymology-only language, too. byName[match(name, "^(.*) languages$")] end

if not code and allowFamily then byName = conditionalRequire("Module:families/canonical names", useRequire) code = byName and byName[name] or			byName[match(name, "^(.*) languages$")] end

local retval = code and export.getByCode(code, errorIfInvalid, allowEtymLang, allowFamily, useRequire)

if not retval and errorIfInvalid then require("Module:languages/errorGetBy").canonicalName(name, allowEtymLang, allowFamily) end

return retval end

--[==[Used by Module:languages/data/2 (et al.) to add default types to the entities returned.]==] function export.addDefaultTypes(data, regular, ...) local n = arg.n	local types = n > 0 and concat(arg, ",") or "" for _, entity in next, data do -- "regular" encompasses everything that doesn't have another type already assigned. if regular then entity.type = entity.type or "regular" end if n > 0 then entity.type = types .. (entity.type and ("," .. entity.type) or "") end end return data end

--[==[Used by Module:languages/data/2 (et al.) and Module:etymology languages/data to finalize language-related data into the format that is actually returned.]==] function export.finalizeLanguageData(data) -- 4 is scripts. local fields = {4, "ancestors", "link_tr", "override_translit", "type", "wikimedia_codes"} local fields_len = #fields for _, entity in next, data do		for i = 1, fields_len do			local key = fields[i] local field = entity[key] if field and type(field) == "string" then entity[key] = gsub(field, "%s+", "") end end end return data end

--[==[Used by Module:etymology languages/data and Module:families/data/etymology to finalize etymology-related data into the format that is actually returned.]==] function export.finalizeEtymologyData(data) local aliases = {} for _, entity in next, data do -- Move parent to 5 and family to 3. entity[5] = entity[3] entity[3] = entity.family entity.family = nil end for code, alias in next, aliases do		data[code] = alias end return data end

--[==[For backwards compatibility only; modules should require the error themselves.]==] function export.err(lang_code, param, code_desc, template_tag, not_real_lang) return require("Module:languages/error")(lang_code, param, code_desc, template_tag, not_real_lang) end

return export