Module:User:Benwing2/links

local export = {}

--[=[	Unsupported titles, pages with high memory usage, extraction modules and part-of-speech names are listed at Module:links/data.

Other modules used: Module:script utilities Module:scripts Module:languages and its submodules Module:gender and number Module:debug/track ]=] local m_str_utils = require("Module:string utilities")

local anchorEncode = mw.uri.anchorEncode local concat = table.concat local decode_entities = m_str_utils.decode_entities local decode_uri = m_str_utils.decode_uri local find = string.find local encode_entities = require("Module:string/encode entities") -- Can't yet replace, as the Module:string utilities version no longer has automatic double-encoding prevention, which requires changes here to account for. local insert = table.insert local match = string.match local new_title = mw.title.new local remove = table.remove local shallowcopy = require("Module:table").shallowcopy local split = m_str_utils.split local sub = string.sub local toNFC = mw.ustring.toNFC local trim -- defined below local ulower = m_str_utils.lower local umatch = m_str_utils.match local unstrip = mw.text.unstrip local u = m_str_utils.char local TEMP_UNDERSCORE = u(0xFFF0) local pron_qualifier_module = "Module:pron qualifier"

local function track(page, code) local tracking_page = "links/" .. page if code then require("Module:debug/track"){tracking_page, tracking_page .. "/" .. code} else require("Module:debug/track")(tracking_page) end return true end

do local _trim = m_str_utils.trim -- Unconditionally trimmed charset. local always_trim = "\194\128-\194\159" ..			-- U+0080-009F (C1 control characters) "\194\173" ..					-- U+00AD (soft hyphen) "\226\128\170-\226\128\174" ..	-- U+202A-202E (directionality formatting characters) "\226\129\166-\226\129\169"		-- U+2066-2069 (directionality formatting characters) -- Standard trimmed charset. local standard_trim = "%s" .. -- (default whitespace charset) "\226\128\139-\226\128\141" ..	-- U+200B-200D (zero-width spaces) always_trim -- If there are non-whitespace characters, trim all characters in `standard_trim`. -- Otherwise, only trim the characters in `always_trim`. trim = function(text) if text == "" then return text end local trimmed = _trim(text, standard_trim) if trimmed ~= "" then return trimmed end return _trim(text, always_trim) end end

local function escape(text, str) local rep repeat text, rep = text:gsub("\\\\(\\*" .. str .. ")", "\5%1")	until rep == 0 return (text:gsub("\\" .. str, "\6")) end

local function unescape(text, str) return (text		:gsub("\5", "\\")		:gsub("\6", str)) end

-- Remove bold, italics, soft hyphens, strip markers and HTML tags. local function remove_formatting(str) str = str :gsub("('*)(.-'*)", "%1%2") :gsub("('*)(.-'*)", "%1%2") :gsub("­", "") return (unstrip(str)		:gsub("<[^<>]+>", "")) end

--[==[Takes an input and splits on a double slash (taking account of escaping backslashes).]==] function export.split_on_slashes(text) text = split(escape(text, "//"), "//", true) or {} for i, v in ipairs(text) do		text[i] = unescape(v, "//") if v == "" then text[i] = false end end return text end

-- Does the work of export.get_fragment, but can be called directly to avoid unnecessary checks for embedded links. local function get_fragment(text) text = escape(text, "#") -- Replace numeric character references with the corresponding character (&#29; → '), -- as they contain #, which causes the numeric character reference to be -- misparsed (wa'a → wa&#29;a → pagename wa&, fragment 29;a). text = decode_entities(text) local target, fragment = text:match("^(..-)#(.+)$") target = target or text target = unescape(target, "#") fragment = fragment and unescape(fragment, "#") return target, fragment end

--[==[Takes a link target and outputs the actual target and the fragment (if any).]==] function export.get_fragment(text) -- If there are no embedded links, process input. local open = find(text, "", 1, true)	if not open then		return get_fragment(text)	end	local close = find(text, "", open + 2, true) if not close then return get_fragment(text) -- If there is one, but it's redundant (i.e. encloses everything with no pipe), remove and process. elseif open == 1 and close == #text - 1 and not find(text, "|", 3, true) then return get_fragment(sub(text, 3, -3)) end -- Otherwise, return the input. return text end

local pos_tags function export.get_link_page(target, lang, sc, plain) if not target then return nil end target = remove_formatting(target)

-- Check if the target is an interwiki link. if target:match(":") and target ~= ":" then -- If this is an a link to another namespace or an interwiki link, ensure there's an initial colon and then return what we have (so that it works as a conventional link, and doesn't do anything weird like add the term to a category.) local prefix = target:gsub("^:*(.-):.*", ulower) if (			mw.loadData("Module:data/namespaces")[prefix] or			mw.loadData("Module:data/interwikis")[prefix]		) then return ":" .. target:gsub("^:+", ""), nil, {} end -- Convert any escaped colons target = target:gsub("\\:", ":") end

-- Check if the term is reconstructed and remove any asterisk. Otherwise, handle the escapes. local reconstructed, escaped if not plain then target, reconstructed = target:gsub("^%*(.)", "%1") end target, escaped = target:gsub("^(\\-)\\%*", "%1*")

if not require("Module:utilities").check_object("script", true, sc) or sc:getCode == "None" then sc = lang:findBestScript(target) end

-- Remove carets if they are used to capitalize parts of transliterations (unless they have been escaped). if (not sc:hasCapitalization) and sc:isTransliterated and target:match("%^") then target = escape(target, "^") :gsub("%^", "") target = unescape(target, "^") end

-- Get the entry name for the language. target = lang:makeEntryName(target, sc)

-- If the link contains unexpanded template parameters, then don't create a link. if target:find("{{{") then return nil end

if target:sub(1, 1) == "/" then return ":" .. target

elseif target:find("^Reconstruction:") then return target

-- Link to appendix for reconstructed terms and terms in appendix-only languages. Plain links interpret * literally, however. elseif reconstructed == 1 then if lang:getFullCode == "und" then return nil else target = "Reconstruction:" .. lang:getFullName .. "/" .. target end -- Reconstructed languages and substrates require an initial *. elseif lang:hasType("reconstructed") or lang:getFamilyCode == "qfa-sub" then local check = target:match("^:*([^:]*):") check = check and ulower(check) if (			mw.loadData("Module:data/namespaces")[check] or			mw.loadData("Module:data/interwikis")[check]		) then return target else error("The specified language " .. lang:getCanonicalName				.. " is unattested, while the given word is not marked with '*' to indicate that it is reconstructed.") end

elseif lang:hasType("appendix-constructed") then target = "Appendix:" .. lang:getFullName .. "/" .. target end

return target, escaped > 0 end

-- Make a link from a given link's parts local function make_link(link, lang, sc, id, isolated, plain, cats, no_alt_ast) -- Convert percent encoding to plaintext. link.target = decode_uri(link.target, "PATH") link.fragment = link.fragment and decode_uri(link.fragment, "PATH") -- Find fragments (if one isn't already set). -- Prevents {{l|en|word#Etymology 2|word}} from linking to word. -- # can be escaped as \#. if link.target and link.fragment == nil then link.target, link.fragment = get_fragment(link.target) end

-- Create a default display form. local auto_display = link.target -- Process the target local escaped link.target, escaped = export.get_link_page(link.target, lang, sc, plain)

-- If the display is the target and the reconstruction * has been escaped, remove the escaping backslash. if escaped then auto_display = auto_display:gsub("\\([^\\]*%*)", "%1", 1) end -- Process the display form. if link.display then local orig_display = link.display link.display = lang:makeDisplayText(link.display, sc, true) if cats then auto_display = lang:makeDisplayText(auto_display, sc) -- If the alt text is the same as what would have been automatically generated, then the alt parameter is redundant (e.g. {{l|en|foo|foo}}, {{l|en|w:foo|foo}}, but not {{l|en|w:foo|w:foo}}). -- If they're different, but the alt text could have been entered as the term parameter without it affecting the target page, then the target parameter is redundant (e.g. {{l|ru|фу|фу́}}). -- If `no_alt_ast` is true, use pcall to catch the error which will be thrown if this is a reconstructed lang and the alt text doesn't have *. if link.display == auto_display then insert(cats, lang:getFullName .. " links with redundant alt parameters") else local ok, check if no_alt_ast then ok, check = pcall(export.get_link_page, orig_display, lang, sc, plain) else ok = true check = export.get_link_page(orig_display, lang, sc, plain) end if ok and link.target == check then insert(cats, lang:getFullName .. " links with redundant target parameters") end end end else link.display = lang:makeDisplayText(auto_display, sc) end if not link.target then return link.display end -- If the target is the same as the current page, there is no sense id	-- and either the language code is "und" or the current L2 is the current -- language then return a "self-link" like the software does. if not id and link.target == mw.title.getCurrentTitle.prefixedText and (		lang:getFullCode == "und" or		require("Module:utilities").get_current_L2 == lang:getCanonicalName	) then return tostring(mw.html.create("strong")			:addClass("selflink")			:wikitext(link.display)) end

-- Add fragment. Do not add a section link to "Undetermined", as such sections do not exist and are invalid. -- TabbedLanguages handles links without a section by linking to the "last visited" section, but adding -- "Undetermined" would break that feature. For localized prefixes that make syntax error, please use the -- format: ["xyz"] = true. local prefix = link.target:match("^:*([^:]+):") prefix = prefix and ulower(prefix)

if prefix ~= "category" and not (prefix and mw.loadData("Module:data/interwikis")[prefix]) then if (link.fragment or link.target:find("#$")) and not plain then track("fragment", lang:getFullCode) if cats then insert(cats, lang:getFullName .. " links with manual fragments") end end

if (not link.fragment) and lang:getFullCode ~= "und" then if id then link.fragment = require("Module:senseid").anchor(lang, id) elseif not (link.target:find("^Appendix:") or link.target:find("^Reconstruction:") or plain) then link.fragment = lang:getFullName end elseif plain and id then link.fragment = id		end end -- Put inward-facing square brackets around a link to isolated spacing character(s). if isolated and #link.display > 0 and not umatch(decode_entities(link.display), "%S") then link.display = "&#x5D;" .. link.display .. "&#x5B;" end

link.target = link.target:gsub("^(:?)(.*)", function(m1, m2)		return m1 .. encode_entities(m2, "#%&+/:<=>@[\\]_{|}")	end) link.fragment = link.fragment and encode_entities(remove_formatting(link.fragment), "#%&+/:<=>@[\\]_{|}")

return "" .. link.display .. "" end

-- Split a link into its parts local function parse_link(linktext) local link = {target = linktext}

local target = link.target link.target, link.display = target:match("^(..-)|(.+)$") if not link.target then link.target = target link.display = target end

-- There's no point in processing these, as they aren't real links. local target_lower = link.target:lower for _, false_positive in ipairs({"category", "cat", "file", "image"}) do if target_lower:match("^" .. false_positive .. ":") then return nil end end

link.display = decode_entities(link.display) link.target, link.fragment = get_fragment(link.target)

-- So that make_link does not look for a fragment again. if not link.fragment then link.fragment = false end

return link end

-- Find embedded links and ensure they link to the correct section. local function process_embedded_links(text, data, plain) -- Process the non-linked text. text = data.lang:makeDisplayText(text, data.sc[1], true)

-- If the text begins with * and another character, then act as if each link begins with *. However, don't do this if the * is contained within a link at the start. E.g. `|*foo` would set all_reconstructed to true, while `|*foo` would not. local all_reconstructed = false if not plain then -- anchorEncode removes links etc.		if anchorEncode(text):sub(1, 1) == "*" then all_reconstructed = true end -- Otherwise, handle any escapes. text = text:gsub("^(\\-)\\%*", "%1*") end

if data.alt then track("alt-ignored") mw.log("(from Module:links)", "text with embedded wikilinks:", text,			"ignored alt:", data.alt, "lang:", data.lang:getFullCode) if data.cats then insert(data.cats, data.lang:getFullName .. " links with ignored alt parameters") end end if data.id then track("id-ignored") mw.log("(from Module:links)", "text with embedded wikilinks:", text,			"ignored id:", data.id, "lang:", data.lang:getFullCode) if data.cats then insert(data.cats, data.lang:getFullName .. " links with ignored id parameters") end end

local function process_link(space1, linktext, space2) local capture = "" .. linktext .. ""

local link = parse_link(linktext)

--Return unprocessed false positives untouched (e.g. categories). if not link then return capture end

if all_reconstructed and not link.target:find("^%*") then link.target = "*" .. link.target end

linktext = make_link(link, data.lang, data.sc, data.id, false, plain) :gsub("^%[%[", "\3") :gsub("%]%]$", "\4")

return space1 .. linktext .. space2 end

-- Use chars 1 and 2 as temporary substitutions, so that we can use charsets. These are converted to chars 3 and 4 by process_link, which means we can convert any remaining chars 1 and 2 back to square brackets (i.e. those not part of a link). text = text :gsub("%[%[", "\1") :gsub("%]%]", "\2") -- If the script uses ^ to capitalize transliterations, make sure that any carets preceding links are on the inside, so that they get processed with the following text. if text:match("%^") and not data.sc:hasCapitalization and data.sc:isTransliterated then text = escape(text, "^") :gsub("%^\1", "\1%^") text = unescape(text, "^") end text = text:gsub("\1(%s*)([^\1\2]-)(%s*)\2", process_link)

-- Remove the extra * at the beginning of a language link if it's immediately followed by a link whose display begins with * too. if all_reconstructed then text = text:gsub("^%*\3([^|\1-\4]+)|%*", "\3%1|*") end

return (text		:gsub("[\1\3]", "")		:gsub("[\2\4]", "")) end

local function handle_redundant_wikilink(data) local text = data.term if ( -- Filters out anything but "..." with no intermediate "" or "".		not match(text, "^%[%[") or -- Faster than sub(text, 1, 2) ~= "".		find(text, "[[", 3, true) or		find(text, "", 3, true) ~= #text - 1	) then		return	end	text = sub(text, 3, -3)	-- A pipe at the start of an embedded link is treated as part of the target (e.g. |foo: "|foo").	-- FIXME: This should be handled via a proper escape sequence.	local pipe = find(text, "|", 2, true)	local term = pipe and sub(text, 1, pipe - 1) or text	local title = new_title(term)	if title then		local namespace = title.namespace		-- Categories and files are false-positives.		if namespace == 6 or namespace == 14 then			return		end	end	data.term, data.alt = term, pipe and pipe ~= #text and sub(text, pipe + 1) or nil	if data.cats then		insert(data.cats, data.lang:getFullName .. " links with redundant wikilinks")	end end

--[==[Creates a basic link to the given term. It links to the language section (such as ), but it does not add language and script wrappers, so any code that uses this function should call the   from Module:script utilities to add such wrappers itself at some point. The first argument,, may contain the following items, a subset of the items used in the   argument of. If any other items are included, they are ignored. { {	term = entry_to_link_to, alt = link_text_or_displayed_text, lang = language_object, id = sense_id, } }
 * Text to turn into a link. This is generally the name of a page. The text can contain wikilinks already embedded in it. These are processed individually just like a single link would be. The  argument is ignored in this case.
 * Text to turn into a link. This is generally the name of a page. The text can contain wikilinks already embedded in it. These are processed individually just like a single link would be. The  argument is ignored in this case.


 * (optional)
 * The alternative display for the link, if different from the linked page. If this is {{code|lua|nil}}, the  argument is used instead (much like regular wikilinks). If   contains wikilinks in it, this argument is ignored and has no effect. (Links in which the alt is ignored are tracked with the tracking template {{whatlinkshere|tracking=links/alt-ignored}}.)


 * The language object for the term being linked. If this argument is defined, the function will determine the language's canonical name (see Template:language data documentation), and point the link or links in the  to the language's section of an entry, or to a language-specific senseid if the   argument is defined.
 * The language object for the term being linked. If this argument is defined, the function will determine the language's canonical name (see Template:language data documentation), and point the link or links in the  to the language's section of an entry, or to a language-specific senseid if the   argument is defined.


 * (optional)
 * Sense id string. If this argument is defined, the link will point to a language-specific sense id ({{ll|en|identifier|id=HTML}}) created by the template {{temp|senseid}}. A sense id consists of the language's canonical name, a hyphen, and the string that was supplied as the  argument. This is useful when a term has more than one sense in a language. If the   argument contains wikilinks, this argument is ignored. (Links in which the sense id is ignored are tracked with the tracking template {{whatlinkshere|tracking=links/id-ignored}}.)

The second argument is as follows:
 * If {{code|lua|true}}, the function will also generate links to the current page. The default ({{code|lua|false}}) will not generate a link but generate a bolded "self link" instead.
 * If {{code|lua|true}}, the function will also generate links to the current page. The default ({{code|lua|false}}) will not generate a link but generate a bolded "self link" instead.

The following special options are processed for each link (both simple text and with embedded wikilinks): function export.language_link(data) if type(data) ~= "table" then error("The first argument to the function language_link must be a table. See Module:links/documentation for more information.") -- Do we have a redundant wikilink? If so, remove it. elseif data.term then handle_redundant_wikilink(data) -- Nothing to process, return nil. elseif not data.alt then return nil end local text = data.term -- If we don't have a script, get one. if not data.sc then data.sc = data.lang:findBestScript(data.alt or text) end -- Do we have embedded wikilinks? If so, they need to be processed individually. if text then local open = find(text, "", 1, true)		if open and find(text, "", open + 2, true) then return process_embedded_links(text, data) end end -- If not, make a link using the parameters. text = text and trim(text) data.alt = data.alt and trim(data.alt) return make_link({target = text, display = data.alt, fragment = data.fragment}, data.lang, data.sc, data.id, true, nil, data.cats, data.no_alt_ast) end
 * The target page name will be processed to generate the correct entry name. This is done by the makeEntryName function in Module:languages, using the  replacements in the language's data file (see Template:language data documentation for more information). This function is generally used to automatically strip dictionary-only diacritics that are not part of the normal written form of a language.
 * If the text starts with, then the term is considered a reconstructed term, and a link to the Reconstruction: namespace will be created. If the text contains embedded wikilinks, then   is automatically applied to each one individually, while preserving the displayed form of each link as it was given. This allows linking to phrases containing multiple reconstructed terms, while only showing the * once at the beginning.
 * If the text starts with, then the link is treated as "raw" and the above steps are skipped. This can be used in rare cases where the page name begins with   or if diacritics should not be stripped. For example:
 * {{temp|l|en|*nix}} links to the nonexistent page Reconstruction:English/nix ( is interpreted as a reconstruction), but {{temp|l|en|:*nix}} links to *nix.
 * {{temp|l|sl|Franche-Comté}} links to the nonexistent page Franche-Comte ( is converted to   by  ), but {{temp|l|sl|:Franche-Comté}} links to Franche-Comté.]==]

function export.plain_link(data) if type(data) ~= "table" then error("The first argument to the function language_link must be a table. See Module:links/documentation for more information.") -- Do we have a redundant wikilink? If so, remove it. elseif data.term then handle_redundant_wikilink(data) -- Only have alt (or nothing), just return it. else return data.alt end -- Make sure the language is "und". local lang = data.lang if not lang or lang:getCode ~= "und" then data.lang = require("Module:languages").getByCode("und") end local text = data.term -- If we don't have a script, get one. if not data.sc then data.sc = require("Module:scripts").findBestScriptWithoutLang(data.alt or text) end -- Do we have embedded wikilinks? If so, they need to be processed individually. local open = find(text, "", 1, true)	if open and find(text, "", open + 2, true) then return process_embedded_links(text, data) end -- If not, make a link using the parameters. text = trim(text) data.alt = data.alt and trim(data.alt) return make_link({target = text, display = data.alt, fragment = data.fragment}, data.lang, data.sc, data.id, true, true) end

--[==[Replace any links with links to the correct section, but don't link the whole text if no embedded links are found. Returns the display text form.]==] function export.embedded_language_links(data) if type(data) ~= "table" then error("The first argument to the function language_link must be a table. See Module:links/documentation for more information.") end

local text = data.term -- If we don't have a script, get one. if not data.sc then data.sc = data.lang:findBestScript(text) end -- Do we have embedded wikilinks? If so, they need to be processed individually. local open = find(text, "", 1, true)	if open and find(text, "", open + 2, true) then return process_embedded_links(text, data) end -- If not, return the display text. text = trim(text) -- FIXME: Double-escape any percent-signs, because we don't want to treat non-linked text as having percent-encoded characters. This is a hack: percent-decoding should come out of Module:languages and only dealt with in this module, as it's specific to links. text = text:gsub("%%", "%%25") return (data.lang:makeDisplayText(text, data.sc, true)) end

function export.mark(text, item_type, face, lang) local tag = { "", "" }

if item_type == "gloss" then tag = { ' “ ', ' ” ' }	elseif item_type == "tr" then if face == "term" then tag = { '', ' ' }		else tag = { '', ' ' } end elseif item_type == "ts" then -- \226\129\160 = word joiner (zero-width non-breaking space) U+2060 tag = { ' /\226\129\160', '\226\129\160/ ' } elseif item_type == "pos" then tag = { ' ', ' ' } elseif item_type == "annotations" then tag = { ' ( ',			' ) ' } end

if type(text) == "string" then return tag[1] .. text .. tag[2] else return "" end end

--[==[Formats the annotations that are displayed with a link created by {{code|lua|full_link}}. Annotations are the extra bits of information that are displayed following the linked term, and include things such as gender, transliteration, gloss and so on. function export.format_link_annotations(data, face) local output = {}
 * The first argument is a table possessing some or all of the following keys:
 * Table containing a list of gender specifications in the style of Module:gender and number.
 * Transliteration.
 * Gloss that translates the term in the link, or gives some other descriptive information.
 * Part of speech of the linked term. If the given argument matches one of the templates in Category:Part of speech tags, then call that to show a part-of-speech tag. Otherwise, just show the given text as it is.
 * Literal meaning of the term, if the usual meaning is figurative or idiomatic.
 * Any of the above values can be omitted from the  argument. If a completely empty table is given (with no annotations at all), then an empty string is returned.
 * The second argument is a string. Valid values are listed in Module:script utilities/data "data.translit" table.]==]
 * Part of speech of the linked term. If the given argument matches one of the templates in Category:Part of speech tags, then call that to show a part-of-speech tag. Otherwise, just show the given text as it is.
 * Literal meaning of the term, if the usual meaning is figurative or idiomatic.
 * Any of the above values can be omitted from the  argument. If a completely empty table is given (with no annotations at all), then an empty string is returned.
 * The second argument is a string. Valid values are listed in Module:script utilities/data "data.translit" table.]==]
 * The second argument is a string. Valid values are listed in Module:script utilities/data "data.translit" table.]==]

-- Interwiki link if data.interwiki then insert(output, data.interwiki) end

-- Genders if type(data.genders) ~= "table" then data.genders = { data.genders } end

if data.genders and #data.genders > 0 then local m_gen = require("Module:gender and number") insert(output, " " .. m_gen.format_list(data.genders, data.lang)) end

local annotations = {}

-- Transliteration and transcription if data.tr and data.tr[1] or data.ts and data.ts[1] then local kind if face == "term" then kind = face else kind = "default" end

if data.tr[1] and data.ts[1] then insert(annotations,				require("Module:script utilities").tag_translit(data.tr[1], data.lang, kind)				.. " " .. export.mark(data.ts[1], "ts")) elseif data.ts[1] then insert(annotations, export.mark(data.ts[1], "ts")) else insert(annotations,				require("Module:script utilities").tag_translit(data.tr[1], data.lang, kind)) end end

-- Gloss/translation if data.gloss then insert(annotations, export.mark(data.gloss, "gloss")) end

-- Part of speech if data.pos then -- debug category for pos= containing transcriptions if data.pos:find("/[^><]*/") then data.pos = data.pos .. ""		end

pos_tags = pos_tags or mw.loadData("Module:links/data").pos_tags insert(annotations, export.mark(pos_tags[data.pos] or data.pos, "pos")) end

-- Literal/sum-of-parts meaning if data.lit then insert(annotations, "literally " .. export.mark(data.lit, "gloss")) end

if #annotations > 0 then insert(output, " " .. export.mark(concat(annotations, ", "), "annotations")) end

return concat(output) end

-- Add any left or right regular or accent qualifiers, labels or references to a formatted term. `data` is the object -- specifying the term, which should optionally contain: -- * a language object in `lang`; required if any accent qualifiers or labels are given; -- * left regular qualifiers in `q` (an array of strings or a single string); an empty array or blank string will be --  ignored; -- * right regular qualifiers in `qq` (an array of strings or a single string); an empty array or blank string will be --  ignored; -- * left accent qualifiers in `a` (an array of strings); an empty array will be ignored; -- * right accent qualifiers in `aa` (an array of strings); an empty array will be ignored; -- * left labels in `l` (an array of strings); an empty array will be ignored; -- * right labels in `ll` (an array of strings); an empty array will be ignored; -- * references in `refs`, an array either of strings (formatted reference text) or objects containing fields `text` --  (formatted reference text) and optionally `name` and/or `group`. -- `formatted` is the formatted version of the term itself. local function add_qualifiers_and_refs_to_term(data, formatted) local q = data.q	if type(q) == "string" then q = {q} end local qq = data.qq	if type(qq) == "string" then qq = {qq} end if q and q[1] or qq and qq[1] or data.a and data.a[1] or data.aa and data.aa[1] or data.l and data.l[1] or		data.ll and data.ll[1] or data.refs and data.refs[1] then formatted = require(pron_qualifier_module).format_qualifiers { lang = data.lang, text = formatted, q = q,			qq = qq, a = data.a,			aa = data.aa, l = data.l,			ll = data.ll, refs = data.refs, }	end

return formatted end

--[==[Creates a full link, with annotations (see ), in the style of {{temp|l}} or {{temp|m}}. The first argument,, must be a table. It contains the various elements that can be supplied as parameters to {{temp|l}} or {{temp|m}}: { {	term = entry_to_link_to, alt = link_text_or_displayed_text, lang = language_object, sc = script_object, track_sc = boolean, fragment = link_fragment id = sense_id, genders = { "gender1", "gender2", ... },	tr = transliteration, ts = transcription, gloss = gloss, pos = part_of_speech_tag, lit = literal_translation, no_alt_ast = boolean, accel = {accelerated_creation_tags}, interwiki = interwiki, q = { "left_qualifier1", "left_qualifier2", ...} or "left_qualifier", qq = { "right_qualifier1", "right_qualifier2", ...} or "right_qualifier", refs = { "formatted_ref1", "formatted_ref2", ...} or { {text = "text", name = "name", group = "group"}, ... }, } } Any one of the items in the  table may be {{code|lua|nil}}, but an error will be shown if neither   nor   nor   is present. Thus, calling {{code|lua|2=full_link{ term = term, lang = lang, sc = sc } }}, where  is an entry name,    is a language object from Module:languages, and   is a script object from Module:scripts, will give a plain link similar to the one produced by the template {{temp|l}}, and calling {{code|lua|2=full_link( { term = term, lang = lang, sc = sc }, "term" )}} will give a link similar to the one produced by the template {{temp|m}}. The function will: function export.full_link(data, face, allow_self_link, show_qualifiers) -- Prevent data from being destructively modified. local data = shallowcopy(data)
 * Try to determine the script, based on the characters found in the term or alt argument, if the script was not given. If a script is given and  is {{code|lua|true}}, it will check whether the input script is the same as the one which would have been automatically generated and add the category Category:Terms with redundant script codes if yes, or Category:Terms with non-redundant manual script codes if no. This should be used when the input script object is directly determined by a template's   parameter.
 * Call  on the term or alt forms, to remove diacritics in the page name, process any embedded wikilinks and create links to Reconstruction or Appendix pages when necessary.
 * Call  to add the appropriate language and script tags to the term, and to italicize terms written in the Latin script if necessary. Accelerated creation tags, as used by WT:ACCEL, are included.
 * Generate a transliteration, based on the alt or term arguments, if the script is not Latin and no transliteration was provided.
 * Add the annotations (transliteration, gender, gloss etc.) after the link.
 * If  is specified, then the alt text does not need to contain an asterisk if the language is reconstructed. This should only be used by modules which really need to allow links to reconstructions that don't display asterisks (e.g. number boxes).
 * If  is specified, left and right qualifiers and references will be displayed. (This is for compatibility reasons, since a fair amount of code stores qualifiers and/or references in these fields and displays them itself, expecting {{code|lua|full_link}} to ignore them.]==]

if type(data) ~= "table" then error("The first argument to the function full_link must be a table. "			.. "See Module:links/documentation for more information.") end

local terms = {true}

-- Generate multiple forms if applicable. for _, param in ipairs{"term", "alt"} do		if type(data[param]) == "string" and data[param]:find("//") then data[param] = export.split_on_slashes(data[param]) elseif type(data[param]) == "string" and not (type(data.term) == "string" and data.term:find("//")) then data[param] = data.lang:generateForms(data[param]) else data[param] = {} end end

for _, param in ipairs{"sc", "tr", "ts"} do		data[param] = {data[param]} end

for _, param in ipairs{"term", "alt", "sc", "tr", "ts"} do		for i in pairs(data[param]) do			terms[i] = true end end -- Create the link local output = {} data.cats = {} local link = "" local annotations

for i in ipairs(terms) do -- Is there any text to show? if (data.term[i] or data.alt[i]) then -- Try to detect the script if it was not provided local display_term = data.alt[i] or data.term[i] local best = data.lang:findBestScript(display_term) if (				best:getCode == "None" and				require("Module:scripts").findBestScriptWithoutLang(display_term):getCode ~= "None"			) then insert(data.cats, data.lang:getFullName .. " terms in nonstandard scripts") end if not data.sc[i] then data.sc[i] = best -- Track uses of sc parameter. elseif data.track_sc then if data.sc[i]:getCode == best:getCode then insert(data.cats, data.lang:getFullName .. " terms with redundant script codes") else insert(data.cats, data.lang:getFullName .. " terms with non-redundant manual script codes") end end

-- If using a discouraged character sequence, add to maintenance category if data.sc[i]:hasNormalizationFixes == true then if (data.term[i] and data.sc[i]:fixDiscouragedSequences(toNFC(data.term[i])) ~= toNFC(data.term[i])) or (data.alt[i] and data.sc[i]:fixDiscouragedSequences(toNFC(data.alt[i])) ~= toNFC(data.alt[i])) then insert(data.cats, "Pages using discouraged character sequences") end end

local class = ""

-- Encode certain characters to avoid various delimiter-related issues at various stages. We need to encode -- because they end up forming part of CSS class names inside of and will interfere with finding the end -- of the HTML tag. I first tried converting them to URL encoding, i.e. %3C and %3E; they then appear in the URL as -- %253C and %253E, which get mapped back to %3C and %3E when passed to Module:accel. But mapping them to &lt; -- and &gt; somehow works magically without any further work; they appear in the URL as, and get passed to -- Module:accel as. I have no idea who along the chain of calls is doing the encoding and decoding. If -- someone knows, please modify this comment appropriately! local encode_accel_char_map = { ["%"] = ".",				[" "] = "_",				["_"] = TEMP_UNDERSCORE, ["<"] = "&lt;", [">"] = "&gt;", }			local function encode_accel_param_chars(param) local retval = param:gsub("[% <>_]", encode_accel_char_map) -- discard second return value return retval end

local function encode_accel_param(prefix, param) if not param then return "" end if type(param) == "table" then local filled_params = {} -- There may be gaps in the sequence, especially for translit params. local maxindex = 0 for k, v in pairs(param) do						if type(k) == "number" and k > maxindex then maxindex = k						end end for i=1,maxindex do						filled_params[i] = param[i] or "" end -- Module:accel splits these up again. param = concat(filled_params, "*~!") end -- This is decoded again by WT:ACCEL. return prefix .. encode_accel_param_chars(param) end

if data.accel then local form = data.accel.form and encode_accel_param_chars(data.accel.form) .. "-form-of" or "" local gender = encode_accel_param("gender-", data.accel.gender) local pos = encode_accel_param("pos-", data.accel.pos) local translit = encode_accel_param("transliteration-",					data.accel.translit or (data.tr[i] ~= "-" and data.tr[i] or nil)) local target = encode_accel_param("target-", data.accel.target) local lemma = encode_accel_param("origin-", data.accel.lemma) local lemma_translit = encode_accel_param("origin_transliteration-", data.accel.lemma_translit) local no_store = data.accel.no_store and "form-of-nostore" or ""

local accel = form .. " " ..					gender .. " " ..					pos .. " " ..					translit .. " " ..					target .. " " ..					lemma .. " " ..					lemma_translit .. " " ..					no_store .. " "

class = "form-of lang-" .. data.lang:getFullCode .. " " .. accel end

-- Only make a link if the term has been given, otherwise just show the alt text without a link local term_data = { term = data.term[i], alt = data.alt[i], lang = data.lang, sc = data.sc[i], fragment = data.fragment, id = data.id, genders = data.genders, tr = data.tr[i], ts = data.ts[i], gloss = data.gloss, pos = data.pos, lit = data.lit, accel = data.accel, interwiki = data.interwiki, cats = data.cats, no_alt_ast = data.no_alt_ast }			link = require("Module:script utilities").tag_text(				data.term[i] and export.language_link(term_data)				or data.alt[i], data.lang, data.sc[i], face, class) else --	No term to show.					Is there at least a transliteration we can work from? link = require("Module:script utilities").request_script(data.lang, data.sc[i]) -- No link to show, and no transliteration either. Show a term request (unless it's a substrate, as they rarely take terms). if (link == "" or (not data.tr[i]) or data.tr[i] == "-") and data.lang:getFamilyCode ~= "qfa-sub" then -- If there are multiple terms, break the loop instead. if i > 1 then remove(output) break elseif mw.title.getCurrentTitle.nsText ~= "Template" then insert(data.cats, data.lang:getFullName .. " term requests") end link = " [Term?] " end end insert(output, link) if i < #terms then insert(output, "／ ") end end

-- TODO: Currently only handles the first transliteration, pending consensus on how to handle multiple translits for multiple forms, as this is not always desirable (e.g. traditional/simplified Chinese). if data.tr[1] == "" or data.tr[1] == "-" then data.tr[1] = nil

else local phonetic_extraction = mw.loadData("Module:links/data").phonetic_extraction phonetic_extraction = phonetic_extraction[data.lang:getCode] or phonetic_extraction[data.lang:getFullCode]

if phonetic_extraction then data.tr[1] = data.tr[1] or require(phonetic_extraction).getTranslit(export.remove_links(data.alt[1] or data.term[1]))

elseif (data.term[1] or data.alt[1]) and data.sc[1]:isTransliterated then -- Track whenever there is manual translit. The categories below like 'terms with redundant transliterations' -- aren't sufficient because they only work with reference to automatic translit and won't operate at all in -- languages without any automatic translit, like Persian and Hebrew. if data.tr[1] then track("manual-tr", data.lang:getFullCode) end

-- Try to generate a transliteration, unless transliteration has been supplied and data.no_check_redundant_translit is -- given. (Checking for redundant transliteration can use up significant amounts of memory so we don't want to do it			-- if memory is tight. `no_check_redundant_translit` is currently set when called ultimately from			-- {{multitrans|...|no-check-redundant-translit=1}}.) if not (data.tr[1] and data.no_check_redundant_translit) then local text = data.alt[1] or data.term[1] if not data.lang:link_tr(data.sc[1]) then text = export.remove_links(text, true) end

local automated_tr, tr_categories automated_tr, data.tr_fail, tr_categories = data.lang:transliterate(text, data.sc[1])

if automated_tr or data.tr_fail then local manual_tr = data.tr[1]

if manual_tr then if (export.remove_links(manual_tr) == export.remove_links(automated_tr)) and (not data.tr_fail) then insert(data.cats, data.lang:getFullName .. " terms with redundant transliterations") elseif not data.tr_fail then -- Prevents Arabic root categories from flooding the tracking categories. if mw.title.getCurrentTitle.nsText ~= "Category" then insert(data.cats, data.lang:getFullName .. " terms with non-redundant manual transliterations") end end end if (not manual_tr) or data.lang:overrideManualTranslit(data.sc[1]) then data.tr[1] = automated_tr for _, category in ipairs(tr_categories) do							insert(data.cats, category) end end end end end end

-- Link to the transliteration entry for languages that require this if data.tr[1] and data.lang:link_tr(data.sc[1]) and not (data.tr[1]:match("%[%[(.-)%]%]") or data.tr_fail) then data.tr[1] = export.language_link{ lang = data.lang, term = data.tr[1], sc = require("Module:scripts").getByCode("Latn") }	elseif data.tr[1] and not (data.lang:link_tr(data.sc[1]) or data.tr_fail) then -- Remove the pseudo-HTML tags added by remove_links. data.tr[1] = data.tr[1]:gsub("", "") end if data.tr[1] and not umatch(data.tr[1], "[^%s%p]") then data.tr[1] = nil end

insert(output, export.format_link_annotations(data, face))

local categories = #data.cats > 0 and require("Module:utilities").format_categories(data.cats, data.lang, "-", nil, nil, data.sc) or ""

output = concat(output) if show_qualifiers then output = add_qualifiers_and_refs_to_term(data, output) end return output .. categories end

--[==[Replaces all wikilinks with their displayed text, and removes any categories. This function can be invoked either from a template or from another module. -- Strips links: deletes category links, the targets of piped links, and any double square brackets involved in links (other than file links, which are untouched). If `tag` is set, then any links removed will be given pseudo-HTML tags, which allow the substitution functions in Module:languages to properly subdivide the text in order to reduce the chance of substitution failures in modules which scrape pages like Module:zh-translit. -- FIXME: This is quite hacky. We probably want this to be integrated into Module:languages, but we can't do that until we know that nothing is pushing pipe linked transliterations through it for languages which don't have link_tr set. function export.remove_links(text, tag) if type(text) == "table" then text = text.args[1] end
 * &rarr;
 * &rarr;
 * &rarr; (nothing)]==]

if not text or text == "" then return "" end

text = text :gsub("%[%[", "\1") :gsub("%]%]", "\2")

-- Parse internal links for the display text. text = text:gsub("(\1)([^\1\2]-)(\2)",		function(c1, c2, c3)			-- Don't remove files.			for _, false_positive in ipairs({"file", "image"}) do				if c2:lower:match("^" .. false_positive .. ":") then return c1 .. c2 .. c3 end			end			-- Remove categories completely.			for _, false_positive in ipairs({"category", "cat"}) do				if c2:lower:match("^" .. false_positive .. ":") then return "" end			end			-- In piped links, remove all text before the pipe, unless it's the final character (i.e. the pipe trick), in which case just remove the pipe.			c2 = c2:match("^[^|]*|(.+)") or c2:match("([^|]+)|$") or c2			if tag then				return " " .. c2 .. " "			else				return c2			end		end)

text = text :gsub("\1", "")		:gsub("\2", "")

return text end

--[=[ This decodes old section encodings. For example, Norwegian_Bokm.C3.A5l → Norwegian_Bokmål. It isn't picky about whether the section encodings represent the UTF-8 encoding of a real Unicode character, so it will mangle section names that contain a period followed by two uppercase hex characters. At least such section names are probably pretty rare.

Wiktionary adds an additional id="" attribute for sections using a legacy encoding, if it is different from the modern minimally modified attribute. It is like percent encoding (URI or URL encoding) except with "." instead of "%". See mw:Manual:$wgFragmentMode and the code that does the encoding at https://gerrit.wikimedia.org/r/plugins/gitiles/mediawiki/core/+/7bf779524ab1fd8e1d74f79ea4840564d48eea4d/includes/parser/Sanitizer.php#893 ]=]

function export.section_link(link) if type(link) ~= "string" then error("The first argument to section_link was a " .. type(link) .. ", but it should be a string.") end link = link:gsub("_", " ") local target, section = link:match("(.-)#(.*)") if not target then error("The function “section_link” could not find a number sign marking a section name.") end return export.plain_link{ term = target, fragment = section, alt = link:gsub("#", " § ", 1) } end

return export