Dokumentasjon for modulen kan opprettast på Modul:links/dok

local export = {}

--[=[
	[[Unsupported titles]], pages with high memory usage,
	extraction modules and part-of-speech names are listed
	at [[Module:links/data]].

	Other modules used:
		[[Module:script utilities]]
		[[Module:scripts]]
		[[Module:languages]] and its submodules
		[[Module:gender and number]]
		[[Module:debug/track]]
]=]

-- These are prefixed with u to avoid confusion with the default string methods
-- of the same name.
local toNFC = mw.ustring.toNFC
local uchar = mw.ustring.char
local ufind = mw.ustring.find
local ugsub = mw.ustring.gsub
local usub = mw.ustring.sub

local function track(page, code)
	local tracking_page = "links/" .. page
	if code then
		require("Module:debug/track"){tracking_page, tracking_page .. "/" .. code}
	else
		require("Module:debug/track")(tracking_page)
	end
	return true
end

local get_entities = require("Module:utilities").get_entities

local table_insert = table.insert
local table_concat = table.concat

local ignore_cap
local pos_tags
function export.getLinkPage(target, lang, sc)
	
	-- Remove diacritics from the page name
	target = (lang:makeEntryName(target, sc))
	
	-- If the link contains unexpanded template parameters, then don't create a link.
	if target:find("{{{") then
		return nil
	end
	
	if target:sub(1, 1) == "/" then
		return ":" .. target
	
	elseif get_entities(target):find("^%*?Reconstruction:") then
		return target:gsub("^%*", "")
		
	-- Link to appendix for reconstructed terms and terms in appendix-only languages
	elseif target:sub(1, 1) == "*" and #target > 1 then
		if lang:getCode() == "und" then
			return nil
		end

		target = "Reconstruction:" .. lang:getCanonicalName() .. "/" .. usub(target, 2)
		
	elseif lang:getType() == "reconstructed" then
		local check, m_utildata = target:match("^:*([^:]*):"), mw.loadData("Module:utilities/data")
		if m_utildata.interwikis[check] or m_utildata.namespaces[check] then
			return target
		else
			error("The specified language " .. lang:getCanonicalName()
				.. " is unattested, while the given word is not marked with '*' to indicate that it is reconstructed.")
		end
		
	elseif lang:getType() == "appendix-constructed" then
		target = "Appendix:" .. lang:getCanonicalName() .. "/" .. target
	end

	return target
end

-- Make a language-specific link from given link's parts
local function makeLangLink(link, lang, sc, id, allow_self_link)
	
	-- Find fragments (when link didn't come from parseLink).
	-- Prevents {{l|en|word#Etymology 2|word}} from linking to [[word#Etymology 2#English]].
	if link.fragment == nil then
		-- Replace numeric character references with the corresponding character ( → '),
		-- as they contain #, which causes the numeric character reference to be
		-- misparsed (wa'a → waa → pagename wa&, fragment 29;a).
		link.target = link.target:find("&#[^;]+;") and get_entities(link.target) or link.target
		local first, second = link.target:match("^([^#]+)#(.+)$")
		if first then
			link.target, link.fragment = first, second
		end
	end

	-- If there is no display form, then create a default one.
	if not link.display then link.display = link.target end
	
	-- Process the target
	link.target = export.getLinkPage(link.target, lang, sc)
	
	-- Process the display form.
	link.display = (lang:makeDisplayText(link.display, sc))

	if not link.target then
		return link.display
	end

	-- If the target is the same as the current page and there is no sense id
	-- and linking to the same page hasn't been turned on, then return a "self-link"
	-- like the software does.
	local selfCheckTarget = get_entities(mw.uri.decode(link.target):gsub("^:", ""))
	local selfCheckCurrentTitle = mw.title.getCurrentTitle().prefixedText
	if selfCheckTarget:find("[#%%&+/:<=>?@[\\%]_{|}]") then
		selfCheckTarget = (lang:makeEntryName(selfCheckTarget))
		selfCheckCurrentTitle = (lang:makeEntryName(selfCheckCurrentTitle))
	end
	if not (allow_self_link or id) and selfCheckTarget == selfCheckCurrentTitle then
		return "<strong class=\"selflink\">" .. link.display .. "</strong>"
	end

	--[[
		Add fragment
		Do not add a section link to "Undetermined", as such sections do not exist and are invalid.
		TabbedLanguages handles links without a section by linking to the "last visited" section,
		but adding "Undetermined" would break that feature.
		For localized prefixes that make syntax error, please use the format: ["xyz"] = true,
	]]
	local prefix = link.target:match("^:?([^:]+):")
	local prefixes = {
		w = true,
		wikipedia = true,
		Category = true,
	}

	if not prefixes[prefix] then
		if link.fragment or link.target:find("#$") then
			track("fragment", lang:getCode())
		end

		if not link.fragment and lang:getCode() ~= "und" then
			if id then
				link.fragment = require("Module:senseid").anchor(lang, id)
			elseif not ufind(link.target, "^Appendix:")
					and not ufind(link.target, "^Reconstruction:") then
				link.fragment = lang:getCanonicalName()
			end
		end
	end

	return "[[" .. link.target .. (link.fragment and "#" .. link.fragment or "") .. "|" .. link.display .. "]]"
end


-- Split a link into its parts
local function parseLink(linktext)
	local link = { target = linktext }
	local first, second = link.target:match("^([^|]+)|(.+)$")
	
	-- Prevent characters whose HTML entities are unsupported titles from being incorrectly recognised as the entity if they are in a link being re-parsed (e.g. "&" becomes "&amp;" when returned, but "&amp;" is also an unsupported title. If "&" is given as a link which is then re-parsed, we don't want it to be perceived as "&amp;".)
	if link.target:match("&[^;]+;") then
		local unsupported_titles = mw.loadData("Module:links/data").unsupported_titles
		if unsupported_titles[second] and unsupported_titles[second] ~= first then
			link.target = get_entities(link.target)
			first, second = link.target:match("^([^|]+)|(.+)$")
		end
	end

	if first then
		link.target = first
		link.display = second
	else
		link.display = link.target
	end
	
	-- There's no point in processing these, as they aren't real links.
	for _, falsePositive in ipairs({"Category", "File", "Image"}) do
		if link.target:match("^" .. falsePositive .. ":") then return nil end
	end

	first, second = link.target:match("^(.+)#(.+)$")

	if first then
		link.target = first
		link.fragment = second
	else
		-- So that makeLangLink does not look for a fragment again
		link.fragment = false
	end

	return link
end

-- Find embedded links and ensure they link to the correct section.
local function process_embedded_links(text, data, allow_self_link, allReconstructed)
	if data.alt then
		track("alt-ignored")
		mw.log("(from Module:links)", "text with embedded wikilinks:", text,
			"ignored alt:", data.alt, "lang:", data.lang:getCode())
	end

	if data.id then
		track("id-ignored")
		mw.log("(from Module:links)", "text with embedded wikilinks:", text,
			"ignored id:", data.id, "lang:", data.lang:getCode())
	end

	-- Begins and ends with a wikilink tag
	if text:find("^%[%[(.+)%]%]$") then
		-- There are no [ ] in between.
		-- This makes the wikilink tag redundant.
		if text:find("^%[%[[^%[%]]+%]%]$") then
			track("redundant wikilink")
		else
			local temp = text:gsub("^%[%[(.+)%]%]$", "%1")
			temp = temp:gsub("%]%], %[%[", "|")

			if not temp:find("[%[%]]") then
				track("list")
			end
		end
	end

	text = text:gsub("(%[%[([^%]]+)%]%])",
		function(capture, linktext)
			local link = parseLink(linktext)
			
			--Return unprocessed false positives untouched (e.g. categories).
			if not link then return capture end
			
			if allReconstructed and not link.target:find("^%*") then
				link.target = "*" .. link.target
			end
			
			return makeLangLink(link, data.lang, data.sc, data.id, allow_self_link)
		end)

	-- Remove the extra * at the beginning if it's immediately followed by a link whose display begins with * too
	if allReconstructed then
		text = text:gsub("^%*%[%[([^|%]]+)|%*", "[[%1|*")
	end
	
	local patterns = {"^(%s*)(.-)(%s*%[%[)", "(%]%]%s*)(.-)(%s*%[%[)", "(.*%[%[.-%]%]%s*)(.-)(%s*)$"}
	
	return (data.lang:gsubSubstitutions(text, data.sc[1], "makeDisplayText", patterns))
end

-- Creates a basic wikilink to the given term. If the text already contains links, these are replaced with links to the correct section.
function export.language_link(data, allow_self_link)
	if type(data) ~= "table" then
		error("The first argument to the function language_link must be a table. See Module:links/documentation for more information.")
	end
	
	local text = data.term
	data.sc = data.sc or data.lang:findBestScript(text)

	ignore_cap = ignore_cap or mw.loadData("Module:links/data").ignore_cap
	if ignore_cap[data.lang:getCode()] and text then
		text = text:gsub("%^", "")
	end
	
	-- If the text begins with * and another character,
	-- then act as if each link begins with *
	local allReconstructed = false

	if require("Module:utilities").get_plaintext(text):match("^*.") then
		allReconstructed = true
	end

	-- Do we have embedded wikilinks?
	if text:find("%[%[") then
		--[=[
		[[Special:WhatLinksHere/Template:tracking/links/alt-ignored]]
		[[Special:WhatLinksHere/Template:tracking/links/id-ignored]]
		]=]
		text = process_embedded_links(text, data, allow_self_link, allReconstructed)
	else
		-- If there are no embedded wikilinks, make a link using the parameters.
		text = makeLangLink({ target = text, display = data.alt }, data.lang, data.sc, data.id, allow_self_link)
	end

	return text
end

-- Replace any links with links to the correct section, but don't link the whole text if no embedded links are found. Returns the display text form.
function export.embedded_language_links(data, allow_self_link)
	if type(data) ~= "table" then
		error("The first argument to the function language_link must be a table. See Module:links/documentation for more information.")
	end
	
	local text = data.term
	data.sc = data.sc or data.lang:findBestScript(text)
	
	-- Do we have embedded wikilinks?
	if text:find("%[%[") then
		--[=[
		[[Special:WhatLinksHere/Template:tracking/links/alt-ignored]]
		[[Special:WhatLinksHere/Template:tracking/links/id-ignored]]
		]=]
		text = process_embedded_links(text, data, allow_self_link, allReconstructed)
	else
		-- If there are no embedded wikilinks, return the display text.
		text = (data.lang:makeDisplayText(text, data.sc))
	end
	
	return text
end

function export.mark(text, itemType, face, lang)
	local tag = { "", "" }

	if itemType == "gloss" then
		tag = { '<span class="mention-gloss-double-quote">“</span><span class="mention-gloss">',
			'</span><span class="mention-gloss-double-quote">”</span>' }
	elseif itemType == "tr" then
		if face == "term" then
			tag = { '<span lang="' .. lang:getCode() .. '" class="tr mention-tr Latn">',
				'</span>' }
		else
			tag = { '<span lang="' .. lang:getCode() .. '" class="tr Latn">', '</span>' }
		end
	elseif itemType == "ts" then
		tag = { '<span class="ts mention-ts Latn">/', '/</span>' }
	elseif itemType == "pos" then
		tag = { '<span class="ann-pos">', '</span>' }
	elseif itemType == "annotations" then
		tag = { '<span class="mention-gloss-paren annotation-paren">(</span>',
			'<span class="mention-gloss-paren annotation-paren">)</span>' }
	end

	if type(text) == "string" then
		return tag[1] .. text .. tag[2]
	else
		return ""
	end
end

-- Format the annotations (things following the linked term)
function export.format_link_annotations(data, face)
	local output = {}

	-- Interwiki link
	if data.interwiki then
		table_insert(output, data.interwiki)
	end

	-- Genders
	if type(data.genders) ~= "table" then
		data.genders = { data.genders }
	end

	if data.genders and #data.genders > 0 then
		local m_gen = require("Module:gender and number")
		table_insert(output, "&nbsp;" .. m_gen.format_list(data.genders, data.lang))
	end

	local annotations = {}

	-- Transliteration and transcription
	if data.tr and data.tr[1] or data.ts and data.ts[1] then
		local kind
		if face == "term" then
			kind = face
		else
			kind = "default"
		end

		if data.tr[1] and data.ts[1] then
			table_insert(annotations,
				require("Module:script utilities").tag_translit(data.tr[1], data.lang, kind)
				.. " " .. export.mark(data.ts[1], "ts"))
		elseif data.ts[1] then
			table_insert(annotations, export.mark(data.ts[1], "ts"))
		else
			table_insert(annotations,
				require("Module:script utilities").tag_translit(data.tr[1], data.lang, kind))
		end
	end

	-- Gloss/translation
	if data.gloss then
		table_insert(annotations, export.mark(data.gloss, "gloss"))
	end

	-- Part of speech
	if data.pos then
		-- debug category for pos= containing transcriptions
		if data.pos:find("/[^><]*/") then
			data.pos = data.pos .. "[[Category:links likely containing transcriptions in pos]]"
		end

		pos_tags = pos_tags or mw.loadData("Module:links/data").pos_tags
		table_insert(annotations, export.mark(pos_tags[data.pos] or data.pos, "pos"))
	end

	-- Literal/sum-of-parts meaning
	if data.lit then
		table_insert(annotations, "literally " .. export.mark(data.lit, "gloss"))
	end

	if #annotations > 0 then
		table_insert(output, " " .. export.mark(table_concat(annotations, ", "), "annotations"))
	end

	return table_concat(output)
end

-- A version of {{l}} or {{m}} that can be called from other modules too
function export.full_link(data, face, allow_self_link, no_check_redundant_translit)
	if type(data) ~= "table" then
		error("The first argument to the function full_link must be a table. "
			.. "See Module:links/documentation for more information.")
	end
	
	local multiparams = {"term", "alt", "sc", "tr", "ts"}
	local terms = {true}
	
	-- Generate multiple forms if applicable.
	for i = 1, 2 do
		if (type(data[multiparams[i]]) == "string" and data[multiparams[i]]:find("//")) then
			local u = uchar
			data[multiparams[i]] = data[multiparams[i]]
				:gsub("\\\\//", u(0xE000) .. "//")
				:gsub("\\//", u(0xE001))
			data[multiparams[i]] = mw.text.split(data[multiparams[i]], "//") or {}
			for j, subparam in ipairs(data[multiparams[i]]) do
				data[multiparams[i]][j] = subparam
					:gsub(u(0xE000), "\\")
					:gsub(u(0xE001), "//")
				if subparam == "" then data[multiparams[i]][j] = nil end
			end
		elseif type(data[multiparams[i]]) == "string" and not (type(data.term) == "string" and data.term:find("//")) then
			data[multiparams[i]] = data.lang:generateForms(data[multiparams[i]])
		elseif type(data[multiparams[i]]) == "table" and #data[multiparams[i]] == 1 then
			data[multiparams[i]] = data.lang:generateForms(data[multiparams[i]][1])
		end
	end
	
	for i, multiparam in ipairs(multiparams) do
		data[multiparam] = data[multiparam] or {}
		if type(data[multiparam]) == "string" then
			data[multiparam] = {data[multiparam]}
		elseif data[multiparam] and data[multiparam]._type == "script object" then
			data[multiparam] = {data[multiparam]}
		end
		for i, subparam in pairs(data[multiparam]) do
			terms[i] = true
		end
	end
	
	-- Create the link
	local output = {}
	local categories = {}
	local link = ""
	local annotations
	
	local phonetic_extraction = mw.loadData("Module:links/data").phonetic_extraction
	
	for i in ipairs(terms) do
		-- Is there any text to show?
		if (data.term[i] or data.alt[i]) then
			-- Try to detect the script if it was not provided
			local best = data.lang:findBestScript(data.alt[i] or data.term[i])
			if not data.sc[i] then
				data.sc[i] = best
			else
				-- Track uses of sc parameter
				track("sc")

				if data.sc[i]:getCode() == best:getCode() then
					track("sc/redundant", data.sc[i]:getCode())
				else
					track("sc/needed", data.sc[i]:getCode())
				end
			end
			
			-- If using a discouraged character sequence, add to maintenance category
			if data.sc[i]:hasNormalizationFixes() == true then
				if (data.term[i] and data.sc[i]:fixDiscouragedSequences(toNFC(data.term[i])) ~= toNFC(data.term[i])) or (data.alt[i] and data.sc[i]:fixDiscouragedSequences(toNFC(data.alt[i])) ~= toNFC(data.alt[i])) then
					table.insert(categories, "[[Category:Pages using discouraged character sequences]]")
				end
			end

			local class = ""

			-- Encode certain characters to avoid various delimiter-related issues at various stages. We need to encode < and >
			-- because they end up forming part of CSS class names inside of <span ...> and will interfere with finding the end
			-- of the HTML tag. I first tried converting them to URL encoding, i.e. %3C and %3E; they then appear in the URL as
			-- %253C and %253E, which get mapped back to %3C and %3E when passed to [[Module:accel]]. But mapping them to &lt;
			-- and &gt; somehow works magically without any further work; they appear in the URL as < and >, and get passed to
			-- [[Module:accel]] as < and >. I have no idea who along the chain of calls is doing the encoding and decoding. If
			-- someone knows, please modify this comment appropriately!
			local encode_accel_char_map = {
				["%"] = ".",
				[" "] = "_",
				["<"] = "&lt;",
				[">"] = "&gt;",
			}
			local function encode_accel_param_chars(param)
				local retval = param:gsub("[% <>]", encode_accel_char_map) -- discard second return value
				return retval
			end

			local function encode_accel_param(prefix, param)
				if not param then
					return ""
				end
				if type(param) == "table" then
					local filled_params = {}
					-- There may be gaps in the sequence, especially for translit params.
					local maxindex = 0
					for k, v in pairs(param) do
						if type(k) == "number" and k > maxindex then
							maxindex = k
						end
					end
					for i=1,maxindex do
						filled_params[i] = param[i] or ""
					end
					-- [[Module:accel]] splits these up again.
					param = table.concat(filled_params, "*~!")
				end
				-- This is decoded again by [[WT:ACCEL]].
				return prefix .. encode_accel_param_chars(param)
			end

			if data.accel then
				local form = data.accel.form and encode_accel_param_chars(data.accel.form) .. "-form-of" or ""
				local gender = encode_accel_param("gender-", data.accel.gender)
				local pos = encode_accel_param("pos-", data.accel.pos)
				local translit = encode_accel_param("transliteration-",
					data.accel.translit or (data.tr[i] ~= "-" and data.tr[i] or nil))
				local target = encode_accel_param("target-", data.accel.target)
				local lemma = encode_accel_param("origin-", data.accel.lemma)
				local lemma_translit = encode_accel_param("origin_transliteration-", data.accel.lemma_translit)
				local no_store = data.accel.no_store and "form-of-nostore" or ""

				local accel =
					form .. " " ..
					gender .. " " ..
					pos .. " " ..
					translit .. " " ..
					target .. " " ..
					lemma .. " " ..
					lemma_translit .. " " ..
					no_store .. " "

				class = "form-of lang-" .. data.lang:getCode() .. " " .. accel
			end
			
			-- Only make a link if the term has been given, otherwise just show the alt text without a link
			local term_data = {term = data.term[i], alt = data.alt[i], lang = data.lang, sc = data.sc[i], id = data.id, genders = data.genders, tr = data.tr[i], ts = data.ts[i], gloss = data.gloss, pos = data.pos, lit = data.lit, accel = data.accel, interwiki = data.interwiki}
			
			link = require("Module:script utilities").tag_text(
				data.term[i] and export.language_link(term_data, allow_self_link)
				or data.alt[i], data.lang, data.sc[i], face, class)
		else
			--[[	No term to show.
					Is there at least a transliteration we can work from?	]]
			link = require("Module:script utilities").request_script(data.lang, data.sc[i])

			if link == "" or not data.tr[i] or data.tr[i] == "-" then
				-- No link to show, and no transliteration either. Show a term request.
				local category = ""

				if mw.title.getCurrentTitle().nsText ~= "Template" then
					table_insert(categories, "[[Category:" .. data.lang:getCanonicalName() .. " term requests]]")
				end
				
				link = "<small>[ord?]</small>"
			end
		end
		table_insert(output, link)
		if i < #terms then table_insert(output, "<span class=\"Zsym mention\" style=\"font-size:100%;\">/</span>") end
	end

	-- TODO: Currently only handles the first transliteration, pending consensus on how to handle multiple translits for multiple forms, as this is not always desirable (e.g. traditional/simplified Chinese).
	if data.tr[1] == "" or data.tr[1] == "-" then
		data.tr[1] = nil

	elseif phonetic_extraction[data.lang:getCode()] then
		local m_phonetic = require(phonetic_extraction[data.lang:getCode()])
		data.tr[1] = data.tr[1] or m_phonetic.getTranslit(export.remove_links(data.alt[1] or data.term[1]))

	elseif (data.term[1] or data.alt[1]) and data.sc[1]:isTransliterated() then
		-- Track whenever there is manual translit. The categories below like 'terms with redundant transliterations'
		-- aren't sufficient because they only work with reference to automatic translit and won't operate at all in
		-- languages without any automatic translit, like Persian and Hebrew.
		if data.tr[1] then
			track("manual-tr", data.lang:getCode())
		end

		-- Try to generate a transliteration, unless transliteration has been supplied and no_check_redundant_translit is
		-- given. (Checking for redundant transliteration can use up significant amounts of memory so we don't want to do
		-- it if memory is tight. `no_check_redundant_translit` is currently set when called ultimately from
		-- {{multitrans|...|no-check-redundant-translit=1}}.)
		if not (data.tr[1] and no_check_redundant_translit) then
			local text = data.alt[1] or data.term[1]
			if not data.lang:link_tr() then
				text = export.remove_links(text, true)
			end
			
			local automated_tr, tr_fail, tr_categories = data.lang:transliterate(text, data.sc[1])
			
			if automated_tr or tr_fail then
				local manual_tr = data.tr[1]

				if manual_tr then
					if (export.remove_links(manual_tr) == export.remove_links(automated_tr)) and (not tr_fail) then
						table_insert(categories,
							"[[Category:Terms with redundant transliterations]]"
									.. "[[Category:Terms with redundant transliterations/" .. data.lang:getCode() .. "]]")
					elseif not tr_fail then
						-- Prevents Arabic root categories from flooding the tracking categories.
						if mw.title.getCurrentTitle().nsText ~= "Category" then
							table_insert(categories,
								"[[Category:Terms with manual transliterations different from the automated ones]]"
										.. "[[Category:Terms with manual transliterations different from the automated ones/" .. data.lang:getCode() .. "]]")
						end
					end
				end

				if (not manual_tr) or data.lang:overrideManualTranslit() then
					data.tr[1] = automated_tr
					for _, category in ipairs(tr_categories) do
						table_insert(categories, category)
					end
				end
			end
		end
	end
	
	-- Link to the transliteration entry for languages that require this
	if data.tr[1] and data.lang:link_tr() and not (data.tr[1]:match("%[%[(.-)%]%]") or tr_fail) then
		data.tr[1] = export.language_link{lang = data.lang, term = data.tr[1]}
	elseif data.tr[1] and not data.lang:link_tr() and not tr_fail then
		-- Remove the pseudo-HTML tags added by remove_links.
		data.tr[1] = data.tr[1]:gsub("</?link>", "")
	end
	if data.tr[1] and data.tr[1]:gsub("[%s%p]", "") == "" then data.tr[1] = nil end
	
	table_insert(output, export.format_link_annotations(data, face))

	return table_concat(output) .. table_concat(categories)
end


-- Strips links: deletes category links, the targets of piped links, and any double square brackets involved in links (other than file links, which are untouched). If `tag` is set, then any links removed will be given pseudo-HTML tags, which allow the substitution functions in [[Module:languages]] to properly subdivide the text in order to reduce the chance of substitution failures in modules which scrape pages like [[Module:zh-translit]].
-- FIXME: This is quite hacky. We probably want this to be integrated into [[Module:languages]], but we can't do that until we know that nothing is pushing pipe linked transliterations through it for languages which don't have link_tr set.
function export.remove_links(text, tag)
	if type(text) == "table" then
		text = text.args[1]
	end

	if not text or text == "" then
		return ""
	end
	
	text = text
		:gsub("%[%[", "\1")
		:gsub("%]%]", "\2")

	-- Parse internal links for the display text.
	text = text:gsub("(\1)([^\1\2]-)(\2)",
		function(c1, c2, c3)
			-- If <>[]{} are in the target or there's only a pipe, then this is an invalid link, so should be returned in full.
			if c2:match("^[^|]*[<>%[%]{}][^|]*|") or c2:match("^[^|]*[<>%[%]{}][^|]*$") or c2 == "|" then
				return c1 .. c2 .. c3
			end
			-- Don't remove files.
			for _, falsePositive in ipairs({"File", "Image"}) do
				if c2:match("^" .. falsePositive .. ":") then return c1 .. c2 .. c3 end
			end
			-- Remove categories completely.
			if c2:match("^Category:") then return "" end
			-- In piped links, remove all text before the pipe, unless it's the final character (i.e. the pipe trick), in which case just remove the pipe.
			c2 = c2:match("^[^|]*|(.+)") or c2:match("([^|]+)|$") or c2
			if tag then
				return "<link>" .. c2 .. "</link>"
			else
				return c2
			end
		end)
		
	text = text
		:gsub("\1", "[[")
		:gsub("\2", "]]")

	return text
end

--[=[
This decodes old section encodings.
For example, Norwegian_Bokm.C3.A5l → Norwegian_Bokmål.
It isn't picky about whether the section encodings represent the UTF-8 encoding
of a real Unicode character, so it will mangle section names that contain
a period followed by two uppercase hex characters. At least such section names
are probably pretty rare.

Wiktionary adds an additional id="" attribute for sections
using a legacy encoding, if it is different from the modern minimally modified attribute.
It is like percent encoding (URI or URL encoding) except with "." instead of "%".
See [[mw:Manual:$wgFragmentMode]] and the code that does the encoding at
https://gerrit.wikimedia.org/r/plugins/gitiles/mediawiki/core/+/7bf779524ab1fd8e1d74f79ea4840564d48eea4d/includes/parser/Sanitizer.php#893
]=]

-- The character class %x should not be used, as it includes the characters a-f,
-- which do not occur in these anchor encodings.
local capitalHex = "[0-9A-F]"

local function decodeAnchor(anchor)
	return (anchor:gsub("%.(" .. capitalHex .. capitalHex .. ")",
		function(hexByte)
			return string.char(tonumber(hexByte, 16))
		end))
end

function export.section_link(link)
	if type(link) ~= "string" then
		error("The first argument to section_link was a " .. type(link) .. ", but it should be a string.")
	end

	link = link:gsub("_", " ")

	local numberSigns = select(2, link:gsub("#", ""))

	if numberSigns > 1 then
		error("The section link should only contain one number sign (#).")
	end

	link = mw.uri.decode(link, "WIKI")
	local page, section = link:match("^([^#]*)#(.+)$")
	if page == "" then
		page = nil
	end

	if section then
		section = decodeAnchor(section)

		-- URI-encode (percent-encode) section to allow square brackets and
		-- other dodgy characters in section name.
		-- If not percent-encoded, they prevent the parser from creating a link.
		-- Decode percent-encoding in the displayed text
		if page then
			return "[[" .. page .. "#" .. mw.uri.encode(section, "WIKI")
				.. "|" .. page .. " §&nbsp;" .. section .. "]]"
		else
			return "[[#" .. mw.uri.encode(section, "WIKI")
				.. "|§&nbsp;" .. section .. "]]"
		end
	else
		error("The function “section_link” could not find a number sign marking a section name.")
	end
end

return export