103 lines
3.0 KiB
Python
103 lines
3.0 KiB
Python
|
# Process *this* and _that_
|
||
|
#
|
||
|
from __future__ import annotations
|
||
|
|
||
|
from .state_inline import Delimiter, StateInline
|
||
|
|
||
|
|
||
|
def tokenize(state: StateInline, silent: bool) -> bool:
|
||
|
"""Insert each marker as a separate text token, and add it to delimiter list"""
|
||
|
start = state.pos
|
||
|
marker = state.src[start]
|
||
|
|
||
|
if silent:
|
||
|
return False
|
||
|
|
||
|
if marker not in ("_", "*"):
|
||
|
return False
|
||
|
|
||
|
scanned = state.scanDelims(state.pos, marker == "*")
|
||
|
|
||
|
for _ in range(scanned.length):
|
||
|
token = state.push("text", "", 0)
|
||
|
token.content = marker
|
||
|
state.delimiters.append(
|
||
|
Delimiter(
|
||
|
marker=ord(marker),
|
||
|
length=scanned.length,
|
||
|
token=len(state.tokens) - 1,
|
||
|
end=-1,
|
||
|
open=scanned.can_open,
|
||
|
close=scanned.can_close,
|
||
|
)
|
||
|
)
|
||
|
|
||
|
state.pos += scanned.length
|
||
|
|
||
|
return True
|
||
|
|
||
|
|
||
|
def _postProcess(state: StateInline, delimiters: list[Delimiter]) -> None:
|
||
|
i = len(delimiters) - 1
|
||
|
while i >= 0:
|
||
|
startDelim = delimiters[i]
|
||
|
|
||
|
# /* _ */ /* * */
|
||
|
if startDelim.marker != 0x5F and startDelim.marker != 0x2A:
|
||
|
i -= 1
|
||
|
continue
|
||
|
|
||
|
# Process only opening markers
|
||
|
if startDelim.end == -1:
|
||
|
i -= 1
|
||
|
continue
|
||
|
|
||
|
endDelim = delimiters[startDelim.end]
|
||
|
|
||
|
# If the previous delimiter has the same marker and is adjacent to this one,
|
||
|
# merge those into one strong delimiter.
|
||
|
#
|
||
|
# `<em><em>whatever</em></em>` -> `<strong>whatever</strong>`
|
||
|
#
|
||
|
isStrong = (
|
||
|
i > 0
|
||
|
and delimiters[i - 1].end == startDelim.end + 1
|
||
|
# check that first two markers match and adjacent
|
||
|
and delimiters[i - 1].marker == startDelim.marker
|
||
|
and delimiters[i - 1].token == startDelim.token - 1
|
||
|
# check that last two markers are adjacent (we can safely assume they match)
|
||
|
and delimiters[startDelim.end + 1].token == endDelim.token + 1
|
||
|
)
|
||
|
|
||
|
ch = chr(startDelim.marker)
|
||
|
|
||
|
token = state.tokens[startDelim.token]
|
||
|
token.type = "strong_open" if isStrong else "em_open"
|
||
|
token.tag = "strong" if isStrong else "em"
|
||
|
token.nesting = 1
|
||
|
token.markup = ch + ch if isStrong else ch
|
||
|
token.content = ""
|
||
|
|
||
|
token = state.tokens[endDelim.token]
|
||
|
token.type = "strong_close" if isStrong else "em_close"
|
||
|
token.tag = "strong" if isStrong else "em"
|
||
|
token.nesting = -1
|
||
|
token.markup = ch + ch if isStrong else ch
|
||
|
token.content = ""
|
||
|
|
||
|
if isStrong:
|
||
|
state.tokens[delimiters[i - 1].token].content = ""
|
||
|
state.tokens[delimiters[startDelim.end + 1].token].content = ""
|
||
|
i -= 1
|
||
|
|
||
|
i -= 1
|
||
|
|
||
|
|
||
|
def postProcess(state: StateInline) -> None:
|
||
|
"""Walk through delimiter list and replace text tokens with tags."""
|
||
|
_postProcess(state, state.delimiters)
|
||
|
|
||
|
for token in state.tokens_meta:
|
||
|
if token and "delimiters" in token:
|
||
|
_postProcess(state, token["delimiters"])
|