mirror of
https://github.com/Sevichecc/Urara-Blog.git
synced 2025-05-05 17:39:30 +08:00
234 lines
5.9 KiB
Text
234 lines
5.9 KiB
Text
/**
|
|
* @typedef {import('micromark-util-types').Construct} Construct
|
|
* @typedef {import('micromark-util-types').Tokenizer} Tokenizer
|
|
* @typedef {import('micromark-util-types').State} State
|
|
* @typedef {import('micromark-util-types').Code} Code
|
|
*/
|
|
|
|
import {ok as assert} from 'uvu/assert'
|
|
import {factorySpace} from 'micromark-factory-space'
|
|
import {
|
|
markdownLineEnding,
|
|
markdownLineEndingOrSpace
|
|
} from 'micromark-util-character'
|
|
import {codes} from 'micromark-util-symbol/codes.js'
|
|
import {constants} from 'micromark-util-symbol/constants.js'
|
|
import {types} from 'micromark-util-symbol/types.js'
|
|
|
|
/** @type {Construct} */
|
|
export const codeFenced = {
|
|
name: 'codeFenced',
|
|
tokenize: tokenizeCodeFenced,
|
|
concrete: true
|
|
}
|
|
|
|
/** @type {Tokenizer} */
|
|
function tokenizeCodeFenced(effects, ok, nok) {
|
|
const self = this
|
|
/** @type {Construct} */
|
|
const closingFenceConstruct = {tokenize: tokenizeClosingFence, partial: true}
|
|
/** @type {Construct} */
|
|
const nonLazyLine = {tokenize: tokenizeNonLazyLine, partial: true}
|
|
const tail = this.events[this.events.length - 1]
|
|
const initialPrefix =
|
|
tail && tail[1].type === types.linePrefix
|
|
? tail[2].sliceSerialize(tail[1], true).length
|
|
: 0
|
|
let sizeOpen = 0
|
|
/** @type {NonNullable<Code>} */
|
|
let marker
|
|
|
|
return start
|
|
|
|
/** @type {State} */
|
|
function start(code) {
|
|
assert(
|
|
code === codes.graveAccent || code === codes.tilde,
|
|
'expected `` ` `` or `~`'
|
|
)
|
|
effects.enter(types.codeFenced)
|
|
effects.enter(types.codeFencedFence)
|
|
effects.enter(types.codeFencedFenceSequence)
|
|
marker = code
|
|
return sequenceOpen(code)
|
|
}
|
|
|
|
/** @type {State} */
|
|
function sequenceOpen(code) {
|
|
if (code === marker) {
|
|
effects.consume(code)
|
|
sizeOpen++
|
|
return sequenceOpen
|
|
}
|
|
|
|
effects.exit(types.codeFencedFenceSequence)
|
|
return sizeOpen < constants.codeFencedSequenceSizeMin
|
|
? nok(code)
|
|
: factorySpace(effects, infoOpen, types.whitespace)(code)
|
|
}
|
|
|
|
/** @type {State} */
|
|
function infoOpen(code) {
|
|
if (code === codes.eof || markdownLineEnding(code)) {
|
|
return openAfter(code)
|
|
}
|
|
|
|
effects.enter(types.codeFencedFenceInfo)
|
|
effects.enter(types.chunkString, {contentType: constants.contentTypeString})
|
|
return info(code)
|
|
}
|
|
|
|
/** @type {State} */
|
|
function info(code) {
|
|
if (code === codes.eof || markdownLineEndingOrSpace(code)) {
|
|
effects.exit(types.chunkString)
|
|
effects.exit(types.codeFencedFenceInfo)
|
|
return factorySpace(effects, infoAfter, types.whitespace)(code)
|
|
}
|
|
|
|
if (code === codes.graveAccent && code === marker) return nok(code)
|
|
effects.consume(code)
|
|
return info
|
|
}
|
|
|
|
/** @type {State} */
|
|
function infoAfter(code) {
|
|
if (code === codes.eof || markdownLineEnding(code)) {
|
|
return openAfter(code)
|
|
}
|
|
|
|
effects.enter(types.codeFencedFenceMeta)
|
|
effects.enter(types.chunkString, {contentType: constants.contentTypeString})
|
|
return meta(code)
|
|
}
|
|
|
|
/** @type {State} */
|
|
function meta(code) {
|
|
if (code === codes.eof || markdownLineEnding(code)) {
|
|
effects.exit(types.chunkString)
|
|
effects.exit(types.codeFencedFenceMeta)
|
|
return openAfter(code)
|
|
}
|
|
|
|
if (code === codes.graveAccent && code === marker) return nok(code)
|
|
effects.consume(code)
|
|
return meta
|
|
}
|
|
|
|
/** @type {State} */
|
|
function openAfter(code) {
|
|
effects.exit(types.codeFencedFence)
|
|
return self.interrupt ? ok(code) : contentStart(code)
|
|
}
|
|
|
|
/** @type {State} */
|
|
function contentStart(code) {
|
|
if (code === codes.eof) {
|
|
return after(code)
|
|
}
|
|
|
|
if (markdownLineEnding(code)) {
|
|
return effects.attempt(
|
|
nonLazyLine,
|
|
effects.attempt(
|
|
closingFenceConstruct,
|
|
after,
|
|
initialPrefix
|
|
? factorySpace(
|
|
effects,
|
|
contentStart,
|
|
types.linePrefix,
|
|
initialPrefix + 1
|
|
)
|
|
: contentStart
|
|
),
|
|
after
|
|
)(code)
|
|
}
|
|
|
|
effects.enter(types.codeFlowValue)
|
|
return contentContinue(code)
|
|
}
|
|
|
|
/** @type {State} */
|
|
function contentContinue(code) {
|
|
if (code === codes.eof || markdownLineEnding(code)) {
|
|
effects.exit(types.codeFlowValue)
|
|
return contentStart(code)
|
|
}
|
|
|
|
effects.consume(code)
|
|
return contentContinue
|
|
}
|
|
|
|
/** @type {State} */
|
|
function after(code) {
|
|
effects.exit(types.codeFenced)
|
|
return ok(code)
|
|
}
|
|
|
|
/** @type {Tokenizer} */
|
|
function tokenizeNonLazyLine(effects, ok, nok) {
|
|
const self = this
|
|
|
|
return start
|
|
|
|
/** @type {State} */
|
|
function start(code) {
|
|
assert(markdownLineEnding(code), 'expected eol')
|
|
effects.enter(types.lineEnding)
|
|
effects.consume(code)
|
|
effects.exit(types.lineEnding)
|
|
return lineStart
|
|
}
|
|
|
|
/** @type {State} */
|
|
function lineStart(code) {
|
|
return self.parser.lazy[self.now().line] ? nok(code) : ok(code)
|
|
}
|
|
}
|
|
|
|
/** @type {Tokenizer} */
|
|
function tokenizeClosingFence(effects, ok, nok) {
|
|
let size = 0
|
|
|
|
return factorySpace(
|
|
effects,
|
|
closingSequenceStart,
|
|
types.linePrefix,
|
|
this.parser.constructs.disable.null.includes('codeIndented')
|
|
? undefined
|
|
: constants.tabSize
|
|
)
|
|
|
|
/** @type {State} */
|
|
function closingSequenceStart(code) {
|
|
effects.enter(types.codeFencedFence)
|
|
effects.enter(types.codeFencedFenceSequence)
|
|
return closingSequence(code)
|
|
}
|
|
|
|
/** @type {State} */
|
|
function closingSequence(code) {
|
|
if (code === marker) {
|
|
effects.consume(code)
|
|
size++
|
|
return closingSequence
|
|
}
|
|
|
|
if (size < sizeOpen) return nok(code)
|
|
effects.exit(types.codeFencedFenceSequence)
|
|
return factorySpace(effects, closingSequenceEnd, types.whitespace)(code)
|
|
}
|
|
|
|
/** @type {State} */
|
|
function closingSequenceEnd(code) {
|
|
if (code === codes.eof || markdownLineEnding(code)) {
|
|
effects.exit(types.codeFencedFence)
|
|
return ok(code)
|
|
}
|
|
|
|
return nok(code)
|
|
}
|
|
}
|
|
}
|