123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181 |
- /**
- * @typedef {import('micromark-util-types').Construct} Construct
- * @typedef {import('micromark-util-types').Resolver} Resolver
- * @typedef {import('micromark-util-types').State} State
- * @typedef {import('micromark-util-types').Token} Token
- * @typedef {import('micromark-util-types').TokenizeContext} TokenizeContext
- * @typedef {import('micromark-util-types').Tokenizer} Tokenizer
- */
- import {factorySpace} from 'micromark-factory-space'
- import {markdownLineEnding} from 'micromark-util-character'
- import {subtokenize} from 'micromark-util-subtokenize'
- import {codes, constants, types} from 'micromark-util-symbol'
- import {ok as assert} from 'devlop'
- /**
- * No name because it must not be turned off.
- * @type {Construct}
- */
- export const content = {tokenize: tokenizeContent, resolve: resolveContent}
- /** @type {Construct} */
- const continuationConstruct = {tokenize: tokenizeContinuation, partial: true}
- /**
- * Content is transparent: it’s parsed right now. That way, definitions are also
- * parsed right now: before text in paragraphs (specifically, media) are parsed.
- *
- * @type {Resolver}
- */
- function resolveContent(events) {
- subtokenize(events)
- return events
- }
- /**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
- function tokenizeContent(effects, ok) {
- /** @type {Token | undefined} */
- let previous
- return chunkStart
- /**
- * Before a content chunk.
- *
- * ```markdown
- * > | abc
- * ^
- * ```
- *
- * @type {State}
- */
- function chunkStart(code) {
- assert(
- code !== codes.eof && !markdownLineEnding(code),
- 'expected no eof or eol'
- )
- effects.enter(types.content)
- previous = effects.enter(types.chunkContent, {
- contentType: constants.contentTypeContent
- })
- return chunkInside(code)
- }
- /**
- * In a content chunk.
- *
- * ```markdown
- * > | abc
- * ^^^
- * ```
- *
- * @type {State}
- */
- function chunkInside(code) {
- if (code === codes.eof) {
- return contentEnd(code)
- }
- // To do: in `markdown-rs`, each line is parsed on its own, and everything
- // is stitched together resolving.
- if (markdownLineEnding(code)) {
- return effects.check(
- continuationConstruct,
- contentContinue,
- contentEnd
- )(code)
- }
- // Data.
- effects.consume(code)
- return chunkInside
- }
- /**
- *
- *
- * @type {State}
- */
- function contentEnd(code) {
- effects.exit(types.chunkContent)
- effects.exit(types.content)
- return ok(code)
- }
- /**
- *
- *
- * @type {State}
- */
- function contentContinue(code) {
- assert(markdownLineEnding(code), 'expected eol')
- effects.consume(code)
- effects.exit(types.chunkContent)
- assert(previous, 'expected previous token')
- previous.next = effects.enter(types.chunkContent, {
- contentType: constants.contentTypeContent,
- previous
- })
- previous = previous.next
- return chunkInside
- }
- }
- /**
- * @this {TokenizeContext}
- * @type {Tokenizer}
- */
- function tokenizeContinuation(effects, ok, nok) {
- const self = this
- return startLookahead
- /**
- *
- *
- * @type {State}
- */
- function startLookahead(code) {
- assert(markdownLineEnding(code), 'expected a line ending')
- effects.exit(types.chunkContent)
- effects.enter(types.lineEnding)
- effects.consume(code)
- effects.exit(types.lineEnding)
- return factorySpace(effects, prefixed, types.linePrefix)
- }
- /**
- *
- *
- * @type {State}
- */
- function prefixed(code) {
- if (code === codes.eof || markdownLineEnding(code)) {
- return nok(code)
- }
- // Always populated by defaults.
- assert(
- self.parser.constructs.disable.null,
- 'expected `disable.null` to be populated'
- )
- const tail = self.events[self.events.length - 1]
- if (
- !self.parser.constructs.disable.null.includes('codeIndented') &&
- tail &&
- tail[1].type === types.linePrefix &&
- tail[2].sliceSerialize(tail[1], true).length >= constants.tabSize
- ) {
- return ok(code)
- }
- return effects.interrupt(self.parser.constructs.flow, nok, ok)(code)
- }
- }
|