Compare commits
No commits in common. "c4a8ebebf38a766bfc41f9b303d3bb090140ea3c" and "86d5c87e9a20a05583d831bfef65fe6bcf6413bc" have entirely different histories.
c4a8ebebf3
...
86d5c87e9a
|
@ -6,8 +6,8 @@ use nom::character::complete::{
|
||||||
alpha1, alphanumeric1, anychar, char as one_char, line_ending, not_line_ending, one_of,
|
alpha1, alphanumeric1, anychar, char as one_char, line_ending, not_line_ending, one_of,
|
||||||
satisfy, space1, tab,
|
satisfy, space1, tab,
|
||||||
};
|
};
|
||||||
use nom::combinator::{eof, fail, map, not, opt, peek, recognize};
|
use nom::combinator::{eof, fail, map, not, opt, recognize};
|
||||||
use nom::error::ErrorKind;
|
use nom::error::{ErrorKind, ParseError};
|
||||||
use nom::multi::{many0, many0_count, many1, many1_count, many_till, separated_list1};
|
use nom::multi::{many0, many0_count, many1, many1_count, many_till, separated_list1};
|
||||||
use nom::sequence::tuple;
|
use nom::sequence::tuple;
|
||||||
use nom::{IResult, Offset, Parser, Slice};
|
use nom::{IResult, Offset, Parser, Slice};
|
||||||
|
@ -273,11 +273,8 @@ fn collect_char_sequence(
|
||||||
}
|
}
|
||||||
|
|
||||||
#[inline]
|
#[inline]
|
||||||
fn space1_unicode(input: Span) -> IResult<Span, Span> {
|
fn alpha1_unicode(input: Span) -> IResult<Span, Span> {
|
||||||
recognize(many1_count(tuple((
|
recognize(many1_count(satisfy(char::is_alphanumeric)))(input)
|
||||||
not(line_ending),
|
|
||||||
satisfy(char::is_whitespace),
|
|
||||||
))))(input)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[inline]
|
#[inline]
|
||||||
|
@ -773,14 +770,6 @@ impl Context {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if let FlankingRule::Strict = opening_rule {
|
|
||||||
let (input, pre) =
|
|
||||||
opt(recognize(tuple((alphanumeric1_unicode, &opening_tag))))(input)?;
|
|
||||||
if let Some(pre_text) = pre {
|
|
||||||
return Ok((input, Token::PlainText(pre_text.into_fragment().into())));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let begin = input;
|
let begin = input;
|
||||||
let (post_open, _) = opening_tag(input)?;
|
let (post_open, _) = opening_tag(input)?;
|
||||||
|
|
||||||
|
@ -843,12 +832,8 @@ impl Context {
|
||||||
true
|
true
|
||||||
};
|
};
|
||||||
|
|
||||||
let (input, alphanum) = opt(peek(alphanumeric1_unicode))(input)?;
|
// TODO: Unfinished flanking rules
|
||||||
let correct_right_outer =
|
let correct_flanking = correct_left_flanking && correct_right_flanking;
|
||||||
alphanum.is_none() || !matches!(closing_rule, FlankingRule::Strict);
|
|
||||||
|
|
||||||
let correct_flanking =
|
|
||||||
correct_left_flanking && correct_right_flanking && correct_right_outer;
|
|
||||||
|
|
||||||
if !correct_flanking {
|
if !correct_flanking {
|
||||||
return Ok((
|
return Ok((
|
||||||
|
@ -1308,12 +1293,7 @@ impl Context {
|
||||||
}
|
}
|
||||||
|
|
||||||
fn tag_hashtag<'a>(&self, input: Span<'a>) -> IResult<Span<'a>, Token> {
|
fn tag_hashtag<'a>(&self, input: Span<'a>) -> IResult<Span<'a>, Token> {
|
||||||
let (input, maybe_preceded) =
|
// TODO: Skip when preceded by alphanumerics
|
||||||
opt(recognize(tuple((alphanumeric1_unicode, tag("#")))))(input)?;
|
|
||||||
|
|
||||||
if let Some(preceded) = maybe_preceded {
|
|
||||||
return Ok((input, Token::PlainText(preceded.into_fragment().into())));
|
|
||||||
}
|
|
||||||
|
|
||||||
let (input, _) = tag("#")(input)?;
|
let (input, _) = tag("#")(input)?;
|
||||||
|
|
||||||
|
@ -1367,8 +1347,8 @@ impl Context {
|
||||||
tag(")"),
|
tag(")"),
|
||||||
))),
|
))),
|
||||||
recognize(tuple((
|
recognize(tuple((
|
||||||
not(space1_unicode),
|
not(space1),
|
||||||
not(line_ending),
|
not_line_ending,
|
||||||
not(one_of(".,:;!?#?/[]【】()「」()<>")),
|
not(one_of(".,:;!?#?/[]【】()「」()<>")),
|
||||||
anychar,
|
anychar,
|
||||||
))),
|
))),
|
||||||
|
@ -1521,20 +1501,6 @@ mod test {
|
||||||
Token::PlainText("* italic *".into())
|
Token::PlainText("* italic *".into())
|
||||||
);
|
);
|
||||||
|
|
||||||
assert_eq!(
|
|
||||||
parse_full("snake_case_variable"),
|
|
||||||
Token::PlainText("snake_case_variable".into())
|
|
||||||
);
|
|
||||||
|
|
||||||
assert_eq!(
|
|
||||||
parse_full("intra*word*italic"),
|
|
||||||
Token::Sequence(vec![
|
|
||||||
Token::PlainText("intra".into()),
|
|
||||||
Token::Italic(Box::new(Token::PlainText("word".into()))),
|
|
||||||
Token::PlainText("italic".into())
|
|
||||||
])
|
|
||||||
);
|
|
||||||
|
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
parse_full(r#"_ italic *"#),
|
parse_full(r#"_ italic *"#),
|
||||||
Token::PlainText("_ italic *".into())
|
Token::PlainText("_ italic *".into())
|
||||||
|
@ -1733,20 +1699,6 @@ text</center>"#
|
||||||
])
|
])
|
||||||
);
|
);
|
||||||
|
|
||||||
assert_eq!(
|
|
||||||
parse_full("test #hashtag tail"),
|
|
||||||
Token::Sequence(vec![
|
|
||||||
Token::PlainText("test ".into()),
|
|
||||||
Token::Hashtag("hashtag".into()),
|
|
||||||
Token::PlainText(" tail".into())
|
|
||||||
])
|
|
||||||
);
|
|
||||||
|
|
||||||
assert_eq!(
|
|
||||||
parse_full("not#hashtag tail"),
|
|
||||||
Token::PlainText("not#hashtag tail".into())
|
|
||||||
);
|
|
||||||
|
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
parse_full("<https://example.com>"),
|
parse_full("<https://example.com>"),
|
||||||
Token::UrlNoEmbed("https://example.com".into())
|
Token::UrlNoEmbed("https://example.com".into())
|
||||||
|
|
Loading…
Reference in New Issue