1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495 |
- package parser
- import (
- "testing"
- "github.com/stretchr/testify/require"
- "github.com/usememos/memos/plugin/gomark/parser/tokenizer"
- )
- func TestHeadingParser(t *testing.T) {
- tests := []struct {
- text string
- heading *HeadingParser
- }{
- {
- text: "*Hello world",
- heading: nil,
- },
- {
- text: "## Hello World",
- heading: &HeadingParser{
- Level: 2,
- ContentTokens: []*tokenizer.Token{
- {
- Type: tokenizer.Text,
- Value: "Hello",
- },
- {
- Type: tokenizer.Space,
- Value: " ",
- },
- {
- Type: tokenizer.Text,
- Value: "World",
- },
- },
- },
- },
- {
- text: "# # Hello World",
- heading: &HeadingParser{
- Level: 1,
- ContentTokens: []*tokenizer.Token{
- {
- Type: tokenizer.Hash,
- Value: "#",
- },
- {
- Type: tokenizer.Space,
- Value: " ",
- },
- {
- Type: tokenizer.Text,
- Value: "Hello",
- },
- {
- Type: tokenizer.Space,
- Value: " ",
- },
- {
- Type: tokenizer.Text,
- Value: "World",
- },
- },
- },
- },
- {
- text: " # 123123 Hello World",
- heading: nil,
- },
- {
- text: `# 123
- Hello World`,
- heading: &HeadingParser{
- Level: 1,
- ContentTokens: []*tokenizer.Token{
- {
- Type: tokenizer.Text,
- Value: "123",
- },
- {
- Type: tokenizer.Space,
- Value: " ",
- },
- },
- },
- },
- }
- for _, test := range tests {
- tokens := tokenizer.Tokenize(test.text)
- heading := NewHeadingParser()
- require.Equal(t, test.heading, heading.Match(tokens))
- }
- }
|