mirror of
https://github.com/usememos/memos.git
synced 2024-12-22 18:51:31 +03:00
86 lines
1.4 KiB
Go
86 lines
1.4 KiB
Go
|
package parser
|
||
|
|
||
|
import (
|
||
|
"testing"
|
||
|
|
||
|
"github.com/stretchr/testify/require"
|
||
|
"github.com/usememos/memos/plugin/gomark/parser/tokenizer"
|
||
|
)
|
||
|
|
||
|
func TestParagraphParser(t *testing.T) {
|
||
|
tests := []struct {
|
||
|
text string
|
||
|
paragraph *ParagraphParser
|
||
|
}{
|
||
|
{
|
||
|
text: "",
|
||
|
paragraph: nil,
|
||
|
},
|
||
|
{
|
||
|
text: "Hello world!",
|
||
|
paragraph: &ParagraphParser{
|
||
|
ContentTokens: []*tokenizer.Token{
|
||
|
{
|
||
|
Type: tokenizer.Text,
|
||
|
Value: "Hello",
|
||
|
},
|
||
|
{
|
||
|
Type: tokenizer.Space,
|
||
|
Value: " ",
|
||
|
},
|
||
|
{
|
||
|
Type: tokenizer.Text,
|
||
|
Value: "world!",
|
||
|
},
|
||
|
},
|
||
|
},
|
||
|
},
|
||
|
{
|
||
|
text: `Hello
|
||
|
world!`,
|
||
|
paragraph: &ParagraphParser{
|
||
|
ContentTokens: []*tokenizer.Token{
|
||
|
{
|
||
|
Type: tokenizer.Text,
|
||
|
Value: "Hello",
|
||
|
},
|
||
|
{
|
||
|
Type: tokenizer.Space,
|
||
|
Value: " ",
|
||
|
},
|
||
|
},
|
||
|
},
|
||
|
},
|
||
|
{
|
||
|
text: `Hello \n
|
||
|
world!`,
|
||
|
paragraph: &ParagraphParser{
|
||
|
ContentTokens: []*tokenizer.Token{
|
||
|
{
|
||
|
Type: tokenizer.Text,
|
||
|
Value: "Hello",
|
||
|
},
|
||
|
{
|
||
|
Type: tokenizer.Space,
|
||
|
Value: " ",
|
||
|
},
|
||
|
{
|
||
|
Type: tokenizer.Text,
|
||
|
Value: `\n`,
|
||
|
},
|
||
|
{
|
||
|
Type: tokenizer.Space,
|
||
|
Value: " ",
|
||
|
},
|
||
|
},
|
||
|
},
|
||
|
},
|
||
|
}
|
||
|
|
||
|
for _, test := range tests {
|
||
|
tokens := tokenizer.Tokenize(test.text)
|
||
|
paragraph := NewParagraphParser()
|
||
|
require.Equal(t, test.paragraph, paragraph.Match(tokens))
|
||
|
}
|
||
|
}
|