2
0
mirror of https://github.com/tenrok/BBob.git synced 2026-05-15 11:59:37 +03:00

add parser tests

This commit is contained in:
Nikolay Kostyurin
2018-06-12 18:07:13 +02:00
parent cacfda9685
commit 77ddfca271
4 changed files with 47 additions and 21 deletions
+1 -1
View File
@@ -17,7 +17,7 @@ const Tokenizer = require('./Tokenizer');
const TokenChar = Tokenizer.CHAR;
const getChar = Tokenizer.getChar;
const createTagNode = (name, attrs = {}, content = []) => ({ tag: name, attrs, content });
const createTagNode = (tag, attrs = {}, content = []) => ({ tag, attrs, content });
/**
*
+27 -3
View File
@@ -1,11 +1,35 @@
const Parser = require('./Parser');
const TOKEN = require('./token');
const parse = input => (new Parser(input).parse());
describe('Parser', () => {
test('parse paired tags tokens', () => {
const parser = new Parser([
[TOKEN.TYPE_TAG, 'ch'],
[TOKEN.TYPE_TAG, '/ch'],
const input = [
[TOKEN.TYPE_TAG, 'best'],
[TOKEN.TYPE_ATTR_NAME, 'name'],
[TOKEN.TYPE_ATTR_VALUE, 'value'],
[TOKEN.TYPE_WORD, 'Foo'],
[TOKEN.TYPE_SPACE, ' '],
[TOKEN.TYPE_WORD, 'Bar'],
[TOKEN.TYPE_TAG, '/best'],
];
const ast = parse(input);
expect(ast).toBeInstanceOf(Array);
expect(ast).toEqual([
{
tag: 'best',
attrs: {
name: 'value',
},
content: [
'Foo',
' ',
'Bar',
],
},
]);
});
});
+8 -6
View File
@@ -2,10 +2,12 @@ const Tokenizer = require('./Tokenizer');
const TYPE = Tokenizer.TYPE;
const tokenize = input => (new Tokenizer(input).tokenize());
describe('Tokenizer', () => {
test('tokenize single tag', () => {
const input = '[SingleTag]';
const tokens = new Tokenizer(input).tokenize();
const tokens = tokenize(input);
expect(tokens).toBeInstanceOf(Array);
expect(tokens).toEqual([
@@ -15,7 +17,7 @@ describe('Tokenizer', () => {
test('tokenize single tag with spaces', () => {
const input = '[Single Tag]';
const tokens = new Tokenizer(input).tokenize();
const tokens = tokenize(input);
expect(tokens).toBeInstanceOf(Array);
expect(tokens).toEqual([
@@ -25,7 +27,7 @@ describe('Tokenizer', () => {
test('tokenize tag as param', () => {
const input = '[color="#ff0000"]Text[/color]';
const tokens = new Tokenizer(input).tokenize();
const tokens = tokenize(input);
expect(tokens).toBeInstanceOf(Array);
expect(tokens).toEqual([
@@ -38,7 +40,7 @@ describe('Tokenizer', () => {
test('tokenize tag param without quotemarks', () => {
const input = '[style color=#ff0000]Text[/style]';
const tokens = new Tokenizer(input).tokenize();
const tokens = tokenize(input);
expect(tokens).toBeInstanceOf(Array);
expect(tokens).toEqual([
@@ -57,7 +59,7 @@ describe('Tokenizer', () => {
[*] Item 3.
[/list]`;
const tokens = new Tokenizer(input).tokenize();
const tokens = tokenize(input);
expect(tokens).toBeInstanceOf(Array);
expect(tokens).toEqual([
@@ -136,7 +138,7 @@ describe('Tokenizer', () => {
];
inputs.forEach((input, idx) => {
const tokens = new Tokenizer(input).tokenize();
const tokens = tokenize(input);
expect(tokens).toBeInstanceOf(Array);
expect(tokens).toEqual(asserts[idx]);
+11 -11
View File
@@ -11,14 +11,14 @@ const TOKEN_TYPE_SPACE = 'space';
const TOKEN_TYPE_NEW_LINE = 'new-line';
module.exports = {
TYPE_ID: TOKEN_TYPE_ID,
VALUE_ID: TOKEN_VALUE_ID,
LINE_ID: TOKEN_LINE_ID,
COLUMN_ID: TOKEN_COLUMN_ID,
TYPE_WORD: TOKEN_TYPE_WORD,
TYPE_TAG: TOKEN_TYPE_TAG,
TYPE_ATTR_NAME: TOKEN_TYPE_ATTR_NAME,
TYPE_ATTR_VALUE: TOKEN_TYPE_ATTR_VALUE,
TYPE_SPACE: TOKEN_TYPE_SPACE,
TYPE_NEW_LINE: TOKEN_TYPE_NEW_LINE
};
TYPE_ID: TOKEN_TYPE_ID,
VALUE_ID: TOKEN_VALUE_ID,
LINE_ID: TOKEN_LINE_ID,
COLUMN_ID: TOKEN_COLUMN_ID,
TYPE_WORD: TOKEN_TYPE_WORD,
TYPE_TAG: TOKEN_TYPE_TAG,
TYPE_ATTR_NAME: TOKEN_TYPE_ATTR_NAME,
TYPE_ATTR_VALUE: TOKEN_TYPE_ATTR_VALUE,
TYPE_SPACE: TOKEN_TYPE_SPACE,
TYPE_NEW_LINE: TOKEN_TYPE_NEW_LINE
};