MindMap/frontend/node_modules/mammoth/test/styles/parser/tokeniser.tests.js

105 lines
2.9 KiB
JavaScript

var hamjest = require("hamjest");
var assertThat = hamjest.assertThat;
var contains = hamjest.contains;
var hasProperties = hamjest.hasProperties;
var tokenise = require("../../../lib/styles/parser/tokeniser").tokenise;
var test = require("../../test")(module);
test("unknown tokens are tokenised", function() {
assertTokens("~", [isToken("unrecognisedCharacter", "~")]);
});
test("empty string is tokenised to end of file token", function() {
assertTokens("", []);
});
test("whitespace is tokenised", function() {
assertTokens(" \t\t ", [isToken("whitespace")]);
});
test("identifiers are tokenised", function() {
assertTokens("Overture", [isToken("identifier", "Overture")]);
});
test("integers are tokenised", function() {
assertTokens("123", [isToken("integer", "123")]);
});
test("strings are tokenised", function() {
assertTokens("'Tristan'", [isToken("string", "Tristan")]);
});
test("unterminated strings are tokenised", function() {
assertTokens("'Tristan", [isToken("unterminated-string", "Tristan")]);
});
test("arrows are tokenised", function() {
assertTokens("=>", [isToken("arrow")]);
});
test("classes are tokenised", function() {
assertTokens(".overture", [isToken("dot"), isToken("identifier", "overture")]);
});
test("colons are tokenised", function() {
assertTokens("::", [isToken("colon"), isToken("colon")]);
});
test("greater thans are tokenised", function() {
assertTokens(">>", [isToken("gt"), isToken("gt")]);
});
test("equals are tokenised", function() {
assertTokens("==", [isToken("equals"), isToken("equals")]);
});
test("startsWith symbols are tokenised", function() {
assertTokens("^=^=", [isToken("startsWith"), isToken("startsWith")]);
});
test("open parens are tokenised", function() {
assertTokens("((", [isToken("open-paren"), isToken("open-paren")]);
});
test("close parens are tokenised", function() {
assertTokens("))", [isToken("close-paren"), isToken("close-paren")]);
});
test("open square brackets are tokenised", function() {
assertTokens("[[", [isToken("open-square-bracket"), isToken("open-square-bracket")]);
});
test("close square brackets are tokenised", function() {
assertTokens("]]", [isToken("close-square-bracket"), isToken("close-square-bracket")]);
});
test("choices are tokenised", function() {
assertTokens("||", [isToken("choice"), isToken("choice")]);
});
test("can tokenise multiple tokens", function() {
assertTokens("The Magic Position", [
isToken("identifier", "The"),
isToken("whitespace"),
isToken("identifier", "Magic"),
isToken("whitespace"),
isToken("identifier", "Position")
]);
});
function assertTokens(input, expectedTokens) {
assertThat(
tokenise(input),
contains.apply(null, expectedTokens.concat([isToken("end", null)]))
);
}
function isToken(tokenType, value) {
return hasProperties({
name: tokenType,
value: value
});
}