1769 lines
52 KiB
JavaScript
1769 lines
52 KiB
JavaScript
|
"use strict";
|
|||
|
|
|||
|
Object.defineProperty(exports, "__esModule", {
|
|||
|
value: true
|
|||
|
});
|
|||
|
exports.parse = parse;
|
|||
|
|
|||
|
var _helperCodeFrame = require("@webassemblyjs/helper-code-frame");
|
|||
|
|
|||
|
var t = _interopRequireWildcard(require("@webassemblyjs/ast"));
|
|||
|
|
|||
|
var _numberLiterals = require("./number-literals");
|
|||
|
|
|||
|
var _stringLiterals = require("./string-literals");
|
|||
|
|
|||
|
var _tokenizer = require("./tokenizer");
|
|||
|
|
|||
|
function _interopRequireWildcard(obj) { if (obj && obj.__esModule) { return obj; } else { var newObj = {}; if (obj != null) { for (var key in obj) { if (Object.prototype.hasOwnProperty.call(obj, key)) { var desc = Object.defineProperty && Object.getOwnPropertyDescriptor ? Object.getOwnPropertyDescriptor(obj, key) : {}; if (desc.get || desc.set) { Object.defineProperty(newObj, key, desc); } else { newObj[key] = obj[key]; } } } } newObj.default = obj; return newObj; } }
|
|||
|
|
|||
|
function _typeof(obj) { if (typeof Symbol === "function" && typeof Symbol.iterator === "symbol") { _typeof = function _typeof(obj) { return typeof obj; }; } else { _typeof = function _typeof(obj) { return obj && typeof Symbol === "function" && obj.constructor === Symbol && obj !== Symbol.prototype ? "symbol" : typeof obj; }; } return _typeof(obj); }
|
|||
|
|
|||
|
function _toConsumableArray(arr) { if (Array.isArray(arr)) { for (var i = 0, arr2 = new Array(arr.length); i < arr.length; i++) { arr2[i] = arr[i]; } return arr2; } else { return Array.from(arr); } }
|
|||
|
|
|||
|
function hasPlugin(name) {
|
|||
|
if (name !== "wast") throw new Error("unknow plugin");
|
|||
|
return true;
|
|||
|
}
|
|||
|
|
|||
|
function isKeyword(token, id) {
|
|||
|
return token.type === _tokenizer.tokens.keyword && token.value === id;
|
|||
|
}
|
|||
|
|
|||
|
function tokenToString(token) {
|
|||
|
if (token.type === "keyword") {
|
|||
|
return "keyword (".concat(token.value, ")");
|
|||
|
}
|
|||
|
|
|||
|
return token.type;
|
|||
|
}
|
|||
|
|
|||
|
function identifierFromToken(token) {
|
|||
|
var _token$loc = token.loc,
|
|||
|
end = _token$loc.end,
|
|||
|
start = _token$loc.start;
|
|||
|
return t.withLoc(t.identifier(token.value), end, start);
|
|||
|
}
|
|||
|
|
|||
|
function parse(tokensList, source) {
|
|||
|
var current = 0;
|
|||
|
var getUniqueName = t.getUniqueNameGenerator();
|
|||
|
var state = {
|
|||
|
registredExportedElements: []
|
|||
|
}; // But this time we're going to use recursion instead of a `while` loop. So we
|
|||
|
// define a `walk` function.
|
|||
|
|
|||
|
function walk() {
|
|||
|
var token = tokensList[current];
|
|||
|
|
|||
|
function eatToken() {
|
|||
|
token = tokensList[++current];
|
|||
|
}
|
|||
|
|
|||
|
function getEndLoc() {
|
|||
|
var currentToken = token;
|
|||
|
|
|||
|
if (typeof currentToken === "undefined") {
|
|||
|
var lastToken = tokensList[tokensList.length - 1];
|
|||
|
currentToken = lastToken;
|
|||
|
}
|
|||
|
|
|||
|
return currentToken.loc.end;
|
|||
|
}
|
|||
|
|
|||
|
function getStartLoc() {
|
|||
|
return token.loc.start;
|
|||
|
}
|
|||
|
|
|||
|
function eatTokenOfType(type) {
|
|||
|
if (token.type !== type) {
|
|||
|
throw new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "Assertion error: expected token of type " + type + ", given " + tokenToString(token));
|
|||
|
}
|
|||
|
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
|
|||
|
function parseExportIndex(token) {
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
var index = identifierFromToken(token);
|
|||
|
eatToken();
|
|||
|
return index;
|
|||
|
} else if (token.type === _tokenizer.tokens.number) {
|
|||
|
var _index = t.numberLiteralFromRaw(token.value);
|
|||
|
|
|||
|
eatToken();
|
|||
|
return _index;
|
|||
|
} else {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "unknown export index" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
function lookaheadAndCheck() {
|
|||
|
var len = arguments.length;
|
|||
|
|
|||
|
for (var i = 0; i < len; i++) {
|
|||
|
var tokenAhead = tokensList[current + i];
|
|||
|
var expectedToken = i < 0 || arguments.length <= i ? undefined : arguments[i];
|
|||
|
|
|||
|
if (tokenAhead.type === "keyword") {
|
|||
|
if (isKeyword(tokenAhead, expectedToken) === false) {
|
|||
|
return false;
|
|||
|
}
|
|||
|
} else if (expectedToken !== tokenAhead.type) {
|
|||
|
return false;
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
return true;
|
|||
|
} // TODO(sven): there is probably a better way to do this
|
|||
|
// can refactor it if it get out of hands
|
|||
|
|
|||
|
|
|||
|
function maybeIgnoreComment() {
|
|||
|
if (typeof token === "undefined") {
|
|||
|
// Ignore
|
|||
|
return;
|
|||
|
}
|
|||
|
|
|||
|
while (token.type === _tokenizer.tokens.comment) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
if (typeof token === "undefined") {
|
|||
|
// Hit the end
|
|||
|
break;
|
|||
|
}
|
|||
|
}
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses a memory instruction
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* memory: ( memory <name>? <memory_sig> )
|
|||
|
* ( memory <name>? ( export <string> ) <...> )
|
|||
|
* ( memory <name>? ( import <string> <string> ) <memory_sig> )
|
|||
|
* ( memory <name>? ( export <string> )* ( data <string>* )
|
|||
|
* memory_sig: <nat> <nat>?
|
|||
|
*
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseMemory() {
|
|||
|
var id = t.identifier(getUniqueName("memory"));
|
|||
|
var limits = t.limit(0);
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.string || token.type === _tokenizer.tokens.identifier) {
|
|||
|
id = t.identifier(token.value);
|
|||
|
eatToken();
|
|||
|
} else {
|
|||
|
id = t.withRaw(id, ""); // preserve anonymous
|
|||
|
}
|
|||
|
/**
|
|||
|
* Maybe data
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.data)) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
eatToken(); // data
|
|||
|
// TODO(sven): do something with the data collected here
|
|||
|
|
|||
|
var stringInitializer = token.value;
|
|||
|
eatTokenOfType(_tokenizer.tokens.string); // Update limits accordingly
|
|||
|
|
|||
|
limits = t.limit(stringInitializer.length);
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
/**
|
|||
|
* Maybe export
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.export)) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
eatToken(); // export
|
|||
|
|
|||
|
if (token.type !== _tokenizer.tokens.string) {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Expected string in export" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
var _name = token.value;
|
|||
|
eatToken();
|
|||
|
state.registredExportedElements.push({
|
|||
|
exportType: "Memory",
|
|||
|
name: _name,
|
|||
|
id: id
|
|||
|
});
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
/**
|
|||
|
* Memory signature
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.number) {
|
|||
|
limits = t.limit((0, _numberLiterals.parse32I)(token.value));
|
|||
|
eatToken();
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.number) {
|
|||
|
limits.max = (0, _numberLiterals.parse32I)(token.value);
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
return t.memory(limits, id);
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses a data section
|
|||
|
* https://webassembly.github.io/spec/core/text/modules.html#data-segments
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* data: ( data <index>? <offset> <string> )
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseData() {
|
|||
|
// optional memory index
|
|||
|
var memidx = 0;
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.number) {
|
|||
|
memidx = token.value;
|
|||
|
eatTokenOfType(_tokenizer.tokens.number); // .
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.openParen);
|
|||
|
var offset;
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.valtype) {
|
|||
|
eatTokenOfType(_tokenizer.tokens.valtype); // i32
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.dot); // .
|
|||
|
|
|||
|
if (token.value !== "const") {
|
|||
|
throw new Error("constant expression required");
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.name); // const
|
|||
|
|
|||
|
var numberLiteral = t.numberLiteralFromRaw(token.value, "i32");
|
|||
|
offset = t.objectInstruction("const", "i32", [numberLiteral]);
|
|||
|
eatToken();
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
} else {
|
|||
|
eatTokenOfType(_tokenizer.tokens.name); // get_global
|
|||
|
|
|||
|
var _numberLiteral = t.numberLiteralFromRaw(token.value, "i32");
|
|||
|
|
|||
|
offset = t.instruction("get_global", [_numberLiteral]);
|
|||
|
eatToken();
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
var byteArray = (0, _stringLiterals.parseString)(token.value);
|
|||
|
eatToken(); // "string"
|
|||
|
|
|||
|
return t.data(t.memIndexLiteral(memidx), offset, t.byteArray(byteArray));
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses a table instruction
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* table: ( table <name>? <table_type> )
|
|||
|
* ( table <name>? ( export <string> ) <...> )
|
|||
|
* ( table <name>? ( import <string> <string> ) <table_type> )
|
|||
|
* ( table <name>? ( export <string> )* <elem_type> ( elem <var>* ) )
|
|||
|
*
|
|||
|
* table_type: <nat> <nat>? <elem_type>
|
|||
|
* elem_type: anyfunc
|
|||
|
*
|
|||
|
* elem: ( elem <var>? (offset <instr>* ) <var>* )
|
|||
|
* ( elem <var>? <expr> <var>* )
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseTable() {
|
|||
|
var name = t.identifier(getUniqueName("table"));
|
|||
|
var limit = t.limit(0);
|
|||
|
var elemIndices = [];
|
|||
|
var elemType = "anyfunc";
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.string || token.type === _tokenizer.tokens.identifier) {
|
|||
|
name = identifierFromToken(token);
|
|||
|
eatToken();
|
|||
|
} else {
|
|||
|
name = t.withRaw(name, ""); // preserve anonymous
|
|||
|
}
|
|||
|
|
|||
|
while (token.type !== _tokenizer.tokens.closeParen) {
|
|||
|
/**
|
|||
|
* Maybe export
|
|||
|
*/
|
|||
|
if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.elem)) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
eatToken(); // elem
|
|||
|
|
|||
|
while (token.type === _tokenizer.tokens.identifier) {
|
|||
|
elemIndices.push(t.identifier(token.value));
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
} else if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.export)) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
eatToken(); // export
|
|||
|
|
|||
|
if (token.type !== _tokenizer.tokens.string) {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Expected string in export" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
var exportName = token.value;
|
|||
|
eatToken();
|
|||
|
state.registredExportedElements.push({
|
|||
|
exportType: "Table",
|
|||
|
name: exportName,
|
|||
|
id: name
|
|||
|
});
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
} else if (isKeyword(token, _tokenizer.keywords.anyfunc)) {
|
|||
|
// It's the default value, we can ignore it
|
|||
|
eatToken(); // anyfunc
|
|||
|
} else if (token.type === _tokenizer.tokens.number) {
|
|||
|
/**
|
|||
|
* Table type
|
|||
|
*/
|
|||
|
var min = parseInt(token.value);
|
|||
|
eatToken();
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.number) {
|
|||
|
var max = parseInt(token.value);
|
|||
|
eatToken();
|
|||
|
limit = t.limit(min, max);
|
|||
|
} else {
|
|||
|
limit = t.limit(min);
|
|||
|
}
|
|||
|
|
|||
|
eatToken();
|
|||
|
} else {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
if (elemIndices.length > 0) {
|
|||
|
return t.table(elemType, limit, name, elemIndices);
|
|||
|
} else {
|
|||
|
return t.table(elemType, limit, name);
|
|||
|
}
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses an import statement
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* import: ( import <string> <string> <imkind> )
|
|||
|
* imkind: ( func <name>? <func_sig> )
|
|||
|
* ( global <name>? <global_sig> )
|
|||
|
* ( table <name>? <table_sig> )
|
|||
|
* ( memory <name>? <memory_sig> )
|
|||
|
*
|
|||
|
* global_sig: <type> | ( mut <type> )
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseImport() {
|
|||
|
if (token.type !== _tokenizer.tokens.string) {
|
|||
|
throw new Error("Expected a string, " + token.type + " given.");
|
|||
|
}
|
|||
|
|
|||
|
var moduleName = token.value;
|
|||
|
eatToken();
|
|||
|
|
|||
|
if (token.type !== _tokenizer.tokens.string) {
|
|||
|
throw new Error("Expected a string, " + token.type + " given.");
|
|||
|
}
|
|||
|
|
|||
|
var name = token.value;
|
|||
|
eatToken();
|
|||
|
eatTokenOfType(_tokenizer.tokens.openParen);
|
|||
|
var descr;
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.func)) {
|
|||
|
eatToken(); // keyword
|
|||
|
|
|||
|
var fnParams = [];
|
|||
|
var fnResult = [];
|
|||
|
var typeRef;
|
|||
|
var fnName = t.identifier(getUniqueName("func"));
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
fnName = identifierFromToken(token);
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
|
|||
|
while (token.type === _tokenizer.tokens.openParen) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.keywords.type) === true) {
|
|||
|
eatToken();
|
|||
|
typeRef = parseTypeReference();
|
|||
|
} else if (lookaheadAndCheck(_tokenizer.keywords.param) === true) {
|
|||
|
eatToken();
|
|||
|
fnParams.push.apply(fnParams, _toConsumableArray(parseFuncParam()));
|
|||
|
} else if (lookaheadAndCheck(_tokenizer.keywords.result) === true) {
|
|||
|
eatToken();
|
|||
|
fnResult.push.apply(fnResult, _toConsumableArray(parseFuncResult()));
|
|||
|
} else {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in import of type" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
if (typeof fnName === "undefined") {
|
|||
|
throw new Error("Imported function must have a name");
|
|||
|
}
|
|||
|
|
|||
|
descr = t.funcImportDescr(fnName, typeRef !== undefined ? typeRef : t.signature(fnParams, fnResult));
|
|||
|
} else if (isKeyword(token, _tokenizer.keywords.global)) {
|
|||
|
eatToken(); // keyword
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.openParen) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.keyword); // mut keyword
|
|||
|
|
|||
|
var valtype = token.value;
|
|||
|
eatToken();
|
|||
|
descr = t.globalType(valtype, "var");
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
} else {
|
|||
|
var _valtype = token.value;
|
|||
|
eatTokenOfType(_tokenizer.tokens.valtype);
|
|||
|
descr = t.globalType(_valtype, "const");
|
|||
|
}
|
|||
|
} else if (isKeyword(token, _tokenizer.keywords.memory) === true) {
|
|||
|
eatToken(); // Keyword
|
|||
|
|
|||
|
descr = parseMemory();
|
|||
|
} else if (isKeyword(token, _tokenizer.keywords.table) === true) {
|
|||
|
eatToken(); // Keyword
|
|||
|
|
|||
|
descr = parseTable();
|
|||
|
} else {
|
|||
|
throw new Error("Unsupported import type: " + tokenToString(token));
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.moduleImport(moduleName, name, descr);
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses a block instruction
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* expr: ( block <name>? <block_sig> <instr>* )
|
|||
|
* instr: block <name>? <block_sig> <instr>* end <name>?
|
|||
|
* block_sig : ( result <type>* )*
|
|||
|
*
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseBlock() {
|
|||
|
var label = t.identifier(getUniqueName("block"));
|
|||
|
var blockResult = null;
|
|||
|
var instr = [];
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
label = identifierFromToken(token);
|
|||
|
eatToken();
|
|||
|
} else {
|
|||
|
label = t.withRaw(label, ""); // preserve anonymous
|
|||
|
}
|
|||
|
|
|||
|
while (token.type === _tokenizer.tokens.openParen) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.keywords.result) === true) {
|
|||
|
eatToken();
|
|||
|
blockResult = token.value;
|
|||
|
eatToken();
|
|||
|
} else if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword
|
|||
|
) {
|
|||
|
// Instruction
|
|||
|
instr.push(parseFuncInstr());
|
|||
|
} else {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in block body of type" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
maybeIgnoreComment();
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
return t.blockInstruction(label, instr, blockResult);
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses a if instruction
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* expr:
|
|||
|
* ( if <name>? <block_sig> ( then <instr>* ) ( else <instr>* )? )
|
|||
|
* ( if <name>? <block_sig> <expr>+ ( then <instr>* ) ( else <instr>* )? )
|
|||
|
*
|
|||
|
* instr:
|
|||
|
* if <name>? <block_sig> <instr>* end <name>?
|
|||
|
* if <name>? <block_sig> <instr>* else <name>? <instr>* end <name>?
|
|||
|
*
|
|||
|
* block_sig : ( result <type>* )*
|
|||
|
*
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseIf() {
|
|||
|
var blockResult = null;
|
|||
|
var label = t.identifier(getUniqueName("if"));
|
|||
|
var testInstrs = [];
|
|||
|
var consequent = [];
|
|||
|
var alternate = [];
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
label = identifierFromToken(token);
|
|||
|
eatToken();
|
|||
|
} else {
|
|||
|
label = t.withRaw(label, ""); // preserve anonymous
|
|||
|
}
|
|||
|
|
|||
|
while (token.type === _tokenizer.tokens.openParen) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
/**
|
|||
|
* Block signature
|
|||
|
*/
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.result) === true) {
|
|||
|
eatToken();
|
|||
|
blockResult = token.value;
|
|||
|
eatTokenOfType(_tokenizer.tokens.valtype);
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
continue;
|
|||
|
}
|
|||
|
/**
|
|||
|
* Then
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.then) === true) {
|
|||
|
eatToken(); // then
|
|||
|
|
|||
|
while (token.type === _tokenizer.tokens.openParen) {
|
|||
|
eatToken(); // Instruction
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword
|
|||
|
) {
|
|||
|
consequent.push(parseFuncInstr());
|
|||
|
} else {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in consequent body of type" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
continue;
|
|||
|
}
|
|||
|
/**
|
|||
|
* Alternate
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.else)) {
|
|||
|
eatToken(); // else
|
|||
|
|
|||
|
while (token.type === _tokenizer.tokens.openParen) {
|
|||
|
eatToken(); // Instruction
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword
|
|||
|
) {
|
|||
|
alternate.push(parseFuncInstr());
|
|||
|
} else {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in alternate body of type" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
continue;
|
|||
|
}
|
|||
|
/**
|
|||
|
* Test instruction
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword
|
|||
|
) {
|
|||
|
testInstrs.push(parseFuncInstr());
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
continue;
|
|||
|
}
|
|||
|
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in if body" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
return t.ifInstruction(label, testInstrs, blockResult, consequent, alternate);
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses a loop instruction
|
|||
|
*
|
|||
|
* WAT:
|
|||
|
*
|
|||
|
* blockinstr :: 'loop' I:label rt:resulttype (in:instr*) 'end' id?
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* instr :: loop <name>? <block_sig> <instr>* end <name>?
|
|||
|
* expr :: ( loop <name>? <block_sig> <instr>* )
|
|||
|
* block_sig :: ( result <type>* )*
|
|||
|
*
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseLoop() {
|
|||
|
var label = t.identifier(getUniqueName("loop"));
|
|||
|
var blockResult;
|
|||
|
var instr = [];
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
label = identifierFromToken(token);
|
|||
|
eatToken();
|
|||
|
} else {
|
|||
|
label = t.withRaw(label, ""); // preserve anonymous
|
|||
|
}
|
|||
|
|
|||
|
while (token.type === _tokenizer.tokens.openParen) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.keywords.result) === true) {
|
|||
|
eatToken();
|
|||
|
blockResult = token.value;
|
|||
|
eatToken();
|
|||
|
} else if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword
|
|||
|
) {
|
|||
|
// Instruction
|
|||
|
instr.push(parseFuncInstr());
|
|||
|
} else {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in loop body" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
return t.loopInstruction(label, blockResult, instr);
|
|||
|
}
|
|||
|
|
|||
|
function parseCallIndirect() {
|
|||
|
var typeRef;
|
|||
|
var params = [];
|
|||
|
var results = [];
|
|||
|
var instrs = [];
|
|||
|
|
|||
|
while (token.type !== _tokenizer.tokens.closeParen) {
|
|||
|
if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.type)) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
eatToken(); // type
|
|||
|
|
|||
|
typeRef = parseTypeReference();
|
|||
|
} else if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.param)) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
eatToken(); // param
|
|||
|
|
|||
|
/**
|
|||
|
* Params can be empty:
|
|||
|
* (params)`
|
|||
|
*/
|
|||
|
|
|||
|
if (token.type !== _tokenizer.tokens.closeParen) {
|
|||
|
params.push.apply(params, _toConsumableArray(parseFuncParam()));
|
|||
|
}
|
|||
|
} else if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.result)) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
eatToken(); // result
|
|||
|
|
|||
|
/**
|
|||
|
* Results can be empty:
|
|||
|
* (result)`
|
|||
|
*/
|
|||
|
|
|||
|
if (token.type !== _tokenizer.tokens.closeParen) {
|
|||
|
results.push.apply(results, _toConsumableArray(parseFuncResult()));
|
|||
|
}
|
|||
|
} else {
|
|||
|
eatTokenOfType(_tokenizer.tokens.openParen);
|
|||
|
instrs.push(parseFuncInstr());
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
return t.callIndirectInstruction(typeRef !== undefined ? typeRef : t.signature(params, results), instrs);
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses an export instruction
|
|||
|
*
|
|||
|
* WAT:
|
|||
|
*
|
|||
|
* export: ( export <string> <exkind> )
|
|||
|
* exkind: ( func <var> )
|
|||
|
* ( global <var> )
|
|||
|
* ( table <var> )
|
|||
|
* ( memory <var> )
|
|||
|
* var: <nat> | <name>
|
|||
|
*
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseExport() {
|
|||
|
if (token.type !== _tokenizer.tokens.string) {
|
|||
|
throw new Error("Expected string after export, got: " + token.type);
|
|||
|
}
|
|||
|
|
|||
|
var name = token.value;
|
|||
|
eatToken();
|
|||
|
var moduleExportDescr = parseModuleExportDescr();
|
|||
|
return t.moduleExport(name, moduleExportDescr);
|
|||
|
}
|
|||
|
|
|||
|
function parseModuleExportDescr() {
|
|||
|
var startLoc = getStartLoc();
|
|||
|
var type = "";
|
|||
|
var index;
|
|||
|
eatTokenOfType(_tokenizer.tokens.openParen);
|
|||
|
|
|||
|
while (token.type !== _tokenizer.tokens.closeParen) {
|
|||
|
if (isKeyword(token, _tokenizer.keywords.func)) {
|
|||
|
type = "Func";
|
|||
|
eatToken();
|
|||
|
index = parseExportIndex(token);
|
|||
|
} else if (isKeyword(token, _tokenizer.keywords.table)) {
|
|||
|
type = "Table";
|
|||
|
eatToken();
|
|||
|
index = parseExportIndex(token);
|
|||
|
} else if (isKeyword(token, _tokenizer.keywords.global)) {
|
|||
|
type = "Global";
|
|||
|
eatToken();
|
|||
|
index = parseExportIndex(token);
|
|||
|
} else if (isKeyword(token, _tokenizer.keywords.memory)) {
|
|||
|
type = "Memory";
|
|||
|
eatToken();
|
|||
|
index = parseExportIndex(token);
|
|||
|
}
|
|||
|
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
|
|||
|
if (type === "") {
|
|||
|
throw new Error("Unknown export type");
|
|||
|
}
|
|||
|
|
|||
|
if (index === undefined) {
|
|||
|
throw new Error("Exported function must have a name");
|
|||
|
}
|
|||
|
|
|||
|
var node = t.moduleExportDescr(type, index);
|
|||
|
var endLoc = getEndLoc();
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.withLoc(node, endLoc, startLoc);
|
|||
|
}
|
|||
|
|
|||
|
function parseModule() {
|
|||
|
var name = null;
|
|||
|
var isBinary = false;
|
|||
|
var isQuote = false;
|
|||
|
var moduleFields = [];
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
name = token.value;
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
|
|||
|
if (hasPlugin("wast") && token.type === _tokenizer.tokens.name && token.value === "binary") {
|
|||
|
eatToken();
|
|||
|
isBinary = true;
|
|||
|
}
|
|||
|
|
|||
|
if (hasPlugin("wast") && token.type === _tokenizer.tokens.name && token.value === "quote") {
|
|||
|
eatToken();
|
|||
|
isQuote = true;
|
|||
|
}
|
|||
|
|
|||
|
if (isBinary === true) {
|
|||
|
var blob = [];
|
|||
|
|
|||
|
while (token.type === _tokenizer.tokens.string) {
|
|||
|
blob.push(token.value);
|
|||
|
eatToken();
|
|||
|
maybeIgnoreComment();
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.binaryModule(name, blob);
|
|||
|
}
|
|||
|
|
|||
|
if (isQuote === true) {
|
|||
|
var string = [];
|
|||
|
|
|||
|
while (token.type === _tokenizer.tokens.string) {
|
|||
|
string.push(token.value);
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.quoteModule(name, string);
|
|||
|
}
|
|||
|
|
|||
|
while (token.type !== _tokenizer.tokens.closeParen) {
|
|||
|
moduleFields.push(walk());
|
|||
|
|
|||
|
if (state.registredExportedElements.length > 0) {
|
|||
|
state.registredExportedElements.forEach(function (decl) {
|
|||
|
moduleFields.push(t.moduleExport(decl.name, t.moduleExportDescr(decl.exportType, decl.id)));
|
|||
|
});
|
|||
|
state.registredExportedElements = [];
|
|||
|
}
|
|||
|
|
|||
|
token = tokensList[current];
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.module(name, moduleFields);
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses the arguments of an instruction
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseFuncInstrArguments(signature) {
|
|||
|
var args = [];
|
|||
|
var namedArgs = {};
|
|||
|
var signaturePtr = 0;
|
|||
|
|
|||
|
while (token.type === _tokenizer.tokens.name || isKeyword(token, _tokenizer.keywords.offset)) {
|
|||
|
var key = token.value;
|
|||
|
eatToken();
|
|||
|
eatTokenOfType(_tokenizer.tokens.equal);
|
|||
|
var value = void 0;
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.number) {
|
|||
|
value = t.numberLiteralFromRaw(token.value);
|
|||
|
} else {
|
|||
|
throw new Error("Unexpected type for argument: " + token.type);
|
|||
|
}
|
|||
|
|
|||
|
namedArgs[key] = value;
|
|||
|
eatToken();
|
|||
|
} // $FlowIgnore
|
|||
|
|
|||
|
|
|||
|
var signatureLength = signature.vector ? Infinity : signature.length;
|
|||
|
|
|||
|
while (token.type !== _tokenizer.tokens.closeParen && ( // $FlowIgnore
|
|||
|
token.type === _tokenizer.tokens.openParen || signaturePtr < signatureLength)) {
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
args.push(t.identifier(token.value));
|
|||
|
eatToken();
|
|||
|
} else if (token.type === _tokenizer.tokens.valtype) {
|
|||
|
// Handle locals
|
|||
|
args.push(t.valtypeLiteral(token.value));
|
|||
|
eatToken();
|
|||
|
} else if (token.type === _tokenizer.tokens.string) {
|
|||
|
args.push(t.stringLiteral(token.value));
|
|||
|
eatToken();
|
|||
|
} else if (token.type === _tokenizer.tokens.number) {
|
|||
|
args.push( // TODO(sven): refactor the type signature handling
|
|||
|
// https://github.com/xtuc/webassemblyjs/pull/129 is a good start
|
|||
|
t.numberLiteralFromRaw(token.value, // $FlowIgnore
|
|||
|
signature[signaturePtr] || "f64")); // $FlowIgnore
|
|||
|
|
|||
|
if (!signature.vector) {
|
|||
|
++signaturePtr;
|
|||
|
}
|
|||
|
|
|||
|
eatToken();
|
|||
|
} else if (token.type === _tokenizer.tokens.openParen) {
|
|||
|
/**
|
|||
|
* Maybe some nested instructions
|
|||
|
*/
|
|||
|
eatToken(); // Instruction
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword
|
|||
|
) {
|
|||
|
// $FlowIgnore
|
|||
|
args.push(parseFuncInstr());
|
|||
|
} else {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in nested instruction" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.closeParen) {
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
} else {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in instruction argument" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
return {
|
|||
|
args: args,
|
|||
|
namedArgs: namedArgs
|
|||
|
};
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses an instruction
|
|||
|
*
|
|||
|
* WAT:
|
|||
|
*
|
|||
|
* instr :: plaininst
|
|||
|
* blockinstr
|
|||
|
*
|
|||
|
* blockinstr :: 'block' I:label rt:resulttype (in:instr*) 'end' id?
|
|||
|
* 'loop' I:label rt:resulttype (in:instr*) 'end' id?
|
|||
|
* 'if' I:label rt:resulttype (in:instr*) 'else' id? (in2:intr*) 'end' id?
|
|||
|
*
|
|||
|
* plaininst :: 'unreachable'
|
|||
|
* 'nop'
|
|||
|
* 'br' l:labelidx
|
|||
|
* 'br_if' l:labelidx
|
|||
|
* 'br_table' l*:vec(labelidx) ln:labelidx
|
|||
|
* 'return'
|
|||
|
* 'call' x:funcidx
|
|||
|
* 'call_indirect' x, I:typeuse
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* instr:
|
|||
|
* <expr>
|
|||
|
* <op>
|
|||
|
* block <name>? <block_sig> <instr>* end <name>?
|
|||
|
* loop <name>? <block_sig> <instr>* end <name>?
|
|||
|
* if <name>? <block_sig> <instr>* end <name>?
|
|||
|
* if <name>? <block_sig> <instr>* else <name>? <instr>* end <name>?
|
|||
|
*
|
|||
|
* expr:
|
|||
|
* ( <op> )
|
|||
|
* ( <op> <expr>+ )
|
|||
|
* ( block <name>? <block_sig> <instr>* )
|
|||
|
* ( loop <name>? <block_sig> <instr>* )
|
|||
|
* ( if <name>? <block_sig> ( then <instr>* ) ( else <instr>* )? )
|
|||
|
* ( if <name>? <block_sig> <expr>+ ( then <instr>* ) ( else <instr>* )? )
|
|||
|
*
|
|||
|
* op:
|
|||
|
* unreachable
|
|||
|
* nop
|
|||
|
* br <var>
|
|||
|
* br_if <var>
|
|||
|
* br_table <var>+
|
|||
|
* return
|
|||
|
* call <var>
|
|||
|
* call_indirect <func_sig>
|
|||
|
* drop
|
|||
|
* select
|
|||
|
* get_local <var>
|
|||
|
* set_local <var>
|
|||
|
* tee_local <var>
|
|||
|
* get_global <var>
|
|||
|
* set_global <var>
|
|||
|
* <type>.load((8|16|32)_<sign>)? <offset>? <align>?
|
|||
|
* <type>.store(8|16|32)? <offset>? <align>?
|
|||
|
* current_memory
|
|||
|
* grow_memory
|
|||
|
* <type>.const <value>
|
|||
|
* <type>.<unop>
|
|||
|
* <type>.<binop>
|
|||
|
* <type>.<testop>
|
|||
|
* <type>.<relop>
|
|||
|
* <type>.<cvtop>/<type>
|
|||
|
*
|
|||
|
* func_type: ( type <var> )? <param>* <result>*
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseFuncInstr() {
|
|||
|
var startLoc = getStartLoc();
|
|||
|
maybeIgnoreComment();
|
|||
|
/**
|
|||
|
* A simple instruction
|
|||
|
*/
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.name || token.type === _tokenizer.tokens.valtype) {
|
|||
|
var _name2 = token.value;
|
|||
|
var object;
|
|||
|
eatToken();
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.dot) {
|
|||
|
object = _name2;
|
|||
|
eatToken();
|
|||
|
|
|||
|
if (token.type !== _tokenizer.tokens.name) {
|
|||
|
throw new TypeError("Unknown token: " + token.type + ", name expected");
|
|||
|
}
|
|||
|
|
|||
|
_name2 = token.value;
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.closeParen) {
|
|||
|
var _endLoc = token.loc.end;
|
|||
|
|
|||
|
if (typeof object === "undefined") {
|
|||
|
return t.withLoc(t.instruction(_name2), _endLoc, startLoc);
|
|||
|
} else {
|
|||
|
return t.withLoc(t.objectInstruction(_name2, object, []), _endLoc, startLoc);
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
var signature = t.signatureForOpcode(object || "", _name2);
|
|||
|
|
|||
|
var _parseFuncInstrArgume = parseFuncInstrArguments(signature),
|
|||
|
_args = _parseFuncInstrArgume.args,
|
|||
|
_namedArgs = _parseFuncInstrArgume.namedArgs;
|
|||
|
|
|||
|
var endLoc = token.loc.end;
|
|||
|
|
|||
|
if (typeof object === "undefined") {
|
|||
|
return t.withLoc(t.instruction(_name2, _args, _namedArgs), endLoc, startLoc);
|
|||
|
} else {
|
|||
|
return t.withLoc(t.objectInstruction(_name2, object, _args, _namedArgs), endLoc, startLoc);
|
|||
|
}
|
|||
|
} else if (isKeyword(token, _tokenizer.keywords.loop)) {
|
|||
|
/**
|
|||
|
* Else a instruction with a keyword (loop or block)
|
|||
|
*/
|
|||
|
eatToken(); // keyword
|
|||
|
|
|||
|
return parseLoop();
|
|||
|
} else if (isKeyword(token, _tokenizer.keywords.block)) {
|
|||
|
eatToken(); // keyword
|
|||
|
|
|||
|
return parseBlock();
|
|||
|
} else if (isKeyword(token, _tokenizer.keywords.call_indirect)) {
|
|||
|
eatToken(); // keyword
|
|||
|
|
|||
|
return parseCallIndirect();
|
|||
|
} else if (isKeyword(token, _tokenizer.keywords.call)) {
|
|||
|
eatToken(); // keyword
|
|||
|
|
|||
|
var index;
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
index = identifierFromToken(token);
|
|||
|
eatToken();
|
|||
|
} else if (token.type === _tokenizer.tokens.number) {
|
|||
|
index = t.indexLiteral(token.value);
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
|
|||
|
var instrArgs = []; // Nested instruction
|
|||
|
|
|||
|
while (token.type === _tokenizer.tokens.openParen) {
|
|||
|
eatToken();
|
|||
|
instrArgs.push(parseFuncInstr());
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
if (typeof index === "undefined") {
|
|||
|
throw new Error("Missing argument in call instruciton");
|
|||
|
}
|
|||
|
|
|||
|
if (instrArgs.length > 0) {
|
|||
|
return t.callInstruction(index, instrArgs);
|
|||
|
} else {
|
|||
|
return t.callInstruction(index);
|
|||
|
}
|
|||
|
} else if (isKeyword(token, _tokenizer.keywords.if)) {
|
|||
|
eatToken(); // Keyword
|
|||
|
|
|||
|
return parseIf();
|
|||
|
} else if (isKeyword(token, _tokenizer.keywords.module) && hasPlugin("wast")) {
|
|||
|
eatToken(); // In WAST you can have a module as an instruction's argument
|
|||
|
// we will cast it into a instruction to not break the flow
|
|||
|
// $FlowIgnore
|
|||
|
|
|||
|
var module = parseModule();
|
|||
|
return module;
|
|||
|
} else {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected instruction in function body" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
}
|
|||
|
/*
|
|||
|
* Parses a function
|
|||
|
*
|
|||
|
* WAT:
|
|||
|
*
|
|||
|
* functype :: ( 'func' t1:vec(param) t2:vec(result) )
|
|||
|
* param :: ( 'param' id? t:valtype )
|
|||
|
* result :: ( 'result' t:valtype )
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* func :: ( func <name>? <func_sig> <local>* <instr>* )
|
|||
|
* ( func <name>? ( export <string> ) <...> )
|
|||
|
* ( func <name>? ( import <string> <string> ) <func_sig> )
|
|||
|
* func_sig :: ( type <var> )? <param>* <result>*
|
|||
|
* param :: ( param <type>* ) | ( param <name> <type> )
|
|||
|
* result :: ( result <type>* )
|
|||
|
* local :: ( local <type>* ) | ( local <name> <type> )
|
|||
|
*
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseFunc() {
|
|||
|
var fnName = t.identifier(getUniqueName("func"));
|
|||
|
var typeRef;
|
|||
|
var fnBody = [];
|
|||
|
var fnParams = [];
|
|||
|
var fnResult = []; // name
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
fnName = identifierFromToken(token);
|
|||
|
eatToken();
|
|||
|
} else {
|
|||
|
fnName = t.withRaw(fnName, ""); // preserve anonymous
|
|||
|
}
|
|||
|
|
|||
|
maybeIgnoreComment();
|
|||
|
|
|||
|
while (token.type === _tokenizer.tokens.openParen || token.type === _tokenizer.tokens.name || token.type === _tokenizer.tokens.valtype) {
|
|||
|
// Instructions without parens
|
|||
|
if (token.type === _tokenizer.tokens.name || token.type === _tokenizer.tokens.valtype) {
|
|||
|
fnBody.push(parseFuncInstr());
|
|||
|
continue;
|
|||
|
}
|
|||
|
|
|||
|
eatToken();
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.keywords.param) === true) {
|
|||
|
eatToken();
|
|||
|
fnParams.push.apply(fnParams, _toConsumableArray(parseFuncParam()));
|
|||
|
} else if (lookaheadAndCheck(_tokenizer.keywords.result) === true) {
|
|||
|
eatToken();
|
|||
|
fnResult.push.apply(fnResult, _toConsumableArray(parseFuncResult()));
|
|||
|
} else if (lookaheadAndCheck(_tokenizer.keywords.export) === true) {
|
|||
|
eatToken();
|
|||
|
parseFuncExport(fnName);
|
|||
|
} else if (lookaheadAndCheck(_tokenizer.keywords.type) === true) {
|
|||
|
eatToken();
|
|||
|
typeRef = parseTypeReference();
|
|||
|
} else if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword
|
|||
|
) {
|
|||
|
// Instruction
|
|||
|
fnBody.push(parseFuncInstr());
|
|||
|
} else {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in func body" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
return t.func(fnName, typeRef !== undefined ? typeRef : t.signature(fnParams, fnResult), fnBody);
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses shorthand export in func
|
|||
|
*
|
|||
|
* export :: ( export <string> )
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseFuncExport(funcId) {
|
|||
|
if (token.type !== _tokenizer.tokens.string) {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Function export expected a string" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
var name = token.value;
|
|||
|
eatToken();
|
|||
|
/**
|
|||
|
* Func export shorthand, we trait it as a syntaxic sugar.
|
|||
|
* A export ModuleField will be added later.
|
|||
|
*
|
|||
|
* We give the anonymous function a generated name and export it.
|
|||
|
*/
|
|||
|
|
|||
|
var id = t.identifier(funcId.value);
|
|||
|
state.registredExportedElements.push({
|
|||
|
exportType: "Func",
|
|||
|
name: name,
|
|||
|
id: id
|
|||
|
});
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses a type instruction
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* typedef: ( type <name>? ( func <param>* <result>* ) )
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseType() {
|
|||
|
var id;
|
|||
|
var params = [];
|
|||
|
var result = [];
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
id = identifierFromToken(token);
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.func)) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
eatToken(); // func
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.closeParen) {
|
|||
|
eatToken(); // function with an empty signature, we can abort here
|
|||
|
|
|||
|
return t.typeInstruction(id, t.signature([], []));
|
|||
|
}
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.param)) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
eatToken(); // param
|
|||
|
|
|||
|
params = parseFuncParam();
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.result)) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
eatToken(); // result
|
|||
|
|
|||
|
result = parseFuncResult();
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
return t.typeInstruction(id, t.signature(params, result));
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses a function result
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* result :: ( result <type>* )
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseFuncResult() {
|
|||
|
var results = [];
|
|||
|
|
|||
|
while (token.type !== _tokenizer.tokens.closeParen) {
|
|||
|
if (token.type !== _tokenizer.tokens.valtype) {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in func result" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
var valtype = token.value;
|
|||
|
eatToken();
|
|||
|
results.push(valtype);
|
|||
|
}
|
|||
|
|
|||
|
return results;
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses a type reference
|
|||
|
*
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseTypeReference() {
|
|||
|
var ref;
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
ref = identifierFromToken(token);
|
|||
|
eatToken();
|
|||
|
} else if (token.type === _tokenizer.tokens.number) {
|
|||
|
ref = t.numberLiteralFromRaw(token.value);
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
|
|||
|
return ref;
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses a global instruction
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* global: ( global <name>? <global_sig> <instr>* )
|
|||
|
* ( global <name>? ( export <string> ) <...> )
|
|||
|
* ( global <name>? ( import <string> <string> ) <global_sig> )
|
|||
|
*
|
|||
|
* global_sig: <type> | ( mut <type> )
|
|||
|
*
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseGlobal() {
|
|||
|
var name = t.identifier(getUniqueName("global"));
|
|||
|
var type; // Keep informations in case of a shorthand import
|
|||
|
|
|||
|
var importing = null;
|
|||
|
maybeIgnoreComment();
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
name = identifierFromToken(token);
|
|||
|
eatToken();
|
|||
|
} else {
|
|||
|
name = t.withRaw(name, ""); // preserve anonymous
|
|||
|
}
|
|||
|
/**
|
|||
|
* maybe export
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.export)) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
eatToken(); // export
|
|||
|
|
|||
|
var exportName = token.value;
|
|||
|
eatTokenOfType(_tokenizer.tokens.string);
|
|||
|
state.registredExportedElements.push({
|
|||
|
exportType: "Global",
|
|||
|
name: exportName,
|
|||
|
id: name
|
|||
|
});
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
/**
|
|||
|
* maybe import
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.import)) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
eatToken(); // import
|
|||
|
|
|||
|
var moduleName = token.value;
|
|||
|
eatTokenOfType(_tokenizer.tokens.string);
|
|||
|
var _name3 = token.value;
|
|||
|
eatTokenOfType(_tokenizer.tokens.string);
|
|||
|
importing = {
|
|||
|
module: moduleName,
|
|||
|
name: _name3,
|
|||
|
descr: undefined
|
|||
|
};
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
/**
|
|||
|
* global_sig
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.valtype) {
|
|||
|
type = t.globalType(token.value, "const");
|
|||
|
eatToken();
|
|||
|
} else if (token.type === _tokenizer.tokens.openParen) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.mut) === false) {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unsupported global type, expected mut" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
eatToken(); // mut
|
|||
|
|
|||
|
type = t.globalType(token.value, "var");
|
|||
|
eatToken();
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
if (type === undefined) {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Could not determine global type" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
maybeIgnoreComment();
|
|||
|
var init = [];
|
|||
|
|
|||
|
if (importing != null) {
|
|||
|
importing.descr = type;
|
|||
|
init.push(t.moduleImport(importing.module, importing.name, importing.descr));
|
|||
|
}
|
|||
|
/**
|
|||
|
* instr*
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
while (token.type === _tokenizer.tokens.openParen) {
|
|||
|
eatToken();
|
|||
|
init.push(parseFuncInstr());
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
return t.global(type, init, name);
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses a function param
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* param :: ( param <type>* ) | ( param <name> <type> )
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseFuncParam() {
|
|||
|
var params = [];
|
|||
|
var id;
|
|||
|
var valtype;
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
id = token.value;
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.valtype) {
|
|||
|
valtype = token.value;
|
|||
|
eatToken();
|
|||
|
params.push({
|
|||
|
id: id,
|
|||
|
valtype: valtype
|
|||
|
});
|
|||
|
/**
|
|||
|
* Shorthand notation for multiple anonymous parameters
|
|||
|
* @see https://webassembly.github.io/spec/core/text/types.html#function-types
|
|||
|
* @see https://github.com/xtuc/webassemblyjs/issues/6
|
|||
|
*/
|
|||
|
|
|||
|
if (id === undefined) {
|
|||
|
while (token.type === _tokenizer.tokens.valtype) {
|
|||
|
valtype = token.value;
|
|||
|
eatToken();
|
|||
|
params.push({
|
|||
|
id: undefined,
|
|||
|
valtype: valtype
|
|||
|
});
|
|||
|
}
|
|||
|
}
|
|||
|
} else {// ignore
|
|||
|
}
|
|||
|
|
|||
|
return params;
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses an element segments instruction
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* elem: ( elem <var>? (offset <instr>* ) <var>* )
|
|||
|
* ( elem <var>? <expr> <var>* )
|
|||
|
*
|
|||
|
* var: <nat> | <name>
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseElem() {
|
|||
|
var tableIndex = t.indexLiteral(0);
|
|||
|
var offset = [];
|
|||
|
var funcs = [];
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
tableIndex = identifierFromToken(token);
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.number) {
|
|||
|
tableIndex = t.indexLiteral(token.value);
|
|||
|
eatToken();
|
|||
|
}
|
|||
|
|
|||
|
while (token.type !== _tokenizer.tokens.closeParen) {
|
|||
|
if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.offset)) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
eatToken(); // offset
|
|||
|
|
|||
|
while (token.type !== _tokenizer.tokens.closeParen) {
|
|||
|
eatTokenOfType(_tokenizer.tokens.openParen);
|
|||
|
offset.push(parseFuncInstr());
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
} else if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
funcs.push(t.identifier(token.value));
|
|||
|
eatToken();
|
|||
|
} else if (token.type === _tokenizer.tokens.number) {
|
|||
|
funcs.push(t.indexLiteral(token.value));
|
|||
|
eatToken();
|
|||
|
} else if (token.type === _tokenizer.tokens.openParen) {
|
|||
|
eatToken(); // (
|
|||
|
|
|||
|
offset.push(parseFuncInstr());
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
} else {
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unsupported token in elem" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
return t.elem(tableIndex, offset, funcs);
|
|||
|
}
|
|||
|
/**
|
|||
|
* Parses the start instruction in a module
|
|||
|
*
|
|||
|
* WAST:
|
|||
|
*
|
|||
|
* start: ( start <var> )
|
|||
|
* var: <nat> | <name>
|
|||
|
*
|
|||
|
* WAT:
|
|||
|
* start ::= ‘(’ ‘start’ x:funcidx ‘)’
|
|||
|
*/
|
|||
|
|
|||
|
|
|||
|
function parseStart() {
|
|||
|
if (token.type === _tokenizer.tokens.identifier) {
|
|||
|
var index = identifierFromToken(token);
|
|||
|
eatToken();
|
|||
|
return t.start(index);
|
|||
|
}
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.number) {
|
|||
|
var _index2 = t.indexLiteral(token.value);
|
|||
|
|
|||
|
eatToken();
|
|||
|
return t.start(_index2);
|
|||
|
}
|
|||
|
|
|||
|
throw new Error("Unknown start, token: " + tokenToString(token));
|
|||
|
}
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.openParen) {
|
|||
|
eatToken();
|
|||
|
var startLoc = getStartLoc();
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.export)) {
|
|||
|
eatToken();
|
|||
|
var node = parseExport();
|
|||
|
|
|||
|
var _endLoc2 = getEndLoc();
|
|||
|
|
|||
|
return t.withLoc(node, _endLoc2, startLoc);
|
|||
|
}
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.loop)) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
var _node = parseLoop();
|
|||
|
|
|||
|
var _endLoc3 = getEndLoc();
|
|||
|
|
|||
|
return t.withLoc(_node, _endLoc3, startLoc);
|
|||
|
}
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.func)) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
var _node2 = parseFunc();
|
|||
|
|
|||
|
var _endLoc4 = getEndLoc();
|
|||
|
|
|||
|
maybeIgnoreComment();
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.withLoc(_node2, _endLoc4, startLoc);
|
|||
|
}
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.module)) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
var _node3 = parseModule();
|
|||
|
|
|||
|
var _endLoc5 = getEndLoc();
|
|||
|
|
|||
|
return t.withLoc(_node3, _endLoc5, startLoc);
|
|||
|
}
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.import)) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
var _node4 = parseImport();
|
|||
|
|
|||
|
var _endLoc6 = getEndLoc();
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.withLoc(_node4, _endLoc6, startLoc);
|
|||
|
}
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.block)) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
var _node5 = parseBlock();
|
|||
|
|
|||
|
var _endLoc7 = getEndLoc();
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.withLoc(_node5, _endLoc7, startLoc);
|
|||
|
}
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.memory)) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
var _node6 = parseMemory();
|
|||
|
|
|||
|
var _endLoc8 = getEndLoc();
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.withLoc(_node6, _endLoc8, startLoc);
|
|||
|
}
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.data)) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
var _node7 = parseData();
|
|||
|
|
|||
|
var _endLoc9 = getEndLoc();
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.withLoc(_node7, _endLoc9, startLoc);
|
|||
|
}
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.table)) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
var _node8 = parseTable();
|
|||
|
|
|||
|
var _endLoc10 = getEndLoc();
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.withLoc(_node8, _endLoc10, startLoc);
|
|||
|
}
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.global)) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
var _node9 = parseGlobal();
|
|||
|
|
|||
|
var _endLoc11 = getEndLoc();
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.withLoc(_node9, _endLoc11, startLoc);
|
|||
|
}
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.type)) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
var _node10 = parseType();
|
|||
|
|
|||
|
var _endLoc12 = getEndLoc();
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.withLoc(_node10, _endLoc12, startLoc);
|
|||
|
}
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.start)) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
var _node11 = parseStart();
|
|||
|
|
|||
|
var _endLoc13 = getEndLoc();
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.withLoc(_node11, _endLoc13, startLoc);
|
|||
|
}
|
|||
|
|
|||
|
if (isKeyword(token, _tokenizer.keywords.elem)) {
|
|||
|
eatToken();
|
|||
|
|
|||
|
var _node12 = parseElem();
|
|||
|
|
|||
|
var _endLoc14 = getEndLoc();
|
|||
|
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
return t.withLoc(_node12, _endLoc14, startLoc);
|
|||
|
}
|
|||
|
|
|||
|
var instruction = parseFuncInstr();
|
|||
|
var endLoc = getEndLoc();
|
|||
|
maybeIgnoreComment();
|
|||
|
|
|||
|
if (_typeof(instruction) === "object") {
|
|||
|
if (typeof token !== "undefined") {
|
|||
|
eatTokenOfType(_tokenizer.tokens.closeParen);
|
|||
|
}
|
|||
|
|
|||
|
return t.withLoc(instruction, endLoc, startLoc);
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
if (token.type === _tokenizer.tokens.comment) {
|
|||
|
var _startLoc = getStartLoc();
|
|||
|
|
|||
|
var builder = token.opts.type === "leading" ? t.leadingComment : t.blockComment;
|
|||
|
|
|||
|
var _node13 = builder(token.value);
|
|||
|
|
|||
|
eatToken(); // comment
|
|||
|
|
|||
|
var _endLoc15 = getEndLoc();
|
|||
|
|
|||
|
return t.withLoc(_node13, _endLoc15, _startLoc);
|
|||
|
}
|
|||
|
|
|||
|
throw function () {
|
|||
|
return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unknown token" + ", given " + tokenToString(token));
|
|||
|
}();
|
|||
|
}
|
|||
|
|
|||
|
var body = [];
|
|||
|
|
|||
|
while (current < tokensList.length) {
|
|||
|
body.push(walk());
|
|||
|
}
|
|||
|
|
|||
|
return t.program(body);
|
|||
|
}
|