|
|
- "use strict";
-
- Object.defineProperty(exports, "__esModule", {
- value: true
- });
- exports.parse = parse;
-
- var _helperCodeFrame = require("@webassemblyjs/helper-code-frame");
-
- var t = _interopRequireWildcard(require("@webassemblyjs/ast"));
-
- var _numberLiterals = require("./number-literals");
-
- var _stringLiterals = require("./string-literals");
-
- var _tokenizer = require("./tokenizer");
-
- function _interopRequireWildcard(obj) { if (obj && obj.__esModule) { return obj; } else { var newObj = {}; if (obj != null) { for (var key in obj) { if (Object.prototype.hasOwnProperty.call(obj, key)) { var desc = Object.defineProperty && Object.getOwnPropertyDescriptor ? Object.getOwnPropertyDescriptor(obj, key) : {}; if (desc.get || desc.set) { Object.defineProperty(newObj, key, desc); } else { newObj[key] = obj[key]; } } } } newObj.default = obj; return newObj; } }
-
- function _typeof(obj) { if (typeof Symbol === "function" && typeof Symbol.iterator === "symbol") { _typeof = function _typeof(obj) { return typeof obj; }; } else { _typeof = function _typeof(obj) { return obj && typeof Symbol === "function" && obj.constructor === Symbol && obj !== Symbol.prototype ? "symbol" : typeof obj; }; } return _typeof(obj); }
-
- function _toConsumableArray(arr) { if (Array.isArray(arr)) { for (var i = 0, arr2 = new Array(arr.length); i < arr.length; i++) { arr2[i] = arr[i]; } return arr2; } else { return Array.from(arr); } }
-
- function hasPlugin(name) {
- if (name !== "wast") throw new Error("unknow plugin");
- return true;
- }
-
- function isKeyword(token, id) {
- return token.type === _tokenizer.tokens.keyword && token.value === id;
- }
-
- function tokenToString(token) {
- if (token.type === "keyword") {
- return "keyword (".concat(token.value, ")");
- }
-
- return token.type;
- }
-
- function identifierFromToken(token) {
- var _token$loc = token.loc,
- end = _token$loc.end,
- start = _token$loc.start;
- return t.withLoc(t.identifier(token.value), end, start);
- }
-
- function parse(tokensList, source) {
- var current = 0;
- var getUniqueName = t.getUniqueNameGenerator();
- var state = {
- registredExportedElements: []
- }; // But this time we're going to use recursion instead of a `while` loop. So we
- // define a `walk` function.
-
- function walk() {
- var token = tokensList[current];
-
- function eatToken() {
- token = tokensList[++current];
- }
-
- function getEndLoc() {
- var currentToken = token;
-
- if (typeof currentToken === "undefined") {
- var lastToken = tokensList[tokensList.length - 1];
- currentToken = lastToken;
- }
-
- return currentToken.loc.end;
- }
-
- function getStartLoc() {
- return token.loc.start;
- }
-
- function eatTokenOfType(type) {
- if (token.type !== type) {
- throw new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "Assertion error: expected token of type " + type + ", given " + tokenToString(token));
- }
-
- eatToken();
- }
-
- function parseExportIndex(token) {
- if (token.type === _tokenizer.tokens.identifier) {
- var index = identifierFromToken(token);
- eatToken();
- return index;
- } else if (token.type === _tokenizer.tokens.number) {
- var _index = t.numberLiteralFromRaw(token.value);
-
- eatToken();
- return _index;
- } else {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "unknown export index" + ", given " + tokenToString(token));
- }();
- }
- }
-
- function lookaheadAndCheck() {
- var len = arguments.length;
-
- for (var i = 0; i < len; i++) {
- var tokenAhead = tokensList[current + i];
- var expectedToken = i < 0 || arguments.length <= i ? undefined : arguments[i];
-
- if (tokenAhead.type === "keyword") {
- if (isKeyword(tokenAhead, expectedToken) === false) {
- return false;
- }
- } else if (expectedToken !== tokenAhead.type) {
- return false;
- }
- }
-
- return true;
- } // TODO(sven): there is probably a better way to do this
- // can refactor it if it get out of hands
-
-
- function maybeIgnoreComment() {
- if (typeof token === "undefined") {
- // Ignore
- return;
- }
-
- while (token.type === _tokenizer.tokens.comment) {
- eatToken();
-
- if (typeof token === "undefined") {
- // Hit the end
- break;
- }
- }
- }
- /**
- * Parses a memory instruction
- *
- * WAST:
- *
- * memory: ( memory <name>? <memory_sig> )
- * ( memory <name>? ( export <string> ) <...> )
- * ( memory <name>? ( import <string> <string> ) <memory_sig> )
- * ( memory <name>? ( export <string> )* ( data <string>* )
- * memory_sig: <nat> <nat>?
- *
- */
-
-
- function parseMemory() {
- var id = t.identifier(getUniqueName("memory"));
- var limits = t.limit(0);
-
- if (token.type === _tokenizer.tokens.string || token.type === _tokenizer.tokens.identifier) {
- id = t.identifier(token.value);
- eatToken();
- } else {
- id = t.withRaw(id, ""); // preserve anonymous
- }
- /**
- * Maybe data
- */
-
-
- if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.data)) {
- eatToken(); // (
-
- eatToken(); // data
- // TODO(sven): do something with the data collected here
-
- var stringInitializer = token.value;
- eatTokenOfType(_tokenizer.tokens.string); // Update limits accordingly
-
- limits = t.limit(stringInitializer.length);
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
- /**
- * Maybe export
- */
-
-
- if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.export)) {
- eatToken(); // (
-
- eatToken(); // export
-
- if (token.type !== _tokenizer.tokens.string) {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Expected string in export" + ", given " + tokenToString(token));
- }();
- }
-
- var _name = token.value;
- eatToken();
- state.registredExportedElements.push({
- exportType: "Memory",
- name: _name,
- id: id
- });
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
- /**
- * Memory signature
- */
-
-
- if (token.type === _tokenizer.tokens.number) {
- limits = t.limit((0, _numberLiterals.parse32I)(token.value));
- eatToken();
-
- if (token.type === _tokenizer.tokens.number) {
- limits.max = (0, _numberLiterals.parse32I)(token.value);
- eatToken();
- }
- }
-
- return t.memory(limits, id);
- }
- /**
- * Parses a data section
- * https://webassembly.github.io/spec/core/text/modules.html#data-segments
- *
- * WAST:
- *
- * data: ( data <index>? <offset> <string> )
- */
-
-
- function parseData() {
- // optional memory index
- var memidx = 0;
-
- if (token.type === _tokenizer.tokens.number) {
- memidx = token.value;
- eatTokenOfType(_tokenizer.tokens.number); // .
- }
-
- eatTokenOfType(_tokenizer.tokens.openParen);
- var offset;
-
- if (token.type === _tokenizer.tokens.valtype) {
- eatTokenOfType(_tokenizer.tokens.valtype); // i32
-
- eatTokenOfType(_tokenizer.tokens.dot); // .
-
- if (token.value !== "const") {
- throw new Error("constant expression required");
- }
-
- eatTokenOfType(_tokenizer.tokens.name); // const
-
- var numberLiteral = t.numberLiteralFromRaw(token.value, "i32");
- offset = t.objectInstruction("const", "i32", [numberLiteral]);
- eatToken();
- eatTokenOfType(_tokenizer.tokens.closeParen);
- } else {
- eatTokenOfType(_tokenizer.tokens.name); // get_global
-
- var _numberLiteral = t.numberLiteralFromRaw(token.value, "i32");
-
- offset = t.instruction("get_global", [_numberLiteral]);
- eatToken();
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- var byteArray = (0, _stringLiterals.parseString)(token.value);
- eatToken(); // "string"
-
- return t.data(t.memIndexLiteral(memidx), offset, t.byteArray(byteArray));
- }
- /**
- * Parses a table instruction
- *
- * WAST:
- *
- * table: ( table <name>? <table_type> )
- * ( table <name>? ( export <string> ) <...> )
- * ( table <name>? ( import <string> <string> ) <table_type> )
- * ( table <name>? ( export <string> )* <elem_type> ( elem <var>* ) )
- *
- * table_type: <nat> <nat>? <elem_type>
- * elem_type: anyfunc
- *
- * elem: ( elem <var>? (offset <instr>* ) <var>* )
- * ( elem <var>? <expr> <var>* )
- */
-
-
- function parseTable() {
- var name = t.identifier(getUniqueName("table"));
- var limit = t.limit(0);
- var elemIndices = [];
- var elemType = "anyfunc";
-
- if (token.type === _tokenizer.tokens.string || token.type === _tokenizer.tokens.identifier) {
- name = identifierFromToken(token);
- eatToken();
- } else {
- name = t.withRaw(name, ""); // preserve anonymous
- }
-
- while (token.type !== _tokenizer.tokens.closeParen) {
- /**
- * Maybe export
- */
- if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.elem)) {
- eatToken(); // (
-
- eatToken(); // elem
-
- while (token.type === _tokenizer.tokens.identifier) {
- elemIndices.push(t.identifier(token.value));
- eatToken();
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- } else if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.export)) {
- eatToken(); // (
-
- eatToken(); // export
-
- if (token.type !== _tokenizer.tokens.string) {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Expected string in export" + ", given " + tokenToString(token));
- }();
- }
-
- var exportName = token.value;
- eatToken();
- state.registredExportedElements.push({
- exportType: "Table",
- name: exportName,
- id: name
- });
- eatTokenOfType(_tokenizer.tokens.closeParen);
- } else if (isKeyword(token, _tokenizer.keywords.anyfunc)) {
- // It's the default value, we can ignore it
- eatToken(); // anyfunc
- } else if (token.type === _tokenizer.tokens.number) {
- /**
- * Table type
- */
- var min = parseInt(token.value);
- eatToken();
-
- if (token.type === _tokenizer.tokens.number) {
- var max = parseInt(token.value);
- eatToken();
- limit = t.limit(min, max);
- } else {
- limit = t.limit(min);
- }
-
- eatToken();
- } else {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token" + ", given " + tokenToString(token));
- }();
- }
- }
-
- if (elemIndices.length > 0) {
- return t.table(elemType, limit, name, elemIndices);
- } else {
- return t.table(elemType, limit, name);
- }
- }
- /**
- * Parses an import statement
- *
- * WAST:
- *
- * import: ( import <string> <string> <imkind> )
- * imkind: ( func <name>? <func_sig> )
- * ( global <name>? <global_sig> )
- * ( table <name>? <table_sig> )
- * ( memory <name>? <memory_sig> )
- *
- * global_sig: <type> | ( mut <type> )
- */
-
-
- function parseImport() {
- if (token.type !== _tokenizer.tokens.string) {
- throw new Error("Expected a string, " + token.type + " given.");
- }
-
- var moduleName = token.value;
- eatToken();
-
- if (token.type !== _tokenizer.tokens.string) {
- throw new Error("Expected a string, " + token.type + " given.");
- }
-
- var name = token.value;
- eatToken();
- eatTokenOfType(_tokenizer.tokens.openParen);
- var descr;
-
- if (isKeyword(token, _tokenizer.keywords.func)) {
- eatToken(); // keyword
-
- var fnParams = [];
- var fnResult = [];
- var typeRef;
- var fnName = t.identifier(getUniqueName("func"));
-
- if (token.type === _tokenizer.tokens.identifier) {
- fnName = identifierFromToken(token);
- eatToken();
- }
-
- while (token.type === _tokenizer.tokens.openParen) {
- eatToken();
-
- if (lookaheadAndCheck(_tokenizer.keywords.type) === true) {
- eatToken();
- typeRef = parseTypeReference();
- } else if (lookaheadAndCheck(_tokenizer.keywords.param) === true) {
- eatToken();
- fnParams.push.apply(fnParams, _toConsumableArray(parseFuncParam()));
- } else if (lookaheadAndCheck(_tokenizer.keywords.result) === true) {
- eatToken();
- fnResult.push.apply(fnResult, _toConsumableArray(parseFuncResult()));
- } else {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in import of type" + ", given " + tokenToString(token));
- }();
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- if (typeof fnName === "undefined") {
- throw new Error("Imported function must have a name");
- }
-
- descr = t.funcImportDescr(fnName, typeRef !== undefined ? typeRef : t.signature(fnParams, fnResult));
- } else if (isKeyword(token, _tokenizer.keywords.global)) {
- eatToken(); // keyword
-
- if (token.type === _tokenizer.tokens.openParen) {
- eatToken(); // (
-
- eatTokenOfType(_tokenizer.tokens.keyword); // mut keyword
-
- var valtype = token.value;
- eatToken();
- descr = t.globalType(valtype, "var");
- eatTokenOfType(_tokenizer.tokens.closeParen);
- } else {
- var _valtype = token.value;
- eatTokenOfType(_tokenizer.tokens.valtype);
- descr = t.globalType(_valtype, "const");
- }
- } else if (isKeyword(token, _tokenizer.keywords.memory) === true) {
- eatToken(); // Keyword
-
- descr = parseMemory();
- } else if (isKeyword(token, _tokenizer.keywords.table) === true) {
- eatToken(); // Keyword
-
- descr = parseTable();
- } else {
- throw new Error("Unsupported import type: " + tokenToString(token));
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.moduleImport(moduleName, name, descr);
- }
- /**
- * Parses a block instruction
- *
- * WAST:
- *
- * expr: ( block <name>? <block_sig> <instr>* )
- * instr: block <name>? <block_sig> <instr>* end <name>?
- * block_sig : ( result <type>* )*
- *
- */
-
-
- function parseBlock() {
- var label = t.identifier(getUniqueName("block"));
- var blockResult = null;
- var instr = [];
-
- if (token.type === _tokenizer.tokens.identifier) {
- label = identifierFromToken(token);
- eatToken();
- } else {
- label = t.withRaw(label, ""); // preserve anonymous
- }
-
- while (token.type === _tokenizer.tokens.openParen) {
- eatToken();
-
- if (lookaheadAndCheck(_tokenizer.keywords.result) === true) {
- eatToken();
- blockResult = token.value;
- eatToken();
- } else if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword
- ) {
- // Instruction
- instr.push(parseFuncInstr());
- } else {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in block body of type" + ", given " + tokenToString(token));
- }();
- }
-
- maybeIgnoreComment();
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- return t.blockInstruction(label, instr, blockResult);
- }
- /**
- * Parses a if instruction
- *
- * WAST:
- *
- * expr:
- * ( if <name>? <block_sig> ( then <instr>* ) ( else <instr>* )? )
- * ( if <name>? <block_sig> <expr>+ ( then <instr>* ) ( else <instr>* )? )
- *
- * instr:
- * if <name>? <block_sig> <instr>* end <name>?
- * if <name>? <block_sig> <instr>* else <name>? <instr>* end <name>?
- *
- * block_sig : ( result <type>* )*
- *
- */
-
-
- function parseIf() {
- var blockResult = null;
- var label = t.identifier(getUniqueName("if"));
- var testInstrs = [];
- var consequent = [];
- var alternate = [];
-
- if (token.type === _tokenizer.tokens.identifier) {
- label = identifierFromToken(token);
- eatToken();
- } else {
- label = t.withRaw(label, ""); // preserve anonymous
- }
-
- while (token.type === _tokenizer.tokens.openParen) {
- eatToken(); // (
-
- /**
- * Block signature
- */
-
- if (isKeyword(token, _tokenizer.keywords.result) === true) {
- eatToken();
- blockResult = token.value;
- eatTokenOfType(_tokenizer.tokens.valtype);
- eatTokenOfType(_tokenizer.tokens.closeParen);
- continue;
- }
- /**
- * Then
- */
-
-
- if (isKeyword(token, _tokenizer.keywords.then) === true) {
- eatToken(); // then
-
- while (token.type === _tokenizer.tokens.openParen) {
- eatToken(); // Instruction
-
- if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword
- ) {
- consequent.push(parseFuncInstr());
- } else {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in consequent body of type" + ", given " + tokenToString(token));
- }();
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- continue;
- }
- /**
- * Alternate
- */
-
-
- if (isKeyword(token, _tokenizer.keywords.else)) {
- eatToken(); // else
-
- while (token.type === _tokenizer.tokens.openParen) {
- eatToken(); // Instruction
-
- if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword
- ) {
- alternate.push(parseFuncInstr());
- } else {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in alternate body of type" + ", given " + tokenToString(token));
- }();
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- continue;
- }
- /**
- * Test instruction
- */
-
-
- if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword
- ) {
- testInstrs.push(parseFuncInstr());
- eatTokenOfType(_tokenizer.tokens.closeParen);
- continue;
- }
-
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in if body" + ", given " + tokenToString(token));
- }();
- }
-
- return t.ifInstruction(label, testInstrs, blockResult, consequent, alternate);
- }
- /**
- * Parses a loop instruction
- *
- * WAT:
- *
- * blockinstr :: 'loop' I:label rt:resulttype (in:instr*) 'end' id?
- *
- * WAST:
- *
- * instr :: loop <name>? <block_sig> <instr>* end <name>?
- * expr :: ( loop <name>? <block_sig> <instr>* )
- * block_sig :: ( result <type>* )*
- *
- */
-
-
- function parseLoop() {
- var label = t.identifier(getUniqueName("loop"));
- var blockResult;
- var instr = [];
-
- if (token.type === _tokenizer.tokens.identifier) {
- label = identifierFromToken(token);
- eatToken();
- } else {
- label = t.withRaw(label, ""); // preserve anonymous
- }
-
- while (token.type === _tokenizer.tokens.openParen) {
- eatToken();
-
- if (lookaheadAndCheck(_tokenizer.keywords.result) === true) {
- eatToken();
- blockResult = token.value;
- eatToken();
- } else if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword
- ) {
- // Instruction
- instr.push(parseFuncInstr());
- } else {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in loop body" + ", given " + tokenToString(token));
- }();
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- return t.loopInstruction(label, blockResult, instr);
- }
-
- function parseCallIndirect() {
- var typeRef;
- var params = [];
- var results = [];
- var instrs = [];
-
- while (token.type !== _tokenizer.tokens.closeParen) {
- if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.type)) {
- eatToken(); // (
-
- eatToken(); // type
-
- typeRef = parseTypeReference();
- } else if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.param)) {
- eatToken(); // (
-
- eatToken(); // param
-
- /**
- * Params can be empty:
- * (params)`
- */
-
- if (token.type !== _tokenizer.tokens.closeParen) {
- params.push.apply(params, _toConsumableArray(parseFuncParam()));
- }
- } else if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.result)) {
- eatToken(); // (
-
- eatToken(); // result
-
- /**
- * Results can be empty:
- * (result)`
- */
-
- if (token.type !== _tokenizer.tokens.closeParen) {
- results.push.apply(results, _toConsumableArray(parseFuncResult()));
- }
- } else {
- eatTokenOfType(_tokenizer.tokens.openParen);
- instrs.push(parseFuncInstr());
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- return t.callIndirectInstruction(typeRef !== undefined ? typeRef : t.signature(params, results), instrs);
- }
- /**
- * Parses an export instruction
- *
- * WAT:
- *
- * export: ( export <string> <exkind> )
- * exkind: ( func <var> )
- * ( global <var> )
- * ( table <var> )
- * ( memory <var> )
- * var: <nat> | <name>
- *
- */
-
-
- function parseExport() {
- if (token.type !== _tokenizer.tokens.string) {
- throw new Error("Expected string after export, got: " + token.type);
- }
-
- var name = token.value;
- eatToken();
- var moduleExportDescr = parseModuleExportDescr();
- return t.moduleExport(name, moduleExportDescr);
- }
-
- function parseModuleExportDescr() {
- var startLoc = getStartLoc();
- var type = "";
- var index;
- eatTokenOfType(_tokenizer.tokens.openParen);
-
- while (token.type !== _tokenizer.tokens.closeParen) {
- if (isKeyword(token, _tokenizer.keywords.func)) {
- type = "Func";
- eatToken();
- index = parseExportIndex(token);
- } else if (isKeyword(token, _tokenizer.keywords.table)) {
- type = "Table";
- eatToken();
- index = parseExportIndex(token);
- } else if (isKeyword(token, _tokenizer.keywords.global)) {
- type = "Global";
- eatToken();
- index = parseExportIndex(token);
- } else if (isKeyword(token, _tokenizer.keywords.memory)) {
- type = "Memory";
- eatToken();
- index = parseExportIndex(token);
- }
-
- eatToken();
- }
-
- if (type === "") {
- throw new Error("Unknown export type");
- }
-
- if (index === undefined) {
- throw new Error("Exported function must have a name");
- }
-
- var node = t.moduleExportDescr(type, index);
- var endLoc = getEndLoc();
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.withLoc(node, endLoc, startLoc);
- }
-
- function parseModule() {
- var name = null;
- var isBinary = false;
- var isQuote = false;
- var moduleFields = [];
-
- if (token.type === _tokenizer.tokens.identifier) {
- name = token.value;
- eatToken();
- }
-
- if (hasPlugin("wast") && token.type === _tokenizer.tokens.name && token.value === "binary") {
- eatToken();
- isBinary = true;
- }
-
- if (hasPlugin("wast") && token.type === _tokenizer.tokens.name && token.value === "quote") {
- eatToken();
- isQuote = true;
- }
-
- if (isBinary === true) {
- var blob = [];
-
- while (token.type === _tokenizer.tokens.string) {
- blob.push(token.value);
- eatToken();
- maybeIgnoreComment();
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.binaryModule(name, blob);
- }
-
- if (isQuote === true) {
- var string = [];
-
- while (token.type === _tokenizer.tokens.string) {
- string.push(token.value);
- eatToken();
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.quoteModule(name, string);
- }
-
- while (token.type !== _tokenizer.tokens.closeParen) {
- moduleFields.push(walk());
-
- if (state.registredExportedElements.length > 0) {
- state.registredExportedElements.forEach(function (decl) {
- moduleFields.push(t.moduleExport(decl.name, t.moduleExportDescr(decl.exportType, decl.id)));
- });
- state.registredExportedElements = [];
- }
-
- token = tokensList[current];
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.module(name, moduleFields);
- }
- /**
- * Parses the arguments of an instruction
- */
-
-
- function parseFuncInstrArguments(signature) {
- var args = [];
- var namedArgs = {};
- var signaturePtr = 0;
-
- while (token.type === _tokenizer.tokens.name || isKeyword(token, _tokenizer.keywords.offset)) {
- var key = token.value;
- eatToken();
- eatTokenOfType(_tokenizer.tokens.equal);
- var value = void 0;
-
- if (token.type === _tokenizer.tokens.number) {
- value = t.numberLiteralFromRaw(token.value);
- } else {
- throw new Error("Unexpected type for argument: " + token.type);
- }
-
- namedArgs[key] = value;
- eatToken();
- } // $FlowIgnore
-
-
- var signatureLength = signature.vector ? Infinity : signature.length;
-
- while (token.type !== _tokenizer.tokens.closeParen && ( // $FlowIgnore
- token.type === _tokenizer.tokens.openParen || signaturePtr < signatureLength)) {
- if (token.type === _tokenizer.tokens.identifier) {
- args.push(t.identifier(token.value));
- eatToken();
- } else if (token.type === _tokenizer.tokens.valtype) {
- // Handle locals
- args.push(t.valtypeLiteral(token.value));
- eatToken();
- } else if (token.type === _tokenizer.tokens.string) {
- args.push(t.stringLiteral(token.value));
- eatToken();
- } else if (token.type === _tokenizer.tokens.number) {
- args.push( // TODO(sven): refactor the type signature handling
- // https://github.com/xtuc/webassemblyjs/pull/129 is a good start
- t.numberLiteralFromRaw(token.value, // $FlowIgnore
- signature[signaturePtr] || "f64")); // $FlowIgnore
-
- if (!signature.vector) {
- ++signaturePtr;
- }
-
- eatToken();
- } else if (token.type === _tokenizer.tokens.openParen) {
- /**
- * Maybe some nested instructions
- */
- eatToken(); // Instruction
-
- if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword
- ) {
- // $FlowIgnore
- args.push(parseFuncInstr());
- } else {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in nested instruction" + ", given " + tokenToString(token));
- }();
- }
-
- if (token.type === _tokenizer.tokens.closeParen) {
- eatToken();
- }
- } else {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in instruction argument" + ", given " + tokenToString(token));
- }();
- }
- }
-
- return {
- args: args,
- namedArgs: namedArgs
- };
- }
- /**
- * Parses an instruction
- *
- * WAT:
- *
- * instr :: plaininst
- * blockinstr
- *
- * blockinstr :: 'block' I:label rt:resulttype (in:instr*) 'end' id?
- * 'loop' I:label rt:resulttype (in:instr*) 'end' id?
- * 'if' I:label rt:resulttype (in:instr*) 'else' id? (in2:intr*) 'end' id?
- *
- * plaininst :: 'unreachable'
- * 'nop'
- * 'br' l:labelidx
- * 'br_if' l:labelidx
- * 'br_table' l*:vec(labelidx) ln:labelidx
- * 'return'
- * 'call' x:funcidx
- * 'call_indirect' x, I:typeuse
- *
- * WAST:
- *
- * instr:
- * <expr>
- * <op>
- * block <name>? <block_sig> <instr>* end <name>?
- * loop <name>? <block_sig> <instr>* end <name>?
- * if <name>? <block_sig> <instr>* end <name>?
- * if <name>? <block_sig> <instr>* else <name>? <instr>* end <name>?
- *
- * expr:
- * ( <op> )
- * ( <op> <expr>+ )
- * ( block <name>? <block_sig> <instr>* )
- * ( loop <name>? <block_sig> <instr>* )
- * ( if <name>? <block_sig> ( then <instr>* ) ( else <instr>* )? )
- * ( if <name>? <block_sig> <expr>+ ( then <instr>* ) ( else <instr>* )? )
- *
- * op:
- * unreachable
- * nop
- * br <var>
- * br_if <var>
- * br_table <var>+
- * return
- * call <var>
- * call_indirect <func_sig>
- * drop
- * select
- * get_local <var>
- * set_local <var>
- * tee_local <var>
- * get_global <var>
- * set_global <var>
- * <type>.load((8|16|32)_<sign>)? <offset>? <align>?
- * <type>.store(8|16|32)? <offset>? <align>?
- * current_memory
- * grow_memory
- * <type>.const <value>
- * <type>.<unop>
- * <type>.<binop>
- * <type>.<testop>
- * <type>.<relop>
- * <type>.<cvtop>/<type>
- *
- * func_type: ( type <var> )? <param>* <result>*
- */
-
-
- function parseFuncInstr() {
- var startLoc = getStartLoc();
- maybeIgnoreComment();
- /**
- * A simple instruction
- */
-
- if (token.type === _tokenizer.tokens.name || token.type === _tokenizer.tokens.valtype) {
- var _name2 = token.value;
- var object;
- eatToken();
-
- if (token.type === _tokenizer.tokens.dot) {
- object = _name2;
- eatToken();
-
- if (token.type !== _tokenizer.tokens.name) {
- throw new TypeError("Unknown token: " + token.type + ", name expected");
- }
-
- _name2 = token.value;
- eatToken();
- }
-
- if (token.type === _tokenizer.tokens.closeParen) {
- var _endLoc = token.loc.end;
-
- if (typeof object === "undefined") {
- return t.withLoc(t.instruction(_name2), _endLoc, startLoc);
- } else {
- return t.withLoc(t.objectInstruction(_name2, object, []), _endLoc, startLoc);
- }
- }
-
- var signature = t.signatureForOpcode(object || "", _name2);
-
- var _parseFuncInstrArgume = parseFuncInstrArguments(signature),
- _args = _parseFuncInstrArgume.args,
- _namedArgs = _parseFuncInstrArgume.namedArgs;
-
- var endLoc = token.loc.end;
-
- if (typeof object === "undefined") {
- return t.withLoc(t.instruction(_name2, _args, _namedArgs), endLoc, startLoc);
- } else {
- return t.withLoc(t.objectInstruction(_name2, object, _args, _namedArgs), endLoc, startLoc);
- }
- } else if (isKeyword(token, _tokenizer.keywords.loop)) {
- /**
- * Else a instruction with a keyword (loop or block)
- */
- eatToken(); // keyword
-
- return parseLoop();
- } else if (isKeyword(token, _tokenizer.keywords.block)) {
- eatToken(); // keyword
-
- return parseBlock();
- } else if (isKeyword(token, _tokenizer.keywords.call_indirect)) {
- eatToken(); // keyword
-
- return parseCallIndirect();
- } else if (isKeyword(token, _tokenizer.keywords.call)) {
- eatToken(); // keyword
-
- var index;
-
- if (token.type === _tokenizer.tokens.identifier) {
- index = identifierFromToken(token);
- eatToken();
- } else if (token.type === _tokenizer.tokens.number) {
- index = t.indexLiteral(token.value);
- eatToken();
- }
-
- var instrArgs = []; // Nested instruction
-
- while (token.type === _tokenizer.tokens.openParen) {
- eatToken();
- instrArgs.push(parseFuncInstr());
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- if (typeof index === "undefined") {
- throw new Error("Missing argument in call instruciton");
- }
-
- if (instrArgs.length > 0) {
- return t.callInstruction(index, instrArgs);
- } else {
- return t.callInstruction(index);
- }
- } else if (isKeyword(token, _tokenizer.keywords.if)) {
- eatToken(); // Keyword
-
- return parseIf();
- } else if (isKeyword(token, _tokenizer.keywords.module) && hasPlugin("wast")) {
- eatToken(); // In WAST you can have a module as an instruction's argument
- // we will cast it into a instruction to not break the flow
- // $FlowIgnore
-
- var module = parseModule();
- return module;
- } else {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected instruction in function body" + ", given " + tokenToString(token));
- }();
- }
- }
- /*
- * Parses a function
- *
- * WAT:
- *
- * functype :: ( 'func' t1:vec(param) t2:vec(result) )
- * param :: ( 'param' id? t:valtype )
- * result :: ( 'result' t:valtype )
- *
- * WAST:
- *
- * func :: ( func <name>? <func_sig> <local>* <instr>* )
- * ( func <name>? ( export <string> ) <...> )
- * ( func <name>? ( import <string> <string> ) <func_sig> )
- * func_sig :: ( type <var> )? <param>* <result>*
- * param :: ( param <type>* ) | ( param <name> <type> )
- * result :: ( result <type>* )
- * local :: ( local <type>* ) | ( local <name> <type> )
- *
- */
-
-
- function parseFunc() {
- var fnName = t.identifier(getUniqueName("func"));
- var typeRef;
- var fnBody = [];
- var fnParams = [];
- var fnResult = []; // name
-
- if (token.type === _tokenizer.tokens.identifier) {
- fnName = identifierFromToken(token);
- eatToken();
- } else {
- fnName = t.withRaw(fnName, ""); // preserve anonymous
- }
-
- maybeIgnoreComment();
-
- while (token.type === _tokenizer.tokens.openParen || token.type === _tokenizer.tokens.name || token.type === _tokenizer.tokens.valtype) {
- // Instructions without parens
- if (token.type === _tokenizer.tokens.name || token.type === _tokenizer.tokens.valtype) {
- fnBody.push(parseFuncInstr());
- continue;
- }
-
- eatToken();
-
- if (lookaheadAndCheck(_tokenizer.keywords.param) === true) {
- eatToken();
- fnParams.push.apply(fnParams, _toConsumableArray(parseFuncParam()));
- } else if (lookaheadAndCheck(_tokenizer.keywords.result) === true) {
- eatToken();
- fnResult.push.apply(fnResult, _toConsumableArray(parseFuncResult()));
- } else if (lookaheadAndCheck(_tokenizer.keywords.export) === true) {
- eatToken();
- parseFuncExport(fnName);
- } else if (lookaheadAndCheck(_tokenizer.keywords.type) === true) {
- eatToken();
- typeRef = parseTypeReference();
- } else if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword
- ) {
- // Instruction
- fnBody.push(parseFuncInstr());
- } else {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in func body" + ", given " + tokenToString(token));
- }();
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- return t.func(fnName, typeRef !== undefined ? typeRef : t.signature(fnParams, fnResult), fnBody);
- }
- /**
- * Parses shorthand export in func
- *
- * export :: ( export <string> )
- */
-
-
- function parseFuncExport(funcId) {
- if (token.type !== _tokenizer.tokens.string) {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Function export expected a string" + ", given " + tokenToString(token));
- }();
- }
-
- var name = token.value;
- eatToken();
- /**
- * Func export shorthand, we trait it as a syntaxic sugar.
- * A export ModuleField will be added later.
- *
- * We give the anonymous function a generated name and export it.
- */
-
- var id = t.identifier(funcId.value);
- state.registredExportedElements.push({
- exportType: "Func",
- name: name,
- id: id
- });
- }
- /**
- * Parses a type instruction
- *
- * WAST:
- *
- * typedef: ( type <name>? ( func <param>* <result>* ) )
- */
-
-
- function parseType() {
- var id;
- var params = [];
- var result = [];
-
- if (token.type === _tokenizer.tokens.identifier) {
- id = identifierFromToken(token);
- eatToken();
- }
-
- if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.func)) {
- eatToken(); // (
-
- eatToken(); // func
-
- if (token.type === _tokenizer.tokens.closeParen) {
- eatToken(); // function with an empty signature, we can abort here
-
- return t.typeInstruction(id, t.signature([], []));
- }
-
- if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.param)) {
- eatToken(); // (
-
- eatToken(); // param
-
- params = parseFuncParam();
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.result)) {
- eatToken(); // (
-
- eatToken(); // result
-
- result = parseFuncResult();
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- return t.typeInstruction(id, t.signature(params, result));
- }
- /**
- * Parses a function result
- *
- * WAST:
- *
- * result :: ( result <type>* )
- */
-
-
- function parseFuncResult() {
- var results = [];
-
- while (token.type !== _tokenizer.tokens.closeParen) {
- if (token.type !== _tokenizer.tokens.valtype) {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in func result" + ", given " + tokenToString(token));
- }();
- }
-
- var valtype = token.value;
- eatToken();
- results.push(valtype);
- }
-
- return results;
- }
- /**
- * Parses a type reference
- *
- */
-
-
- function parseTypeReference() {
- var ref;
-
- if (token.type === _tokenizer.tokens.identifier) {
- ref = identifierFromToken(token);
- eatToken();
- } else if (token.type === _tokenizer.tokens.number) {
- ref = t.numberLiteralFromRaw(token.value);
- eatToken();
- }
-
- return ref;
- }
- /**
- * Parses a global instruction
- *
- * WAST:
- *
- * global: ( global <name>? <global_sig> <instr>* )
- * ( global <name>? ( export <string> ) <...> )
- * ( global <name>? ( import <string> <string> ) <global_sig> )
- *
- * global_sig: <type> | ( mut <type> )
- *
- */
-
-
- function parseGlobal() {
- var name = t.identifier(getUniqueName("global"));
- var type; // Keep informations in case of a shorthand import
-
- var importing = null;
- maybeIgnoreComment();
-
- if (token.type === _tokenizer.tokens.identifier) {
- name = identifierFromToken(token);
- eatToken();
- } else {
- name = t.withRaw(name, ""); // preserve anonymous
- }
- /**
- * maybe export
- */
-
-
- if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.export)) {
- eatToken(); // (
-
- eatToken(); // export
-
- var exportName = token.value;
- eatTokenOfType(_tokenizer.tokens.string);
- state.registredExportedElements.push({
- exportType: "Global",
- name: exportName,
- id: name
- });
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
- /**
- * maybe import
- */
-
-
- if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.import)) {
- eatToken(); // (
-
- eatToken(); // import
-
- var moduleName = token.value;
- eatTokenOfType(_tokenizer.tokens.string);
- var _name3 = token.value;
- eatTokenOfType(_tokenizer.tokens.string);
- importing = {
- module: moduleName,
- name: _name3,
- descr: undefined
- };
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
- /**
- * global_sig
- */
-
-
- if (token.type === _tokenizer.tokens.valtype) {
- type = t.globalType(token.value, "const");
- eatToken();
- } else if (token.type === _tokenizer.tokens.openParen) {
- eatToken(); // (
-
- if (isKeyword(token, _tokenizer.keywords.mut) === false) {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unsupported global type, expected mut" + ", given " + tokenToString(token));
- }();
- }
-
- eatToken(); // mut
-
- type = t.globalType(token.value, "var");
- eatToken();
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- if (type === undefined) {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Could not determine global type" + ", given " + tokenToString(token));
- }();
- }
-
- maybeIgnoreComment();
- var init = [];
-
- if (importing != null) {
- importing.descr = type;
- init.push(t.moduleImport(importing.module, importing.name, importing.descr));
- }
- /**
- * instr*
- */
-
-
- while (token.type === _tokenizer.tokens.openParen) {
- eatToken();
- init.push(parseFuncInstr());
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- return t.global(type, init, name);
- }
- /**
- * Parses a function param
- *
- * WAST:
- *
- * param :: ( param <type>* ) | ( param <name> <type> )
- */
-
-
- function parseFuncParam() {
- var params = [];
- var id;
- var valtype;
-
- if (token.type === _tokenizer.tokens.identifier) {
- id = token.value;
- eatToken();
- }
-
- if (token.type === _tokenizer.tokens.valtype) {
- valtype = token.value;
- eatToken();
- params.push({
- id: id,
- valtype: valtype
- });
- /**
- * Shorthand notation for multiple anonymous parameters
- * @see https://webassembly.github.io/spec/core/text/types.html#function-types
- * @see https://github.com/xtuc/webassemblyjs/issues/6
- */
-
- if (id === undefined) {
- while (token.type === _tokenizer.tokens.valtype) {
- valtype = token.value;
- eatToken();
- params.push({
- id: undefined,
- valtype: valtype
- });
- }
- }
- } else {// ignore
- }
-
- return params;
- }
- /**
- * Parses an element segments instruction
- *
- * WAST:
- *
- * elem: ( elem <var>? (offset <instr>* ) <var>* )
- * ( elem <var>? <expr> <var>* )
- *
- * var: <nat> | <name>
- */
-
-
- function parseElem() {
- var tableIndex = t.indexLiteral(0);
- var offset = [];
- var funcs = [];
-
- if (token.type === _tokenizer.tokens.identifier) {
- tableIndex = identifierFromToken(token);
- eatToken();
- }
-
- if (token.type === _tokenizer.tokens.number) {
- tableIndex = t.indexLiteral(token.value);
- eatToken();
- }
-
- while (token.type !== _tokenizer.tokens.closeParen) {
- if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.offset)) {
- eatToken(); // (
-
- eatToken(); // offset
-
- while (token.type !== _tokenizer.tokens.closeParen) {
- eatTokenOfType(_tokenizer.tokens.openParen);
- offset.push(parseFuncInstr());
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- } else if (token.type === _tokenizer.tokens.identifier) {
- funcs.push(t.identifier(token.value));
- eatToken();
- } else if (token.type === _tokenizer.tokens.number) {
- funcs.push(t.indexLiteral(token.value));
- eatToken();
- } else if (token.type === _tokenizer.tokens.openParen) {
- eatToken(); // (
-
- offset.push(parseFuncInstr());
- eatTokenOfType(_tokenizer.tokens.closeParen);
- } else {
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unsupported token in elem" + ", given " + tokenToString(token));
- }();
- }
- }
-
- return t.elem(tableIndex, offset, funcs);
- }
- /**
- * Parses the start instruction in a module
- *
- * WAST:
- *
- * start: ( start <var> )
- * var: <nat> | <name>
- *
- * WAT:
- * start ::= ‘(’ ‘start’ x:funcidx ‘)’
- */
-
-
- function parseStart() {
- if (token.type === _tokenizer.tokens.identifier) {
- var index = identifierFromToken(token);
- eatToken();
- return t.start(index);
- }
-
- if (token.type === _tokenizer.tokens.number) {
- var _index2 = t.indexLiteral(token.value);
-
- eatToken();
- return t.start(_index2);
- }
-
- throw new Error("Unknown start, token: " + tokenToString(token));
- }
-
- if (token.type === _tokenizer.tokens.openParen) {
- eatToken();
- var startLoc = getStartLoc();
-
- if (isKeyword(token, _tokenizer.keywords.export)) {
- eatToken();
- var node = parseExport();
-
- var _endLoc2 = getEndLoc();
-
- return t.withLoc(node, _endLoc2, startLoc);
- }
-
- if (isKeyword(token, _tokenizer.keywords.loop)) {
- eatToken();
-
- var _node = parseLoop();
-
- var _endLoc3 = getEndLoc();
-
- return t.withLoc(_node, _endLoc3, startLoc);
- }
-
- if (isKeyword(token, _tokenizer.keywords.func)) {
- eatToken();
-
- var _node2 = parseFunc();
-
- var _endLoc4 = getEndLoc();
-
- maybeIgnoreComment();
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.withLoc(_node2, _endLoc4, startLoc);
- }
-
- if (isKeyword(token, _tokenizer.keywords.module)) {
- eatToken();
-
- var _node3 = parseModule();
-
- var _endLoc5 = getEndLoc();
-
- return t.withLoc(_node3, _endLoc5, startLoc);
- }
-
- if (isKeyword(token, _tokenizer.keywords.import)) {
- eatToken();
-
- var _node4 = parseImport();
-
- var _endLoc6 = getEndLoc();
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.withLoc(_node4, _endLoc6, startLoc);
- }
-
- if (isKeyword(token, _tokenizer.keywords.block)) {
- eatToken();
-
- var _node5 = parseBlock();
-
- var _endLoc7 = getEndLoc();
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.withLoc(_node5, _endLoc7, startLoc);
- }
-
- if (isKeyword(token, _tokenizer.keywords.memory)) {
- eatToken();
-
- var _node6 = parseMemory();
-
- var _endLoc8 = getEndLoc();
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.withLoc(_node6, _endLoc8, startLoc);
- }
-
- if (isKeyword(token, _tokenizer.keywords.data)) {
- eatToken();
-
- var _node7 = parseData();
-
- var _endLoc9 = getEndLoc();
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.withLoc(_node7, _endLoc9, startLoc);
- }
-
- if (isKeyword(token, _tokenizer.keywords.table)) {
- eatToken();
-
- var _node8 = parseTable();
-
- var _endLoc10 = getEndLoc();
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.withLoc(_node8, _endLoc10, startLoc);
- }
-
- if (isKeyword(token, _tokenizer.keywords.global)) {
- eatToken();
-
- var _node9 = parseGlobal();
-
- var _endLoc11 = getEndLoc();
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.withLoc(_node9, _endLoc11, startLoc);
- }
-
- if (isKeyword(token, _tokenizer.keywords.type)) {
- eatToken();
-
- var _node10 = parseType();
-
- var _endLoc12 = getEndLoc();
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.withLoc(_node10, _endLoc12, startLoc);
- }
-
- if (isKeyword(token, _tokenizer.keywords.start)) {
- eatToken();
-
- var _node11 = parseStart();
-
- var _endLoc13 = getEndLoc();
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.withLoc(_node11, _endLoc13, startLoc);
- }
-
- if (isKeyword(token, _tokenizer.keywords.elem)) {
- eatToken();
-
- var _node12 = parseElem();
-
- var _endLoc14 = getEndLoc();
-
- eatTokenOfType(_tokenizer.tokens.closeParen);
- return t.withLoc(_node12, _endLoc14, startLoc);
- }
-
- var instruction = parseFuncInstr();
- var endLoc = getEndLoc();
- maybeIgnoreComment();
-
- if (_typeof(instruction) === "object") {
- if (typeof token !== "undefined") {
- eatTokenOfType(_tokenizer.tokens.closeParen);
- }
-
- return t.withLoc(instruction, endLoc, startLoc);
- }
- }
-
- if (token.type === _tokenizer.tokens.comment) {
- var _startLoc = getStartLoc();
-
- var builder = token.opts.type === "leading" ? t.leadingComment : t.blockComment;
-
- var _node13 = builder(token.value);
-
- eatToken(); // comment
-
- var _endLoc15 = getEndLoc();
-
- return t.withLoc(_node13, _endLoc15, _startLoc);
- }
-
- throw function () {
- return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unknown token" + ", given " + tokenToString(token));
- }();
- }
-
- var body = [];
-
- while (current < tokensList.length) {
- body.push(walk());
- }
-
- return t.program(body);
- }
|