forked from Fediversity/Fediversity
1023 lines
28 KiB
JavaScript
1023 lines
28 KiB
JavaScript
|
/* eslint max-len: 0 */
|
||
|
|
||
|
// A recursive descent parser operates by defining functions for all
|
||
|
// syntactic elements, and recursively calling those, each function
|
||
|
// advancing the input stream and returning an AST node. Precedence
|
||
|
// of constructs (for example, the fact that `!x[1]` means `!(x[1])`
|
||
|
// instead of `(!x)[1]` is handled by the fact that the parser
|
||
|
// function that parses unary prefix operators is called first, and
|
||
|
// in turn calls the function that parses `[]` subscripts — that
|
||
|
// way, it'll receive the node for `x[1]` already parsed, and wraps
|
||
|
// *that* in the unary operator node.
|
||
|
//
|
||
|
// Acorn uses an [operator precedence parser][opp] to handle binary
|
||
|
// operator precedence, because it is much more compact than using
|
||
|
// the technique outlined above, which uses different, nesting
|
||
|
// functions to specify precedence, for all of the ten binary
|
||
|
// precedence levels that JavaScript defines.
|
||
|
//
|
||
|
// [opp]: http://en.wikipedia.org/wiki/Operator-precedence_parser
|
||
|
|
||
|
import {
|
||
|
flowParseArrow,
|
||
|
flowParseFunctionBodyAndFinish,
|
||
|
flowParseMaybeAssign,
|
||
|
flowParseSubscript,
|
||
|
flowParseSubscripts,
|
||
|
flowParseVariance,
|
||
|
flowStartParseAsyncArrowFromCallExpression,
|
||
|
flowStartParseNewArguments,
|
||
|
flowStartParseObjPropValue,
|
||
|
} from "../plugins/flow";
|
||
|
import {jsxParseElement} from "../plugins/jsx/index";
|
||
|
import {typedParseConditional, typedParseParenItem} from "../plugins/types";
|
||
|
import {
|
||
|
tsParseArrow,
|
||
|
tsParseFunctionBodyAndFinish,
|
||
|
tsParseMaybeAssign,
|
||
|
tsParseSubscript,
|
||
|
tsParseType,
|
||
|
tsParseTypeAssertion,
|
||
|
tsStartParseAsyncArrowFromCallExpression,
|
||
|
tsStartParseObjPropValue,
|
||
|
} from "../plugins/typescript";
|
||
|
import {
|
||
|
eat,
|
||
|
IdentifierRole,
|
||
|
lookaheadCharCode,
|
||
|
lookaheadType,
|
||
|
match,
|
||
|
next,
|
||
|
nextTemplateToken,
|
||
|
popTypeContext,
|
||
|
pushTypeContext,
|
||
|
rescan_gt,
|
||
|
retokenizeSlashAsRegex,
|
||
|
} from "../tokenizer/index";
|
||
|
import {ContextualKeyword} from "../tokenizer/keywords";
|
||
|
import {Scope} from "../tokenizer/state";
|
||
|
import {TokenType, TokenType as tt} from "../tokenizer/types";
|
||
|
import {charCodes} from "../util/charcodes";
|
||
|
import {IS_IDENTIFIER_START} from "../util/identifier";
|
||
|
import {getNextContextId, isFlowEnabled, isJSXEnabled, isTypeScriptEnabled, state} from "./base";
|
||
|
import {
|
||
|
markPriorBindingIdentifier,
|
||
|
parseBindingIdentifier,
|
||
|
parseMaybeDefault,
|
||
|
parseRest,
|
||
|
parseSpread,
|
||
|
} from "./lval";
|
||
|
import {
|
||
|
parseBlock,
|
||
|
parseBlockBody,
|
||
|
parseClass,
|
||
|
parseDecorators,
|
||
|
parseFunction,
|
||
|
parseFunctionParams,
|
||
|
} from "./statement";
|
||
|
import {
|
||
|
canInsertSemicolon,
|
||
|
eatContextual,
|
||
|
expect,
|
||
|
expectContextual,
|
||
|
hasFollowingLineBreak,
|
||
|
hasPrecedingLineBreak,
|
||
|
isContextual,
|
||
|
unexpected,
|
||
|
} from "./util";
|
||
|
|
||
|
export class StopState {
|
||
|
|
||
|
constructor(stop) {
|
||
|
this.stop = stop;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// ### Expression parsing
|
||
|
|
||
|
// These nest, from the most general expression type at the top to
|
||
|
// 'atomic', nondivisible expression types at the bottom. Most of
|
||
|
// the functions will simply let the function (s) below them parse,
|
||
|
// and, *if* the syntactic construct they handle is present, wrap
|
||
|
// the AST node that the inner parser gave them in another node.
|
||
|
export function parseExpression(noIn = false) {
|
||
|
parseMaybeAssign(noIn);
|
||
|
if (match(tt.comma)) {
|
||
|
while (eat(tt.comma)) {
|
||
|
parseMaybeAssign(noIn);
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
/**
|
||
|
* noIn is used when parsing a for loop so that we don't interpret a following "in" as the binary
|
||
|
* operatior.
|
||
|
* isWithinParens is used to indicate that we're parsing something that might be a comma expression
|
||
|
* or might be an arrow function or might be a Flow type assertion (which requires explicit parens).
|
||
|
* In these cases, we should allow : and ?: after the initial "left" part.
|
||
|
*/
|
||
|
export function parseMaybeAssign(noIn = false, isWithinParens = false) {
|
||
|
if (isTypeScriptEnabled) {
|
||
|
return tsParseMaybeAssign(noIn, isWithinParens);
|
||
|
} else if (isFlowEnabled) {
|
||
|
return flowParseMaybeAssign(noIn, isWithinParens);
|
||
|
} else {
|
||
|
return baseParseMaybeAssign(noIn, isWithinParens);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Parse an assignment expression. This includes applications of
|
||
|
// operators like `+=`.
|
||
|
// Returns true if the expression was an arrow function.
|
||
|
export function baseParseMaybeAssign(noIn, isWithinParens) {
|
||
|
if (match(tt._yield)) {
|
||
|
parseYield();
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
if (match(tt.parenL) || match(tt.name) || match(tt._yield)) {
|
||
|
state.potentialArrowAt = state.start;
|
||
|
}
|
||
|
|
||
|
const wasArrow = parseMaybeConditional(noIn);
|
||
|
if (isWithinParens) {
|
||
|
parseParenItem();
|
||
|
}
|
||
|
if (state.type & TokenType.IS_ASSIGN) {
|
||
|
next();
|
||
|
parseMaybeAssign(noIn);
|
||
|
return false;
|
||
|
}
|
||
|
return wasArrow;
|
||
|
}
|
||
|
|
||
|
// Parse a ternary conditional (`?:`) operator.
|
||
|
// Returns true if the expression was an arrow function.
|
||
|
function parseMaybeConditional(noIn) {
|
||
|
const wasArrow = parseExprOps(noIn);
|
||
|
if (wasArrow) {
|
||
|
return true;
|
||
|
}
|
||
|
parseConditional(noIn);
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
function parseConditional(noIn) {
|
||
|
if (isTypeScriptEnabled || isFlowEnabled) {
|
||
|
typedParseConditional(noIn);
|
||
|
} else {
|
||
|
baseParseConditional(noIn);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
export function baseParseConditional(noIn) {
|
||
|
if (eat(tt.question)) {
|
||
|
parseMaybeAssign();
|
||
|
expect(tt.colon);
|
||
|
parseMaybeAssign(noIn);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Start the precedence parser.
|
||
|
// Returns true if this was an arrow function
|
||
|
function parseExprOps(noIn) {
|
||
|
const startTokenIndex = state.tokens.length;
|
||
|
const wasArrow = parseMaybeUnary();
|
||
|
if (wasArrow) {
|
||
|
return true;
|
||
|
}
|
||
|
parseExprOp(startTokenIndex, -1, noIn);
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
// Parse binary operators with the operator precedence parsing
|
||
|
// algorithm. `left` is the left-hand side of the operator.
|
||
|
// `minPrec` provides context that allows the function to stop and
|
||
|
// defer further parser to one of its callers when it encounters an
|
||
|
// operator that has a lower precedence than the set it is parsing.
|
||
|
function parseExprOp(startTokenIndex, minPrec, noIn) {
|
||
|
if (
|
||
|
isTypeScriptEnabled &&
|
||
|
(tt._in & TokenType.PRECEDENCE_MASK) > minPrec &&
|
||
|
!hasPrecedingLineBreak() &&
|
||
|
(eatContextual(ContextualKeyword._as) || eatContextual(ContextualKeyword._satisfies))
|
||
|
) {
|
||
|
const oldIsType = pushTypeContext(1);
|
||
|
tsParseType();
|
||
|
popTypeContext(oldIsType);
|
||
|
rescan_gt();
|
||
|
parseExprOp(startTokenIndex, minPrec, noIn);
|
||
|
return;
|
||
|
}
|
||
|
|
||
|
const prec = state.type & TokenType.PRECEDENCE_MASK;
|
||
|
if (prec > 0 && (!noIn || !match(tt._in))) {
|
||
|
if (prec > minPrec) {
|
||
|
const op = state.type;
|
||
|
next();
|
||
|
if (op === tt.nullishCoalescing) {
|
||
|
state.tokens[state.tokens.length - 1].nullishStartIndex = startTokenIndex;
|
||
|
}
|
||
|
|
||
|
const rhsStartTokenIndex = state.tokens.length;
|
||
|
parseMaybeUnary();
|
||
|
// Extend the right operand of this operator if possible.
|
||
|
parseExprOp(rhsStartTokenIndex, op & TokenType.IS_RIGHT_ASSOCIATIVE ? prec - 1 : prec, noIn);
|
||
|
if (op === tt.nullishCoalescing) {
|
||
|
state.tokens[startTokenIndex].numNullishCoalesceStarts++;
|
||
|
state.tokens[state.tokens.length - 1].numNullishCoalesceEnds++;
|
||
|
}
|
||
|
// Continue with any future operator holding this expression as the left operand.
|
||
|
parseExprOp(startTokenIndex, minPrec, noIn);
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Parse unary operators, both prefix and postfix.
|
||
|
// Returns true if this was an arrow function.
|
||
|
export function parseMaybeUnary() {
|
||
|
if (isTypeScriptEnabled && !isJSXEnabled && eat(tt.lessThan)) {
|
||
|
tsParseTypeAssertion();
|
||
|
return false;
|
||
|
}
|
||
|
if (
|
||
|
isContextual(ContextualKeyword._module) &&
|
||
|
lookaheadCharCode() === charCodes.leftCurlyBrace &&
|
||
|
!hasFollowingLineBreak()
|
||
|
) {
|
||
|
parseModuleExpression();
|
||
|
return false;
|
||
|
}
|
||
|
if (state.type & TokenType.IS_PREFIX) {
|
||
|
next();
|
||
|
parseMaybeUnary();
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
const wasArrow = parseExprSubscripts();
|
||
|
if (wasArrow) {
|
||
|
return true;
|
||
|
}
|
||
|
while (state.type & TokenType.IS_POSTFIX && !canInsertSemicolon()) {
|
||
|
// The tokenizer calls everything a preincrement, so make it a postincrement when
|
||
|
// we see it in that context.
|
||
|
if (state.type === tt.preIncDec) {
|
||
|
state.type = tt.postIncDec;
|
||
|
}
|
||
|
next();
|
||
|
}
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
// Parse call, dot, and `[]`-subscript expressions.
|
||
|
// Returns true if this was an arrow function.
|
||
|
export function parseExprSubscripts() {
|
||
|
const startTokenIndex = state.tokens.length;
|
||
|
const wasArrow = parseExprAtom();
|
||
|
if (wasArrow) {
|
||
|
return true;
|
||
|
}
|
||
|
parseSubscripts(startTokenIndex);
|
||
|
// If there was any optional chain operation, the start token would be marked
|
||
|
// as such, so also mark the end now.
|
||
|
if (state.tokens.length > startTokenIndex && state.tokens[startTokenIndex].isOptionalChainStart) {
|
||
|
state.tokens[state.tokens.length - 1].isOptionalChainEnd = true;
|
||
|
}
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
function parseSubscripts(startTokenIndex, noCalls = false) {
|
||
|
if (isFlowEnabled) {
|
||
|
flowParseSubscripts(startTokenIndex, noCalls);
|
||
|
} else {
|
||
|
baseParseSubscripts(startTokenIndex, noCalls);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
export function baseParseSubscripts(startTokenIndex, noCalls = false) {
|
||
|
const stopState = new StopState(false);
|
||
|
do {
|
||
|
parseSubscript(startTokenIndex, noCalls, stopState);
|
||
|
} while (!stopState.stop && !state.error);
|
||
|
}
|
||
|
|
||
|
function parseSubscript(startTokenIndex, noCalls, stopState) {
|
||
|
if (isTypeScriptEnabled) {
|
||
|
tsParseSubscript(startTokenIndex, noCalls, stopState);
|
||
|
} else if (isFlowEnabled) {
|
||
|
flowParseSubscript(startTokenIndex, noCalls, stopState);
|
||
|
} else {
|
||
|
baseParseSubscript(startTokenIndex, noCalls, stopState);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
/** Set 'state.stop = true' to indicate that we should stop parsing subscripts. */
|
||
|
export function baseParseSubscript(
|
||
|
startTokenIndex,
|
||
|
noCalls,
|
||
|
stopState,
|
||
|
) {
|
||
|
if (!noCalls && eat(tt.doubleColon)) {
|
||
|
parseNoCallExpr();
|
||
|
stopState.stop = true;
|
||
|
// Propagate startTokenIndex so that `a::b?.()` will keep `a` as the first token. We may want
|
||
|
// to revisit this in the future when fully supporting bind syntax.
|
||
|
parseSubscripts(startTokenIndex, noCalls);
|
||
|
} else if (match(tt.questionDot)) {
|
||
|
state.tokens[startTokenIndex].isOptionalChainStart = true;
|
||
|
if (noCalls && lookaheadType() === tt.parenL) {
|
||
|
stopState.stop = true;
|
||
|
return;
|
||
|
}
|
||
|
next();
|
||
|
state.tokens[state.tokens.length - 1].subscriptStartIndex = startTokenIndex;
|
||
|
|
||
|
if (eat(tt.bracketL)) {
|
||
|
parseExpression();
|
||
|
expect(tt.bracketR);
|
||
|
} else if (eat(tt.parenL)) {
|
||
|
parseCallExpressionArguments();
|
||
|
} else {
|
||
|
parseMaybePrivateName();
|
||
|
}
|
||
|
} else if (eat(tt.dot)) {
|
||
|
state.tokens[state.tokens.length - 1].subscriptStartIndex = startTokenIndex;
|
||
|
parseMaybePrivateName();
|
||
|
} else if (eat(tt.bracketL)) {
|
||
|
state.tokens[state.tokens.length - 1].subscriptStartIndex = startTokenIndex;
|
||
|
parseExpression();
|
||
|
expect(tt.bracketR);
|
||
|
} else if (!noCalls && match(tt.parenL)) {
|
||
|
if (atPossibleAsync()) {
|
||
|
// We see "async", but it's possible it's a usage of the name "async". Parse as if it's a
|
||
|
// function call, and if we see an arrow later, backtrack and re-parse as a parameter list.
|
||
|
const snapshot = state.snapshot();
|
||
|
const asyncStartTokenIndex = state.tokens.length;
|
||
|
next();
|
||
|
state.tokens[state.tokens.length - 1].subscriptStartIndex = startTokenIndex;
|
||
|
|
||
|
const callContextId = getNextContextId();
|
||
|
|
||
|
state.tokens[state.tokens.length - 1].contextId = callContextId;
|
||
|
parseCallExpressionArguments();
|
||
|
state.tokens[state.tokens.length - 1].contextId = callContextId;
|
||
|
|
||
|
if (shouldParseAsyncArrow()) {
|
||
|
// We hit an arrow, so backtrack and start again parsing function parameters.
|
||
|
state.restoreFromSnapshot(snapshot);
|
||
|
stopState.stop = true;
|
||
|
state.scopeDepth++;
|
||
|
|
||
|
parseFunctionParams();
|
||
|
parseAsyncArrowFromCallExpression(asyncStartTokenIndex);
|
||
|
}
|
||
|
} else {
|
||
|
next();
|
||
|
state.tokens[state.tokens.length - 1].subscriptStartIndex = startTokenIndex;
|
||
|
const callContextId = getNextContextId();
|
||
|
state.tokens[state.tokens.length - 1].contextId = callContextId;
|
||
|
parseCallExpressionArguments();
|
||
|
state.tokens[state.tokens.length - 1].contextId = callContextId;
|
||
|
}
|
||
|
} else if (match(tt.backQuote)) {
|
||
|
// Tagged template expression.
|
||
|
parseTemplate();
|
||
|
} else {
|
||
|
stopState.stop = true;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
export function atPossibleAsync() {
|
||
|
// This was made less strict than the original version to avoid passing around nodes, but it
|
||
|
// should be safe to have rare false positives here.
|
||
|
return (
|
||
|
state.tokens[state.tokens.length - 1].contextualKeyword === ContextualKeyword._async &&
|
||
|
!canInsertSemicolon()
|
||
|
);
|
||
|
}
|
||
|
|
||
|
export function parseCallExpressionArguments() {
|
||
|
let first = true;
|
||
|
while (!eat(tt.parenR) && !state.error) {
|
||
|
if (first) {
|
||
|
first = false;
|
||
|
} else {
|
||
|
expect(tt.comma);
|
||
|
if (eat(tt.parenR)) {
|
||
|
break;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
parseExprListItem(false);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
function shouldParseAsyncArrow() {
|
||
|
return match(tt.colon) || match(tt.arrow);
|
||
|
}
|
||
|
|
||
|
function parseAsyncArrowFromCallExpression(startTokenIndex) {
|
||
|
if (isTypeScriptEnabled) {
|
||
|
tsStartParseAsyncArrowFromCallExpression();
|
||
|
} else if (isFlowEnabled) {
|
||
|
flowStartParseAsyncArrowFromCallExpression();
|
||
|
}
|
||
|
expect(tt.arrow);
|
||
|
parseArrowExpression(startTokenIndex);
|
||
|
}
|
||
|
|
||
|
// Parse a no-call expression (like argument of `new` or `::` operators).
|
||
|
|
||
|
function parseNoCallExpr() {
|
||
|
const startTokenIndex = state.tokens.length;
|
||
|
parseExprAtom();
|
||
|
parseSubscripts(startTokenIndex, true);
|
||
|
}
|
||
|
|
||
|
// Parse an atomic expression — either a single token that is an
|
||
|
// expression, an expression started by a keyword like `function` or
|
||
|
// `new`, or an expression wrapped in punctuation like `()`, `[]`,
|
||
|
// or `{}`.
|
||
|
// Returns true if the parsed expression was an arrow function.
|
||
|
export function parseExprAtom() {
|
||
|
if (eat(tt.modulo)) {
|
||
|
// V8 intrinsic expression. Just parse the identifier, and the function invocation is parsed
|
||
|
// naturally.
|
||
|
parseIdentifier();
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
if (match(tt.jsxText) || match(tt.jsxEmptyText)) {
|
||
|
parseLiteral();
|
||
|
return false;
|
||
|
} else if (match(tt.lessThan) && isJSXEnabled) {
|
||
|
state.type = tt.jsxTagStart;
|
||
|
jsxParseElement();
|
||
|
next();
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
const canBeArrow = state.potentialArrowAt === state.start;
|
||
|
switch (state.type) {
|
||
|
case tt.slash:
|
||
|
case tt.assign:
|
||
|
retokenizeSlashAsRegex();
|
||
|
// Fall through.
|
||
|
|
||
|
case tt._super:
|
||
|
case tt._this:
|
||
|
case tt.regexp:
|
||
|
case tt.num:
|
||
|
case tt.bigint:
|
||
|
case tt.decimal:
|
||
|
case tt.string:
|
||
|
case tt._null:
|
||
|
case tt._true:
|
||
|
case tt._false:
|
||
|
next();
|
||
|
return false;
|
||
|
|
||
|
case tt._import:
|
||
|
next();
|
||
|
if (match(tt.dot)) {
|
||
|
// import.meta
|
||
|
state.tokens[state.tokens.length - 1].type = tt.name;
|
||
|
next();
|
||
|
parseIdentifier();
|
||
|
}
|
||
|
return false;
|
||
|
|
||
|
case tt.name: {
|
||
|
const startTokenIndex = state.tokens.length;
|
||
|
const functionStart = state.start;
|
||
|
const contextualKeyword = state.contextualKeyword;
|
||
|
parseIdentifier();
|
||
|
if (contextualKeyword === ContextualKeyword._await) {
|
||
|
parseAwait();
|
||
|
return false;
|
||
|
} else if (
|
||
|
contextualKeyword === ContextualKeyword._async &&
|
||
|
match(tt._function) &&
|
||
|
!canInsertSemicolon()
|
||
|
) {
|
||
|
next();
|
||
|
parseFunction(functionStart, false);
|
||
|
return false;
|
||
|
} else if (
|
||
|
canBeArrow &&
|
||
|
contextualKeyword === ContextualKeyword._async &&
|
||
|
!canInsertSemicolon() &&
|
||
|
match(tt.name)
|
||
|
) {
|
||
|
state.scopeDepth++;
|
||
|
parseBindingIdentifier(false);
|
||
|
expect(tt.arrow);
|
||
|
// let foo = async bar => {};
|
||
|
parseArrowExpression(startTokenIndex);
|
||
|
return true;
|
||
|
} else if (match(tt._do) && !canInsertSemicolon()) {
|
||
|
next();
|
||
|
parseBlock();
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
if (canBeArrow && !canInsertSemicolon() && match(tt.arrow)) {
|
||
|
state.scopeDepth++;
|
||
|
markPriorBindingIdentifier(false);
|
||
|
expect(tt.arrow);
|
||
|
parseArrowExpression(startTokenIndex);
|
||
|
return true;
|
||
|
}
|
||
|
|
||
|
state.tokens[state.tokens.length - 1].identifierRole = IdentifierRole.Access;
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
case tt._do: {
|
||
|
next();
|
||
|
parseBlock();
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
case tt.parenL: {
|
||
|
const wasArrow = parseParenAndDistinguishExpression(canBeArrow);
|
||
|
return wasArrow;
|
||
|
}
|
||
|
|
||
|
case tt.bracketL:
|
||
|
next();
|
||
|
parseExprList(tt.bracketR, true);
|
||
|
return false;
|
||
|
|
||
|
case tt.braceL:
|
||
|
parseObj(false, false);
|
||
|
return false;
|
||
|
|
||
|
case tt._function:
|
||
|
parseFunctionExpression();
|
||
|
return false;
|
||
|
|
||
|
case tt.at:
|
||
|
parseDecorators();
|
||
|
// Fall through.
|
||
|
|
||
|
case tt._class:
|
||
|
parseClass(false);
|
||
|
return false;
|
||
|
|
||
|
case tt._new:
|
||
|
parseNew();
|
||
|
return false;
|
||
|
|
||
|
case tt.backQuote:
|
||
|
parseTemplate();
|
||
|
return false;
|
||
|
|
||
|
case tt.doubleColon: {
|
||
|
next();
|
||
|
parseNoCallExpr();
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
case tt.hash: {
|
||
|
const code = lookaheadCharCode();
|
||
|
if (IS_IDENTIFIER_START[code] || code === charCodes.backslash) {
|
||
|
parseMaybePrivateName();
|
||
|
} else {
|
||
|
next();
|
||
|
}
|
||
|
// Smart pipeline topic reference.
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
default:
|
||
|
unexpected();
|
||
|
return false;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
function parseMaybePrivateName() {
|
||
|
eat(tt.hash);
|
||
|
parseIdentifier();
|
||
|
}
|
||
|
|
||
|
function parseFunctionExpression() {
|
||
|
const functionStart = state.start;
|
||
|
parseIdentifier();
|
||
|
if (eat(tt.dot)) {
|
||
|
// function.sent
|
||
|
parseIdentifier();
|
||
|
}
|
||
|
parseFunction(functionStart, false);
|
||
|
}
|
||
|
|
||
|
export function parseLiteral() {
|
||
|
next();
|
||
|
}
|
||
|
|
||
|
export function parseParenExpression() {
|
||
|
expect(tt.parenL);
|
||
|
parseExpression();
|
||
|
expect(tt.parenR);
|
||
|
}
|
||
|
|
||
|
// Returns true if this was an arrow expression.
|
||
|
function parseParenAndDistinguishExpression(canBeArrow) {
|
||
|
// Assume this is a normal parenthesized expression, but if we see an arrow, we'll bail and
|
||
|
// start over as a parameter list.
|
||
|
const snapshot = state.snapshot();
|
||
|
|
||
|
const startTokenIndex = state.tokens.length;
|
||
|
expect(tt.parenL);
|
||
|
|
||
|
let first = true;
|
||
|
|
||
|
while (!match(tt.parenR) && !state.error) {
|
||
|
if (first) {
|
||
|
first = false;
|
||
|
} else {
|
||
|
expect(tt.comma);
|
||
|
if (match(tt.parenR)) {
|
||
|
break;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
if (match(tt.ellipsis)) {
|
||
|
parseRest(false /* isBlockScope */);
|
||
|
parseParenItem();
|
||
|
break;
|
||
|
} else {
|
||
|
parseMaybeAssign(false, true);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
expect(tt.parenR);
|
||
|
|
||
|
if (canBeArrow && shouldParseArrow()) {
|
||
|
const wasArrow = parseArrow();
|
||
|
if (wasArrow) {
|
||
|
// It was an arrow function this whole time, so start over and parse it as params so that we
|
||
|
// get proper token annotations.
|
||
|
state.restoreFromSnapshot(snapshot);
|
||
|
state.scopeDepth++;
|
||
|
// Don't specify a context ID because arrow functions don't need a context ID.
|
||
|
parseFunctionParams();
|
||
|
parseArrow();
|
||
|
parseArrowExpression(startTokenIndex);
|
||
|
if (state.error) {
|
||
|
// Nevermind! This must have been something that looks very much like an
|
||
|
// arrow function but where its "parameter list" isn't actually a valid
|
||
|
// parameter list. Force non-arrow parsing.
|
||
|
// See https://github.com/alangpierce/sucrase/issues/666 for an example.
|
||
|
state.restoreFromSnapshot(snapshot);
|
||
|
parseParenAndDistinguishExpression(false);
|
||
|
return false;
|
||
|
}
|
||
|
return true;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
function shouldParseArrow() {
|
||
|
return match(tt.colon) || !canInsertSemicolon();
|
||
|
}
|
||
|
|
||
|
// Returns whether there was an arrow token.
|
||
|
export function parseArrow() {
|
||
|
if (isTypeScriptEnabled) {
|
||
|
return tsParseArrow();
|
||
|
} else if (isFlowEnabled) {
|
||
|
return flowParseArrow();
|
||
|
} else {
|
||
|
return eat(tt.arrow);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
function parseParenItem() {
|
||
|
if (isTypeScriptEnabled || isFlowEnabled) {
|
||
|
typedParseParenItem();
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// New's precedence is slightly tricky. It must allow its argument to
|
||
|
// be a `[]` or dot subscript expression, but not a call — at least,
|
||
|
// not without wrapping it in parentheses. Thus, it uses the noCalls
|
||
|
// argument to parseSubscripts to prevent it from consuming the
|
||
|
// argument list.
|
||
|
function parseNew() {
|
||
|
expect(tt._new);
|
||
|
if (eat(tt.dot)) {
|
||
|
// new.target
|
||
|
parseIdentifier();
|
||
|
return;
|
||
|
}
|
||
|
parseNewCallee();
|
||
|
if (isFlowEnabled) {
|
||
|
flowStartParseNewArguments();
|
||
|
}
|
||
|
if (eat(tt.parenL)) {
|
||
|
parseExprList(tt.parenR);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
function parseNewCallee() {
|
||
|
parseNoCallExpr();
|
||
|
eat(tt.questionDot);
|
||
|
}
|
||
|
|
||
|
export function parseTemplate() {
|
||
|
// Finish `, read quasi
|
||
|
nextTemplateToken();
|
||
|
// Finish quasi, read ${
|
||
|
nextTemplateToken();
|
||
|
while (!match(tt.backQuote) && !state.error) {
|
||
|
expect(tt.dollarBraceL);
|
||
|
parseExpression();
|
||
|
// Finish }, read quasi
|
||
|
nextTemplateToken();
|
||
|
// Finish quasi, read either ${ or `
|
||
|
nextTemplateToken();
|
||
|
}
|
||
|
next();
|
||
|
}
|
||
|
|
||
|
// Parse an object literal or binding pattern.
|
||
|
export function parseObj(isPattern, isBlockScope) {
|
||
|
// Attach a context ID to the object open and close brace and each object key.
|
||
|
const contextId = getNextContextId();
|
||
|
let first = true;
|
||
|
|
||
|
next();
|
||
|
state.tokens[state.tokens.length - 1].contextId = contextId;
|
||
|
|
||
|
while (!eat(tt.braceR) && !state.error) {
|
||
|
if (first) {
|
||
|
first = false;
|
||
|
} else {
|
||
|
expect(tt.comma);
|
||
|
if (eat(tt.braceR)) {
|
||
|
break;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
let isGenerator = false;
|
||
|
if (match(tt.ellipsis)) {
|
||
|
const previousIndex = state.tokens.length;
|
||
|
parseSpread();
|
||
|
if (isPattern) {
|
||
|
// Mark role when the only thing being spread over is an identifier.
|
||
|
if (state.tokens.length === previousIndex + 2) {
|
||
|
markPriorBindingIdentifier(isBlockScope);
|
||
|
}
|
||
|
if (eat(tt.braceR)) {
|
||
|
break;
|
||
|
}
|
||
|
}
|
||
|
continue;
|
||
|
}
|
||
|
|
||
|
if (!isPattern) {
|
||
|
isGenerator = eat(tt.star);
|
||
|
}
|
||
|
|
||
|
if (!isPattern && isContextual(ContextualKeyword._async)) {
|
||
|
if (isGenerator) unexpected();
|
||
|
|
||
|
parseIdentifier();
|
||
|
if (
|
||
|
match(tt.colon) ||
|
||
|
match(tt.parenL) ||
|
||
|
match(tt.braceR) ||
|
||
|
match(tt.eq) ||
|
||
|
match(tt.comma)
|
||
|
) {
|
||
|
// This is a key called "async" rather than an async function.
|
||
|
} else {
|
||
|
if (match(tt.star)) {
|
||
|
next();
|
||
|
isGenerator = true;
|
||
|
}
|
||
|
parsePropertyName(contextId);
|
||
|
}
|
||
|
} else {
|
||
|
parsePropertyName(contextId);
|
||
|
}
|
||
|
|
||
|
parseObjPropValue(isPattern, isBlockScope, contextId);
|
||
|
}
|
||
|
|
||
|
state.tokens[state.tokens.length - 1].contextId = contextId;
|
||
|
}
|
||
|
|
||
|
function isGetterOrSetterMethod(isPattern) {
|
||
|
// We go off of the next and don't bother checking if the node key is actually "get" or "set".
|
||
|
// This lets us avoid generating a node, and should only make the validation worse.
|
||
|
return (
|
||
|
!isPattern &&
|
||
|
(match(tt.string) || // get "string"() {}
|
||
|
match(tt.num) || // get 1() {}
|
||
|
match(tt.bracketL) || // get ["string"]() {}
|
||
|
match(tt.name) || // get foo() {}
|
||
|
!!(state.type & TokenType.IS_KEYWORD)) // get debugger() {}
|
||
|
);
|
||
|
}
|
||
|
|
||
|
// Returns true if this was a method.
|
||
|
function parseObjectMethod(isPattern, objectContextId) {
|
||
|
// We don't need to worry about modifiers because object methods can't have optional bodies, so
|
||
|
// the start will never be used.
|
||
|
const functionStart = state.start;
|
||
|
if (match(tt.parenL)) {
|
||
|
if (isPattern) unexpected();
|
||
|
parseMethod(functionStart, /* isConstructor */ false);
|
||
|
return true;
|
||
|
}
|
||
|
|
||
|
if (isGetterOrSetterMethod(isPattern)) {
|
||
|
parsePropertyName(objectContextId);
|
||
|
parseMethod(functionStart, /* isConstructor */ false);
|
||
|
return true;
|
||
|
}
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
function parseObjectProperty(isPattern, isBlockScope) {
|
||
|
if (eat(tt.colon)) {
|
||
|
if (isPattern) {
|
||
|
parseMaybeDefault(isBlockScope);
|
||
|
} else {
|
||
|
parseMaybeAssign(false);
|
||
|
}
|
||
|
return;
|
||
|
}
|
||
|
|
||
|
// Since there's no colon, we assume this is an object shorthand.
|
||
|
|
||
|
// If we're in a destructuring, we've now discovered that the key was actually an assignee, so
|
||
|
// we need to tag it as a declaration with the appropriate scope. Otherwise, we might need to
|
||
|
// transform it on access, so mark it as a normal object shorthand.
|
||
|
let identifierRole;
|
||
|
if (isPattern) {
|
||
|
if (state.scopeDepth === 0) {
|
||
|
identifierRole = IdentifierRole.ObjectShorthandTopLevelDeclaration;
|
||
|
} else if (isBlockScope) {
|
||
|
identifierRole = IdentifierRole.ObjectShorthandBlockScopedDeclaration;
|
||
|
} else {
|
||
|
identifierRole = IdentifierRole.ObjectShorthandFunctionScopedDeclaration;
|
||
|
}
|
||
|
} else {
|
||
|
identifierRole = IdentifierRole.ObjectShorthand;
|
||
|
}
|
||
|
state.tokens[state.tokens.length - 1].identifierRole = identifierRole;
|
||
|
|
||
|
// Regardless of whether we know this to be a pattern or if we're in an ambiguous context, allow
|
||
|
// parsing as if there's a default value.
|
||
|
parseMaybeDefault(isBlockScope, true);
|
||
|
}
|
||
|
|
||
|
function parseObjPropValue(
|
||
|
isPattern,
|
||
|
isBlockScope,
|
||
|
objectContextId,
|
||
|
) {
|
||
|
if (isTypeScriptEnabled) {
|
||
|
tsStartParseObjPropValue();
|
||
|
} else if (isFlowEnabled) {
|
||
|
flowStartParseObjPropValue();
|
||
|
}
|
||
|
const wasMethod = parseObjectMethod(isPattern, objectContextId);
|
||
|
if (!wasMethod) {
|
||
|
parseObjectProperty(isPattern, isBlockScope);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
export function parsePropertyName(objectContextId) {
|
||
|
if (isFlowEnabled) {
|
||
|
flowParseVariance();
|
||
|
}
|
||
|
if (eat(tt.bracketL)) {
|
||
|
state.tokens[state.tokens.length - 1].contextId = objectContextId;
|
||
|
parseMaybeAssign();
|
||
|
expect(tt.bracketR);
|
||
|
state.tokens[state.tokens.length - 1].contextId = objectContextId;
|
||
|
} else {
|
||
|
if (match(tt.num) || match(tt.string) || match(tt.bigint) || match(tt.decimal)) {
|
||
|
parseExprAtom();
|
||
|
} else {
|
||
|
parseMaybePrivateName();
|
||
|
}
|
||
|
|
||
|
state.tokens[state.tokens.length - 1].identifierRole = IdentifierRole.ObjectKey;
|
||
|
state.tokens[state.tokens.length - 1].contextId = objectContextId;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Parse object or class method.
|
||
|
export function parseMethod(functionStart, isConstructor) {
|
||
|
const funcContextId = getNextContextId();
|
||
|
|
||
|
state.scopeDepth++;
|
||
|
const startTokenIndex = state.tokens.length;
|
||
|
const allowModifiers = isConstructor; // For TypeScript parameter properties
|
||
|
parseFunctionParams(allowModifiers, funcContextId);
|
||
|
parseFunctionBodyAndFinish(functionStart, funcContextId);
|
||
|
const endTokenIndex = state.tokens.length;
|
||
|
state.scopes.push(new Scope(startTokenIndex, endTokenIndex, true));
|
||
|
state.scopeDepth--;
|
||
|
}
|
||
|
|
||
|
// Parse arrow function expression.
|
||
|
// If the parameters are provided, they will be converted to an
|
||
|
// assignable list.
|
||
|
export function parseArrowExpression(startTokenIndex) {
|
||
|
parseFunctionBody(true);
|
||
|
const endTokenIndex = state.tokens.length;
|
||
|
state.scopes.push(new Scope(startTokenIndex, endTokenIndex, true));
|
||
|
state.scopeDepth--;
|
||
|
}
|
||
|
|
||
|
export function parseFunctionBodyAndFinish(functionStart, funcContextId = 0) {
|
||
|
if (isTypeScriptEnabled) {
|
||
|
tsParseFunctionBodyAndFinish(functionStart, funcContextId);
|
||
|
} else if (isFlowEnabled) {
|
||
|
flowParseFunctionBodyAndFinish(funcContextId);
|
||
|
} else {
|
||
|
parseFunctionBody(false, funcContextId);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
export function parseFunctionBody(allowExpression, funcContextId = 0) {
|
||
|
const isExpression = allowExpression && !match(tt.braceL);
|
||
|
|
||
|
if (isExpression) {
|
||
|
parseMaybeAssign();
|
||
|
} else {
|
||
|
parseBlock(true /* isFunctionScope */, funcContextId);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Parses a comma-separated list of expressions, and returns them as
|
||
|
// an array. `close` is the token type that ends the list, and
|
||
|
// `allowEmpty` can be turned on to allow subsequent commas with
|
||
|
// nothing in between them to be parsed as `null` (which is needed
|
||
|
// for array literals).
|
||
|
|
||
|
function parseExprList(close, allowEmpty = false) {
|
||
|
let first = true;
|
||
|
while (!eat(close) && !state.error) {
|
||
|
if (first) {
|
||
|
first = false;
|
||
|
} else {
|
||
|
expect(tt.comma);
|
||
|
if (eat(close)) break;
|
||
|
}
|
||
|
parseExprListItem(allowEmpty);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
function parseExprListItem(allowEmpty) {
|
||
|
if (allowEmpty && match(tt.comma)) {
|
||
|
// Empty item; nothing more to parse for this item.
|
||
|
} else if (match(tt.ellipsis)) {
|
||
|
parseSpread();
|
||
|
parseParenItem();
|
||
|
} else if (match(tt.question)) {
|
||
|
// Partial function application proposal.
|
||
|
next();
|
||
|
} else {
|
||
|
parseMaybeAssign(false, true);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Parse the next token as an identifier.
|
||
|
export function parseIdentifier() {
|
||
|
next();
|
||
|
state.tokens[state.tokens.length - 1].type = tt.name;
|
||
|
}
|
||
|
|
||
|
// Parses await expression inside async function.
|
||
|
function parseAwait() {
|
||
|
parseMaybeUnary();
|
||
|
}
|
||
|
|
||
|
// Parses yield expression inside generator.
|
||
|
function parseYield() {
|
||
|
next();
|
||
|
if (!match(tt.semi) && !canInsertSemicolon()) {
|
||
|
eat(tt.star);
|
||
|
parseMaybeAssign();
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// https://github.com/tc39/proposal-js-module-blocks
|
||
|
function parseModuleExpression() {
|
||
|
expectContextual(ContextualKeyword._module);
|
||
|
expect(tt.braceL);
|
||
|
// For now, just call parseBlockBody to parse the block. In the future when we
|
||
|
// implement full support, we'll want to emit scopes and possibly other
|
||
|
// information.
|
||
|
parseBlockBody(tt.braceR);
|
||
|
}
|