 85bf1341f3
			
		
	
	85bf1341f3
	
	
	
		
			
			Frontend Enhancements: - Complete React TypeScript frontend with modern UI components - Distributed workflows management interface with real-time updates - Socket.IO integration for live agent status monitoring - Agent management dashboard with cluster visualization - Project management interface with metrics and task tracking - Responsive design with proper error handling and loading states Backend Infrastructure: - Distributed coordinator for multi-agent workflow orchestration - Cluster management API with comprehensive agent operations - Enhanced database models for agents and projects - Project service for filesystem-based project discovery - Performance monitoring and metrics collection - Comprehensive API documentation and error handling Documentation: - Complete distributed development guide (README_DISTRIBUTED.md) - Comprehensive development report with architecture insights - System configuration templates and deployment guides The platform now provides a complete web interface for managing the distributed AI cluster with real-time monitoring, workflow orchestration, and agent coordination capabilities. 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude <noreply@anthropic.com>
		
			
				
	
	
		
			1023 lines
		
	
	
		
			28 KiB
		
	
	
	
		
			JavaScript
		
	
	
	
	
	
			
		
		
	
	
			1023 lines
		
	
	
		
			28 KiB
		
	
	
	
		
			JavaScript
		
	
	
	
	
	
| /* eslint max-len: 0 */
 | |
| 
 | |
| // A recursive descent parser operates by defining functions for all
 | |
| // syntactic elements, and recursively calling those, each function
 | |
| // advancing the input stream and returning an AST node. Precedence
 | |
| // of constructs (for example, the fact that `!x[1]` means `!(x[1])`
 | |
| // instead of `(!x)[1]` is handled by the fact that the parser
 | |
| // function that parses unary prefix operators is called first, and
 | |
| // in turn calls the function that parses `[]` subscripts — that
 | |
| // way, it'll receive the node for `x[1]` already parsed, and wraps
 | |
| // *that* in the unary operator node.
 | |
| //
 | |
| // Acorn uses an [operator precedence parser][opp] to handle binary
 | |
| // operator precedence, because it is much more compact than using
 | |
| // the technique outlined above, which uses different, nesting
 | |
| // functions to specify precedence, for all of the ten binary
 | |
| // precedence levels that JavaScript defines.
 | |
| //
 | |
| // [opp]: http://en.wikipedia.org/wiki/Operator-precedence_parser
 | |
| 
 | |
| import {
 | |
|   flowParseArrow,
 | |
|   flowParseFunctionBodyAndFinish,
 | |
|   flowParseMaybeAssign,
 | |
|   flowParseSubscript,
 | |
|   flowParseSubscripts,
 | |
|   flowParseVariance,
 | |
|   flowStartParseAsyncArrowFromCallExpression,
 | |
|   flowStartParseNewArguments,
 | |
|   flowStartParseObjPropValue,
 | |
| } from "../plugins/flow";
 | |
| import {jsxParseElement} from "../plugins/jsx/index";
 | |
| import {typedParseConditional, typedParseParenItem} from "../plugins/types";
 | |
| import {
 | |
|   tsParseArrow,
 | |
|   tsParseFunctionBodyAndFinish,
 | |
|   tsParseMaybeAssign,
 | |
|   tsParseSubscript,
 | |
|   tsParseType,
 | |
|   tsParseTypeAssertion,
 | |
|   tsStartParseAsyncArrowFromCallExpression,
 | |
|   tsStartParseObjPropValue,
 | |
| } from "../plugins/typescript";
 | |
| import {
 | |
|   eat,
 | |
|   IdentifierRole,
 | |
|   lookaheadCharCode,
 | |
|   lookaheadType,
 | |
|   match,
 | |
|   next,
 | |
|   nextTemplateToken,
 | |
|   popTypeContext,
 | |
|   pushTypeContext,
 | |
|   rescan_gt,
 | |
|   retokenizeSlashAsRegex,
 | |
| } from "../tokenizer/index";
 | |
| import {ContextualKeyword} from "../tokenizer/keywords";
 | |
| import {Scope} from "../tokenizer/state";
 | |
| import {TokenType, TokenType as tt} from "../tokenizer/types";
 | |
| import {charCodes} from "../util/charcodes";
 | |
| import {IS_IDENTIFIER_START} from "../util/identifier";
 | |
| import {getNextContextId, isFlowEnabled, isJSXEnabled, isTypeScriptEnabled, state} from "./base";
 | |
| import {
 | |
|   markPriorBindingIdentifier,
 | |
|   parseBindingIdentifier,
 | |
|   parseMaybeDefault,
 | |
|   parseRest,
 | |
|   parseSpread,
 | |
| } from "./lval";
 | |
| import {
 | |
|   parseBlock,
 | |
|   parseBlockBody,
 | |
|   parseClass,
 | |
|   parseDecorators,
 | |
|   parseFunction,
 | |
|   parseFunctionParams,
 | |
| } from "./statement";
 | |
| import {
 | |
|   canInsertSemicolon,
 | |
|   eatContextual,
 | |
|   expect,
 | |
|   expectContextual,
 | |
|   hasFollowingLineBreak,
 | |
|   hasPrecedingLineBreak,
 | |
|   isContextual,
 | |
|   unexpected,
 | |
| } from "./util";
 | |
| 
 | |
| export class StopState {
 | |
|   
 | |
|   constructor(stop) {
 | |
|     this.stop = stop;
 | |
|   }
 | |
| }
 | |
| 
 | |
| // ### Expression parsing
 | |
| 
 | |
| // These nest, from the most general expression type at the top to
 | |
| // 'atomic', nondivisible expression types at the bottom. Most of
 | |
| // the functions will simply let the function (s) below them parse,
 | |
| // and, *if* the syntactic construct they handle is present, wrap
 | |
| // the AST node that the inner parser gave them in another node.
 | |
| export function parseExpression(noIn = false) {
 | |
|   parseMaybeAssign(noIn);
 | |
|   if (match(tt.comma)) {
 | |
|     while (eat(tt.comma)) {
 | |
|       parseMaybeAssign(noIn);
 | |
|     }
 | |
|   }
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * noIn is used when parsing a for loop so that we don't interpret a following "in" as the binary
 | |
|  * operatior.
 | |
|  * isWithinParens is used to indicate that we're parsing something that might be a comma expression
 | |
|  * or might be an arrow function or might be a Flow type assertion (which requires explicit parens).
 | |
|  * In these cases, we should allow : and ?: after the initial "left" part.
 | |
|  */
 | |
| export function parseMaybeAssign(noIn = false, isWithinParens = false) {
 | |
|   if (isTypeScriptEnabled) {
 | |
|     return tsParseMaybeAssign(noIn, isWithinParens);
 | |
|   } else if (isFlowEnabled) {
 | |
|     return flowParseMaybeAssign(noIn, isWithinParens);
 | |
|   } else {
 | |
|     return baseParseMaybeAssign(noIn, isWithinParens);
 | |
|   }
 | |
| }
 | |
| 
 | |
| // Parse an assignment expression. This includes applications of
 | |
| // operators like `+=`.
 | |
| // Returns true if the expression was an arrow function.
 | |
| export function baseParseMaybeAssign(noIn, isWithinParens) {
 | |
|   if (match(tt._yield)) {
 | |
|     parseYield();
 | |
|     return false;
 | |
|   }
 | |
| 
 | |
|   if (match(tt.parenL) || match(tt.name) || match(tt._yield)) {
 | |
|     state.potentialArrowAt = state.start;
 | |
|   }
 | |
| 
 | |
|   const wasArrow = parseMaybeConditional(noIn);
 | |
|   if (isWithinParens) {
 | |
|     parseParenItem();
 | |
|   }
 | |
|   if (state.type & TokenType.IS_ASSIGN) {
 | |
|     next();
 | |
|     parseMaybeAssign(noIn);
 | |
|     return false;
 | |
|   }
 | |
|   return wasArrow;
 | |
| }
 | |
| 
 | |
| // Parse a ternary conditional (`?:`) operator.
 | |
| // Returns true if the expression was an arrow function.
 | |
| function parseMaybeConditional(noIn) {
 | |
|   const wasArrow = parseExprOps(noIn);
 | |
|   if (wasArrow) {
 | |
|     return true;
 | |
|   }
 | |
|   parseConditional(noIn);
 | |
|   return false;
 | |
| }
 | |
| 
 | |
| function parseConditional(noIn) {
 | |
|   if (isTypeScriptEnabled || isFlowEnabled) {
 | |
|     typedParseConditional(noIn);
 | |
|   } else {
 | |
|     baseParseConditional(noIn);
 | |
|   }
 | |
| }
 | |
| 
 | |
| export function baseParseConditional(noIn) {
 | |
|   if (eat(tt.question)) {
 | |
|     parseMaybeAssign();
 | |
|     expect(tt.colon);
 | |
|     parseMaybeAssign(noIn);
 | |
|   }
 | |
| }
 | |
| 
 | |
| // Start the precedence parser.
 | |
| // Returns true if this was an arrow function
 | |
| function parseExprOps(noIn) {
 | |
|   const startTokenIndex = state.tokens.length;
 | |
|   const wasArrow = parseMaybeUnary();
 | |
|   if (wasArrow) {
 | |
|     return true;
 | |
|   }
 | |
|   parseExprOp(startTokenIndex, -1, noIn);
 | |
|   return false;
 | |
| }
 | |
| 
 | |
| // Parse binary operators with the operator precedence parsing
 | |
| // algorithm. `left` is the left-hand side of the operator.
 | |
| // `minPrec` provides context that allows the function to stop and
 | |
| // defer further parser to one of its callers when it encounters an
 | |
| // operator that has a lower precedence than the set it is parsing.
 | |
| function parseExprOp(startTokenIndex, minPrec, noIn) {
 | |
|   if (
 | |
|     isTypeScriptEnabled &&
 | |
|     (tt._in & TokenType.PRECEDENCE_MASK) > minPrec &&
 | |
|     !hasPrecedingLineBreak() &&
 | |
|     (eatContextual(ContextualKeyword._as) || eatContextual(ContextualKeyword._satisfies))
 | |
|   ) {
 | |
|     const oldIsType = pushTypeContext(1);
 | |
|     tsParseType();
 | |
|     popTypeContext(oldIsType);
 | |
|     rescan_gt();
 | |
|     parseExprOp(startTokenIndex, minPrec, noIn);
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   const prec = state.type & TokenType.PRECEDENCE_MASK;
 | |
|   if (prec > 0 && (!noIn || !match(tt._in))) {
 | |
|     if (prec > minPrec) {
 | |
|       const op = state.type;
 | |
|       next();
 | |
|       if (op === tt.nullishCoalescing) {
 | |
|         state.tokens[state.tokens.length - 1].nullishStartIndex = startTokenIndex;
 | |
|       }
 | |
| 
 | |
|       const rhsStartTokenIndex = state.tokens.length;
 | |
|       parseMaybeUnary();
 | |
|       // Extend the right operand of this operator if possible.
 | |
|       parseExprOp(rhsStartTokenIndex, op & TokenType.IS_RIGHT_ASSOCIATIVE ? prec - 1 : prec, noIn);
 | |
|       if (op === tt.nullishCoalescing) {
 | |
|         state.tokens[startTokenIndex].numNullishCoalesceStarts++;
 | |
|         state.tokens[state.tokens.length - 1].numNullishCoalesceEnds++;
 | |
|       }
 | |
|       // Continue with any future operator holding this expression as the left operand.
 | |
|       parseExprOp(startTokenIndex, minPrec, noIn);
 | |
|     }
 | |
|   }
 | |
| }
 | |
| 
 | |
| // Parse unary operators, both prefix and postfix.
 | |
| // Returns true if this was an arrow function.
 | |
| export function parseMaybeUnary() {
 | |
|   if (isTypeScriptEnabled && !isJSXEnabled && eat(tt.lessThan)) {
 | |
|     tsParseTypeAssertion();
 | |
|     return false;
 | |
|   }
 | |
|   if (
 | |
|     isContextual(ContextualKeyword._module) &&
 | |
|     lookaheadCharCode() === charCodes.leftCurlyBrace &&
 | |
|     !hasFollowingLineBreak()
 | |
|   ) {
 | |
|     parseModuleExpression();
 | |
|     return false;
 | |
|   }
 | |
|   if (state.type & TokenType.IS_PREFIX) {
 | |
|     next();
 | |
|     parseMaybeUnary();
 | |
|     return false;
 | |
|   }
 | |
| 
 | |
|   const wasArrow = parseExprSubscripts();
 | |
|   if (wasArrow) {
 | |
|     return true;
 | |
|   }
 | |
|   while (state.type & TokenType.IS_POSTFIX && !canInsertSemicolon()) {
 | |
|     // The tokenizer calls everything a preincrement, so make it a postincrement when
 | |
|     // we see it in that context.
 | |
|     if (state.type === tt.preIncDec) {
 | |
|       state.type = tt.postIncDec;
 | |
|     }
 | |
|     next();
 | |
|   }
 | |
|   return false;
 | |
| }
 | |
| 
 | |
| // Parse call, dot, and `[]`-subscript expressions.
 | |
| // Returns true if this was an arrow function.
 | |
| export function parseExprSubscripts() {
 | |
|   const startTokenIndex = state.tokens.length;
 | |
|   const wasArrow = parseExprAtom();
 | |
|   if (wasArrow) {
 | |
|     return true;
 | |
|   }
 | |
|   parseSubscripts(startTokenIndex);
 | |
|   // If there was any optional chain operation, the start token would be marked
 | |
|   // as such, so also mark the end now.
 | |
|   if (state.tokens.length > startTokenIndex && state.tokens[startTokenIndex].isOptionalChainStart) {
 | |
|     state.tokens[state.tokens.length - 1].isOptionalChainEnd = true;
 | |
|   }
 | |
|   return false;
 | |
| }
 | |
| 
 | |
| function parseSubscripts(startTokenIndex, noCalls = false) {
 | |
|   if (isFlowEnabled) {
 | |
|     flowParseSubscripts(startTokenIndex, noCalls);
 | |
|   } else {
 | |
|     baseParseSubscripts(startTokenIndex, noCalls);
 | |
|   }
 | |
| }
 | |
| 
 | |
| export function baseParseSubscripts(startTokenIndex, noCalls = false) {
 | |
|   const stopState = new StopState(false);
 | |
|   do {
 | |
|     parseSubscript(startTokenIndex, noCalls, stopState);
 | |
|   } while (!stopState.stop && !state.error);
 | |
| }
 | |
| 
 | |
| function parseSubscript(startTokenIndex, noCalls, stopState) {
 | |
|   if (isTypeScriptEnabled) {
 | |
|     tsParseSubscript(startTokenIndex, noCalls, stopState);
 | |
|   } else if (isFlowEnabled) {
 | |
|     flowParseSubscript(startTokenIndex, noCalls, stopState);
 | |
|   } else {
 | |
|     baseParseSubscript(startTokenIndex, noCalls, stopState);
 | |
|   }
 | |
| }
 | |
| 
 | |
| /** Set 'state.stop = true' to indicate that we should stop parsing subscripts. */
 | |
| export function baseParseSubscript(
 | |
|   startTokenIndex,
 | |
|   noCalls,
 | |
|   stopState,
 | |
| ) {
 | |
|   if (!noCalls && eat(tt.doubleColon)) {
 | |
|     parseNoCallExpr();
 | |
|     stopState.stop = true;
 | |
|     // Propagate startTokenIndex so that `a::b?.()` will keep `a` as the first token. We may want
 | |
|     // to revisit this in the future when fully supporting bind syntax.
 | |
|     parseSubscripts(startTokenIndex, noCalls);
 | |
|   } else if (match(tt.questionDot)) {
 | |
|     state.tokens[startTokenIndex].isOptionalChainStart = true;
 | |
|     if (noCalls && lookaheadType() === tt.parenL) {
 | |
|       stopState.stop = true;
 | |
|       return;
 | |
|     }
 | |
|     next();
 | |
|     state.tokens[state.tokens.length - 1].subscriptStartIndex = startTokenIndex;
 | |
| 
 | |
|     if (eat(tt.bracketL)) {
 | |
|       parseExpression();
 | |
|       expect(tt.bracketR);
 | |
|     } else if (eat(tt.parenL)) {
 | |
|       parseCallExpressionArguments();
 | |
|     } else {
 | |
|       parseMaybePrivateName();
 | |
|     }
 | |
|   } else if (eat(tt.dot)) {
 | |
|     state.tokens[state.tokens.length - 1].subscriptStartIndex = startTokenIndex;
 | |
|     parseMaybePrivateName();
 | |
|   } else if (eat(tt.bracketL)) {
 | |
|     state.tokens[state.tokens.length - 1].subscriptStartIndex = startTokenIndex;
 | |
|     parseExpression();
 | |
|     expect(tt.bracketR);
 | |
|   } else if (!noCalls && match(tt.parenL)) {
 | |
|     if (atPossibleAsync()) {
 | |
|       // We see "async", but it's possible it's a usage of the name "async". Parse as if it's a
 | |
|       // function call, and if we see an arrow later, backtrack and re-parse as a parameter list.
 | |
|       const snapshot = state.snapshot();
 | |
|       const asyncStartTokenIndex = state.tokens.length;
 | |
|       next();
 | |
|       state.tokens[state.tokens.length - 1].subscriptStartIndex = startTokenIndex;
 | |
| 
 | |
|       const callContextId = getNextContextId();
 | |
| 
 | |
|       state.tokens[state.tokens.length - 1].contextId = callContextId;
 | |
|       parseCallExpressionArguments();
 | |
|       state.tokens[state.tokens.length - 1].contextId = callContextId;
 | |
| 
 | |
|       if (shouldParseAsyncArrow()) {
 | |
|         // We hit an arrow, so backtrack and start again parsing function parameters.
 | |
|         state.restoreFromSnapshot(snapshot);
 | |
|         stopState.stop = true;
 | |
|         state.scopeDepth++;
 | |
| 
 | |
|         parseFunctionParams();
 | |
|         parseAsyncArrowFromCallExpression(asyncStartTokenIndex);
 | |
|       }
 | |
|     } else {
 | |
|       next();
 | |
|       state.tokens[state.tokens.length - 1].subscriptStartIndex = startTokenIndex;
 | |
|       const callContextId = getNextContextId();
 | |
|       state.tokens[state.tokens.length - 1].contextId = callContextId;
 | |
|       parseCallExpressionArguments();
 | |
|       state.tokens[state.tokens.length - 1].contextId = callContextId;
 | |
|     }
 | |
|   } else if (match(tt.backQuote)) {
 | |
|     // Tagged template expression.
 | |
|     parseTemplate();
 | |
|   } else {
 | |
|     stopState.stop = true;
 | |
|   }
 | |
| }
 | |
| 
 | |
| export function atPossibleAsync() {
 | |
|   // This was made less strict than the original version to avoid passing around nodes, but it
 | |
|   // should be safe to have rare false positives here.
 | |
|   return (
 | |
|     state.tokens[state.tokens.length - 1].contextualKeyword === ContextualKeyword._async &&
 | |
|     !canInsertSemicolon()
 | |
|   );
 | |
| }
 | |
| 
 | |
| export function parseCallExpressionArguments() {
 | |
|   let first = true;
 | |
|   while (!eat(tt.parenR) && !state.error) {
 | |
|     if (first) {
 | |
|       first = false;
 | |
|     } else {
 | |
|       expect(tt.comma);
 | |
|       if (eat(tt.parenR)) {
 | |
|         break;
 | |
|       }
 | |
|     }
 | |
| 
 | |
|     parseExprListItem(false);
 | |
|   }
 | |
| }
 | |
| 
 | |
| function shouldParseAsyncArrow() {
 | |
|   return match(tt.colon) || match(tt.arrow);
 | |
| }
 | |
| 
 | |
| function parseAsyncArrowFromCallExpression(startTokenIndex) {
 | |
|   if (isTypeScriptEnabled) {
 | |
|     tsStartParseAsyncArrowFromCallExpression();
 | |
|   } else if (isFlowEnabled) {
 | |
|     flowStartParseAsyncArrowFromCallExpression();
 | |
|   }
 | |
|   expect(tt.arrow);
 | |
|   parseArrowExpression(startTokenIndex);
 | |
| }
 | |
| 
 | |
| // Parse a no-call expression (like argument of `new` or `::` operators).
 | |
| 
 | |
| function parseNoCallExpr() {
 | |
|   const startTokenIndex = state.tokens.length;
 | |
|   parseExprAtom();
 | |
|   parseSubscripts(startTokenIndex, true);
 | |
| }
 | |
| 
 | |
| // Parse an atomic expression — either a single token that is an
 | |
| // expression, an expression started by a keyword like `function` or
 | |
| // `new`, or an expression wrapped in punctuation like `()`, `[]`,
 | |
| // or `{}`.
 | |
| // Returns true if the parsed expression was an arrow function.
 | |
| export function parseExprAtom() {
 | |
|   if (eat(tt.modulo)) {
 | |
|     // V8 intrinsic expression. Just parse the identifier, and the function invocation is parsed
 | |
|     // naturally.
 | |
|     parseIdentifier();
 | |
|     return false;
 | |
|   }
 | |
| 
 | |
|   if (match(tt.jsxText) || match(tt.jsxEmptyText)) {
 | |
|     parseLiteral();
 | |
|     return false;
 | |
|   } else if (match(tt.lessThan) && isJSXEnabled) {
 | |
|     state.type = tt.jsxTagStart;
 | |
|     jsxParseElement();
 | |
|     next();
 | |
|     return false;
 | |
|   }
 | |
| 
 | |
|   const canBeArrow = state.potentialArrowAt === state.start;
 | |
|   switch (state.type) {
 | |
|     case tt.slash:
 | |
|     case tt.assign:
 | |
|       retokenizeSlashAsRegex();
 | |
|     // Fall through.
 | |
| 
 | |
|     case tt._super:
 | |
|     case tt._this:
 | |
|     case tt.regexp:
 | |
|     case tt.num:
 | |
|     case tt.bigint:
 | |
|     case tt.decimal:
 | |
|     case tt.string:
 | |
|     case tt._null:
 | |
|     case tt._true:
 | |
|     case tt._false:
 | |
|       next();
 | |
|       return false;
 | |
| 
 | |
|     case tt._import:
 | |
|       next();
 | |
|       if (match(tt.dot)) {
 | |
|         // import.meta
 | |
|         state.tokens[state.tokens.length - 1].type = tt.name;
 | |
|         next();
 | |
|         parseIdentifier();
 | |
|       }
 | |
|       return false;
 | |
| 
 | |
|     case tt.name: {
 | |
|       const startTokenIndex = state.tokens.length;
 | |
|       const functionStart = state.start;
 | |
|       const contextualKeyword = state.contextualKeyword;
 | |
|       parseIdentifier();
 | |
|       if (contextualKeyword === ContextualKeyword._await) {
 | |
|         parseAwait();
 | |
|         return false;
 | |
|       } else if (
 | |
|         contextualKeyword === ContextualKeyword._async &&
 | |
|         match(tt._function) &&
 | |
|         !canInsertSemicolon()
 | |
|       ) {
 | |
|         next();
 | |
|         parseFunction(functionStart, false);
 | |
|         return false;
 | |
|       } else if (
 | |
|         canBeArrow &&
 | |
|         contextualKeyword === ContextualKeyword._async &&
 | |
|         !canInsertSemicolon() &&
 | |
|         match(tt.name)
 | |
|       ) {
 | |
|         state.scopeDepth++;
 | |
|         parseBindingIdentifier(false);
 | |
|         expect(tt.arrow);
 | |
|         // let foo = async bar => {};
 | |
|         parseArrowExpression(startTokenIndex);
 | |
|         return true;
 | |
|       } else if (match(tt._do) && !canInsertSemicolon()) {
 | |
|         next();
 | |
|         parseBlock();
 | |
|         return false;
 | |
|       }
 | |
| 
 | |
|       if (canBeArrow && !canInsertSemicolon() && match(tt.arrow)) {
 | |
|         state.scopeDepth++;
 | |
|         markPriorBindingIdentifier(false);
 | |
|         expect(tt.arrow);
 | |
|         parseArrowExpression(startTokenIndex);
 | |
|         return true;
 | |
|       }
 | |
| 
 | |
|       state.tokens[state.tokens.length - 1].identifierRole = IdentifierRole.Access;
 | |
|       return false;
 | |
|     }
 | |
| 
 | |
|     case tt._do: {
 | |
|       next();
 | |
|       parseBlock();
 | |
|       return false;
 | |
|     }
 | |
| 
 | |
|     case tt.parenL: {
 | |
|       const wasArrow = parseParenAndDistinguishExpression(canBeArrow);
 | |
|       return wasArrow;
 | |
|     }
 | |
| 
 | |
|     case tt.bracketL:
 | |
|       next();
 | |
|       parseExprList(tt.bracketR, true);
 | |
|       return false;
 | |
| 
 | |
|     case tt.braceL:
 | |
|       parseObj(false, false);
 | |
|       return false;
 | |
| 
 | |
|     case tt._function:
 | |
|       parseFunctionExpression();
 | |
|       return false;
 | |
| 
 | |
|     case tt.at:
 | |
|       parseDecorators();
 | |
|     // Fall through.
 | |
| 
 | |
|     case tt._class:
 | |
|       parseClass(false);
 | |
|       return false;
 | |
| 
 | |
|     case tt._new:
 | |
|       parseNew();
 | |
|       return false;
 | |
| 
 | |
|     case tt.backQuote:
 | |
|       parseTemplate();
 | |
|       return false;
 | |
| 
 | |
|     case tt.doubleColon: {
 | |
|       next();
 | |
|       parseNoCallExpr();
 | |
|       return false;
 | |
|     }
 | |
| 
 | |
|     case tt.hash: {
 | |
|       const code = lookaheadCharCode();
 | |
|       if (IS_IDENTIFIER_START[code] || code === charCodes.backslash) {
 | |
|         parseMaybePrivateName();
 | |
|       } else {
 | |
|         next();
 | |
|       }
 | |
|       // Smart pipeline topic reference.
 | |
|       return false;
 | |
|     }
 | |
| 
 | |
|     default:
 | |
|       unexpected();
 | |
|       return false;
 | |
|   }
 | |
| }
 | |
| 
 | |
| function parseMaybePrivateName() {
 | |
|   eat(tt.hash);
 | |
|   parseIdentifier();
 | |
| }
 | |
| 
 | |
| function parseFunctionExpression() {
 | |
|   const functionStart = state.start;
 | |
|   parseIdentifier();
 | |
|   if (eat(tt.dot)) {
 | |
|     // function.sent
 | |
|     parseIdentifier();
 | |
|   }
 | |
|   parseFunction(functionStart, false);
 | |
| }
 | |
| 
 | |
| export function parseLiteral() {
 | |
|   next();
 | |
| }
 | |
| 
 | |
| export function parseParenExpression() {
 | |
|   expect(tt.parenL);
 | |
|   parseExpression();
 | |
|   expect(tt.parenR);
 | |
| }
 | |
| 
 | |
| // Returns true if this was an arrow expression.
 | |
| function parseParenAndDistinguishExpression(canBeArrow) {
 | |
|   // Assume this is a normal parenthesized expression, but if we see an arrow, we'll bail and
 | |
|   // start over as a parameter list.
 | |
|   const snapshot = state.snapshot();
 | |
| 
 | |
|   const startTokenIndex = state.tokens.length;
 | |
|   expect(tt.parenL);
 | |
| 
 | |
|   let first = true;
 | |
| 
 | |
|   while (!match(tt.parenR) && !state.error) {
 | |
|     if (first) {
 | |
|       first = false;
 | |
|     } else {
 | |
|       expect(tt.comma);
 | |
|       if (match(tt.parenR)) {
 | |
|         break;
 | |
|       }
 | |
|     }
 | |
| 
 | |
|     if (match(tt.ellipsis)) {
 | |
|       parseRest(false /* isBlockScope */);
 | |
|       parseParenItem();
 | |
|       break;
 | |
|     } else {
 | |
|       parseMaybeAssign(false, true);
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   expect(tt.parenR);
 | |
| 
 | |
|   if (canBeArrow && shouldParseArrow()) {
 | |
|     const wasArrow = parseArrow();
 | |
|     if (wasArrow) {
 | |
|       // It was an arrow function this whole time, so start over and parse it as params so that we
 | |
|       // get proper token annotations.
 | |
|       state.restoreFromSnapshot(snapshot);
 | |
|       state.scopeDepth++;
 | |
|       // Don't specify a context ID because arrow functions don't need a context ID.
 | |
|       parseFunctionParams();
 | |
|       parseArrow();
 | |
|       parseArrowExpression(startTokenIndex);
 | |
|       if (state.error) {
 | |
|         // Nevermind! This must have been something that looks very much like an
 | |
|         // arrow function but where its "parameter list" isn't actually a valid
 | |
|         // parameter list. Force non-arrow parsing.
 | |
|         // See https://github.com/alangpierce/sucrase/issues/666 for an example.
 | |
|         state.restoreFromSnapshot(snapshot);
 | |
|         parseParenAndDistinguishExpression(false);
 | |
|         return false;
 | |
|       }
 | |
|       return true;
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   return false;
 | |
| }
 | |
| 
 | |
| function shouldParseArrow() {
 | |
|   return match(tt.colon) || !canInsertSemicolon();
 | |
| }
 | |
| 
 | |
| // Returns whether there was an arrow token.
 | |
| export function parseArrow() {
 | |
|   if (isTypeScriptEnabled) {
 | |
|     return tsParseArrow();
 | |
|   } else if (isFlowEnabled) {
 | |
|     return flowParseArrow();
 | |
|   } else {
 | |
|     return eat(tt.arrow);
 | |
|   }
 | |
| }
 | |
| 
 | |
| function parseParenItem() {
 | |
|   if (isTypeScriptEnabled || isFlowEnabled) {
 | |
|     typedParseParenItem();
 | |
|   }
 | |
| }
 | |
| 
 | |
| // New's precedence is slightly tricky. It must allow its argument to
 | |
| // be a `[]` or dot subscript expression, but not a call — at least,
 | |
| // not without wrapping it in parentheses. Thus, it uses the noCalls
 | |
| // argument to parseSubscripts to prevent it from consuming the
 | |
| // argument list.
 | |
| function parseNew() {
 | |
|   expect(tt._new);
 | |
|   if (eat(tt.dot)) {
 | |
|     // new.target
 | |
|     parseIdentifier();
 | |
|     return;
 | |
|   }
 | |
|   parseNewCallee();
 | |
|   if (isFlowEnabled) {
 | |
|     flowStartParseNewArguments();
 | |
|   }
 | |
|   if (eat(tt.parenL)) {
 | |
|     parseExprList(tt.parenR);
 | |
|   }
 | |
| }
 | |
| 
 | |
| function parseNewCallee() {
 | |
|   parseNoCallExpr();
 | |
|   eat(tt.questionDot);
 | |
| }
 | |
| 
 | |
| export function parseTemplate() {
 | |
|   // Finish `, read quasi
 | |
|   nextTemplateToken();
 | |
|   // Finish quasi, read ${
 | |
|   nextTemplateToken();
 | |
|   while (!match(tt.backQuote) && !state.error) {
 | |
|     expect(tt.dollarBraceL);
 | |
|     parseExpression();
 | |
|     // Finish }, read quasi
 | |
|     nextTemplateToken();
 | |
|     // Finish quasi, read either ${ or `
 | |
|     nextTemplateToken();
 | |
|   }
 | |
|   next();
 | |
| }
 | |
| 
 | |
| // Parse an object literal or binding pattern.
 | |
| export function parseObj(isPattern, isBlockScope) {
 | |
|   // Attach a context ID to the object open and close brace and each object key.
 | |
|   const contextId = getNextContextId();
 | |
|   let first = true;
 | |
| 
 | |
|   next();
 | |
|   state.tokens[state.tokens.length - 1].contextId = contextId;
 | |
| 
 | |
|   while (!eat(tt.braceR) && !state.error) {
 | |
|     if (first) {
 | |
|       first = false;
 | |
|     } else {
 | |
|       expect(tt.comma);
 | |
|       if (eat(tt.braceR)) {
 | |
|         break;
 | |
|       }
 | |
|     }
 | |
| 
 | |
|     let isGenerator = false;
 | |
|     if (match(tt.ellipsis)) {
 | |
|       const previousIndex = state.tokens.length;
 | |
|       parseSpread();
 | |
|       if (isPattern) {
 | |
|         // Mark role when the only thing being spread over is an identifier.
 | |
|         if (state.tokens.length === previousIndex + 2) {
 | |
|           markPriorBindingIdentifier(isBlockScope);
 | |
|         }
 | |
|         if (eat(tt.braceR)) {
 | |
|           break;
 | |
|         }
 | |
|       }
 | |
|       continue;
 | |
|     }
 | |
| 
 | |
|     if (!isPattern) {
 | |
|       isGenerator = eat(tt.star);
 | |
|     }
 | |
| 
 | |
|     if (!isPattern && isContextual(ContextualKeyword._async)) {
 | |
|       if (isGenerator) unexpected();
 | |
| 
 | |
|       parseIdentifier();
 | |
|       if (
 | |
|         match(tt.colon) ||
 | |
|         match(tt.parenL) ||
 | |
|         match(tt.braceR) ||
 | |
|         match(tt.eq) ||
 | |
|         match(tt.comma)
 | |
|       ) {
 | |
|         // This is a key called "async" rather than an async function.
 | |
|       } else {
 | |
|         if (match(tt.star)) {
 | |
|           next();
 | |
|           isGenerator = true;
 | |
|         }
 | |
|         parsePropertyName(contextId);
 | |
|       }
 | |
|     } else {
 | |
|       parsePropertyName(contextId);
 | |
|     }
 | |
| 
 | |
|     parseObjPropValue(isPattern, isBlockScope, contextId);
 | |
|   }
 | |
| 
 | |
|   state.tokens[state.tokens.length - 1].contextId = contextId;
 | |
| }
 | |
| 
 | |
| function isGetterOrSetterMethod(isPattern) {
 | |
|   // We go off of the next and don't bother checking if the node key is actually "get" or "set".
 | |
|   // This lets us avoid generating a node, and should only make the validation worse.
 | |
|   return (
 | |
|     !isPattern &&
 | |
|     (match(tt.string) || // get "string"() {}
 | |
|       match(tt.num) || // get 1() {}
 | |
|       match(tt.bracketL) || // get ["string"]() {}
 | |
|       match(tt.name) || // get foo() {}
 | |
|       !!(state.type & TokenType.IS_KEYWORD)) // get debugger() {}
 | |
|   );
 | |
| }
 | |
| 
 | |
| // Returns true if this was a method.
 | |
| function parseObjectMethod(isPattern, objectContextId) {
 | |
|   // We don't need to worry about modifiers because object methods can't have optional bodies, so
 | |
|   // the start will never be used.
 | |
|   const functionStart = state.start;
 | |
|   if (match(tt.parenL)) {
 | |
|     if (isPattern) unexpected();
 | |
|     parseMethod(functionStart, /* isConstructor */ false);
 | |
|     return true;
 | |
|   }
 | |
| 
 | |
|   if (isGetterOrSetterMethod(isPattern)) {
 | |
|     parsePropertyName(objectContextId);
 | |
|     parseMethod(functionStart, /* isConstructor */ false);
 | |
|     return true;
 | |
|   }
 | |
|   return false;
 | |
| }
 | |
| 
 | |
| function parseObjectProperty(isPattern, isBlockScope) {
 | |
|   if (eat(tt.colon)) {
 | |
|     if (isPattern) {
 | |
|       parseMaybeDefault(isBlockScope);
 | |
|     } else {
 | |
|       parseMaybeAssign(false);
 | |
|     }
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   // Since there's no colon, we assume this is an object shorthand.
 | |
| 
 | |
|   // If we're in a destructuring, we've now discovered that the key was actually an assignee, so
 | |
|   // we need to tag it as a declaration with the appropriate scope. Otherwise, we might need to
 | |
|   // transform it on access, so mark it as a normal object shorthand.
 | |
|   let identifierRole;
 | |
|   if (isPattern) {
 | |
|     if (state.scopeDepth === 0) {
 | |
|       identifierRole = IdentifierRole.ObjectShorthandTopLevelDeclaration;
 | |
|     } else if (isBlockScope) {
 | |
|       identifierRole = IdentifierRole.ObjectShorthandBlockScopedDeclaration;
 | |
|     } else {
 | |
|       identifierRole = IdentifierRole.ObjectShorthandFunctionScopedDeclaration;
 | |
|     }
 | |
|   } else {
 | |
|     identifierRole = IdentifierRole.ObjectShorthand;
 | |
|   }
 | |
|   state.tokens[state.tokens.length - 1].identifierRole = identifierRole;
 | |
| 
 | |
|   // Regardless of whether we know this to be a pattern or if we're in an ambiguous context, allow
 | |
|   // parsing as if there's a default value.
 | |
|   parseMaybeDefault(isBlockScope, true);
 | |
| }
 | |
| 
 | |
| function parseObjPropValue(
 | |
|   isPattern,
 | |
|   isBlockScope,
 | |
|   objectContextId,
 | |
| ) {
 | |
|   if (isTypeScriptEnabled) {
 | |
|     tsStartParseObjPropValue();
 | |
|   } else if (isFlowEnabled) {
 | |
|     flowStartParseObjPropValue();
 | |
|   }
 | |
|   const wasMethod = parseObjectMethod(isPattern, objectContextId);
 | |
|   if (!wasMethod) {
 | |
|     parseObjectProperty(isPattern, isBlockScope);
 | |
|   }
 | |
| }
 | |
| 
 | |
| export function parsePropertyName(objectContextId) {
 | |
|   if (isFlowEnabled) {
 | |
|     flowParseVariance();
 | |
|   }
 | |
|   if (eat(tt.bracketL)) {
 | |
|     state.tokens[state.tokens.length - 1].contextId = objectContextId;
 | |
|     parseMaybeAssign();
 | |
|     expect(tt.bracketR);
 | |
|     state.tokens[state.tokens.length - 1].contextId = objectContextId;
 | |
|   } else {
 | |
|     if (match(tt.num) || match(tt.string) || match(tt.bigint) || match(tt.decimal)) {
 | |
|       parseExprAtom();
 | |
|     } else {
 | |
|       parseMaybePrivateName();
 | |
|     }
 | |
| 
 | |
|     state.tokens[state.tokens.length - 1].identifierRole = IdentifierRole.ObjectKey;
 | |
|     state.tokens[state.tokens.length - 1].contextId = objectContextId;
 | |
|   }
 | |
| }
 | |
| 
 | |
| // Parse object or class method.
 | |
| export function parseMethod(functionStart, isConstructor) {
 | |
|   const funcContextId = getNextContextId();
 | |
| 
 | |
|   state.scopeDepth++;
 | |
|   const startTokenIndex = state.tokens.length;
 | |
|   const allowModifiers = isConstructor; // For TypeScript parameter properties
 | |
|   parseFunctionParams(allowModifiers, funcContextId);
 | |
|   parseFunctionBodyAndFinish(functionStart, funcContextId);
 | |
|   const endTokenIndex = state.tokens.length;
 | |
|   state.scopes.push(new Scope(startTokenIndex, endTokenIndex, true));
 | |
|   state.scopeDepth--;
 | |
| }
 | |
| 
 | |
| // Parse arrow function expression.
 | |
| // If the parameters are provided, they will be converted to an
 | |
| // assignable list.
 | |
| export function parseArrowExpression(startTokenIndex) {
 | |
|   parseFunctionBody(true);
 | |
|   const endTokenIndex = state.tokens.length;
 | |
|   state.scopes.push(new Scope(startTokenIndex, endTokenIndex, true));
 | |
|   state.scopeDepth--;
 | |
| }
 | |
| 
 | |
| export function parseFunctionBodyAndFinish(functionStart, funcContextId = 0) {
 | |
|   if (isTypeScriptEnabled) {
 | |
|     tsParseFunctionBodyAndFinish(functionStart, funcContextId);
 | |
|   } else if (isFlowEnabled) {
 | |
|     flowParseFunctionBodyAndFinish(funcContextId);
 | |
|   } else {
 | |
|     parseFunctionBody(false, funcContextId);
 | |
|   }
 | |
| }
 | |
| 
 | |
| export function parseFunctionBody(allowExpression, funcContextId = 0) {
 | |
|   const isExpression = allowExpression && !match(tt.braceL);
 | |
| 
 | |
|   if (isExpression) {
 | |
|     parseMaybeAssign();
 | |
|   } else {
 | |
|     parseBlock(true /* isFunctionScope */, funcContextId);
 | |
|   }
 | |
| }
 | |
| 
 | |
| // Parses a comma-separated list of expressions, and returns them as
 | |
| // an array. `close` is the token type that ends the list, and
 | |
| // `allowEmpty` can be turned on to allow subsequent commas with
 | |
| // nothing in between them to be parsed as `null` (which is needed
 | |
| // for array literals).
 | |
| 
 | |
| function parseExprList(close, allowEmpty = false) {
 | |
|   let first = true;
 | |
|   while (!eat(close) && !state.error) {
 | |
|     if (first) {
 | |
|       first = false;
 | |
|     } else {
 | |
|       expect(tt.comma);
 | |
|       if (eat(close)) break;
 | |
|     }
 | |
|     parseExprListItem(allowEmpty);
 | |
|   }
 | |
| }
 | |
| 
 | |
| function parseExprListItem(allowEmpty) {
 | |
|   if (allowEmpty && match(tt.comma)) {
 | |
|     // Empty item; nothing more to parse for this item.
 | |
|   } else if (match(tt.ellipsis)) {
 | |
|     parseSpread();
 | |
|     parseParenItem();
 | |
|   } else if (match(tt.question)) {
 | |
|     // Partial function application proposal.
 | |
|     next();
 | |
|   } else {
 | |
|     parseMaybeAssign(false, true);
 | |
|   }
 | |
| }
 | |
| 
 | |
| // Parse the next token as an identifier.
 | |
| export function parseIdentifier() {
 | |
|   next();
 | |
|   state.tokens[state.tokens.length - 1].type = tt.name;
 | |
| }
 | |
| 
 | |
| // Parses await expression inside async function.
 | |
| function parseAwait() {
 | |
|   parseMaybeUnary();
 | |
| }
 | |
| 
 | |
| // Parses yield expression inside generator.
 | |
| function parseYield() {
 | |
|   next();
 | |
|   if (!match(tt.semi) && !canInsertSemicolon()) {
 | |
|     eat(tt.star);
 | |
|     parseMaybeAssign();
 | |
|   }
 | |
| }
 | |
| 
 | |
| // https://github.com/tc39/proposal-js-module-blocks
 | |
| function parseModuleExpression() {
 | |
|   expectContextual(ContextualKeyword._module);
 | |
|   expect(tt.braceL);
 | |
|   // For now, just call parseBlockBody to parse the block. In the future when we
 | |
|   // implement full support, we'll want to emit scopes and possibly other
 | |
|   // information.
 | |
|   parseBlockBody(tt.braceR);
 | |
| }
 |