 85bf1341f3
			
		
	
	85bf1341f3
	
	
	
		
			
			Frontend Enhancements: - Complete React TypeScript frontend with modern UI components - Distributed workflows management interface with real-time updates - Socket.IO integration for live agent status monitoring - Agent management dashboard with cluster visualization - Project management interface with metrics and task tracking - Responsive design with proper error handling and loading states Backend Infrastructure: - Distributed coordinator for multi-agent workflow orchestration - Cluster management API with comprehensive agent operations - Enhanced database models for agents and projects - Project service for filesystem-based project discovery - Performance monitoring and metrics collection - Comprehensive API documentation and error handling Documentation: - Complete distributed development guide (README_DISTRIBUTED.md) - Comprehensive development report with architecture insights - System configuration templates and deployment guides The platform now provides a complete web interface for managing the distributed AI cluster with real-time monitoring, workflow orchestration, and agent coordination capabilities. 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude <noreply@anthropic.com>
		
			
				
	
	
		
			192 lines
		
	
	
		
			5.4 KiB
		
	
	
	
		
			JavaScript
		
	
	
	
	
	
			
		
		
	
	
			192 lines
		
	
	
		
			5.4 KiB
		
	
	
	
		
			JavaScript
		
	
	
	
	
	
| "use strict";
 | |
| 
 | |
| Object.defineProperty(exports, "__esModule", {
 | |
|   value: true
 | |
| });
 | |
| exports.TokenMap = void 0;
 | |
| var _t = require("@babel/types");
 | |
| const {
 | |
|   traverseFast,
 | |
|   VISITOR_KEYS
 | |
| } = _t;
 | |
| class TokenMap {
 | |
|   constructor(ast, tokens, source) {
 | |
|     this._tokens = void 0;
 | |
|     this._source = void 0;
 | |
|     this._nodesToTokenIndexes = new Map();
 | |
|     this._nodesOccurrencesCountCache = new Map();
 | |
|     this._tokensCache = new Map();
 | |
|     this._tokens = tokens;
 | |
|     this._source = source;
 | |
|     traverseFast(ast, node => {
 | |
|       const indexes = this._getTokensIndexesOfNode(node);
 | |
|       if (indexes.length > 0) this._nodesToTokenIndexes.set(node, indexes);
 | |
|     });
 | |
|     this._tokensCache = null;
 | |
|   }
 | |
|   has(node) {
 | |
|     return this._nodesToTokenIndexes.has(node);
 | |
|   }
 | |
|   getIndexes(node) {
 | |
|     return this._nodesToTokenIndexes.get(node);
 | |
|   }
 | |
|   find(node, condition) {
 | |
|     const indexes = this._nodesToTokenIndexes.get(node);
 | |
|     if (indexes) {
 | |
|       for (let k = 0; k < indexes.length; k++) {
 | |
|         const index = indexes[k];
 | |
|         const tok = this._tokens[index];
 | |
|         if (condition(tok, index)) return tok;
 | |
|       }
 | |
|     }
 | |
|     return null;
 | |
|   }
 | |
|   findLastIndex(node, condition) {
 | |
|     const indexes = this._nodesToTokenIndexes.get(node);
 | |
|     if (indexes) {
 | |
|       for (let k = indexes.length - 1; k >= 0; k--) {
 | |
|         const index = indexes[k];
 | |
|         const tok = this._tokens[index];
 | |
|         if (condition(tok, index)) return index;
 | |
|       }
 | |
|     }
 | |
|     return -1;
 | |
|   }
 | |
|   findMatching(node, test, occurrenceCount = 0) {
 | |
|     const indexes = this._nodesToTokenIndexes.get(node);
 | |
|     if (indexes) {
 | |
|       let i = 0;
 | |
|       const count = occurrenceCount;
 | |
|       if (count > 1) {
 | |
|         const cache = this._nodesOccurrencesCountCache.get(node);
 | |
|         if (cache && cache.test === test && cache.count < count) {
 | |
|           i = cache.i + 1;
 | |
|           occurrenceCount -= cache.count + 1;
 | |
|         }
 | |
|       }
 | |
|       for (; i < indexes.length; i++) {
 | |
|         const tok = this._tokens[indexes[i]];
 | |
|         if (this.matchesOriginal(tok, test)) {
 | |
|           if (occurrenceCount === 0) {
 | |
|             if (count > 0) {
 | |
|               this._nodesOccurrencesCountCache.set(node, {
 | |
|                 test,
 | |
|                 count,
 | |
|                 i
 | |
|               });
 | |
|             }
 | |
|             return tok;
 | |
|           }
 | |
|           occurrenceCount--;
 | |
|         }
 | |
|       }
 | |
|     }
 | |
|     return null;
 | |
|   }
 | |
|   matchesOriginal(token, test) {
 | |
|     if (token.end - token.start !== test.length) return false;
 | |
|     if (token.value != null) return token.value === test;
 | |
|     return this._source.startsWith(test, token.start);
 | |
|   }
 | |
|   startMatches(node, test) {
 | |
|     const indexes = this._nodesToTokenIndexes.get(node);
 | |
|     if (!indexes) return false;
 | |
|     const tok = this._tokens[indexes[0]];
 | |
|     if (tok.start !== node.start) return false;
 | |
|     return this.matchesOriginal(tok, test);
 | |
|   }
 | |
|   endMatches(node, test) {
 | |
|     const indexes = this._nodesToTokenIndexes.get(node);
 | |
|     if (!indexes) return false;
 | |
|     const tok = this._tokens[indexes[indexes.length - 1]];
 | |
|     if (tok.end !== node.end) return false;
 | |
|     return this.matchesOriginal(tok, test);
 | |
|   }
 | |
|   _getTokensIndexesOfNode(node) {
 | |
|     if (node.start == null || node.end == null) return [];
 | |
|     const {
 | |
|       first,
 | |
|       last
 | |
|     } = this._findTokensOfNode(node, 0, this._tokens.length - 1);
 | |
|     let low = first;
 | |
|     const children = childrenIterator(node);
 | |
|     if ((node.type === "ExportNamedDeclaration" || node.type === "ExportDefaultDeclaration") && node.declaration && node.declaration.type === "ClassDeclaration") {
 | |
|       children.next();
 | |
|     }
 | |
|     const indexes = [];
 | |
|     for (const child of children) {
 | |
|       if (child == null) continue;
 | |
|       if (child.start == null || child.end == null) continue;
 | |
|       const childTok = this._findTokensOfNode(child, low, last);
 | |
|       const high = childTok.first;
 | |
|       for (let k = low; k < high; k++) indexes.push(k);
 | |
|       low = childTok.last + 1;
 | |
|     }
 | |
|     for (let k = low; k <= last; k++) indexes.push(k);
 | |
|     return indexes;
 | |
|   }
 | |
|   _findTokensOfNode(node, low, high) {
 | |
|     const cached = this._tokensCache.get(node);
 | |
|     if (cached) return cached;
 | |
|     const first = this._findFirstTokenOfNode(node.start, low, high);
 | |
|     const last = this._findLastTokenOfNode(node.end, first, high);
 | |
|     this._tokensCache.set(node, {
 | |
|       first,
 | |
|       last
 | |
|     });
 | |
|     return {
 | |
|       first,
 | |
|       last
 | |
|     };
 | |
|   }
 | |
|   _findFirstTokenOfNode(start, low, high) {
 | |
|     while (low <= high) {
 | |
|       const mid = high + low >> 1;
 | |
|       if (start < this._tokens[mid].start) {
 | |
|         high = mid - 1;
 | |
|       } else if (start > this._tokens[mid].start) {
 | |
|         low = mid + 1;
 | |
|       } else {
 | |
|         return mid;
 | |
|       }
 | |
|     }
 | |
|     return low;
 | |
|   }
 | |
|   _findLastTokenOfNode(end, low, high) {
 | |
|     while (low <= high) {
 | |
|       const mid = high + low >> 1;
 | |
|       if (end < this._tokens[mid].end) {
 | |
|         high = mid - 1;
 | |
|       } else if (end > this._tokens[mid].end) {
 | |
|         low = mid + 1;
 | |
|       } else {
 | |
|         return mid;
 | |
|       }
 | |
|     }
 | |
|     return high;
 | |
|   }
 | |
| }
 | |
| exports.TokenMap = TokenMap;
 | |
| function* childrenIterator(node) {
 | |
|   if (node.type === "TemplateLiteral") {
 | |
|     yield node.quasis[0];
 | |
|     for (let i = 1; i < node.quasis.length; i++) {
 | |
|       yield node.expressions[i - 1];
 | |
|       yield node.quasis[i];
 | |
|     }
 | |
|     return;
 | |
|   }
 | |
|   const keys = VISITOR_KEYS[node.type];
 | |
|   for (const key of keys) {
 | |
|     const child = node[key];
 | |
|     if (!child) continue;
 | |
|     if (Array.isArray(child)) {
 | |
|       yield* child;
 | |
|     } else {
 | |
|       yield child;
 | |
|     }
 | |
|   }
 | |
| }
 | |
| 
 | |
| //# sourceMappingURL=token-map.js.map
 |