Files
hive/backend/app/main.py
anthonyrawlins 7af5b47477 Implement complete Bearer Token and API key authentication system
- Create comprehensive authentication backend with JWT and API key support
- Add database models for users, API keys, and tokens with proper security
- Implement authentication middleware and API endpoints
- Build complete frontend authentication UI with:
  - LoginForm component with JWT authentication
  - APIKeyManager for creating and managing API keys
  - AuthDashboard for comprehensive auth management
  - AuthContext for state management and authenticated requests
- Initialize database with default admin user (admin/admin123)
- Add proper token refresh, validation, and blacklisting
- Implement scope-based API key authorization system

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-07-10 20:52:58 +10:00

309 lines
10 KiB
Python

from fastapi import FastAPI, Depends, HTTPException
from fastapi.middleware.cors import CORSMiddleware
from fastapi.staticfiles import StaticFiles
from contextlib import asynccontextmanager
import json
import asyncio
import uvicorn
from datetime import datetime
from pathlib import Path
import socketio
from .core.hive_coordinator import HiveCoordinator
from .core.distributed_coordinator import DistributedCoordinator
from .core.database import engine, get_db, init_database_with_retry, test_database_connection
from .api import agents, workflows, executions, monitoring, projects, tasks, cluster, distributed_workflows, cli_agents, auth
# from .mcp.distributed_mcp_server import get_mcp_server
from .models.user import Base
from .models import agent, project # Import the new agent and project models
# Global coordinator instances
hive_coordinator = HiveCoordinator()
distributed_coordinator = DistributedCoordinator()
@asynccontextmanager
async def lifespan(app: FastAPI):
"""Enhanced application lifespan manager with proper error handling"""
startup_success = False
try:
# Startup
print("🚀 Starting Hive Orchestrator...")
# Initialize database with retry logic
print("📊 Initializing database...")
init_database_with_retry()
# Initialize auth database tables and initial data
print("🔐 Initializing authentication system...")
from .core.init_db import initialize_database
initialize_database()
# Test database connection
if not test_database_connection():
raise Exception("Database connection test failed")
# Initialize coordinators with error handling
print("🤖 Initializing AI coordinator...")
await hive_coordinator.initialize()
print("🌐 Initializing distributed coordinator...")
await distributed_coordinator.start()
# Initialize MCP server
# print("🔌 Initializing MCP server...")
# mcp_server = get_mcp_server()
# await mcp_server.initialize(distributed_coordinator)
startup_success = True
print("✅ Hive Orchestrator with distributed workflows started successfully!")
yield
except Exception as e:
print(f"❌ Startup failed: {e}")
if startup_success:
# If we got past startup, try to shutdown cleanly
try:
await hive_coordinator.shutdown()
await distributed_coordinator.stop()
except Exception as shutdown_error:
print(f"Shutdown error during startup failure: {shutdown_error}")
raise
finally:
# Shutdown
print("🛑 Shutting down Hive Orchestrator...")
try:
await hive_coordinator.shutdown()
await distributed_coordinator.stop()
print("✅ Hive Orchestrator stopped")
except Exception as e:
print(f"❌ Shutdown error: {e}")
# Create FastAPI application
app = FastAPI(
title="Hive API",
description="Unified Distributed AI Orchestration Platform",
version="1.0.0",
lifespan=lifespan
)
# Enhanced CORS configuration for production
app.add_middleware(
CORSMiddleware,
allow_origins=[
"http://localhost:3000",
"http://localhost:3001",
"https://hive.home.deepblack.cloud",
"http://hive.home.deepblack.cloud"
],
allow_credentials=True,
allow_methods=["*"],
allow_headers=["*"],
)
# Include API routes
app.include_router(auth.router, prefix="/api/auth", tags=["authentication"])
app.include_router(agents.router, prefix="/api", tags=["agents"])
app.include_router(workflows.router, prefix="/api", tags=["workflows"])
app.include_router(executions.router, prefix="/api", tags=["executions"])
app.include_router(monitoring.router, prefix="/api", tags=["monitoring"])
app.include_router(projects.router, prefix="/api", tags=["projects"])
app.include_router(tasks.router, prefix="/api", tags=["tasks"])
app.include_router(cluster.router, prefix="/api", tags=["cluster"])
app.include_router(distributed_workflows.router, tags=["distributed-workflows"])
app.include_router(cli_agents.router, tags=["cli-agents"])
# Set coordinator reference in tasks module
tasks.set_coordinator(hive_coordinator)
# Socket.IO server setup
sio = socketio.AsyncServer(
async_mode='asgi',
cors_allowed_origins="*",
logger=True,
engineio_logger=False
)
# Socket.IO event handlers
@sio.event
async def connect(sid, environ):
"""Handle client connection"""
print(f"🔌 Socket.IO client {sid} connected")
await sio.emit('connection_confirmed', {
'status': 'connected',
'timestamp': datetime.now().isoformat(),
'message': 'Connected to Hive Socket.IO server'
}, room=sid)
@sio.event
async def disconnect(sid):
"""Handle client disconnection"""
print(f"🔌 Socket.IO client {sid} disconnected")
@sio.event
async def join_room(sid, data):
"""Handle client joining a room/topic"""
room = data.get('room', 'general')
await sio.enter_room(sid, room)
print(f"🔌 Client {sid} joined room: {room}")
await sio.emit('room_joined', {
'room': room,
'timestamp': datetime.now().isoformat(),
'message': f'Successfully joined {room} room'
}, room=sid)
@sio.event
async def leave_room(sid, data):
"""Handle client leaving a room/topic"""
room = data.get('room', 'general')
await sio.leave_room(sid, room)
print(f"🔌 Client {sid} left room: {room}")
await sio.emit('room_left', {
'room': room,
'timestamp': datetime.now().isoformat(),
'message': f'Successfully left {room} room'
}, room=sid)
@sio.event
async def subscribe(sid, data):
"""Handle event subscription"""
events = data.get('events', [])
room = data.get('room', 'general')
# Join the room if not already joined
await sio.enter_room(sid, room)
print(f"🔌 Client {sid} subscribed to events: {events} in room: {room}")
await sio.emit('subscription_confirmed', {
'events': events,
'room': room,
'timestamp': datetime.now().isoformat(),
'message': f'Subscribed to {len(events)} events in {room} room'
}, room=sid)
@sio.event
async def ping(sid):
"""Handle ping from client"""
await sio.emit('pong', {
'timestamp': datetime.now().isoformat()
}, room=sid)
# Socket.IO connection manager
class SocketIOManager:
def __init__(self, socketio_server):
self.sio = socketio_server
async def send_to_room(self, room: str, event: str, data: dict):
"""Send event to all clients in a room"""
try:
await self.sio.emit(event, data, room=room)
except Exception as e:
print(f"Error sending to room {room}: {e}")
async def broadcast(self, event: str, data: dict):
"""Broadcast event to all connected clients"""
try:
await self.sio.emit(event, data)
except Exception as e:
print(f"Error broadcasting event {event}: {e}")
async def send_to_client(self, sid: str, event: str, data: dict):
"""Send event to a specific client"""
try:
await self.sio.emit(event, data, room=sid)
except Exception as e:
print(f"Error sending to client {sid}: {e}")
manager = SocketIOManager(sio)
# Socket.IO integration with FastAPI
# The socket.io server is integrated below in the app creation
@app.get("/")
async def root():
"""Root endpoint"""
return {
"message": "🐝 Welcome to Hive - Distributed AI Orchestration Platform",
"status": "operational",
"version": "1.0.0",
"api_docs": "/docs",
"timestamp": datetime.now().isoformat()
}
@app.get("/health")
async def health_check_internal():
"""Internal health check endpoint for Docker and monitoring"""
return {"status": "healthy", "timestamp": datetime.now().isoformat()}
@app.get("/api/health")
async def health_check():
"""Enhanced health check endpoint with comprehensive status"""
health_status = {
"status": "healthy",
"timestamp": datetime.now().isoformat(),
"version": "1.0.0",
"components": {
"api": "operational",
"database": "unknown",
"coordinator": "unknown",
"agents": {}
}
}
# Test database connection
try:
if test_database_connection():
health_status["components"]["database"] = "operational"
else:
health_status["components"]["database"] = "unhealthy"
health_status["status"] = "degraded"
except Exception as e:
health_status["components"]["database"] = f"error: {str(e)}"
health_status["status"] = "degraded"
# Test coordinator health
try:
coordinator_status = await hive_coordinator.get_health_status()
health_status["components"]["coordinator"] = coordinator_status.get("status", "unknown")
health_status["components"]["agents"] = coordinator_status.get("agents", {})
except Exception as e:
health_status["components"]["coordinator"] = f"error: {str(e)}"
health_status["status"] = "degraded"
# Return appropriate status code
if health_status["status"] == "degraded":
raise HTTPException(status_code=503, detail=health_status)
return health_status
@app.get("/api/status")
async def get_system_status():
"""Get comprehensive system status"""
return await hive_coordinator.get_comprehensive_status()
@app.get("/api/metrics")
async def get_metrics():
"""Prometheus metrics endpoint"""
return await hive_coordinator.get_prometheus_metrics()
# Make manager and coordinators available to other modules
app.state.socketio_manager = manager
app.state.hive_coordinator = hive_coordinator
app.state.distributed_coordinator = distributed_coordinator
# Create Socket.IO ASGI app
socket_app = socketio.ASGIApp(sio, other_asgi_app=app, socketio_path='/socket.io')
if __name__ == "__main__":
uvicorn.run(
"app.main:socket_app",
host="0.0.0.0",
port=8000,
reload=True,
log_level="info"
)