mirror of
https://github.com/cpacker/MemGPT.git
synced 2025-06-03 04:30:22 +00:00

Co-authored-by: Charles Packer <packercharles@gmail.com> Co-authored-by: Shubham Naik <shubham.naik10@gmail.com> Co-authored-by: Shubham Naik <shub@memgpt.ai>
194 lines
6.6 KiB
Python
194 lines
6.6 KiB
Python
import json
|
|
import logging
|
|
import secrets
|
|
from pathlib import Path
|
|
from typing import Optional
|
|
|
|
import typer
|
|
import uvicorn
|
|
from fastapi import FastAPI, Request
|
|
from fastapi.responses import JSONResponse
|
|
from starlette.middleware.cors import CORSMiddleware
|
|
|
|
from letta.server.constants import REST_DEFAULT_PORT
|
|
|
|
# NOTE(charles): these are extra routes that are not part of v1 but we still need to mount to pass tests
|
|
from letta.server.rest_api.auth.index import (
|
|
setup_auth_router, # TODO: probably remove right?
|
|
)
|
|
from letta.server.rest_api.interface import StreamingServerInterface
|
|
from letta.server.rest_api.routers.openai.assistants.assistants import (
|
|
router as openai_assistants_router,
|
|
)
|
|
from letta.server.rest_api.routers.openai.assistants.threads import (
|
|
router as openai_threads_router,
|
|
)
|
|
from letta.server.rest_api.routers.openai.chat_completions.chat_completions import (
|
|
router as openai_chat_completions_router,
|
|
)
|
|
|
|
# from letta.orm.utilities import get_db_session # TODO(ethan) reenable once we merge ORM
|
|
from letta.server.rest_api.routers.v1 import ROUTERS as v1_routes
|
|
from letta.server.rest_api.routers.v1.organizations import (
|
|
router as organizations_router,
|
|
)
|
|
from letta.server.rest_api.routers.v1.users import (
|
|
router as users_router, # TODO: decide on admin
|
|
)
|
|
from letta.server.rest_api.static_files import mount_static_files
|
|
from letta.server.server import SyncServer
|
|
from letta.settings import settings
|
|
|
|
# TODO(ethan)
|
|
# NOTE(charles): @ethan I had to add this to get the global as the bottom to work
|
|
interface: StreamingServerInterface = StreamingServerInterface
|
|
server = SyncServer(default_interface_factory=lambda: interface())
|
|
|
|
# TODO: remove
|
|
password = None
|
|
## TODO(ethan): eventuall remove
|
|
# if password := settings.server_pass:
|
|
# # if the pass was specified in the environment, use it
|
|
# print(f"Using existing admin server password from environment.")
|
|
# else:
|
|
# # Autogenerate a password for this session and dump it to stdout
|
|
# password = secrets.token_urlsafe(16)
|
|
# #typer.secho(f"Generated admin server password for this session: {password}", fg=typer.colors.GREEN)
|
|
|
|
|
|
ADMIN_PREFIX = "/v1/admin"
|
|
API_PREFIX = "/v1"
|
|
OPENAI_API_PREFIX = "/openai"
|
|
|
|
|
|
def create_application() -> "FastAPI":
|
|
"""the application start routine"""
|
|
# global server
|
|
# server = SyncServer(default_interface_factory=lambda: interface())
|
|
|
|
app = FastAPI(
|
|
swagger_ui_parameters={"docExpansion": "none"},
|
|
# openapi_tags=TAGS_METADATA,
|
|
title="Letta",
|
|
summary="Create LLM agents with long-term memory and custom tools 📚🦙",
|
|
version="1.0.0", # TODO wire this up to the version in the package
|
|
)
|
|
app.add_middleware(
|
|
CORSMiddleware,
|
|
allow_origins=settings.cors_origins,
|
|
allow_credentials=True,
|
|
allow_methods=["*"],
|
|
allow_headers=["*"],
|
|
)
|
|
|
|
@app.middleware("http")
|
|
async def set_current_user_middleware(request: Request, call_next):
|
|
user_id = request.headers.get("user_id")
|
|
if user_id:
|
|
try:
|
|
server.set_current_user(user_id)
|
|
except ValueError as e:
|
|
# Return an HTTP 401 Unauthorized response
|
|
# raise HTTPException(status_code=401, detail=str(e))
|
|
return JSONResponse(status_code=401, content={"detail": str(e)})
|
|
else:
|
|
server.set_current_user(None)
|
|
response = await call_next(request)
|
|
return response
|
|
|
|
for route in v1_routes:
|
|
app.include_router(route, prefix=API_PREFIX)
|
|
# this gives undocumented routes for "latest" and bare api calls.
|
|
# we should always tie this to the newest version of the api.
|
|
# app.include_router(route, prefix="", include_in_schema=False)
|
|
app.include_router(route, prefix="/latest", include_in_schema=False)
|
|
|
|
# NOTE: ethan these are the extra routes
|
|
# TODO(ethan) remove
|
|
|
|
# admin/users
|
|
app.include_router(users_router, prefix=ADMIN_PREFIX)
|
|
app.include_router(organizations_router, prefix=ADMIN_PREFIX)
|
|
|
|
# openai
|
|
app.include_router(openai_assistants_router, prefix=OPENAI_API_PREFIX)
|
|
app.include_router(openai_threads_router, prefix=OPENAI_API_PREFIX)
|
|
app.include_router(openai_chat_completions_router, prefix=OPENAI_API_PREFIX)
|
|
|
|
# /api/auth endpoints
|
|
app.include_router(setup_auth_router(server, interface, password), prefix=API_PREFIX)
|
|
|
|
# / static files
|
|
mount_static_files(app)
|
|
|
|
@app.on_event("startup")
|
|
def on_startup():
|
|
# load the default tools
|
|
# from letta.orm.tool import Tool
|
|
|
|
# Tool.load_default_tools(get_db_session())
|
|
|
|
# Update the OpenAPI schema
|
|
if not app.openapi_schema:
|
|
app.openapi_schema = app.openapi()
|
|
|
|
openai_docs, letta_docs = [app.openapi_schema.copy() for _ in range(2)]
|
|
|
|
openai_docs["paths"] = {k: v for k, v in openai_docs["paths"].items() if k.startswith("/openai")}
|
|
openai_docs["info"]["title"] = "OpenAI Assistants API"
|
|
letta_docs["paths"] = {k: v for k, v in letta_docs["paths"].items() if not k.startswith("/openai")}
|
|
letta_docs["info"]["title"] = "Letta API"
|
|
|
|
# Split the API docs into Letta API, and OpenAI Assistants compatible API
|
|
for name, docs in [
|
|
(
|
|
"openai",
|
|
openai_docs,
|
|
),
|
|
(
|
|
"letta",
|
|
letta_docs,
|
|
),
|
|
]:
|
|
if settings.cors_origins:
|
|
docs["servers"] = [{"url": host} for host in settings.cors_origins]
|
|
Path(f"openapi_{name}.json").write_text(json.dumps(docs, indent=2))
|
|
|
|
@app.on_event("shutdown")
|
|
def on_shutdown():
|
|
global server
|
|
server.save_agents()
|
|
# server = None
|
|
|
|
return app
|
|
|
|
|
|
app = create_application()
|
|
|
|
|
|
def start_server(
|
|
port: Optional[int] = None,
|
|
host: Optional[str] = None,
|
|
debug: bool = False,
|
|
):
|
|
"""Convenience method to start the server from within Python"""
|
|
if debug:
|
|
from letta.server.server import logger as server_logger
|
|
|
|
# Set the logging level
|
|
server_logger.setLevel(logging.DEBUG)
|
|
# Create a StreamHandler
|
|
stream_handler = logging.StreamHandler()
|
|
# Set the formatter (optional)
|
|
formatter = logging.Formatter("%(asctime)s - %(name)s - %(levelname)s - %(message)s")
|
|
stream_handler.setFormatter(formatter)
|
|
# Add the handler to the logger
|
|
server_logger.addHandler(stream_handler)
|
|
|
|
print(f"Running: uvicorn server:app --host {host or 'localhost'} --port {port or REST_DEFAULT_PORT}")
|
|
uvicorn.run(
|
|
app,
|
|
host=host or "localhost",
|
|
port=port or REST_DEFAULT_PORT,
|
|
)
|