chore: deprecate letta configure and remove config defaults (#1841)

Co-authored-by: cpacker <packercharles@gmail.com>
This commit is contained in:
Sarah Wooders 2024-10-09 14:32:08 -07:00 committed by GitHub
parent 2492ad780d
commit e094e18580
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
19 changed files with 33 additions and 1224 deletions

View File

@ -1,7 +0,0 @@
{
"embedding_endpoint": "https://embeddings.memgpt.ai",
"embedding_model": "BAAI/bge-large-en-v1.5",
"embedding_dim": 1024,
"embedding_chunk_size": 300,
"embedding_endpoint_type": "hugging-face"
}

View File

@ -1,7 +0,0 @@
{
"embedding_endpoint": null,
"embedding_model": "BAAI/bge-small-en-v1.5",
"embedding_dim": 384,
"embedding_chunk_size": 300,
"embedding_endpoint_type": "local"
}

View File

@ -1,7 +0,0 @@
{
"embedding_endpoint_type": "ollama",
"embedding_endpoint": "http://127.0.0.1:11434",
"embedding_model": "mxbai-embed-large",
"embedding_dim": 512,
"embedding_chunk_size": 200
}

View File

@ -1,7 +0,0 @@
{
"embedding_endpoint_type": "openai",
"embedding_endpoint": "https://api.openai.com/v1",
"embedding_model": "text-embedding-ada-002",
"embedding_dim": 1536,
"embedding_chunk_size": 300
}

View File

@ -1,11 +0,0 @@
{
"context_window": 8192,
"model_endpoint_type": "openai",
"model_endpoint": "https://inference.memgpt.ai",
"model": "memgpt-openai",
"embedding_endpoint_type": "hugging-face",
"embedding_endpoint": "https://embeddings.memgpt.ai",
"embedding_model": "BAAI/bge-large-en-v1.5",
"embedding_dim": 1024,
"embedding_chunk_size": 300
}

View File

@ -1,7 +0,0 @@
{
"context_window": 200000,
"model": "claude-3-opus-20240229",
"model_endpoint_type": "anthropic",
"model_endpoint": "https://api.anthropic.com/v1",
"model_wrapper": null
}

View File

@ -1,7 +0,0 @@
{
"context_window": 8192,
"model": "gpt-4",
"model_endpoint_type": "openai",
"model_endpoint": "https://api.openai.com/v1",
"model_wrapper": null
}

View File

@ -1,7 +0,0 @@
{
"context_window": 8192,
"model": "llama3-groq-70b-8192-tool-use-preview",
"model_endpoint_type": "groq",
"model_endpoint": "https://api.groq.com/openai/v1",
"model_wrapper": null
}

View File

@ -1,6 +0,0 @@
{
"context_window": 16384,
"model_endpoint_type": "openai",
"model_endpoint": "https://inference.memgpt.ai",
"model": "memgpt-openai"
}

View File

@ -1,6 +0,0 @@
{
"context_window": 8192,
"model_endpoint_type": "ollama",
"model_endpoint": "http://127.0.0.1:11434",
"model": "dolphin2.2-mistral:7b-q6_K"
}

View File

@ -1,12 +0,0 @@
{
"context_window": 8192,
"model": "gpt-4",
"model_endpoint_type": "openai",
"model_endpoint": "https://api.openai.com/v1",
"model_wrapper": null,
"embedding_endpoint_type": "openai",
"embedding_endpoint": "https://api.openai.com/v1",
"embedding_model": "text-embedding-ada-002",
"embedding_dim": 1536,
"embedding_chunk_size": 300
}

File diff suppressed because it is too large Load Diff

View File

@ -1,13 +0,0 @@
{
"context_window": 200000,
"model": "claude-3-opus-20240229",
"model_endpoint_type": "anthropic",
"model_endpoint": "https://api.anthropic.com/v1",
"model_wrapper": null,
"embedding_endpoint_type": "hugging-face",
"embedding_endpoint": "https://embeddings.memgpt.ai",
"embedding_model": "BAAI/bge-large-en-v1.5",
"embedding_dim": 1024,
"embedding_chunk_size": 300
}

View File

@ -1,11 +0,0 @@
{
"context_window": 8192,
"model_endpoint_type": "openai",
"model_endpoint": "https://inference.memgpt.ai",
"model": "memgpt-openai",
"embedding_endpoint_type": "hugging-face",
"embedding_endpoint": "https://embeddings.memgpt.ai",
"embedding_model": "BAAI/bge-large-en-v1.5",
"embedding_dim": 1024,
"embedding_chunk_size": 300
}

View File

@ -1,12 +0,0 @@
{
"context_window": 8192,
"model": "gpt-4",
"model_endpoint_type": "openai",
"model_endpoint": "https://api.openai.com/v1",
"model_wrapper": null,
"embedding_endpoint_type": "openai",
"embedding_endpoint": "https://api.openai.com/v1",
"embedding_model": "text-embedding-ada-002",
"embedding_dim": 1536,
"embedding_chunk_size": 300
}

View File

@ -13,7 +13,6 @@ from letta.schemas.llm_config import LLMConfig
class Provider(BaseModel):
base_url: str
def list_llm_models(self):
return []
@ -25,6 +24,32 @@ class Provider(BaseModel):
pass
class LettaProvider(Provider):
name: str = "letta"
def list_llm_models(self) -> List[LLMConfig]:
return [
LLMConfig(
model="memgpt-openai",
model_endpoint_type="openai",
model_endpoint="https://inference.memgpt.ai",
context_window=16384,
)
]
def list_embedding_models(self):
return [
EmbeddingConfig(
embedding_model="BAAI/bge-large-en-v1.5",
embedding_endpoint_type="hugging-face",
embedding_endpoint="https://embeddings.memgpt.ai",
embedding_dim=1024,
embedding_chunk_size=300,
)
]
class OpenAIProvider(Provider):
name: str = "openai"
api_key: str = Field(..., description="API key for the OpenAI API.")

View File

@ -47,6 +47,7 @@ from letta.providers import (
AnthropicProvider,
AzureProvider,
GoogleAIProvider,
LettaProvider,
OllamaProvider,
OpenAIProvider,
VLLMProvider,
@ -259,8 +260,8 @@ class SyncServer(Server):
# add global default tools (for admin)
self.add_default_tools(module_name="base")
# collect providers
self._enabled_providers = []
# collect providers (always has Letta as a default)
self._enabled_providers = [LettaProvider()]
if model_settings.openai_api_key:
self._enabled_providers.append(OpenAIProvider(api_key=model_settings.openai_api_key))
if model_settings.anthropic_api_key:

View File

@ -49,8 +49,8 @@ namespace = uuid.NAMESPACE_DNS
agent_uuid = str(uuid.uuid5(namespace, "test-endpoints-agent"))
# defaults (letta hosted)
EMBEDDING_CONFIG_PATH = "configs/embedding_model_configs/letta-hosted.json"
LLM_CONFIG_PATH = "configs/llm_model_configs/letta-hosted.json"
EMBEDDING_CONFIG_PATH = "tests/configs/embedding_model_configs/letta-hosted.json"
LLM_CONFIG_PATH = "tests/configs/llm_model_configs/letta-hosted.json"
# ======================================================================================================================

View File

@ -13,7 +13,7 @@ from tests.helpers.endpoints_helper import (
)
# directories
embedding_config_dir = "configs/embedding_model_configs"
embedding_config_dir = "tests/configs/embedding_model_configs"
llm_config_dir = "tests/configs/llm_model_configs"