mirror of
https://github.com/cpacker/MemGPT.git
synced 2025-06-03 04:30:22 +00:00
chore: deprecate letta configure
and remove config defaults (#1841)
Co-authored-by: cpacker <packercharles@gmail.com>
This commit is contained in:
parent
2492ad780d
commit
e094e18580
@ -1,7 +0,0 @@
|
||||
{
|
||||
"embedding_endpoint": "https://embeddings.memgpt.ai",
|
||||
"embedding_model": "BAAI/bge-large-en-v1.5",
|
||||
"embedding_dim": 1024,
|
||||
"embedding_chunk_size": 300,
|
||||
"embedding_endpoint_type": "hugging-face"
|
||||
}
|
@ -1,7 +0,0 @@
|
||||
{
|
||||
"embedding_endpoint": null,
|
||||
"embedding_model": "BAAI/bge-small-en-v1.5",
|
||||
"embedding_dim": 384,
|
||||
"embedding_chunk_size": 300,
|
||||
"embedding_endpoint_type": "local"
|
||||
}
|
@ -1,7 +0,0 @@
|
||||
{
|
||||
"embedding_endpoint_type": "ollama",
|
||||
"embedding_endpoint": "http://127.0.0.1:11434",
|
||||
"embedding_model": "mxbai-embed-large",
|
||||
"embedding_dim": 512,
|
||||
"embedding_chunk_size": 200
|
||||
}
|
@ -1,7 +0,0 @@
|
||||
{
|
||||
"embedding_endpoint_type": "openai",
|
||||
"embedding_endpoint": "https://api.openai.com/v1",
|
||||
"embedding_model": "text-embedding-ada-002",
|
||||
"embedding_dim": 1536,
|
||||
"embedding_chunk_size": 300
|
||||
}
|
@ -1,11 +0,0 @@
|
||||
{
|
||||
"context_window": 8192,
|
||||
"model_endpoint_type": "openai",
|
||||
"model_endpoint": "https://inference.memgpt.ai",
|
||||
"model": "memgpt-openai",
|
||||
"embedding_endpoint_type": "hugging-face",
|
||||
"embedding_endpoint": "https://embeddings.memgpt.ai",
|
||||
"embedding_model": "BAAI/bge-large-en-v1.5",
|
||||
"embedding_dim": 1024,
|
||||
"embedding_chunk_size": 300
|
||||
}
|
@ -1,7 +0,0 @@
|
||||
{
|
||||
"context_window": 200000,
|
||||
"model": "claude-3-opus-20240229",
|
||||
"model_endpoint_type": "anthropic",
|
||||
"model_endpoint": "https://api.anthropic.com/v1",
|
||||
"model_wrapper": null
|
||||
}
|
@ -1,7 +0,0 @@
|
||||
{
|
||||
"context_window": 8192,
|
||||
"model": "gpt-4",
|
||||
"model_endpoint_type": "openai",
|
||||
"model_endpoint": "https://api.openai.com/v1",
|
||||
"model_wrapper": null
|
||||
}
|
@ -1,7 +0,0 @@
|
||||
{
|
||||
"context_window": 8192,
|
||||
"model": "llama3-groq-70b-8192-tool-use-preview",
|
||||
"model_endpoint_type": "groq",
|
||||
"model_endpoint": "https://api.groq.com/openai/v1",
|
||||
"model_wrapper": null
|
||||
}
|
@ -1,6 +0,0 @@
|
||||
{
|
||||
"context_window": 16384,
|
||||
"model_endpoint_type": "openai",
|
||||
"model_endpoint": "https://inference.memgpt.ai",
|
||||
"model": "memgpt-openai"
|
||||
}
|
@ -1,6 +0,0 @@
|
||||
{
|
||||
"context_window": 8192,
|
||||
"model_endpoint_type": "ollama",
|
||||
"model_endpoint": "http://127.0.0.1:11434",
|
||||
"model": "dolphin2.2-mistral:7b-q6_K"
|
||||
}
|
@ -1,12 +0,0 @@
|
||||
{
|
||||
"context_window": 8192,
|
||||
"model": "gpt-4",
|
||||
"model_endpoint_type": "openai",
|
||||
"model_endpoint": "https://api.openai.com/v1",
|
||||
"model_wrapper": null,
|
||||
"embedding_endpoint_type": "openai",
|
||||
"embedding_endpoint": "https://api.openai.com/v1",
|
||||
"embedding_model": "text-embedding-ada-002",
|
||||
"embedding_dim": 1536,
|
||||
"embedding_chunk_size": 300
|
||||
}
|
File diff suppressed because it is too large
Load Diff
@ -1,13 +0,0 @@
|
||||
{
|
||||
"context_window": 200000,
|
||||
"model": "claude-3-opus-20240229",
|
||||
"model_endpoint_type": "anthropic",
|
||||
"model_endpoint": "https://api.anthropic.com/v1",
|
||||
"model_wrapper": null,
|
||||
"embedding_endpoint_type": "hugging-face",
|
||||
"embedding_endpoint": "https://embeddings.memgpt.ai",
|
||||
"embedding_model": "BAAI/bge-large-en-v1.5",
|
||||
"embedding_dim": 1024,
|
||||
"embedding_chunk_size": 300
|
||||
|
||||
}
|
@ -1,11 +0,0 @@
|
||||
{
|
||||
"context_window": 8192,
|
||||
"model_endpoint_type": "openai",
|
||||
"model_endpoint": "https://inference.memgpt.ai",
|
||||
"model": "memgpt-openai",
|
||||
"embedding_endpoint_type": "hugging-face",
|
||||
"embedding_endpoint": "https://embeddings.memgpt.ai",
|
||||
"embedding_model": "BAAI/bge-large-en-v1.5",
|
||||
"embedding_dim": 1024,
|
||||
"embedding_chunk_size": 300
|
||||
}
|
@ -1,12 +0,0 @@
|
||||
{
|
||||
"context_window": 8192,
|
||||
"model": "gpt-4",
|
||||
"model_endpoint_type": "openai",
|
||||
"model_endpoint": "https://api.openai.com/v1",
|
||||
"model_wrapper": null,
|
||||
"embedding_endpoint_type": "openai",
|
||||
"embedding_endpoint": "https://api.openai.com/v1",
|
||||
"embedding_model": "text-embedding-ada-002",
|
||||
"embedding_dim": 1536,
|
||||
"embedding_chunk_size": 300
|
||||
}
|
@ -13,7 +13,6 @@ from letta.schemas.llm_config import LLMConfig
|
||||
|
||||
|
||||
class Provider(BaseModel):
|
||||
base_url: str
|
||||
|
||||
def list_llm_models(self):
|
||||
return []
|
||||
@ -25,6 +24,32 @@ class Provider(BaseModel):
|
||||
pass
|
||||
|
||||
|
||||
class LettaProvider(Provider):
|
||||
|
||||
name: str = "letta"
|
||||
|
||||
def list_llm_models(self) -> List[LLMConfig]:
|
||||
return [
|
||||
LLMConfig(
|
||||
model="memgpt-openai",
|
||||
model_endpoint_type="openai",
|
||||
model_endpoint="https://inference.memgpt.ai",
|
||||
context_window=16384,
|
||||
)
|
||||
]
|
||||
|
||||
def list_embedding_models(self):
|
||||
return [
|
||||
EmbeddingConfig(
|
||||
embedding_model="BAAI/bge-large-en-v1.5",
|
||||
embedding_endpoint_type="hugging-face",
|
||||
embedding_endpoint="https://embeddings.memgpt.ai",
|
||||
embedding_dim=1024,
|
||||
embedding_chunk_size=300,
|
||||
)
|
||||
]
|
||||
|
||||
|
||||
class OpenAIProvider(Provider):
|
||||
name: str = "openai"
|
||||
api_key: str = Field(..., description="API key for the OpenAI API.")
|
||||
|
@ -47,6 +47,7 @@ from letta.providers import (
|
||||
AnthropicProvider,
|
||||
AzureProvider,
|
||||
GoogleAIProvider,
|
||||
LettaProvider,
|
||||
OllamaProvider,
|
||||
OpenAIProvider,
|
||||
VLLMProvider,
|
||||
@ -259,8 +260,8 @@ class SyncServer(Server):
|
||||
# add global default tools (for admin)
|
||||
self.add_default_tools(module_name="base")
|
||||
|
||||
# collect providers
|
||||
self._enabled_providers = []
|
||||
# collect providers (always has Letta as a default)
|
||||
self._enabled_providers = [LettaProvider()]
|
||||
if model_settings.openai_api_key:
|
||||
self._enabled_providers.append(OpenAIProvider(api_key=model_settings.openai_api_key))
|
||||
if model_settings.anthropic_api_key:
|
||||
|
@ -49,8 +49,8 @@ namespace = uuid.NAMESPACE_DNS
|
||||
agent_uuid = str(uuid.uuid5(namespace, "test-endpoints-agent"))
|
||||
|
||||
# defaults (letta hosted)
|
||||
EMBEDDING_CONFIG_PATH = "configs/embedding_model_configs/letta-hosted.json"
|
||||
LLM_CONFIG_PATH = "configs/llm_model_configs/letta-hosted.json"
|
||||
EMBEDDING_CONFIG_PATH = "tests/configs/embedding_model_configs/letta-hosted.json"
|
||||
LLM_CONFIG_PATH = "tests/configs/llm_model_configs/letta-hosted.json"
|
||||
|
||||
|
||||
# ======================================================================================================================
|
||||
|
@ -13,7 +13,7 @@ from tests.helpers.endpoints_helper import (
|
||||
)
|
||||
|
||||
# directories
|
||||
embedding_config_dir = "configs/embedding_model_configs"
|
||||
embedding_config_dir = "tests/configs/embedding_model_configs"
|
||||
llm_config_dir = "tests/configs/llm_model_configs"
|
||||
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user