Python Client: llama-stack-client for Python 3.12+ with full agent and model APIs
Vector Store APIs: Create and query vector stores from the client, including PGVector-backed stores when the server is configured with ENABLE_PGVECTOR=true
REST-Friendly: Server exposes APIs for inference, agents, and tool runtime; can be wrapped in FastAPI or other web frameworks for production use