Drop-in replacement for OpenAI APIs running on dedicated Apple Silicon. Same code, complete privacy, zero per-token costs.
Click any prompt to try it in the live demo below
Generate functions, classes, and complete modules with context-aware AI assistance.
Find bugs, security issues, and performance problems in your codebase.
Generate API docs, READMEs, and inline documentation from your code.
Get help with system design, database schemas, and architectural decisions.
Test the AI Metal Cluster API
Running locally. Your code never leaves this server.
Just change the base URL
from openai import OpenAI
# Before: OpenAI cloud
# client = OpenAI(api_key="sk-...")
# After: AI Metal Cluster
client = OpenAI(
base_url="http://your-cluster:5001/v1",
api_key="not-needed" # Local, no auth required
)
response = client.chat.completions.create(
model="llama-3.2-70b",
messages=[{"role": "user", "content": "Hello!"}]
)
print(response.choices[0].message.content)