Nscale (EU Sovereign)
https://docs.nscale.com/docs/inference/chat
tip
We support ALL Nscale models, just set model=nscale/<any-model-on-nscale> as a prefix when sending litellm requests
| Property | Details | 
|---|---|
| Description | European-domiciled full-stack AI cloud platform for LLMs and image generation. | 
| Provider Route on LiteLLM | nscale/ | 
| Supported Endpoints | /chat/completions,/images/generations | 
| API Reference | Nscale docs | 
Required Variablesโ
Environment Variables
os.environ["NSCALE_API_KEY"] = ""  # your Nscale API key
Explore Available Modelsโ
Explore our full list of text and multimodal AI models โ all available at highly competitive pricing: ๐ Full List of Models
Key Featuresโ
- EU Sovereign: Full data sovereignty and compliance with European regulations
- Ultra-Low Cost (starting at $0.01 / M tokens): Extremely competitive pricing for both text and image generation models
- Production Grade: Reliable serverless deployments with full isolation
- No Setup Required: Instant access to compute without infrastructure management
- Full Control: Your data remains private and isolated
Usage - LiteLLM Python SDKโ
Text Generationโ
Nscale Text Generation
from litellm import completion
import os
os.environ["NSCALE_API_KEY"] = ""  # your Nscale API key
response = completion(
    model="nscale/meta-llama/Llama-4-Scout-17B-16E-Instruct",
    messages=[{"role": "user", "content": "What is LiteLLM?"}]
)
print(response)
Nscale Text Generation - Streaming
from litellm import completion
import os
os.environ["NSCALE_API_KEY"] = ""  # your Nscale API key
stream = completion(
    model="nscale/meta-llama/Llama-4-Scout-17B-16E-Instruct",
    messages=[{"role": "user", "content": "What is LiteLLM?"}],
    stream=True
)
for chunk in stream:
    if chunk.choices[0].delta.content is not None:
        print(chunk.choices[0].delta.content, end="")
Image Generationโ
Nscale Image Generation
from litellm import image_generation
import os
os.environ["NSCALE_API_KEY"] = ""  # your Nscale API key
response = image_generation(
    model="nscale/stabilityai/stable-diffusion-xl-base-1.0",
    prompt="A beautiful sunset over mountains",
    n=1,
    size="1024x1024"
)
print(response)
Usage - LiteLLM Proxyโ
Add the following to your LiteLLM Proxy configuration file:
config.yaml
model_list:
  - model_name: nscale/meta-llama/Llama-4-Scout-17B-16E-Instruct
    litellm_params:
      model: nscale/meta-llama/Llama-4-Scout-17B-16E-Instruct
      api_key: os.environ/NSCALE_API_KEY
  - model_name: nscale/meta-llama/Llama-3.3-70B-Instruct
    litellm_params:
      model: nscale/meta-llama/Llama-3.3-70B-Instruct
      api_key: os.environ/NSCALE_API_KEY
  - model_name: nscale/stabilityai/stable-diffusion-xl-base-1.0
    litellm_params:
      model: nscale/stabilityai/stable-diffusion-xl-base-1.0
      api_key: os.environ/NSCALE_API_KEY
Start your LiteLLM Proxy server:
Start LiteLLM Proxy
litellm --config config.yaml
# RUNNING on http://0.0.0.0:4000
- OpenAI SDK
- LiteLLM SDK
- cURL
Nscale via Proxy - Non-streaming
from openai import OpenAI
# Initialize client with your proxy URL
client = OpenAI(
    base_url="http://localhost:4000",  # Your proxy URL
    api_key="your-proxy-api-key"       # Your proxy API key
)
# Non-streaming response
response = client.chat.completions.create(
    model="nscale/meta-llama/Llama-4-Scout-17B-16E-Instruct",
    messages=[{"role": "user", "content": "What is LiteLLM?"}]
)
print(response.choices[0].message.content)
Nscale via Proxy - LiteLLM SDK
import litellm
# Configure LiteLLM to use your proxy
response = litellm.completion(
    model="litellm_proxy/nscale/meta-llama/Llama-4-Scout-17B-16E-Instruct",
    messages=[{"role": "user", "content": "What is LiteLLM?"}],
    api_base="http://localhost:4000",
    api_key="your-proxy-api-key"
)
print(response.choices[0].message.content)
Nscale via Proxy - cURL
curl http://localhost:4000/v1/chat/completions \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer your-proxy-api-key" \
  -d '{
    "model": "nscale/meta-llama/Llama-4-Scout-17B-16E-Instruct",
    "messages": [{"role": "user", "content": "What is LiteLLM?"}]
  }'
Getting Startedโ
- Create an account at console.nscale.com
- Claim free credit
- Create an API key in settings
- Start making API calls using LiteLLM