Quick Start
Get your first prompt running in Scope in under 5 minutes. This guide walks you through signing in, configuring a provider, creating a prompt, testing it, and promoting it to production.
Prerequisites​
- A base14 account with access to Scope
- An API key from at least one LLM provider (e.g., OpenAI, Anthropic)
Step 1: Sign in to Scope​
Navigate to your base14 dashboard and open Scope. Scope is available to all base14 accounts — no additional setup is required.
Step 2: Configure a Provider​
Before you can test prompts, Scope needs credentials for at least one LLM provider.
- Go to Settings > Providers
- Click Add Provider
- Select your provider (e.g., OpenAI)
- Enter your API key
- Click Test Connection to verify the key is valid
- Save the provider configuration
Once connected, enable the models you want to use (e.g., gpt-4o,
claude-3-opus).
You can configure multiple providers and switch between them when testing prompts. See Configure Providers for details.
Step 3: Create a Prompt​
- Click New Prompt from the prompt list
- Enter a name (e.g.,
greeting) - Add your prompt content:
You are a friendly assistant. Greet the user by name and suggest
a fun activity for them to try today.
User name: {{name}}
Activity preference: {{preference}}
Scope automatically detects {{name}} and {{preference}} as variables.
- Click Create — this creates version
v1in draft status
Step 4: Test the Prompt​
- Open the prompt you just created
- Click Test to open the test panel
- Select a provider and model (e.g., OpenAI /
gpt-4o) - Fill in the variable values:
name:Alicepreference:outdoor
- Click Run
The test panel shows the LLM response along with token usage, latency, and estimated cost. Adjust your prompt and re-run until you're satisfied.
Step 5: Promote to Production​
When the prompt is ready:
- Click Promote on the version you want to deploy
- Add optional promotion notes (e.g., "Initial release")
- Confirm the promotion
The version status changes from draft to published. Any application
using the SDK or API will now receive this version when requesting the
greeting prompt.
Step 6: Fetch from Your Application​
Install the Scope SDK and fetch the prompt at runtime:
- Python
- Ruby
from scope_client import ScopeClient, ApiKeyCredentials
credentials = ApiKeyCredentials.from_env()
client = ScopeClient(credentials=credentials)
version = client.get_prompt_version("greeting")
rendered = version.render(name="Alice", preference="outdoor")
print(rendered)
require "scope_client"
credentials = ScopeClient::Credentials::ApiKey.from_env
client = ScopeClient::Client.new(credentials: credentials)
version = client.get_prompt_version("greeting")
rendered = version.render(name: "Alice", preference: "outdoor")
puts rendered
Pass the rendered string to your LLM provider of choice. See the SDK Quickstart for a full end-to-end example.
Next Steps​
- Core Concepts — understand prompts, versions, variables, and promotion
- Configure Providers — add and manage LLM providers
- SDK Installation — install the SDK for Python or Ruby
- API Reference — use the REST API directly