OpenAI
OpenAI provides powerful AI services including language models, embeddings, and other machine learning capabilities. Rill supports connecting to OpenAI services using your own API key and configuration parameters to enable AI-powered conversations and data analysis features.
API Key
Rill will use your configured OpenAI connector if available, or fall back to its built-in LLM service if no custom configuration is provided. Once configured, your API key will be used for AI features in both Rill Developer and Rill Cloud, including the AI Agent builder. You can configure your API key in your project's .env file and reference the credentials in a connector YAML.
OpenAI API Key
To configure OpenAI access, you'll need to obtain an API key and configure it in your project.
-
Obtain your OpenAI API key from the OpenAI Platform.
-
Create the connector YAML:
Create
connectors/openai.yamlin your project:type: connector
driver: openai
api_key: "{{ .env.OPENAI_API_KEY }}" -
Set up environment variable:
If configuring manually, ensure your project's
.envfile contains the key before starting Rill:OPENAI_API_KEY=sk-... -
Configure OpenAI as the default AI connector:
Add the following to your
rill.yamlto use OpenAI as the AI provider for your project:ai_connector: openai
For details on managing credentials across environments, see Configure Local Credentials.
Configuration Options
For additional configuration options (model, base URL, API type, etc.), see the OpenAI connector reference.
Deploy to Rill Cloud
Rill requires you to explicitly provide an OpenAI API key to use the OpenAI connector. See the connector reference for details.
For details on pushing and pulling credentials between environments, see Configure Local Credentials.