Cloud Providers

While On Device AI is designed for 100% offline operation, you can optionally connect to cloud AI providers when you need access to larger models or additional capabilities. Cloud is always opt-in and off by default.

Supported Providers

On Device AI supports the following cloud and local server providers:

Setting Up a Provider

  1. Open Settings → Cloud Providers

    Navigate to the Cloud Providers section in app settings.

  2. Select a provider

    Choose the provider you want to connect to.

  3. Enter your credentials

    For most providers, paste your API key from the provider's dashboard. For AWS Bedrock, enter your Access Key ID and Secret Access Key. For Qwen Portal, you can also use an OAuth refresh token. All credentials are stored securely in your device's Keychain — never in plain text.

  4. Select a model

    Browse available models from the provider and select the one you want to use. For providers without automatic model listing (Bedrock, Kimi), you can enter a model ID manually.

⚠️ Important

When using cloud providers, your conversation data is transmitted to the provider's servers. The app does not control how providers handle your data. Review each provider's privacy policy before use.

AWS Bedrock

AWS Bedrock requires AWS credentials rather than a simple API key:

  1. Set your AWS Region

    Enter the AWS region where Bedrock is enabled (e.g. us-east-1, us-west-2). This is saved in your app configuration, not in the Keychain.

  2. Enter AWS Credentials

    Tap Enter Credentials and provide your AWS Access Key ID and Secret Access Key. An optional session token is supported for temporary credentials (AWS STS).

  3. Enter a Bedrock Model ID

    Use the manual model entry field to type the Bedrock model ID, e.g. anthropic.claude-3-sonnet-20240229-v1:0 or amazon.titan-text-premier-v1:0.

ℹ️ How it works

Requests to Bedrock are authenticated using AWS Signature Version 4 (SigV4) — signed directly on your device. No credentials are ever transmitted to any proxy server.

Qwen Portal

Qwen Portal supports two authentication modes:

Select your preferred authentication mode in Settings → Cloud Providers → Qwen Portal before entering your credentials.

Switching Between Local & Cloud

You can switch between local and cloud models at any time, even within the same conversation:

ℹ️ Note

When you switch from a local model to a cloud model mid-conversation, your conversation history is sent to the cloud provider. Consider starting a new conversation if you have sensitive content.

Privacy Considerations

On Device AI is designed with privacy first:

Local Servers (Ollama & LM Studio)

Ollama and LM Studio are special cases — they run AI models on your own hardware (Mac, PC, or server) rather than in the cloud. This gives you the power of larger models while maintaining privacy:

Your data stays on your local network when using these providers. This is a great option for running larger models on a powerful Mac while chatting from your iPhone.

💡 Tip

On Device AI can also serve as a remote inference server itself (macOS). Connect your iPhone to your Mac running On Device AI for the best of both worlds.