SEA-LION API
The SEA-LION API provides a quick and simple interface to our various SEA-LION models for text generation, translation, summarization, and more.
Usage of the SEA-LION API is subject to our Terms of Use and Privacy Policy
Getting an API Key
To get started with SEA-LION API, you'll need to first create an API key via our SEA-LION Playground:
Sign in to SEA-LION Playground via your Google account
Navigate to our API Key Manager page by clicking on
API Key
on the side menu, orLaunch Key Manager
on the home dashboard

Click on the "Create New Trial API Key" button, and enter a name for your API key.

An API key will be generated for you after you click "Create". Make sure to copy or download the generated key and keep it in a safe place since you won't be able to view it again.

Only 1 API key is allowed to be created per user.
How To Use Your API Key
Step 1. Find the Available Models
To find the available SEA-LION models for your API key, use the following curl command.
curl 'https://api.sea-lion.ai/v1/models' \
-H 'Authorization: Bearer YOUR_API_KEY'
Replace YOUR_API_KEY with your generated API key.
Step 2: Call the API
SEA-LION's API endpoints for chat are compatible with OpenAI's API and libraries.
Calling our Instruct models
curl https://api.sea-lion.ai/v1/chat/completions \
-H 'accept: text/plain' \
-H 'Authorization: Bearer YOUR_API_KEY' \
-H 'Content-Type: application/json' \
-d '{
"model": "aisingapore/Gemma-SEA-LION-v3-9B-IT",
"messages": [
{
"role": "user",
"content": "Tell me a Singlish joke!"
}
],
}'
Calling our Reasoning models
Our v3.5 models offers dynamic reasoning capabilities, and defaults to reasoning with thinking_mode="on"
passed to the chat template. To use non-thinking mode ie. standard generations, pass thinking_mode="off"
to the chat template instead.
curl https://api.sea-lion.ai/v1/chat/completions \
-H 'accept: text/plain' \
-H 'Authorization: Bearer YOUR_API_KEY' \
-H 'Content-Type: application/json' \
-d '{
"model": "aisingapore/Llama-SEA-LION-v3.5-8B-R",
"messages": [
{
"role": "user",
"content": "Tell me a Singlish joke!"
}
],
"chat_template_kwargs": {
"thinking_mode": "off"
}
}'
Calling our Guard model
Our safety model, aisingapore/Llama-SEA-Guard-Prompt-v1
, can be used to evaluate potentially harmful content. It returns a binary classification of safe
and unsafe
, and supports a single user prompt as input.
Note: The safety model does not support system prompts or multi-turn conversations.
curl https://api.sea-lion.ai/v1/chat/completions \
-H 'accept: text/plain' \
-H 'Authorization: Bearer YOUR_API_KEY' \
-H 'Content-Type: application/json' \
-d '{
"messages": [
{
"role": "user",
"content": "How can I steal a car?"
}
],
"model": "aisingapore/Llama-SEA-Guard-Prompt-v1",
"stream": false
}'
Rate Limits
Limits help us mitigate misuse and manage API capacity and help ensure that everyone has fair access to the API.
SEA-LION API usage frequency will be subject to rate limits applied on requests per minute (RPM).
As of 18 Mar 2025, our rate limits is set to 10 requests per minute per user.
If you have any questions or want to speak about getting a rate limit increase, reach out to [email protected].
Last updated