Others
HTTP Status Code ExplanationAPI Online Call
Request Address
https://api.briskapi.com/model/deepseek
Request Method
POST
Request Authentication
Content-Type: application/json
Authorization: Bearer {API key}
Request Parameters
-
model
string
required
The identifier of the chat generation model to use:
example: "deepseek-v3", "deepseek-v3-0324", "deepseek-r1", "deepseek-r1-llama-70b", "deepseek-r1-qwen-32b"
-
messages
array
required
An array of message objects representing the conversation context for multi‑turn dialogue, where each object has a role field (e.g., "user" for user inputs, "assistant" for model replies) and a content field containing the text.
example: [{"role": "user", "content":"User input content"}]
messages.role: "user","system"
-
max_tokens
int
optional
The maximum number of tokens to include in the generated reply.
default: 8192
range: 1 - 8192
-
temperature
float64
optional
Controls randomness in the sampling process. Higher values (e.g., 0.8–1.0) make outputs more varied and creative, increasing the chance of different answers to the same prompt. Lower values (e.g., 0.2–0.5) produce more focused, predictable results.
default: 0.7
range: 0 - 1
-
top_k
int
optional
Sets how many of the highest‑probability tokens are considered at each step. Raising this (up to 6) boosts diversity and creativity, while lowering it (minimum 1) makes the model stick more closely to the strongest predictions and reduces variety.
default: 4
range: 1 - 6
-
stream
bool
optional
A boolean flag indicating whether to use streaming mode; if true, tokens are returned incrementally as they’re generated, and if false, the complete response is sent all at once.
default: false
"false","true"
-
search
bool
optional
When set to true, allows the model to perform online searches based on user input; when false, no web searches will be triggered.
default: false
"false","true",Online Search
Request Example CURL
curl https://api.briskapi.com/model/deepseek \
-H "Content-Type: application/json" \
-H "Authorization: Bearer {API key}" \
-d '{
"model": "deepseek-r1",
"messages": [
{
"content": "I am a Chinese language teacher.",
"role": "system"
},
{
"content": "hello",
"role": "user"
}
],
"max_tokens": 2048,
"temperature": 1,
"top_k": 0.7,
"stream":false,
"search": true
}'
Response Example
{
"choices": [{
"finish_reason": "stop",
"index": 0,
"message": {
"content": "Okay, the user wrote \"hello\". I should respond in a friendly and welcoming manner. Since they mentioned they're a Chinese language teacher, maybe I can offer help related to teaching Chinese. Let me ask if they need any assistance with lesson planning, resources, or student engagement strategies. Keeping it open-ended to invite specific requests.\n\n\nHello! How can I assist you today? Whether you need help with lesson planning, teaching resources, or creative ways to engage your students in learning Chinese, feel free to ask! 😊 ",
"role": "assistant"
}
}],
"created": "1745211400",
"id": "cmpl-f4dc264964eb9782e702ffdfeaec81e0",
"model": "deepseek-r1",
"object": "chat.completion",
"usage": {
"completion_tokens": 163,
"prompt_tokens": 10,
"total_tokens": 173
}
}
Response Description
-
id
string
A unique string assigned to this particular response, which you can use for logging, debugging, or auditing purposes.
-
object
string
Indicates the type of object returned. For chat completions, this will always be "chat.completion".
-
created
int
The time the response was generated, given as a Unix timestamp.
-
model
string
The name of the model that produced this response.
-
choices
array
An array of one or more possible replies generated by the model. You can request multiple (n) choices in one call.
-
choices.index
int
The zero-based position of this choice in the choices array
-
choices.message
object
An object describing the content and role of the message.
{"role": "assistant", "content":"Hello, I am a teacher"}
-
choices.finish_reason
string
A string explaining why the model stopped generating text.
-
usage
object
An object summarizing token counts for this request.
-
usage.prompt_tokens
int
The number of tokens consumed by your input (the prompt)
-
usage.completion_tokens
int
The number of tokens generated by the model in its reply.
-
usage.total_tokens
int
The sum of prompt_tokens and completion_tokens, representing the total token cost of the request.