Inference API
Generate
Generate tokens and wait for completion
POST
/
generate
{
"details": {
"best_of_sequences": [
{
"finish_reason": "length",
"generated_text": "test",
"generated_tokens": 1,
"prefill": [
{
"id": 0,
"logprob": -0.34,
"text": "test"
}
],
"seed": 42,
"tokens": [
{
"id": 0,
"logprob": -0.34,
"special": "false",
"text": "test"
}
]
}
],
"finish_reason": "length",
"prompt_tokens": 1,
"generated_tokens": 1,
"prefill": [
{
"id": 0,
"logprob": -0.34,
"text": "test"
}
],
"seed": 42,
"tokens": [
{
"id": 0,
"logprob": -0.34,
"special": "false",
"text": "test"
}
]
},
"generated_text": "test"
}
Body
application/json
Response
200
application/json
Generated Text
The response is of type object
.
{
"details": {
"best_of_sequences": [
{
"finish_reason": "length",
"generated_text": "test",
"generated_tokens": 1,
"prefill": [
{
"id": 0,
"logprob": -0.34,
"text": "test"
}
],
"seed": 42,
"tokens": [
{
"id": 0,
"logprob": -0.34,
"special": "false",
"text": "test"
}
]
}
],
"finish_reason": "length",
"prompt_tokens": 1,
"generated_tokens": 1,
"prefill": [
{
"id": 0,
"logprob": -0.34,
"text": "test"
}
],
"seed": 42,
"tokens": [
{
"id": 0,
"logprob": -0.34,
"special": "false",
"text": "test"
}
]
},
"generated_text": "test"
}
Assistant
Responses are generated using AI and may contain mistakes.