Skip to main content
GET
/
api
/
v3
/
evaluations
/
{evaluation_id}
/
results
Get Evaluation Results
curl --request GET \
  --url https://api.nugen.in/api/v3/evaluations/{evaluation_id}/results \
  --header 'Authorization: Bearer <token>'
{
  "evaluation_id": "<string>",
  "model_id": "<string>",
  "benchmark_id": "<string>",
  "status": "<string>",
  "raw_answers_count": 123,
  "completed_at": "<string>",
  "metrics": {},
  "method": "<string>",
  "model_id_2": "<string>",
  "base_model": {},
  "eval_model": {},
  "comparison": {}
}

Authorizations

Authorization
string
header
required

Bearer authentication header of the form Bearer <token>, where <token> is your auth token.

Path Parameters

evaluation_id
string
required

Response

Returns detailed evaluation metrics and scores for a completed evaluation. This endpoint provides comprehensive results for a finished evaluation, including all relevant metrics, scores, and comparison data if applicable. Use this to analyze the performance of the evaluated model(s) against the benchmark once the evaluation is complete.

Response schema for evaluation results

evaluation_id
string
required

Unique identifier for the evaluation

model_id
string
required

ID of the model that was evaluated

benchmark_id
string
required

Benchmark ID used

status
string
required

Evaluation status

raw_answers_count
integer
required

Number of raw answers generated

completed_at
string
required

ISO timestamp when evaluation completed

metrics
Metrics · object

Evaluation metrics and scores (single model)

method
string | null

Evaluation method: 'eval' or 'eval-compare'

model_id_2
string | null

ID of second model (for comparison)

base_model
Base Model · object

Base model results (comparison mode)

eval_model
Eval Model · object

Eval model results (comparison mode)

comparison
Comparison · object

Comparison results between models