OpenAPI Contract Test Report
Summary
- Total Tests: 36
- Passed: ✅ 36
- Failed: ❌ 0
Test #1 ✅
📋 Test case from OpenAPI example
Endpoint: GET /health
Expected Response
Status: 200
Actual Response
Status: 200
Test #2 ✅
📋 Test case from OpenAPI example
Endpoint: GET /private/v1/providers/{provider}/max-context-window
Expected Response
Status: 200
Actual Response
Status: 404
Test #3 ✅
📋 Test case from OpenAPI example
Endpoint: GET /private/v1/providers
Expected Response
Status: 200
{
"available": [
"pixtral",
"gemma3_on_vpn"
],
"default": "pixtral",
"total": 2,
"status": "ready"
}
Actual Response
Status: 200
{
"available": [
"local_gemma3_270m"
],
"default": "local_gemma3_270m",
"total": 12,
"status": "one_provider_available"
}
Test #4 ✅
📋 Test case from OpenAPI example
Endpoint: POST /v1/agent/chat
Request Body
{
"message": "What's the weather?",
"conversation_id": "conv-123",
"user_id": "user-456",
"stream": false
}
Expected Response
Status: 200
{
"conversation_id": "conv-123",
"user_id": "user-456",
"message": "The weather is sunny today!",
"role": "assistant",
"created": 1703347200,
"usage": {
"prompt_tokens": 56,
"completion_tokens": 31,
"total_tokens": 87
}
}
Actual Response
Status: 200
{
"conversation_id": "conv-123",
"user_id": "user-456",
"message": "The weather is sunny with a high of 78 degrees Fahrenheit.\n",
"role": "assistant",
"created": 1769205864,
"usage": {
"prompt_tokens": 46,
"completion_tokens": 16,
"total_tokens": 62
}
}
Test #5 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/agent/chat
Request Body
{
"message": "Lorem ipsum dolor sit amet",
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Lorem ipsum dolor sit amet",
"stream": true
}
Expected Response
Status: 200
{
"conversation_id": "conv-123",
"user_id": "user-456",
"message": "The weather is sunny today!",
"role": "assistant",
"created": 1703347200,
"usage": {
"prompt_tokens": 56,
"completion_tokens": 31,
"total_tokens": 87
}
}
Actual Response
Status: 501
Test #6 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/agent/chat
Request Body
{
"message": "Lorem ipsum dolor sit amet",
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Lorem ipsum dolor sit amet",
"stream": false
}
Expected Response
Status: 200
{
"conversation_id": "conv-123",
"user_id": "user-456",
"message": "The weather is sunny today!",
"role": "assistant",
"created": 1703347200,
"usage": {
"prompt_tokens": 56,
"completion_tokens": 31,
"total_tokens": 87
}
}
Actual Response
Status: 200
{
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Lorem ipsum dolor sit amet",
"message": "Lorem ipsum dolor sit amet.\n",
"role": "assistant",
"created": 1769205865,
"usage": {
"prompt_tokens": 45,
"completion_tokens": 8,
"total_tokens": 53
}
}
Test #7 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/agent/chat
Request Body
{
"message": "Lorem ipsum dolor sit amet",
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Test with 'single' quotes",
"stream": true
}
Expected Response
Status: 200
{
"conversation_id": "conv-123",
"user_id": "user-456",
"message": "The weather is sunny today!",
"role": "assistant",
"created": 1703347200,
"usage": {
"prompt_tokens": 56,
"completion_tokens": 31,
"total_tokens": 87
}
}
Actual Response
Status: 501
Test #8 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/agent/chat
Request Body
{
"message": "Lorem ipsum dolor sit amet",
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Test with 'single' quotes",
"stream": false
}
Expected Response
Status: 200
{
"conversation_id": "conv-123",
"user_id": "user-456",
"message": "The weather is sunny today!",
"role": "assistant",
"created": 1703347200,
"usage": {
"prompt_tokens": 56,
"completion_tokens": 31,
"total_tokens": 87
}
}
Actual Response
Status: 200
{
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Test with 'single' quotes",
"message": "",
"role": "assistant",
"created": 1769205865,
"usage": {
"prompt_tokens": 45,
"completion_tokens": 1,
"total_tokens": 46
}
}
Test #9 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/agent/chat
Request Body
{
"message": "Lorem ipsum dolor sit amet",
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Test with \"double\" quotes",
"stream": true
}
Expected Response
Status: 200
{
"conversation_id": "conv-123",
"user_id": "user-456",
"message": "The weather is sunny today!",
"role": "assistant",
"created": 1703347200,
"usage": {
"prompt_tokens": 56,
"completion_tokens": 31,
"total_tokens": 87
}
}
Actual Response
Status: 501
Test #10 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/agent/chat
Request Body
{
"message": "Lorem ipsum dolor sit amet",
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Test with \"double\" quotes",
"stream": false
}
Expected Response
Status: 200
{
"conversation_id": "conv-123",
"user_id": "user-456",
"message": "The weather is sunny today!",
"role": "assistant",
"created": 1703347200,
"usage": {
"prompt_tokens": 56,
"completion_tokens": 31,
"total_tokens": 87
}
}
Actual Response
Status: 200
{
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Test with \"double\" quotes",
"message": "Lorem ipsum dolor sit amet\n",
"role": "assistant",
"created": 1769205866,
"usage": {
"prompt_tokens": 45,
"completion_tokens": 7,
"total_tokens": 52
}
}
Test #11 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/agent/chat
Request Body
{
"message": "Lorem ipsum dolor sit amet",
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Test:with:colons",
"stream": true
}
Expected Response
Status: 200
{
"conversation_id": "conv-123",
"user_id": "user-456",
"message": "The weather is sunny today!",
"role": "assistant",
"created": 1703347200,
"usage": {
"prompt_tokens": 56,
"completion_tokens": 31,
"total_tokens": 87
}
}
Actual Response
Status: 501
Test #12 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/agent/chat
Request Body
{
"message": "Lorem ipsum dolor sit amet",
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Test:with:colons",
"stream": false
}
Expected Response
Status: 200
{
"conversation_id": "conv-123",
"user_id": "user-456",
"message": "The weather is sunny today!",
"role": "assistant",
"created": 1703347200,
"usage": {
"prompt_tokens": 56,
"completion_tokens": 31,
"total_tokens": 87
}
}
Actual Response
Status: 200
{
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Test:with:colons",
"message": "Lorem ipsum dolor sit amet.\n",
"role": "assistant",
"created": 1769205866,
"usage": {
"prompt_tokens": 45,
"completion_tokens": 8,
"total_tokens": 53
}
}
Test #13 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/agent/chat
Request Body
{
"message": "Lorem ipsum dolor sit amet",
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Test\\with\\backslashes",
"stream": true
}
Expected Response
Status: 200
{
"conversation_id": "conv-123",
"user_id": "user-456",
"message": "The weather is sunny today!",
"role": "assistant",
"created": 1703347200,
"usage": {
"prompt_tokens": 56,
"completion_tokens": 31,
"total_tokens": 87
}
}
Actual Response
Status: 501
Test #14 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/agent/chat
Request Body
{
"message": "Lorem ipsum dolor sit amet",
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Test\\with\\backslashes",
"stream": false
}
Expected Response
Status: 200
{
"conversation_id": "conv-123",
"user_id": "user-456",
"message": "The weather is sunny today!",
"role": "assistant",
"created": 1703347200,
"usage": {
"prompt_tokens": 56,
"completion_tokens": 31,
"total_tokens": 87
}
}
Actual Response
Status: 200
{
"conversation_id": "Lorem ipsum dolor sit amet",
"user_id": "Test\\with\\backslashes",
"message": "Lorem ipsum dolor sit amet.",
"role": "assistant",
"created": 1769205867,
"usage": {
"prompt_tokens": 45,
"completion_tokens": 7,
"total_tokens": 52
}
}
Test #15 ✅
📋 Test case from OpenAPI example
Endpoint: POST /v1/chat/completions
Request Body
{
"model": "pixtral",
"messages": [
{
"role": "user",
"content": "What is the capital of France?"
}
],
"temperature": 0.7,
"max_tokens": 100
}
Expected Response
Status: 200
{
"id": "chatcmpl-abc123",
"object": "chat.completion",
"created": 1734700000,
"model": "pixtral",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello! How can I help you today?"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 20,
"completion_tokens": 10,
"total_tokens": 30
}
}
Actual Response
Status: 200
{
"id": "chatcmpl-7b6b1d8ba96446759f767f64",
"object": "chat.completion",
"created": 1769205867,
"model": "pixtral",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "The capital of France is Paris.\n"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 22,
"completion_tokens": 9,
"total_tokens": 31
}
}
Test #16 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/chat/completions
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"messages": [
{
"role": "system",
"content": "Lorem ipsum dolor sit amet"
}
],
"temperature": 0.0,
"max_tokens": 1,
"stream": true,
"top_p": 0.0,
"stop": []
}
Expected Response
Status: 200
{
"id": "chatcmpl-abc123",
"object": "chat.completion",
"created": 1734700000,
"model": "pixtral",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello! How can I help you today?"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 20,
"completion_tokens": 10,
"total_tokens": 30
}
}
Actual Response
Status: 501
Test #17 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/chat/completions
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"messages": [
{
"role": "system",
"content": "Lorem ipsum dolor sit amet"
}
],
"temperature": 0.0,
"max_tokens": 1,
"stream": true,
"top_p": 0.0,
"stop": [
"Lorem ipsum dolor sit amet"
]
}
Expected Response
Status: 200
{
"id": "chatcmpl-abc123",
"object": "chat.completion",
"created": 1734700000,
"model": "pixtral",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello! How can I help you today?"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 20,
"completion_tokens": 10,
"total_tokens": 30
}
}
Actual Response
Status: 501
Test #18 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/chat/completions
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"messages": [
{
"role": "system",
"content": "Lorem ipsum dolor sit amet"
}
],
"temperature": 0.0,
"max_tokens": 1,
"stream": true,
"top_p": 0.0,
"stop": [
"Lorem ipsum dolor sit amet",
"Test with 'single' quotes",
"Test with \"double\" quotes"
]
}
Expected Response
Status: 200
{
"id": "chatcmpl-abc123",
"object": "chat.completion",
"created": 1734700000,
"model": "pixtral",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello! How can I help you today?"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 20,
"completion_tokens": 10,
"total_tokens": 30
}
}
Actual Response
Status: 501
Test #19 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/chat/completions
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"messages": [
{
"role": "system",
"content": "Lorem ipsum dolor sit amet"
}
],
"temperature": 0.0,
"max_tokens": 1,
"stream": true,
"top_p": 0.5,
"stop": []
}
Expected Response
Status: 200
{
"id": "chatcmpl-abc123",
"object": "chat.completion",
"created": 1734700000,
"model": "pixtral",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello! How can I help you today?"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 20,
"completion_tokens": 10,
"total_tokens": 30
}
}
Actual Response
Status: 501
Test #20 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/chat/completions
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"messages": [
{
"role": "system",
"content": "Lorem ipsum dolor sit amet"
}
],
"temperature": 0.0,
"max_tokens": 1,
"stream": true,
"top_p": 0.5,
"stop": [
"Lorem ipsum dolor sit amet"
]
}
Expected Response
Status: 200
{
"id": "chatcmpl-abc123",
"object": "chat.completion",
"created": 1734700000,
"model": "pixtral",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello! How can I help you today?"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 20,
"completion_tokens": 10,
"total_tokens": 30
}
}
Actual Response
Status: 501
Test #21 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/chat/completions
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"messages": [
{
"role": "system",
"content": "Lorem ipsum dolor sit amet"
}
],
"temperature": 0.0,
"max_tokens": 1,
"stream": true,
"top_p": 0.5,
"stop": [
"Lorem ipsum dolor sit amet",
"Test with 'single' quotes",
"Test with \"double\" quotes"
]
}
Expected Response
Status: 200
{
"id": "chatcmpl-abc123",
"object": "chat.completion",
"created": 1734700000,
"model": "pixtral",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello! How can I help you today?"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 20,
"completion_tokens": 10,
"total_tokens": 30
}
}
Actual Response
Status: 501
Test #22 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/chat/completions
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"messages": [
{
"role": "system",
"content": "Lorem ipsum dolor sit amet"
}
],
"temperature": 0.0,
"max_tokens": 1,
"stream": true,
"top_p": 1.0,
"stop": []
}
Expected Response
Status: 200
{
"id": "chatcmpl-abc123",
"object": "chat.completion",
"created": 1734700000,
"model": "pixtral",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello! How can I help you today?"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 20,
"completion_tokens": 10,
"total_tokens": 30
}
}
Actual Response
Status: 501
Test #23 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/chat/completions
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"messages": [
{
"role": "system",
"content": "Lorem ipsum dolor sit amet"
}
],
"temperature": 0.0,
"max_tokens": 1,
"stream": true,
"top_p": 1.0,
"stop": [
"Lorem ipsum dolor sit amet"
]
}
Expected Response
Status: 200
{
"id": "chatcmpl-abc123",
"object": "chat.completion",
"created": 1734700000,
"model": "pixtral",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello! How can I help you today?"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 20,
"completion_tokens": 10,
"total_tokens": 30
}
}
Actual Response
Status: 501
Test #24 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/chat/completions
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"messages": [
{
"role": "system",
"content": "Lorem ipsum dolor sit amet"
}
],
"temperature": 0.0,
"max_tokens": 1,
"stream": true,
"top_p": 1.0,
"stop": [
"Lorem ipsum dolor sit amet",
"Test with 'single' quotes",
"Test with \"double\" quotes"
]
}
Expected Response
Status: 200
{
"id": "chatcmpl-abc123",
"object": "chat.completion",
"created": 1734700000,
"model": "pixtral",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello! How can I help you today?"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 20,
"completion_tokens": 10,
"total_tokens": 30
}
}
Actual Response
Status: 501
Test #25 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/chat/completions
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"messages": [
{
"role": "system",
"content": "Lorem ipsum dolor sit amet"
}
],
"temperature": 0.0,
"max_tokens": 1,
"stream": false,
"top_p": 0.0,
"stop": []
}
Expected Response
Status: 200
{
"id": "chatcmpl-abc123",
"object": "chat.completion",
"created": 1734700000,
"model": "pixtral",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Hello! How can I help you today?"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 20,
"completion_tokens": 10,
"total_tokens": 30
}
}
Actual Response
Status: 200
{
"id": "chatcmpl-1dd6396394bc40619ced8970",
"object": "chat.completion",
"created": 1769205867,
"model": "Lorem ipsum dolor sit amet",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": ""
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 20,
"completion_tokens": 1,
"total_tokens": 21
}
}
Test #26 ✅
📋 Test case from OpenAPI example
Endpoint: POST /v1/api/generate
Request Body
{
"model": "gemma3:4b",
"prompt": "What is the capital of France?",
"stream": false,
"temperature": 0.7
}
Expected Response
Status: 200
{
"model": "gemma3:4b",
"created_at": "2024-12-20T00:00:00.000000Z",
"response": "The capital of France is Paris.",
"done": true,
"context": [],
"total_duration": 0,
"load_duration": 0,
"prompt_eval_count": 10,
"prompt_eval_duration": 0,
"eval_count": 10,
"eval_duration": 0
}
Actual Response
Status: 200
{
"model": "gemma3:4b",
"created_at": "2026-01-23T22:04:28.350676Z",
"response": "The capital of France is Paris.",
"done": true,
"context": [],
"total_duration": 0,
"load_duration": 0,
"prompt_eval_count": 22,
"prompt_eval_duration": 0,
"eval_count": 8,
"eval_duration": 0
}
Test #27 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/api/generate
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"prompt": "Lorem ipsum dolor sit amet",
"stream": true,
"temperature": 0.0,
"top_p": 0.0,
"top_k": 1
}
Expected Response
Status: 200
{
"model": "gemma3:4b",
"created_at": "2024-12-20T00:00:00.000000Z",
"response": "The capital of France is Paris.",
"done": true,
"context": [],
"total_duration": 0,
"load_duration": 0,
"prompt_eval_count": 10,
"prompt_eval_duration": 0,
"eval_count": 10,
"eval_duration": 0
}
Actual Response
Status: 501
Test #28 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/api/generate
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"prompt": "Lorem ipsum dolor sit amet",
"stream": true,
"temperature": 0.0,
"top_p": 0.0,
"top_k": 500000
}
Expected Response
Status: 200
{
"model": "gemma3:4b",
"created_at": "2024-12-20T00:00:00.000000Z",
"response": "The capital of France is Paris.",
"done": true,
"context": [],
"total_duration": 0,
"load_duration": 0,
"prompt_eval_count": 10,
"prompt_eval_duration": 0,
"eval_count": 10,
"eval_duration": 0
}
Actual Response
Status: 501
Test #29 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/api/generate
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"prompt": "Lorem ipsum dolor sit amet",
"stream": true,
"temperature": 0.0,
"top_p": 0.0,
"top_k": 1000000
}
Expected Response
Status: 200
{
"model": "gemma3:4b",
"created_at": "2024-12-20T00:00:00.000000Z",
"response": "The capital of France is Paris.",
"done": true,
"context": [],
"total_duration": 0,
"load_duration": 0,
"prompt_eval_count": 10,
"prompt_eval_duration": 0,
"eval_count": 10,
"eval_duration": 0
}
Actual Response
Status: 501
Test #30 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/api/generate
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"prompt": "Lorem ipsum dolor sit amet",
"stream": true,
"temperature": 0.0,
"top_p": 0.5,
"top_k": 1
}
Expected Response
Status: 200
{
"model": "gemma3:4b",
"created_at": "2024-12-20T00:00:00.000000Z",
"response": "The capital of France is Paris.",
"done": true,
"context": [],
"total_duration": 0,
"load_duration": 0,
"prompt_eval_count": 10,
"prompt_eval_duration": 0,
"eval_count": 10,
"eval_duration": 0
}
Actual Response
Status: 501
Test #31 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/api/generate
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"prompt": "Lorem ipsum dolor sit amet",
"stream": true,
"temperature": 0.0,
"top_p": 0.5,
"top_k": 500000
}
Expected Response
Status: 200
{
"model": "gemma3:4b",
"created_at": "2024-12-20T00:00:00.000000Z",
"response": "The capital of France is Paris.",
"done": true,
"context": [],
"total_duration": 0,
"load_duration": 0,
"prompt_eval_count": 10,
"prompt_eval_duration": 0,
"eval_count": 10,
"eval_duration": 0
}
Actual Response
Status: 501
Test #32 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/api/generate
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"prompt": "Lorem ipsum dolor sit amet",
"stream": true,
"temperature": 0.0,
"top_p": 0.5,
"top_k": 1000000
}
Expected Response
Status: 200
{
"model": "gemma3:4b",
"created_at": "2024-12-20T00:00:00.000000Z",
"response": "The capital of France is Paris.",
"done": true,
"context": [],
"total_duration": 0,
"load_duration": 0,
"prompt_eval_count": 10,
"prompt_eval_duration": 0,
"eval_count": 10,
"eval_duration": 0
}
Actual Response
Status: 501
Test #33 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/api/generate
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"prompt": "Lorem ipsum dolor sit amet",
"stream": true,
"temperature": 0.0,
"top_p": 1.0,
"top_k": 1
}
Expected Response
Status: 200
{
"model": "gemma3:4b",
"created_at": "2024-12-20T00:00:00.000000Z",
"response": "The capital of France is Paris.",
"done": true,
"context": [],
"total_duration": 0,
"load_duration": 0,
"prompt_eval_count": 10,
"prompt_eval_duration": 0,
"eval_count": 10,
"eval_duration": 0
}
Actual Response
Status: 501
Test #34 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/api/generate
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"prompt": "Lorem ipsum dolor sit amet",
"stream": true,
"temperature": 0.0,
"top_p": 1.0,
"top_k": 500000
}
Expected Response
Status: 200
{
"model": "gemma3:4b",
"created_at": "2024-12-20T00:00:00.000000Z",
"response": "The capital of France is Paris.",
"done": true,
"context": [],
"total_duration": 0,
"load_duration": 0,
"prompt_eval_count": 10,
"prompt_eval_duration": 0,
"eval_count": 10,
"eval_duration": 0
}
Actual Response
Status: 501
Test #35 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/api/generate
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"prompt": "Lorem ipsum dolor sit amet",
"stream": true,
"temperature": 0.0,
"top_p": 1.0,
"top_k": 1000000
}
Expected Response
Status: 200
{
"model": "gemma3:4b",
"created_at": "2024-12-20T00:00:00.000000Z",
"response": "The capital of France is Paris.",
"done": true,
"context": [],
"total_duration": 0,
"load_duration": 0,
"prompt_eval_count": 10,
"prompt_eval_duration": 0,
"eval_count": 10,
"eval_duration": 0
}
Actual Response
Status: 501
Test #36 ✅
🔧 Test case generated from schema
Endpoint: POST /v1/api/generate
Request Body
{
"model": "Lorem ipsum dolor sit amet",
"prompt": "Lorem ipsum dolor sit amet",
"stream": true,
"temperature": 0.123456789,
"top_p": 0.0,
"top_k": 1
}
Expected Response
Status: 200
{
"model": "gemma3:4b",
"created_at": "2024-12-20T00:00:00.000000Z",
"response": "The capital of France is Paris.",
"done": true,
"context": [],
"total_duration": 0,
"load_duration": 0,
"prompt_eval_count": 10,
"prompt_eval_duration": 0,
"eval_count": 10,
"eval_duration": 0
}
Actual Response
Status: 501