Skip to content

OpenAPI Contract Test Report

Summary

  • Total Tests: 36
  • Passed: ✅ 36
  • Failed: ❌ 0

Test #1 ✅

📋 Test case from OpenAPI example

Endpoint: GET /health

Expected Response

Status: 200

{
  "status": "ok",
  "providers_loading": false
}

Actual Response

Status: 200

{
  "status": "ok",
  "providers_loading": false
}

Test #2 ✅

📋 Test case from OpenAPI example

Endpoint: GET /private/v1/providers/{provider}/max-context-window

Expected Response

Status: 200

{
  "provider": "pixtral",
  "max_context_window": 128000
}

Actual Response

Status: 404

{
  "detail": "Provider '{provider}' not found"
}

Test #3 ✅

📋 Test case from OpenAPI example

Endpoint: GET /private/v1/providers

Expected Response

Status: 200

{
  "available": [
    "pixtral",
    "gemma3_on_vpn"
  ],
  "default": "pixtral",
  "total": 2,
  "status": "ready"
}

Actual Response

Status: 200

{
  "available": [
    "local_gemma3_270m"
  ],
  "default": "local_gemma3_270m",
  "total": 12,
  "status": "one_provider_available"
}

Test #4 ✅

📋 Test case from OpenAPI example

Endpoint: POST /v1/agent/chat

Request Body

{
  "message": "What's the weather?",
  "conversation_id": "conv-123",
  "user_id": "user-456",
  "stream": false
}

Expected Response

Status: 200

{
  "conversation_id": "conv-123",
  "user_id": "user-456",
  "message": "The weather is sunny today!",
  "role": "assistant",
  "created": 1703347200,
  "usage": {
    "prompt_tokens": 56,
    "completion_tokens": 31,
    "total_tokens": 87
  }
}

Actual Response

Status: 200

{
  "conversation_id": "conv-123",
  "user_id": "user-456",
  "message": "The weather is sunny with a high of 78 degrees Fahrenheit.\n",
  "role": "assistant",
  "created": 1769205864,
  "usage": {
    "prompt_tokens": 46,
    "completion_tokens": 16,
    "total_tokens": 62
  }
}

Test #5 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/agent/chat

Request Body

{
  "message": "Lorem ipsum dolor sit amet",
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Lorem ipsum dolor sit amet",
  "stream": true
}

Expected Response

Status: 200

{
  "conversation_id": "conv-123",
  "user_id": "user-456",
  "message": "The weather is sunny today!",
  "role": "assistant",
  "created": 1703347200,
  "usage": {
    "prompt_tokens": 56,
    "completion_tokens": 31,
    "total_tokens": 87
  }
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #6 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/agent/chat

Request Body

{
  "message": "Lorem ipsum dolor sit amet",
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Lorem ipsum dolor sit amet",
  "stream": false
}

Expected Response

Status: 200

{
  "conversation_id": "conv-123",
  "user_id": "user-456",
  "message": "The weather is sunny today!",
  "role": "assistant",
  "created": 1703347200,
  "usage": {
    "prompt_tokens": 56,
    "completion_tokens": 31,
    "total_tokens": 87
  }
}

Actual Response

Status: 200

{
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Lorem ipsum dolor sit amet",
  "message": "Lorem ipsum dolor sit amet.\n",
  "role": "assistant",
  "created": 1769205865,
  "usage": {
    "prompt_tokens": 45,
    "completion_tokens": 8,
    "total_tokens": 53
  }
}

Test #7 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/agent/chat

Request Body

{
  "message": "Lorem ipsum dolor sit amet",
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Test with 'single' quotes",
  "stream": true
}

Expected Response

Status: 200

{
  "conversation_id": "conv-123",
  "user_id": "user-456",
  "message": "The weather is sunny today!",
  "role": "assistant",
  "created": 1703347200,
  "usage": {
    "prompt_tokens": 56,
    "completion_tokens": 31,
    "total_tokens": 87
  }
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #8 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/agent/chat

Request Body

{
  "message": "Lorem ipsum dolor sit amet",
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Test with 'single' quotes",
  "stream": false
}

Expected Response

Status: 200

{
  "conversation_id": "conv-123",
  "user_id": "user-456",
  "message": "The weather is sunny today!",
  "role": "assistant",
  "created": 1703347200,
  "usage": {
    "prompt_tokens": 56,
    "completion_tokens": 31,
    "total_tokens": 87
  }
}

Actual Response

Status: 200

{
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Test with 'single' quotes",
  "message": "",
  "role": "assistant",
  "created": 1769205865,
  "usage": {
    "prompt_tokens": 45,
    "completion_tokens": 1,
    "total_tokens": 46
  }
}

Test #9 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/agent/chat

Request Body

{
  "message": "Lorem ipsum dolor sit amet",
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Test with \"double\" quotes",
  "stream": true
}

Expected Response

Status: 200

{
  "conversation_id": "conv-123",
  "user_id": "user-456",
  "message": "The weather is sunny today!",
  "role": "assistant",
  "created": 1703347200,
  "usage": {
    "prompt_tokens": 56,
    "completion_tokens": 31,
    "total_tokens": 87
  }
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #10 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/agent/chat

Request Body

{
  "message": "Lorem ipsum dolor sit amet",
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Test with \"double\" quotes",
  "stream": false
}

Expected Response

Status: 200

{
  "conversation_id": "conv-123",
  "user_id": "user-456",
  "message": "The weather is sunny today!",
  "role": "assistant",
  "created": 1703347200,
  "usage": {
    "prompt_tokens": 56,
    "completion_tokens": 31,
    "total_tokens": 87
  }
}

Actual Response

Status: 200

{
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Test with \"double\" quotes",
  "message": "Lorem ipsum dolor sit amet\n",
  "role": "assistant",
  "created": 1769205866,
  "usage": {
    "prompt_tokens": 45,
    "completion_tokens": 7,
    "total_tokens": 52
  }
}

Test #11 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/agent/chat

Request Body

{
  "message": "Lorem ipsum dolor sit amet",
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Test:with:colons",
  "stream": true
}

Expected Response

Status: 200

{
  "conversation_id": "conv-123",
  "user_id": "user-456",
  "message": "The weather is sunny today!",
  "role": "assistant",
  "created": 1703347200,
  "usage": {
    "prompt_tokens": 56,
    "completion_tokens": 31,
    "total_tokens": 87
  }
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #12 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/agent/chat

Request Body

{
  "message": "Lorem ipsum dolor sit amet",
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Test:with:colons",
  "stream": false
}

Expected Response

Status: 200

{
  "conversation_id": "conv-123",
  "user_id": "user-456",
  "message": "The weather is sunny today!",
  "role": "assistant",
  "created": 1703347200,
  "usage": {
    "prompt_tokens": 56,
    "completion_tokens": 31,
    "total_tokens": 87
  }
}

Actual Response

Status: 200

{
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Test:with:colons",
  "message": "Lorem ipsum dolor sit amet.\n",
  "role": "assistant",
  "created": 1769205866,
  "usage": {
    "prompt_tokens": 45,
    "completion_tokens": 8,
    "total_tokens": 53
  }
}

Test #13 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/agent/chat

Request Body

{
  "message": "Lorem ipsum dolor sit amet",
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Test\\with\\backslashes",
  "stream": true
}

Expected Response

Status: 200

{
  "conversation_id": "conv-123",
  "user_id": "user-456",
  "message": "The weather is sunny today!",
  "role": "assistant",
  "created": 1703347200,
  "usage": {
    "prompt_tokens": 56,
    "completion_tokens": 31,
    "total_tokens": 87
  }
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #14 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/agent/chat

Request Body

{
  "message": "Lorem ipsum dolor sit amet",
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Test\\with\\backslashes",
  "stream": false
}

Expected Response

Status: 200

{
  "conversation_id": "conv-123",
  "user_id": "user-456",
  "message": "The weather is sunny today!",
  "role": "assistant",
  "created": 1703347200,
  "usage": {
    "prompt_tokens": 56,
    "completion_tokens": 31,
    "total_tokens": 87
  }
}

Actual Response

Status: 200

{
  "conversation_id": "Lorem ipsum dolor sit amet",
  "user_id": "Test\\with\\backslashes",
  "message": "Lorem ipsum dolor sit amet.",
  "role": "assistant",
  "created": 1769205867,
  "usage": {
    "prompt_tokens": 45,
    "completion_tokens": 7,
    "total_tokens": 52
  }
}

Test #15 ✅

📋 Test case from OpenAPI example

Endpoint: POST /v1/chat/completions

Request Body

{
  "model": "pixtral",
  "messages": [
    {
      "role": "user",
      "content": "What is the capital of France?"
    }
  ],
  "temperature": 0.7,
  "max_tokens": 100
}

Expected Response

Status: 200

{
  "id": "chatcmpl-abc123",
  "object": "chat.completion",
  "created": 1734700000,
  "model": "pixtral",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "Hello! How can I help you today?"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 20,
    "completion_tokens": 10,
    "total_tokens": 30
  }
}

Actual Response

Status: 200

{
  "id": "chatcmpl-7b6b1d8ba96446759f767f64",
  "object": "chat.completion",
  "created": 1769205867,
  "model": "pixtral",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "The capital of France is Paris.\n"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 22,
    "completion_tokens": 9,
    "total_tokens": 31
  }
}

Test #16 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/chat/completions

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "messages": [
    {
      "role": "system",
      "content": "Lorem ipsum dolor sit amet"
    }
  ],
  "temperature": 0.0,
  "max_tokens": 1,
  "stream": true,
  "top_p": 0.0,
  "stop": []
}

Expected Response

Status: 200

{
  "id": "chatcmpl-abc123",
  "object": "chat.completion",
  "created": 1734700000,
  "model": "pixtral",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "Hello! How can I help you today?"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 20,
    "completion_tokens": 10,
    "total_tokens": 30
  }
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #17 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/chat/completions

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "messages": [
    {
      "role": "system",
      "content": "Lorem ipsum dolor sit amet"
    }
  ],
  "temperature": 0.0,
  "max_tokens": 1,
  "stream": true,
  "top_p": 0.0,
  "stop": [
    "Lorem ipsum dolor sit amet"
  ]
}

Expected Response

Status: 200

{
  "id": "chatcmpl-abc123",
  "object": "chat.completion",
  "created": 1734700000,
  "model": "pixtral",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "Hello! How can I help you today?"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 20,
    "completion_tokens": 10,
    "total_tokens": 30
  }
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #18 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/chat/completions

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "messages": [
    {
      "role": "system",
      "content": "Lorem ipsum dolor sit amet"
    }
  ],
  "temperature": 0.0,
  "max_tokens": 1,
  "stream": true,
  "top_p": 0.0,
  "stop": [
    "Lorem ipsum dolor sit amet",
    "Test with 'single' quotes",
    "Test with \"double\" quotes"
  ]
}

Expected Response

Status: 200

{
  "id": "chatcmpl-abc123",
  "object": "chat.completion",
  "created": 1734700000,
  "model": "pixtral",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "Hello! How can I help you today?"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 20,
    "completion_tokens": 10,
    "total_tokens": 30
  }
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #19 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/chat/completions

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "messages": [
    {
      "role": "system",
      "content": "Lorem ipsum dolor sit amet"
    }
  ],
  "temperature": 0.0,
  "max_tokens": 1,
  "stream": true,
  "top_p": 0.5,
  "stop": []
}

Expected Response

Status: 200

{
  "id": "chatcmpl-abc123",
  "object": "chat.completion",
  "created": 1734700000,
  "model": "pixtral",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "Hello! How can I help you today?"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 20,
    "completion_tokens": 10,
    "total_tokens": 30
  }
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #20 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/chat/completions

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "messages": [
    {
      "role": "system",
      "content": "Lorem ipsum dolor sit amet"
    }
  ],
  "temperature": 0.0,
  "max_tokens": 1,
  "stream": true,
  "top_p": 0.5,
  "stop": [
    "Lorem ipsum dolor sit amet"
  ]
}

Expected Response

Status: 200

{
  "id": "chatcmpl-abc123",
  "object": "chat.completion",
  "created": 1734700000,
  "model": "pixtral",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "Hello! How can I help you today?"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 20,
    "completion_tokens": 10,
    "total_tokens": 30
  }
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #21 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/chat/completions

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "messages": [
    {
      "role": "system",
      "content": "Lorem ipsum dolor sit amet"
    }
  ],
  "temperature": 0.0,
  "max_tokens": 1,
  "stream": true,
  "top_p": 0.5,
  "stop": [
    "Lorem ipsum dolor sit amet",
    "Test with 'single' quotes",
    "Test with \"double\" quotes"
  ]
}

Expected Response

Status: 200

{
  "id": "chatcmpl-abc123",
  "object": "chat.completion",
  "created": 1734700000,
  "model": "pixtral",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "Hello! How can I help you today?"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 20,
    "completion_tokens": 10,
    "total_tokens": 30
  }
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #22 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/chat/completions

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "messages": [
    {
      "role": "system",
      "content": "Lorem ipsum dolor sit amet"
    }
  ],
  "temperature": 0.0,
  "max_tokens": 1,
  "stream": true,
  "top_p": 1.0,
  "stop": []
}

Expected Response

Status: 200

{
  "id": "chatcmpl-abc123",
  "object": "chat.completion",
  "created": 1734700000,
  "model": "pixtral",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "Hello! How can I help you today?"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 20,
    "completion_tokens": 10,
    "total_tokens": 30
  }
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #23 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/chat/completions

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "messages": [
    {
      "role": "system",
      "content": "Lorem ipsum dolor sit amet"
    }
  ],
  "temperature": 0.0,
  "max_tokens": 1,
  "stream": true,
  "top_p": 1.0,
  "stop": [
    "Lorem ipsum dolor sit amet"
  ]
}

Expected Response

Status: 200

{
  "id": "chatcmpl-abc123",
  "object": "chat.completion",
  "created": 1734700000,
  "model": "pixtral",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "Hello! How can I help you today?"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 20,
    "completion_tokens": 10,
    "total_tokens": 30
  }
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #24 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/chat/completions

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "messages": [
    {
      "role": "system",
      "content": "Lorem ipsum dolor sit amet"
    }
  ],
  "temperature": 0.0,
  "max_tokens": 1,
  "stream": true,
  "top_p": 1.0,
  "stop": [
    "Lorem ipsum dolor sit amet",
    "Test with 'single' quotes",
    "Test with \"double\" quotes"
  ]
}

Expected Response

Status: 200

{
  "id": "chatcmpl-abc123",
  "object": "chat.completion",
  "created": 1734700000,
  "model": "pixtral",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "Hello! How can I help you today?"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 20,
    "completion_tokens": 10,
    "total_tokens": 30
  }
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #25 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/chat/completions

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "messages": [
    {
      "role": "system",
      "content": "Lorem ipsum dolor sit amet"
    }
  ],
  "temperature": 0.0,
  "max_tokens": 1,
  "stream": false,
  "top_p": 0.0,
  "stop": []
}

Expected Response

Status: 200

{
  "id": "chatcmpl-abc123",
  "object": "chat.completion",
  "created": 1734700000,
  "model": "pixtral",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "Hello! How can I help you today?"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 20,
    "completion_tokens": 10,
    "total_tokens": 30
  }
}

Actual Response

Status: 200

{
  "id": "chatcmpl-1dd6396394bc40619ced8970",
  "object": "chat.completion",
  "created": 1769205867,
  "model": "Lorem ipsum dolor sit amet",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": ""
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 20,
    "completion_tokens": 1,
    "total_tokens": 21
  }
}

Test #26 ✅

📋 Test case from OpenAPI example

Endpoint: POST /v1/api/generate

Request Body

{
  "model": "gemma3:4b",
  "prompt": "What is the capital of France?",
  "stream": false,
  "temperature": 0.7
}

Expected Response

Status: 200

{
  "model": "gemma3:4b",
  "created_at": "2024-12-20T00:00:00.000000Z",
  "response": "The capital of France is Paris.",
  "done": true,
  "context": [],
  "total_duration": 0,
  "load_duration": 0,
  "prompt_eval_count": 10,
  "prompt_eval_duration": 0,
  "eval_count": 10,
  "eval_duration": 0
}

Actual Response

Status: 200

{
  "model": "gemma3:4b",
  "created_at": "2026-01-23T22:04:28.350676Z",
  "response": "The capital of France is Paris.",
  "done": true,
  "context": [],
  "total_duration": 0,
  "load_duration": 0,
  "prompt_eval_count": 22,
  "prompt_eval_duration": 0,
  "eval_count": 8,
  "eval_duration": 0
}

Test #27 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/api/generate

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "prompt": "Lorem ipsum dolor sit amet",
  "stream": true,
  "temperature": 0.0,
  "top_p": 0.0,
  "top_k": 1
}

Expected Response

Status: 200

{
  "model": "gemma3:4b",
  "created_at": "2024-12-20T00:00:00.000000Z",
  "response": "The capital of France is Paris.",
  "done": true,
  "context": [],
  "total_duration": 0,
  "load_duration": 0,
  "prompt_eval_count": 10,
  "prompt_eval_duration": 0,
  "eval_count": 10,
  "eval_duration": 0
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #28 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/api/generate

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "prompt": "Lorem ipsum dolor sit amet",
  "stream": true,
  "temperature": 0.0,
  "top_p": 0.0,
  "top_k": 500000
}

Expected Response

Status: 200

{
  "model": "gemma3:4b",
  "created_at": "2024-12-20T00:00:00.000000Z",
  "response": "The capital of France is Paris.",
  "done": true,
  "context": [],
  "total_duration": 0,
  "load_duration": 0,
  "prompt_eval_count": 10,
  "prompt_eval_duration": 0,
  "eval_count": 10,
  "eval_duration": 0
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #29 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/api/generate

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "prompt": "Lorem ipsum dolor sit amet",
  "stream": true,
  "temperature": 0.0,
  "top_p": 0.0,
  "top_k": 1000000
}

Expected Response

Status: 200

{
  "model": "gemma3:4b",
  "created_at": "2024-12-20T00:00:00.000000Z",
  "response": "The capital of France is Paris.",
  "done": true,
  "context": [],
  "total_duration": 0,
  "load_duration": 0,
  "prompt_eval_count": 10,
  "prompt_eval_duration": 0,
  "eval_count": 10,
  "eval_duration": 0
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #30 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/api/generate

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "prompt": "Lorem ipsum dolor sit amet",
  "stream": true,
  "temperature": 0.0,
  "top_p": 0.5,
  "top_k": 1
}

Expected Response

Status: 200

{
  "model": "gemma3:4b",
  "created_at": "2024-12-20T00:00:00.000000Z",
  "response": "The capital of France is Paris.",
  "done": true,
  "context": [],
  "total_duration": 0,
  "load_duration": 0,
  "prompt_eval_count": 10,
  "prompt_eval_duration": 0,
  "eval_count": 10,
  "eval_duration": 0
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #31 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/api/generate

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "prompt": "Lorem ipsum dolor sit amet",
  "stream": true,
  "temperature": 0.0,
  "top_p": 0.5,
  "top_k": 500000
}

Expected Response

Status: 200

{
  "model": "gemma3:4b",
  "created_at": "2024-12-20T00:00:00.000000Z",
  "response": "The capital of France is Paris.",
  "done": true,
  "context": [],
  "total_duration": 0,
  "load_duration": 0,
  "prompt_eval_count": 10,
  "prompt_eval_duration": 0,
  "eval_count": 10,
  "eval_duration": 0
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #32 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/api/generate

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "prompt": "Lorem ipsum dolor sit amet",
  "stream": true,
  "temperature": 0.0,
  "top_p": 0.5,
  "top_k": 1000000
}

Expected Response

Status: 200

{
  "model": "gemma3:4b",
  "created_at": "2024-12-20T00:00:00.000000Z",
  "response": "The capital of France is Paris.",
  "done": true,
  "context": [],
  "total_duration": 0,
  "load_duration": 0,
  "prompt_eval_count": 10,
  "prompt_eval_duration": 0,
  "eval_count": 10,
  "eval_duration": 0
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #33 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/api/generate

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "prompt": "Lorem ipsum dolor sit amet",
  "stream": true,
  "temperature": 0.0,
  "top_p": 1.0,
  "top_k": 1
}

Expected Response

Status: 200

{
  "model": "gemma3:4b",
  "created_at": "2024-12-20T00:00:00.000000Z",
  "response": "The capital of France is Paris.",
  "done": true,
  "context": [],
  "total_duration": 0,
  "load_duration": 0,
  "prompt_eval_count": 10,
  "prompt_eval_duration": 0,
  "eval_count": 10,
  "eval_duration": 0
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #34 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/api/generate

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "prompt": "Lorem ipsum dolor sit amet",
  "stream": true,
  "temperature": 0.0,
  "top_p": 1.0,
  "top_k": 500000
}

Expected Response

Status: 200

{
  "model": "gemma3:4b",
  "created_at": "2024-12-20T00:00:00.000000Z",
  "response": "The capital of France is Paris.",
  "done": true,
  "context": [],
  "total_duration": 0,
  "load_duration": 0,
  "prompt_eval_count": 10,
  "prompt_eval_duration": 0,
  "eval_count": 10,
  "eval_duration": 0
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #35 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/api/generate

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "prompt": "Lorem ipsum dolor sit amet",
  "stream": true,
  "temperature": 0.0,
  "top_p": 1.0,
  "top_k": 1000000
}

Expected Response

Status: 200

{
  "model": "gemma3:4b",
  "created_at": "2024-12-20T00:00:00.000000Z",
  "response": "The capital of France is Paris.",
  "done": true,
  "context": [],
  "total_duration": 0,
  "load_duration": 0,
  "prompt_eval_count": 10,
  "prompt_eval_duration": 0,
  "eval_count": 10,
  "eval_duration": 0
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}

Test #36 ✅

🔧 Test case generated from schema

Endpoint: POST /v1/api/generate

Request Body

{
  "model": "Lorem ipsum dolor sit amet",
  "prompt": "Lorem ipsum dolor sit amet",
  "stream": true,
  "temperature": 0.123456789,
  "top_p": 0.0,
  "top_k": 1
}

Expected Response

Status: 200

{
  "model": "gemma3:4b",
  "created_at": "2024-12-20T00:00:00.000000Z",
  "response": "The capital of France is Paris.",
  "done": true,
  "context": [],
  "total_duration": 0,
  "load_duration": 0,
  "prompt_eval_count": 10,
  "prompt_eval_duration": 0,
  "eval_count": 10,
  "eval_duration": 0
}

Actual Response

Status: 501

{
  "detail": "Streaming not yet implemented"
}