Frequently Asked Questions

This page answers some of the often asked questions about Schemathesis.

Usage & Configuration

What kind of data does Schemathesis generate?

Schemathesis generates random test data that conforms to the given API schema as well as not-conforming data, depending on the supplied data generation method config value. This data consists of all possible data types from the JSON schema specification in various combinations and different nesting levels.

We can’t guarantee that the generated data will always be accepted by the application under test since there could be validation rules not covered by the API schema. If you found that Schemathesis generated something that doesn’t fit the API schema, consider reporting a bug

How many tests does Schemathesis execute for an API operation?

The total number of tests Schemathesis executes is influenced by the API schema’s complexity, user-defined settings like --hypothesis-max-examples for the maximum tests generated, and the test generation phases (explicit, generate, reuse, and shrink). The process is designed to optimize coverage within a reasonable test budget rather than aiming for exhaustive coverage. For detailed insights and customization options, refer to our data generation docs.

What kind errors Schemathesis is capable to find?

The main two groups of problems that Schemathesis targets are server-side errors and nonconformity to the behavior described in the API schema.

What parts of the application is Schemathesis targeting during its tests?

It depends. The test data that Schemathesis generates is random. Input validation is, therefore, more frequently examined than other parts.

Since Schemathesis generates data that fits the application’s API schema, it can reach the app’s business logic, but it depends on the architecture of each particular application.

What if my application doesn’t have an API schema?

As the first step, you can use schema generators like flasgger for Python, GrapeSwagger for Ruby, or Swashbuckle for ASP.Net. Then, running Schemathesis against the generated API schema will help you to refine its definitions.

How is Schemathesis different from Dredd?

Schemathesis focuses on finding inputs that result in application crash, but it shares the goal of keeping the API documentation up to date with Dredd. Both tools can generate requests to the API under test, but they approach it differently.

Schemathesis uses Property-Based Testing to infer all input values and uses examples defined in the API schema as separate test cases. Dredd uses examples described in the API schema as the primary source of inputs (and requires them to work) and generates data only in some situations.

By using Hypothesis as the underlying testing framework, Schemathesis benefits from all its features like test case reduction and stateful testing. Dredd works more in a way that requires you to write some sort of example-based tests when Schemathesis requires only a valid API schema and will generate tests for you.

There are a lot of features that Dredd has are Schemathesis has not (e.g., API Blueprint support, that powerful hook system, and many more) and probably vice versa. Definitely, Schemathesis can learn a lot from Dredd and if you miss any feature that exists in Dredd but doesn’t exist in Schemathesis, let us know.

Why are no examples generated in Schemathesis when using --hypothesis-phase=explicit?

The --hypothesis-phase=explicit option is designed to test only the examples that are explicitly defined in the API schema. It avoids generating new examples to maintain predictability and adhere strictly to the documented API behavior.

If you need random examples for API operations without explicit examples, consider using the --contrib-openapi-fill-missing-examples CLI option.

How should I run Schemathesis?

There are two main ways to run it — as a part of Python’s test suite, and as a command-line tool.

If you wrote a Python application and you want to utilize the features of an existing test suite, then the in-code option will best suit your needs.

If you wrote your application in a language other than Python, you should use the built-in CLI. Please keep in mind that you will need to have a running application where you can run Schemathesis against.

Should I always have my application running before starting the test suite?

Only in some workflows! In CLI, you can test your AioHTTP / ASGI / WSGI apps with the --app CLI option. For the pytest integration, there is schemathesis.from_pytest_fixture loader where you can postpone API schema loading and start the test application as a part of your test setup. See more information in the Writing Python tests section.

How long does it usually take for Schemathesis to test an app?

It depends on many factors, including the API’s complexity under test, the network connection speed, and the Schemathesis configuration. Usually, it takes from a few seconds to a few minutes to run all the tests. However, there are exceptions where it might take an hour and more.

Can I exclude particular data from being generated?

Yes. Schemathesis’s hooks mechanism allows you to adapt its behavior and generate data that better fits your use case.

Also, if your application fails on some input early in the code, then it’s often a good idea to exclude this input from the next test run so you can explore deeper parts of your codebase.

How can I use database objects IDs in tests?

The case object that is injected in each test can be modified, assuming your URL template is /api/users/{user_id} then in tests, it can be done like this:

schema = ...  # Load the API schema here


@schema.parametrize()
def test_api(case):
    case.path_parameters["user_id"] = 42

Why does Schemathesis fail to parse my API schema generate by FastAPI?

FastAPI uses pydantic, which in turn uses JSON Schema Draft 7. This can lead to compatibility issues as OpenAPI 2.0 and 3.0.x use earlier versions of JSON Schema.

For detailed solutions, please refer to the Compatibility section.

Why Schemathesis generates uniform data for my API schema?

There might be multiple reasons for that, but usually, this behavior occurs when the API schema is complex or deeply nested. Please, refer to the Data generation section in the documentation for more info. If you think that it is not the case, feel free to open an issue.

How different is --request-timeout from --hypothesis-deadline

These CLI parameters both represent some kind of limit for the duration of a certain part of a single test. However, each of them has a different scope.

--hypothesis-deadline counts parts of a single test case execution, including waiting for the API response, and running all checks and relevant hooks for that single test case.

--request-timeout is only relevant for waiting for the API response. If this duration is exceeded, the test is marked as a “Timeout”.

Why Schemathesis reports “Flaky” errors?

When Schemathesis finds a failure, it tries to verify it by re-running the test again. If the same failure is not reproduced, then Schemathesis concludes the test as “Flaky”.

This situation usually happens, when the tested application state is not reset between tests. Let’s imagine that we have an API where the user can create “orders”, then the “Flaky” situation might look like this:

  1. Create order “A” -> 201 with payload that does not conform to the definition in the API schema;

  2. Create order “A” again to verify the failure -> 409 with conformant payload.

With Python tests, you may want to write a context manager that cleans the application state between test runs as suggested by Hypothesis docs.

CLI reports flaky failures as regular failures with a special note about their flakiness. Cleaning the application state could be done via the before_call hook.

Does Schemathesis support Open API discriminators? Schemathesis raises an “Unsatisfiable” error.

The discriminator field does not affect data generation, and Schemathesis work directly with the underlying schemas. Usually, the problem comes from using the oneOf keyword with very permissive sub-schemas. For example:

discriminator:
  propertyName: objectType
oneOf:
  - type: object
    required:
      - objectType
    properties:
      objectType:
        type: string
      foo:
        type: string
  - type: object
    required:
      - objectType
    properties:
      objectType:
        type: string
      bar:
        type: string

Here both schemas do not restrict their additional properties, and for this reason, any object that is valid for the first sub-schema is also valid for the second one, which contradicts the definition of the oneOf keyword behavior, where the value should be valid against exactly one sub-schema.

To solve this problem, you can use anyOf or make your sub-schemas less permissive.

Schemathesis reports conformance issue for schemas with the oneOf keyword. Why?

The oneOf keyword is a tricky one and the validation results might look counterintuitive at first glance. Let’s take a look at an example:

paths:
  /pets:
    patch:
      requestBody:
        content:
          application/json:
            schema:
              oneOf:
                - $ref: '#/components/schemas/Cat'
                - $ref: '#/components/schemas/Dog'
      responses:
        '200':
          description: Updated
components:
  schemas:
    Dog:
      type: object
      properties:
        bark:
          type: boolean
        breed:
          type: string
          enum: [Dingo, Husky, Retriever, Shepherd]
    Cat:
      type: object
      properties:
        hunts:
          type: boolean
        age:
          type: integer

Here we have two possible payload options - Dog and Cat. The following JSON object is valid against the Dog schema:

{
  "bark": true,
  "breed": "Dingo"
}

Though, oneOf requires that the input should be valid against exactly one sub-schema! At first glance it looks like the case, but it is actually not. It happens because the Cat schema does not restrict what properties should always be present and what should not. If the input object does not have the hunts or age properties, then it will be validated as a Cat instance. To prevent this situation you might use required and additionalProperties keywords:

components:
  schemas:
    Dog:
      type: object
      properties:
        bark:
          type: boolean
        breed:
          type: string
          enum: [Dingo, Husky, Retriever, Shepherd]
      required: [bark, breed]      # List all the required properties
      additionalProperties: false  # And forbid any others
    Cat:
      type: object
      properties:
        hunts:
          type: boolean
        age:
          type: integer
      required: [hunts, age]       # List all the required properties
      additionalProperties: false  # And forbid any others

By adding these keywords, any Cat instance will always require the hunts and age properties to be present.

As an alternative, you could use the anyOf keyword instead.

Why Schemathesis does not generate UUIDs for Open API 2.0 / 3.0 even if format: uuid is specified?

Open API 2.0 / 3.0 do not declare the uuid format as built-in. You have two options to enable UUID generation:

  1. Use an extension:

from schemathesis.contrib.openapi import formats

formats.uuid.install()
  1. Enable experimental support for OpenAPI 3.1, which also activates UUID generation. See the Experimental Features section for details.

Why is Schemathesis slower on Windows when using localhost?

When Schemathesis sends a request to http://localhost/, it first attempts to use IPv6. This can cause delays if your server only supports IPv4. This is especially problematic on Windows due to an unavoidable 1-second timeout for refused TCP connections, which the OS may retry up to three times. On Linux, the connection fails immediately if refused, allowing a quick switch to IPv4.

Solution: To avoid this delay, simply use http://127.0.0.1/ instead of http://localhost/. This ensures that Schemathesis will use IPv4 directly.

Why can’t Schemathesis connect to my locally running application when run via Docker on MacOS?

The host has a changing IP address, or none if you have no network access. As a result, the Docker container cannot use localhost to reach the host machine.

Solution: Instead, use host.docker.internal as the hostname to allow Schemathesis to connect to services running on the host.

How to prevent Schemathesis from generating NULL bytes in strings?

By default, Schemathesis generates NULL bytes for all strings in order to cover more possible edge cases.

Solution: To prevent Schemathesis from generating NULL bytes in strings, you need to set the allow_x00 configuration to False.

CLI:

$ st run --generation-allow-x00=false ...

Python:

import schemathesis
from schemathesis import GenerationConfig

schema = schemathesis.from_uri(
    "https://example.schemathesis.io/openapi.json",
    generation_config=GenerationConfig(allow_x00=False),
)

This adjustment ensures that Schemathesis does not include NULL bytes in strings for all your tests, making them compatible with systems that reject such inputs.

Working with API schemas

How to disallow random field names in my schema?

You need to add additionalProperties: false to the relevant object definition. But there is a caveat with emulating inheritance with Open API via allOf.

In this case, it is better to use YAML anchors to share schema parts; otherwise it will prevent valid data from passing the validation.