graphql-engine/server/tests-py
Samir Talwar ffa3f5e3cc server/tests-py: Remove test_horizontal_scale.py.
This removes *test_horizontal_scale.py*, which tests horizontal scaling by running some tests against two HGE servers, both talking to PgBouncer.

This doesn't really test HGE; it's testing whether PgBouncer guarantees the same level of consistency as PostgreSQL. I'm not convinced that it's worthwhile, so I am deleting it.

[NDAT-542]: https://hasurahq.atlassian.net/browse/NDAT-542?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ

PR-URL: https://github.com/hasura/graphql-engine-mono/pull/8821
GitOrigin-RevId: 2029f2fea11c7fd58ed47e724cb3cd50e67d5a43
2023-04-20 09:28:15 +00:00
..
docker server/api-tests: Add a Docker Compose file for the API tests. 2023-01-16 16:44:10 +00:00
fixtures server/tests-py: Parallelize JWT tests. 2023-04-19 10:30:21 +00:00
pgdump tests: add Postgres 14 and Postgres 15 pytests 2022-10-06 09:46:38 +00:00
queries server/tests-py: Parallelize test_logging.py. 2023-04-17 13:09:34 +00:00
remote_schemas/nodejs CI: Add commentary on how the Python integration test scripts work. 2023-03-07 20:59:31 +00:00
test_tests Test result ordering, add --accept test mode to automatically accept changed test cases 2019-11-05 15:15:25 -06:00
webhook/insecure [server] coalesce multiple run_sql calls in tests (#270) 2021-01-06 16:07:22 +00:00
.devsh_version Restructure Pro Parser modules to align with the OSS Parser 2022-07-26 07:33:35 +00:00
.envrc server/tests-py: Load the Python environment in a .envrc file. 2022-07-05 16:52:27 +00:00
.gitignore server/tests-py: Move installing node_modules to the makefile. 2022-08-16 09:27:04 +00:00
.prettierignore server/tests-py: Factor out service URLs as environment variables. 2022-08-15 14:59:05 +00:00
auth_webhook_server.py server/tests-py: Generate random ports for helper services. 2022-10-21 17:34:21 +00:00
conftest.py server/tests-py: Parallelize JWT tests. 2023-04-19 10:30:21 +00:00
context.py server/tests-py: Parallelize JWT tests. 2023-04-19 10:30:21 +00:00
docker-compose.yml server/tests-py: Run TestSubscriptionBasicNoAuth in the new mode. 2023-02-08 11:51:54 +00:00
graphql_server.py server: fix bug with customized remote schema enum variables 2022-11-24 08:47:46 +00:00
jwk_server.py server/tests-py: Generate random ports for a few more helper services. 2022-10-27 11:49:06 +00:00
package-lock.json Upgrade all package-lock.json files to the v2 format. 2022-09-28 08:15:29 +00:00
package.json CI: Add commentary on how the Python integration test scripts work. 2023-03-07 20:59:31 +00:00
ports.py server/tests-py: Start some node.js test services on random ports. 2022-09-07 16:26:10 +00:00
PortToHaskell.py chore: Fix all outstanding hlint hints 2023-02-20 17:43:28 +00:00
pytest.ini server/tests-py: Parallelize JWT tests. 2023-04-19 10:30:21 +00:00
README.md add run-new.sh docs 2022-12-14 16:11:00 +00:00
remote_server.py CI: Run the Python tests in the "new" mode, one HGE per test class. 2023-02-03 14:09:29 +00:00
requirements-top-level.txt CI: Upgrade server-pytest-runner's dependencies. 2023-03-07 17:41:00 +00:00
requirements.txt CI: Upgrade server-pytest-runner's dependencies. 2023-03-07 17:41:00 +00:00
run-new.sh server/tests-py: Get all tests passing with separate HGE binaries. 2022-12-21 15:56:41 +00:00
run.sh server/tests-py: Get all tests passing with separate HGE binaries. 2022-12-21 15:56:41 +00:00
super_classes.py server/tests-py: Reduce the number of locations we check the status code. 2022-07-05 18:01:07 +00:00
tempCodeRunnerFile.py server/MSSQL: Event Delivery System (Incremental PR - 3) 2022-04-21 07:20:34 +00:00
test_actions.py server: fix nullable field omitted from action response when omitted in the webhook response 2023-02-07 15:34:07 +00:00
test_allowlist_queries.py server/tests-py: Generate random ports for helper services. 2022-10-21 17:34:21 +00:00
test_apis_disabled.py server/tests-py: Spawn a new PostgreSQL schema ("database") per test. 2022-11-15 19:08:53 +00:00
test_apollo_federation.py server/tests-py: Provide the admin secret to the HGE server. 2022-09-29 17:20:07 +00:00
test_auth_webhook_cookie.py server/tests-py: Start webhook.py inside the test harness. 2022-10-20 19:00:01 +00:00
test_compat.py server/tests-py: Provide the admin secret to the HGE server. 2022-09-29 17:20:07 +00:00
test_compression.py server: don't compress small responses 2022-12-13 17:50:01 +00:00
test_config_api.py server/tests-py: Parallelize JWT tests. 2023-04-19 10:30:21 +00:00
test_cors.py server/tests-py: Declaratively state the HGE environment variables. 2022-09-28 09:21:02 +00:00
test_dev_endpoints.py server/tests-py: Provide the admin secret to the HGE server. 2022-09-29 17:20:07 +00:00
test_endpoints.py server/tests-py: Set up postgis extensions using a fixture. 2022-08-15 14:30:42 +00:00
test_events.py server/tests-py: Get all tests passing with separate HGE binaries. 2022-12-21 15:56:41 +00:00
test_graphql_introspection.py server/tests-py: Generate random ports for helper services. 2022-10-21 17:34:21 +00:00
test_graphql_mutations.py server/tests-py: Start webhook.py inside the test harness. 2022-10-20 19:00:01 +00:00
test_graphql_queries.py server/tests-py: Parallelize JWT tests. 2023-04-19 10:30:21 +00:00
test_graphql_read_only_source.py server/tests-py: Run test_graphql_read_only_source.py in parallel. 2023-04-12 14:37:21 +00:00
test_heterogeneous.py server/tests-py: Make HGECtxGQLServer a fixture. 2022-09-14 21:42:40 +00:00
test_inconsistent_meta.py server/tests-py: Fix test_inconsistent_meta.py for split databases. 2022-11-23 14:14:01 +00:00
test_jwk.py server/tests-py: Parallelize test_logging.py. 2023-04-17 13:09:34 +00:00
test_jwt_claims_map.py server/tests-py: Parallelize JWT tests. 2023-04-19 10:30:21 +00:00
test_jwt.py server/tests-py: Parallelize JWT tests. 2023-04-19 10:30:21 +00:00
test_logging.py server/tests-py: Parallelize test_logging.py. 2023-04-17 13:09:34 +00:00
test_metadata.py tests: add Postgres 14 and Postgres 15 pytests 2022-10-06 09:46:38 +00:00
test_naming_conventions.py server/tests-py: Get all tests passing with separate HGE binaries. 2022-12-21 15:56:41 +00:00
test_openapi.py server/tests-py: Reduce the number of locations we check the status code. 2022-07-05 18:01:07 +00:00
test_pg_dump.py server/tests-py: Upgrade to postgresql-client-15. 2022-10-20 05:57:30 +00:00
test_query_cache.py server: disable caching for actions with forward client headers enabled 2021-04-13 07:01:34 +00:00
test_remote_relationships.py server/tests-py: Generate random ports for helper services. 2022-10-21 17:34:21 +00:00
test_remote_schema_permissions.py server/tests-py: Generate random ports for helper services. 2022-10-21 17:34:21 +00:00
test_roles_inheritance.py server/tests-py: Generate random ports for helper services. 2022-10-21 17:34:21 +00:00
test_scheduled_triggers.py server/tests-py: Get all tests passing with separate HGE binaries. 2022-12-21 15:56:41 +00:00
test_schema_duplication.py server/tests-py: Declaratively state the HGE environment variables. 2022-09-28 09:21:02 +00:00
test_schema_stitching.py server/tests-py: Generate random ports for helper services. 2022-10-21 17:34:21 +00:00
test_subscriptions.py server/tests-py: Run TestSubscriptionBasicNoAuth in the new mode. 2023-02-08 11:51:54 +00:00
test_tests.py server/tests-py: Start webhook.py inside the test harness. 2022-10-20 19:00:01 +00:00
test_v1_queries.py server/tests-py: Declaratively state the HGE environment variables. 2022-09-28 09:21:02 +00:00
test_v1alpha1_endpoint.py server/tests-py: Start webhook.py inside the test harness. 2022-10-20 19:00:01 +00:00
test_v2_queries.py server/tests-py: Use markers for backends instead of parameterized tests. 2022-08-02 19:33:59 +00:00
test_validation.py server/tests-py: Set up postgis extensions using a fixture. 2022-08-15 14:30:42 +00:00
test_version.py server/tests-py: Move installing node_modules to the makefile. 2022-08-16 09:27:04 +00:00
test_webhook_insecure.py server/tests-py: Start webhook.py inside the test harness. 2022-10-20 19:00:01 +00:00
test_webhook_request_context.py server/tests-py: Generate random ports for a few more helper services. 2022-10-27 11:49:06 +00:00
test_webhook.py Remove redundant basic object query pytests 2023-03-08 17:07:38 +00:00
test_websocket_init_cookie.py server/tests-py: Run test_websocket_init_cookie.py in parallel. 2023-04-12 16:26:22 +00:00
utils.py server/tests-py: Declaratively state the HGE environment variables. 2022-09-28 09:21:02 +00:00
validate.py server/tests-py: Parallelize JWT tests. 2023-04-19 10:30:21 +00:00
webhook.py server/tests-py: Start webhook.py inside the test harness. 2022-10-20 19:00:01 +00:00
webserver.py server/tests-py: Start helper services in the test harness. 2022-10-13 15:45:24 +00:00

Python Integration Test Suite

This document describes the Python integration test suite. Please consult the server/CONTRIBUTING document for general information on the overall test setup and other testing suites.

This document describes running and writing tests, as well as some information on how to update test dependencies.

Running tests

Tests can be run using run-new.sh, dev.sh or directly using pytest.

Please note that running the BigQuery tests requires a few manual steps.

Run and test via run-new.sh

The run-new.sh scripts are an active work in progress, and will eventually replace the dev.sh option below.

The easiest way to run the test suite is to:

Run the Python integration tests with ./server/tests-py/run-new.sh.

Filter on specific test files with ./server/tests-py/run-new.sh -- create_async_action_with_nested_output_and_relation.py

If you have any issues with run-new.sh, please create a GitHub issue and run and test via dev.sh instead.

Running tests via dev.sh

scripts/dev.sh test --integration

NOTE: this only runs the tests for Postgres. If you want to run tests for a different backend, use:

scripts/dev.sh test --integration --backend mssql

Available options are documented in scripts/parse-pytest-backend:

  • postgres (default)
  • bigquery (see section below)
  • citus
  • mssql

Filtering tests

You can filter tests by using -k <name>. Note that <name> is case- insensitive.

scripts/dev.sh test --integration --backend mssql -k MSSQL

Note that you can also use expressions here, for example:

scripts/dev.sh test --integration --backend mssql -k "MSSQL and not Permission"

See pytest docs for more details.

Failures

If you want to stop after the first test failure you can pass -x:

scripts/dev.sh test --integration --backend mssql -k MSSQL -x

Verbosity

You can increase or decrease the log verbosity by adding -v or -q to the command.

Running tests directly

WARNING: running tests manually will force skipping of some tests. dev.sh deals with setting up some environment variables which decide how and if some of the tests are executed.

  1. To run the Python tests, youll need to install the necessary Python dependencies first. It is recommended that you do this in a self-contained Python venv, which is supported by Python 3.3+ out of the box. To create one, run:

    python3 -m venv .python-venv
    

    (The second argument names a directory where the venv sandbox will be created; it can be anything you like, but .python-venv is .gitignored.)

    With the venv created, you can enter into it in your current shell session by running:

    source .python-venv/bin/activate
    

    (Source .python-venv/bin/activate.fish instead if you are using fish as your shell.)

  2. Install the necessary Python dependencies into the sandbox:

    pip3 install -r tests-py/requirements.txt
    
  3. Install the dependencies for the Node server used by the remote schema tests:

    (cd tests-py/remote_schemas/nodejs && npm ci)
    
  4. Start an instance of graphql-engine for the test suite to use:

    env EVENT_WEBHOOK_HEADER=MyEnvValue \
        EVENT_WEBHOOK_HANDLER=http://localhost:5592 \
        SCHEDULED_TRIGGERS_WEBHOOK_DOMAIN=http://127.0.0.1:5594 \
      cabal new-run -- exe:graphql-engine \
        --database-url='postgres://<user>:<password>@<host>:<port>/<dbname>' \
        serve --stringify-numeric-types
    

    Optionally, replace the --database-url parameter with --metadata-database-url to enable testing against multiple sources.

    The environment variables are needed for a couple of tests, and the --stringify-numeric-types option is used to avoid the need to do floating-point comparisons.

  5. Optionally, add more sources to test against:

    If the tests include more sources (e.g., by using -k MSSQL), then you can use the following commands to add sources to your running graphql instance:

    # Add a Postgres source
    curl "$METADATA_URL" \
    --data-raw '{"type":"pg_add_source","args":{"name":"default","configuration":{"connection_info":{"database_url":"'"$POSTGRES_DB_URL"'","pool_settings":{}}}}}'
    
    # Add a SQL Server source
    curl "$METADATA_URL" \
    --data-raw '{"type":"mssql_add_source","args":{"name":"mssql","configuration":{"connection_info":{"connection_string":"'"$MSSQL_DB_URL"'","pool_settings":{}}}}}'
    
    # Optionally verify sources have been added
    curl "$METADATA_URL" --data-raw '{"type":"export_metadata","args":{}}'
    
  6. With the server running, run the test suite:

    cd tests-py
    pytest --hge-urls http://localhost:8080 \
           --pg-urls 'postgres://<user>:<password>@<host>:<port>/<dbname>'
    

This will run all the tests, which can take a couple minutes (especially since some of the tests are slow). You can configure pytest to run only a subset of the tests; see the pytest documentation for more details.

Some other useful points of note:

  • It is recommended to use a separate Postgres database for testing, since the tests will drop and recreate the hdb_catalog schema, and they may fail if certain tables already exist. (Its also useful to be able to just drop and recreate the entire test database if it somehow gets into a bad state.)

  • You can pass the -v or -vv options to pytest to enable more verbose output while running the tests and in test failures. You can also pass the -l option to display the current values of Python local variables in test failures.

  • Tests can be run against a specific backend (defaulting to Postgres) with the backend flag, for example:

      pytest --hge-urls http://localhost:8080 \
             --pg-urls 'postgres://<user>:<password>@<host>:<port>/<dbname>'
             --backend mssql -k TestGraphQLQueryBasicCommon
    

For more details, please consult pytest --help.

Running BigQuery tests

Running integration tests against a BigQuery data source is a little more involved due to the necessary service account requirements:

HASURA_BIGQUERY_PROJECT_ID=# the project ID of the service account
HASURA_BIGQUERY_SERVICE_ACCOUNT_EMAIL=# eg. "<<SERVICE_ACCOUNT_NAME>>@<<PROJECT_NAME>>.iam.gserviceaccount.com"
HASURA_BIGQUERY_SERVICE_KEY=# the service account key

Before running the test suite:

  1. Ensure you have access to a Google Cloud Console service account. Store the project ID and account email in HASURA_BIGQUERY_PROJECT_ID variable.
  2. Create and download a new service account key. Store the contents of file in a HASURA_BIGQUERY_SERVICE_KEY variable.
    export HASURA_BIGQUERY_SERVICE_KEY=$(cat /path/to/service/account)
    
  3. Login and activate the service account, if it is not already activated.
  4. Verify the service account is accessible via the BigQuery API:
    1. Run the following command:
    source scripts/verify-bigquery-creds.sh $HASURA_BIGQUERY_PROJECT_ID $HASURA_BIGQUERY_SERVICE_KEY $HASURA_BIGQUERY_SERVICE_ACCOUNT_EMAIL
    
    If the query succeeds, the service account is setup correctly to run tests against BigQuery locally.
  5. Finally, run the BigQuery test suite with HASURA_BIGQUERY_SERVICE_KEY and HASURA_BIGQUERY_PROJECT_ID environment variables set. For example:
scripts/dev.sh test --integration --backend bigquery -k TestGraphQLQueryBasicBigquery

Note to Hasura team: a service account is already setup for internal use, please check the wiki for further details.

Tests structure

  • Tests are grouped as test classes in test modules (names starting with test_)

  • The configuration files (if needed) for the tests in a class are usually kept in one folder.

    • The folder name is usually either the dir variable or the dir() function
  • Some tests (like in test_graphql_queries.py) requires a setup and teardown per class.

    • Here we are extending the DefaultTestSelectQueries class.
    • This class defines a fixture which will run the configurations in setup.yaml and teardown.yaml once per class
    • Extending test class should define a function name dir(), which returns the configuration folder
  • For mutation tests (like in test_graphql_mutations.py)

    • We need a schema_setup and schema_teardown per class
    • And values_setup and values_teardown per test
    • Doing schema setup and teardown per test is expensive.
    • We are extending the DefaultTestMutations class for this.
    • This class defines a fixture which will run the configuration in setup.yaml and teardown.yaml once per class.
    • Another fixture defined in this class runs the configuration in values_setup.yaml and values_teardown.yaml once per class.

Writing python tests

  1. Check whether the test you intend to write already exists in the test suite, so that there will be no duplicate tests or the existing test will just need to be modified.

  2. All the tests use setup and teardown, the setup step is used to initialize the graphql-engine and the database in a certain state after which the tests should be run. After the tests are run, the state needs to be cleared, which should be done in the teardown step. The setup and teardown is localised for every python test class.

    See TestCreateAndDelete in test_events.py for reference.

  3. The setup and teardown can be configured to run before and after every test in a test class or run before and after running all the tests in a class. Depending on the use case, there are different fixtures like per_class_tests_db_state,per_method_tests_db_state defined in the conftest.py file.

  4. Sometimes, it's required to run the graphql-engine with in a different configuration only for a particular set of tests. In this case, these tests should be run only when the graphql-engine is run with the said configuration and should be skipped in other graphql-engine configurations. This can be done by accepting a new command-line flag from the pytest command and depending on the value or presence of the flag, the tests should be run accordingly. After adding this kind of a test, a new section needs to be added in the test-server.sh. This new section's name should also be added in the server-test-names.txt file, otherwise the test will not be run in the CI.

    For example,

    The tests in the test_remote_schema_permissions.py are only to be run when the remote schema permissions are enabled in the graphql-engine and when it's not set, these tests should be skipped. Now, to run these tests we parse a command line option from pytest called (--enable-remote-schema-permissions) and the presence of this flag means that we need to run these tests. When the tests are run with this command line option, it's assumed that the server has enabled remote schema permissions.

Adding test support for a new backend

The current workflow for supporting a new backend in integration tests is as follows:

  1. Add functions to launch and cleanup a server for the new backend. Example.
  2. Augment dev.sh to support the new backend. Example.
  3. Connect the GraphQL Engine to the database you've just launched. Example.
  4. Add setup and teardown files:
    1. setup_<backend>: for v1/query or metadata queries such as <backend>_track_table. Example.
    2. schema_setup_<backend>: for v2/query queries such as <backend>_run_sql. Example.
    3. teardown_<backend> and cleardb_<backend>
    4. Important: filename suffixes should be the same as the value thats being passed to —backend; that's how the files are looked up.
  5. Specify a backend parameter for the per_backend_test_class and per_backend_test_function fixtures, parameterised by backend. Example.

Note: When teardown is not disabled (via skip_teardown(*) , in which case, this phase is skipped entirely), teardown.yaml always runs before schema_teardown.yaml, even if the tests fail. See setup_and_teardown in server/tests-py/conftest.py for the full source code/logic.

(*): See setup_and_teardown_v1q and setup_and_teardown_v2q in conftest.py for more details.

This means, for example, that if teardown.yaml untracks a table, and schema_teardown.yaml runs raw SQL to drop the table, both would succeed (assuming the table is tracked/exists).

Test suite naming convention The current convention is to indicate the backend(s) tests can be run against in the class name. For example:

  • TestGraphQLQueryBasicMSSQL for tests that can only be run against a SQL Server backend
  • TestGraphQLQueryBasicCommon for tests that can be run against more than one backend
  • If a test class doesn't have a suffix specifying the backend, nor does its name end in Common, then it is likely a test written pre-v2.0 that can only be run on Postgres

This naming convention enables easier test filtering with pytest command line flags.

The backend-specific and common test suites are disjoint; for example, run pytest --integration -k "Common or MSSQL" --backend mssql to run all MSSQL tests.

Note that --backend does not interact with the selection of tests. You will generally have to combine --backend with -k.

Updating Python requirements

The packages/requirements are documented in two files:

  • server/tests-py/requirements-top-level.txt
  • server/tests-py/requirements.txt

The server/tests-py/requirements-top-level.txt file is the main file. It contains the direct dependencies along with version requirements we know we should be careful about.

The server/tests-py/requirements.txt file is the lock file. It holds version numbers for all direct and transitive dependencies. This file can be re-generated by:

  1. alter server/tests-py/requirements-top-level.txt
  2. remove server/tests-py/requirements.txt
  3. run dev.sh test --integration
  4. update DEVSH_VERSION in scripts/dev.sh to force reinstall these dependencies

Steps 3 can be done manually:

pip3 install -r requirements-top-level.txt
pip3 freeze > requirements.txt