GitOrigin-RevId: 9e99047f7ab83f06c94c4741c7c0591eb2ca35e8
12 KiB
Contributing
This guide explains how to set up the graphql-engine server for development on your own machine and how to contribute.
Pre-requisites
- GHC 8.10.2 and cabal-install
- There are various ways these can be installed, but ghcup is a good choice if you’re not sure.
- Node.js (>= v8.9)
- npm >= 5.7
- gsutil
- libpq-dev
- libkrb5-dev
- openssl and libssl-dev
- python >= 3.5 with pip3 and virtualenv
The last few prerequisites can be installed on Debian or Ubuntu with:
$ sudo apt install libpq-dev libkrb5-dev python3 python3-pip python3-venv openssl libssl-dev
Additionally, you will need a way to run a Postgres database server. The dev.sh
script (described below) can set up a Postgres instance for you via Docker, but if you want to run it yourself, you’ll need:
- PostgreSQL >= 9.5
- postgis
Upgrading npm
If your npm is too old (>= 5.7 required):
$ npm install -g npm@latest # sudo may be required
or update your nodejs.
Development workflow
You should fork the repo on github and then git clone https://github.com/<your-username>/graphql-engine
.
After making your changes
Compile
...console assets:
$ cd console
$ npm ci
$ npm run server-build
$ cd ..
...and the server:
$ cd server
$ ln -s cabal.project.dev cabal.project.local
$ cabal new-update
$ cabal new-build
To set up the project configuration to coincide with the testing scripts below, thus avoiding recompilation when testing locally, rather use cabal.project.dev-sh.local
instead of cabal.project.dev
:
$ ln -s cabal.project.dev-sh.local cabal.project.local
IDE Support
You may want to use hls/ghcide if your editor has LSP support. A sample configuration has been provided which can be used as follows:
ln -s sample.hie.yaml hie.yaml
If you have to customise any of the options for ghcide/hls, you should instead copy the sample file and make necessary changes in hie.yaml
file. Note that hie.yaml
is gitignored so the changes will be specific to your machine.
cp sample.hie.yaml hie.yaml
Run and test via dev.sh
The dev.sh
script in the top-level scripts/
directory is a turnkey solution to build, run, and
test graphql-engine
using a Docker container to run a Postgres database. Docker is necessary to
use dev.sh
.
To use dev.sh
, first launch a new postgres container with:
$ scripts/dev.sh postgres
Then in a new terminal launch graphql-engine
in dev mode with:
$ scripts/dev.sh graphql-engine
The dev.sh
will print some helpful information and logs from both services
will be printed to screen.
You can run the test suite with:
$ scripts/dev.sh test
This should run in isolation. The output format is described in the pytest documentation. Errors and failures are indicated by F
s and E
s.
Optionally, launch a new container for alternative (MSSQL) backend with:
$ scripts/dev.sh mssql
Tests can be run against a specific backend (defaulting to Postgres) with the backend
flag, for example:
$ scripts/dev.sh test --integration -k TestGraphQLQueryBasicCommon --backend mssql
Run and test manually
If you want, you can also run the server and test suite manually against an instance of your choosing.
Run
The following command can be used to build and launch a local graphql-engine
instance:
cabal new-run -- exe:graphql-engine \
--database-url='postgres://<user>:<password>@<host>:<port>/<dbname>' \
serve --enable-console --console-assets-dir=../console/static/dist
This will launch a server on port 8080, and it will serve the console assets if they were built with npm run server-build
as mentioned above.
Test
graphql-engine
has two test suites:
-
A small set of unit tests and integration tests written in Haskell.
-
An extensive set of end-to-end tests written in Python.
Both sets of tests require a running Postgres database.
Running the Haskell test suite
cabal new-run -- test:graphql-engine-tests \
--database-url='postgres://<user>:<password>@<host>:<port>/<dbname>'
Running the Python test suite
-
To run the Python tests, you’ll need to install the necessary Python dependencies first. It is recommended that you do this in a self-contained Python venv, which is supported by Python 3.3+ out of the box. To create one, run:
python3 -m venv .python-venv
(The second argument names a directory where the venv sandbox will be created; it can be anything you like, but
.python-venv
is.gitignore
d.)With the venv created, you can enter into it in your current shell session by running:
source .python-venv/bin/activate
(Source
.python-venv/bin/activate.fish
instead if you are usingfish
as your shell.) -
Install the necessary Python dependencies into the sandbox:
pip3 install -r tests-py/requirements.txt
-
Install the dependencies for the Node server used by the remote schema tests:
(cd tests-py/remote_schemas/nodejs && npm ci)
-
Start an instance of
graphql-engine
for the test suite to use:env EVENT_WEBHOOK_HEADER=MyEnvValue \ WEBHOOK_FROM_ENV=http://localhost:5592/ \ SCHEDULED_TRIGGERS_WEBHOOK_DOMAIN=http://127.0.0.1:5594 \ cabal new-run -- exe:graphql-engine \ --database-url='postgres://<user>:<password>@<host>:<port>/<dbname>' \ serve --stringify-numeric-types
The environment variables are needed for a couple tests, and the
--stringify-numeric-types
option is used to avoid the need to do floating-point comparisons. -
With the server running, run the test suite:
cd tests-py pytest --hge-urls http://localhost:8080 \ --pg-urls 'postgres://<user>:<password>@<host>:<port>/<dbname>'
This will run all the tests, which can take a couple minutes (especially since some of the tests are slow). You can configure pytest
to run only a subset of the tests; see the pytest
documentation for more details.
Some other useful points of note:
-
It is recommended to use a separate Postgres database for testing, since the tests will drop and recreate the
hdb_catalog
schema, and they may fail if certain tables already exist. (It’s also useful to be able to just drop and recreate the entire test database if it somehow gets into a bad state.) -
You can pass the
-v
or-vv
options topytest
to enable more verbose output while running the tests and in test failures. You can also pass the-l
option to display the current values of Python local variables in test failures.
Guide on writing python tests
-
Check whether the test you intend to write already exists in the test suite, so that there will be no duplicate tests or the existing test will just need to be modified.
-
All the tests use setup and teardown, the setup step is used to initialize the graphql-engine and the database in a certain state after which the tests should be run. After the tests are run, the state needs to be cleared, which should be done in the teardown step. The setup and teardown is localised for every python test class.
See
TestCreateAndDelete
in test_events.py for reference. -
The setup and teardown can be configured to run before and after every test in a test class or run before and after running all the tests in a class. Depending on the use case, there are different fixtures like
per_class_tests_db_state
,per_method_tests_db_state
defined in the conftest.py file. -
Sometimes, it's required to run the graphql-engine with in a different configuration only for a particular set of tests. In this case, these tests should be run only when the graphql-engine is run with the said configuration and should be skipped in other graphql-engine configurations. This can be done by accepting a new command-line flag from the
pytest
command and depending on the value or presence of the flag, the tests should be run accordingly. After adding this kind of a test, a new section needs to be added in the test-server.sh. This new section's name should also be added in theserver-test-names.txt
file, otherwise the test will not be run in the CI.For example,
The tests in the test_remote_schema_permissions.py are only to be run when the remote schema permissions are enabled in the graphql-engine and when it's not set, these tests should be skipped. Now, to run these tests we parse a command line option from pytest called (
--enable-remote-schema-permissions
) and the presence of this flag means that we need to run these tests. When the tests are run with this command line option, it's assumed that the server has enabled remote schema permissions.
Adding test support for a new backend
The current workflow for supporting a new backend in integration tests is as follows:
- Add functions to launch and cleanup a server for the new backend. Example.
- Connect to the database you've just launched. Example.
- Add setup and teardown files:
setup_<backend>
: forv1/query
or metadata queries such as<backend>_track_table
. Example.schema_setup_<backend>
: forv2/query
queries such as<backend>_run_sql
. Example.teardown_<backend>
andcleardb_<backend>
- important: filename suffixes should be the same as the value that’s being passed to
—backend
; that's how the files are looked up.
- Write test using the
per_backend_tests
fixture, parameterised by backend. Example. - Optional: Run the existing (Postgres) test suite against the new backend to identify and group common and backend-specific tests into their own classes.
Tests against alternative backends aren't yet run/supported in CI, so please test locally.
Create Pull Request
- Make sure your commit messages meet the guidelines.
- If you changed the versions of any dependencies, run
cabal new-freeze
to update the freeze file. - Create a pull request from your forked repo to the main repo.
- Every pull request will automatically build and run the tests.
Code conventions
This helps enforce a uniform style for all committers.
- Compiler warnings are turned on, make sure your code has no warnings.
- Use hlint to make sure your code has no warnings.
You can use our custom hlint config with
$ hlint --hint=server/.hlint.yaml .
- Use stylish-haskell to format your code.