-
Notifications
You must be signed in to change notification settings - Fork 4.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
New source: Zenhub Graphql #33986
New source: Zenhub Graphql #33986
Changes from all commits
e9e3c1d
365d432
6e05c77
1b6572f
8784bb3
80ee7bf
3eaf3d3
c130ab2
7e16b6c
9d22ef3
dfb52e7
f6b56c8
f105de1
0b414f5
469b229
706c101
5baa44d
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,166 @@ | ||
# Zenhub Graphql Source | ||
|
||
This is the repository for the Zenhub Graphql source connector, written in Python. | ||
For information about how to use this connector within Airbyte, see [the documentation](https://docs.airbyte.com/integrations/sources/zenhub-graphql). | ||
|
||
## Local development | ||
|
||
### Prerequisites | ||
**To iterate on this connector, make sure to complete this prerequisites section.** | ||
|
||
#### Minimum Python version required `= 3.9.0` | ||
|
||
#### Activate Virtual Environment and install dependencies | ||
From this connector directory, create a virtual environment: | ||
``` | ||
python -m venv .venv | ||
``` | ||
|
||
This will generate a virtualenv for this module in `.venv/`. Make sure this venv is active in your | ||
development environment of choice. To activate it from the terminal, run: | ||
``` | ||
source .venv/bin/activate | ||
pip install -r requirements.txt | ||
pip install '.[tests]' | ||
``` | ||
If you are in an IDE, follow your IDE's instructions to activate the virtualenv. | ||
|
||
Note that while we are installing dependencies from `requirements.txt`, you should only edit `setup.py` for your dependencies. `requirements.txt` is | ||
used for editable installs (`pip install -e`) to pull in Python dependencies from the monorepo and will call `setup.py`. | ||
If this is mumbo jumbo to you, don't worry about it, just put your deps in `setup.py` but install using `pip install -r requirements.txt` and everything | ||
should work as you expect. | ||
|
||
#### Create credentials | ||
**If you are a community contributor**, follow the instructions in the [documentation](https://docs.airbyte.com/integrations/sources/zenhub-graphql) | ||
to generate the necessary credentials. Then create a file `secrets/config.json` conforming to the `source_zenhub_graphql/spec.yaml` file. | ||
Note that any directory named `secrets` is gitignored across the entire Airbyte repo, so there is no danger of accidentally checking in sensitive information. | ||
See `integration_tests/sample_config.json` for a sample config file. | ||
|
||
**If you are an Airbyte core member**, copy the credentials in Lastpass under the secret name `source zenhub-graphql test creds` | ||
and place them into `secrets/config.json`. | ||
|
||
### Locally running the connector | ||
``` | ||
python main.py spec | ||
python main.py check --config secrets/config.json | ||
python main.py discover --config secrets/config.json | ||
python main.py read --config secrets/config.json --catalog integration_tests/configured_catalog.json | ||
``` | ||
|
||
### Locally running the connector docker image | ||
|
||
#### Use `airbyte-ci` to build your connector | ||
The Airbyte way of building this connector is to use our `airbyte-ci` tool. | ||
You can follow install instructions [here](https://github.com/airbytehq/airbyte/blob/master/airbyte-ci/connectors/pipelines/README.md#L1). | ||
Then running the following command will build your connector: | ||
|
||
```bash | ||
airbyte-ci connectors --name source-zenhub-graphql build | ||
``` | ||
Once the command is done, you will find your connector image in your local docker registry: `airbyte/source-zenhub-graphql:dev`. | ||
|
||
##### Customizing our build process | ||
When contributing on our connector you might need to customize the build process to add a system dependency or set an env var. | ||
You can customize our build process by adding a `build_customization.py` module to your connector. | ||
This module should contain a `pre_connector_install` and `post_connector_install` async function that will mutate the base image and the connector container respectively. | ||
It will be imported at runtime by our build process and the functions will be called if they exist. | ||
|
||
Here is an example of a `build_customization.py` module: | ||
```python | ||
from __future__ import annotations | ||
|
||
from typing import TYPE_CHECKING | ||
|
||
if TYPE_CHECKING: | ||
# Feel free to check the dagger documentation for more information on the Container object and its methods. | ||
# https://dagger-io.readthedocs.io/en/sdk-python-v0.6.4/ | ||
from dagger import Container | ||
|
||
|
||
async def pre_connector_install(base_image_container: Container) -> Container: | ||
return await base_image_container.with_env_variable("MY_PRE_BUILD_ENV_VAR", "my_pre_build_env_var_value") | ||
|
||
async def post_connector_install(connector_container: Container) -> Container: | ||
return await connector_container.with_env_variable("MY_POST_BUILD_ENV_VAR", "my_post_build_env_var_value") | ||
``` | ||
|
||
#### Build your own connector image | ||
This connector is built using our dynamic built process in `airbyte-ci`. | ||
The base image used to build it is defined within the metadata.yaml file under the `connectorBuildOptions`. | ||
The build logic is defined using [Dagger](https://dagger.io/) [here](https://github.com/airbytehq/airbyte/blob/master/airbyte-ci/connectors/pipelines/pipelines/builds/python_connectors.py). | ||
It does not rely on a Dockerfile. | ||
|
||
If you would like to patch our connector and build your own a simple approach would be to: | ||
|
||
1. Create your own Dockerfile based on the latest version of the connector image. | ||
```Dockerfile | ||
FROM airbyte/source-zenhub-graphql:latest | ||
|
||
COPY . ./airbyte/integration_code | ||
RUN pip install ./airbyte/integration_code | ||
|
||
# The entrypoint and default env vars are already set in the base image | ||
# ENV AIRBYTE_ENTRYPOINT "python /airbyte/integration_code/main.py" | ||
# ENTRYPOINT ["python", "/airbyte/integration_code/main.py"] | ||
``` | ||
Please use this as an example. This is not optimized. | ||
|
||
2. Build your image: | ||
```bash | ||
docker build -t airbyte/source-zenhub-graphql:dev . | ||
# Running the spec command against your patched connector | ||
docker run airbyte/source-zenhub-graphql:dev spec | ||
```` | ||
|
||
#### Run | ||
Then run any of the connector commands as follows: | ||
``` | ||
docker run --rm airbyte/source-zenhub-graphql:dev spec | ||
docker run --rm -v $(pwd)/secrets:/secrets airbyte/source-zenhub-graphql:dev check --config /secrets/config.json | ||
docker run --rm -v $(pwd)/secrets:/secrets airbyte/source-zenhub-graphql:dev discover --config /secrets/config.json | ||
docker run --rm -v $(pwd)/secrets:/secrets -v $(pwd)/integration_tests:/integration_tests airbyte/source-zenhub-graphql:dev read --config /secrets/config.json --catalog /integration_tests/configured_catalog.json | ||
``` | ||
## Testing | ||
Make sure to familiarize yourself with [pytest test discovery](https://docs.pytest.org/en/latest/goodpractices.html#test-discovery) to know how your test files and methods should be named. | ||
First install test dependencies into your virtual environment: | ||
``` | ||
pip install .[tests] | ||
``` | ||
### Unit Tests | ||
To run unit tests locally, from the connector directory run: | ||
``` | ||
python -m pytest unit_tests | ||
``` | ||
|
||
### Integration Tests | ||
There are two types of integration tests: Acceptance Tests (Airbyte's test suite for all source connectors) and custom integration tests (which are specific to this connector). | ||
#### Custom Integration tests | ||
Place custom tests inside `integration_tests/` folder, then, from the connector root, run | ||
``` | ||
python -m pytest integration_tests | ||
``` | ||
|
||
### Acceptance Tests | ||
Customize `acceptance-test-config.yml` file to configure tests. See [Connector Acceptance Tests](https://docs.airbyte.com/connector-development/testing-connectors/connector-acceptance-tests-reference) for more information. | ||
If your connector requires to create or destroy resources for use during acceptance tests create fixtures for it and place them inside integration_tests/acceptance.py. | ||
Please run acceptance tests via [airbyte-ci](https://github.com/airbytehq/airbyte/blob/master/airbyte-ci/connectors/pipelines/README.md#connectors-test-command): | ||
```bash | ||
airbyte-ci connectors --name source-zenhub-graphql test | ||
``` | ||
|
||
## Dependency Management | ||
All of your dependencies should go in `setup.py`, NOT `requirements.txt`. The requirements file is only used to connect internal Airbyte dependencies in the monorepo for local development. | ||
We split dependencies between two groups, dependencies that are: | ||
* required for your connector to work need to go to `MAIN_REQUIREMENTS` list. | ||
* required for the testing need to go to `TEST_REQUIREMENTS` list | ||
|
||
### Publishing a new version of the connector | ||
You've checked out the repo, implemented a million dollar feature, and you're ready to share your changes with the world. Now what? | ||
1. Make sure your changes are passing our test suite: `airbyte-ci connectors --name=source-zenhub-graphql test` | ||
2. Bump the connector version in `metadata.yaml`: increment the `dockerImageTag` value. Please follow [semantic versioning for connectors](https://docs.airbyte.com/contributing-to-airbyte/resources/pull-requests-handbook/#semantic-versioning-for-connectors). | ||
3. Make sure the `metadata.yaml` content is up to date. | ||
4. Make the connector documentation and its changelog is up to date (`docs/integrations/sources/zenhub-graphql.md`). | ||
5. Create a Pull Request: use [our PR naming conventions](https://docs.airbyte.com/contributing-to-airbyte/resources/pull-requests-handbook/#pull-request-title-convention). | ||
6. Pat yourself on the back for being an awesome contributor. | ||
7. Someone from Airbyte will take a look at your PR and iterate with you to merge it into master. | ||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,48 @@ | ||
# See [Connector Acceptance Tests](https://docs.airbyte.com/connector-development/testing-connectors/connector-acceptance-tests-reference) | ||
# for more information about how to configure these tests | ||
connector_image: airbyte/source-zenhub-graphql:dev | ||
acceptance_tests: | ||
spec: | ||
tests: | ||
- spec_path: "source_zenhub_graphql/spec.yaml" | ||
connection: | ||
tests: | ||
- config_path: "secrets/config.json" | ||
status: "succeed" | ||
- config_path: "integration_tests/invalid_config.json" | ||
status: "failed" | ||
discovery: | ||
tests: | ||
- config_path: "secrets/config.json" | ||
basic_read: | ||
tests: | ||
- config_path: "secrets/config.json" | ||
configured_catalog_path: "integration_tests/configured_catalog.json" | ||
empty_streams: [] | ||
# expect_records: | ||
# path: "integration_tests/expected_records.jsonl" | ||
# extra_fields: no | ||
# exact_order: no | ||
# extra_records: yes | ||
# ignored_fields: | ||
# zenhub_issues: | ||
# - name: updatedAt | ||
# bypass_reason: "Dates could change" | ||
# - name: title | ||
# bypass_reason: "Could have changes in the title" | ||
# - name: body | ||
# bypass_reason: "The content will be changing constantly" | ||
# - name: state | ||
# bypass_reason: "The state changes over time (e.g Open to Closed)" | ||
# - name: pipeline_name | ||
# bypass_reason: "The state changes over time (e.g Open to Closed)" | ||
# - name: priority_name | ||
# bypass_reason: "The priority changes depending on th eurgency of the task" | ||
|
||
incremental: | ||
bypass_reason: "This connector does not implement incremental sync" | ||
|
||
full_refresh: | ||
tests: | ||
- config_path: "secrets/config.json" | ||
configured_catalog_path: "integration_tests/configured_catalog.json" |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,3 @@ | ||
# | ||
# Copyright (c) 2023 Airbyte, Inc., all rights reserved. | ||
# |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,5 @@ | ||
{ | ||
"ZenhubWorkspace": { | ||
"": "todo-abnormal-value" | ||
} | ||
} |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,16 @@ | ||
# | ||
# Copyright (c) 2023 Airbyte, Inc., all rights reserved. | ||
# | ||
|
||
|
||
import pytest | ||
|
||
pytest_plugins = ("connector_acceptance_test.plugin",) | ||
|
||
|
||
@pytest.fixture(scope="session", autouse=True) | ||
def connector_setup(): | ||
"""This fixture is a placeholder for external resources that acceptance test might require.""" | ||
# TODO: setup test dependencies if needed. otherwise remove the TODO comments | ||
yield | ||
# TODO: clean up test dependencies |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,31 @@ | ||
{ | ||
"streams": [ | ||
{ | ||
"stream": { | ||
"name": "zenhub_workspace", | ||
"json_schema": {}, | ||
"supported_sync_modes": ["full_refresh"] | ||
}, | ||
"sync_mode": "full_refresh", | ||
"destination_sync_mode": "overwrite" | ||
}, | ||
{ | ||
"stream": { | ||
"name": "zenhub_pipelines", | ||
"json_schema": {}, | ||
"supported_sync_modes": ["full_refresh", "incremental"] | ||
}, | ||
"sync_mode": "full_refresh", | ||
"destination_sync_mode": "overwrite" | ||
}, | ||
{ | ||
"stream": { | ||
"name": "zenhub_issues", | ||
"json_schema": {}, | ||
"supported_sync_modes": ["full_refresh", "incremental"] | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Do you know how you are going to do incremental syncs here? There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Yes, I am thinking on using updated_at |
||
}, | ||
"sync_mode": "full_refresh", | ||
"destination_sync_mode": "overwrite" | ||
} | ||
] | ||
} |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,5 @@ | ||
{ | ||
"access_token": "thisismytokne123" | ||
, "workspace_name": "Python Connectors" | ||
, "repo_names": ["air"] | ||
} |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,5 @@ | ||
{ | ||
"access_token": "<user's_access_token>" | ||
, "workspace_name": "<your_workspace>" | ||
, "repo_names": ["yout_list","of_repos"] | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. typos |
||
} |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,5 @@ | ||
{ | ||
"todo-stream-name": { | ||
"todo-field-name": "value" | ||
} | ||
} |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,7 @@ | ||
# | ||
# Copyright (c) 2023 Airbyte, Inc., all rights reserved. | ||
# | ||
from source_zenhub_graphql.run import run | ||
|
||
if __name__ == "__main__": | ||
run() |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
why do we need a new source? Can't we update the existing zenhub connector?
edit: derp. there's no zenhub connector. Is the fact that this is using graphql relevant to the user? Why not call this the zenhub source?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, you are right. It is not relevant for the user. I will rename it to be just Zenhub
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
+1