Skip to content

Commit

Permalink
Allow chaining of 'add_file' calls
Browse files Browse the repository at this point in the history
  • Loading branch information
koenvo committed Sep 24, 2024
1 parent 15a1284 commit b6e3732
Show file tree
Hide file tree
Showing 2 changed files with 54 additions and 51 deletions.
3 changes: 3 additions & 0 deletions ingestify/domain/models/resources/dataset_resource.py
Original file line number Diff line number Diff line change
Expand Up @@ -94,3 +94,6 @@ def add_file(
)

self.files[file_id] = file_resource

# Allow chaining
return self
102 changes: 51 additions & 51 deletions ingestify/tests/test_engine.py
Original file line number Diff line number Diff line change
Expand Up @@ -66,35 +66,36 @@ def find_datasets(
season_id,
**kwargs
):
dataset_resource = DatasetResource(
dict(
competition_id=competition_id,
season_id=season_id,
),
provider="fake",
dataset_type="match",
name="Test Dataset",
)

last_modified = datetime.now(pytz.utc)

dataset_resource.add_file(
last_modified=last_modified,
data_feed_key="file1",
data_spec_version="v1",
file_loader=file_loader,
)
dataset_resource.add_file(
last_modified=last_modified,
data_feed_key="file2",
data_spec_version="v1",
file_loader=file_loader,
)
dataset_resource.add_file(
last_modified=last_modified,
data_feed_key="file3",
data_spec_version="v1",
json_content={"test": "some-content"},
yield (
DatasetResource(
dict(
competition_id=competition_id,
season_id=season_id,
),
provider="fake",
dataset_type="match",
name="Test Dataset",
)
.add_file(
last_modified=last_modified,
data_feed_key="file1",
data_spec_version="v1",
file_loader=file_loader,
)
.add_file(
last_modified=last_modified,
data_feed_key="file2",
data_spec_version="v1",
file_loader=file_loader,
)
.add_file(
last_modified=last_modified,
data_feed_key="file3",
data_spec_version="v1",
json_content={"test": "some-content"},
)
)
# dataset_resource.add_file(
# last_modified=last_modified,
Expand All @@ -104,8 +105,6 @@ def find_datasets(
# data_serialization_format="json"
# )

yield dataset_resource


class BatchSource(Source):
def __init__(self, name, callback):
Expand All @@ -128,30 +127,31 @@ def find_datasets(
for i in range(10):
match_id = self.idx
self.idx += 1
dataset_resource = DatasetResource(
dict(
competition_id=competition_id,
season_id=season_id,
match_id=match_id,
),
name="Test dataset",
provider="fake",
dataset_type="match",
)

last_modified = datetime.now(pytz.utc)

dataset_resource.add_file(
last_modified=last_modified,
data_feed_key="file1",
data_spec_version="v1",
file_loader=file_loader,
)
dataset_resource.add_file(
last_modified=last_modified,
data_feed_key="file2",
data_spec_version="v1",
file_loader=file_loader,
dataset_resource = (
DatasetResource(
dict(
competition_id=competition_id,
season_id=season_id,
match_id=match_id,
),
name="Test dataset",
provider="fake",
dataset_type="match",
)
.add_file(
last_modified=last_modified,
data_feed_key="file1",
data_spec_version="v1",
file_loader=file_loader,
)
.add_file(
last_modified=last_modified,
data_feed_key="file2",
data_spec_version="v1",
file_loader=file_loader,
)
)

items.append(dataset_resource)
Expand Down

0 comments on commit b6e3732

Please sign in to comment.