|
2 | 2 |
|
3 | 3 | from typing import TYPE_CHECKING |
4 | 4 |
|
5 | | -from apify_shared.consts import ApifyEnvVars |
6 | | - |
7 | | -from ._utils import generate_unique_resource_name |
8 | 5 | from apify import Actor |
9 | 6 |
|
10 | 7 | if TYPE_CHECKING: |
11 | | - import pytest |
12 | | - |
13 | | - from apify_client import ApifyClientAsync |
14 | | - |
15 | 8 | from .conftest import MakeActorFunction, RunActorFunction |
16 | 9 |
|
17 | 10 |
|
18 | | -async def test_push_and_verify_data_in_default_dataset( |
19 | | - make_actor: MakeActorFunction, |
20 | | - run_actor: RunActorFunction, |
21 | | -) -> None: |
22 | | - desired_item_count = 100 # Also change inside main() if you're changing this |
23 | | - |
24 | | - async def main() -> None: |
25 | | - desired_item_count = 100 |
26 | | - async with Actor: |
27 | | - await Actor.push_data([{'id': i} for i in range(desired_item_count)]) |
28 | | - |
29 | | - actor = await make_actor(label='push-data', main_func=main) |
30 | | - run_result = await run_actor(actor) |
31 | | - |
32 | | - assert run_result.status == 'SUCCEEDED' |
33 | | - |
34 | | - list_page = await actor.last_run().dataset().list_items() |
35 | | - assert list_page.items[0]['id'] == 0 |
36 | | - assert list_page.items[-1]['id'] == desired_item_count - 1 |
37 | | - assert len(list_page.items) == list_page.count == desired_item_count |
38 | | - |
39 | | - |
40 | | -async def test_push_large_data_chunks_over_9mb( |
41 | | - make_actor: MakeActorFunction, |
42 | | - run_actor: RunActorFunction, |
43 | | -) -> None: |
44 | | - async def main() -> None: |
45 | | - async with Actor: |
46 | | - await Actor.push_data([{'str': 'x' * 10000} for _ in range(5000)]) # ~50MB |
47 | | - |
48 | | - actor = await make_actor(label='push-data-over-9mb', main_func=main) |
49 | | - run_result = await run_actor(actor) |
50 | | - |
51 | | - assert run_result.status == 'SUCCEEDED' |
52 | | - |
53 | | - async for item in actor.last_run().dataset().iterate_items(): |
54 | | - assert item['str'] == 'x' * 10000 |
55 | | - |
56 | | - |
57 | | -async def test_dataset_iter_items( |
58 | | - make_actor: MakeActorFunction, |
59 | | - run_actor: RunActorFunction, |
60 | | -) -> None: |
61 | | - async def main() -> None: |
62 | | - inserted_data = {'Something': 'something else'} |
63 | | - |
64 | | - async with Actor: |
65 | | - dataset = await Actor.open_dataset() |
66 | | - await dataset.push_data(inserted_data) |
67 | | - requested_data = [item async for item in dataset.iterate_items()] |
68 | | - |
69 | | - assert len(requested_data) == 1 |
70 | | - assert requested_data[0] == inserted_data |
71 | | - |
72 | | - actor = await make_actor(label='test_dataset_iter_items', main_func=main) |
73 | | - run_result = await run_actor(actor) |
74 | | - assert run_result.status == 'SUCCEEDED' |
75 | | - |
76 | | - |
77 | | -async def test_same_references_in_default_dataset( |
78 | | - make_actor: MakeActorFunction, |
79 | | - run_actor: RunActorFunction, |
80 | | -) -> None: |
81 | | - async def main() -> None: |
82 | | - async with Actor: |
83 | | - dataset1 = await Actor.open_dataset() |
84 | | - dataset2 = await Actor.open_dataset() |
85 | | - assert dataset1 is dataset2 |
86 | | - |
87 | | - actor = await make_actor(label='dataset-same-ref-default', main_func=main) |
88 | | - run_result = await run_actor(actor) |
89 | | - |
90 | | - assert run_result.status == 'SUCCEEDED' |
91 | | - |
92 | | - |
93 | | -async def test_same_references_in_named_dataset( |
94 | | - make_actor: MakeActorFunction, |
95 | | - run_actor: RunActorFunction, |
96 | | -) -> None: |
97 | | - dataset_name = generate_unique_resource_name('dataset') |
98 | | - |
99 | | - async def main() -> None: |
100 | | - async with Actor: |
101 | | - input_object = await Actor.get_input() |
102 | | - dataset_name = input_object['datasetName'] |
103 | | - dataset_by_name_1 = await Actor.open_dataset(name=dataset_name) |
104 | | - dataset_by_name_2 = await Actor.open_dataset(name=dataset_name) |
105 | | - assert dataset_by_name_1 is dataset_by_name_2 |
106 | | - |
107 | | - dataset_1_metadata = await dataset_by_name_1.get_metadata() |
108 | | - dataset_by_id_1 = await Actor.open_dataset(id=dataset_1_metadata.id) |
109 | | - dataset_by_id_2 = await Actor.open_dataset(id=dataset_1_metadata.id) |
110 | | - assert dataset_by_id_1 is dataset_by_name_1 |
111 | | - assert dataset_by_id_2 is dataset_by_id_1 |
112 | | - |
113 | | - await dataset_by_name_1.drop() |
114 | | - |
115 | | - actor = await make_actor(label='dataset-same-ref-named', main_func=main) |
116 | | - run_result = await run_actor(actor, run_input={'datasetName': dataset_name}) |
117 | | - |
118 | | - assert run_result.status == 'SUCCEEDED' |
119 | | - |
120 | | - |
121 | | -async def test_force_cloud( |
122 | | - apify_client_async: ApifyClientAsync, |
123 | | - monkeypatch: pytest.MonkeyPatch, |
124 | | -) -> None: |
125 | | - assert apify_client_async.token is not None |
126 | | - monkeypatch.setenv(ApifyEnvVars.TOKEN, apify_client_async.token) |
127 | | - |
128 | | - dataset_name = generate_unique_resource_name('dataset') |
129 | | - dataset_item = {'foo': 'bar'} |
130 | | - |
131 | | - async with Actor: |
132 | | - dataset = await Actor.open_dataset(name=dataset_name, force_cloud=True) |
133 | | - dataset_id = (await dataset.get_metadata()).id |
134 | | - |
135 | | - await dataset.push_data(dataset_item) |
136 | | - |
137 | | - dataset_client = apify_client_async.dataset(dataset_id) |
138 | | - |
139 | | - try: |
140 | | - dataset_details = await dataset_client.get() |
141 | | - assert dataset_details is not None |
142 | | - assert dataset_details.get('name') == dataset_name |
143 | | - |
144 | | - dataset_items = await dataset_client.list_items() |
145 | | - assert dataset_items.items == [dataset_item] |
146 | | - finally: |
147 | | - await dataset_client.delete() |
148 | | - |
149 | | - |
150 | 11 | async def test_dataset_defaults( |
151 | 12 | make_actor: MakeActorFunction, |
152 | 13 | run_actor: RunActorFunction, |
|
0 commit comments