Skip to content

Commit

Permalink
Adapt more API tests around history contents to use JSON payload
Browse files Browse the repository at this point in the history
See galaxyproject#12152 for details
  • Loading branch information
davelopez committed Nov 1, 2021
1 parent 646b33a commit ffac0c4
Show file tree
Hide file tree
Showing 4 changed files with 12 additions and 12 deletions.
4 changes: 2 additions & 2 deletions lib/galaxy_test/api/test_jobs.py
Original file line number Diff line number Diff line change
Expand Up @@ -554,7 +554,7 @@ def test_search(self, history_id):
# We first copy the datasets, so that the update time is lower than the job creation time
new_history_id = self.dataset_populator.new_history()
copy_payload = {"content": dataset_id, "source": "hda", "type": "dataset"}
copy_response = self._post(f"histories/{new_history_id}/contents", data=copy_payload)
copy_response = self._post(f"histories/{new_history_id}/contents", data=copy_payload, json=True)
self._assert_status_code_is(copy_response, 200)
inputs = json.dumps({
'input1': {'src': 'hda', 'id': dataset_id}
Expand Down Expand Up @@ -661,7 +661,7 @@ def test_search_with_hdca_pair_input(self, history_id):
# We test that a job can be found even if the collection has been copied to another history
new_history_id = self.dataset_populator.new_history()
copy_payload = {"content": list_id_a, "source": "hdca", "type": "dataset_collection"}
copy_response = self._post(f"histories/{new_history_id}/contents", data=copy_payload)
copy_response = self._post(f"histories/{new_history_id}/contents", data=copy_payload, json=True)
self._assert_status_code_is(copy_response, 200)
new_list_a = copy_response.json()['id']
copied_inputs = json.dumps({
Expand Down
12 changes: 6 additions & 6 deletions lib/galaxy_test/api/test_libraries.py
Original file line number Diff line number Diff line change
Expand Up @@ -424,17 +424,17 @@ def test_import_paired_collection(self):
'name': collection_name,
'collection_type': 'list:paired',
"type": "dataset_collection",
'element_identifiers': json.dumps([
'element_identifiers': [
{
'src': 'new_collection',
'name': 'pair1',
'collection_type': 'paired',
'element_identifiers': [{'name': 'forward', 'src': 'ldda', 'id': ld['id']},
{'name': 'reverse', 'src': 'ldda', 'id': ld['id']}]
}
])
]
}
new_collection = self._post(url, payload).json()
new_collection = self._post(url, payload, json=True).json()
assert new_collection['name'] == collection_name
pair = new_collection['elements'][0]
assert pair['element_identifier'] == 'pair1'
Expand All @@ -453,14 +453,14 @@ def _import_to_history(self, visible=True):
"history_id": history_id,
"name": collection_name,
"hide_source_items": not visible,
"element_identifiers": json.dumps([{
"element_identifiers": [{
"id": ld['id'],
"name": element_identifer,
"src": "ldda"}]),
"src": "ldda"}],
"type": "dataset_collection",
"elements": []
}
new_collection = self._post(url, payload).json()
new_collection = self._post(url, payload, json=True).json()
assert new_collection['name'] == collection_name
assert new_collection['element_count'] == 1
element = new_collection['elements'][0]
Expand Down
4 changes: 2 additions & 2 deletions lib/galaxy_test/api/test_workflow_extraction.py
Original file line number Diff line number Diff line change
Expand Up @@ -454,14 +454,14 @@ def __copy_content_to_history(self, history_id, content):
source="hda",
content=content["id"]
)
response = self._post(f"histories/{history_id}/contents/datasets", payload)
response = self._post(f"histories/{history_id}/contents/datasets", payload, json=True)

else:
payload = dict(
source="hdca",
content=content["id"]
)
response = self._post(f"histories/{history_id}/contents/dataset_collections", payload)
response = self._post(f"histories/{history_id}/contents/dataset_collections", payload, json=True)
self._assert_status_code_is(response, 200)
return response.json()

Expand Down
4 changes: 2 additions & 2 deletions lib/galaxy_test/api/test_workflows.py
Original file line number Diff line number Diff line change
Expand Up @@ -2887,7 +2887,7 @@ def test_workflow_rerun_with_use_cached_job(self):
new_ds_map = json.loads(new_workflow_request['ds_map'])
for key, input_values in invocation_1['inputs'].items():
copy_payload = {"content": input_values['id'], "source": "hda", "type": "dataset"}
copy_response = self._post(f"histories/{history_id_two}/contents", data=copy_payload).json()
copy_response = self._post(f"histories/{history_id_two}/contents", data=copy_payload, json=True).json()
new_ds_map[key]['id'] = copy_response['id']
new_workflow_request['ds_map'] = json.dumps(new_ds_map, sort_keys=True)
new_workflow_request['history'] = f"hist_id={history_id_two}"
Expand Down Expand Up @@ -2924,7 +2924,7 @@ def test_nested_workflow_rerun_with_use_cached_job(self):
dataset_type = inputs['outer_input']['src']
dataset_id = inputs['outer_input']['id']
copy_payload = {"content": dataset_id, "source": dataset_type, "type": "dataset"}
copy_response = self._post(f"histories/{history_id_two}/contents", data=copy_payload)
copy_response = self._post(f"histories/{history_id_two}/contents", data=copy_payload, json=True)
self._assert_status_code_is(copy_response, 200)
new_dataset_id = copy_response.json()['id']
inputs['outer_input']['id'] = new_dataset_id
Expand Down

0 comments on commit ffac0c4

Please sign in to comment.