Newer
Older
# can be run on vagrant like this:
# vagrant ssh obisserver -c 'cd /vagrant_python/OBis/integration_tests && pytest ./integration_tests.py'
import subprocess
from contextlib import contextmanager
from pybis import Openbis
# 0. pybis login
o = Openbis('https://localhost:8443', verify_certificates=False)
o.login('admin', 'admin', save_token=True)
# 1. Global configuration
result = run('./01_global_config.sh', tmpdir)
config = json.loads(run('./00_get_config_global.sh'))
assert config['openbis_url'] == 'https://localhost:8443'
assert config['user'] == 'admin'
assert config['data_set_type'] == 'UNKNOWN'
assert config['verify_certificates'] == False
# 2. First commit
result = run('./02_first_commit_1_create_repository.sh', tmpdir)
assert '?? .obis/config.json' in result
assert '?? file' in result
result = run('./02_first_commit_2_commit.sh', tmpdir)
config = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data1'))
assert config['external_dms_id'].startswith('ADMIN-' + socket.gethostname().upper())
assert len(config['repository_id']) == 36
assert "Created data set {}.".format(config['data_set_id']) in result
data_set = o.get_dataset(config['data_set_id']).data
assert_matching(config, data_set, tmpdir, 'obis_data/data1')
# 3. Second commit
config_before = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data1'))
result = run('./03_second_commit_1_commit.sh', tmpdir)
config = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data1'))
assert config['data_set_id'] != config_before['data_set_id']
assert config['external_dms_id'].startswith('ADMIN-' + socket.gethostname().upper())
assert config['external_dms_id'] == config_before['external_dms_id']
assert config['repository_id'] == config_before['repository_id']
assert "Created data set {}.".format(config['data_set_id']) in result
result = run('./03_second_commit_2_git_annex_info.sh', tmpdir)
assert 'file: big_file' in result
assert 'key: SHA256E-s1000000--d29751f2649b32ff572b5e0a9f541ea660a50f94ff0beedfb0b692b924cc8025' in result
assert 'present: true' in result
data_set = o.get_dataset(config['data_set_id']).data
assert_matching(config, data_set, tmpdir, 'obis_data/data1')
assert data_set['parents'][0]['code'] == config_before['data_set_id']
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
# 4. Second repository
result = run('./04_second_repository.sh', tmpdir)
config_data1 = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data1'))
config = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data2'))
assert config['external_dms_id'].startswith('ADMIN-' + socket.gethostname().upper())
assert config['external_dms_id'] == config_data1['external_dms_id']
assert len(config['repository_id']) == 36
assert config['repository_id'] != config_data1['repository_id']
assert "Created data set {}.".format(config['data_set_id']) in result
data_set = o.get_dataset(config['data_set_id']).data
assert_matching(config, data_set, tmpdir, 'obis_data/data2')
# 5. Second external dms
result = run('./05_second_external_dms.sh', tmpdir)
config_data1 = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data1'))
config = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data_b/data3'))
assert config['external_dms_id'].startswith('ADMIN-' + socket.gethostname().upper())
assert config['external_dms_id'] != config_data1['external_dms_id']
assert len(config['repository_id']) == 36
assert config['repository_id'] != config_data1['repository_id']
assert "Created data set {}.".format(config['data_set_id']) in result
data_set = o.get_dataset(config['data_set_id']).data
assert_matching(config, data_set, tmpdir, 'obis_data_b/data3')
# 6. Error on first commit
result = run('./06_error_on_first_commit_1_error.sh', tmpdir)
assert 'Missing configuration settings for [\'object_id\', \'collection_id\'].' in result
result = run('./06_error_on_first_commit_2_status.sh', tmpdir)
assert '?? file' in result
result = run('./06_error_on_first_commit_3_commit.sh', tmpdir)
config = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data4'))
assert "Created data set {}.".format(config['data_set_id']) in result
data_set = o.get_dataset(config['data_set_id']).data
assert_matching(config, data_set, tmpdir, 'obis_data/data4')
# 7. Attach data set to a collection
result = run('./07_attach_to_collection.sh', tmpdir)
config = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data5'))
assert config['external_dms_id'].startswith('ADMIN-' + socket.gethostname().upper())
assert len(config['repository_id']) == 36
assert "Created data set {}.".format(config['data_set_id']) in result
data_set = o.get_dataset(config['data_set_id']).data
assert_matching(config, data_set, tmpdir, 'obis_data/data5')
# 8. Addref
result = run('./08_addref_1_success.sh', tmpdir)
config_data1 = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data1'))
config_data6 = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data6'))
assert config_data6 == config_data1
result = run('./08_addref_2_duplicate.sh', tmpdir)
assert 'DataSet already exists in the database' in result
result = run('./08_addref_3_non-existent.sh', tmpdir)
assert 'Invalid value' in result
data_set = o.get_dataset(config_data6['data_set_id']).data
assert_matching(config_data6, data_set, tmpdir, 'obis_data/data6')
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
# 9. Local clone
config_data2 = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data2'))
result = run('./09_local_clone.sh', tmpdir, [config_data2['data_set_id']])
config_data2_clone = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data_b/data2'))
assert config_data2_clone['external_dms_id'].startswith('ADMIN-' + socket.gethostname().upper())
assert config_data2_clone['external_dms_id'] != config_data2['external_dms_id']
data_set = o.get_dataset(config_data2_clone['data_set_id']).data
assert_matching(config_data2_clone, data_set, tmpdir, 'obis_data_b/data2')
del config_data2['external_dms_id']
del config_data2_clone['external_dms_id']
assert config_data2_clone == config_data2
# 11. Init analysis
result = run('./11_init_analysis_1_external.sh', tmpdir, [config_data2['data_set_id']])
config_data1 = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data1'))
config_analysis1 = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/analysis1'))
assert "Created data set {}.".format(config_analysis1['data_set_id']) in result
assert len(config_analysis1['repository_id']) == 36
assert config_analysis1['repository_id'] != config_data1['repository_id']
assert config_analysis1['data_set_id'] != config_data1['data_set_id']
data_set = o.get_dataset(config_analysis1['data_set_id']).data
assert_matching(config_analysis1, data_set, tmpdir, 'obis_data/analysis1')
assert data_set['parents'][0]['code'] == config_data1['data_set_id']
result = run('./11_init_analysis_2_internal.sh', tmpdir)
config_analysis2 = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data1/analysis2'))
assert "Created data set {}.".format(config_analysis2['data_set_id']) in result
assert len(config_analysis2['repository_id']) == 36
assert config_analysis2['repository_id'] != config_data1['repository_id']
assert config_analysis2['data_set_id'] != config_data1['data_set_id']
result = run('./11_init_analysis_3_git_check_ignore.sh', tmpdir)
assert 'analysis2' in result
data_set = o.get_dataset(config_analysis2['data_set_id']).data
assert_matching(config_analysis2, data_set, tmpdir, 'obis_data/data1/analysis2')
assert data_set['parents'][0]['code'] == config_data1['data_set_id']
# 12. Metadata only commit
result = run('./12_metadata_only_1_commit.sh', tmpdir)
config = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data7'))
assert "Created data set {}.".format(config['data_set_id']) in result
data_set = o.get_dataset(config['data_set_id']).data
assert_matching(config, data_set, tmpdir, 'obis_data/data7')
result = run('./12_metadata_only_2_metadata_commit.sh', tmpdir)
config = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data7'))
assert "Created data set {}.".format(config['data_set_id']) in result
data_set = o.get_dataset(config['data_set_id']).data
assert_matching(config, data_set, tmpdir, 'obis_data/data7')
# 13. obis sync
result = run('./13_sync_1_git_commit_and_sync.sh', tmpdir)
config = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data7'))
assert "Created data set {}.".format(config['data_set_id']) in result
data_set = o.get_dataset(config['data_set_id']).data
assert_matching(config, data_set, tmpdir, 'obis_data/data7')
result = run('./13_sync_2_only_sync.sh', tmpdir)
assert 'Nothing to sync' in result
# 14. Configure data set properties
result = run('./14_config_data_set_properties_1.sh', tmpdir)
config = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data8'))
assert config['data_set_properties'] == { 'A': '0' }
result = run('./14_config_data_set_properties_2.sh', tmpdir)
config = json.loads(run('./00_get_config.sh', tmpdir + '/obis_data/data8'))
assert config['data_set_properties'] == { 'A': '0', 'B': '1', 'C': '3' }
result = run('./14_config_data_set_properties_3.sh', tmpdir)
assert 'Duplicate key after capitalizing JSON config: A' in result
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
# 15. Removeref
with cd(tmpdir + '/obis_data'):
config = get_config('data6')
content_copies = get_data_set(o, config)['linkedData']['contentCopies']
assert len(content_copies) == 2
cmd('obis removeref data6')
content_copies = get_data_set(o, config)['linkedData']['contentCopies']
assert len(content_copies) == 1
assert content_copies[0]['path'].endswith('data1')
cmd('obis addref data6')
cmd('obis removeref data1')
content_copies = get_data_set(o, config)['linkedData']['contentCopies']
assert len(content_copies) == 1
assert content_copies[0]['path'].endswith('data6')
result = cmd('obis removeref data1')
assert 'Matching content copy not fount in data set' in result
cmd('obis addref data1')
def get_config(repository_folder):
with cd(repository_folder):
return json.loads(cmd('obis config'))
def get_data_set(o, config):
return o.get_dataset(config['data_set_id']).data
@contextmanager
def cd(newdir):
"""Safe cd -- return to original dir after execution, even if an exception is raised."""
prevdir = os.getcwd()
os.chdir(os.path.expanduser(newdir))
try:
yield
finally:
os.chdir(prevdir)
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
def run(cmd, tmpdir="", params=[]):
completed_process = subprocess.run([cmd, tmpdir] + params, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
return get_cmd_result(completed_process, tmpdir)
def cmd(cmd):
cmd_split = cmd.split(' ')
completed_process = subprocess.run(cmd_split, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
return get_cmd_result(completed_process)
def get_cmd_result(completed_process, tmpdir=''):
result = ''
if completed_process.stderr:
result += completed_process.stderr.decode('utf-8').strip()
if completed_process.stdout:
result += completed_process.stdout.decode('utf-8').strip()
return result
def assert_matching(config, data_set, tmpdir, path):
content_copies = data_set['linkedData']['contentCopies']
content_copy = list(filter(lambda cc: cc['path'].endswith(path) == 1, content_copies))[0]
assert data_set['type']['code'] == config['data_set_type']
assert content_copy['externalDms']['code'] == config['external_dms_id']
assert content_copy['gitCommitHash'] == run('./00_get_commit_hash.sh', str(tmpdir) + '/' + path)
assert content_copy['gitRepositoryId'] == config['repository_id']
if config['object_id'] is not None:
assert data_set['sample']['identifier']['identifier'] == config['object_id']
if config['collection_id'] is not None:
assert data_set['experiment']['identifier']['identifier'] == config['collection_id']