98
98
# Work around https://github.com/pypa/pip/issues/9542
99
99
- script : ' pip install -U numpy~=1.21.0'
100
100
displayName : ' Upgrade numpy'
101
-
102
- - script : ' pip install pytest pytest-runner jupyter jupyter-client nbconvert nbformat seaborn xgboost tqdm && python setup.py pytest'
101
+
102
+ - script : ' pip install pytest pytest-runner jupyter jupyter-client nbconvert nbformat seaborn xgboost tqdm && pip list && python setup.py pytest'
103
103
displayName : ' Unit tests'
104
104
env :
105
105
PYTEST_ADDOPTS : ' -m "notebook"'
@@ -126,12 +126,6 @@ jobs:
126
126
# Work around https://github.com/pypa/pip/issues/9542
127
127
- script : ' pip install -U numpy~=1.21.0'
128
128
displayName : ' Upgrade numpy'
129
-
130
- # shap 0.39 and sklearn 1.0 interact badly in these notebooks
131
- # shap 0.40 has a bug in waterfall (https://github.com/slundberg/shap/issues/2283) that breaks our main tests
132
- # but fixes the interaction here...
133
- - script : ' pip install -U shap~=0.40.0'
134
- displayName : ' Upgrade shap'
135
129
136
130
- script : ' pip install pytest pytest-runner jupyter jupyter-client nbconvert nbformat seaborn xgboost tqdm && python setup.py pytest'
137
131
displayName : ' Unit tests'
@@ -207,7 +201,7 @@ jobs:
207
201
- script : ' pip install pytest pytest-runner && python setup.py pytest'
208
202
displayName : ' Unit tests'
209
203
env :
210
- PYTEST_ADDOPTS : ' -m "not (notebook or automl or dml or causal )" -n 2'
204
+ PYTEST_ADDOPTS : ' -m "not (notebook or automl or dml or serial or cate_api )" -n 2'
211
205
COVERAGE_PROCESS_START : ' setup.cfg'
212
206
- task : PublishTestResults@2
213
207
displayName : ' Publish Test Results **/test-results.xml'
@@ -253,15 +247,44 @@ jobs:
253
247
parameters :
254
248
package : ' -e .[tf,plt]'
255
249
job :
256
- job : Tests_causal
250
+ job : Tests_serial
257
251
dependsOn : ' EvalChanges'
258
252
condition : eq(dependencies.EvalChanges.outputs['output.testCode'], 'True')
259
- displayName : ' Run tests (Causal )'
253
+ displayName : ' Run tests (Serial )'
260
254
steps :
261
255
- script : ' pip install pytest pytest-runner && python setup.py pytest'
262
256
displayName : ' Unit tests'
263
257
env :
264
- PYTEST_ADDOPTS : ' -m "causal" -n 1'
258
+ PYTEST_ADDOPTS : ' -m "serial" -n 1'
259
+ COVERAGE_PROCESS_START : ' setup.cfg'
260
+ - task : PublishTestResults@2
261
+ displayName : ' Publish Test Results **/test-results.xml'
262
+ inputs :
263
+ testResultsFiles : ' **/test-results.xml'
264
+ testRunTitle : ' Python $(python.version), image $(imageName)'
265
+ condition : succeededOrFailed()
266
+
267
+ - task : PublishCodeCoverageResults@1
268
+ displayName : ' Publish Code Coverage Results'
269
+ inputs :
270
+ codeCoverageTool : Cobertura
271
+ summaryFileLocation : ' $(System.DefaultWorkingDirectory)/**/coverage.xml'
272
+
273
+ - template : azure-pipelines-steps.yml
274
+ parameters :
275
+ package : ' -e .[tf,plt]'
276
+ job :
277
+ job : Tests_CATE_API
278
+ dependsOn : ' EvalChanges'
279
+ condition : eq(dependencies.EvalChanges.outputs['output.testCode'], 'True')
280
+ displayName : ' Run tests (Other)'
281
+ steps :
282
+ - script : ' pip install pytest pytest-runner'
283
+ displayName : ' Install pytest'
284
+ - script : ' python setup.py pytest'
285
+ displayName : ' CATE Unit tests'
286
+ env :
287
+ PYTEST_ADDOPTS : ' -m "cate_api" -n auto'
265
288
COVERAGE_PROCESS_START : ' setup.cfg'
266
289
- task : PublishTestResults@2
267
290
displayName : ' Publish Test Results **/test-results.xml'
0 commit comments