Skip to content

[Test Stability] PR to Test BinaryClassifierSymSgdTest #4722

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
Changes from all commits
Commits
Show all changes
30 commits
Select commit Hold shift + click to select a range
9d0d18c
Enable running only specific tests
antoniovs1029 Jan 29, 2020
71c690a
Enable on Mac
antoniovs1029 Jan 29, 2020
fa57c31
Added RunSpecific
antoniovs1029 Jan 29, 2020
d2131bd
Removed x64Fact
antoniovs1029 Jan 29, 2020
42c73e8
Added Theory for a 100 iterations
antoniovs1029 Jan 29, 2020
ed1b679
Enable Conditional Numerical Reproducibility for tests
sharwell Dec 12, 2019
134f67b
Update baselines for failing tests
sharwell Dec 12, 2019
383b062
Skip BinaryClassifierSymSgdTest on Linux due to numerical differences
sharwell Jan 16, 2020
83e7dce
Updated to master
antoniovs1029 Jan 31, 2020
b11218b
Updated to PR 4569
antoniovs1029 Jan 31, 2020
8cbee2b
runSpecific to true
antoniovs1029 Jan 31, 2020
a280811
disable innerLoop
antoniovs1029 Feb 1, 2020
0dd4cb0
runSpecific to false
antoniovs1029 Feb 1, 2020
00e071a
Disable on Mac and Linux
antoniovs1029 Feb 3, 2020
86fefd7
Revert "Disable on Mac and Linux"
antoniovs1029 Feb 4, 2020
8810fb5
Only one iteration and fail fast to get the baselines
antoniovs1029 Feb 4, 2020
a8876fe
Revert "runSpecific to false"
antoniovs1029 Feb 4, 2020
823f996
runSpecific true 2
antoniovs1029 Feb 4, 2020
6cbc207
Revert "Revert "runSpecific to false""
antoniovs1029 Feb 4, 2020
0af804d
innerLoop = true
antoniovs1029 Feb 4, 2020
4e42864
Merge remote-tracking branch 'upstream/master' into tst02BinaryClassi…
antoniovs1029 Feb 4, 2020
56e5d1b
Added baselines for each platform
antoniovs1029 Feb 4, 2020
2900510
run 100 iterations
antoniovs1029 Feb 4, 2020
305a4f4
Changes Is64OperatingSystem to Is64BitProcess to check if it's window…
antoniovs1029 Feb 4, 2020
e829096
Run all tests (including disabled ones)
antoniovs1029 Feb 4, 2020
a028997
Updated to masters
antoniovs1029 Feb 5, 2020
6e0a64c
Revert "Run all tests (including disabled ones)"
antoniovs1029 Feb 6, 2020
6de4273
Disable CNR to get baselines
antoniovs1029 Feb 6, 2020
9524464
Return to windows 64 original baselines
antoniovs1029 Feb 6, 2020
54e4ecc
Use linux baseline for windows 32 and windows 64 to previous baseline
antoniovs1029 Feb 6, 2020
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
8 changes: 4 additions & 4 deletions config.json
Original file line number Diff line number Diff line change
@@ -92,14 +92,14 @@
},
"RunNightlyBuildTests": {
"description": "MsBuild target that run the nightly build tests. Call this after building.",
"valueType": "target",
"values": [],
"defaultValue": ""
"valueType": "target",
"values": [],
"defaultValue": ""
},
"Coverage": {
"description": "Turn on code coverge.",
"valueType": "property",
"values": ["false", "true"],
"values": [ "false", "true" ],
"defaultValue": "false"
},
"CleanAllProjects": {
24 changes: 12 additions & 12 deletions test/BaselineOutput/Common/SymSGD/SymSGD-CV-breast-cancer-out.txt
Original file line number Diff line number Diff line change
@@ -2,7 +2,7 @@ maml.exe CV tr=SymSGD{nt=1} threads=- norm=No dout=%Output% data=%Data% seed=1
Not adding a normalizer.
Data fully loaded into memory.
Initial learning rate is tuned to 100.000000
Bias: -467.9297, Weights: [5.415065,76.39395,22.35155,-11.98839,-28.26446,44.58415,22.72012,11.13254,2.851256]
Bias: -468.3528, Weights: [4.515409,75.74901,22.2914,-10.50209,-28.58107,44.81024,23.8734,13.20304,2.448269]
Not training a calibrator because it is not needed.
Not adding a normalizer.
Data fully loaded into memory.
@@ -15,15 +15,15 @@ Confusion table
||======================
PREDICTED || positive | negative | Recall
TRUTH ||======================
positive || 130 | 4 | 0.9701
positive || 132 | 2 | 0.9851
negative || 8 | 212 | 0.9636
||======================
Precision || 0.9420 | 0.9815 |
OVERALL 0/1 ACCURACY: 0.966102
Precision || 0.9429 | 0.9907 |
OVERALL 0/1 ACCURACY: 0.971751
LOG LOSS/instance: Infinity
Test-set entropy (prior Log-Loss/instance): 0.956998
LOG-LOSS REDUCTION (RIG): -Infinity
AUC: 0.990706
AUC: 0.991045
Warning: The predictor produced non-finite prediction values on 8 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable.
TEST POSITIVE RATIO: 0.3191 (105.0/(105.0+224.0))
Confusion table
@@ -42,16 +42,16 @@ AUC: 0.963435

OVERALL RESULTS
---------------------------------------
AUC: 0.977070 (0.0136)
Accuracy: 0.952656 (0.0134)
Positive precision: 0.919613 (0.0224)
Positive recall: 0.942217 (0.0279)
Negative precision: 0.970470 (0.0110)
AUC: 0.977240 (0.0138)
Accuracy: 0.955481 (0.0163)
Positive precision: 0.920027 (0.0228)
Positive recall: 0.949680 (0.0354)
Negative precision: 0.975057 (0.0156)
Negative recall: 0.957265 (0.0064)
Log-loss: Infinity (NaN)
Log-loss reduction: -Infinity (NaN)
F1 Score: 0.930771 (0.0251)
AUPRC: 0.964018 (0.0164)
F1 Score: 0.934582 (0.0289)
AUPRC: 0.964431 (0.0168)

---------------------------------------
Physical memory usage(MB): %Number%
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
SymSGD
AUC Accuracy Positive precision Positive recall Negative precision Negative recall Log-loss Log-loss reduction F1 Score AUPRC /nt Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings
0.97707 0.952656 0.919613 0.942217 0.97047 0.957265 Infinity -Infinity 0.930771 0.964018 1 SymSGD %Data% %Output% 99 0 0 maml.exe CV tr=SymSGD{nt=1} threads=- norm=No dout=%Output% data=%Data% seed=1 /nt:1
0.97724 0.955481 0.920027 0.94968 0.975057 0.957265 Infinity -Infinity 0.934582 0.964431 1 SymSGD %Data% %Output% 99 0 0 maml.exe CV tr=SymSGD{nt=1} threads=- norm=No dout=%Output% data=%Data% seed=1 /nt:1

Original file line number Diff line number Diff line change
@@ -0,0 +1,69 @@
maml.exe CV tr=SymSGD{nt=1} threads=- norm=No dout=%Output% data=%Data% seed=1
Not adding a normalizer.
Data fully loaded into memory.
Initial learning rate is tuned to 100.000000
Bias: -465.74258, Weights: [10.088505,67.45707,18.099556,-11.573266,-32.295338,52.25842,17.802341,14.058811,-2.0404801]
Not training a calibrator because it is not needed.
Not adding a normalizer.
Data fully loaded into memory.
Initial learning rate is tuned to 100.000000
Bias: -484.28625, Weights: [-12.7873125,140.42897,121.93798,37.52739,-129.81442,70.90551,-89.37083,81.64296,-32.327835]
Not training a calibrator because it is not needed.
Warning: The predictor produced non-finite prediction values on 8 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable.
TEST POSITIVE RATIO: 0.3785 (134.0/(134.0+220.0))
Confusion table
||======================
PREDICTED || positive | negative | Recall
TRUTH ||======================
positive || 130 | 4 | 0.9701
negative || 7 | 213 | 0.9682
||======================
Precision || 0.9489 | 0.9816 |
OVERALL 0/1 ACCURACY: 0.968927
LOG LOSS/instance: Infinity
Test-set entropy (prior Log-Loss/instance): 0.956998
LOG-LOSS REDUCTION (RIG): -Infinity
AUC: 0.990095
Warning: The predictor produced non-finite prediction values on 8 instances during testing. Possible causes: abnormal data or the predictor is numerically unstable.
TEST POSITIVE RATIO: 0.3191 (105.0/(105.0+224.0))
Confusion table
||======================
PREDICTED || positive | negative | Recall
TRUTH ||======================
positive || 96 | 9 | 0.9143
negative || 11 | 213 | 0.9509
||======================
Precision || 0.8972 | 0.9595 |
OVERALL 0/1 ACCURACY: 0.939210
LOG LOSS/instance: Infinity
Test-set entropy (prior Log-Loss/instance): 0.903454
LOG-LOSS REDUCTION (RIG): -Infinity
AUC: 0.963435

OVERALL RESULTS
---------------------------------------
AUC: 0.976765 (0.0133)
Accuracy: 0.954068 (0.0149)
Positive precision: 0.923051 (0.0259)
Positive recall: 0.942217 (0.0279)
Negative precision: 0.970513 (0.0111)
Negative recall: 0.959537 (0.0086)
Log-loss: Infinity (NaN)
Log-loss reduction: -Infinity (NaN)
F1 Score: 0.932535 (0.0269)
AUPRC: 0.963544 (0.0159)

---------------------------------------
Physical memory usage(MB): %Number%
Virtual memory usage(MB): %Number%
%DateTime% Time elapsed(s): %Number%

--- Progress log ---
[1] 'Preprocessing' started.
[1] 'Preprocessing' finished in %Time%.
[2] 'Training' started.
[2] 'Training' finished in %Time%.
[3] 'Preprocessing #2' started.
[3] 'Preprocessing #2' finished in %Time%.
[4] 'Training #2' started.
[4] 'Training #2' finished in %Time%.
Loading