You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We are currently using version 0.50 in creating some classifier models but are seeing some strange behavior. We are currently setting the number of threads in our classifiers due to #217 and wanting to be able to control the CPU usage on the server.
So when using a classifier like so .. var algo = new StochasticDualCoordinateAscentClassifier() { Caching = CachingOptions.Disk, MaxIterations = 100, LossFunction = new SmoothedHingeLossSDCAClassificationLossFunction(), Shuffle = false, NumThreads = System.Environment.ProcessorCount - 1 //We use one less than the number of processors available, };
What we are noticing is that if we run this from a box with 4 cores on it then we get a decent model where the microaccuracy is above 90%. However, when we move this same code over to a larger server with 8 cores we are getting wildly different results. The microaccuracy drops down to around <60%.
Yikes!
Is there possibly something we are missing in the documentation that would address this?
The text was updated successfully, but these errors were encountered:
The datasets are around 15k entries and should be sufficiently sized.
Last night we ran tests constraining the number of CPUs from 7 - 1 … Anything under 4 comes out with the expected accuracy ….and anything greater goes to south. This was with running and rerunning with the same training data.
We are just trying to understand what the reason for this is, so that we can properly gauge things of scale such as how long it takes to train a model. When we are using 7 cores we get good throughput ~5 minutes training time but lousy results. When we use 4 we get slower ~30+ minutes training time but much better results.
We are currently using version 0.50 in creating some classifier models but are seeing some strange behavior. We are currently setting the number of threads in our classifiers due to #217 and wanting to be able to control the CPU usage on the server.
So when using a classifier like so ..
var algo = new StochasticDualCoordinateAscentClassifier() { Caching = CachingOptions.Disk, MaxIterations = 100, LossFunction = new SmoothedHingeLossSDCAClassificationLossFunction(), Shuffle = false, NumThreads = System.Environment.ProcessorCount - 1 //We use one less than the number of processors available, };
What we are noticing is that if we run this from a box with 4 cores on it then we get a decent model where the microaccuracy is above 90%. However, when we move this same code over to a larger server with 8 cores we are getting wildly different results. The microaccuracy drops down to around <60%.
Yikes!
Is there possibly something we are missing in the documentation that would address this?
The text was updated successfully, but these errors were encountered: