-
Notifications
You must be signed in to change notification settings - Fork 1.9k
LightGBMMulticlass doesn't normalize its Score column #4450
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
With 100k rows, you are going to be in TrainTest mode instead of CV. My guess for the non-normalized
|
When I run the sample code in LightGBM.cs I do see that the scores are being normalized. Let me take a look at your sample repro and get back to you. |
I downloaded your project file and tried to build it and ran into the following error:
Running nuget restore isn't helping. |
After messing with the nuget package sources, I was able to build and run the project. But I can't reproduce the issue. This is my output:
I see that my run produces only 1 model and that the out normalizes to 1. Do you get the same results if you download the code from github and run? |
@harishsk: You'll need additional runtime. As you note, only one model was tested. With longer runtime, LightGBM will have a chance to run and be selected as the winning model. Try changing Program.cs line 53 to from 5s to 60s: With the longer runtime, I see LightGBM chosen as the winning model; and the predicted |
When I change the run time to 60 seconds I get 5 models, with the best model as FastTreeOva and normalized output. But if I change the run time to 30 seconds, I get the output similar to yours. |
@LittleLittleCloud Thank you for raising this issue. |
@harishsk: Wow. That had to be difficult to locate the source of the issue. Thanks. |
Related: #3647 ? |
It is very likely that this PR fixes that issue as well. Let me verify tomorrow and report back. |
#4051 could also be related here as well. |
Thanks @harishsk for the fast fix. |
when I try AutoML v0.16.0 multi-classification, sometimes I get prediction score <0, and the summation of all scores doesn't equal to 1. which is unexpected. Maybe some trainer's output doesn't normalize.
I publish a mini-reproducible project in this repo for facilitating debug.
The text was updated successfully, but these errors were encountered: