You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I would like to use different subset of features in each iteration. I'm aware that there is colsample_bytree but instead random features selection I would like to write custom Python code to select features.
Motivation
I hope that my LightGBM will get better generalization when trained on custom subsets of features.
Description
I was trying to use Booster class, with the following code, but it crashed:
[LightGBM] [Warning] Auto-choosing col-wise multi-threading, the overhead of testing was 0.199411 seconds.
You can set `force_col_wise=true` to remove the overhead.
[LightGBM] [Info] Total Bins 25500
[LightGBM] [Info] Number of data points in the train set: 1000, number of used features: 100
0
/home/piotr/sandbox/numerai_v4/v4/lib/python3.8/site-packages/lightgbm/basic.py:506: UserWarning: Usage of np.ndarray subset (sliced data) is not recommended due to it will double the peak memory cost in LightGBM.
_log_warning("Usage of np.ndarray subset (sliced data) is not recommended "
[LightGBM] [Info] Start training from score 0.505111
Segmentation fault (core dumped)
The text was updated successfully, but these errors were encountered:
Summary
I would like to use different subset of features in each iteration. I'm aware that there is
colsample_bytree
but instead random features selection I would like to write custom Python code to select features.Motivation
I hope that my LightGBM will get better generalization when trained on custom subsets of features.
Description
I was trying to use
Booster
class, with the following code, but it crashed:I got output:
The text was updated successfully, but these errors were encountered: