-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is KAN (pykan) sufficient for classification tasks? #477
Comments
@SaranDS Here code with 1 output: `from kan import KAN model = KAN(width=[16, 3, 1], grid=3, k=3) def train_acc(): def test_acc(): start_time = time.time() print(custom_dataset['train_input'].dtype) results = model.fit(custom_dataset, opt="LBFGS", steps=100, results['train_acc'][-1], results['test_acc'][-1]` Code with 2 output: `from kan import KAN model = KAN(width=[16, 5, 3, 2], grid=5, k=3) start_time = time.time() results = model.fit(custom_dataset, opt="LBFGS", steps=100, |
@YuriyKabanenko The results obtained for this dataset [not only this dataset, I used 3 other different dataset related to same domain, but each dataset comprise different features and datapoints] are exceptionally good, which raises concerns about the possibility of test set leakage during training. To verify this, I included print statements to check the sizes of the training, validation, and test sets. The output confirmed that the dataset was split correctly according to the specified sizes. However, I am unsure of any further methods to validate these results. |
@KindXiaoming |
The implemented following code snippet for binary classification on tabular data, using stratified K-fold cross-validation (K=10). The performance results seem exceptionally good. Can someone help review and suggest improvements to the implementation?
`model = KAN(width=[38,5,3, 2], grid=5, k=3)
for train_idx, test_idx in (kf.split(X_scaled, y)):
`
Dataset Description :
Features - 39
data points - 16,900 (after SMOTE - 32,900)
The text was updated successfully, but these errors were encountered: