Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[ENH] Assessing Performance #134

Open
samihamdan opened this issue May 7, 2021 · 5 comments
Open

[ENH] Assessing Performance #134

samihamdan opened this issue May 7, 2021 · 5 comments
Assignees
Labels
enhancement New feature or request needs thinking We need more discussion around this topic

Comments

@samihamdan
Copy link
Collaborator

Problem
I noticed that sometimes performance of julearn seems to be not that great. I am not sure whether this is a real trend or just normal frustration with the speed of ml.

Solution
I am not sure whether this is actually a problem, but it would be nice to in general assess performance to keep and eye on how much overhead we add to sklearn. Even if we do not change the speed it is good to make realistic expectations of potential users.

Considerations
How does it change with more data or more transformers.
It could be that each transformation from np.array to pd.DataFrame has a big impact. On the other hand the implementation of confound removal could also be the reason for long computation times in real world observations.

Screenshot
I did one very simple observation with only one transformer.
If I use 4x of the data I still have a similar 3x worse performance of julearn.

image

@samihamdan samihamdan added the enhancement New feature or request label May 7, 2021
@fraimondo
Copy link
Contributor

fraimondo commented May 7, 2021 via email

@samihamdan
Copy link
Collaborator Author

For this simplistic example no difference between both julearn APIs.
But good point for the general assessment of speed.

@fraimondo fraimondo added the needs thinking We need more discussion around this topic label Jul 21, 2022
@fraimondo
Copy link
Contributor

What about adding some benchmark tests to compare if a PR makes a huge mess with performance? Can it be done using CI? Maybe @synchon can help with this one.

@synchon
Copy link
Member

synchon commented Jul 21, 2022

I'll take a look.

@fraimondo
Copy link
Contributor

@synchon will take a look at it soon

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request needs thinking We need more discussion around this topic
Projects
None yet
Development

No branches or pull requests

3 participants