-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
lesser powerful gpu #140
Comments
how long will it take on h100? and have you tried on other cheaper GPUs? |
@xprabhudayal You can ask the agent to do more small scale experiments targeting fine-grained issues (e.g. activation functions, classical ML with HPO, easy NLP optimization) rather than reaching all the way into DL territory. When it comes to idea generation others have said it takes "a few hours" (2-4 hours) for more compute-heavy concepts. #145 @NathanaelTamirat Check is (currently it is made for data-centric experiments not physical research) #145 (comment) |
I have run this agent on Google Colab for free by using the T4 GPU and during the experiment part it has took me around 2 to 3 hours for Like the only part which utilizers the GPU is the experiment agent (when using LLM API) which has been seen in the diagram (in the middle section) |
@xprabhudayal so only the GPU is used for both writing code and experimentation? Could you share the setup and the time needed for each process (paper fetching, ideation, code-gen, experiment, drafting, reviewing)? |
ive been running this model on a high end instance(h100) for a while & i have some doubts..
The text was updated successfully, but these errors were encountered: