Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why does it take so long before the training really starts? #75

Open
jannichorst opened this issue Jan 11, 2024 · 2 comments
Open

Why does it take so long before the training really starts? #75

jannichorst opened this issue Jan 11, 2024 · 2 comments

Comments

@jannichorst
Copy link

Using the run.py with standard parameters It takes roughly 20-30min before my GPU starts working, before that only 4 CPU cores have constant spikes. And judging by the timing of the logs It seems that a lot processing is done before the actual training starts... Could you provide any insight into what is happening and if there is something that can be done to optimize? I run this on a 12C/24T CPU @4,7 GhZ and an RTX4090 (cuda enabled, and it starts using the GPU when the training starts).

@tianzhou2011
Copy link
Collaborator

tianzhou2011 commented Jan 12, 2024 via email

@jannichorst
Copy link
Author

jannichorst commented Jan 13, 2024

I'll test it and get back to you! Already thanks for the answer.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants