Giter Club home page Giter Club logo

Comments (14)

hkchengrex avatar hkchengrex commented on June 2, 2024

Can your dataloaders catch up? i.e. are the GPUs at (almost) full load all the time?
The reported training time is very rough (we used a mix of hardware at different times). We will re-train again and give a better estimate in the next revision of the paper.

In any case, it should take much less than 100h for s0 even with 2x 1080Ti. The most probable reason is dataloader bottleneck.

from stcn.

PinxueGuo avatar PinxueGuo commented on June 2, 2024

I think it's highly likely, cause my GPU sometimes far lower than 100%.
Can you give me some suggestions to solve it? Should I change OMP_NUM_THREADS (in command) or num_works (pytorch dataloader) bigger?
Thank you!

from stcn.

PinxueGuo avatar PinxueGuo commented on June 2, 2024

I find bigger num_works really speed things up in my case.

from stcn.

hkchengrex avatar hkchengrex commented on June 2, 2024

I think the general wisdom is to use higher OMP_NUM_THREADS and num_workers when you have more free CPU cores available.
That's great to hear.

from stcn.

PinxueGuo avatar PinxueGuo commented on June 2, 2024

OK. More num_works exactly helps, and OMP_NUM_THREADS=4 (1,8,16 will be slower even) as your original setting is the fastest in my case.
Thank you for your great work and quick reply !

from stcn.

hkchengrex avatar hkchengrex commented on June 2, 2024

BTW you can try adding the --benchmark flag.

from stcn.

PinxueGuo avatar PinxueGuo commented on June 2, 2024

Thank you, I tried it but not really effective.
And bigger num_workers only bring 10% speed improvement .
Could you tell me what's your time consuming in log "retrain_s0 - It ******* [TRAIN] [time ]: ?". In my case , time≈1.0+.

from stcn.

PinxueGuo avatar PinxueGuo commented on June 2, 2024

Sorry, it should be a 25% improvement about speed. (bigger num_workers=16, 1*3090, --nproc_per_node=1, bs=16).
log:retrain_s0 - It 51300 [TRAIN] [time ]: 1.0771173

from stcn.

hkchengrex avatar hkchengrex commented on June 2, 2024

With 1x 3090 I am getting around 0.7 for [time].
2x 2080Ti should be faster than 1x 3090.

from stcn.

hkchengrex avatar hkchengrex commented on June 2, 2024

Hmm, it's actually 0.7 around the start of training and stabilizes around 0.5.

from stcn.

PinxueGuo avatar PinxueGuo commented on June 2, 2024

I compared 22080ti with 13090, and result is 1×3090 is a little faster than 2×2080ti.
If the [time] is round 0.5, s0 need 45 hours right?
In my case, s3 exactly need 30 hours. So I wanna to confirm "Regular training without BL30K takes around 30 hours"(in paper), 30-hour is refer to s3 or s0+s3?

from stcn.

hkchengrex avatar hkchengrex commented on June 2, 2024

It refers to s0+s3. I guess hardware infrastructure affects the training speed a lot.

from stcn.

PinxueGuo avatar PinxueGuo commented on June 2, 2024

Ok. Thank you !

from stcn.

zhouweii234 avatar zhouweii234 commented on June 2, 2024

May I ask the training time for stage 0 after you use bigger num_workers? (num_workers=16, 1*3090, --nproc_per_node=1, bs=16). @BWYWTB

from stcn.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.