Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Decrease nighlty tests upper limitation to 7 [skip ci] #4701

Merged

Conversation

pxLi
Copy link
Collaborator

@pxLi pxLi commented Feb 7, 2022

Signed-off-by: Peixin Li pxli@nyu.edu

fix #4675, 8 as the upper bound of parallelism is not good any more. Decrease it to 7 to save around 40-60 mins

@pxLi pxLi added bug Something isn't working build Related to CI / CD or cleanly building labels Feb 7, 2022
@pxLi pxLi changed the base branch from branch-22.04 to branch-22.02 February 7, 2022 02:17
Signed-off-by: Peixin Li <pxli@nyu.edu>
@pxLi pxLi force-pushed the decrease-nightly-parallelism-upper-limit-7 branch from 3a1aa1b to ecbd8cc Compare February 7, 2022 02:18
@pxLi pxLi changed the title [BUG] Decrease nighlty tests upper limitation to 7 [BUG] Decrease nighlty tests upper limitation to 7 [skip ci] Feb 7, 2022
@pxLi
Copy link
Collaborator Author

pxLi commented Feb 7, 2022

build

@pxLi pxLi merged commit 403c0f1 into NVIDIA:branch-22.02 Feb 7, 2022
@pxLi pxLi changed the title [BUG] Decrease nighlty tests upper limitation to 7 [skip ci] Decrease nighlty tests upper limitation to 7 [skip ci] Feb 7, 2022
@tgravescs
Copy link
Collaborator

I'm a bit unclear by the description here, decreasing the parallelism makes it faster why? We are allowing more gpu memory and not spilling as much?

@pxLi
Copy link
Collaborator Author

pxLi commented Feb 9, 2022

I'm a bit unclear by the description here, decreasing the parallelism makes it faster why? We are allowing more gpu memory and not spilling as much?

Per the test results, its not always the larger parallelism, the faster total tests run. Even we have more gpu mem to split tests, the parallel run could much slower when too many cases were sharing the cuda cores (competing the SM? context switching? sry, I am not a cuda expert)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working build Related to CI / CD or cleanly building
Projects
None yet
Development

Successfully merging this pull request may close these issues.

[BUG] Jenkins integration build timed out at 10 hours
3 participants