Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix verify build after integration tests are run #657

Merged
merged 4 commits into from
Sep 4, 2020

Conversation

revans2
Copy link
Collaborator

@revans2 revans2 commented Sep 3, 2020

Not perfect but tries to deal with fixing some issues with hive metastore files in the integration tests.

This fixes #656

Signed-off-by: Robert (Bobby) Evans <bobby@apache.org>
@revans2
Copy link
Collaborator Author

revans2 commented Sep 3, 2020

build

@jlowe jlowe added bug Something isn't working test Only impacts tests labels Sep 3, 2020
@jlowe jlowe added this to the Aug 31 - Sep 11 milestone Sep 3, 2020
@jlowe
Copy link
Member

jlowe commented Sep 3, 2020

Removing these directories would be a nasty surprised to anyone who ran spark from the top of the plugin repo. Granted may not be a common thing, but still destruction of data the tests didn't necessarily create.

It would be preferable to have the integration tests run from a directory somewhere underneath integration_tests/target/ so Maven's clean will clean it up by default and the user doesn't expect things there to be unaffected during builds. Would it be better to update run_pyspark_from_build.sh to do something like that?

kuhushukla
kuhushukla previously approved these changes Sep 3, 2020
@revans2
Copy link
Collaborator Author

revans2 commented Sep 4, 2020

It would be preferable to have the integration tests run from a directory somewhere underneath integration_tests/target/

This script only works being run from the main test directory. So having someone try to run this script from another directory already will run into issues with not being able to find dependencies. Which because of the set -x should cause the script to fail when spark-submit fails so the files would not be removed.

I can try to play around with some things in pytest, but generally pytest assumes that you are running from the base directory with how it tries to play games with importing things.

@revans2
Copy link
Collaborator Author

revans2 commented Sep 4, 2020

build

@revans2
Copy link
Collaborator Author

revans2 commented Sep 4, 2020

@jlowe the tests now run in a sub-directory of target by default.

@jlowe jlowe changed the title Fix leaked files Fix verify build after integration tests are run Sep 4, 2020
@jlowe jlowe merged commit 2d88789 into NVIDIA:branch-0.2 Sep 4, 2020
nartal1 pushed a commit to nartal1/spark-rapids that referenced this pull request Jun 9, 2021
* Fix leaked files

Signed-off-by: Robert (Bobby) Evans <bobby@apache.org>

* Addressed review comments

* Cleanup
nartal1 pushed a commit to nartal1/spark-rapids that referenced this pull request Jun 9, 2021
* Fix leaked files

Signed-off-by: Robert (Bobby) Evans <bobby@apache.org>

* Addressed review comments

* Cleanup
tgravescs pushed a commit to tgravescs/spark-rapids that referenced this pull request Nov 30, 2023
Signed-off-by: Peixin Li <pxli@nyu.edu>

Signed-off-by: Peixin Li <pxli@nyu.edu>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working test Only impacts tests
Projects
None yet
Development

Successfully merging this pull request may close these issues.

[BUG] integration tests produce hive metadata files
3 participants