Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add support for Apache Spark(Core module). #20

Open
wants to merge 20 commits into
base: spark
Choose a base branch
from

Conversation

ZHLOLin
Copy link

@ZHLOLin ZHLOLin commented Dec 2, 2022

Add support for Apache Spark(Core module).

Details:

Logging Configuration API

-Code changes
Since Spark is written in Scala and uses the scalatest plugin for unit testing, there are many variations in the runner and collector.
For runner.py:
Add write_report function to generate reports for parsing since scalatest do not generate reports that contains ctest logging information.
Modify some condition statement to support spark.
For collector.py:
Update some function to support projects that allow white space in test name.
Scripts and Const:
modify add_project.sh, identify_param.sh, constant.py, and constant.py to for setting up and generate mapping for Spark.

-Data collected
conf_params.txt, test_method_list.json, spark-core-default.tsv, and opensource-spark-core.json

Intercept Configuration API:

-Code changes
For inject.py
Update injection scripts to support Spark.
The spark store all default configuration as static singleton ConfigEntry objects. The The SparkConf object will load the user-specified configuration from the system properties of the JVM, So I modified the POM file to specify the system properties used by the Scalatest plugin to override the configuration loading.
For run_test_utils.py
Add maven command for running Spark
For run_test.py
Update run_test.py according to the methods changed in utils.

Default value for configuration.
Collected from spark's official document.
https://spark.apache.org/docs/latest/configuration.html
Update constant to support Spark Core module.
LOCAL_SUREFIRE_PATH for spark-core is not added since Scalatest will not generate surefire-reports that contains CTest log.
The reports for parsing is generated by runner scripts.
The Scalatest module allows the use of white space(" ") in test method names. To support the spark-core module, I have updated the variable used in split methods called in lines 28, 29, 39, and 40.
Update runner script.
Add command for running Scala test in Maven.
Add write report method to generate CTest report for parsing.
Add support for spark core module
Add support for spark-core module.
CTEST_SPARK_DIR is the directory for POM file.
Add support for spark core module.
Injection is done by adding system configuration to the pom file.
Many methods in run_test_utils do not support scala test command.
Update run_test.py according to the methods changed in utils.
Add report generator.
@ZHLOLin ZHLOLin changed the title Spark core Add support for Apache Spark(Core module). Dec 2, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant