Skip to content

332 improve spark integration #333

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 5 commits into from
Dec 10, 2024

Conversation

xuwenyihust
Copy link
Owner

No description provided.

…n for improved documentation. Removed redundant markdown cell and updated execution metadata for clarity. This enhances the user experience by providing better context and organization within the notebook.
- Renamed `create_spark_dev` to `create_spark` in `startup.py` to simplify the function name.
- Made `create_spark` available in IPython's global namespace for easier access.
- Removed the default Spark instance creation to allow for manual session management.
- Updated `SparkModel.js` to use the new `create_spark` function for initializing Spark sessions, enhancing integration with the backend API.
- Updated `create_spark` function in `startup.py` to accept an optional `notebook_path` parameter, allowing for dynamic configuration retrieval based on the notebook context.
- Improved error handling in `create_spark` to log errors and use default Spark configuration when the API request fails.
- Modified `SparkModel.js` to pass the `notebookPath` to the `create_spark` function, ensuring proper session initialization.
- Cleaned up the demo notebook by removing outdated code cells, enhancing clarity and usability.
- Updated the configuration keys in the create_spark function within startup.py to use a more concise naming convention, changing 'spark.executor.memory', 'spark.executor.cores', and 'spark.executor.instances' to 'executor_memory', 'executor_cores', and 'executor_instances' respectively.
- Adjusted the corresponding references in the Spark session creation logic to align with the new key names, improving consistency and readability of the configuration handling.
Updated the configuration keys in the create_spark function within startup.py to use the correct Spark naming convention, changing 'executor_memory', 'executor_cores', and 'executor_instances' to 'spark.executor.memory', 'spark.executor.cores', and 'spark.executor.instances'. Adjusted the corresponding references in the Spark session creation logic for improved consistency and clarity.
@xuwenyihust xuwenyihust linked an issue Dec 10, 2024 that may be closed by this pull request
@xuwenyihust xuwenyihust merged commit 02e3486 into release-0.6.1 Dec 10, 2024
4 checks passed
@xuwenyihust xuwenyihust deleted the 332-improve-spark-integration branch December 10, 2024 07:03
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Improve Spark Integration
1 participant