You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm seeing that in production the scikit-learn version is 1.0.2 which corresponds to the DBR version of my workflow cluster (12.2).
Steps to Reproduce (for bugs)
Context
I'm using the DBX workflow to build my package using poetry and run certain code from it on Databricks clusters as jobs using DBX workflow definitions. I am expecting that the code running in the Databricks cluster, after being executed via DBX, will have the same environment as I do locally given that locally I am building my environment using poetry, and DBX also builds my wheel using poetry (same dependency specifications). I am attempting to use DBX as a way to centralize my projects configurations so that I can use the same dependency versions in both my development and production environments. Is there anyway to force DBR clusters to use my poetry-specified dependencies with DBX (as in, override the cluster installed packages through DBX)?
Your Environment
dbx version used: 0.8.18
Databricks Runtime version: 12.2 LTS ML GPU
The text was updated successfully, but these errors were encountered:
Edwardp17
changed the title
Production runtime doesn't matter development
Production runtime doesn't match development
Sep 19, 2023
Edwardp17
changed the title
Production runtime doesn't match development
How to override DBR cluster installed packages with DBX
Sep 20, 2023
Edwardp17
changed the title
How to override DBR cluster installed packages with DBX
Overriding DBR cluster installed packages with DBX
Sep 20, 2023
Expected Behavior
Production
scikit-learn
version to be 1.3.0.Current Behavior
I'm seeing that in production the
scikit-learn
version is 1.0.2 which corresponds to the DBR version of my workflow cluster (12.2).Steps to Reproduce (for bugs)
Context
I'm using the DBX workflow to build my package using poetry and run certain code from it on Databricks clusters as jobs using DBX workflow definitions. I am expecting that the code running in the Databricks cluster, after being executed via DBX, will have the same environment as I do locally given that locally I am building my environment using poetry, and DBX also builds my wheel using poetry (same dependency specifications). I am attempting to use DBX as a way to centralize my projects configurations so that I can use the same dependency versions in both my development and production environments. Is there anyway to force DBR clusters to use my poetry-specified dependencies with DBX (as in, override the cluster installed packages through DBX)?
Your Environment
The text was updated successfully, but these errors were encountered: