Skip to content

Commit 1448caa

Browse files
committed
fixedddd build
1 parent ba64a26 commit 1448caa

File tree

2 files changed

+14
-22
lines changed

2 files changed

+14
-22
lines changed

Dockerfile.tmpl

Lines changed: 11 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -31,14 +31,16 @@ RUN uv pip uninstall --system google-cloud-bigquery-storage
3131
# b/394382016: sigstore (dependency of kagglehub) requires a prerelease packages, installing separate.
3232
RUN uv pip install --system --force-reinstall --prerelease=allow kagglehub[pandas-datasets,hf-datasets,signing]>=0.3.9
3333

34+
# google-cloud-automl 2.0.0 introduced incompatible API changes, need to pin to 1.0.1
35+
# Keras 3.6 broke test_keras.py > test_train > keras.datasets.mnist.load_data():
36+
# See https://github.com/keras-team/keras/commit/dcefb139863505d166dd1325066f329b3033d45a
37+
RUN uv pip install --system google-cloud-automl==1.0.1 google-cloud-aiplatform google-cloud-translate==3.12.1 \
38+
google-cloud-videointelligence google-cloud-vision google-genai "keras<3.6"
39+
3440
# uv cannot install this in requirements.txt without --no-build-isolation
3541
# to avoid affecting the larger build, we'll post-install it.
3642
RUN uv pip install --no-build-isolation --system "git+https://github.com/Kaggle/learntools"
3743

38-
# b/385161357 Latest Colab uses tf 2.17.1, but tf decision forests only has a version for 2.17.0.
39-
# Instead, we'll install tfdf with its deps and hope that 2.17.0 compat tfdf works with tf 2.17.1.
40-
RUN uv pip install --system --no-deps tensorflow-decision-forests==1.10.0 wurlitzer==3.1.1 ydf==0.9.0
41-
4244
# b/385145217 Latest Colab lacks mkl numpy, install it.
4345
RUN uv pip install --system --force-reinstall -i https://pypi.anaconda.org/intel/simple numpy
4446

@@ -52,8 +54,8 @@ ADD clean-layer.sh /tmp/clean-layer.sh
5254
ADD patches/nbconvert-extensions.tpl /opt/kaggle/nbconvert-extensions.tpl
5355
ADD patches/template_conf.json /opt/kaggle/conf.json
5456

55-
# /opt/conda/lib/python3.10/site-packages
56-
ARG PACKAGE_PATH=/usr/local/lib/python3.10/dist-packages
57+
# /opt/conda/lib/python3.11/site-packages
58+
ARG PACKAGE_PATH=/usr/local/lib/python3.11/dist-packages
5759

5860
# Install GPU-specific non-pip packages.
5961
{{ if eq .Accelerator "gpu" }}
@@ -108,6 +110,7 @@ RUN apt-get install -y libfreetype6-dev && \
108110
apt-get install -y libglib2.0-0 libxext6 libsm6 libxrender1 libfontconfig1 --fix-missing
109111

110112
# NLTK Project datasets
113+
RUN uv pip install --system --force-reinstall "nltk>=3.9.1"
111114
RUN mkdir -p /usr/share/nltk_data && \
112115
# NLTK Downloader no longer continues smoothly after an error, so we explicitly list
113116
# the corpuses that work
@@ -120,7 +123,7 @@ RUN mkdir -p /usr/share/nltk_data && \
120123
masc_tagged maxent_ne_chunker maxent_treebank_pos_tagger moses_sample movie_reviews \
121124
mte_teip5 names nps_chat omw opinion_lexicon paradigms \
122125
pil pl196x porter_test ppattach problem_reports product_reviews_1 product_reviews_2 propbank \
123-
pros_cons ptb punkt qc reuters rslp rte sample_grammars semcor senseval sentence_polarity \
126+
pros_cons ptb punkt punkt_tab qc reuters rslp rte sample_grammars semcor senseval sentence_polarity \
124127
sentiwordnet shakespeare sinica_treebank smultron snowball_data spanish_grammars \
125128
state_union stopwords subjectivity swadesh switchboard tagsets timit toolbox treebank \
126129
twitter_samples udhr2 udhr unicode_samples universal_tagset universal_treebanks_v20 \
@@ -198,7 +201,7 @@ ADD patches/kaggle_gcp.py \
198201

199202
# Figure out why this is in a different place?
200203
# Found by doing a export PYTHONVERBOSE=1 and then running python and checking for where it looked for it.
201-
ADD patches/sitecustomize.py /usr/lib/python3.10/sitecustomize.py
204+
ADD patches/sitecustomize.py /usr/lib/python3.11/sitecustomize.py
202205

203206
ARG GIT_COMMIT=unknown \
204207
BUILD_DATE=unknown

kaggle_requirements.txt

Lines changed: 3 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ catboost
2323
category-encoders
2424
cesium
2525
comm
26-
cuml-cu12
26+
cuml-cu12>=24.12.0
2727
cytoolz
2828
dask-expr
2929
# Older versions of datasets fail with "Loading a dataset cached in a LocalFileSystem is not supported"
@@ -45,14 +45,6 @@ fuzzywuzzy
4545
geojson
4646
# geopandas > v0.14.4 breaks learn tools
4747
geopandas==v0.14.4
48-
google-cloud-aiplatform
49-
# google-cloud-automl 2.0.0 introduced incompatible API changes, need to pin to 1.0.1
50-
google-cloud-automl==1.0.1
51-
# b/315753846: Unpin translate package.
52-
google-cloud-translate==3.12.1
53-
google-cloud-videointelligence
54-
google-cloud-vision
55-
google-genai
5648
gpxpy
5749
h2o
5850
haversine
@@ -69,15 +61,11 @@ jupyter_server==2.12.5
6961
jupyterlab
7062
jupyterlab-lsp
7163
kaggle-environments
72-
# Keras 3.6 broke test_keras.py > test_train > keras.datasets.mnist.load_data():
73-
# See https://github.com/keras-team/keras/commit/dcefb139863505d166dd1325066f329b3033d45a
74-
keras
7564
keras-cv
7665
keras-nlp
7766
keras-tuner
7867
kornia
7968
langid
80-
leven
8169
# b/328788268: libpysal 4.10 seems to fail with "module 'shapely' has no attribute 'Geometry'. Did you mean: 'geometry'"
8270
libpysal<=4.9.2
8371
lime
@@ -134,14 +122,15 @@ scikit-plot
134122
scikit-surprise
135123
# Also pinning seaborn for learntools
136124
seaborn==0.12.2
125+
git+https://github.com/facebookresearch/segment-anything.git
137126
# b/329869023 shap 0.45.0 breaks learntools
138127
shap==0.44.1
139128
squarify
140129
tensorflow-cloud
141130
tensorflow-io
142131
tensorflow-text
143132
# b/385161357: tf 2.17.1 does not have matching tensorflow_decision_forests release
144-
# tensorflow_decision_forests
133+
tensorflow_decision_forests
145134
timm
146135
torchinfo
147136
torchmetrics

0 commit comments

Comments
 (0)