-
Notifications
You must be signed in to change notification settings - Fork 59
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Newest Data Job Base Images are Broken #2391
Comments
Looking at the differences between the two python. base images , the only changes are in PYTHON_VERSION and PYTHON_PIP_VERSION. PYTHON_VERSION=3.11.4 vs PYTHON_VERSION=3.11.3 |
Similar issues: GoogleContainerTools/kaniko#1395 and GoogleContainerTools/kaniko#1045 As far as I can see this issue is caused by trying to use separate image and copying kaniko executor into it. The way we are preparing our job-builder image is like this:
This is apparently known issue : https://github.com/GoogleContainerTools/kaniko#known-issues
So instead making sure we built from kaniko image directly should fix the issue |
The way we are preparing our job-builder image is like this: ``` FROM gcr.io/kaniko-project/executor FROM alpine COPY --from=0 /kaniko /kaniko ``` This is apparently known issue : https://github.com/GoogleContainerTools/kaniko#known-issues > Running kaniko in any Docker image other than the official kaniko image is not supported (ie YMMV). > This includes copying the kaniko executables from the official image into another image. So instead making sure we built from kaniko image directly should fix the issue See #2391
The way we are preparing our job-builder image is like this: ``` FROM gcr.io/kaniko-project/executor FROM alpine COPY --from=0 /kaniko /kaniko ``` This is apparently known issue : https://github.com/GoogleContainerTools/kaniko#known-issues > Running kaniko in any Docker image other than the official kaniko image is not supported (ie YMMV). > This includes copying the kaniko executables from the official image into another image. So instead making sure we built from kaniko image directly should fix the issue See #2391
The way we are preparing our job-builder image is like this: ``` FROM gcr.io/kaniko-project/executor FROM alpine COPY --from=0 /kaniko /kaniko ``` This is apparently known issue : https://github.com/GoogleContainerTools/kaniko#known-issues > Running kaniko in any Docker image other than the official kaniko image is not supported (ie YMMV). > This includes copying the kaniko executables from the official image into another image. So instead making sure we built from kaniko image directly should fix the issue See #2391 Google Java Format control-service: job-builder uisng kaniko fix
The way we are preparing our job-builder image is like this: ``` FROM gcr.io/kaniko-project/executor FROM alpine COPY --from=0 /kaniko /kaniko ``` This is apparently known issue : https://github.com/GoogleContainerTools/kaniko#known-issues > Running kaniko in any Docker image other than the official kaniko image is not supported (ie YMMV). > This includes copying the kaniko executables from the official image into another image. So instead making sure we built from kaniko image directly should fix the issue See #2391 Google Java Format control-service: job-builder uisng kaniko fix
The way we are preparing our job-builder image is like this: ``` FROM gcr.io/kaniko-project/executor FROM alpine COPY --from=0 /kaniko /kaniko ``` This is apparently known issue : https://github.com/GoogleContainerTools/kaniko#known-issues > Running kaniko in any Docker image other than the official kaniko image is not supported (ie YMMV). > This includes copying the kaniko executables from the official image into another image. So instead making sure we built from kaniko image directly should fix the issue See #2391 Google Java Format control-service: job-builder uisng kaniko fix
The way we are preparing our job-builder image is like this: ``` FROM gcr.io/kaniko-project/executor FROM alpine COPY --from=0 /kaniko /kaniko ``` This is apparently known issue : https://github.com/GoogleContainerTools/kaniko#known-issues > Running kaniko in any Docker image other than the official kaniko image is not supported (ie YMMV). > This includes copying the kaniko executables from the official image into another image. So instead making sure we built from kaniko image directly should fix the issue See #2391 Google Java Format control-service: job-builder uisng kaniko fix
The way we are preparing our job-builder image is like this: ``` FROM gcr.io/kaniko-project/executor FROM alpine COPY --from=0 /kaniko /kaniko ``` This is apparently known issue : https://github.com/GoogleContainerTools/kaniko#known-issues > Running kaniko in any Docker image other than the official kaniko image is not supported (ie YMMV). > This includes copying the kaniko executables from the official image into another image. So instead making sure we built from kaniko image directly should fix the issue See #2391 Google Java Format control-service: job-builder uisng kaniko fix
The way we are preparing our job-builder image is like this: ``` FROM gcr.io/kaniko-project/executor FROM alpine COPY --from=0 /kaniko /kaniko ``` This is apparently known issue and caused this outage described in #2391 https://github.com/GoogleContainerTools/kaniko#known-issues > Running kaniko in any Docker image other than the official kaniko image is not supported (ie YMMV). > This includes copying the kaniko executables from the official image into another image. So instead making sure we built from kaniko image directly should fix the issue See #2391 Testing Done: Beyond automated tests I also tested it in one of environments where the issue in #2391 reproduce and verified with the new image the jobs are correctly built. --------- Co-authored-by: github-actions <> Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Upon `vdk deploy` the Control Service is responsible for installing all external dependecies defined in requirements.txt of the data job so they can be used during a cloud run. A recent outage in one of our customers caused data jobs to ignore their requirements.txt and not install any external libraries. It was trace to be due to a release of [data job base image](https://hub.docker.com/layers/versatiledatakit/data-job-base-python-3.11) CICD of the open source VDK did not catch this. I am extending VDK heartbeat to test for the issue so we can catch it in the future. See also #2391
Describe the bug
Latest data job base images do not work properly due to ssl errors caused by the underlying OS packages. Building data job images succeeds, but subsequently job executions fail with User Errors, similar to the one below:
There are some errors in the job builder logs, but these do not seem to stop the process of deployment, and job images are built and pushed successfully to the registry. Sample error logs:
Steps To Reproduce
Steps to reproduce the behavior:
data-job-base-python-3.9:latest
for example.Module Not Found
Expected behavior
The data job executes successfully without errors.
Version (please complete the following information):
Additional context
We managed to workaround the error by using version
1.832469684
of the job base images (e.g., data-job-base-python-3.9:1.832469684)An observation we made was that the newest job base images use python version 3.9.17, which was released on July 6, https://hub.docker.com/layers/library/python/3.9.17/images/sha256-46d99870b9c25e64c5583a59fec411df04191ab6b14cf81a72b9e4a20a78b659?context=explore
The text was updated successfully, but these errors were encountered: