billowy-sundown-29516
06/03/2024, 3:14 PMapache/spark:3.5.0-python3
as the Docker image for the step that is running the Spark job. When I run the workflow, the spark step fails with an error message like this:
24/06/03 15:07:57 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
/usr/bin/python3: can't open file '/opt/venv/bin/entrypoint.py': [Errno 2] No such file or directory
The regular Spark image doesn't seem to have that entrypoint.py script, so I can only assume this integration is supposed to use a Flyte-specific Spark image of some sort. I tried using the Flyte image <http://ghcr.io/flyteorg/flytekit:py3.12-1.12.0|ghcr.io/flyteorg/flytekit:py3.12-1.12.0>
, which I use for my regular Python tasks, but that one doesn't seem to have Spark installed on it.
Is there a particular image that must be used with Spark tasks in Flyte? I can't figure out where this entrypoint script that the Flyte task is using should be coming from.high-accountant-32689
06/03/2024, 10:09 PM<http://ghcr.io/flyteorg/flytekit:spark-1.12.0|ghcr.io/flyteorg/flytekit:spark-1.12.0>
a try?tall-lock-23197
billowy-sundown-29516
06/04/2024, 3:26 PM24/06/04 15:24:04 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
/usr/bin/python3: can't open file '/opt/venv/bin/entrypoint.py': [Errno 2] No such file or directory
Do you expect that script to be present in this image?