prehistoric-mechanic-34647
08/11/2022, 10:36 PM.pymakemake download_datamake_preprocess_datash.pymake run_entire_pipeline.py@task@workflow.mrxs.tiff.czi.jpeg.pngfreezing-airport-6809
freezing-airport-6809
Obviously this rough pipeline misses a few benefits such as caching earlier steps such that they do not need to be executed if they’ve already been performed (e.g. no need to re-download data on a subsequent model training pipeline run if the data has already been downloaded on an earlier run of the pipeline and if there have been no changes in that data).The benefits you get is - failure tolerance, distributed execution, caching and isolation. Today you do not get the benefit of re-using data that has been downloaded already
What is the best way to migrate this Make-based workflow into a Flyte-based workflow? Specifically is there a way to map eachThere are 2 ways 1. Use the ShellTask to model what you had today with little more data passing. Thus model it as a Flyte workflow 2. Or update your scripts to have ascripts to a.pywhen building a@taskpipeline in Flyte? I learned about the Flyte “Script mode”, and it sounds somewhat akin to what I’m trying to do, but I’m totally new to Flyte. Thanks for any help and direction.@workflow
task@task
def foo(...):
  globals ...@workflow@taskI’m working with very large digital pathology whole slide image (WSI) images, BTW. Does Flyte support inputs of the WSI variety? I.e.Any type of File can be handled using the FlyteFile It will automatically upload and download files to S3/GCS etc Example: Workfing Withe Files,.mrxs,.tiff,.czi,.jpeg, etc.png
prehistoric-mechanic-34647
08/12/2022, 2:57 PMShellTaskShellTask{inputs.input_name}{outputs.output_name}{inputs.myHydraConfigObject}{inputs.hydra_object.learning_rate}{inputs.hydra_object.num_epochs}freezing-airport-6809
late-pencil-3873
08/15/2022, 7:07 AMtall-lock-23197
prehistoric-mechanic-34647
08/16/2022, 10:02 PMhydra-coreflytefreezing-airport-6809
pyflyte run