gray-vr-17331
09/01/2023, 12:18 AMjsonl.gz
file, and write it to object storage in a different cloud account.
I can accomplish this with custom tasks that connect to the source and target object storage systems, but that seems like a lot of boilerplate. It sure seems like Flyte is not designed for these types of ETL efforts and a different tool should be used for those. Am I reading the situation right?freezing-airport-6809
freezing-airport-6809
freezing-airport-6809
freezing-airport-6809
gray-vr-17331
09/01/2023, 2:56 AMjsonl.gz
file to a local file (maybe the Flyte data
3. Uploads the data to the s3 bucket
All this would work fine, of course, but that's the boilerplate I'm referring too. Since neither the source nor destination would be the FlyteObject store configured for the environment, I think we'd need to use some 3rd party SDKs to interact with the object stores.gray-vr-17331
09/01/2023, 2:58 AMfreezing-airport-6809
freezing-airport-6809
gray-vr-17331
09/01/2023, 3:03 AMFlyteFile
/ FlyteDirectory
are good fits for that type of activity - that's greatfreezing-airport-6809
freezing-airport-6809
freezing-airport-6809
gray-vr-17331
09/01/2023, 3:12 AMglamorous-carpet-83516
09/01/2023, 3:13 AMfrom flytekit.sensor.file_sensor import FileSensor
from flytekit import task, workflow
sensor = FileSensor(name="test_sensor")
@task()
def t1():
print("flyte")
@workflow
def wf():
sensor(tmp_file.name) >> t1()
gray-vr-17331
09/01/2023, 3:36 AMFlyteFile
and FlyteDirectory
to interact with external object stores).freezing-airport-6809
thankful-minister-83577