Hi, I heard that offline batch inference is not re...
# flyte-support
s
Hi, I heard that offline batch inference is not recommended in flyte, and I better to deploy model to end-point and send requests there. Is it true?
f
where did you hear this from
this is infact one of the biggest uses of flyte in the world today
infact you should not deploy an endpoint
s
Great, thank you! It popped up, In our internal conversations, maybe a misunderstanding of something. And I couldn't google any direct answer Just to be completely clear for our internal discussion. Even if the model requires GPU, right?
f
that is indeed true
flyte works really nicely with gpus
now you may want to use unions (reusable) container if you have very short running jobs
s
Awesome, thank you!