We did use Ariflow Operators, but we don't really need this kind of
feature at this moment - we're pretty comfortable calling our
Hive/Presto and HDFS from Python directly.
As for Argo and Events, there were few issues we had with it:
1. It's hard, what we needed really is to trigger workflow every
crontab. With Argo Events that would require creating Gatewat, Sensor
and then in Sensor have nested workflow, which makes it super hard to
read. What we did is simple operator that takes new CRD CronWorkflow,
which is really just workflow definition + 1 spec field - crontab. It
then creates CronJob that will execute workflow in spec. Pretty
simple, all and all ~200 lines of python and makes our lives much
easier.
2. Argo UI really needs some way to filter workflows. One flat list is
unmanagable for people who run them in cron + ad hoc + CI:) We've
created our own UI that pretty much allows listing and searching
workflows. It also integrates with CronWorkflow and allows to show all
executions of particular cronjob with single click. React in official
argo ui proven to be too much for infra engineers and data scienists
:(
On Sun, 24 Mar 2019 at 21:20, Kapas, Mukulika <
Mukulik...@intuit.com> wrote:
>
> Hi Michal,
>
> This is great feedback that data scientists are fine with YAML and finding it easy to use Argo.
>
> We got feedback from others also for Argo events and would look into how better to integrate eventing with Argo workflows. Will also give feedback to Blackrock team who are maintaining Argo events.
>
> Were you guys using airflow operators? Are you planning to build similar connectors for data sources that can be invoked from Argo?
>
> Please let us know if you have any questions.
>
> Any talk /blog on this topic would be great whenever you are comfortable.
>
> Jeremy, thanks for getting us connected.
>
> Regards,
> Mukulika
>
> On 3/24/19, 4:49 PM, "Michał Jastrzębski" <
inc...@gmail.com> wrote:
>
> This email is from an external sender.