Thanks for the response! Currently, we're doing a pilot project with Debezium where our pipeline is based heavily on Google's CDC parent dataflow project which I'm sure you're aware of.
Our source is Oracle, though, so we are doing Oracle - Debezium Server - Pubsub (all messages pushed to single topic) - Dataflow - Bigquery
We're also not interested in having all of the transactions, so similar to Google's project our Dataflow process takes all the CDC output and merges it into a target table so that we only have the most recent version of a particular record.
This works fine! However, we've discussed internally that it'd be useful for near real-time analytics use-cases to have a process which is just Oracle - Debezium Server - BigQuery (via streaming inserts). We can then clean up the CDC output in views for end-users and not have all of the overhead of running Dataflow.
We could do something like Oracle - Debezium Oracle Connector - Google Bigquery Kafka connect plugin, but Kafka is not widely used in our organisation and so it's cool to have a kafka-less solution.
Apache Camel looks cool, but looks like no Oracle component currently. The Apache Iceberg use-case is also interesting. I think we'd be willing to do a similar thing for BigQuery to see if it'd gain any traction in the community.