Being new to Confluent Kafka Connect and the whole ecosystem, I have some questions about how this would fit/work with my proposed architecture in a multi-tenant data ingestion platform.
Users will have the ability to use various data endpoint types (files, SQL, etc.) and define endpoints through a dashboard. Thus, the endpoint data metadata will differ from one endpoint to another, and will change through the course of a Schema Registry server being activated.
When defining an endpoint, the user may not yet know the schema, so maybe retrieving sample data and implementing automation or a manual process to define the schema would be necessary. Thoughts/advice on this?
This is why Confluent’s stream data platform includes a Schema Registry. We don’t want to force anyone to use them, and if you decide that Schemas are not applicable for your use-case, you can still use Kafka and the rest of the stream processing platform. But we believe that in many cases, a schema registry is a best practice and want to encourage its use.
It appears that Schemas are not required. Given this use case, what would the best course of action be?
Is there a workflow/process for allowing the Schema Registry configuration to be dynamic? (hopefully I am asking the question correctly here).