Hi guys,
is any real difference in the performance between Logminer and XStream?
I have recently tested a 5 million event changes using logminer and it take up to 16 hours to replicate the changes into a Snowflake database.
XStream is a go-to when we talk about near real-time changes?
Or is that the real bottleneck is usually the destination database processing the queries?
--
You received this message because you are subscribed to the Google Groups "debezium" group.
To unsubscribe from this group and stop receiving emails from it, send an email to debezium+u...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/debezium/dfe11ac9-f75c-ec32-f773-d4a070121ee2%40gmail.com.
Hello Carlos,I am using logminer in an environment with over 20k changes/second and the logminer part is perfectly able to handle those amounts of changes. So 16 hours for 5 million events seems extremely long.
Debezium though is not optimized for those numbers. Parsing the SQL statement is very expensive with the used library. I shared my implementation of the parsing with the Debezium project and am in the process to Open-Sourcing our setup based on Kubernetes.
When you run the poll often enough the number of log files you have to open becomes very low. Probably only the redo-log. In my case the setup to start mining is in the order of milliseconds.
Golden Gate is nice but very expensive if you use it in a large environment. Hiring a couple of engineers to maintain a solution like Debezium is often cheaper. And the near-realtime is not that different. Even with polling my delay stays well within a couple of seconds. This should be possible with Debezium as well.
MAG,Milo
To view this discussion on the web visit https://groups.google.com/d/msgid/debezium/CAO5pNX%2B0zxF1-N83ZRvQNaVHa2gDd8HRjJAGc_D%3Dqdf%3DsOWR_A%40mail.gmail.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/debezium/CADGJaX8-yQrQSLCwjo7Ttj%2BtyZAi4M9k4Cv9vB%3Dd5VGfNKXwDg%40mail.gmail.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/debezium/CAO5pNX%2BQMqRMwzFah-5dDh_Nf9-1Dad42QLy7nePbt4c7cPdnA%40mail.gmail.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/debezium/ff7837f0-681f-4907-84e7-ce0f1231f107n%40googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/debezium/ed738ff7-5e24-41f3-1080-76d6af5a08ae%40gmail.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/debezium/CAO5pNXLHTXrFfy34LCotA4qFMu%2BiSdOWqpPRF9xBD2Y3tEKw3w%40mail.gmail.com.
You received this message because you are subscribed to a topic in the Google Groups "debezium" group.
To unsubscribe from this topic, visit https://groups.google.com/d/topic/debezium/3b23GyOtY74/unsubscribe.
To unsubscribe from this group and all its topics, send an email to debezium+u...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/debezium/6058d197-f75e-fee6-2c1f-ce192261a19a%40gmail.com.
Hello Chris,
I do read the dictionary from the database. I had a lot of issues when using the log files. Also because our log files are removed after the backup so the set of logs is sometimes very limited. I did not put much effort in this. If I read your comment then I assume it would be problematic in our situation. We have very frequent log switches. Even though the log is 1GB each.
It's not that it would be problematic but the fact that the streaming of events would have this small 25-30s blip as the current mining session is swapped out with a new one when using redo logs dictionaries. When using the database for the dictionary, this took a matter of 1-2s. Worse case scenario, I would expect the connector to require several of the most recent redo logs or perhaps an archive log in conjunction with those; however, that shouldn't be an issue.
It makes sense that the diskspeed is of influence on the speed of reading the log miner results. In our situation we did not see any impact on the database and reading is fast enough.
What do you mean by disabling the log and still parsing it with the connector? Then you won't get much data... I enabled the ALL and PK flags and obviously that creates much more data and so is slower.But I might misunderstand your comment?
I wonder if supporting DDL changes is worth the effort. A change procedure like:- stop changes- finish reading archives- do DDL change- start changes- start miningSomething like this would work even without supporting DDL changes and reading the dictionary from the online catalog.
To view this discussion on the web visit https://groups.google.com/d/msgid/debezium/2f389207-03fd-c717-63be-0463a2fdaa7d%40gmail.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/debezium/fa7c63d8-3345-4e1c-bfb5-3051b08c27a2n%40googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/debezium/fa7c63d8-3345-4e1c-bfb5-3051b08c27a2n%40googlegroups.com.