Hi Huy,
I may help you for some suggestion, because of handling over 100,000 records/seconds...
1. For monitoring of missing of records, we have 2 ways like this:
* watching Fluentd's logs with configuration such like '<match fluent.**>'
* flow counting by 'fluent-plugin-flowcounter' and put it to graphs
2a. log-forwarder node are used to provide these features:
* buffering in downtimes of network between nodes and log-aggregators
* forwarding with load balancing and/or active-standby supports
(some of fluent-logger libraries doesn't have these features)
2b. Of course, yes.
3. Scaling up of fluentd throughput is very hard, and scaling out
by load balancing is very simple and easy to do. (And we do so.)
2013年10月15日火曜日 18時40分54秒 UTC+9 Huy Nguyen: