--
You received this message because you are subscribed to the Google Groups "OpenTracing" group.
To unsubscribe from this group and stop receiving emails from it, send an email to opentracing+unsubscribe@googlegroups.com.
To post to this group, send email to opent...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/opentracing/d261d6da-8f07-400a-b517-458a83c9a73f%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
I think this was an attempt to do that, not sure how far it went: https://github.com/opentracing-contrib/perfevents
On Mon, Apr 23, 2018 at 4:09 PM, <loic....@gmail.com> wrote:
Hi everyone,I am working in the DORSAL Lab (https://www.dorsal.polymtl.ca/en) which focuses on understanding the performance of complex systems using tracing. The lab had been developing the LTTng kernel and user-space tracer, which is now supported by EfficiOS. Analyzing kernel traces made it very useful to understand complex contention problem, boot-up time of virtual machines or virtual machines performance. Of course, the problem of the current approach taken by LTTng is that analyzing the trace is performed offline, which means the integration with monitoring tools or dashboards is poor.I was thinking of working on integrating the LTTng kernel traces into other tracing or monitoring tools so as to add useful insight into the behavior of distributed systems based on microservices and containers. From what I understand from the OpenTracing ecosystem, the tools currently available make it possible to detect complex performance problems, but not always to explain them. Adding chosen traces from the Linux kernel and analyzing them live along with OpenTracing events could make it possible to explain subtle bottlenecks or errors in complex systems.I have a few options in mind for that integration, but I'd like to know if some people here already thought about kernel tracing as an additionnal source of information in analyses. Comments of that are very welcome! Also, if someone has an example of performance problem that might be hard to explain using the current tools, I'd be very happy to know.Cheers,Loïc.
--
You received this message because you are subscribed to the Google Groups "OpenTracing" group.
To unsubscribe from this group and stop receiving emails from it, send an email to opentracing...@googlegroups.com.
To post to this group, send email to opent...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/opentracing/d261d6da-8f07-400a-b517-458a83c9a73f%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
--
You received this message because you are subscribed to the Google Groups "OpenTracing" group.
To unsubscribe from this group and stop receiving emails from it, send an email to opentracing...@googlegroups.com.
To post to this group, send email to opent...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/opentracing/CAJ9wD%2BpELDAH5SXEcVzjVZ_Zt0%2BT_U071cW_gOa3kxiE8Usssw%40mail.gmail.com.
I think this was an attempt to do that, not sure how far it went: https://github.com/opentracing-contrib/perfevents
On Mon, Apr 23, 2018 at 4:09 PM, <loic....@gmail.com> wrote:
Hi everyone,I am working in the DORSAL Lab (https://www.dorsal.polymtl.ca/en) which focuses on understanding the performance of complex systems using tracing. The lab had been developing the LTTng kernel and user-space tracer, which is now supported by EfficiOS. Analyzing kernel traces made it very useful to understand complex contention problem, boot-up time of virtual machines or virtual machines performance. Of course, the problem of the current approach taken by LTTng is that analyzing the trace is performed offline, which means the integration with monitoring tools or dashboards is poor.I was thinking of working on integrating the LTTng kernel traces into other tracing or monitoring tools so as to add useful insight into the behavior of distributed systems based on microservices and containers. From what I understand from the OpenTracing ecosystem, the tools currently available make it possible to detect complex performance problems, but not always to explain them. Adding chosen traces from the Linux kernel and analyzing them live along with OpenTracing events could make it possible to explain subtle bottlenecks or errors in complex systems.I have a few options in mind for that integration, but I'd like to know if some people here already thought about kernel tracing as an additionnal source of information in analyses. Comments of that are very welcome! Also, if someone has an example of performance problem that might be hard to explain using the current tools, I'd be very happy to know.Cheers,Loïc.
--
You received this message because you are subscribed to the Google Groups "OpenTracing" group.
To unsubscribe from this group and stop receiving emails from it, send an email to opentracing...@googlegroups.com.
Hi Loïc,
What you are proposing sounds really cool and I think it’d be very useful.
I do have two (amazingly smart) high-school students working on extending end-to-end tracing into the kernel. They are propagating the trace context into the kernel and modifying LTTNG’s trace points to accept it. I’m not sure about the performance overheads of this approach yet…but, their final project is due in a month, so we should know by then :).
You might also want to check out this recent paper from Google: http://bit.ly/2qUXONS . Instead of propagating context into the kernel, they use a specific pattern of syscalls to demarcate the start and end of application-level work (i.e., spans) within kernel traces. (See Section 3.3.2.) This allows them to attribute work done by syscalls executed between start and ends of a span to that span. (See the paper for how they handle context switches.)
Anyway, I’m happy to chat more and hear about what direction you decide to go in.
Regards,Raja
Le 24 avr. 2018 à 02:07, Sambasivan, Raja R <r...@bu.edu> a écrit :
Hi Loïc,
What you are proposing sounds really cool and I think it’d be very useful.
I do have two (amazingly smart) high-school students working on extending end-to-end tracing into the kernel. They are propagating the trace context into the kernel and modifying LTTNG’s trace points to accept it. I’m not sure about the performance overheads of this approach yet…but, their final project is due in a month, so we should know by then :).
You might also want to check out this recent paper from Google: http://bit.ly/2qUXONS . Instead of propagating context into the kernel, they use a specific pattern of syscalls to demarcate the start and end of application-level work (i.e., spans) within kernel traces. (See Section 3.3.2.) This allows them to attribute work done by syscalls executed between start and ends of a span to that span. (See the paper for how they handle context switches.)
Anyway, I’m happy to chat more and hear about what direction you decide to go in.
Regards,Raja
On Apr 23, 2018, at 6:54 PM, Ben Sigelman <b...@lightstep.com> wrote: