I am trying to load 100 billion multi-dimensional time series datapoints into Graphite from a CSV file with the following format:
I tried to find a fast loading method on the official documentation and here's how I am currently doing the insertion (my codebase is in Python):
As the code above shows, my code is reading the dataset CSV file and preparing batches of 65000 data points, then sending the datapoints using requests.post.
However, this method is not very efficient. In fact, I am trying to load 100 billion data points and this is taking way longer than expected, loading only 3 Million rows with 100 columns each has been running for 29 hours and still has 991 hours to finish!!!!
I am certain there is a better way to load the dataset into KairosDB. Any suggestions for a faster loading better.