Hi,
I'm developing a logging ingestion application that has the special twist of concentrating more on "how many" rather than "what".
As such, I'm curious if I can make it scale to millions of requests per second, as the plaintext benchmarks indicate I should be able to.
The plaintext benchmarks, however, send multiple requests at once if I understand that correctly (that's what the pipeline.lua script is for), and also there's the question if one wrk instance generating all the load is even a realistic - I reckon TCP connections will be reused, for example.
I'm fairly confident I can optimize my own app enough to handle the load (
ASP.NET Core and not even doing json parsing), but I have no idea what stress OS connection bookkeeping and whatever else there may be is an issue.
And I don't know how I would create more realistic load and whether it's worthwhile pursuing.
Existing load services have pitifully low rps and are more geared towards saturating a PHP script.
Any advice from the experts would be appreciated.
Thanks, Jens