Hi Jeong,
I agree with you, realistically no production ML system runs on a single CPU. I am not sure what the number of CPUs we use for our service is, but I am pretty sure it's >1.
At the same time, the point of this challenge is not maximise metrics by throwing more hardware at the problem, but to fund creative ways to deal with constraints.
Finally keep in mind that the amount of real-time Twitter data we need to process is orders of magnitude bigger than this dataset. If we would not constrain ourselves at this stage, we would run into all sorts of problems at scale.
We are thinking about how we can have people use their hardware for the test set after the challenge is over, if there are any particular methods they want to test.
Stay tuned for more!
Luca