Web Search Using Mobile Cores: Quantifying and Mitigating the Price of
Efficiency
Authors
Vijay Janapa Reddi, Benjamin Lee, Trishul Chilimbi, Kushagra Vaid
Date
ISCA'10 - International Symposium on Computer Architecture, 2010
Novel Idea
The work describes and quantifies what the authors call "the price of
efficiency" - how increasingly stringent demands (in a search engine
context) on low cores impact the general QoS.
Main Results
The paper compares the performance of Xeon and Atom processors under a
particular search engine workload and measure the impact that low
cores have on the QoS. They also propose some strategies to lighten
the "price of efficiency".
Impact
The paper promotes the discussion about data center efficiency in a
particular workload scenario to the processor micro-architecture
level. Therefore, I believe that their results give important
characterizations both to the hardware design industry and to managers
that need to quantify investments in data centers for this particular
scenario.
Evidence
The authors present multiple characterizations of "the price of
efficiency": how processors adhere to QoS restrictions, how they adapt
to bursts of demands (including some metrics particularly important
for web-searching, as the latency cutoff), complexity, throughput, and
others.
I think that they have an excellent evaluation section.
Prior Work
They mention some articles that identify change trends in data
centers, the increasing awareness about power, and systems that also
propose using low-power cores in a similar scenario (including FAWN).
Competitive Work
They mention the Piranha and Niagara, as well as FAWN (in which case
they note that their discussion is on the processor micro-architecture
level instead on the consumption measurement level).
Reproducibility
They use industry processors and their mention the tools they employ
to their evaluations. I believe that their results are reproducible.
Questions + Criticism
[Criticism] When the authors are discussing about the issue of
"over-provisioning and under utilization", I think that they should
have explored the possibility of just employing processors in which we
can dynamically scale the voltage/frequency. [Question] Would such
approach provide the appropriate performance on low power and low
demand scenario, as well as being flexible in the sense that when more
load is required, as the frequency can be adapted?
Ideas for Further Work
Trying out middle range processors and dynamic voltage and frequency scaling.
Authors: Vijay Janapa Reddi, Benjamin Lee, Trishul Chilimbi, and Kushagra Vaid
Date: ISCA 2010
Novel idea: As the authors clearly state, "advances in processors
comprised of small cores provide opportunities for power efficiency."
Main results: Search is 5x more efficient on Atom than on Xeon
per-core on the basis of queries per second per watt. For search,
query latency is higher on Atom than on Xeon; Xeon is able to absorb
activity spikes more smoothly than Atom; and Atom is more reliable and
energy-efficient.
Impact: Switching to processors such as Atom may provide TCO benefits.
Evidence: The authors analyzed the performance of both Atom and Xeon
processors running their search workload using VTune, a toolbox that
provides access to processor hardware counters.
Prior work: The authors cite inspiration from Lim et al's unified
systems as well as FAWN.
Competitive work: In contrast to previous approaches, the authors take
into account microarchitectural details of processor performance in
addition to higher-level metric such as energy consumption.
Reproducibility: Assuming you had a large number of computers with
some kind of similar search workload (with both Atom and Xeon
processors), it should be possible to collect similar metrics.