SPEC awards, power performance
By walterbays on May 22, 2009
SPECtacular awards. The SPECpower committee has been busy. They
released version 1.10 of
the SPECpower_ssj2008 benchmark as a no-cost upgrade to existing
licensees. It adds support for measurement of multi-node (blade)
servers, improves usability, and adds a graphical display of power
data during benchmark execution. Review and publication of benchmark
results continues apace, with a spirited competition for first place,
and with ever more power
analyzers accepted for testing, and more test labs qualified for
independent publication. They have also been assisting several other
benchmark committees inside SPEC, and other industry
standard benchmark organizations, to implement energy measurement for
their benchmarks. SPECpower is more than just a benchmark; it is a
and the methodology is modified and expanded as necessary over time
to accommodate energy measurements for all the different workloads
which are relevant to the real world in those market segments. In
alphabetical order SPEC recognizes:
Chris Boire (Sun Microsystems) – As release manager he coordinated and integrated development activities to keep the deliverables on schedule.
David Schmidt (HP) – He created stand-alone and network integrated tools for automated results checking to help insure that results submissions are correct and complete.
Greg Darnell (Dell) – Author of the PTDaemon, he helped many other groups get started measuring power for their benchmarks. He helps out with whatever needs to be done, technical or organizational.
Hansfried Block (Fujitsu Technology Solutions) - He automated the process of determining power analyzer precision, handled the acceptance of several new power analyzers, and was instrumental in getting multi-channel analyzers accepted.
Harry Li (Intel) – He was primary developer of the Visual Activity Monitor, giving an unique view of the system's activity.
Jeremy Arnold (IBM) – If I tried to recount all the accomplishments Jeremy was cited for I'd probably run into some internal blog size limit. Suffice it to say he is a primary developer on many parts of the code, who never turns down a plea for help, and who is never satisfied until the entire benchmark package is right.
Karl Huppler (IBM) – As primary author/editor of the Power and Performance Methodology, he organized the document to capture deep technical consensus in the committee, and made it readable and understandable for people new to the field.
Matthew Galloway (HP) – He designed the control software to drive multiple JVMs, enabling multi node (blade) testing.
An engineer (AMD) – Who created and maintained much of the web content explaining the benchmark and methodology to the public.