skip to main content
10.1145/2818613.2818743acmconferencesArticle/Chapter ViewAbstractPublication PagessospConference Proceedingsconference-collections
research-article

Modeling communication costs in blade servers

Published:04 October 2015Publication History

ABSTRACT

Datacenters demand big memory servers for big data. For blade servers, which disaggregate memory across multiple blades, we derive technology and architectural models to estimate communication delay and energy. These models permit new case studies in refusal scheduling to mitigate NUMA and improve the energy efficiency of data movement. Preliminary results show that our model helps researchers coordinate NUMA mitigation and queueing dynamics. We find that judiciously permitting NUMA reduces queueing time, benefiting throughput, latency and energy efficiency for datacenter workloads like Spark. These findings highlight blade servers' strengths and opportunities when building distributed shared memory machines for data analytics.

References

  1. AMD. The truth about power consumption starts here. In AMD White Paper: Power Consumption, 2009.Google ScholarGoogle Scholar
  2. J. Antony et al. Exploring thread and memory placement on numa architectures. In HiPC, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. B. Goglin and N. Furmento. Enabling high-performance memory migration for multithreaded applications on LINUX. In IPDPS, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. T. Ham et al. Disintegrated control for power-efficient and heterogeneous memory systems. In HPCA, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. B. Holden. Latency comparison between HyperTransport and PCI-Express in communication systems. In HyperTransport Consortium Technical Note, 2006.Google ScholarGoogle Scholar
  6. R. Hou et al. Cost effective data center servers. In HPCA, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. A. Joy et al. Analog-DFE-based 16Gb/s SerDes in 40nm CMOS that operates across 34dB loss channels at Nyquist with a baud rate CDR and 1.2Vpp voltage-mode driver. In ISSCC, 2011.Google ScholarGoogle Scholar
  8. A. Kazmi. Minimizing PCI Express power consumption. In PCI-SIG Developers Conference, 2007.Google ScholarGoogle Scholar
  9. T. Li et al. Efficient operating system scheduling for performance-asymmetric multi-core architectures. In SC, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. K. Lim et al. Disaggregated memory for expansion and sharing in blade servers. In ISCA, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. K. Lim et al. System-level implications of disaggregated memory. In HPCA, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. H. Löf et al. Affinity-on-next-touch: Increasing the performance of an industrial PDE solver on a cc-NUMA system. In ICS, 2005.Google ScholarGoogle Scholar
  13. K. Malladi et al. Towards energy-proportional datacenter memory with mobile DRAM. In ISCA, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. D. Meisner et al. BigHouse: A simulation infrastructure for data center systems. In ISPASS, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Micron. Calculating memory system power for DDR3. In Technicalj Note TN-41-01, 2007.Google ScholarGoogle Scholar
  16. PLX Technology. PLX PCIe switch power consumption. In PLX Technology White Paper, 2008.Google ScholarGoogle Scholar
  17. J. Rao, K. Wang, et al. Optimizing virtual machine scheduling in NUMA multicore systems. In HPCA, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. J. Regula. Using non-transparent bridging in PCI Express systems. In PLX Technology White Paper, 2004.Google ScholarGoogle Scholar
  19. L. Tang et al. Optimizing Google's warehouse scale computers: The NUMA experience. In HPCA, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. M. Zaharia et al. Delay scheduling. In EuroSys, 2010.Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Modeling communication costs in blade servers

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Conferences
            HotPower '15: Proceedings of the Workshop on Power-Aware Computing and Systems
            October 2015
            37 pages
            ISBN:9781450339469
            DOI:10.1145/2818613

            Copyright © 2015 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 4 October 2015

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article

            Acceptance Rates

            HotPower '15 Paper Acceptance Rate7of12submissions,58%Overall Acceptance Rate20of50submissions,40%

            Upcoming Conference

            SOSP '24

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader