Full Text:   <1069>

Summary:  <485>

CLC number: TP393; TK12

On-line Access: 2015-01-29

Received: 2014-05-13

Revision Accepted: 2014-08-08

Crosschecked: 2014-12-30

Cited: 4

Clicked: 2666

Citations:  Bibtex RefMan EndNote GB/T7714


Muhammad Tayyab Chaudhry


-   Go to

Article info.
Open peer comments

Frontiers of Information Technology & Electronic Engineering  2015 Vol.16 No.2 P.119-134


Thermal-aware relocation of servers in green data centers

Author(s):  Muhammad Tayyab Chaudhry, T. C. Ling, S. A. Hussain, Xin-zhu Lu

Affiliation(s):  Department of Computer System & Technology, Faculty of Computer Science & Information Technology, University of Malaya, Kuala Lumpur 50603, Malaysia; more

Corresponding email(s):   mtayyabch@yahoo.com, tchaw@um.edu.my, asadhussain@ciitlahore.edu.pk, luxinzhu2013@siswa.um.edu.my

Key Words:  Servers, Green data center, Thermal-aware, Relocation

Muhammad Tayyab Chaudhry, T. C. Ling, S. A. Hussain, Xin-zhu Lu. Thermal-aware relocation of servers in green data centers[J]. Frontiers of Information Technology & Electronic Engineering, 2015, 16(2): 119-134.

@article{title="Thermal-aware relocation of servers in green data centers",
author="Muhammad Tayyab Chaudhry, T. C. Ling, S. A. Hussain, Xin-zhu Lu",
journal="Frontiers of Information Technology & Electronic Engineering",
publisher="Zhejiang University Press & Springer",

%0 Journal Article
%T Thermal-aware relocation of servers in green data centers
%A Muhammad Tayyab Chaudhry
%A T. C. Ling
%A S. A. Hussain
%A Xin-zhu Lu
%J Frontiers of Information Technology & Electronic Engineering
%V 16
%N 2
%P 119-134
%@ 2095-9184
%D 2015
%I Zhejiang University Press & Springer
%DOI 10.1631/FITEE.1400174

T1 - Thermal-aware relocation of servers in green data centers
A1 - Muhammad Tayyab Chaudhry
A1 - T. C. Ling
A1 - S. A. Hussain
A1 - Xin-zhu Lu
J0 - Frontiers of Information Technology & Electronic Engineering
VL - 16
IS - 2
SP - 119
EP - 134
%@ 2095-9184
Y1 - 2015
PB - Zhejiang University Press & Springer
ER -
DOI - 10.1631/FITEE.1400174

Rise in inlet air temperature increases the corresponding outlet air temperature from the server. As an added effect of rise in inlet air temperature, some active servers may start exhaling intensely hot air to form a hotspot. Increase in hot air temperature and occasional hotspots are an added burden on the cooling mechanism and result in energy wastage in data centers. The increase in inlet air temperature may also result in failure of server hardware. Identifying and comparing the thermal sensitivity to inlet air temperature for various servers helps in the thermal-aware arrangement and location switching of servers to minimize the cooling energy wastage. The peak outlet temperature among the relocated servers can be lowered and even be homogenized to reduce the cooling load and chances of hotspots. Based upon mutual comparison of inlet temperature sensitivity of heterogeneous servers, this paper presents a proactive approach for thermal-aware relocation of data center servers. The experimental results show that each relocation operation has a cooling energy saving of as much as 2.1 kW·h and lowers the chances of hotspots by over 77%. Thus, the thermal-aware relocation of servers helps in the establishment of green data centers.

The paper addresses the interesting problem of cooling-aware placement of servers in the data center in order to reduce the total energy consumption. The basic idea in the methodology adopted is to relocate servers in such a way to reduce the occurrence of hot spots (i.e., parts in the data centre with unusually high air temperature). This goal is obtained by relocating servers so that the outlet temperatures for the different servers are as homogeneous as possible, while satisfying other constraints, mainly related to the maximum temperature in the data center. The solution to the cooling aware problem presented in the paper is innovative. The algorithm presented in the paper is simple yet efficient. The quality of the results are validated empirically on a real case data center.


结论:所提热感知再定位方法应用于数据服务中心服务器可实现节能2.1 kW˙h。再定位之后的服务器出风口温度同构化程度更高。对于每对再定位服务器,可以减少77%热点产生可能性。


Darkslateblue:Affiliate; Royal Blue:Author; Turquoise:Article


[1]ABB, 2013. Efficient DC Power Supply for Data Centers. Available from http://www.electricalreview.co.uk/features/9475-efficient-dc-power-supply-for-data-centres.

[2]Ahuja, N., 2012. Datacenter power savings through high ambient datacenter operation: CFD modeling study. Proc. 28th Annual IEEE Semiconductor Thermal Measurement and Management Symp., p.104-107.

[3]Ahuja, N., Rego, C., Ahuja, S., et al., 2011. Data center efficiency with higher ambient temperatures and optimized cooling control. Proc. 27th Annual IEEE Semiconductor Thermal Measurement and Management Symp., p.105-109.

[4]ASHRAE, 2011. 2011 Thermal Guidelines for Data Processing Environments—Expanded Data Center Classes and Usage Guidance. Available from http://ecoinfo.cnrs.fr/IMG/pdf/ashrae_2011_thermal_guidelines_data_center.pdf.

[5]Banerjee, A., Mukherjee, T., Varsamopoulos, G., et al., 2010. Cooling-aware and thermal-aware workload placement for green HPC data centers. Int. Green Computing Conf., p.245-256.

[6]Banerjee, A., Mukherjee, T., Varsamopoulos, G., et al., 2011. Integrating cooling awareness with thermal aware workload placement for HPC data centers. Sustain. Comput. Inform. Syst., 1(2):134-150.

[7]BBC, 2014. Energy Transfer and Storage. Available from http://www.bbc.co.uk/bitesize/ks3/science/energy_electricity_forces/energy_transfer_storage/ revision/1/.

[8]Corradi, A., Fanelli, M., Foschini, L., 2011. Increasing cloud power efficiency through consolidation techniques. Proc. IEEE Symp. on Computers and Communications, p.129-134.

[9]Huck, S., 2011. Measuring Processor Power TDP vs. ACP. Available from http://www.intel.com/content/dam/doc/white-paper/resources-xeon-measuring-processor-power-paper.pdf.

[10]Jonas, M., Varsamopoulos, G., Gupta, S.K.S., 2007. On developing a fast, cost-effective and non-invasive method to derive data center thermal maps. Proc. IEEE Int. Conf. on Cluster Computing, p.474-475.

[11]Jonas, M., Varsamopoulos, G., Gupta, S.K.S., 2010. Non-invasive thermal modeling techniques using ambient sensors for greening data centers. Proc. 39th Int. Conf. on Parallel Processing Workshops, p.453-460.

[12]Koomey, J., 2011. Growth in Data Center Electricity Use 2005 to 2010. Analytics Press, Oakland, CA.

[13]Kusic, D., Kephart, J.O., Hanson, J.E., et al., 2009. Power and performance management of virtualized computing environments via lookahead control. Cluster Comput., 12(1):1-15.

[14]LD Didactic Gmbh. Converting Electrical Energy into Heat Energy—Measuring with the Joule and Wattmeter. LD DIDACTIC GmbH, Germany.

[15]Lee, E.K., Kulkarni, I., Pompili, D., et al., 2012. Proactive thermal management in green datacenters. J. Supercomput., 60(2):165-195.

[16]Liu, Z., Chen, Y., Bash, C., et al., 2012. Renewable and cooling aware workload management for sustainable data centers. ACM SIGMETRICS Perform. Eval. Rev., 40(1):175-186.

[17]Masiero, M., 2012. CPU Charts 2012: 86 Processors from AMD and Intel, Tested. Tom’s Hardware.

[18]Mersenne Research, Inc., 2012. Great Internet Mersenne Prime Search (GIMPS). Available from http://www.mersenne.org/freesoft/.

[19]Moore, J., Chase, J., Ranganathan, P., et al., 2005. Making scheduling “cool”: temperature-aware workload placement in data centers. Proc. USENIX Annual Technical Conf., p.61-75.

[20]Mukherjee, T., Banerjee, A., Varsamopoulos, G., et al., 2009. Spatio-temporal thermal-aware job scheduling to minimize energy consumption in virtualized heterogeneous data centers. Comput. Netw., 53(17):2888-2904.

[21]Rodero, I., Lee, E.K., Pompili, D., et al., 2010. Towards energy-efficient reactive thermal management in instrumented datacenters. Proc. 11th IEEE/ACM Int. Conf. on Grid Computing, p.321-328.

[22]Rodero, I., Viswanathan, H., Lee, E.K., et al., 2012. Energy-efficient thermal-aware autonomic management of virtualized HPC cloud infrastructure. J. Grid Comput., 10(3):447-473.

[23]Sansottera, A., Cremonesi, P., 2011. Cooling-aware workload placement with performance constraints. Perform. Eval., 68(11):1232-1246.

[24]Tang, Q., Mukherjee, T., Gupta, S.K.S., et al., 2006. Sensor-based fast thermal evaluation model for energy efficient high-performance datacenters. Proc. 4th Int. Conf. on Intelligent Sensing and Information Processing, p.203-208.

[25]Tang, Q., Gupta, S.K.S., Varsamopoulos, G., 2007. Thermal-aware task scheduling for data centers through minimizing heat recirculation. Proc. IEEE Int. Conf. on Cluster Computing, p.129-138.

[26]Tu, C.Y., Kuo, W.C., Teng, W.H., et al., 2010. A power-aware cloud architecture with smart metering. Proc. 39th Int. Conf. on Parallel Processing Workshops, p.497-503.

[27]U.S. Environmental Protection Agency, 2007. Report to Congress on Server and Data Center Energy Efficiency Public Law 109-431. Available from http://www.energystar.gov/ia/partners/prod_development/downloads/EPA_Datacenter_ Report_Congress_Final1.pdf?6133-414f.

[28]VMware Inc., 2009. VMware vSphere Basics ESXi 5.0 vCenter Server 5.0 (in Chinese). Available from http://pubs.vmware.com/vsphere-50/topic/com.vmware.ICbase/PDF/vsphere-esxi-vcenter-server-50-basics-guide.pdf.

[29]Wang, L., von Laszewski, G., Dayal, J., et al., 2009a. Thermal aware workload scheduling with backfilling for green data centers. Proc. IEEE 28th Int. Performance Computing and Communications Conf., p.289-296.

[30]Wang, L., von Laszewski, G., Dayal, J., et al., 2009b. Towards thermal aware workload scheduling in a data center. Proc. 10th Int. Symp. on Pervasive Systems, Algorithms, and Networks, p.116-122.

[31]Wang, L., Khan, S.U., Dayal, J., 2012. Thermal aware workload placement with task-temperature profiles in a data center. J. Supercomput., 61(3):780-803.

Open peer comments: Debate/Discuss/Question/Opinion


Please provide your name, email address and a comment

Journal of Zhejiang University-SCIENCE, 38 Zheda Road, Hangzhou 310027, China
Tel: +86-571-87952783; E-mail: cjzhang@zju.edu.cn
Copyright © 2000 - Journal of Zhejiang University-SCIENCE