Archive for the 'Uncategorized' Category

24
Apr
14

Current Trends in Power

‘Current,’ Get it?!

http://dailycaller.com/2014/04/23/americas-power-grid-at-the-limit-the-road-to-electrical-blackouts/

This is a light-weight article on electricity with a political agenda, but it does raise some interesting questions.  Data center clients are often drawn to northern climates for moderate temperatures and free cooling.  One consideration that I’ve personally never made is the level of competition for power during extreme cold temperatures, and how this might subject a data center to volatile power costs and the specter of lost power.

The author is correct that older plants are not in compliance with environmental regulations are being shut down.  The question becomes how willing or able are companies to provide new capacity to meet demand?  Could this be a boon for ERCOT in spite of Texas’ climate?

04
Mar
14

Is It Really That Bad?

http://www.nytimes.com/2012/09/23/technology/data-centers-waste-vast-amounts-of-energy-belying-industry-image.html?pagewanted=all&_r=0

Some time has passed since this article was published, so it should be less controversial now than it was when it first came out.  Needless to say, many data center designers, operators and owners had their feathers ruffled a little bit.  I’d like to add my own take on the content of this article, now that I’ve had time to count to 10 and take deep breaths (jk).

I’ll preface this by saying that most everything in the article is true.  Data centers consume a LOT of energy, they often generate power from fossil fuels and even have tailpipes!  The notion that the power consumption is inordinate or inefficient is really a value judgement on the part of Mr. James Glanz, and is the kind of editorializing that sadly passes for news in this day and age.  But, to some extent, his article is right.  Unfortunately, his article is myopic in its understanding of data centers and their function or role in our modern economy.

When I think about how much energy some of these data centers consume, I like to liken them to diesel buses in a public transportation system.  They are big, clunky, and by golly, the consume a lot of fuel.  For the sake of discussion, let’s assume that they get a meager 10 miles to the gallon.  This gas mileage seems abhorrent in the age of Hybrids and high efficiency diesel engines.  But is it terrible for the environment?  It depends.  A hybrid is a very efficient way to get around in a car as a single person.  Let’s say for this discussion that they get 50 miles to the gallon.  If there is a single passenger in the Hybrid, then it also gets 50 passenger miles per gallon.  If a dirty old city bus that gets 10 miles to the gallon has five passengers, then it too gets 50 passenger miles per gallon.  At peak traffic, a bus may hold as many as 30 or more people, meaning that it actually gets 300 passenger miles to the gallon.  That would be 6 times as efficient as the Hybrid.

Something similar happens in Data Centers as well.  When something takes down as much power and cooling as one of these facilities, they actually achieve an economy of scale.  To achieve the equivalent processing power dispersed among end users, there would be considerably more transmission and transformation loss of power, it would be less reliable, and the cooling would be considerably less efficient.  So it’s important to consider power costs per process, which are by far and away less in a Data Center than they would be on a desk top or laptop (if those things even possessed enough processing power to compare).  This is why Cloud Computing is starting to catch on.  You don’t need to have a powerful machine sitting on your desk if you only use it for process intensive purposes every once in a while.  Turning on a top-of-the-line Alienware to check email is terribly inefficient.  Data Centers can possibly allow users to outsource these processes and only use them when they need them.  With increasing use of mobile devices as thin clients to access cloud computing software, we are actually making the world far more efficient.  Imagine having the power of a super computer in a data center at your disposal when you need it, controlled from your iPad, which sips electricity compared to a traditional PC.

But to further address issues of efficiency, Data Center designers and operators are focused on a set of metrics that outline how efficient their installation is.  The reason?  Electricity is expensive, and at the quantities that we are talking about, the incentives are quite real; on the order of 7 and 8 figures real.  One metric (that may be falling out of favor) is the PUE, or Power Utilization Effectiveness.  This is just a ratio to determine how much of the electricity is actually dedicated to computing purposes rather than supporting functions.  The closer to 1 a facility is, the closer it is to having no other loads on the building but the IT load.  Google, a couple of years back, set the bar high at 1.21, which means that a very small percentage is dedicated to support loads (even better numbers have since been reported).  When I was starting out with Data Centers, we planned around a PUE of 2.0, meaning that the total load of the facility, including the IT load, was twice that of the IT load alone.  Now, we plan for an average of 1.6, and that drop has been relatively recent (certainly prior to Mr. Glanz’ article).

The larger issue that concerns Mr. Glanz is the blistering pace of growth for demand in data/processing.  He’s right to suggest that new found gains in efficiency or processing power are quickly met with increased demand for processes and data.  This is an economic issue that escapes cursory concerns about the use of limited resources.  When a process is made more efficient, it is done not to permanently block off the use of that resource for the future, but rather to make this newly freed resource available for other productive uses.  For example, let’s say that a company does something earning $1.00 per process, but that process currently consumes 1 watt of power.  That means that a 1MW facility will return $1,000,000 for every megawatt consumed.  Now, lets say you roll out a new, faster and more efficient processor that reduces the power consumption to .5 watts per process.  Your facility still has a maximum power capacity of 1MW, do you honestly think that the company is going to cap their usage at .5 Megawatts and be satisfied with $1,000,000 of processes?  No, they are going to use a full Megawatt, and earn $2,000,000 by serving more customers, or offering better service to existing customers.  Put another way, does your laptop or desktop really need a shiny new look with fancy graphics and faster processing?  We could probably write this article with DOS or a stripped down text editor in Linux without an X server, and consume WAY less electricity in the process.  Those bouncy icons annoying you on the task bar on your new Apple laptop?  Wasting resources in this view.  Who needs Angry Birds on an iPhone when we could revive Legend of the Red Dragon (those under 30 should probably use Google to learn what this is, or not and save resources) I jest, of course, but these things make the experience better.  If you ever had to stay up late creating a text document in Word Perfect  before it was native to Windows, staring at a blue screen with grey text, or creating a spreadsheet in Lotus 1-2-3 running directly out of DOS, then you don’t need me to explain why things are better now, despite the fact that they require more resources.

This brings up the other economic consideration, that our view of consumption of resources is not temporal, but immediate.  What I mean by this is that if we were to gather computers from 1989 and run the kinds of processes we see today, the cost would be absolutely immense (assuming of course that it would have been possible in the first place).  today, we are capable of producing far more at a fraction of what the previous costs would have been.  This is the result of the cost of a given process being reduced over time in terms of resources consumed.  So, as is often the case, a business function may be developed some years in the past, for previous systems that were far less efficient per process.  Technology has advanced but this legacy software, which still meets the needs of a certain business function remains in use, but is now running on a virtual server that is more efficient per process.  This is the inverse of the last problem, where rather than grabbing more energy to accomplish the same task, they actually grab increasingly less over time, freeing up resources for new business functions.

All of this is to say that this industry, its experts and visionaries, possess a sum total of knowledge and understanding greater than myself, or Mr. Glanz.  At the end of the day, I think the big question should be:  If this stuff is so inefficient, then why is everybody doing it?  I suppose that it could be that we are all just getting it wrong, including the NYT, which hosts the article in question in a Data Center.  But I think that the landscape of Data Centers today is the sum of innumerable decisions by companies and people that make the most sense, and in all of these cases, if they were losing propositions with respect to resources, they would almost certainly be losing propositions with money as well.  It’s well and good to extol the virtues of efficiency and not being wasteful.  But I think these are lessons that businesses already know all too well.

12
Aug
13

Understanding ‘Flight Paths’

Usually before we draw the first wall or column for a new data center, or start demolition in an existing facility, we are asked to aid in site due diligence, which includes evaluating threat assessments for a potential site. These may include nearby hazardous cargo routes, rail ways, nuclear power plants, and the list goes on. Among the most common concerns during this phase is the matter of air travel, which is often brought up as determining ‘flight paths’ or ‘glide slopes’ for airplanes at a nearby airport.

As you can imagine, this is easier asked than answered, and the answers can be quite complicated. I’ll start with the obvious, which I think everyone understands, that airplanes navigate in 3 dimensions and not two. They are physically capable of moving just about anywhere air is and solids or liquids are not. That being said, there are some areas where traffic is more concentrated.

Before I get into where or what those areas are, lets discuss airspace. In the United States, all air above the ground is divided into categories, and in general these categories divide air traffic into different navigation and flight styles (any pilot reading this is probably rolling his eyes right now, but please bear with me, I’m working from scratch here). These categories are assigned a letter, A through G with some more exotic stuff sprinkled throughout. The easiest to understand is Class A, which is a blanket of airspace covering the entire country that starts at 18,000 feet above the ground and extends to the limits of our atmosphere in space. This is reserved for aircraft with instruments and transponders, and is intended to be the highway for air travel.

I should note here that aircraft may travel pretty much anywhere in this space, with the guidance of air traffic control, but there are ‘lanes’ referred to as ‘Victor Lines’ or airways which tend to garner more traffic. These are lines that cross the country and are approximately 8 miles wide, so in this sense, it is possible that your site could be under a ‘flight path,’ and this is probably the closest thing to a defined flight path. It would be good to point out here that every site in the country is under some kind of navigable airspace.

So, if Class A is the ‘highway’ of air travel, where or what are the off ramps? Well, those would be Class B, C and D air spaces. These are ranked in descending order for the volume or quantity of air traffic for a given airport. So for reference, DFW International Airport has Class B airspace around it, Austin has Class C, and Waco has class D. It’s hard to visualize the airspace around an airport, but the best description would be a multi-tiered wedding cake, turned upside down, with the smallest portion extending from the ground at the airport to some altitude above it. The further away you get, the higher the bottom of the tier will be, but usually the top is the same (For example, around DFW, the bottom is at ground level and the top is at 11,000 altitude. If you go out a few miles, the class B airspace would start at 2,000 and go up to 11,000, so forth and so on.)

Class C works much the same way, but usually with only two tiers, and Class D tends to be a cylinder. These air spaces are used to free up the area for traffic that is landing or taking off. It gives planes room to enter a landing pattern around the airport, and keeps the air relatively clear for takeoff. If you are interested in seeing how these spaces are marked off, I recommend checking out some sectional charts that identify these airspaces.

We can start to see that airspace tends to represent a concentration of air traffic at airports. But how do we know how the planes will approach the airport or enter a traffic pattern? The short answer is that we don’t. Where and how planes will fly around airports has a lot to do with conditions at that airport, which can change from day to day. One day, the traffic may land and take off out of the South, the next day, the North. If the airport is exceptionally busy, the pattern may be larger and more expansive.

I should note here that TIA 942 has a recommendation for data centers to be located near, but not in, major metropolitan areas (“Not greater than 16km/10miles” for a Tier 4 Facility) and also provides guidance for proximity to airports (“Not less than .8km/1/2 mile or greater than 30 miles” for a Tier 4 Facility). TIA 942 also says that the facility “should not be in the flight path of any nearby airports.” As we’ve noted, if you are under controlled airspace, or any airspace for that matter, you might be in the ‘flight path’ of a nearby airport, especially if you take the recommendation of the TIA to locate no further than 30 miles from an airport in a big city.

For an area like DFW, the Class B airspace is quite expansive and would certainly cover data centers that were located per TIA. From a due diligence perspective, I think it would be fair to say that you would not want to locate near areas of greater air traffic density, but there is no public site in America that will be free of concerns over air traffic. Furthermore, it would probably be good for organizations like the TIA, and for anyone seeking out a site to develop a more robust understanding of air traffic when trying to assess risk.

26
Jan
12

Distributed Cooling Through Distributed Process?

A coworker of mine forwarded an article to me about Google’s data center in Finland:

http://www.wired.com/wiredenterprise/2012/01/google-finland/

It’s a great read about a new facility that uses sea water that from the pictures appears to make a direct heat exchange with their chilled water system rather than using chillers.  What is more fascinating to me though is the use of a software platform called ‘Spanner’.  The article discusses a data center that Google operates in Belgium where the temperatures are mild, and they seldom need expensive amounts of cooling.

In previous posts, I’ve mentioned how having multiple sites and multi-facility redundancy have eliminated the need for massive CapEx on any one facility for expensive backup power systems.  This scheme by Google shifts processes to another facility when the cooling bill gets too high, effectively using the planet as a kind of massive economizer.  There is always cooling available somewhere around the world.  Of course, I have no doubt that this scheme carries costs.  There is a reason why Google chose to locate a Data Center in Belgium (probably multiple reasons), and shifting a process temporarily might result in slightly less efficient or expedient processes due to transfer time and possibly latency across high speed networks.  

But still, it showcases the classic economic principle of achieving an economy of scale.  They grow more efficient with each facility they build.  Simply amazing.

 

03
Jan
12

They Don’t Build’em Like They Used To

2011 was a good year for me and my wife.  We bought an old home that was likely built in 1915.  We don’t know that for certain, but the records for the original deed and plans were lost in a fire in Dallas in 1930, so all homes in our neighborhood are listed as having been built in 1930, when the earliest remaining records are dated.  We did this because the home is made of old growth pine, old and strong, and as they say it has ‘good bones’ and charm to boot.

I happen to come from a family of builders in the city of McComb, Mississippi.  They built homes and commercial buildings that were contemporaries of my house.  My father still marvels at their perfection, noting that not a crack exists in the foundations that they poured and that the bricks were instructive for any mason living today.  These things are true because they didn’t have the benefits of construction science or building codes as we know them today.  Everything was tremendously over-built to compensate for this lack of knowledge.  To build a building like that today would cost a fortune.

Nowadays, things are built to suit, with as much economy as possible.  This is not to say that buildings are ‘cheap’, but rather that they more closely reflect a near perfect pairing of need and use with the built form.  For example, there is no need to provide a structure that would bear three times the load proposed in the occupancy of the building. The majority of the data center work I have done in the last few years has been in existing facilities, ranging in age from 60+ years to 20+.  Many people see a building that has space and proximity to critical infrastructure and the appropriate setbacks and security measures, and think that it might be suited to a data center occupancy.

I am happy to say that there isn’t a (non-condemnable) structure that can’t accommodate this use.  But this gets into an architectural topic that was vogue in the late 80’s and early 90’s: conversion of program.  The program has quietly become more crucial as buildings have become more efficient for a given purpose.  The ideas passed back in the day were that program and occupancy were largely just social assignments to buildings that could be swapped out in compelling and interesting ways (Like how a McDonalds could be converted into a house, all the essential elements were there, right?).  What this project type has taught me is that this is often easier said than done.  This to me is a axiomatic moment, it is an exhange of economy for flexibility.  Because the building is no longer over built, it may not immediately be suitable for a given occupancy without some work first.  If we spent our time designing and building for a multitude of programs, then a building’s cost would rapidly rise.

So what does all this mean for data centers in existing buildings?  Nothing more than that there are challenges with such a demanding project type.  Most facilities don’t have built-in infrastructure, or structural capacity to fully accommodate the program.  These things can be added.  And then there is an ‘X’ factor that is unique to each building.  I always tell my superiors that something ‘funky’ will come up during the process, and without fail it happens; I have yet to pick up a pattern to the chaos.  These facilities just require a kind of patience and flexibility, because they will challenge even the most meticulous plans.  But I can’t really put into words the satisfaction I feel after a successful data center is inserted into an existing building.

08
Nov
11

Green Power’s Positives

http://www.datacenterdynamics.com/focus/archive/2011/11/winddata-ready-to-break-ground-on-wind-powered-data-center

When I was in architecture school, we had a semester where ideas that revolved around solar power were explored.  The timing could not have been better as the assignment was right on the heels of hurricane Katrina.  At that time, solar power was not as cost competitive as it is today, and it still has a ways to go.  As part of our due diligence when considering this system, we brought in a vendor, and he said something that I would always remember: “It’s nice to get a smaller electricity bill than all of my neighbors, but I know these reductions won’t pay for the system, even after tax rebates.  The real value is when I have power and my neighbors don’t because they rely on the power grid.”

It’s truly anyone’s guess as to whether a solar power plant or wind farm will pay off in the long run.  For all anyone knows, there could be a massive oil reserve discovered next year driving down the costs of energy.  In the short term, it appears that green power simply costs more than traditional forms like coal and oil.  For data center users, this matters because a small shift in the costs per kilowatt hour can translate into large sums of expenditures from operations that consume quite a lot of electricity.  However, in spite of the fact that green power is generally more costly at present, more and more data centers are looking hard at this technology.

If it costs more money, why the interest?  It might be for several reasons.  First, it is a power supply that can be entirely under the control of a private facility owner, and this power supply might be more reliable or add to the reliability of power feeding a facility.  This also removes uncertainties generated from competing interests for electricity.  If capacity were ever reached on a power grid, tough choices would have to be made on how power was distributed to customers; California can attest to that.

Another reason might be the level of cost predictability that comes along with these systems.  The cost of oil or energy can swing wildly in the face of civil unrest, military action or energy cartels.   Renewable energy generation carries a known cost upfront, with a predictable useful life and maintenance cost that can more easily be predicted than the whims of a commodities market.

We might also be looking at renewable energy sources that are for the first time approaching cost competitiveness with traditional power.  If the costs of power continue to rise, then these systems will start to yield cost savings.

In short, the power supplied by these systems might cost more, but cost isn’t really the only factor.  Having the power that you want, that you control, when and where you need it, can be a compelling factor for considering these.  It can take a lot of the volatility and unpredictability out of the cost and supply of energy.

18
Oct
11

Tornadoes and Other High Winds

It goes without saying that  Data Center clients are always looking to locate in an area that fits certain favorable criteria for data centers, not the least of which is a lack of major weather events that might contribute to downtime.  If you look on the SuperNAP website, they are very proud of the fact that they are located in a zone of the country that has virtually no disruptive weather events (Las Vegas).  There is little rain, no hurricanes, earthquakes, volcanoes or few tornadoes.  Of course, weather can’t be the only deciding factor, others might include tax incentives or the cost of utilities, or proximity or connectivity to regional or national locales.  For all the clients that cannot locate their entire IT operation in Las Vegas, weather events will eventually come up as part of the risk assessment.

I am by no means an expert on chance, nor am I an insurance adjuster, but it is still an interesting exercise to watch.  If you have ever played roulette, you have probably seen the sign above the wheel that shows the results of the last few spins.  It should be obvious to all of us that it is purely a game of chance, and that the numbers on the sign are not an indicator that a certain number or color is due because it keeps showing up or doesn’t show up at all.  If the last three spins were 18, then the odds that the next spin will be 18 will still be one in 38 (on an American wheel, for you gambling enthusiasts).  So it is with tornadoes.  We look at historical data, and to some degree we can say with certainty that particular areas of the country will receive more tornadoes than others.  Beyond that, it gets a little fuzzier, and this stems from an evolving process for studying and understanding tornadoes.

The currently accepted measure of severity for tornadoes is the Enhanced Fujita Scale, based on research from the National Weather Service, American Meteorological Society and the Wind Science and Engineering Research Center at Texas Tech.  With time and more exhaustive research, this scale will become more accurate and useful, however, I see two major issues with the scale as it exists currently.  First, the scale is based on observed damage, not on direct measurement of a tornado event.  This means that they look at the damage, and then try to figure out what kind of wind would be required to cause that much damage.  Second, our records of tornado events are incomplete and inconsistent earlier than 1950, when these events were tracked in a national registry.

I’ll address the windspeed first.  The windspeed factor is crucial because this is the end result that we are trying to understand.  When we design a data center, the windspeed which it should resist is among the first decisions made.  We look at the frequency of tornadoes of all scales and pick a design wind speed in excess of that which is required by local building code.  The danger here is that, since the Fujita scale is based on observed damage, it is conceivable that an owner either over builds and thus wastes resources that could have gone into critical infrastructure, or underestimates the risk, exposing the facility to more risk than was assumed based on historical data.

And this brings up the second issue which is historical data.  Consider this:  America is a large country, and is not densely populated from coast to coast, there are gaps between population centers.  If a tornado were to strike one of these areas and not cause any noted damage, would there be damage to assess a windspeed, or would the tornado even be registered?  I suspect that the recorded data probably under-counts the number of tornadoes in a given state or county.  The other question is how accurate past accounts of damage were when being assessed on the Fujita scale.

Ultimately, the decision rests with the owner.  Just like with investments, past performance is no guarantee of future returns, and the decision will come down the owner’s tolerance for risk at that facility.  I have heard of people locating in bunkers, and others who take a riskier approach per facility by distributing processes.