Inside a Google data center

Inside a Google data center

MALE SPEAKER 1: A data centre’s
the brains of the Internet. MALE SPEAKER 2: The
engine of the Internet. FEMALE SPEAKER 1: It is a giant
building with a lot of power, a lot of cooling and
a lot of computers. MALE SPEAKER 3: It’s row, upon
row, upon row of machines, all working together to
provide the services that make Google function. JOE KAVA: I love building
and operating data centres. I’m Joe Kava, Vice-President
of Data Centres at Google. I’m responsible for managing
the teams globally that design, build and operate
Google’s data centres. We’re also responsible for
the environmental health and safety, sustainability
and carbon offsets for our data centres. This data centre, here
in South Carolina, is one node in a larger
network of data centres all over the world. Of all the employees at Google,
a very, very small percentage of those employees are
authorised to even enter a data centre campus. The men and women
who run these data centres and keep them up 24
hours a day, seven days a week, they are incredibly passionate
about what they’re doing. MALE SPEAKER 2: In layman’s
terms, what do I do here? FEMALE SPEAKER 1: I
typically refer to myself as the herder of cats. MALE SPEAKER 4: I’m an engineer. MALE SPEAKER 3: Hardware
site operations manager. MALE SPEAKER 2: We
keep the lights on. MALE SPEAKER 1: And
we enjoy doing it. JOE KAVA: And they
work very hard, so we like to provide them with a fun
environment where they can also play hard as well. FEMALE SPEAKER 2: We just went
past the three-million-man-hour mark for zero
lost-time incidents. Three million man-hours
is a really long time, and with the number of people
we have on site, that is an amazing accomplishment. JOE KAVA: I think that the
Google data centres really can offer a level of security
that almost no other company can match. We have an information
security team that is truly second to none. You have the expression,
“they wrote the book on that.” Well, there are many of
our information security team members who
really have written the books on best practices
in information security. Protecting the security
and the privacy of our users’ information is
our foremost design criterion. We use various layers
of higher-level security the closer into the centre
of the campus you get. So, just to enter
this campus, my badge had to be on a
pre-authorised access list. Then, to come into
the building, that was another level of security. To get into the secure corridor
that leads to the data centre, that’s a higher
level of security. And the data centre and
the networking rooms have the highest
level of security. And the technologies
that we use are different. Like, for instance, in
our highest-level areas, we even use underfloor
intrusion detection via laser beams. So, I’m going to demonstrate
going into the secure corridor now. One, my badge has to be on the
authorised list. And then two, I use a
biometric iris scanner to verify that it truly is me. OK, here we are on
the data centre floor. The first thing that
I notice is that it’s a little warm in here. It’s about 80
degrees Fahrenheit. Google runs our
data centres warmer than most because it
helps with the efficiency. You’ll notice that we have
overhead power distribution. Coming from the yard outside, we
bring in the high-voltage power distributed across the bus bars
to all of the customised bus taps that are basically
plugs, where we plug in all the extension cords. All of our racks don’t really
look like a traditional server rack. These are custom designed
and built for Google so that we can
optimise the servers for hyper-efficiency and
high-performance computing. It’s true that
sometimes drives fail, and we have to replace
them to upgrade them, because maybe they’re no
longer efficient to run. We have a very thorough
end-to-end chain-of-custody process for managing
those drives from the time that they’re
checked out from the server til they’re brought to an
ultra-secure cage, where they’re erased and
crushed if necessary. So any drive that can’t
be verified as 100% clean, we crush it
first and then we take it to an
industrial wood chipper, where it’s shredded into
these little pieces like this. In the time that I’ve been
at Google – for almost six and a half years
now – we have changed our cooling technologies
at least five times. Most data centres have
air-conditioning units along the perimeter walls that
force cold air under the floor. It then rises up in
front of the servers and cools the servers. With our solution, we
take the server racks and we butt them right up against
our air-conditioning unit. We just use cool water
flowing through those copper coils that you see there. So the hot air from the servers
is contained in that hot aisle. It rises up, passes
across those coils, where the heat from
the air transfers to the water in
those coils, and then that warm water is then
brought outside the data centre to our cooling plant,
where it is cooled down through our cooling
towers and returned to the data centre. And that process is just
repeated over and over again. To me, the thing that amazes
me about Google and the data centres is the
pace of innovation and always challenging the
way we’re doing things. So, when people say that
innovation in a certain area is over, that we’ve kind of
reached the pinnacle of what can be achieved, I just laugh. [MUSIC PLAYING]

100 thoughts on “Inside a Google data center

  1. Remember, Google: As impressive as this technology may be it doesn't give you the right to curtail our democratic right of free speech.

  2. I would hope that they could use the heated water some how. Like to generate electricity with metamaterials or do something with it. I am sure there is a use they could recycle the heat generated from all the servers instead of just cooling it off.

  3. @0:59 Only a very selective of our employees are even allowed inside……but ofcourse a pug 🐶is allowed

  4. As an IT manager of a small company with only 250 computers I remember I got grey hairs worrying about my tinny data center all the time ………..

  5. servers produces heat. temparature normally 6 degree. long back i entered one data center for work

  6. First individual: Data center is the brain of the internet.
    Second individual: The engine of the internet
    Third individual: iTs A GiaNT bUilDiNG










  8. am just asking a question why the google map its not up date to view the place i cant view my house in google map😂🤔

  9. Google needs to be careful and stop censorship of conservative ideas / media or it’s going to be regulated

  10. Hmm look s like it will be difficult for them to change all these and YouTube server to ARM based systems….

  11. it's a great job that google people make. But if you wish some advise, as lately has happened things like the recently earthquakes on california (were silicon valey is), and also the ny last blackouts has happened, I think that there should plann who could face those things, in case of service,and also medical assistance if it could be required.

  12. …have really written the books on information sec…Oh. I mean privacy invasion, anti-capitalist monopoly practices and election tampering. Our new motto: [just the sound of Darth Vader breathing]

  13. Wow I work in a Data center In NJ and we have our air conditioner units around the perimeter. Its cool to see google changing it up. Thanks Joe

Leave a Reply

Your email address will not be published. Required fields are marked *