Contemporary language models are sustained not only by a global network of human labor, but by physical infrastructure. The computer warehouses where language models are trained are large: a modern data center can be millions of square feet (the size of several football stadiums) and require a lot of water to prevent the machines from overheating. For instance, a data center outside of Des Moines, Iowa, identified as the “birthplace” of GPT-4, used 11.5 million gallons of water for cooling in 2022, drawn from rivers that also provide the city’s drinking water. These challenges have led to decisions to build data centers in regions with cooler climates with more water to draw from; some companies have experimented with putting data centers underwater. (Data centers are used for a lot more than language models, of course; the entire internet lives on these machines.)
Ok, this is terrible environmentally. We have found yet another way to drain away our natural resources. This is infuriating. Instead of putting so much and effort into a non-essential, harmful technology, how about we try to solve the problems that are impacting basic quality of life for millions rather than contributing to the problems.