blog posts

Environmental Risks of Artificial Intelligence

Environmental Risks of Artificial Intelligence

We will see how energy consumption by artificial intelligence creates environmental problems.

Data centers and large artificial intelligence models use large amounts of energy and harm the environment. Jobs can take action to reduce their environmental impact.

Companies can be fascinated by the idea that they need an advanced deep learning system that can do everything. However, they do not need an advanced system if they want to deal with a centralized use, such as billing automation.

Training an advanced AI model requires high-quality time, cost, and data. It also consumes a lot of energy.

Between storing information in large-scale data centers and then using this Data to teach a machine learning model or deep learning, AI energy consumption is high. While an AI system may work monetization, it is an environmental problem.

Consumption of Artificial Intelligence Energy during training

For example, consider some of the most popular language models.

OpenAI trained its GPT-3 model on 45 terabytes of data. To teach the final version of MegatronLM, a similar language model but smaller than the GPT-3, Nvidia 512 ran the V100 GPU for nine days.

Data quality is even more important for big data applications, which include much larger volume, variety, and data speed.

A V100 graphics unit can consume between 250 and 300 watts. Assuming 250 watts, the 512 V100 GPUS consumes 128,000 watts or 128 kilowatts (kW). The nine-day run means that MegatronLM training cost 27,648 kWh.

According to the US Energy Information Administration, the average household consumes 10,649 kWh per year. Therefore, the training of the final version of MegatronLM consumed almost as much energy as the energy of three houses in one year.

There is still a widespread assumption that digital is inherently green, which is far from the truth.

Energy consumption of Datacenters

As AI becomes more sophisticated, expect some models to use more data. This is a problem because data centers use an incredible amount of energy.

“Data centers will be one of the most impactful on the environment,” says Alan Plz Sharp, founder of Deep Analysis. Although artificial intelligence has many benefits for businesses, it does create problems for the environment.

Meteorological company IBM processes about 400 terabytes of data daily so that its models can predict the weather in the coming days around the world in advance. Facebook generates about 4 petabytes (4,000 terabytes) of data daily.

In 2020, people generated about 64.2 zettabytes of information. This figure is about 58,389,559,853 terabytes.

While data centers have become more electrically efficient over the past decade, experts believe that electricity accounts for only 10 percent of a data center’s CO2 emissions. Datacenter infrastructure, including buildings and cooling systems, also generate a lot of CO2.

Data centers can be considered the “brain” of the Internet.

Their job is to process, store and communicate data behind the various information services we rely on every day, whether video streaming, email, social media, online collaboration, or scientific computing. Data centers use various information technology (IT) devices to provide these services, all of which use electricity. Servers provide information, calculations, and logic in response to requests, while storage drives store the files and data needed to meet those requests.

Network devices connect data centers to the Internet and enable incoming and outgoing data streams. The electricity used by these IT devices is eventually converted into heat, which must be removed from the data center using electrically operated cooling equipment. On average, the largest share of direct power consumption is in data centers, servers, and cooling systems, followed by storage drives and network devices.

Artificial intelligence is high in energy, and the greater the demand for artificial intelligence, the more power we use. Electricity is not just for teaching artificial intelligence. It is building supercomputers. It also collects and stores data.

Between 2010 and 2018, global IP traffic – the amount of data transmitted over the Internet – more than tenfold.

While the storage capacity of global data centers has increased in parallel 25 times, during the same period, the number of computational instances running on the world’s servers – the size of the total number of hosted applications – increased more than sixfold.

These strong growth trends are expected to continue as the world consumes more data. And new forms of intelligence services, such as artificial intelligence (AI), which are highly computationally intensive, may accelerate demand growth. Therefore, quantifying and predicting data center energy use is a top priority for energy and climate policy.

With 600 super-large data centers globally – data centers with more than 5,000 servers and 10,000 square feet – it’s not clear how much energy is needed to store all of our data, but it can be staggering.

Image: A Google data center. Google Data Center in Douglas County, Georgia.

Artificial intelligence, Data, and environment

Energy consumption, CO2, is the main greenhouse gas emitter by humans. In the atmosphere, greenhouse gases such as CO2 trap heat near the Earth’s surface, raising the Earth’s temperature and upsetting delicate ecosystems.

“We have an energy crisis,” says Gary McGovern, author of The Worldwide Waste Book. “AI is full of energy, and the more demand there is for artificial intelligence, the more power we use,” he said.

“Electricity is not just for teaching artificial intelligence,” he said. “This usage is building supercomputers. This usage is also collecting and storing data.” McGovern estimates that by 2035, humans will generate more than 2,000 zettabytes of data. “The energy stored for this will be astronomical,” he said.
According to Alan Plz Sharp, founder of Deep Analysis, data centers will be one of the most environmentally friendly.

On average, the largest share of direct power consumption is in data centers, servers, and cooling systems, followed by storage drives and network devices.

Environmental effects of occupations

While regular businesses can not change the way, corporadataata is stored. Businesses concerned about their environmental footprint can create quality, not high-quality data. They can delete data they no longer use, for example. According to McGovern, businesses 90 percent of Data is 90 days after storage; according to t, data centers require more than 100 megawatts of capacity, which is enough to power about 80,000 American households.

Data quality can be a major challenge in any data management and analysis project. Problems can arise from misspellings, different naming conventions, and data integration problems. But data quality is even more important for big data applications, including much more data volume, variety, and speed.

But why is data quality important for big data?

Big data quality issues can not only lead to incorrect algorithms. But also lead to serious accidents and damage and consequently real-world systemic consequences.

At the very least, business users will be less likely to trust the data sets and applications created on them. In addition, companies may be under government scrutiny if data quality and accuracy play a role in early business decisions.

As AI becomes more sophisticated, expect some models to use more data. This is a concern because data centers utilize an extraordinary amount of energy.

Jobs can also adjust how they use artificial intelligence or the type of artificial intelligence they use.

Organizations can think about the specific uses they want to make and choose an AI or automation technology dedicated to that use. However, different types of AI have additional AI energy costs.

Companies can be fascinated by the idea that they need an advanced deep learning system that can do everything, said Plz Sharp. However, they do not need an advanced system if they want to deal with a centralized use, such as billing automation. These systems are costly and utilize a lot of data, which indicates they have a high carbon footmark.

A dedicated system is trained in much smaller volumes of data while probably complementing a specific use and a more general system.