Photo by Hafidh Satyanto on Unsplash
From analyzing environmental data to powering cleaner technologies, computers undeniably have a significant role to play in tackling the climate crises. Yet, while increasing computational power has long been cited as a driving force for improving efficiency and reducing emissions, in reality, things are not quite that simple. Instead, the demand for computing technology has resulted in a paradox: the technologies supposed to empower us are leading to significant waste, insatiable consumer demand, and increased pollution.
A study by Lancaster University suggests that ICT currently accounts for around 2.1–3.9% of global emissions. And crucially, these emissions (despite the increased efficiency of computers) are growing. If we assume continued relative growth, ICT’s relative contribution will exceed 14% of the 2016-level worldwide GHGE by 2040. In other words, around 10 GtCO₂e, or about 80% of the acceptable CO₂ emissions budget.
In response, tech optimists often argue that innovation will save us from disaster. However, even a near-future utopia of super-efficient computation does not solve the problem. That is because, as computer science professor Wim Vanderbauwhede writes:
“We can’t rely on next-generation hardware technologies to save energy: the production of this next generation of devices will create more emissions than any operational gains can offset.”
Instead, in the face of ecological crises and the potential collapse of the current global order, we need to drastically change our approaches to technology to prevent disaster or prepare for it. And in this vain, attempts have been made to theorize, investigate and implement an alternative technological culture that centers environmental and sociological concerns through ‘radically sustainable computing.’ Increasingly central to this premise is the concept and community of practice known as “permacomputing.”