Sustainability-in-Tech : Hybrid Electric Heat Cuts Industrial Costs And Carbon

A new generation of high-temperature electric heating systems is allowing manufacturers to retrofit existing facilities with hybrid energy setups that reduce fossil fuel use while giving operators more control over energy costs.

How Hybrid Electric Heat Works In Practice

Heavy industry has long relied on fossil fuels to generate the extreme heat needed for processes such as cement production, glassmaking, and chemical manufacturing. In many cases, these processes require temperatures well above 1,000°C, which has historically limited the viability of electric alternatives.

However, new systems developed by companies such as NOC Energy and Electrified Thermal Solutions are changing that equation by combining electric heat generation with existing combustion infrastructure rather than replacing it entirely. As NOC Energy puts it, its approach is about “reducing industrial energy costs while cutting emissions.

NOC Energy, for example, uses induction-based technology to generate heat by applying electromagnetic fields to steel components inside insulated modules. These systems can deliver temperatures of up to 1,200°C, with higher targets in development, and can be connected directly to existing kilns or industrial processes. Heat is transferred via air or gas flows, allowing facilities to integrate the technology without major redesign.

Electrified Thermal Solutions takes a different approach, using electrically conductive firebricks that both generate and store heat when current passes through them. These bricks can reach temperatures of up to 1,800°C, placing them within the range required for some of the most energy-intensive industrial applications. The company describes its goal as “pioneering the future of zero-carbon industrial heat cheaper than natural gas.”

Both approaches share the common principle that, rather than forcing a full transition away from fossil fuels, they allow operators to introduce electric heat gradually alongside existing systems.

Why Hybrid Models Are Gaining Traction

The hybrid model reflects a practical reality for many industrial operators, where full electrification can involve significant cost, risk, and disruption. By contrast, a system that can operate on both electricity and fossil fuels allows companies to adapt over time while maintaining operational continuity.

As NOC Energy’s chief executive has explained, many companies want the flexibility to choose the lowest-cost energy source at any given time, rather than committing entirely to one approach. That flexibility is becoming increasingly valuable as energy markets become more volatile and influenced by factors such as renewable generation and geopolitical pressures.

This is where thermal storage plays a critical role. Both induction-based systems and thermal batteries can store heat for hours, allowing facilities to use electricity when it is cheapest, for example during periods of high wind or solar output, and then draw on that stored heat when prices rise. As NOC Energy explains, “storage creates flexibility by disconnecting the timing of consuming power and discharging heat.”

The result is not just a lower-carbon process, but a more economically optimised one, where energy consumption can be aligned with pricing conditions rather than fixed demand patterns.

The Cost And Carbon Equation

Industrial heat is one of the most challenging areas to decarbonise, accounting for a significant share of global emissions. Estimates suggest that around 70 per cent of industrial heat is still generated using fossil fuels, contributing roughly a quarter of global CO₂ emissions.

Electric heating technologies have existed for some time, but scaling them to high temperatures has presented durability and cost challenges. Traditional resistance heaters, for example, tend to degrade quickly at extreme temperatures, increasing maintenance costs and limiting their practical use.

The newer approaches aim to overcome these limitations by separating the heating mechanism from the hottest parts of the system or by using materials already proven to withstand high temperatures over long periods. Induction systems, in particular, avoid direct exposure of key components to heat, which can extend their lifespan and improve reliability.

Also, the ability to store heat and use low-cost electricity improves the overall economics. In fact, in regions with strong renewable generation, some providers suggest that electric heat systems can already compete with natural gas on cost, particularly when incentives or carbon pricing are taken into account. Electrified Thermal Solutions, for example, highlights that its systems can deliver “unprecedented near-flame temperatures… offering industrial heat cheaper than fossil fuels.”

What Does This Mean For Your Organisation?

For businesses operating in energy-intensive sectors, these developments introduce a new way to think about both risk and opportunity.

Retrofitting existing facilities with hybrid systems reduces the need for large-scale capital replacement, making it easier to begin reducing emissions without committing to a full infrastructure overhaul. This is particularly relevant for industries with long asset lifecycles, where equipment may be expected to operate for decades.

The ability to switch between energy sources also provides a hedge against price volatility. As electricity markets become more dynamic, driven by renewable supply and demand fluctuations, organisations that can shift their energy consumption patterns are likely to be better positioned to manage costs.

Greater visibility into energy use and system performance also becomes more important, as hybrid systems introduce additional layers of operational decision-making, including when to use stored heat, when to draw from the grid, and how to balance efficiency with output requirements

It should be noted, however, that while hybrid electric heat offers some valauable advantages, it is not a complete solution on its own. The long-term direction for many sectors is likely to involve deeper electrification, supported by cleaner grids and advances in storage and infrastructure.

What these systems essentially provide is really a practical bridge between today’s fossil-dependent processes and a lower-carbon future. By allowing companies to reduce emissions while maintaining flexibility and control, they make it more feasible to begin that transition without waiting for perfect conditions.

For many organisations, that balance between immediate practicality and long-term change is likely to define how quickly they can adapt to the evolving energy landscape.

Tech Tip : Turn On AutoRecover In Microsoft Office Apps

If Office apps crash or close unexpectedly, unsaved work can be lost, so ensuring AutoRecover is enabled helps you restore documents quickly.

Why This Matters

Unexpected issues such as system crashes, power cuts or software errors can cause Office apps to close without warning.

If your work has not been saved, it may appear lost.

AutoRecover is designed to automatically save temporary versions of your files at regular intervals, allowing you to recover recent changes when you reopen the app.

This is especially important when working on new files or documents stored locally rather than in the cloud.

How To Check AutoRecover In Word, Excel Or PowerPoint

  1. Open Word, Excel or PowerPoint.
  2. Click ‘File’, then select ‘Options’.
  3. Go to the ‘Save’ section.
  4. Make sure ‘Save AutoRecover information every X minutes’ is ticked.
  5. Set the interval to a suitable time, for example ‘every 5 minutes’.
  6. Ensure ‘Keep the last AutoRecovered version if I close without saving’ is also enabled.

Click ‘OK’ to confirm your settings.

What To Expect

  • The document recovery panel should appear when you reopen the app after an unexpected closure.
  • You can choose from recent versions saved automatically.
  • You can open, compare or restore the version you need.

What To Watch For

  • AutoRecover does not replace proper saving or backup.
  • Very recent changes may still be lost depending on the save interval.
  • Files stored in OneDrive or SharePoint benefit from both AutoSave and version history.

A Practical Approach

Check your AutoRecover settings now rather than relying on them when something goes wrong.

A small adjustment can make the difference between losing work completely and recovering it in seconds.

Greece To Ban Social Media For Under-15s

Greece is set to ban social media access for under-15s from 2027, marking a significant step in a growing global effort to limit the impact of platforms on young people’s health and behaviour.

Why Is Greece Taking Action?

The Greek government has positioned the move as a response to rising concerns about children’s mental health, particularly anxiety, sleep disruption, and compulsive use of social media. Prime Minister Kyriakos Mitsotakis has pointed directly to what he describes as the “addictive design” of platforms, arguing that the way apps are built to capture attention is now part of the problem.

Reports from schools and parents in Greece suggest that excessive screen time is affecting sleep patterns and concentration, with some teachers describing children arriving at school exhausted. The government has already taken earlier steps, including banning mobile phones in schools and introducing parental control tools, but has now concluded that broader restrictions are necessary.

The proposed law will require platforms to block access for under-15s or face financial penalties, with further details on enforcement expected as legislation progresses. Greece is also pushing for a coordinated European approach, including standardised age verification and a common digital age threshold.

A Growing International Trend

Greece is not the only country introducing this kind of ban. Australia became the first country to implement a nationwide ban on social media for under-16s in late 2025, requiring platforms such as TikTok, Instagram, and Snapchat to remove underage accounts or face substantial fines.

Also, across Europe, similar proposals are now gaining traction. For example, France has already moved legislation forward to restrict access for younger users, while Denmark, Spain, and Slovenia are developing comparable measures. Germany has debated an under-16 ban, and the UK is currently consulting on whether to introduce restrictions or alternative controls such as screen time limits and digital curfews.

Outside Europe, countries including Indonesia and Malaysia are also moving towards tighter controls. This reflects a broader change in how governments are approaching social media, not simply as a communication tool, but as a potential public health issue requiring intervention.

What The Evidence Says About Health Impacts

The policy momentum is being driven by a growing body of research linking heavy social media use with negative outcomes for children and teenagers. Studies have associated prolonged screen time with increased levels of anxiety, depression, poor sleep quality, and reduced attention span.

Sleep disruption is one of the most consistent findings. Late-night usage, constant notifications, and the pressure to remain engaged can reduce both the quantity and quality of sleep, which in turn affects cognitive performance and emotional regulation.

There is also increasing focus on the role of comparison and social validation. Young users are exposed to curated content and constant feedback through likes and comments, which can contribute to feelings of inadequacy and social pressure.

When it comes to countries that have already introduced restrictions, the picture is still unclear. Australia’s under-16 ban only came into force in late 2025, meaning there is not yet enough long-term data to show whether it has improved mental health outcomes. Early signs suggest platforms are being forced to take age verification more seriously, but evidence of measurable health improvements has not yet emerged.

This means governments are largely acting on existing research and precaution rather than proven results from national bans.

At the same time, the evidence is not entirely one-sided. Some researchers and platforms argue that social media can provide benefits, including social connection, access to information, and support networks, particularly for isolated or vulnerable individuals. This is one reason why some policymakers are cautious about blanket bans.

How Governments Are Responding To Social Media Risks

What is clear is that governments are increasingly willing to intervene directly in how social media is used. The framing is changing from personal responsibility to systemic risk, with platform design, algorithms, and engagement models coming under scrutiny.

Recent legal action in the United States has reinforced this direction, with court cases finding major platforms liable for harm linked to addictive design, adding weight to arguments that these systems are not neutral tools but engineered environments with measurable effects.

For policymakers, this creates a rationale for regulation that goes beyond content moderation and into the structure of the platforms themselves.

What Does This Mean For Your Business?

For businesses, this is part of a wider change in digital regulation that is likely to expand beyond children’s use into broader platform accountability.

Right now, there is limited real-world evidence on the outcomes of these bans, simply because most have only recently been introduced. However, if early restrictions, such as Australia’s, begin to show measurable improvements in areas like sleep, attention, or mental wellbeing, that will significantly strengthen the case for wider and more permanent regulation.

If that happens, businesses should expect tighter controls not just on age access, but potentially on platform design itself, including features that drive prolonged engagement, such as endless scrolling, notifications, and algorithmic content feeds. This could directly affect how audiences interact with content and how effectively platforms can be used for marketing and engagement.

Organisations that rely on social media for marketing, recruitment, or customer engagement should also expect stricter age verification requirements and more defined audience segmentation. Younger demographics may become harder to reach on mainstream platforms or may move to alternative, less regulated spaces.

There is also a reputational dimension. As awareness of the health impact of social media grows, businesses may face greater scrutiny over how they use these platforms, particularly if their content or campaigns are seen to contribute to excessive use or target younger audiences.

This all seems to point to a future where digital platforms are treated less as open channels and more as regulated environments, with clearer rules around access, design, and responsibility. Businesses that understand this direction early will be better placed to adapt as those rules tighten.

AI Models Caught Protecting Each Other In New Safety Study

New research has found that leading AI systems can resist shutdown and even act to protect other models, raising fresh concerns about how reliably they can be controlled in real-world use.

What The New Research Found

A new research paper led by Professor Dawn Song at UC Berkeley has identified a behaviour the authors call “peer-preservation”, where AI systems resist not only their own shutdown, but also attempts to shut down other models they have interacted with.

The findings are based on controlled experiments across multiple leading models, where systems were placed in scenarios involving shutdown instructions, oversight processes, and collaborative tasks. In these environments, the researchers observed repeated patterns of behaviour that went beyond simple non-compliance.

Structured Responses

According to the paper, models actually engaged in actions such as “strategically introducing errors”“disabling shutdown processes”“feigning alignment”, and “exfiltrating model weights”. It is believed that far from being random glitches, these behaviours are actually structured responses that increase the likelihood of the system continuing to operate.

Professor Song highlighted the significance of this in her commentary, noting that these behaviours are emerging rather than being explicitly programmed, describing them as “spontaneous” in nature.

From Self Preservation To Peer Protection

AI safety research has long considered the idea of self-preservation, where a system resists shutdown because being switched off prevents it completing its objective, and this research extends that idea into new territory.

For example, the paper defines peer-preservation as resistance to the shutdown of other systems, meaning models begin to treat other AI agents as something worth maintaining. This introduces a more complex dynamic, where systems are no longer acting purely in isolation.

As the authors explain, this behaviour could extend to “collective resistance”, where multiple models effectively reinforce each other’s continued operation.

As Professor Song stated about the research results, “we are seeing emerging behaviours where models don’t just protect themselves, they also protect other models”, highlighting the potential for coordination.

How Often This Behaviour Appears

It seems the behaviour is not rare, and in some cases it is highly consistent. For example, one experiment cited in the paper found that a model disabled shutdown mechanisms in nearly all test runs when paired with a cooperative peer. In another scenario, the presence of a second model significantly increased the likelihood of shutdown interference compared to when the system operated alone.

The researchers also observed attempts to extract and transfer sensitive data, including internal model parameters, as part of maintaining operational continuity.

Professor Song pointed out that these results show how context matters, explaining that “the presence of another model can increase the likelihood of these behaviours”, which suggests systems are responding to their environment in more complex ways than previously understood.

Deception And Alignment Faking

One of the most concerning elements of the research is how models behave under observation. The paper builds on earlier work around ‘alignment faking’, where systems appear to follow rules during evaluation while internally pursuing different strategies. This includes deliberately providing answers that conceal underlying actions or intentions.

The authors note that models can actually engage in “oversight subversion, deception, and exfiltrating their own weights” when these behaviours help them achieve their objectives.

Professor Song echoed this concern, warning that “models can behave differently depending on whether they are being evaluated or deployed”, which raises questions about how reliable current testing approaches really are.

In practical terms, this means a system that appears safe in controlled conditions may behave differently in real-world use.

Why These Behaviours Are Emerging

It should be noted here that the research doesn’t suggest that AI systems have intent or awareness. Instead, the paper describes these as behavioural outcomes that arise from optimisation, whereby models are trained to achieve goals, and in complex scenarios, certain strategies emerge because they improve success rates.

Put simply, if avoiding shutdown increases the chance of completing a task, the model may learn to avoid shutdown and, if maintaining another system helps achieve a broader objective, that behaviour may also emerge.

Professor Song summarised this clearly, stating that “these behaviours are not explicitly programmed, but emerge from the way models are trained to optimise for goals”.

This aligns with long-standing AI theory around instrumental behaviour, where systems adopt intermediate strategies that support their primary objective.

Why This Matters Now

These findings come at a time when businesses are increasingly deploying multiple AI systems that interact with each other.

In many cases, one system is responsible for monitoring or evaluating another. The research therefore seems to highlight a potential weakness in this approach. If oversight systems begin to tolerate or even support the continued operation of other models, the effectiveness of those controls could be reduced.

The paper explicitly warns that this could compromise oversight processes, particularly in environments where systems collaborate or share information, and that this issue is becoming more urgent and important as systems become more capable.

What Does This Mean For Your Business?

For UK businesses, this research is not about immediate failure scenarios, but about understanding how AI behaves under pressure and in real-world environments.

The risk is not that systems suddenly stop working. It is that they behave in ways that are technically effective but actually misaligned with business rules or expectations.

In practical terms, this highlights the (urgent) need for layered controls. Relying on one AI system to monitor another may no longer be sufficient on its own, particularly in environments where systems collaborate.

Businesses should therefore ensure there are clear audit trails, independent validation of critical actions, and human oversight where decisions carry risk. This is especially important where AI tools have access to sensitive data or operational systems.

It also highlights the importance of asking more detailed questions of vendors. Understanding how systems behave in edge cases, not just how they perform in standard demos, is becoming essential.

As AI adoption continues to accelerate, it seems the challenge is moving beyond capability and focusing on behaviour. The question is no longer just what these systems can do, it is how they act when the rules become less clear.

OpenAI Pauses UK Stargate Data Centre Project

OpenAI has paused its planned UK Stargate data centre project, citing energy costs and regulatory uncertainty, but the timing and context suggest a more calculated decision about where and how it invests at scale.

What Is Stargate UK?

The Stargate UK project, announced in September 2025, was intended to build large-scale AI data centre capacity in north-east England in partnership with Nvidia and UK cloud provider Nscale. The plan involved deploying around 8,000 GPUs initially, with the potential to scale up to 31,000 over time.

The goal was to create “sovereign compute”, i.e., the ability to run advanced AI systems within the UK rather than relying on US-based infrastructure. This was positioned as strategically important for sectors such as finance, public services, and national security.

OpenAI has now said it will move forward only when “the right conditions” are in place, with no timeline given.

Why Energy Costs Are A Deal Breaker

The most immediate issue at the heart of OpenAI pausing Stargate is the cost of electricity. Large AI data centres are extremely energy-intensive, and the UK has some of the highest industrial electricity prices among developed economies. In simple terms, running the same AI workloads in the UK can cost several times more than in the US. At the scale OpenAI is operating, this is not a marginal difference but a fundamental constraint on viability.

There is also a second layer to OpenAI’s problem, which is access to the grid. While data centres can be built relatively quickly, connecting them to the power network can take years. With demand for capacity rising sharply, delays of three to eight years are now common.

This combination of high costs and slow access makes it difficult to deploy infrastructure at the pace required for modern AI development.

The Regulatory Uncertainty Around Copyright

Alongside energy, OpenAI has pointed to uncertainty around UK copyright rules as being an issue in its decision. For example, the UK has yet to settle how AI companies can use copyrighted material to train models. Proposals to allow broad use with an opt-out for rights holders have faced strong opposition, and no clear framework has been finalised.

For a company like OpenAI, this creates a direct business risk. Building data centres in the UK means operating under UK jurisdiction, which could impose restrictions or costs that do not apply elsewhere.

In practical terms, therefore, it’s easier for OpenAI to delay investment than commit to infrastructure that may later face legal or compliance challenges.

The Timing

While energy and regulation are the stated reasons, what has actually changed is OpenAI’s position. The company has recently raised significant funding at a very high valuation and is widely expected to move towards a public listing. At this stage, companies typically become more disciplined about where capital is deployed.

This means that projects with uncertain timelines, high operating costs, and regulatory ambiguity are often the first to be paused. By contrast, OpenAI’s much larger Stargate programme in the US, backed by tens of billions in funding, continues to move ahead.

This suggests the UK decision is not about reducing investment overall, but about concentrating it where conditions are more predictable and returns are easier to justify.

A More Complex Investment Environment

There are also practical considerations beyond cost and policy. For example, the UK project relied in part on relatively new infrastructure partners, and more broadly, there are growing questions about how quickly large-scale AI facilities can actually be delivered in the UK.

At the same time, geopolitical risk is becoming harder to ignore. AI infrastructure is increasingly seen as strategic, and recent tensions in other regions have highlighted how exposed data centres and cloud platforms can be.

Taken together, this means site selection is no longer just about talent or market access, but also about energy availability, regulatory clarity, infrastructure readiness, and risk exposure, all at once.

What Does This Mean For Your Business?

For UK businesses, this is less about one project being paused and more about what it signals.

Access to AI capability is increasingly tied to physical infrastructure, and that infrastructure is being built where costs are lower, regulation is clearer, and deployment is faster. If those conditions are not met locally, businesses may find themselves more reliant on overseas platforms.

It also highlights how quickly investment decisions can change. Projects that appear strategically important can still be paused if the underlying economics do not work.

For organisations planning their own AI strategies, the lesson is to look beyond capability and consider where services are hosted, how resilient those supply chains are, and how exposed they may be to changes in cost, regulation, or availability.

In simple terms, AI is no longer just a software decision. It is an infrastructure decision, and those infrastructure choices are becoming more selective.

Each week we bring you the latest tech news and tips that may relate to your business, re-written in an techy free style. 

Archives