Instructure has confirmed that a criminal threat actor accessed Canvas user information and messages, while ShinyHunters claims a far larger education-sector data haul affecting millions of students, teachers, and institutions worldwide.
Trellix says attackers gained unauthorised access to part of its source code repository, but has found no evidence that its release pipeline was affected or that code was exploited.
Stargate has become the clearest warning flare in the AI boom, as Norway, Australia and a handful of hyperscalers turn the race for compute into a high‑stakes battle over who will own, power and ultimately control the global inference economy.
AI image models are no longer competing on visual flair alone. As OpenAI’s GPT Image 1.5 responds to Google’s Nano Banana Pro, the contest shifts to control, safety and who shapes the visual record online, raising new stakes for creators, platforms and public trust.
OpenAI’s GPT Image 1.5 has landed less like a product launch and more like a diplomatic signal to Google. The image race is no longer about spectacle alone. It is about control. Over the past two months, Nano Banana powered Gemini has redrawn the visual AI map, and OpenAI has been forced to close a gap that suddenly matters to advertisers, creators and regulators all at once.
The Nano Banana shock
Google’s Nano Banana Pro arrived in November with what once felt close to heresy in AI art circles: reliably readable text and native 4K output. Reviewers quickly described it as almost too good, erasing much of the remaining visual distance between camera photography and AI generated imagery. Under the hood, Nano Banana Pro fuses Gemini 3’s reasoning stack with a visual engine that understands lighting, physics and character continuity across multiple frames. It then hands creators studio style controls for lenses, colour grading and local edits.
That mix has ignited social platforms and creator tools. Independent front ends such as Nano Banana AI and Higgsfield’s implementation lean heavily into multi character scenes, image fusion and a strong “engineer your reality” message. Influencers, brands and indie filmmakers are using it as a single pipeline for social content, merchandise mock ups and even storyboards. For Google, the strategic win is clear. Gemini stops looking like a chat assistant and starts behaving like a full creative workstation embedded across Android, Docs, YouTube and the wider Google One subscriber base.
OpenAI’s response closes the studio gap
GPT Image 1.5 is OpenAI’s answer to that shift. The model runs up to four times faster than its predecessor, costs less to deploy via API, and finally tackles the issues creators complain about most: prompt accuracy, logo and face preservation, and dense text inside images. The new Images interface inside ChatGPT mirrors Google’s creative studio approach, with preset styles, trending prompts and inline editing that turns the chat window into a visual control room rather than a single shot prompt box.
Introducing ChatGPT Images, powered by our flagship new image generation model.
- Stronger instruction following - Precise editing - Detail preservation - 4x faster than before
Where GPT Image 1.5 really leans in is instruction following rather than cinematic flourish. OpenAI positions it less as a surreal art engine and more as a dependable design tool for brands, marketers and product teams that need exact typography, consistent iconography and rapid iteration. It is also a defensive move. OpenAI cannot afford to let Gemini become the default visual layer of the web while ChatGPT still sits at the centre of daily workflows for hundreds of millions of users.
A crowded and restless image stack
Beyond the OpenAI and Google duel, the wider ecosystem has grown louder and more capable. Independent models such as Flux, Luma Dream Machine and a growing wave of speed focused and aesthetic focused engines prioritise performance, open weights or niche looks suited to anime, product imagery or cinematic trailers. Startups wrap these models in smart user experiences with templates, batch workflows and vertical presets for real estate, fashion or gaming. In doing so, they quietly pull usage away from the largest platforms.
For creators, this fragmentation is both a gift and a burden. There has never been more choice. A creator can sketch a storyboard in Nano Banana Pro, refine brand assets in GPT Image 1.5, then finish everything in a specialised tool tuned for social media formats. At the same time, rights management, visual consistency and disclosure become harder. A single project might pass through several models in a day, each with different training sources, safety filters and commercial terms.
Virality, social platforms and a rising propaganda risk
The Bondi Beach mass shooting, with its explicitly antisemitic targeting, has already become a test case for this new visual environment. Within hours, social platforms filled with a familiar mix of real footage, edited stills, AI enhanced memorials and more disturbingly, synthetic propaganda amplifying conspiracy theories and praise for the violence. As models gain the ability to generate hyper realistic scenes, readable protest signage and convincing broadcast style graphics, the cost of producing persuasive hate imagery has collapsed.
This risk is no longer theoretical. Antisemitic incidents have been rising for years, and Australian authorities had already established a dedicated task force by late 2024. Now, anyone with a prompt window can fabricate visual “evidence” of conspiracies, doctored crowd scenes or pseudo historical images that slot cleanly into existing narratives. Both Nano Banana Pro’s reasoning driven image engine and GPT Image 1.5’s precise instruction handling can be misused to produce targeted visual disinformation that spreads faster than text based fact checking can respond.
Towards an uneasy image détente
The real diplomatic challenge facing AI image generation is no longer whether these systems should exist. It is under what shared rules they can operate in a world this volatile. OpenAI and Google both point to safety filters, watermarking and abuse detection, yet neither system is foolproof. Independent tools built on top of their APIs often weaken or bypass those guardrails entirely.
As antisemitic violence, polarised elections and live conflicts collide with an explosion of photorealistic generative imagery, pressure will mount for a shared baseline. That may include content provenance standards, coordinated incident response across platforms, and firm industry red lines around extremist visual content.
For creators and platforms alike, the trade off is stark. The same tools that let a solo designer in Sydney produce agency quality campaigns from a laptop also allow anonymous actors to mass produce sophisticated hate imagery at scale. GPT Image 1.5 versus Nano Banana Pro is not just a product rivalry. It is the opening act of a much larger negotiation over who gets to shape the visual record of events, and whose prompts are allowed to define what the world believes it has seen.
Get the stories that matter to you. Subscribe to Cyber News Centre and update your preferences to follow our Daily 4min Cyber Update, Innovative AI Startups, The AI Diplomat series, or the main Cyber News Centre newsletter — featuring in-depth analysis on major cyber incidents, tech breakthroughs, global policy, and AI developments.
Sign up for Cyber News Centre
Where cybersecurity meets innovation, the CNC team delivers AI and tech breakthroughs for our digital future. We analyze incidents, data, and insights to keep you informed, secure, and ahead.
Stargate has become the clearest warning flare in the AI boom, as Norway, Australia and a handful of hyperscalers turn the race for compute into a high‑stakes battle over who will own, power and ultimately control the global inference economy.
Australia’s A$25bn AI wager, Bezos’s leap into “physical AI” and Musk’s push to shift data centres into orbit turned this week into a defining moment in the AI global industrial contest, with the Global South emerging as both proving ground and prize in the new AI steel age.
Another week, another frontier model. As Anthropic’s Claude Opus 4.7 chases enterprise depth and OpenAI turns ChatGPT, GPT‑6 and GPT‑Rosalind into the ambient verbs of digital work and lab science, the contest is no longer IQ scores. It is which unseen layer we quietly let sit beneath institutions.
Anthropic’s Mythos clampdown, April’s record Patch Tuesday and Nvidia’s Blackwell‑to‑Rubin GPU roadmap mark a turning point in cyber defence, exposing how deeply allied nations now rely on US‑controlled, agentic AI to detect and counter zero‑day threats.
Where cybersecurity meets innovation, the CNC team delivers AI and tech breakthroughs for our digital future. We analyze incidents, data, and insights to keep you informed, secure, and ahead. Sign up for free!