This week, two insider situations involving tech giants Microsoft and Google raised questions on the responsible development of AI systems and mental property management. 

First, at Microsoft, Shane Jones, a principal software engineering manager with six years of experience, has been independently testing the AI image generator Copilot Designer in his free time. 

Jones told CNBC that he was deeply troubled by the violent, sexual, and copyrighted images the tool was able to generating. “It was an eye-opening moment,” Jones said. “It’s once I first realized, wow this is de facto not a secure model.”

Since November 2022, Jones has been actively testing the product for vulnerabilities, a practice referred to as red-teaming. 

He discovered that Copilot Designer could create images depicting “demons and monsters alongside terminology related to abortion rights, teenagers with assault rifles, sexualized images of girls in violent tableaus, and underage drinking and drug use.” 

Despite reporting his findings to Microsoft in December, Jones said the corporate has been reluctant to remove the product from the market.

Microsoft’s Copilot has acted strangely on occasions, including adopting a “god mode” that saw it vow to pursue world domination. 

In a letter addressed to Federal Trade Commission Chair Lina Khan, Jones wrote, “Over the last three months, I actually have repeatedly urged Microsoft to remove Copilot Designer from public use until higher safeguards could possibly be put in place.” 

He added that since Microsoft has “refused that suggestion,” he urges the corporate so as to add clear disclosures to the product and alter its rating on Google’s Android app to point that it is barely suitable for mature audiences.

Copilot Designer has been reportedly easy to coax into bypassing its guardrails and was accountable for the recent notorious explicit images of Taylor Swift, which circulated to thousands and thousands across social media.

As Jones argues, the ability of AI systems to generate disturbing and potentially harmful images raises serious questions on the efficacy of safety features and the way easy they’re to subvert. 

Insider controversy at Google

Meanwhile, Google is grappling with its own AI-related controversy.

Linwei Ding, also referred to as Leon Ding, a former Google software engineer, was indicted in California on 4 charges related to allegedly stealing trade secrets about AI while secretly working for 2 Chinese firms. 

The Chinese national is accused of stealing over 500 confidential files related to the infrastructure of Google’s supercomputing data centers, which host and train large AI models.

According to the indictment, Google hired Ding in 2019 and started uploading sensitive data from Google’s network to his personal Google account in May 2022. 

These uploads continued periodically for a yr, during which Ding spent a number of months in China working for Beijing Rongshu Lianzhi Technology. This start-up tech company approached him and offered a monthly salary of $14,800 to function their Chief Technology Officer (CTO).

Ding also allegedly founded his own AI company, Shanghai Zhisuan Technology.

US Attorney General Merrick Garland stated, “The Justice Department won’t tolerate the theft of artificial intelligence and other advanced technologies that might put our national security in danger.” FBI Director Christopher Wray added that Ding’s alleged actions “are the newest illustration of the lengths” firms in China will go to “to steal American innovation.”

As the world grapples with AI’s transformative potential, insider controversies in tech firms threaten to fire up dissent. 

Cases at Microsoft and Google highlight the importance of fostering a culture of responsible innovation, including trust and transparency throughout the company itself.

AI is a technology that demands trust, and tech firms need to supply more assurance. This hasn’t all the time been forthcoming.

For instance, a bunch of 100+ tech experts recently co-signed a letter pleading with AI firms to open their doors to independent testing. 

They argued that tech firms are too secretive about their products except when their hand is forced, as we saw when Google pulled Gemini’s image generation model for creating bizarre, historically inaccurate images. 

Right now, plainly AI’s exceptional pace of development often leaves trust and safety in its wake.

The post AI-related insider controversies hit each Microsoft and Google appeared first on DailyAI.

This article was originally published at