Microsoft ‘Cherry-picked’ Examples to Make its AI Seem Functional, Leaked Audio Revealed

According to a report by Business Insiders, Microsoft “cherry-picked” examples of generative AI’s output since the system would frequently “hallucinate” wrong responses. 

The intel came from a leaked audio file of an internal presentation on an early version of Microsoft’s Security Copilot a ChatGPT-like artificial intelligence platform that Microsoft created to assist cybersecurity professionals.

Apparently, the audio consists of a Microsoft researcher addressing the result of “threat hunter” testing, in which the AI examined a Windows security log for any indications of potentially malicious behaviour.

“We had to cherry-pick a little bit to get an example that looked good because it would stray and because it’s a stochastic model, it would give us different answers when we asked it the same questions,” said Lloyd Greenwald, a Microsoft Security Partner giving the presentation, as quoted by BI.

“It wasn’t that easy to get good answers,” he added.

Security Copilot

Security Copilot, like any chatbot, allows users to enter their query into a chat window and receive responses as a customer service reply. Security Copilot is largely built on OpenAI’s GPT-4 large language model (LLM), which also runs Microsoft’s other generative AI forays like the Bing Search assistant. Greenwald claims that these demonstrations were “initial explorations” of the possibilities of GPT-4 and that Microsoft was given early access to the technology.

This article has been indexed from CySecurity News – Latest Information Security and Hacking Incidents

Read the original article:

Microsoft ‘Cherry-picked’ Examples to Make its AI Seem Functional, Leaked Audio Revealed