Apple has reportedly imposed restrictions on the internal use of generative AI tools such as OpenAI’s ChatGPT and GitHub Copilot, according to a recent report by The Wall Street Journal. The move is seen as an effort by Apple to safeguard its confidential data and prevent it from falling into the hands of competitors.
The concern stems from the fact that these AI models are trained on user data, and Apple wants to avoid any possibility of developers who have access to these tools using confidential information. It is worth noting that OpenAI recently launched the official ChatGPT app on iOS, raising the stakes for Apple’s data security measures. Bloomberg reporter Mark Gurman also tweeted that ChatGPT has been on Apple’s restricted software list for several months.
Apple is not the only company taking precautions in this regard. In May, Samsung banned its employees from using generative AI tools like ChatGPT after discovering three separate incidents of proprietary company data being fed into the chatbot.
Other prominent organizations, including banks such as Bank of America, Citi, Deutsche Bank, Goldman Sachs, Wells Fargo, JPMorgan, and retail giant Walmart, along with telecom company Verizon, have also imposed restrictions on their staff from accessing ChatGPT.
While Apple has not provided specific details about its own generative AI models, recent job listings indicate that the company is actively seeking talent in this area. Reports from March revealed that multiple teams at Apple, including the Siri team, are experimenting with language-generating AI.
As the race to develop advanced AI tools intensifies among tech giants, the protection of sensitive data becomes a critical concern. With Apple taking proactive measures to control the use of third-party generative AI tools internally, it signals the company’s commitment to safeguarding user information and maintaining its competitive edge in the market.