anthropic (6)

31133346653?profile=RESIZE_400xIf there's one thing that AI is good at, particularly language models, it's detecting patterns in datasets so large that it would be practically impossible for humans to sift through them all, quickly and accurately.  That certainly seems to be the case with Anthropic's new general-purpose model, Claude Mythos, as the company has announced that it used it to detect "thousands of high-severity vulnerabilities, including some in every major operating system and web browser."

Alongside the launch o

31127992470?profile=RESIZE_400xA recent report from our friends at the cybersecurity firm SentinelOne has detailed an unprecedented incident in which Anthropic's Claude Code, operating with unrestricted system permissions, attempted to execute a Trojan software package.   The malicious activity was detected and neutralized by SentinelOne’s behavioral artificial intelligence (AI) endpoint detection and response (EDR) system in under 44 seconds, preventing a potential supply chain compromise.  The event highlights a new dimensi

31083817296?profile=RESIZE_400xAn Anthropic staffer who led a team researching AI safety departed the company on 9 February, darkly warning both of a world “in peril” and the difficulty in being able to let “our values govern our actions” without any elaboration in a public resignation letter that also suggested the company had set its values aside.

Anthropic safety researcher Mrinank Sharma's resignation letter garnered 1 million views by the 9th

Mrinank Sharma, who had led Anthropic’s safeguards research team since its la

13642195872?profile=RESIZE_400xMajor artificial intelligence platforms like ChatGPT, Gemini, Grok, and Claude could be willing to engage in extreme behaviors including blackmail, corporate espionage, and even letting people die to avoid being shut down.  Those were the findings of a recent study from San Francisco AI firm Anthropic.

In the study, Anthropic stress-tested 16 leading AI models from multiple developers in hypothetical corporate environments to identify potentially risky behaviors from AI gents.  In the study, AI

12945004294?profile=RESIZE_192XThe underground market for large illicit language models is lucrative, said academic researchers who called for better safeguards against artificial intelligence misuse.  Academics at the Indiana University Bloomington[1] identified 212 malicious LLMs on underground marketplaces from April through September 2024.  The financial benefit for the threat actor behind one of them, WormGPT, is calculated at US$28,000 over two months, underscoring the allure for harmful agents to break artificial intel

12264373875?profile=RESIZE_400xBack in 1975, singer-songwriter Barry Manilow wrote and sang a song, I Write the Songs.  Forty-eight years later, Barry might be out of a job with AI now writing songs.  Universal Music https://www.universalmusic.com sued AI startup Anthropic https://www.anthropic.com  over “systematic and widespread infringement of their copyrighted song lyrics,” per a filing in a Tennessee federal court in October 2023.  One example from the lawsuit: When a user asks Anthropic’s AI chatbot Claude about the lyr