
The dataset is even pre-formatted for machine learning.

The dataset is even pre-formatted for machine learning.
Nova Sonic can detect your tone.
WordPress.com launches a free AI-powered website builder
WordPress.com has launched a new AI site builder that allows anyone to create a functioning website using an AI chat-style interface.
WordPress.com has launched a new AI site builder that allows anyone to create a functioning website using an AI chat-style interface.
DeepMind: An Approach to Technical AGI Safety and Security.
We have written a paper on our approach to technical AGI safety and security. This post is primarily a copy of the extended abstract, which summarizeā¦
Artificial General Intelligence (AGI) promises transformative benefits but also presents significant risks. We develop an approach to address the risk of harms consequential enough to significantly harm humanity. We identify four areas of risk: misuse, misalignment, mistakes, and structural risks. Of these, we focus on technical approaches to misuse and misalignment. For misuse, our strategy aims to prevent threat actors from accessing dangerous capabilities, by proactively identifying dangerous capabilities, and implementing robust security, access restrictions, monitoring, and model safety mitigations. To address misalignment, we outline two lines of defense. First, model-level mitigations such as amplified oversight and robust training can help to build an aligned model. Second, system-level security measures such as monitoring and access control can mitigate harm even if the model is misaligned. Techniques from interpretability, uncertainty estimation, and safer design patterns c
An AI avatar tried to argue a case before a New York court. The judges weren't having it.
A man appearing before a New York court got a scolding from a judge after he tried to use an avatar generated by artificial intelligence to argue his case.
The Llama 4 herd: The beginning of a new era of natively multimodal AI innovation.
Weāre introducing Llama 4 Scout and Llama 4 Maverick, the first open-weight natively multimodal models with unprecedented context support and our first built using a mixture-of-experts (MoE) architecture.
- Weāre sharing the first models in the Llama 4 herd, which will enable people to build more personalized multimodal experiences.
- Llama 4 Scout, a 17 billion active parameter model with 16 experts, is the best multimodal model in the world in its class and is more powerful than all previous generation Llama models, while fitting in a single H100 GPU. Additionally, Llama 4 Scout offers an industry-leading context window of 10M and delivers better results than Gemma 3, Gemini 2.0 Flash-Lite, and Mistral 3.1 across a broad range of widely reported benchmarks.
- Llama 4 Maverick, a 17 billion active parameter model with 128 experts, is the best multimodal model in its class, beating GPT-4o and Gemini 2.0 Flash across a broad range of widely reported benchmarks, while achieving comparable results to the new DeepSeek v3 on reasoning and codingāat less than half the active parameters. Llama 4 Maverick offers a best-in-class performance to cost ratio with an experimental chat version s
The new tariff plan is confusing ā and the tech industry is scrambling to make sense of it.
Leaked data exposes a Chinese AI censorship machine
One academic who reviewed the dataset said it was "clear evidence" that China, or its affiliates, wants to use AI to improve repression.
cross-posted from: https://lemmy.sdf.org/post/31892983
TLDR:
- China has developed an Artificial Intelligence (AI) system that adds to its already powerful censorship machine, scanning content for all kinds of topics like corruption, military issues, Taiwan politics, satire
- The discovery was accidental, security researchers found an Elasticsearch database unsecured on the web, hosted by Chinese company Baidu
- Experts highlight that AI-driven censorship is evolving to make state control over public discourse even more sophisticated, especially after recent releases like China's AI model DeepSeek
A complaint about poverty in rural China. A news report about a corrupt Communist Party member. A cry for help about corrupt cops shaking down entrepreneurs.
These are just a few of the 133,000 examples fed into a sophisticat
This blog-post compares the coding capabilites of new Gemini 2.5 Pro experimental and Claude 3.7 Sonnet (thinking)
This blog-post compares the coding capabilites of new Gemini 2.5 Pro experimental and Claude 3.7 Sonnet (thinking)
Is Cursor done?
With Cline + Gemini 2.5 Pro, one can get the exact same feature set that Cursor and Windsurf provide. They only call the APIs of the big LLM Providers without an advanced secret sauce.
Itās even the opposite - they worsen model performance by limiting context size. The key advantage, the fixed monthly costs instead of variable API usage, is now gone with Gemini 2.5 Proā¦
What is left that justifies their ridiculous valuation atm?
The country poured billions into AI infrastructure, but the data center gold rush is unraveling as speculative investments collide with weak demand and DeepSeek shifts AI trends.
We surveyed 730 coders and developers about how (and how often) they use AI chatbots on the job. The results amazed and disturbed us.
The main point is that, for a free society, digital literacy matters more than ever.
The evidence-backed model delivered impressive results, but it doesnāt validate the wave of AI therapy bots flooding the market.
OpenAI's model behavior lead says OpenAI is āshifting from blanket refusals in sensitive areas to a more precise approach focused on preventing real-world harmā.
Joanne Jang leads model behavior at OpenAI. Their release of GPT-4o image generation included some notable relaxation of OpenAI's policies concerning acceptable usage - I [noted some of those](https://simonwillison.net/2025/Mar/25/introducing-4o-image-generation/) the ā¦
Gemini 2.5: Our most intelligent AI model
Gemini 2.5 is our most intelligent AI model, now with thinking.
Gemini 2.5 is our most intelligent AI model, now with thinking.
DeepSeek AI model can easily be breached for malware, security researcher Tenable warns
Tenable Research examines DeepSeek R1 and its capability to develop malware, such as a keylogger and ransomware. We found it provides a useful starting point, but requires additional prompting and debugging.
cross-posted from: https://lemmy.sdf.org/post/31583546
Security researcher Tenable successfully used DeepSeek to create a keylogger that could hide an encrypted log file on disk as well as develop a simple ransomware executable.
At its core, DeepSeek can create the basic structure for malware. However, it is not capable of doing so without additional prompt engineering as well as manual code editing for more advanced features. For instance, DeepSeek struggled with implementing process hiding. "We got the DLL injection code it had generated working, but it required lots of manual intervention," Tenable writes in its report.
**"Nonetheless, DeepSeek provides a useful compilation of techniques and search terms that can help someone with no prior experience in writing malicious code the ability to quickly famil
Trust Report DeepSeek R1: "Critical levels of risk with security and ethics, high levels of risk with privacy, stereotype, toxicity, hallucination, and fairness"
Discover the VIJIL Trust Report for DeepSeek R1, a comprehensive evaluation of security, ethics, privacy, hallucination, and performance risks in this large language model (LLM). Our analysis identifies critical security and ethical risks, high privacy vulnerabilities, and moderate hallucination ris...
cross-posted from: https://lemmy.sdf.org/post/31552333
A Trust Report for DeepSeek R1 by VIJIL, a security resercher company, indicates critical levels of risk with security and ethics, high levels of risk with privacy, stereotype, toxicity, hallucination, and fairness, a moderate level of risk with performance, and a low level of risk with robustness.