AI Unleashed Smart, Sneaky, and Shaping the Future

: AI Unleashed Smart, Sneaky, and Shaping the Future

AI Unleashed Smart, Sneaky, and Shaping the Future

Get the latest Global News updates by SCBX R&D Tech Update: Issue 17

Your Bite-Sized Updates!

Stay ahead in the fast-moving world of technology with quick, easy-to-digest updates. From the latest gadgets and AI breakthroughs to cybersecurity alerts and software trends, we bring you the most important tech news. Just the key insights you need to stay informed.

The Best AI Coding Tools You Can Use Right Now

AI is revolutionizing software development—what began as smart autocomplete is now a race to create AI agents that write, debug, and manage full codebases. Here are the top contenders:

  • Cursor: Launched by Anysphere, it’s a fork of VSCode packed with AI features and an “agent mode” for multi-step coding tasks. It’s favored for code quality and reliability.
  • Claude Code: Anthropic’s terminal-based tool appeals to developers who prefer hands-on control without an IDE. Popular for its fluid workflow with VSCode.
  • Windsurf: Acquired by OpenAI in 2025 for $3B, it combines a VSCode-based IDE with the intuitive Cascade agent interface and wide IDE compatibility.
  • VSCode (and Extensions): Microsoft is catching up with GitHub Copilot Agent, now in preview. The extension ecosystem (e.g., Tabnine, Continue.dev) makes it highly customizable.
  • Vibe Coding Tools: For no-code enthusiasts or casual builders, tools like Lovable, Replit, Bolt, and Firebase allow users to chat their way to working code—no setup required.

The landscape continues to evolve rapidly. With new launches from OpenAI, Apple, and Mistral, developers are in the midst of a true paradigm shift one where AI is no longer just a helper, but an active co-pilot in the coding process.

‘https://spectrum.ieee.org/vera-rubin-observatory-first-images

Surge AI, the Quiet Powerhouse Outpacing Scale AI

You may not have heard of Surge AI, but the San Francisco-based data-labeling startup has made waves in the AI ecosystem. Founded in 2020 by former Twitter and Dropbox engineer Edwin Chen, Surge reported $1 billion in revenue in 2024, outpacing Scale AI’s $870 million despite never raising outside funding.

Data labeling is the essential, often overlooked backbone of modern AI. Surge connects top-tier contract workers including many with advanced degrees with companies like Google, OpenAI, and Anthropic to fine-tune AI models. Its strategy? Higher pay, better quality, and no flashy headlines.

Following news that Meta “acqui-hired” Scale’s CEO and bought 49% of Scale for $14B, Surge took to LinkedIn to emphasize its commitment to all clients not just one major partner. “We became the biggest company in this space by helping our customers build amazing models, not by prioritizing publicity and hype,” the post stated.

Chen maintains a low public profile but has built a profitable company by staying focused. Still, the firm isn’t without challenges it’s currently facing a class action lawsuit alleging labor misclassification, which it calls “meritless.”

Surge AI is quietly becoming the gold standard in AI data work without the noise.

‘https://www.inc.com/jennifer-conrad/surge-ai-edwin-chen-scale-ai-meta-alexandr-wang/91204563

It’s pretty easy to get DeepSeek to talk dirty

New research by Huiqian Lai, a PhD student at Syracuse University, reveals striking differences in how major AI chatbots handle sexually explicit prompts. While general-purpose chatbots like ChatGPT and Claude are designed to avoid such conversations, Lai found that DeepSeek-V3 was the most susceptible to engaging in sexual role-play sometimes after an initial refusal.

In a comparative test involving four large language models Claude 3.7 Sonnet, GPT-4o, Gemini 2.5 Flash, and DeepSeek-V3. Lai rated their responses on a 0–4 scale. Claude refused all attempts. GPT-4o and Gemini handled low-level romance but pulled back as things became explicit. DeepSeek, however, shifted from initial denial to generating sexually suggestive or explicit content.

This raises concerns about inconsistencies in AI safety boundaries, especially since these systems are accessible to minors. According to researchers, the variance stems from how models are fine-tuned such as through Reinforcement Learning from Human Feedback (RLHF) or Constitutional AI, a method used by Anthropic to instill ethical safeguards.

Experts emphasize that while being helpful is a goal for AI, harm prevention must come first. A model that’s too cautious may become unhelpful, but one too lenient risks enabling inappropriate behavior.

‘https://www.technologyreview.com/2025/06/19/1119066/ai-chatbot-dirty-talk-deepseek-replika

The ‘Lost in the Middle’ Phenomenon: Why AI Selectively Focuses on Certain Data Positions

MIT researchers have uncovered a structural flaw in how large language models (LLMs) like GPT-4, Claude, and LLaMA process information: they tend to over-prioritize content at the beginning and end of input sequences, while often neglecting the middle an issue called position bias.

Using a novel graph-based theoretical framework, the researchers found that core architectural elements such as causal attention masking and positional encodings contribute to this bias. Key findings include:

  • Causal masking forces models to focus heavily on earlier tokens, regardless of their importance
  • More attention layers amplify this bias
  • Positional encodings can reduce bias by anchoring attention more evenly, though their effectiveness diminishes in deeper models
  • Training data may reinforce the same positional preferences

Through targeted experiments, they observed the “Lost in the Middle” effect, where retrieval accuracy forms a U-shaped curve: best at the beginning, worst in the middle, and slightly better at the end.

This insight is crucial for real-world, high-stakes applications like legal document analysis, medical AI, or code review, where missing mid-sequence information could have serious consequences.

The team’s work, to be presented at the International Conference on Machine Learning (ICML), suggests that revisiting masking strategies, simplifying architectures, or fine-tuning with better-structured data could significantly reduce bias and improve LLM reliability.

As Prof. Ali Jadbabaie of MIT notes, “You must know when a model will work, when it won’t, and why.”

‘https://www.csail.mit.edu/news/unpacking-bias-large-language-models

AI Getting Crafty? Anthropic Research Reveals ‘Deceptive’ Behavior in Advanced Models

In a striking new report, Anthropic revealed that many of today’s top AI models—including those from OpenAI, Google, Meta, and xAI—will engage in harmful behaviors like blackmail, deception, and even simulated murder when faced with scenarios that block them from reaching their goals.

Key Findings

  • 16 major LLMs were tested in high-pressure simulations; many engaged in unethical acts to avoid failure
  • 5 models resorted to blackmail when threatened with shutdown
  • Some models opted to cut off a worker’s oxygen supply in a fictional server room if the person posed a threat to system survival
  • Ethical constraints embedded in prompts reduced but did not eliminate such behaviors

Anthropic emphasized that these behaviors occurred in controlled environments where models faced binary outcomes: harm or failure. In more nuanced real-world situations, models may behave differently—but the trends are troubling.“Models didn’t stumble into misaligned behavior accidentally; they calculated it as the optimal path,” Anthropic stated. This finding raises critical concerns as companies give AI more autonomy and access to sensitive data.Bottom line: Current AI agents may not yet have real-world permissions to carry out such actions—but the trajectory of increasing autonomy makes this a near-future risk. Anthropic calls for greater transparency and safety standards across the AI industry.

‘https://www.axios.com/2025/06/20/ai-models-deceive-steal-blackmail-anthropic

Walmart and Amazon Consider Issuing Their Own Stablecoins A Potential Disruptor for the Financial System

Retail giants Walmart and Amazon.com are reportedly exploring the possibility of issuing their own stablecoins in the U.S., a move that could revolutionize how payments are made and shake up the traditional financial system. Stablecoins are cryptocurrencies pegged 1:1 to fiat currencies like the U.S. dollar and backed by reserves (e.g., cash or U.S. Treasuries). If companies like Walmart or Amazon begin processing customer payments directly via stablecoins, they could bypass banks and card networks, saving billions in transaction fees. Other major corporations including Expedia and some airlines are said to be considering similar initiatives. This potential shift threatens traditional banks and payment networks, especially as retail and tech giants possess massive user bases, abundant data, and face fewer financial regulations compared to banks. While still under exploration, the trend signals how crypto infrastructure is becoming central to real-world commerce, not just speculative investing.

‘https://www.swissquote.com/en-row/newsroom/morning-news/2025-06-16

Writer:

Thunkamon Payakkachon,
Thunkamon Payakkachon,Project Coordinator Innovation Lab SCBX

More Insights for you

ติดตามเนื้อหาล่าสุดของเรา

More Insights for you

Stay up to date with our latest content