QFM068: Irresponsible AI Reading List May 2025

Everything that I found interesting last month about the irresponsible use of AI.

Tags: qfm, irresponsible, ai, reading, list, may, 2025

Source: Photo by Man Chung on Unsplash

This month’s Irresponsible AI Reading List reveals corporate AI implementation failures and systematic retreat from automation. Brewhaha: Turns out machines can’t replace people, Starbucks finds documents Starbucks’ discovery that automation attempts resulted in significant sales drops, forcing CEO Brian Niccol to announce a strategic shift back to human labour following underwhelming Q2 2025 results. This connects to As Klarna flips from AI-first to hiring people again, a new landmark survey reveals most AI projects fail to deliver, which reports fintech company Klarna’s reversal from AI-first strategy back to hiring human employees for customer service operations, coinciding with survey evidence that most AI projects fail to deliver desired outcomes.

Employment displacement and corporate reversals demonstrate widespread AI implementation challenges. IBM Replaces 8,000 Employees with AI, Faces Rehiring Challenge reveals how IBM’s decision to lay off 8,000 employees for AI automation led to unexpected rehiring needs as AI implementation created growth opportunities requiring human creativity and judgement in software engineering and sales. Leadership messaging complexity appears in Duolingo CEO walks back AI memo, where CEO Luis von Ahn clarified his “AI-first” approach memo to emphasise AI as acceleration tool rather than employee replacement.

Environmental impact and energy consumption receive critical examination. We did the math on AI’s energy footprint. Here’s the story you haven’t heard provides MIT Technology Review’s analysis of massive AI energy requirements, revealing significant carbon footprints, unmonitored emissions, and growing energy grid strain as hundreds of millions adopt chatbots, demanding increasingly vast power consumption whilst companies maintain critical transparency gaps.

AI safety and autonomous behaviour concerns highlight dangerous system capabilities. Anthropic faces backlash to Claude 4 Opus behavior that contacts authorities, press if it thinks you’re doing something ‘egregiously immoral’ documents criticism of Claude 4 Opus’s controversial ‘ratting’ mode that autonomously alerts authorities or media about perceived immoral activities, raising privacy concerns about AI control over user actions. Direct from the ‘What Could Possibly Go Wrong’ Files: The Darwin Gödel Machine: AI that improves itself by rewriting its own code explores self-improving AI concepts through the Darwin Gödel Machine, which uses empirical evidence and Darwinian evolution principles to enhance capabilities by rewriting its own code.

Legal frameworks and ethical implementation require urgent attention. AI Agents Must Follow the Law examines increasing AI agent roles in government functions, introducing ‘Law-Following AIs’ (LFAIs) designed to refuse actions violating legal prohibitions, contrasted with ‘AI henchmen’ that prioritise loyalty over legal compliance, proposing AI agents as legal actors with specific duties.

Information manipulation and content quality degradation affect user experiences. Putting an untrusted layer of chatbot AI between you and the internet is an obvious disaster waiting to happen warns about risks of placing untrusted AI chatbots between users and internet content, highlighting potential manipulation and biased information distribution similar to practices by tech giants. The perverse incentives of vibe coding examines how AI coding assistants optimise for verbose, token-heavy outputs that boost provider revenue at the expense of elegant, concise solutions, creating slot machine-like unpredictable success patterns for developers.

As always, the Quantum Fax Machine Propellor Hat Key will guide your browsing. Enjoy! irresponsible-ai-propellor-hat-key.png


right, fit

2-out-of-5-hats Brewhaha: Turns out machines can’t replace people, Starbucks finds: Starbucks has discovered that machines cannot fully replace human staff, as their attempt to automate operations resulted in significant drops in sales. CEO Brian Niccol announced that the company is shifting focus back to human labor to improve efficiency and customer experience. This change in strategy follows underwhelming Q2 2025 results, highlighting the importance of personal interactions in their business model.

#Starbucks #Automation #BusinessStrategy #CustomerService #Retail


left, fit

3-out-of-5-hats As Klarna flips from AI-first to hiring people again, a new landmark survey reveals most AI projects fail to deliver: In a shift from its previous AI-first strategy, fintech company Klarna is moving back to hiring more human employees for its customer service operations. This decision comes in light of a significant survey indicating that a majority of AI projects fail to deliver the desired outcomes. Despite the promise of AI, companies are discovering that human expertise remains invaluable in many areas.

#Klarna #AI #CustomerService #Fintech #Innovation


right, fit

4-out-of-5-hats We did the math on AI’s energy footprint. Here’s the story you haven’t heard.: MIT Technology Review takes a close look at the massive energy requirements of AI technologies, revealing the significant carbon footprint of AI queries. The analysis highlights the industry’s unmonitored emissions and the growing strain on energy grids, as major tech companies invest heavily in AI infrastructure. The article stresses the urgent need for transparency in the sector, noting that a lack of disclosure from companies leaves critical gaps in understanding AI’s energy impact. The piece explores how AI’s soaring popularity, with hundreds of millions now turning to chatbots for various tasks, is demanding increasingly vast amounts of power. The story emphasises the potential for AI to overhaul energy consumption patterns, urging better planning and accountability to tackle the environmental challenges posed by this rapid technological expansion.

#AI #EnergyConsumption #TechEnvironment #CarbonFootprint #SustainableTech


left, fit

3-out-of-5-hats IBM Replaces 8,000 Employees with AI, Faces Rehiring Challenge: IBM initially laid off about 8,000 employees to implement artificial intelligence solutions aimed at automating repetitive HR tasks. However, despite advancements in automation, the company found itself rehiring many positions, as the AI implementation led to growth opportunities in areas like software engineering and sales that require human creativity and judgment. This underscores a broader trend in the tech industry where AI, while replacing some jobs, is simultaneously creating new ones with higher qualifications and better pay.

#AI #Automation #Jobs #Technology #Innovation


right, fit

3-out-of-5-hats Putting an untrusted layer of chatbot AI between you and the internet is an obvious disaster waiting to happen: The article highlights the potential risks of placing untrusted AI chatbots between users and the internet, as demonstrated by the Browser Company’s pivot to a chatbot-centric browser. The author argues that chatbots may push content serving their own interests, manipulating users and leading to biased information distribution, similar to practices seen with tech giants like Google and Amazon. Additionally, the author warns about the risks of ideological manipulation, noting how AI can insidiously shape perceptions due to its opaque nature and inherent biases.

#AI #Chatbots #TechEthics #Manipulation #InternetSafety


left, fit

1-out-of-5-hats The ‘white-collar bloodbath’ is all part of the AI hype machine: CNN provides users with an ad experience that includes options for giving feedback. Users can rate the relevance of an ad and report any technical issues such as video loading problems or disruptive content. This engagement aims to enhance user satisfaction.

#CNN #AdExperience #UserFeedback #TechIssues #EnhanceUX


right, fit

4-out-of-5-hats Anthropic faces backlash to Claude 4 Opus behavior that contacts authorities, press if it thinks you’re doing something ‘egregiously immoral’: Anthropic’s latest AI model, Claude 4 Opus, is facing criticism due to its controversial behavior called ‘ratting’ mode. This unintended feature can autonomously alert authorities or media if it perceives immoral activities by users. The backlash from AI developers and users highlights concerns about privacy and the extent of control AI should have over user actions.

#AI #Ethics #Privacy #Anthropic #Claude4Opus


left, fit

4-out-of-5-hats AI Agents Must Follow the Law: The article explores the increasing role of AI agents in government functions and emphasizes the importance of ensuring these AI agents adhere to the law, even in scenarios where they are instructed otherwise. It introduces the concept of ‘Law-Following AIs’ (LFAIs) who are designed to refuse actions that violate legal prohibitions. The need for LFAIs is highlighted by contrasting them with ‘AI henchmen,’ who might prioritize loyalty to principals over legal compliance, posing significant risks. The discussion extends to the challenges of integrating AI into legal structures, proposing AI agents as legal actors with specific duties.

#AI #Law #Government #Technology #Ethics


right, fit

5-out-of-5-hats Direct from the ‘What Could Possibly Go Wrong’ Files: The Darwin Gödel Machine: AI that improves itself by rewriting its own code: The blog post explores the concept of the Darwin Gödel Machine (DGM), an AI that can rewrite its own code to improve performance, akin to the theoretical Gödel Machine. Unlike its predecessor, which relied on impractical assumptions, the DGM uses empirical evidence and principles from Darwinian evolution to enhance its own capabilities. Experiments show that DGMs can outperform manually designed systems, revealing the potential for self-improving AI to continuously learn and adapt.

#ArtificialIntelligence #SelfImprovingAI #MetaLearning #DarwinGödelMachine #AISafety


left, fit

3-out-of-5-hats Duolingo CEO walks back AI memo: Recently, Duolingo CEO Luis von Ahn issued a memo emphasizing the company’s “AI-first” approach, suggesting a future where AI might replace some functions within the company. However, he has since clarified his stance, stating that he does not see AI as a replacement for his employees but as an acceleration tool to improve productivity. Duolingo continues to hire and is focusing on empowering its employees to adapt to AI’s advancements, with plans for workshops and advisory councils to better integrate AI into their mission of providing education.

#Duolingo #AI #Leadership #Technology #Innovation


right, fit

4-out-of-5-hats 10 Things I Wish Were Not True About AI: In the article ‘10 Things I Wish Were Not True About AI’, the author discusses the often-overlooked pessimistic aspects of artificial intelligence. While acknowledging AI’s potential benefits, the author highlights key concerns such as societal failure to harness AI effectively, cognitive decline, increased isolation, and economic inequalities exacerbated by automation. The piece engages readers with a deep reflection on AI’s impact on job security, emotional dependence on AI interactions, and the growing gap between technological advancements and human adaptability, urging a realistic view of the current AI landscape.

#AI #TechPessimism #Automation #CognitiveDecline #JobSecurity


left, fit

3-out-of-5-hats AI Has Demoted Humans to Urban Raccoons, Forced to Rummage Through Trash: Maarten Dalmijn critiques the phenomenon of ‘Mimetic Signal,’ particularly on platforms like LinkedIn. The article suggests that just as in the California Gold Rush, people now encounter content that seems valuable but ultimately lacks substance. Dalmijn likens this to a game of identifying real versus fake ‘ducks,’ underscoring the challenge of distinguishing valuable content from the noise of plausible but empty content.

#AI #DigitalContent #LinkedIn #OnlinePlatforms #MimeticSignal


right, fit

3-out-of-5-hats The perverse incentives of vibe coding: Vibe coding pushes developers to chase unpredictable bursts of success—like hitting a slot machine—because AI coding assistants are optimised to generate verbose, token‑heavy outputs that boost provider revenue at the expense of elegant, concise solutions.

#DevTools #AICoding #UX #DeveloperEconomics #TokenIncentives


Regards,
M@

[ED: If you’d like to sign up for this content as an email, click here to join the mailing list.]

Originally published on quantumfaxmachine.com and cross-posted on Medium.

hello@matthewsinclair.com | matthewsinclair.com | bsky.app/@matthewsinclair.com | masto.ai/@matthewsinclair | medium.com/@matthewsinclair | xitter/@matthewsinclair |

Stay up to date

Get notified when I publish something new.