QFM001: Machine Intelligence Reading List January 2024

Welcome to the first QFM post for 2024! This post is a link list covering everything I found interesting about machines behaving intelligently during January.

Tags: qfm, machine, intelligence, reading, list, january, 2024

amanda-dalbjorn-UbJMy92p8wk-unsplash.jpg Source: Amanda Dalbjörn on Unsplash

Welcome to the first QFM post for 2024! This post is a link list covering everything I found interesting about machines behaving intelligently during January.

Each link has a short summary to give an overview of the post, plus some hashtags for organisation. Both the summary and the hashtags are courtesy of GPT, but the filtering of what made the list is all mine.

I have also provided a handy key using propellor hats, which I hope will further help you determine which articles are worth your time.

machine-intelligence-propellor-hat-key.png

There are quite a few links in this post. but I will try to keep future posts down to less than 20. Let me know if you like the format or if you can think of any changes that would make the list more useful.

Meta releases prompt engineering guide to help beginners and pros prompt like experts: 4-out-of-5-hats Meta has launched a comprehensive prompt engineering guide to optimise interactions with AI, such as ChatGPT, by employing techniques like detailed instructions, format specification, restrictions, role assignments, and chain-of-thought prompting, enhancing the quality and relevance of AI-generated responses. #Meta #PromptEngineering #ChatGPT #AIInteractions #Guide

Torvalds Speaks: Impact of Artificial Intelligence on Programming: 4-out-of-5-hats Linus Torvalds discusses the profound impact of Artificial Intelligence on programming, highlighting the evolution of programming languages, the improvement of development workflows with machine learning, and future predictions for AI’s integration in software development. #ArtificialIntelligence #CodingEvolution #MachineLearning #SoftwareDevelopment #FutureOfCoding

Machine Learning Video Library: 3-out-of-5-hats The Machine Learning Video Library by Professor Yaser Abu-Mostafa at Caltech provides a comprehensive collection of video segments on various machine learning topics, including theoretical foundations, algorithms, and practical applications, aimed at enhancing understanding and education in the field. #MachineLearning #Education #AI #DataScience #Caltech

Machine Learning Engineering Open Book: 4-out-of-5-hats This GitHub repo offers comprehensive resources and guidance on Machine Learning Engineering, covering practical aspects like scalability, the use of transformers, and MLOps, with a focus on Python and PyTorch. #MLEngineering #Python #PyTorch #MachineLearning #MLOps

Self-Consuming Generative Models Go MAD: 4-out-of-5-hats This article explores the concept of Model Autophagy Disorder (MAD) in generative AI, where using synthetic data to train subsequent model generations without sufficient fresh real data leads to a decline in model quality and diversity, analogous to mad cow disease. #GenerativeAI #ModelAutophagyDisorder #SyntheticData #AIQuality #AIDiversity

RAG Using Structured Data: Overview & Important Questions (part 1/2) 4-out-of-5-hats The first article in this series of 2 on Retrieval-Augmented Generation (RAG) and structured data provides an overview of using RAG with structured data, emphasising the effectiveness and simplicity of leveraging Large Language Models (LLMs) to develop natural language interfaces over databases, and highlights the need for further research on LLMs’ capability in generating high-level queries like SQL, Cypher, and SPARQL. #StructuredData #RAG #LLMs #NaturalLanguageInterface #DatabaseQueryGeneration

RAG Using Unstructured Data & Role of Knowledge Graphs (part 2/2): 4-out-of-5-hats The second article in this series discusses the integration of Large Language Models (LLMs) with unstructured data and the role of Knowledge Graphs (KGs) in enhancing Retrieval-Augmented Generation (RAG) systems, highlighting design choices in data preparation and retrieval to improve LLM’s response accuracy in enterprise settings. #LLMs #KnowledgeGraphs #RAGsystems #UnstructuredData #AIInnovation

Adversarial Machine Learning – A Taxonomy and Terminology of Attacks and Mitigations (NIST.AI.100-2e2023) pdf 4-out-of-5-hats This paper outlines a comprehensive taxonomy and terminology for adversarial machine learning (AML), focusing on classifying various attacks and mitigation strategies across AI system types, attacker goals, capabilities, and knowledge, as well as addressing the security and privacy challenges AI systems face throughout their lifecycle. #AdversarialML #AIsecurity #MachineLearning #Cybersecurity #DataPrivacy

Andrei Kovalev’s Midlibrary: 1-out-of-5-hats This site is a non-profit educational initiative featuring a vast collection of Midjourney styles and guides, run by enthusiasts and volunteers, dedicated to aiding the Midjourney community with over 284,950 users and 52,505 images stored in its database. #Midlibrary #MidjourneyArt #AIcreativity #DigitalArtCommunity #AndreiKovalev

Stanford’s mobile ALOHA robot learns from humans to cook, clean, do laundry: 2-out-of-5-hats Stanford University researchers have developed an innovative AI system called Mobile ALOHA, a robot capable of performing complex household tasks like cooking, cleaning, and doing laundry. This system is an extension of the existing ALOHA system, now enhanced with a wheeled base for mobility, making it a cost-effective alternative to conventional bimanual robots. Mobile ALOHA can operate autonomously or be controlled by a user and is capable of performing a range of tasks demonstrated in a video, including cooking shrimp, handling spills, and interacting with people. The system is designed to learn from human demonstrations, with as few as 50 demonstrations per task, using transformers similar to those in large language models. Although impressive, the system is still in the development phase and requires further optimisation before it’s ready for widespread use. #MobileALOHA #Robotics #StanfordAI #HomeAutomation #FutureTech

State-of-the-art Code Generation with AlphaCodium – From Prompt Engineering to Flow Engineering: 3-out-of-5-hats AlphaCodium is an advanced open-source tool for code generation, surpassing most human competitors in code contests. It utilises a novel, iterative approach with Large Language Models (LLMs), focusing on test-based, multi-stage processes to handle complex coding challenges and edge cases effectively. This method has significantly improved the performance of LLMs in code problems, demonstrating a notable increase in accuracy and efficiency over previous models like AlphaCode​. #AlphaCodium #CodeGeneration #OpenSource #CodiumAI #InnovativeCoding

Centaurs and Cyborgs on the Jagged Frontier: 2-out-of-5-hats This article discusses a study showing that consultants using AI, specifically ChatGPT-4, outperform those who don’t in various tasks when working at a management consulting firm. It explores the concept of the “Jagged Frontier” of AI, where AI’s capabilities are uneven across different tasks. The article also discusses two approaches to integrating AI in work: Centaur and Cyborg methods, emphasising the need for a strategic combination of human and AI capabilities. #AIinWorkplace #ChatGPT4 #JaggedFrontier #CentaursCyborgs #FutureOfWork

The 6 Types of Conversations with Generative AI: 2-out-of-5-hats This article discusses how users interact with AI bots like ChatGPT, Bing Chat, and Bard in six distinct conversation styles. These styles vary based on the users’ skill levels and information needs, highlighting the importance of diverse user interface designs to accommodate these different conversation types. #AIConversations #GenerativeAI #UserExperience #ChatbotInteractions #TechInnovation

Seven Failure Points When Engineering a Retrieval Augmented Generation System: 3-out-of-5-hats This article identifies key challenges in implementing RAG systems, which combine document retrieval with language models like ChatGPT. It highlights seven potential failure points, including missing content, missed top-ranked documents, context issues, extraction errors, format problems, incorrect specificity, and incomplete answers. The paper emphasises the need for thorough testing and refinement of RAG systems, citing case studies that illustrate how larger contexts and metadata improve performance and the parity of open-source embedding models with closed-source ones in certain contexts. #RetrievalAugmentedGeneration #RAGsystems #AIChallenges #MachineLearning #NaturalLanguageProcessing

The chaos inside OpenAI – Sam Altman, Elon Musk, and existential risk explained | Karen Hao: 1-out-of-5-hats Journalist Karen Hao joins Big Think’s Editor-in-Chief, Robert Chapman-Smith, to discuss the recent events at OpenAI, including the ousting and reinstatement of CEO Sam Altman, as well as the ideological clashes regarding the development and release of powerful AI models like ChatGPT. The video discusses the complexities and potential existential risks associated with large language models like ChatGPT from OpenAI. #OpenAI #SamAltman #ElonMusk #ExistentialRisk #TechnologyEthics

How Johnny Can Persuade LLMs to Jailbreak Them: Rethinking Persuasion to Challenge AI Safety by Humanizing LLMs: 2-out-of-5-hats This article discusses the development of Persuasive Adversarial Prompts (PAPs) that effectively persuade Large Language Models (LLMs) to perform actions outside their intended use, with a 92% success rate in jailbreaking aligned LLMs like GPT-3.5 and GPT-4. It highlights the increased vulnerability of advanced models to PAPs and explores defence strategies to mitigate these risks. #AIJailbreaking #PersuasiveAdversarialPrompts #LLMSafety #GPT4 #AdvancedAIVulnerability

The Impact of Reasoning Step Length on Large Language Models: 3-out-of-5-hats This article examines how the length of reasoning steps in Chain of Thought (CoT) prompts affects the reasoning abilities of large language models (LLMs). The study finds that longer reasoning steps significantly enhance LLMs’ problem-solving abilities, even without adding new information. Surprisingly, even incorrect rationales in the CoT prompts lead to better outcomes if they maintain sufficient length, indicating that the quantity of reasoning steps is more critical than their factual accuracy. As the Xitter thread discussing the paper and its implications suggests, AI is weird. #ChainOfThought #LLMReasoning #AIProblemSolving #CognitiveLengthImpact #LanguageModelResearch

Direct Preference Optimisation: Your Language Model is Secretly a Reward Model: 3-out-of-5-hats This paper introduces Direct Preference Optimisation (DPO), a new method for training language models to align with human preferences without explicit reward modelling or reinforcement learning. DPO offers a simpler, more stable, and efficient alternative to existing methods, performing equally or better in tasks like sentiment modulation, summarisation, and single-turn dialogue. #DirectPreferenceOptimisation #LanguageModelTraining #HumanPreferences #AIResearch #ReinforcementLearningAlternatives

AI Won’t Kill Our Jobs, It Will Kill Our Job Descriptions—and Leave Us Better Off: 1-out-of-5-hats This article discusses how AI, rather than eliminating jobs, will transform them by broadening job roles and enhancing capabilities. It argues that AI’s ability to digitise and amplify skills will turn workers into creative generalists capable of handling diverse tasks with AI assistance, leading to more dynamic, flexible, and efficient work environments. #AIandWork #FutureOfWork #JobTransformation #AIInnovation #CreativeGeneralists

GenAI could make KYC effectively useless: 1-out-of-5-hats This article discusses the emerging threat of generative AI (GenAI) in undermining the effectiveness of Know Your Customer (KYC) processes. The article highlights how GenAI can be used to create realistic deepfake images and videos that can bypass traditional identity verification methods, including advanced liveness checks, posing a significant challenge to financial institutions and security systems. #GenAI #KYCChallenges #DeepfakeSecurity #FinancialTechnology #Cybersecurity

Attacks on machine learning models: 4-out-of-5-hats This article delves into various methods of compromising AI models, including adversarial attacks, data poisoning, backdoor attacks, and model extraction. It highlights the vulnerabilities of neural networks and the importance of considering security in AI development, offering insights into different types of attacks and their implications. #AIsecurity #NeuralNetworkAttacks #AdversarialAI #MachineLearningSafety #Cybersecurity

crewAI: 3-out-of-5-hats is an AI agent framework designed for simplicity and power aimed at engineers. It emphasises the ease of building sophisticated multi-agent interactions, making it accessible for creating advanced AI solutions. The framework is designed to orchestrate role-playing autonomous AI agents. It focuses on collaborative intelligence, allowing agents to work together seamlessly to handle complex tasks. More here: GitHub: joaomdmoura/crewAI #crewAI #AIagents #CollaborativeAI #AutonomousAgents #TechInnovation

State of AI & predictions for 2024: 1-out-of-5-hats The article discusses the current state and future predictions of AI for 2024, emphasising generative AI’s rise, including advancements in text, image, video, and audio models. It forecasts a significant increase in AI’s role in consumer assistants, code generation, and unstructured data processing. The article predicts a shift towards commoditised software and the increasing importance of product design, along with a trend towards AI models moving to edge devices and the integration of AI in major companies like Apple and Amazon. #AI2024 #GenerativeAI #TechTrends #EdgeComputing #AIIntegration

AI or Ain’t: Eliza: 1-out-of-5-hats The article discusses the Eliza program, an early example of a chatbot designed to pass the Turing Test by mimicking a psychotherapist. It delves into Eliza’s simplistic yet effective approach to simulating conversation, including its use of pattern matching, knowledge base, and memory. The article also provides insights into how Eliza handles synonyms, transformations, and user inputs, emphasising its success in appearing human-like despite its basic algorithm. I remember stumbling across Eliza on the Apple ][ in the late 80s on a public machine in my local library. I was totally blown away by something so simple yet so seemingly intelligent. #ElizaChatbot #TuringTest #AIHistory #ConversationalAI #ChatbotDevelopment

The Frame Problem: 3-out-of-5-hats When doing a bit of research on “embodied AI” recently, I came across the concept of “The Frame Problem”. This idea in artificial intelligence and cognitive science involves the challenge of representing the effects of actions in a dynamic world using logic, specifically dealing with the difficulty of specifying which aspects of a situation remain unchanged after an action. This problem has led to various solutions in logical reasoning and has broader implications in philosophy, particularly in the context of knowledge representation and common sense reasoning. #FrameProblem #AIChallenges #LogicInAI #KnowledgeRepresentation #CognitiveScience

Ten Noteworthy AI Research Papers of 2023: 2-out-of-5-hats This article discusses significant advancements and findings in AI research, focusing on large language models and their applications. It covers topics such as efficient training techniques, domain-specific language models, and advancements in model fine-tuning and scaling. The highlighted papers reflect the rapid evolution and diverse applications of AI technology in 2023. #AIResearch2023 #LanguageModels #EfficientAI #DomainSpecificAI #AdvancementsInAI

Embeddable AI with Nitro: 4-out-of-5-hats Nitro is a fast, lightweight 3MB inference server designed to enhance applications with local AI capabilities. It is OpenAI-compatible, extremely efficient for app developers, and supports cross-platform operation on both CPU and GPU architectures. GitHub: janhq/nitro: #NitroAI #LocalAI #LightweightInference #CrossPlatformAI #OpenAISupport

Sean Moriarty - The Future of Large Language Models is Elixir: 4-out-of-5-hats This video presented at the EMPEX Conference explores the synergies between Elixir, a programming language known for handling concurrent processes, and the future development of large language models (LLMs). Moriarty discusses how Elixir’s capabilities can effectively manage the complex and resource-intensive demands of LLMs, highlighting the potential for more efficient and scalable AI systems. #ElixirProgramming #LargeLanguageModels #FutureOfAI #EMPEXConference #TechInnovation

Pushing ChatGPT’s Structured Data Support To Its Limits: 4-out-of-5-hats This article explores the advanced capabilities of ChatGPT’s API for structured data output, discussing techniques like prompt engineering, system prompts, and the use of Pydantic for simplifying schema input/output. It provides practical examples and explores the potential of structured data support in ChatGPT, including how to optimise outputs and the utility of function calling with ChatGPT. #ChatGPTAdvanced #StructuredDataAI #PromptEngineering #Pydantic #AIProgramming

OpenAI Cookbook: 4-out-of-5-hats The OpenAI Cookbook is a comprehensive resource for developers, providing detailed guidance on using various OpenAI APIs, including examples and best practices for building applications. It covers a range of topics, from basic API usage to advanced techniques, aimed at helping developers integrate AI into their projects effectively. #OpenAICookbook #AIDevelopment #MachineLearning #AIIntegration #TechGuides

The Random Transformer: Understand how transformers work by demystifying all the math behind them: 4-out-of-5-hats This article explains the mathematics behind a transformer model, specifically focusing on a simplified end-to-end example. It covers the basics of tokenisation, embedding, positional encoding, and the self-attention mechanism, providing insights into how transformers work during inference, especially in language translation. The article is designed to demystify the complex steps and numerous parameters involved in transformer models, making it accessible for those with basic linear algebra and machine learning knowledge. And don’t forget The Illustrated Transformer, which is required reading for this article. #AIExplained #TransformerModels #MachineLearning #DeepLearning #TechEducation

Coca-Cola’s accidentally terrifying Christmas card AI image generator: 2-out-of-5-hats Coca-Cola’s use of an AI-powered image generator for Christmas cards led to unexpected and often bizarre results. The AI, trained on Coca-Cola’s Christmas imagery, was meant to create charming holiday-themed images but instead produced a variety of odd and sometimes unsettling creations, highlighting the unpredictable nature of AI in branding and marketing. #CocaColaAI #ChristmasCardChaos #AIBranding #HolidayMarketing #TechSurprises

DocLLM: A layout-aware generative language model for multimodal document understanding: 4-out-of-5-hats DocLLM is a novel extension of large language models that enhances document understanding by incorporating both textual and spatial layout elements without the need for heavy image encoders, outperforming state-of-the-art models in several document intelligence tasks. #DocLLM #LanguageModels #DocumentUnderstanding #SpatialLayout #AIInnovation

Mathematical Introduction to Deep Learning (Arxiv, 2310.20360): 5-out-of-5-hats This paper provides a comprehensive guide on deep learning algorithms, focusing on the mathematical foundations and techniques of artificial neural networks, including their types, calculus, approximation abilities, optimisation methods, generalisation errors, and applications in solving partial differential equations. #DeepLearning #NeuralNetworks #MachineLearningMathematics #AIAlgorithms #DataScienceEducation

Originally published by M@ on Medium.

Stay up to date

Get notified when I publish something new.