WEEKLY AI NEWS: RESEARCH, NEWS, RESOURCES, AND PERSPECTIVES

AI & ML news: Week 21–28 April

AI can soon edit the DNA, NVIDIA and Apple starts the great consolidation, and much more

Salvatore Raieli
17 min readApr 29, 2024
Photo by Priscilla Du Preez 🇨🇦 on Unsplash

The most interesting news, repository, articles, and resources of the week

Check and star this repository where the news will be collected and indexed:

You will find the news first in GitHub. Single posts are also collected here:

Weekly AI and ML news - each week the best of the field

49 stories

Research

https://www.robots.ox.ac.uk/~vgg/research/flowsam/
https://github.com/bluedyee/tf-gph
https://ollama.com/blog/llama-3-is-not-very-censored
https://arxiv.org/pdf/2404.12358
https://arxiv.org/pdf/2404.13046v1

News

https://arxiv.org/pdf/2404.14219
  • Llama 3 is not very censored. Llama 3 feels significantly less censored than its predecessor. The Llama 3 models have substantially lower false refusal rates, with less than 1⁄3 the number of false refusals when compared to Llama 2, making it possible to discuss a wider range of interesting topics!
  • OpenAI’s GPT-4 can exploit real vulnerabilities by reading security advisories. Researchers have shown that OpenAI’s GPT-4 model outperforms other models and tools like vulnerability scanners, with an 87% success rate in autonomously exploiting security vulnerabilities listed in CVE advisories.
  • US Air Force confirms first successful AI dogfight. The US Air Force is putting AI in the pilot’s seat. In an update on Thursday, the Defense Advanced Research Projects Agency (DARPA) revealed that an AI-controlled jet successfully faced a human pilot during an in-air dogfight test carried out last year.
https://www.tomshardware.com/pc-components/cpus/intel-completes-assembly-of-first-commercial-high-na-euv-chipmaking-tool-as-it-preps-for-14a-process
https://multibooth.github.io/
  • Gurman: Apple Working on On-Device LLM for Generative AI Features. Writing in his “Power On” newsletter, Gurman said that Apple’s LLM underpins upcoming generative AI features. “All indications” apparently suggest that it will run entirely on-device, rather than via the cloud like most existing AI services.
  • Los Angeles is using AI in a pilot program to try to predict homelessness and allocate aid. In Los Angeles, the Homelessness Prevention Program uses predictive AI to identify individuals and families at risk of becoming homeless, offering aid to help them get stabilized and remain housed.
  • Startup Uses AI To Edit Human Data. A team of researchers at a Berkeley-based startup called Profluent say they’ve used generative AI technologies to edit human DNA. As the New York Times reports, the startup fed huge amounts of biological data into a large language model (LLM) to come up with new editors based on the groundbreaking gene-editing technique CRISPR, as detailed in a yet-to-be-peer-reviewed paper.
  • Apple releases OpenELM: small, open source AI models designed to run on-device. Just as Google, Samsung and Microsoft continue to push their efforts with generative AI on PCs and mobile devices, Apple is moving to join the party with OpenELM, a new family of open-source large language models (LLMs) that can run entirely on a single device rather than having to connect to cloud servers.
  • Eric Schmidt-backed Augment, a GitHub Copilot rival, launches out of stealth with $252M. In a recent StackOverflow poll, 44% of software engineers said that they use AI tools as part of their development processes now and 26% plan to soon. Gartner estimates that over half of organizations are currently piloting or have already deployed AI-driven coding assistants and that 75% of developers will use coding assistants in some form by 2028.
  • Sakana releases Japanese image model. a high-speed image generation model optimized for Japanese language prompts
  • Generative A.I. Arrives in the Gene Editing World of CRISPR. Much as ChatGPT generates poetry, a new A.I. system devises blueprints for microscopic mechanisms that can edit your DNA. Generative A.I. technologies can write poetry and computer programs or create images of teddy bears and videos of cartoon characters that look like something from a Hollywood movie. Now, new A.I. technology is generating blueprints for microscopic biological mechanisms that can edit your DNA, pointing to a future when scientists can battle illness and diseases with even greater precision and speed than they can today.
  • FlexAI Launches with $30 Million in Seed Funding to Deliver Universal AI Compute. Ex-Apple, Intel, NVIDIA, and Tesla veterans rearchitect compute infrastructure to accelerate AI innovation. FlexAI, the universal AI compute company, today launched with $30 million (€28.5 million) in seed funding led by Alpha Intelligence Capital (AIC), Elaia Partners, and Heartcore Capital.
  • Report: Google will update Gemini Nano in time for Galaxy S25. Google’s Gemini AI models are constantly advancing, so it comes as no surprise that a new report claims Google will have a “version 2” of Gemini Nano available by the time the Galaxy S25 launches next year.
https://www.anthropic.com/research/probes-catch-sleeper-agents

Resources

  • Fine-tune Llama 3 with ORPO. ORPO is a new exciting fine-tuning technique that combines the traditional supervised fine-tuning and preference alignment stages into a single process. This reduces the computational resources and time required for training. Moreover, empirical results demonstrate that ORPO outperforms other alignment methods on various model sizes and benchmarks.
  • Mistral Common. Mistral-common is a set of tools to help you work with Mistral models. Our first release contains tokenization. Our tokenizers go beyond the usual text <-> tokens, adding parsing of tools and structured conversation. We also release the validation and normalization code that is used in our API.
  • LongEmbed. This repository is the official implementation for the paper “LongEmbed: Extending Embedding Models for Long Context Retrieval”
https://www.nytimes.com/2024/04/22/technology/generative-ai-gene-editing-crispr.html?unlocked_article_code=1.mk0.JQS0.P95fZ2M-SfYp
  • FineWeb: 15T high quality web tokens. 15T tokens were used to train the most recent Llama 3 models. This new dataset yields high-quality models and includes a large deduplicated corpus from the common crawl.
  • A Visual Guide to Vision Transformers. This is a visual guide to Vision Transformers (ViTs), a class of deep learning models that have achieved state-of-the-art performance on image classification tasks. This guide will walk you through the key components of Vision Transformers in a scroll story format, using visualizations and simple explanations to help you understand how these models work and what the flow of the data through the model looks like.
  • The Cauldron VLM data. 50 language and vision datasets merged into a single format to enable better model training.
  • MAexpA Generic Platform for RL-based Multi-Agent Exploration. MAexp, a generic high-efficiency platform designed for multi-agent exploration, encompassing a diverse range of scenarios and MARL algorithms.
https://github.com/hustvl/mim4d
https://github.com/dwzhu-pku/LongEmbed
  • LLaMA3-Quantization. Given the wide application of low-bit quantization for LLMs in resource-limited scenarios, we explore LLaMa3’s capabilities when quantized to low bit-width. This exploration holds the potential to unveil new insights and challenges for low-bit quantization of LLaMa3 and other forthcoming LLMs, especially in addressing performance degradation problems that suffer in LLM compression.
  • Instructor: Structured LLM Outputs. Instructor is a Python library that makes it a breeze to work with structured outputs from large language models (LLMs). Built on top of Pydantic, it provides a simple, transparent, and user-friendly API to manage validation, retries, and streaming responses. Get ready to supercharge your LLM workflows!
  • How does ChatGPT work? As explained by the ChatGPT team. Sometimes the best explanations of how a technology solution works come from the software engineers who built it. To explain how ChatGPT (and other large language models) operate, I turned to the ChatGPT engineering team.
  • BitBLAS. A collection of GPU-accelerated kernels for BitNet-style model training has been made available by Microsoft. These devices offer a significant reduction in memory usage without sacrificing much accuracy.
https://arxiv.org/pdf/2404.15622v1
  • CoreNet: A library for training deep neural networks. CoreNet is a deep neural network toolkit from Apple that allows researchers and engineers to train standard and novel small and large-scale models for variety of tasks, including foundation models (e.g., CLIP and LLM), object classification, object detection, and semantic segmentation.
  • MaxText. MaxText is a high-performance, highly scalable, open-source LLM written in pure Python/Jax and targeting Google Cloud TPUs and GPUs for training and inference. MaxText achieves high MFUs and scales from single hosts to very large clusters while staying simple and “optimization-free” thanks to the power of Jax and the XLA compiler.
  • Cohere Toolkit. A chat interface with numerous useful capabilities for creating AI-powered chat apps has been made available by Cohere.
https://github.com/duangzhu/maexp
  • BAAI/Bunny-Llama-3–8B-V. Bunny is a family of lightweight but powerful multimodal models. It offers multiple plug-and-play vision encoders, like EVA-CLIP, SigLIP, and language backbones, including Llama-3–8B, Phi-1.5, StableLM-2, and Phi-2. To compensate for the decrease in model size, we construct more informative training data by curated selection from a broader data source.
  • Finetune Llama 3–2x faster + 6x longer context + 68% less VRAM. 6x long context length with dramatically less VRAM usage than HF with flash attention.

Perspectives

  • Self-Reasoning Tokens, teaching models to think ahead. This paper presents “reasoning tokens” for language models, which produce more tokens intended to forecast future tokens instead of the one that is immediately next, improving the model’s anticipatory capacity. Experiments show notable increases in prediction accuracy, indicating that more sophisticated reasoning may be possible without the need for explicit step-by-step training.
  • Looking for AI use-cases. This article explores the potential for transformation and the existing constraints of generative AI, such as ChatGPT. It points out that although ChatGPT performs well on simple tasks like coding and creating drafts, it has trouble with more complicated tasks that call for specialized programming. It emphasizes the necessity of a vision that links AI solutions with useful applications and stresses how difficult it is to find and incorporate these into regular workflows.
https://github.com/Suyimu/WRV2
  • Building reliable systems out of unreliable agents. Although AI agents aren’t always dependable, they can be used to create dependable systems. A few strategies are to start with basic prompts and build an iterative improvement evaluation system; to deploy with observability; to use Retrieval Augmented Generation (RAG); to think about fine-tuning the model; and to use complementary agents to strengthen each other’s weaknesses and increase the overall reliability of the system.
  • AI leads a service-as-software paradigm shift. Many VCs are talking about AI taking a bite out of the services business. Foundation Capital believes there is $4.6 trillion worth of work to be automated, thanks to AI: both for in-house functions and outsourced services. We’re entering the era of Service-as-Software.
  • How AI is improving climate forecasts. Researchers are using various machine-learning strategies to speed up climate modeling, reduce its energy costs and hopefully improve accuracy.
  • Will AI accelerate or delay the race to net-zero emissions? As artificial intelligence transforms the global economy, researchers need to explore scenarios to assess how it can help, rather than harm, the climate.
https://arxiv.org/pdf/2404.15141v1

Meme of the week

What do you think about it? Some news that captured your attention? Let me know in the comments

If you have found this interesting:

You can look for my other articles, and you can also connect or reach me on LinkedIn. Check this repository containing weekly updated ML & AI news. I am open to collaborations and projects and you can reach me on LinkedIn. You can also subscribe for free to get notified when I publish a new story.

Here is the link to my GitHub repository, where I am collecting code and many resources related to machine learning, artificial intelligence, and more.

or you may be interested in one of my recent articles:

--

--

Salvatore Raieli

Senior data scientist | about science, machine learning, and AI. Top writer in Artificial Intelligence