Vol.01 · No.10 Daily Dispatch April 24, 2026

Latest AI News

AI · PapersDaily CurationOpen Access
AI NewsResearch
7 min read

OpenAI releases GPT-5.5 for end-to-end computer work

GPT-5.5 hits ChatGPT with GPT‑5.4‑level latency and stronger coding, browsing, and analysis skills. At the same time, Google’s Gemma 4 and Alibaba’s Qwen3.6‑27B push efficient open models, while new MoE research trims training compute.

Reading Mode

One-Line Summary

Agentic AI steps into daily workflows: OpenAI’s GPT‑5.5 ships for end‑to‑end computer use, while Google’s Gemma 4 and Alibaba’s Qwen3.6‑27B advance efficient open models and MoE research shows how to scale with less compute.

LLM & SOTA Models

OpenAI launches GPT-5.5 for agentic computer use

GPT‑5.5 is designed to take messy, multi‑step tasks and drive apps, tools, and the web with less hand‑holding. OpenAI reports state‑of‑the‑art results like 82.7% on Terminal‑Bench 2.0, 58.6% on SWE‑Bench Pro, 84.9% on GDPval, and 78.7% on OSWorld‑Verified, while matching GPT‑5.4 on per‑token latency and using fewer tokens to complete the same Codex tasks. 1

In coding, OpenAI positions GPT‑5.5 as its strongest “agentic” model, highlighting better long‑horizon planning, tool use, and error recovery across real repositories and command‑line workflows. Early testers cited stronger reasoning and less implementation correction compared to GPT‑5.4, with examples of large refactors and branch merges completed in one pass. 1

For knowledge work, OpenAI cites internal adoption and concrete tasks: teams used Codex with GPT‑5.5 to analyze 24,771 K‑1 tax forms spanning 71,637 pages, cutting two weeks off the process, and achieved 98.0% on Tau2‑bench Telecom without prompt tuning alongside 60.0% on FinanceAgent. The company says more than 85% of staff use Codex weekly across functions from engineering to finance and marketing. 1

OpenAI says GPT‑5.5 rolls out in ChatGPT to Plus, Pro, Business, and Enterprise, with GPT‑5.5 Pro for Pro, Business, and Enterprise; API access follows after additional safeguards. The release is evaluated across the company’s safety and preparedness frameworks with internal/external red‑teaming and targeted testing for advanced cybersecurity and biology, with feedback from nearly 200 early‑access partners. 1

Gemma 4 raises the open baseline for on-device and workstation agents

Gemma 4 is Google DeepMind’s new family of open models meant to run on your own hardware—from phones to developer workstations—while handling advanced reasoning and agent workflows. The lineup includes Effective 2B (E2B), Effective 4B (E4B), a 26B Mixture‑of‑Experts (MoE), and a 31B dense model; Google says the 31B ranks #3 and the 26B ranks #6 among open models on Arena AI’s leaderboard as of Apr 1. The models are released under the Apache 2.0 license. 2

Beyond chat, Gemma 4 emphasizes multi‑step planning, native function calling, structured JSON output, code generation, vision (images/video) and audio inputs (E2B/E4B), and long context: 128K tokens for the edge models and up to 256K for the larger ones, trained across 140+ languages. 2

Google highlights day‑one support across common tools and runtimes (Transformers, TRL, vLLM, llama.cpp, MLX, Ollama, NVIDIA NIM/NeMo, and more) and says weights are available from Hugging Face, Kaggle, or Ollama, positioning Gemma 4 as an accessible starting point for customized, local‑first agents. 2

Qwen3.6-27B targets coding agents with a dense open-weight model

Alibaba’s Qwen team introduces Qwen3.6‑27B, a 27‑billion‑parameter dense model released as open weights under Apache 2.0 and tuned for agentic coding. Reporting cites a hybrid design that mixes Gated DeltaNet linear attention with standard attention to speed long contexts while managing memory. 3

On internal and community benchmarks, the model posts gains: 1,487 on QwenWebBench (vs. 1,068 for Qwen3.5‑27B and 1,397 for Qwen3.6‑35B‑A3B), 36.2 on NL2Repo, 77.2 on SWE‑bench Verified, and 59.3 on Terminal‑Bench 2.0—competitive with larger frontier systems on some tasks. 3

Qwen3.6‑27B adds an optional “Thinking Preservation” mode to carry reasoning traces across a conversation, offers a native 262,144‑token context window (extendable to about 1,000,000 with YaRN), and ships in BF16 and fine‑grained FP8 variants compatible with SGLang, vLLM, KTransformers, and Transformers. 4

Open Source & Repos

screenpipe: background agents that learn from your screen

screenpipe is a desktop app that bills itself as “AI memory for your screen,” running background agents that react to what you do and attempt helpful actions, like summarizing meetings or surfacing context when you switch tasks. 5

The latest release, app‑v2.4.39 (Apr 23), fixes the “Meetings summarize with AI” flow, signaling active iteration on everyday workflows rather than only demos. 5

Meanwhile, the mem0 project (a “universal memory layer for AI agents”) shows an issue proposing to add a “Sunsetting” warning to the OpenMemory README—useful context if you’re evaluating long‑term dependencies for agent memory. 6

Research Papers

Expert Upcycling: expand MoE capacity without raising per-token cost

Mixture‑of‑Experts (MoE) models split a network into many “experts” and route tokens to a few of them, increasing capacity without raising compute per token. Expert Upcycling proposes a practical recipe to grow an MoE during continued pre‑training: duplicate experts, extend the router, keep the top‑K routing fixed (so inference cost stays the same), then let training break symmetry so copies specialize. Think of it as adding seats to a workshop by cloning the best instructors, then letting them develop new specialties as classes proceed. 7

Across 7B–13B total‑parameter experiments, upcycled models match fixed‑size baselines on validation loss while saving 32% of GPU hours. The paper also introduces utility‑based expert selection—using gradient‑based importance scores to choose which experts to duplicate—tripling “gap closure” when the continued pre‑training budget is tight. 7

Complementary ICLR 2026 work argues that well‑designed MoEs can outperform dense models under strictly equal total parameters, compute, and data, and identifies a stable optimal activation rate around 20% across 2B and 7B scales. 8

Another study introduces “Efficiency Leverage,” showing how activation ratio, compute budget, and expert granularity predict MoE gains; a reference MoE‑mini matched a 6.1B dense model with only 0.85B active parameters—over 7× efficiency leverage—on downstream tasks. 9

Community Pulse

Hacker News (1041↑) — Mixed reactions: enthusiasm for capability gains but pushback about heavy guardrails and access.

"Laughed a little to this "We are releasing GPT‑5.5 with our strongest set of safeguards to date [...]" yay MORE guardrails" — Hacker News

"$30 per million output? I thought we were “democratising intelligence”?!" — Hacker News

r/OpenAI (642↑) — Similar split between excitement and frustration about pricing and who gets access first.

"Laughed a little to this "We are releasing GPT‑5.5 with our strongest set of safeguards to date [...]" yay MORE guardrails" — Reddit

"$30 per million output? I thought we were “democratising intelligence”?!" — Reddit

Hacker News (218↑) — screenpipe intrigues users as a persistent agent platform, with questions about usage reporting and calls for stronger privacy defaults.

"I love this idea. But why does it say "70 users run screenpipe 24/7!" What's being reported from the app, that they know this number?" — Hacker News

"They're building something with a lot of potential if they can get good vertical specific apps developed on top of it. I'm sure they'll build more mindful privacy features over time (or they'll have to because of various pressures). I wonder if they could build a LinkedIn prospecting tool that doesn't send spammy, fully automated messages? LinkedIn seems to flag and block accounts using other AI automation tools." — Hacker News

Why It Matters

Closed and open ecosystems are converging on the same goal: agents that can plan, use tools, and operate software end‑to‑end. GPT‑5.5 brings that experience into products many teams already use, while Gemma 4 and Qwen3.6‑27B show how far efficient, locally‑runnable models have come. 1

On the research side, MoE techniques like expert upcycling and new scaling laws explain how to add capacity and save compute, which could lower costs for enterprise deployments or local agents. That combination—stronger agents plus cheaper training and serving—signals broader, more practical adoption. 7

This Week, Try

  1. GPT‑5.5 in ChatGPT: If you have Plus/Pro/Business/Enterprise, switch the model to GPT‑5.5 and give it a multi‑step task across tools (e.g., “find sources, draft slides, and create a spreadsheet”). Details: OpenAI’s announcement. 1
  2. screenpipe desktop: Install from the GitHub releases and test the meeting‑summary flow fixed in v2.4.39 to see always‑on agents in action. 5

Sources 10

Helpful?

Comments (0)