GLM-5.1: The Open-Source Model That Beat GPT-5.4
GLM-5.1 (April 2026): Z.ai's 754B open-weight model scored 58.4% on SWE-bench Pro, beating GPT-5.4 and Claude Opus 4.6 on real coding benchmarks.
GLM-5.1 (April 2026): Z.ai's 754B open-weight model scored 58.4% on SWE-bench Pro, beating GPT-5.4 and Claude Opus 4.6 on real coding benchmarks.
AISLE tested 25+ AI models against Mythos's showcase vulnerabilities. A 3.6B model found the same FreeBSD flaw. Here is what the jagged frontier means.
AirLLM runs 70B LLMs on a single 4GB GPU via layer-wise inference — no quantization needed. We test the claims, measure tradeoffs, and compare alternatives.
Google DeepMind's Gemma 4 family: 31B Dense scores 89.2% on AIME 2026, 26B MoE activates 3.8B/token, E2B fits in 1.5 GB. Apache 2.0, runs on your GPU.
OpenCoder review: the Apache-2.0 code model in 1.5B and 8B variants. 83.5% HumanEval, 79.1% MBPP — a free alternative you can deploy on your own hardware.
LM Studio runs open-source LLMs locally on Windows, Mac (Apple Silicon), and Linux. Setup, GPU (CUDA/Metal/Vulkan/ROCm), model picks, and RAG in one guide.
AI SOC: how intelligent agents reshape the Security Operations Center. Alert triage, automated response, and the tooling ending the alert-fatigue era.
One email per week — courses, deep dives, tools, and AI experiments.
No spam. Unsubscribe anytime.