<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Clw-Llm-Crash on ErrorVault — Developer Error Code Dictionary</title><link>https://errorvault.dev/tags/clw-llm-crash/</link><description>Recent content in Clw-Llm-Crash on ErrorVault — Developer Error Code Dictionary</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Mon, 13 Apr 2026 22:24:49 +0800</lastBuildDate><atom:link href="https://errorvault.dev/tags/clw-llm-crash/feed.xml" rel="self" type="application/rss+xml"/><item><title>Fix clw-llm-crash: OpenClaw LLM inference crashes during model execution</title><link>https://errorvault.dev/openclaw/openclaw-clw-llm-crash-inference-failure/</link><pubDate>Mon, 13 Apr 2026 22:24:49 +0800</pubDate><guid>https://errorvault.dev/openclaw/openclaw-clw-llm-crash-inference-failure/</guid><description>&lt;h2 id="1-symptoms">1. Symptoms&lt;/h2>
&lt;p>The &lt;code>clw-llm-crash&lt;/code> error in OpenClaw manifests as an abrupt termination of the LLM inference process. Common indicators include:&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Core dump or segmentation fault&lt;/strong>: Process exits with signal 11 (SIGSEGV) during model loading or token generation.&lt;/li>
&lt;li>&lt;strong>Log output&lt;/strong>:
&lt;div class="highlight">&lt;pre tabindex="0" style="color:#f8f8f2;background-color:#282a36;-moz-tab-size:4;-o-tab-size:4;tab-size:4;">&lt;code class="language-fallback" data-lang="fallback">&lt;span style="display:flex;">&lt;span>[ERROR] clw-llm-crash: LLM backend failed at tensor allocation (line 1423, llm_engine.cpp)
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>[FATAL] GPU context lost: CUDA error 700 (cudaErrorIllegalAddress)
&lt;/span>&lt;/span>&lt;span style="display:flex;">&lt;span>Aborted (core dumped)
&lt;/span>&lt;/span>&lt;/code>&lt;/pre>&lt;/div>&lt;/li>
&lt;li>&lt;strong>High resource usage spike&lt;/strong>: VRAM usage jumps to 100% followed by OOM killer activation or driver reset.&lt;/li>
&lt;li>&lt;strong>Reproducible on specific models&lt;/strong>: Crashes consistently with quantized GGUF models (e.g., Llama-3-8B-Q4_K_M.gguf) but not unquantized ones.&lt;/li>
&lt;li>&lt;strong>Platform-specific&lt;/strong>: More frequent on NVIDIA GPUs with CUDA 11.x; AMD ROCm users report HIP kernel panics.&lt;/li>
&lt;/ul>
&lt;p>Stack traces often point to &lt;code>llm_engine.cpp&lt;/code> or &lt;code>claw_cuda_backend.cu&lt;/code> in OpenClaw source. Use &lt;code>gdb&lt;/code> or &lt;code>cuda-gdb&lt;/code> for deeper inspection:&lt;/p></description></item></channel></rss>