Skip to content

Conversation

Arcs-ur
Copy link
Contributor

@Arcs-ur Arcs-ur commented Jul 21, 2025

Description

update quickstart md related to llama.cpp/ollama

#### 16. `backend buffer base cannot be NULL` error
If you meet `ggml-backend.c:96: GGML_ASSERT(base != NULL && "backend buffer base cannot be NULL") failed`, simply adding `-c xx` parameter during inference, for example `-c 1024` would resolve this problem.

#### 17. `The program was built for 1 devices[/+]` error
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

17. The program was built for 1 devices error


#### 17. `The program was built for 1 devices[/+]` error

If you meet the error `The program was built for 1 devices[/+]
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

If you meet error looks like The program was built for 1 devices. Build program log for 'Intel(R) Arc(TM) A770 Graphics':, this may be caused by the command set/export SYCL_CACHE_PERSISTENT=1. Please try unset SYCL_CACHE_PERSISTENT in the terminal. If the variable has been written into a configuration file such as ~/.bashrc, you need to manually delete or comment out the conrresponding line.

如果你遇到`ggml-backend.c:96: GGML_ASSERT(base != NULL && "backend buffer base cannot be NULL") failed`错误,在推理时传入参数`-c xx`,如`-c 1024`即可解决。

#### 17. `The program was built for 1 devices` 错误
如果遇到错误`The program was built for 1 devices[/+]
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

refer to above comment.

### Step 2: Runtime Configuration

- Open "Command Prompt" (cmd), and enter the extracted folder through `cd /d PATH\TO\EXTRACTED\FOLDER`
- To use GPU acceleration, several environment variables are required or recommended before running `llama.cpp`.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

remove this line.

### Step 2: Runtime Configuration

- Open a "Terminal", and enter the extracted folder through `cd /PATH/TO/EXTRACTED/FOLDER`
- To use GPU acceleration, several environment variables are required or recommended before running `llama.cpp`.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

remove this line.

### 步骤 2:运行时配置

- 开启一个终端,输入命令 `cd /PATH/TO/EXTRACTED/FOLDER` 进入解压缩后的文件夹。
- 要使用 GPU 加速,在运行 `llama.cpp` 之前,建议设置如下环境变量。
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

remove this line.

1. On Linux, you may run `systemctl stop ollama` to stop all ollama processes, and then rerun `ollama serve` in your current directory.
2. On Windows, you may `set OLLAMA_HOST=0.0.0.0` to ensure that the ollama commands run on the current `ollama serve`.

#### 11. Error `The program was built for 1 devices[/+]` when executing `ollama serve`
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

refer to above comment


#### 11. 执行 `ollama serve` 时报 `The program was built for 1 devices` 错误
当启动`ollama serve`或者`ollama run <model_name>`时,产生报错`The program was built for 1 devices[/+]
[+]Build program log for 'Intel(R) Arc(TM) A770 Graphics':`, 这是因为设置了`SYCL_CACHE_PERSISTENT=1`。请按照以下命令操作:
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

refer to above comment

Copy link
Contributor

@rnwang04 rnwang04 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@rnwang04 rnwang04 merged commit 951c237 into intel:main Jul 21, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants