Ollama on Windows with Intel A770 WSL

Ollama on Windows with Intel A770 WSL

My main pc is a AMD Ryzen 9 7900x with a Intel A770 on Windows 11. With all the hype around AI I wanted to play around with some models locally and without dual booting. Here are the steps I used to get Ollama and Open Webui to work.

Install WSL (Use windows store, or google it) , along install docker desktop or rancher desktop

Disable the AMD Radeon Graphics , if you don't do this WSL Ubuntu instance will not recognized the A770.

Install the packages and drivers required as defined

WSL> sudo apt-get install   intel-opencl-icd   intel-level-zero-gpu level-zero   intel-media-va-driver-non-free libmfx1 libmfxgen1 libvpl2   libegl-mesa0 libegl1-mesa libegl1-mesa-dev libgbm1 libgl1-mesa-dev libgl1-mesa-dri   libglapi-mesa libgles2-mesa-dev libglx-mesa0 libigdgmm11 libxatracker2 mesa-va-drivers   mesa-vdpau-drivers mesa-vulkan-drivers va-driver-all

Restart WSL

PS C:\Users\james> wsl --shutdown

Check to see if you see the video card

PS C:\Users\james> wsl
majerus@7900x:/mnt/c/Users/james$ glxinfo | grep "OpenGL renderer string"
OpenGL renderer string: D3D12 (Intel(R) Arc(TM) A770 Graphics)
majerus@7900x:/mnt/c/Users/james$

Now deploy the container

docker run -itd --net=bridge --device=/dev/dri --privileged -p 11434:11434 -v E:\AI_Models:/root/.ollama/models -v /usr/lib/wsl:/usr/lib/wsl -e ZES_ENABLE_SYSMAN=1 -e no_proxy=localhost,127.0.0.1 --memory="32G" -e DEVICE=Arc --shm-size="16g" -e OLLAMA_HOST=0.0.0.0 -e no_proxy=localhost,127.0.0.1  --name=ipex-llm intelanalytics/ipex-llm-inference-cpp-xpu:latest bash -c "cd /llm/scripts && source ipex-llm-init --gpu --device `$DEVICE && bash start-ollama.sh && tail -f /llm/ollama/ollama.log"

Notice the `$DEVICE keeps powershell from killing command, also models path use whatever makes sense on your system.

Verify container is running

docker ps --filter "name=ipex-llm"

Start up openweb ui container

docker run -d -p 3000:8080 -e OLLAMA_BASE_URL=<MACHINEIPADDRESS>:11434 -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:main

<MACHINEIPADDRESS> should be the machine your running ollama on ie 192.168.1.42

Hit the WebUI and start playing with models.

You will need to create an account, and such which the ui will prompt you for. Once you do that you should see something similar to

Lets add a model

Get models at https://ollama.com/library

Now its time to use the model

You can pick which model you want from the drop down

Overall this was a huge mess to get working over the weekend. There are a bunch of links and things I worked on to get this to the point it is. One key detail -e OLLAMA_INTEL_GPU=true will cause this to crash. Do not enable this setting

Links:

💡
This was a quick write up. Ill add more detail as time goes on