From 708cd1d31d1dd3928365f833fcd9871094036bfc Mon Sep 17 00:00:00 2001 From: gitricko Date: Tue, 31 Mar 2026 03:13:28 -0400 Subject: [PATCH 01/13] Integrate WebAI API as fallback in configuration Added NVIDIA provider and fallback model to the OpenCLAW configuration. --- docs/webai-to-api-integration.md | 49 ++++++++++++++++++++++++++++++++ 1 file changed, 49 insertions(+) create mode 100644 docs/webai-to-api-integration.md diff --git a/docs/webai-to-api-integration.md b/docs/webai-to-api-integration.md new file mode 100644 index 0000000..06978b7 --- /dev/null +++ b/docs/webai-to-api-integration.md @@ -0,0 +1,49 @@ +# Integrate WebAPI-to-API as fallback + +sdsd + + + + + +## Example NVIDIA Build's API as fallback changes + +Only **one file** was actually changed by our work: + +**`/config/.openclaw/openclaw.json`** — edited to add the NVIDIA provider and fallback model. + +The other `.json` files (`agents/main/agent/models.json`, `sessions/sessions.json`, etc.) are runtime state files that get rewritten by the gateway automatically — they weren't touched by us. + +Here's the diff of what changed: + +```diff +--- original ++++ modified +@@ -3,7 +3,8 @@ + "agents": { + "defaults": { + "model": { +- "primary": "ollama/minimax-m2.7:cloud" ++ "primary": "ollama/minimax-m2.7:cloud", ++ "fallbacks": ["nvidia/z-ai/glm4.7"] + }, +... + "models": { + "mode": "merge", + "providers": { ++ "nvidia": { ++ "api": "openai-completions", ++ "apiKey": "nvapi-fakefakefakefake_dummydummy--q17Te7wJ5Cr2Qsz", ++ "baseUrl": "https://integrate.api.nvidia.com/v1", ++ "models": [{ ++ "id": "z-ai/glm4.7", ++ "name": "glm4.7", ++ "reasoning": true, ++ "input": ["text"], ++ "contextWindow": 128000, ++ "maxTokens": 16384, ++ "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 } ++ }] ++ }, + "ollama": { +``` From c5e52c920d364647cf2947bbc248b59cd47e465e Mon Sep 17 00:00:00 2001 From: gitricko Date: Tue, 31 Mar 2026 11:01:27 +0000 Subject: [PATCH 02/13] webai-api configuration and dockerfile --- docker-compose.yml | 15 +++++++++ docs/webai-to-api-integration.md | 58 +++++++++++++++++++++++++++++++- 2 files changed, 72 insertions(+), 1 deletion(-) diff --git a/docker-compose.yml b/docker-compose.yml index 11f1392..b5ebafc 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -41,6 +41,21 @@ services: networks: - openclaw-webtop-net + webai-api: + image: ghcr.io/leolionart/webai-to-api:latest + container_name: webai-api + restart: unless-stopped + ports: + - "6969:6969" + environment: + - PYTHONPATH=/app/src + - CONFIG_PATH=/app/data/config.conf + # volumes: + # - ./backup/config.conf:/app/data/config.conf + command: uvicorn app.main:app --host 0.0.0.0 --port 6969 --workers 1 --log-level info + networks: + - openclaw-webtop-net + networks: openclaw-webtop-net: driver: bridge diff --git a/docs/webai-to-api-integration.md b/docs/webai-to-api-integration.md index 06978b7..1d75e01 100644 --- a/docs/webai-to-api-integration.md +++ b/docs/webai-to-api-integration.md @@ -1,7 +1,63 @@ # Integrate WebAPI-to-API as fallback -sdsd +``` +Hey, i want you to add another provider called webai-api which is a proxy to Gemini-Web. Here are the set of APIs. Please configure the provider and the set of models it supports and the details of the API is below. + +To query the models it supports: http://webai-api:6969/v1/models + +And example of calling it: +curl http://webai-api:6969/v1/chat/completions \ + -H "Content-Type: application/json" \ + -d '{ + "model": "gemini-3.0-flash", + "messages": [{ "role": "user", "content": "Hello!" }] + }' + +For full api spec, you can look at. http://webai-api:6969/openapi.json + +``` +## Example of WebAI-API config + +Here is the **webai-api provider config** I added to `models.providers` in your `openclaw.json`: + +```json +"webai-api": { + "api": "openai-completions", + "apiKey": "webai-api", + "baseUrl": "http://webai-api:6969/v1", + "models": [ + { + "id": "gemini-3.0-pro", + "name": "Gemini 3.0 Pro", + "reasoning": true, + "input": ["text", "image"], + "contextWindow": 1000000, + "maxTokens": 32768, + "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 } + }, + { + "id": "gemini-3.0-flash", + "name": "Gemini 3.0 Flash", + "reasoning": false, + "input": ["text", "image"], + "contextWindow": 1000000, + "maxTokens": 32768, + "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 } + }, + { + "id": "gemini-3.0-flash-thinking", + "name": "Gemini 3.0 Flash Thinking", + "reasoning": true, + "input": ["text", "image"], + "contextWindow": 1000000, + "maxTokens": 32768, + "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 } + } + ] +} +``` +This goes under `models.providers` in your `~/.openclaw/openclaw.json` (or `/config/.openclaw/openclaw.json`). The gateway hot-reloads this automatically. From dd5928537516558ace56c1298215b064c7fa2e55 Mon Sep 17 00:00:00 2001 From: gitricko Date: Tue, 31 Mar 2026 11:18:37 +0000 Subject: [PATCH 03/13] increase worker node --- docker-compose.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docker-compose.yml b/docker-compose.yml index b5ebafc..c344e79 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -52,7 +52,7 @@ services: - CONFIG_PATH=/app/data/config.conf # volumes: # - ./backup/config.conf:/app/data/config.conf - command: uvicorn app.main:app --host 0.0.0.0 --port 6969 --workers 1 --log-level info + command: uvicorn app.main:app --host 0.0.0.0 --port 6969 --workers 3 --log-level info networks: - openclaw-webtop-net From 9c40a603287748148fe03c6812d6bab2ee8d7f18 Mon Sep 17 00:00:00 2001 From: fc_orbit Date: Thu, 2 Apr 2026 09:21:09 -0700 Subject: [PATCH 04/13] Update docs --- README.md | 4 +- docker-compose.yml | 2 +- docs/webai-to-api-integration.md | 142 ++++++++++++++++++++++++++++++- 3 files changed, 143 insertions(+), 5 deletions(-) diff --git a/README.md b/README.md index 1b16910..22aaf64 100644 --- a/README.md +++ b/README.md @@ -65,8 +65,8 @@ Perfect for: 5. Inside the WebTop desktop: - Open a terminal → `ollama signin` (sign in via the Chromium browser that pops up) ollama-signin -- Pull a model: `ollama pull kimi-k2.5:cloud` (or any model you like) -- Launch: `ollama launch openclaw --model kimi-k2.5:cloud --yes` (Use this command for fresh codespace) +- Pull a model: `ollama pull minimax-m2.7:cloud` (or any model you like) +- Launch: `ollama launch openclaw --model minimax-m2.7:cloud --yes` (Use this command for fresh codespace) - (If there is errors/After first launch) `openclaw gateway run` or `openclaw gateway restart` - Finally: `openclaw dashboard` → copy the tokenized URL diff --git a/docker-compose.yml b/docker-compose.yml index c344e79..b5c84d6 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -3,7 +3,7 @@ name: openclaw-webtop services: openclaw-webtop: # Uses the specific Ubuntu-based MATE tag from LinuxServer.io - image: ${DOCKER_URI:-ghcr.io/gitricko/openclaw-webtop:latest} + image: ${DOCKER_URI:-ghcr.io/gitricko/openclaw-webtop:d081941f58c0b8aaf03fbf563c6fd5bbe08626b9} container_name: openclaw-webtop # Optional: Needed for some modern GUI apps to function properly on older hosts/kernels diff --git a/docs/webai-to-api-integration.md b/docs/webai-to-api-integration.md index 1d75e01..6383ddb 100644 --- a/docs/webai-to-api-integration.md +++ b/docs/webai-to-api-integration.md @@ -1,7 +1,7 @@ # Integrate WebAPI-to-API as fallback ``` -Hey, i want you to add another provider called webai-api which is a proxy to Gemini-Web. Here are the set of APIs. Please configure the provider and the set of models it supports and the details of the API is below. +Hey, i want you to add another provider called webai-api which is a proxy to Gemini-Web. Here are the set of APIs. Please configure the provider and the set of models it supports. The details of the API is below. To query the models it supports: http://webai-api:6969/v1/models @@ -15,7 +15,12 @@ curl http://webai-api:6969/v1/chat/completions \ For full api spec, you can look at. http://webai-api:6969/openapi.json +I want you test API to ensure it works as well as all the models it supports too +Note: like ollama, there is no API Key needed for webai-api, so configure it accordingly ``` + + + ## Example of WebAI-API config Here is the **webai-api provider config** I added to `models.providers` in your `openclaw.json`: @@ -61,7 +66,6 @@ This goes under `models.providers` in your `~/.openclaw/openclaw.json` (or `/con - ## Example NVIDIA Build's API as fallback changes Only **one file** was actually changed by our work: @@ -103,3 +107,137 @@ Here's the diff of what changed: + }, "ollama": { ``` + + + +No git repo here, but I can show you exactly what changed. I made **two edits** to `openclaw.json`: + +--- + +### 1. Added new provider `nvidia` (inside `models.providers`) + +```json +"nvidia": { + "api": "openai-completions", + "apiKey": "nvapi-3G-DTG1nfakey-fakey-mIVebaPva22_EYU", + "baseUrl": "https://integrate.api.nvidia.com/v1", + "models": [ + { + "id": "z-ai/glm4.7", + "name": "GLM-4.7 (NVIDIA)", + "input": ["text"], + "contextWindow": 128000, + "maxTokens": 16384, + "reasoning": true, + "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 } + } + ] +} +``` + +--- + +### 2. Updated `agents.defaults.model` (was just `primary`, added `fallbacks`) + +**Before:** +```json +"defaults": { + "model": { + "primary": "ollama/minimax-m2.7:cloud" + }, +``` + +**After:** +```json +"defaults": { + "model": { + "primary": "ollama/minimax-m2.7:cloud", + "fallbacks": ["nvidia/z-ai/glm4.7"] + }, +``` + +--- + +```markdown +## Model Configuration Changes + +### Overview + +| Setting | Value | +|---|---| +| Primary model | `ollama/minimax-m2.7:cloud` | +| Fallback model | `nvidia/z-ai/glm4.7` | +| Provider API | OpenAI Completions compatible | +| Base URL | `https://integrate.api.nvidia.com/v1` | +| Model | Zhipu AI GLM-4.7 | +| Reasoning | Enabled | +| Context window | 128,000 tokens | +| Max output | 16,384 tokens | + +--- + +### Changes to `openclaw.json` + +#### 1. Added `models.providers.nvidia` + +```json +"nvidia": { + "api": "openai-completions", + "apiKey": "nvapi-fakefakefake-s18QUOUMexmIVebaPva22_EYU", + "baseUrl": "https://integrate.api.nvidia.com/v1", + "models": [ + { + "id": "z-ai/glm4.7", + "name": "GLM-4.7 (NVIDIA)", + "input": ["text"], + "contextWindow": 128000, + "maxTokens": 16384, + "reasoning": true, + "cost": { + "input": 0, + "output": 0, + "cacheRead": 0, + "cacheWrite": 0 + } + } + ] +} +``` + +#### 2. Updated `agents.defaults.model` + +```diff +"defaults": { + "model": { + "primary": "ollama/minimax-m2.7:cloud", ++ "fallbacks": ["nvidia/z-ai/glm4.7"] + }, +``` + +--- + +### Failover Behavior + +When the primary model (`ollama/minimax-m2.7:cloud`) fails due to: +- Auth errors +- Rate limits +- Timeouts + +OpenClaw automatically switches to the fallback (`nvidia/z-ai/glm4.7`). + + +### How to install OpenClaw without starting from Ollama launch + +src: https://github.com/ollama/ollama/blob/main/cmd/launch/openclaw.go + +``` +openclaw onboard --non-interactive --accept-risk \ +--auth-choice ollama \ +--custom-base-url http://127.0.0.1:11434 \ +--custom-model-id ollama/minimax-m2.7:cloud \ +--skip-channels --skip-skills \ +--skip-health + +Or —install-daemon +``` + From 90b29fab992f47b07d66a4ad0e46b6e39f2bf81b Mon Sep 17 00:00:00 2001 From: gitricko Date: Sat, 4 Apr 2026 22:09:21 -0400 Subject: [PATCH 05/13] dev --- docs/webai-to-api-integration.md | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/docs/webai-to-api-integration.md b/docs/webai-to-api-integration.md index 6383ddb..a4c4a0b 100644 --- a/docs/webai-to-api-integration.md +++ b/docs/webai-to-api-integration.md @@ -2,6 +2,7 @@ ``` Hey, i want you to add another provider called webai-api which is a proxy to Gemini-Web. Here are the set of APIs. Please configure the provider and the set of models it supports. The details of the API is below. +Hey, i want you to add another provider called webai-api which is a proxy to Gemini-Web. Here are the set of APIs. Please configure the provider and the set of models it supports. The details of the API is below. To query the models it supports: http://webai-api:6969/v1/models @@ -21,6 +22,9 @@ Note: like ollama, there is no API Key needed for webai-api, so configure it acc + + + ## Example of WebAI-API config Here is the **webai-api provider config** I added to `models.providers` in your `openclaw.json`: From cc8a22ace338306dca05103e6a3fccc1d510345c Mon Sep 17 00:00:00 2001 From: gitricko Date: Sat, 4 Apr 2026 22:09:57 -0400 Subject: [PATCH 06/13] dev --- docs/webai-to-api-integration.md | 1 - 1 file changed, 1 deletion(-) diff --git a/docs/webai-to-api-integration.md b/docs/webai-to-api-integration.md index a4c4a0b..ad68260 100644 --- a/docs/webai-to-api-integration.md +++ b/docs/webai-to-api-integration.md @@ -2,7 +2,6 @@ ``` Hey, i want you to add another provider called webai-api which is a proxy to Gemini-Web. Here are the set of APIs. Please configure the provider and the set of models it supports. The details of the API is below. -Hey, i want you to add another provider called webai-api which is a proxy to Gemini-Web. Here are the set of APIs. Please configure the provider and the set of models it supports. The details of the API is below. To query the models it supports: http://webai-api:6969/v1/models From b935928f5677adb163446c7b0a9a127856ae58f9 Mon Sep 17 00:00:00 2001 From: gitricko Date: Sun, 12 Apr 2026 12:52:46 -0400 Subject: [PATCH 07/13] Final version for the documentation --- Makefile | 5 ++++- docker-compose.yml | 4 ++-- docker/Dockerfile | 4 ++-- docs/webai-to-api-integration.md | 19 +++++++++++++------ 4 files changed, 21 insertions(+), 11 deletions(-) diff --git a/Makefile b/Makefile index 883fcae..11dc0a3 100644 --- a/Makefile +++ b/Makefile @@ -26,9 +26,12 @@ start: PUID=$(shell id -u) \ PGID=$(shell id -g) \ docker compose up -d - # while ! curl -s --head http://localhost:18789 | head -n 1 | grep -q "200 OK"; do echo; sleep 5; done; docker compose logs -f +start-webai-api: + docker compose --profile webai-api up -d + docker compose --profile webai-api logs -f + stop: docker compose down diff --git a/docker-compose.yml b/docker-compose.yml index b5c84d6..8999b92 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -47,11 +47,11 @@ services: restart: unless-stopped ports: - "6969:6969" + profiles: + - webai-api environment: - PYTHONPATH=/app/src - CONFIG_PATH=/app/data/config.conf - # volumes: - # - ./backup/config.conf:/app/data/config.conf command: uvicorn app.main:app --host 0.0.0.0 --port 6969 --workers 3 --log-level info networks: - openclaw-webtop-net diff --git a/docker/Dockerfile b/docker/Dockerfile index 3413e5b..b6d55d7 100644 --- a/docker/Dockerfile +++ b/docker/Dockerfile @@ -1,5 +1,5 @@ -ARG OPENCLAW_VERSION=2026.3.28 -ARG OLLAMA_VERSION=0.19.0 +ARG OPENCLAW_VERSION=2026.4.11 +ARG OLLAMA_VERSION=0.20.4 ARG NODE_VERSION=24 # Use the official Ollama image to get the binary diff --git a/docs/webai-to-api-integration.md b/docs/webai-to-api-integration.md index ad68260..a2d06b6 100644 --- a/docs/webai-to-api-integration.md +++ b/docs/webai-to-api-integration.md @@ -1,5 +1,13 @@ # Integrate WebAPI-to-API as fallback +## Configure WebAPI-API + +0. Assuming the initial setup is done +1. Start WebAPI-API service via `make start-webai-api` (this service is not started automatically) +2. In WebTop, open a chrome and navigate to this URI: http://webai-api:6969 +3. Click on configuration tab and copy the cookie from the signin gemini page from chrome's inspector. Click `Import Cookies` +4. You should see "Cookies imported and client connected successsfully!" +5. Go to openclaw webui and copy and paste below prompt to get openclaw to configure itself. ``` Hey, i want you to add another provider called webai-api which is a proxy to Gemini-Web. Here are the set of APIs. Please configure the provider and the set of models it supports. The details of the API is below. @@ -15,14 +23,13 @@ curl http://webai-api:6969/v1/chat/completions \ For full api spec, you can look at. http://webai-api:6969/openapi.json -I want you test API to ensure it works as well as all the models it supports too -Note: like ollama, there is no API Key needed for webai-api, so configure it accordingly +Once the new provider and its models is configured in OpenClaw, I want you to test it thoroughly via seperate agents. If this is successful, i want you to make WebAI-API as my LAST fallback model to my primary provider and model. +Note: like ollama, there is no API Key needed for webai-api, so you can give it a dummy API key and configure it accordingly ``` +6. You can ask OpenClaw to switch model and start chatting with Gemini via OpenClaw. +7. Note: This service currently does not support tools unfortunately. - - - - +--- ## Example of WebAI-API config From ea431593fa16ad003511f69ec129609073557736 Mon Sep 17 00:00:00 2001 From: gitricko Date: Sun, 12 Apr 2026 13:05:47 -0400 Subject: [PATCH 08/13] Update Makefile Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com> --- Makefile | 1 - 1 file changed, 1 deletion(-) diff --git a/Makefile b/Makefile index 11dc0a3..0e1d21f 100644 --- a/Makefile +++ b/Makefile @@ -31,7 +31,6 @@ start: start-webai-api: docker compose --profile webai-api up -d docker compose --profile webai-api logs -f - stop: docker compose down From 0fd3a96aa1827900a3549b4a4a42088d08d0cac7 Mon Sep 17 00:00:00 2001 From: gitricko Date: Sun, 12 Apr 2026 13:08:05 -0400 Subject: [PATCH 09/13] Update docs/webai-to-api-integration.md Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com> --- docs/webai-to-api-integration.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/webai-to-api-integration.md b/docs/webai-to-api-integration.md index a2d06b6..ba1c677 100644 --- a/docs/webai-to-api-integration.md +++ b/docs/webai-to-api-integration.md @@ -23,7 +23,7 @@ curl http://webai-api:6969/v1/chat/completions \ For full api spec, you can look at. http://webai-api:6969/openapi.json -Once the new provider and its models is configured in OpenClaw, I want you to test it thoroughly via seperate agents. If this is successful, i want you to make WebAI-API as my LAST fallback model to my primary provider and model. +Once the new provider and its models is configured in OpenClaw, I want you to test it thoroughly via separate agents. If this is successful, i want you to make WebAI-API as my LAST fallback model to my primary provider and model. Note: like ollama, there is no API Key needed for webai-api, so you can give it a dummy API key and configure it accordingly ``` 6. You can ask OpenClaw to switch model and start chatting with Gemini via OpenClaw. From 6376e86196737c557678654cda5af134a9530dc0 Mon Sep 17 00:00:00 2001 From: gitricko Date: Sun, 12 Apr 2026 13:08:30 -0400 Subject: [PATCH 10/13] Update docs/webai-to-api-integration.md Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com> --- docs/webai-to-api-integration.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/webai-to-api-integration.md b/docs/webai-to-api-integration.md index ba1c677..f87cc6b 100644 --- a/docs/webai-to-api-integration.md +++ b/docs/webai-to-api-integration.md @@ -6,7 +6,7 @@ 1. Start WebAPI-API service via `make start-webai-api` (this service is not started automatically) 2. In WebTop, open a chrome and navigate to this URI: http://webai-api:6969 3. Click on configuration tab and copy the cookie from the signin gemini page from chrome's inspector. Click `Import Cookies` -4. You should see "Cookies imported and client connected successsfully!" +4. You should see "Cookies imported and client connected successfully!" 5. Go to openclaw webui and copy and paste below prompt to get openclaw to configure itself. ``` Hey, i want you to add another provider called webai-api which is a proxy to Gemini-Web. Here are the set of APIs. Please configure the provider and the set of models it supports. The details of the API is below. From 8112d3334fe4b5ea2afc6a20480d2cc77f1b3c4e Mon Sep 17 00:00:00 2001 From: gitricko Date: Sun, 12 Apr 2026 13:08:45 -0400 Subject: [PATCH 11/13] Update docs/webai-to-api-integration.md Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com> --- docs/webai-to-api-integration.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/webai-to-api-integration.md b/docs/webai-to-api-integration.md index f87cc6b..af92a58 100644 --- a/docs/webai-to-api-integration.md +++ b/docs/webai-to-api-integration.md @@ -103,7 +103,7 @@ Here's the diff of what changed: "providers": { + "nvidia": { + "api": "openai-completions", -+ "apiKey": "nvapi-fakefakefakefake_dummydummy--q17Te7wJ5Cr2Qsz", ++ "apiKey": "nvapi-", + "baseUrl": "https://integrate.api.nvidia.com/v1", + "models": [{ + "id": "z-ai/glm4.7", From ff790f0fc75e4f941a3839a8c4f37b5d1183c7ee Mon Sep 17 00:00:00 2001 From: gitricko Date: Sun, 12 Apr 2026 13:16:45 -0400 Subject: [PATCH 12/13] Update docker-publish.yml to ignore certain paths Ignore markdown and docs paths for push and pull_request events --- .github/workflows/docker-publish.yml | 8 +++++++- 1 file changed, 7 insertions(+), 1 deletion(-) diff --git a/.github/workflows/docker-publish.yml b/.github/workflows/docker-publish.yml index fb70899..1cd2c6b 100644 --- a/.github/workflows/docker-publish.yml +++ b/.github/workflows/docker-publish.yml @@ -5,12 +5,18 @@ permissions: on: push: + paths-ignore: + - '**.md' + - 'docs/**' branches: [ main ] paths: - 'docker/**' - '.github/workflows/**' pull_request: + paths-ignore: + - '**.md' + - 'docs/**' branches: [ main ] paths: - 'docker/**' @@ -123,4 +129,4 @@ jobs: tags: | ghcr.io/${{ env.LOWERCASE_REPO }}:${{ env.IMAGE_TAG }} ghcr.io/${{ env.LOWERCASE_REPO }}:${{ github.sha }} - labels: ${{ steps.meta.outputs.labels }} \ No newline at end of file + labels: ${{ steps.meta.outputs.labels }} From 60b642dd788ff1af0a62b3c6e98dc9b816807b22 Mon Sep 17 00:00:00 2001 From: gitricko Date: Sun, 12 Apr 2026 13:20:40 -0400 Subject: [PATCH 13/13] Update docker-publish.yml to remove paths-ignore Removed paths-ignore for markdown and docs in push and pull_request events. --- .github/workflows/docker-publish.yml | 6 ------ 1 file changed, 6 deletions(-) diff --git a/.github/workflows/docker-publish.yml b/.github/workflows/docker-publish.yml index 1cd2c6b..75b140f 100644 --- a/.github/workflows/docker-publish.yml +++ b/.github/workflows/docker-publish.yml @@ -5,18 +5,12 @@ permissions: on: push: - paths-ignore: - - '**.md' - - 'docs/**' branches: [ main ] paths: - 'docker/**' - '.github/workflows/**' pull_request: - paths-ignore: - - '**.md' - - 'docs/**' branches: [ main ] paths: - 'docker/**'