Run whisper on external server.Wyoming server for faster-whisper.

First You need to install STT on  a device that can handle the speech recognition processing. This could be the  a separate server.

You must have found a suitable Whisper Container on Docker hub.

Next Integrate with Home Assistant: Use Home Assistant's  wyoming protocol Integration to send the transcribed text from STT to Home Assistant.

I could easily integrate remote rhasspy/Wyoming-Whisper on vultr to my HomeAssistant installed on a rpi4.

Note

Whisper works but it is slow (also around 15 seconds).

Would be great if it would support GPU, that would make it way faster.

Images option

https://github.com/OpenAI/whisper

https://github.com/SYSTRAN/faster-whisper

https://github.com/rhasspy/wyoming-faster-whisper == docker pull docker pull rhasspy/wyoming-whisper

Whisper Addon 

This add-on  uses wyoming-faster-whisper

This wyoming-faster-whisper  uses SYSTRAN/faster-whisper

This SYSTRAN/faster-whisper  uses OpenAI/whisper

 

Download and set up the Whisper Standalone application from this repository(https://github.com/rhasspy/wyoming-faster-whisper).

 

~# docker search wyoming-whisper

output

NAME                                   DESCRIPTION                                     STARS     OFFICIAL
rhasspy/wyoming-whisper                Wyoming protocol server for faster-whisper s…   10        
abtools/wyoming-whisper-cuda           wyoming-whisper compiled with CUDA support f…   0         
pando85/wyoming-whisper                Wyoming protocol server for faster whisper s…   0         
rhasspy/wyoming-whisper-cpp                                                            0         
dwyschka/wyoming-whisper-cuda                                                          0         
slackr31337/wyoming-whisper-gpu        https://github.com/slackr31337/wyoming-whisp…   0         
grubertech/wyoming-whisper             fast-whisper with large-v3 model built for G…   0         
confusedengineer/wyoming-whisper-gpu   Github:  https://github.com/Confused-Enginee…   0         
robomagus/wyoming-whisper-rk3588                                                       0         
clonyara/wyoming-whisper                                                               0         
xiaozhch5/wyoming-whisper    

 

Run the container

https://github.com/rhasspy/wyoming-faster-whisper

 

docker run -it -p 10300:10300 -v /path/to/local/data:/data rhasspy/wyoming-whisper \
    --model tiny-int8 --language en

 

This is work fine.

docker ps -a
CONTAINER ID   IMAGE                              COMMAND                  CREATED       STATUS        PORTS                                                     NAMES
717fe1966138   yaming116/sherpa-onnx-asr:latest   "python app.py"          2 weeks ago   Up 7 days     5001/tcp, 0.0.0.0:10700->10700/tcp, :::10700->10700/tcp   sherpa-onnx-asr
1ee50dbb7a33   rhasspy/wyoming-piper              "bash /run.sh --voic…"   2 weeks ago   Up 7 days     0.0.0.0:10200->10200/tcp, :::10200->10200/tcp             nice_hofstadter
01b832193e61   rhasspy/wyoming-openwakeword       "bash /run.sh --prel…"   2 weeks ago   Up 7 days     0.0.0.0:10400->10400/tcp, :::10400->10400/tcp             gallant_engelbart
0389595d9f22   rhasspy/wyoming-whisper            "bash /run.sh --mode…"   2 weeks ago   Up 18 hours   0.0.0.0:10300->10300/tcp, :::10300->10300/tcp             nostalgic_jang

 

docker inspect 0389595d9f22

output

docker inspect 0389595d9f22
[
    {
        "Id": "0389595d9f22b751ee07ac9380389192713979a6c7f2b3f11e48c15505055bf0",
        "Created": "2025-01-04T07:35:36.285820254Z",
        "Path": "bash",
        "Args": [
            "/run.sh",
            "--model",
            "tiny-int8",
            "--language",
            "en"
        ],
        "State": {
            "Status": "running",
            "Running": true,
            "Paused": false,
            "Restarting": false,
            "OOMKilled": false,
            "Dead": false,
            "Pid": 82152,
            "ExitCode": 0,
            "Error": "",
            "StartedAt": "2025-01-23T19:30:21.636452101Z",
            "FinishedAt": "2025-01-17T13:24:28.786242942Z"
        },
        "Image": "sha256:07c182a447fb456911f2202293b43868ef9bbbfe48aa06c4067891e2a6c2ea53",
        "ResolvConfPath": "/var/lib/docker/containers/0389595d9f22b751ee07ac9380389192713979a6c7f2b3f11e48c15505055bf0/resolv.conf",
        "HostnamePath": "/var/lib/docker/containers/0389595d9f22b751ee07ac9380389192713979a6c7f2b3f11e48c15505055bf0/hostname",
        "HostsPath": "/var/lib/docker/containers/0389595d9f22b751ee07ac9380389192713979a6c7f2b3f11e48c15505055bf0/hosts",
        "LogPath": "/var/lib/docker/containers/0389595d9f22b751ee07ac9380389192713979a6c7f2b3f11e48c15505055bf0/0389595d9f22b751ee07ac9380389192713979a6c7f2b3f11e48c15505055bf0-json.log",
        "Name": "/nostalgic_jang",
        "RestartCount": 0,
        "Driver": "overlay2",
        "Platform": "linux",
        "MountLabel": "",
        "ProcessLabel": "",
        "AppArmorProfile": "docker-default",
        "ExecIDs": null,
        "HostConfig": {
            "Binds": [
                "/path/to/local/data:/data"
            ],
            "ContainerIDFile": "",
            "LogConfig": {
                "Type": "json-file",
                "Config": {}
            },
            "NetworkMode": "default",
            "PortBindings": {
                "10300/tcp": [
                    {
                        "HostIp": "",
                        "HostPort": "10300"
                    }
                ]
            },
            "RestartPolicy": {
                "Name": "no",
                "MaximumRetryCount": 0
            },
            "AutoRemove": false,
            "VolumeDriver": "",
            "VolumesFrom": null,
            "ConsoleSize": [
                24,
                151
            ],
            "CapAdd": null,
            "CapDrop": null,
            "CgroupnsMode": "private",
            "Dns": [],
            "DnsOptions": [],
            "DnsSearch": [],
            "ExtraHosts": null,
            "GroupAdd": null,
            "IpcMode": "private",
            "Cgroup": "",
            "Links": null,
            "OomScoreAdj": 0,
            "PidMode": "",
            "Privileged": false,
            "PublishAllPorts": false,
            "ReadonlyRootfs": false,
            "SecurityOpt": null,
            "UTSMode": "",
            "UsernsMode": "",
            "ShmSize": 67108864,
            "Runtime": "runc",
            "Isolation": "",
            "CpuShares": 0,
            "Memory": 0,
            "NanoCpus": 0,
            "CgroupParent": "",
            "BlkioWeight": 0,
            "BlkioWeightDevice": [],
            "BlkioDeviceReadBps": [],
            "BlkioDeviceWriteBps": [],
            "BlkioDeviceReadIOps": [],
            "BlkioDeviceWriteIOps": [],
            "CpuPeriod": 0,
            "CpuQuota": 0,
            "CpuRealtimePeriod": 0,
            "CpuRealtimeRuntime": 0,
            "CpusetCpus": "",
            "CpusetMems": "",
            "Devices": [],
            "DeviceCgroupRules": null,
            "DeviceRequests": null,
            "MemoryReservation": 0,
            "MemorySwap": 0,
            "MemorySwappiness": null,
            "OomKillDisable": null,
            "PidsLimit": null,
            "Ulimits": null,
            "CpuCount": 0,
            "CpuPercent": 0,
            "IOMaximumIOps": 0,
            "IOMaximumBandwidth": 0,
            "MaskedPaths": [
                "/proc/asound",
                "/proc/acpi",
                "/proc/kcore",
                "/proc/keys",
                "/proc/latency_stats",
                "/proc/timer_list",
                "/proc/timer_stats",
                "/proc/sched_debug",
                "/proc/scsi",
                "/sys/firmware",
                "/sys/devices/virtual/powercap"
            ],
            "ReadonlyPaths": [
                "/proc/bus",
                "/proc/fs",
                "/proc/irq",
                "/proc/sys",
                "/proc/sysrq-trigger"
            ]
        },
        "GraphDriver": {
            "Data": {
                "LowerDir": "/var/lib/docker/overlay2/a0b16b408aa929012f30886d7a2fcfd0786637961f3a0e844ff7ebbfdf8084e9-init/diff:/var/lib/docker/overlay2/cd058de8a3e532b248d6828a93de3515596f2692b0c4bf465d0259a1c3a3878e/diff:/var/lib/docker/overlay2/bdb5a9796506747e1b2e98e4124a848f5841b35c744d11b310afe2596f858499/diff:/var/lib/docker/overlay2/a09cd2d9fc51d90ef7c321e29f014003aed701f2b16c82ca13e42cfb9d78b0c6/diff:/var/lib/docker/overlay2/8abe9e4a501483cb1fa5009290ca085015f6529e50e3a3414d90ba3a9c20524e/diff",
                "MergedDir": "/var/lib/docker/overlay2/a0b16b408aa929012f30886d7a2fcfd0786637961f3a0e844ff7ebbfdf8084e9/merged",
                "UpperDir": "/var/lib/docker/overlay2/a0b16b408aa929012f30886d7a2fcfd0786637961f3a0e844ff7ebbfdf8084e9/diff",
                "WorkDir": "/var/lib/docker/overlay2/a0b16b408aa929012f30886d7a2fcfd0786637961f3a0e844ff7ebbfdf8084e9/work"
            },
            "Name": "overlay2"
        },
        "Mounts": [
            {
                "Type": "bind",
                "Source": "/path/to/local/data",
                "Destination": "/data",
                "Mode": "",
                "RW": true,
                "Propagation": "rprivate"
            }
        ],
        "Config": {
            "Hostname": "0389595d9f22",
            "Domainname": "",
            "User": "",
            "AttachStdin": true,
            "AttachStdout": true,
            "AttachStderr": true,
            "ExposedPorts": {
                "10300/tcp": {}
            },
            "Tty": true,
            "OpenStdin": true,
            "StdinOnce": true,
            "Env": [
                "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
            ],
            "Cmd": [
                "--model",
                "tiny-int8",
                "--language",
                "en"
            ],
            "Image": "rhasspy/wyoming-whisper",
            "Volumes": null,
            "WorkingDir": "/",
            "Entrypoint": [
                "bash",
                "/run.sh"
            ],
            "OnBuild": null,
            "Labels": {}
        },
        "NetworkSettings": {
            "Bridge": "",
            "SandboxID": "0ab0205986917126b832d57fd78aa7f344900f8fa215c135662266aedba94e83",
            "HairpinMode": false,
            "LinkLocalIPv6Address": "",
            "LinkLocalIPv6PrefixLen": 0,
            "Ports": {
                "10300/tcp": [
                    {
                        "HostIp": "0.0.0.0",
                        "HostPort": "10300"
                    },
                    {
                        "HostIp": "::",
                        "HostPort": "10300"
                    }
                ]
            },
            "SandboxKey": "/var/run/docker/netns/0ab020598691",
            "SecondaryIPAddresses": null,
            "SecondaryIPv6Addresses": null,
            "EndpointID": "164817e434e5fc4c67db24dfeb63738c0982d87e715916e71ff33b7e588b7bf6",
            "Gateway": "172.17.0.1",
            "GlobalIPv6Address": "",
            "GlobalIPv6PrefixLen": 0,
            "IPAddress": "172.17.0.5",
            "IPPrefixLen": 16,
            "IPv6Gateway": "",
            "MacAddress": "02:42:ac:11:00:05",
            "Networks": {
                "bridge": {
                    "IPAMConfig": null,
                    "Links": null,
                    "Aliases": null,
                    "NetworkID": "00ceca59d76906efdaaa76c371b716ff489bed887cb9a5f023855fd076812873",
                    "EndpointID": "164817e434e5fc4c67db24dfeb63738c0982d87e715916e71ff33b7e588b7bf6",
                    "Gateway": "172.17.0.1",
                    "IPAddress": "172.17.0.5",
                    "IPPrefixLen": 16,
                    "IPv6Gateway": "",
                    "GlobalIPv6Address": "",
                    "GlobalIPv6PrefixLen": 0,
                    "MacAddress": "02:42:ac:11:00:05",
                    "DriverOpts": null
                }
            }
        }
    }
]

 

Using it in Home Assistant

The Home Assistant requires that the Wyoming Wshiper Server is exposed. You can set that in the settings.

Settings->Integrations->Add Integration: Add the wyoming protocol and add the IP of wherever you are running Wyoming Wshiper Standalone. The port is 10300

 

When  Add the wyoming protocol Integration for the first time you are asked for some information about the Wyoming Wshiper Server.

Wyoming Protocol-3.png

Once I had these containers running, I added the Whisper in the Home Assistant interface using my Whisper Docker host network’s IP address and ports 10300.

You have to simply enter the ip address  and port of a wyoming application such as  whisper. 

When you press “Submit” HA will detect the application,They then appeared as entities in the Wyoming Protocol integration.

Wyoming Protocol-1.png

 

Where model save?

STT Models are automatically downloaded from https://huggingface.co and put into /path/to/local/data.

docker run -it -p 10300:10300 -v /path/to/local/data:/data rhasspy/wyoming-whisper \

    --model medium-int8 --language en

output

config.json: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2.02k/2.02k [00:00<00:00, 49.7kB/s]
vocabulary.txt: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 460k/460k [00:00<00:00, 3.49MB/s]
model.bin: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 785M/785M [00:21<00:00, 37.1MB/s]
tokenizer.json: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2.48M/2.48M [00:01<00:00, 1.79MB/s]
INFO:__main__:Ready

 

docker run -it -p 10300:10300 -v /path/to/local/data:/data rhasspy/wyoming-whisper     --model large --language en
preprocessor_config.json: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 340/340 [00:00<00:00, 7.51kB/s]
config.json: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2.39k/2.39k [00:00<00:00, 68.2kB/s]
vocabulary.json: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 1.07M/1.07M [00:00<00:00, 6.91MB/s]
tokenizer.json: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2.48M/2.48M [00:00<00:00, 14.1MB/s]
model.bin: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 3.09G/3.09G [01:13<00:00, 42.3MB/s]
/run.sh: line 5:     7 Killed                  python3 -m wyoming_faster_whisper --uri 'tcp://0.0.0.0:10300' --data-dir /data --download-dir /data "$@"█████| 2.48M/2.48M [00:00<00:00, 14.9MB/s]

 

root@vultr:~# cd /path/to/local/data
root@vultr:/path/to/local/data# tree -L 3
.
├── en_US-amy-low.onnx
├── en_US-amy-low.onnx.json
├── en_US-amy-medium.onnx
├── en_US-amy-medium.onnx.json
├── en_US-joe-medium.onnx
├── en_US-joe-medium.onnx.json
├── en_US-lessac-high.onnx
├── en_US-lessac-high.onnx.json
├── en_US-lessac-medium.onnx
├── en_US-lessac-medium.onnx.json
├── en_US-ryan-high.onnx
├── en_US-ryan-high.onnx.json
├── models--rhasspy--faster-whisper-medium-int8
│   ├── blobs
│   │   ├── c9074644d9d1205686f16d411564729461324b75
│   │   ├── e19ba3e83e68b15d480def8185b07fc8af56caa15bb8470f2c47ec8c1faaec44
│   │   └── e86f0ed3287af742920a8ef37577c347a1a8153b
│   ├── refs
│   │   └── main
│   └── snapshots
│       └── c9af9dfadcad017757293c10e4478d42a5a06773
├── models--rhasspy--faster-whisper-tiny-int8
│   ├── blobs
│   │   ├── 6fdeb89775be8b3455197e4685fec5d59ce47874de4fb8fb7724a43ea205a7b0
│   │   ├── c9074644d9d1205686f16d411564729461324b75
│   │   └── e86f0ed3287af742920a8ef37577c347a1a8153b
│   ├── refs
│   │   └── main
│   └── snapshots
│       └── 5b6382e0f4ac867ce9ff24aaa249400a7c6c73d9
└── models--Systran--faster-whisper-large-v3
    ├── blobs
    │   ├── 0adcd01e7c237205d593b707e66dd5d7bc785d2d
    │   ├── 3a5e2ba63acdcac9a19ba56cf9bd27f185bfff61
    │   ├── 69f74147e3334731bc3a76048724833325d2ec74642fb52620eda87352e3d4f1
    │   ├── 75336feae814999bae6ccccdecf177639ffc6f9d
    │   └── 931c77a740890c46365c7ae0c9d350ba3cca908f
    ├── refs
    │   └── main
    └── snapshots
        └── edaa852ec7e145841d8ffdb056a99866b5f0a478

15 directories, 26 files

detail:

https://www.matterxiaomi.com/boards/topic/15742/how-to-manually-install-wyoming-piper-and-whisper-on-home-assistant-core/page/2#60079

 

How  it works?

Wyoming Protocol-2.png

 

Option

way 1. desktop machine

I was also curious to know if whisper add-on will run faster in a desktop machine. Since I did not have a linux server I had to install ubuntu on a VirtualBOX, exposed the virtual machine to my network, installed docker on ubuntu then downloaded whisper. I could easily integrate remote whisper to my HomeAssistant installed on a rpi4. All works as expected. So far all good.

 

I’m running this and a few other containers on

Debian 11 bullseye

AMD Ryzen 7 3800X

32 GB RAM

 

You basically need a GPU to enable accurate STT processing with those larger models. CPU processing is too slow.

 

Wyoming-Whisper in Docker with Nvidia GPU support.

https://github.com/Fraddles/Home-Automation/tree/main/Voice-Assistant

 

useful links

https://community.home-assistant.io/t/run-whisper-on-external-server/567449

https://developer.ibm.com/tutorials/run-a-single-container-speech-to-text-service-on-docker/

Comments


Comments are closed