I’m seeing the same issue; I think the problem is with HuggingFace as using the inference API on the website isn’t returning the full text either. Apparently there were some issues earlier today, this might still be a result of that: https://twitter.com/huggingface/status/1756401140495679871. Using the model locally still seems to work fine as an alternative, but I’ll keep investigating.
My current workaround is to kill the process that uses the port specified in the plugin settings with kill $(lsof -ti :$PORT) on Linux (replace $PORT with the actual port).