I've started using #Wallabag as my replacement for Pocket, with a self-hosted backend. So far so good! It's been able to handle all the articles I've thrown at it thus far, and the setup was pretty easy, though I did have to go rooting around a bit for a sample docker-compose.yml
Theres a new version of bovine the #static #webserver for #forgejo and #gitea out. v0.1.5 introduces redis connection pool configs for performance tuning. Go check it out: https://bovine.squarecows.com/docs/ #selfhosted
Theres a new version of bovine the #static #webserver for #forgejo and #gitea out. v0.1.5 introduces redis connection pool configs for performance tuning. Go check it out: https://bovine.squarecows.com/docs/ #selfhosted
Do you know a non US hosting provider for LLM, and ML model for tool like Ollama, and llamacpp ?
i run some ML and LLM basic tool on my own server for checking update and automatically take action (source download, binary download, automatic bypass of censorship / political block) to maintain my stack up to date as EU as possible, and would like to make this "blockage from US proof"
Thanks