Hello everyone,
To learn more about MCPs, I decided to build a server to run Llama.cpp locally. Currently, it functions as an inference server. I thought others who are interested in learning how this works might find it useful. I didn't find anything similar on the web.
I used this SDK to build it: https://github.com/metoro-io/mcp-golang
Thanks!
Kevin