So far, running LLMs has required a large amount of computing resources, mainly GPUs. Running locally, a simple prompt with a typical LLM takes on an average Mac ...
As per the MCP specification, ping requests should be allowed even before initialization: The client SHOULD NOT send requests other than pings before the server has responded to the initialize request ...