Currently there are stubs for the vllm python backend. Add some actual vllm function to perform actions such as load a model, inference, health check, etc. For now, those 3 things are good.
There probably needs to be an agreement on what to do with the .proto spec here