Easily Migrating LLM Inference Serving from vLLM to Friendli Container