Comparing two LLM serving frameworks: Friendli Inference vs. vLLM