LLM Serving Engine Comparative Analysis: Friendli Inference vs. vLLM vs. TensorRT-LLM