Strategies to Optimize Large Language Model (LLM) Inference Performance