OptiLLM: An Optimizing Inference Proxy for Large Language Models