Optimizing inference proxy for LLMs
Easy-to-use LLM API from a state-of-the-art provider and comparison