1. Ollama's performance is too poor. 2. Can it support local deployment of Qwen3-Embedding? 3. Can it support local deployment of the Qwen3 LLM, which is compatible with OpenAI?