--- library_name: mlx license: apache-2.0 pipeline_tag: text-generation language: - en - de - es - fr - it - pt - pl - nl - tr - sv - cs - el - hu - ro - fi - uk - sl - sk - da - lt - lv - et - bg - 'no' - ca - hr - ga - mt - gl - zh - ru - ko - ja - ar - hi tags: - transformers - mlx - translation base_model: - utter-project/EuroLLM-22B-Instruct-2512 --- # mlx-community/EuroLLM-22B-Instruct-2512-mlx-bf16 The Model [mlx-community/EuroLLM-22B-Instruct-2512-mlx-bf16](https://huggingface.co/mlx-community/EuroLLM-22B-Instruct-2512-mlx-bf16) was converted to MLX format from [utter-project/EuroLLM-22B-Instruct-2512](https://huggingface.co/utter-project/EuroLLM-22B-Instruct-2512) using mlx-lm version **0.28.4**. You can find other similar translation-related MLX model quants for an Apple Mac at https://huggingface.co/bibproj 35 Languages: Bulgarian, Croatian, Czech, Danish, Dutch, English, Estonian, Finnish, French, German, Greek, Hungarian, Irish, Italian, Latvian, Lithuanian, Maltese, Polish, Portuguese, Romanian, Slovak, Slovenian, Spanish, Swedish, Arabic, Catalan, Chinese, Galician, Hindi, Japanese, Korean, Norwegian, Russian, Turkish, and Ukrainian. ## Use with mlx ```bash pip install mlx-lm ``` ```python from mlx_lm import load, generate model, tokenizer = load("mlx-community/EuroLLM-22B-Instruct-2512-mlx-bf16") prompt="Translate from English to French: Hi there!" if hasattr(tokenizer, "apply_chat_template") and tokenizer.chat_template is not None: messages = [{"role": "user", "content": prompt}] prompt = tokenizer.apply_chat_template( messages, tokenize=False, add_generation_prompt=True ) response = generate(model, tokenizer, prompt=prompt, verbose=True) ```