inki-NLP/opus-mt-en-de") >>> tokenizer = AutoTokenizer.from_pretrained("Helsinki-NLP/opus-mt-en-de") >>> text = "My friends are cool but they eat too many carbs." >>> input_ids = tokenizer(text, max_length=64, return_tensors="jax").input_ids >>> sequences = model.generate(input_ids, max_length=64, num_beams=2).sequences >>> outputs = tokenizer.batch_decode(sequences, skip_special_tokens=True) >>> # should give *Meine Freunde sind cool, aber sie essen zu viele Kohlenhydrate.* ``` r