Mistral has released Voxtral, a large language model aimed at speech recognition (ASR) applications that seek to integrate more advanced LLM-based capabilities and go beyond simple transcription.
MiniMax-M1, optimized for long-context reasoning and tool use, features a capacity of 456 billion parameters and innovative 'lightning attention' processing for efficient computation.
LLaVA-Phi leverages the Phi-2 model to deliver effective multi-modal dialogues with only 2.7B parameters, demonstrating that smaller models can achieve high performance.
We train a bilingual Arabic-Hebrew language model using a transliterated version of Arabic texts in Hebrew, ensuring both languages are represented in the same script.