
MedQA: Fine-Tuning a Clinical AI on AMD ROCm — No CUDA Required
A complete walkthrough of LoRA fine-tuning Qwen3-1.7B on MedMCQA using AMD MI300X, built for the AMD Developer Hackathon on lablab.ai.
Category
13 articles in this category

A complete walkthrough of LoRA fine-tuning Qwen3-1.7B on MedMCQA using AMD MI300X, built for the AMD Developer Hackathon on lablab.ai.

OpenAI has released three new audio models through its Realtime API, enabling developers to build more sophisticated voice applications with capabilities like live speech translation and streaming transcription.

OpenAI unveils three new voice models that enable real-time reasoning, translation across 70+ languages, and live speech transcription.

Zyphra releases ZAYA1-8B, an open, efficient reasoning model with 8 billion parameters, trained on AMD Instinct MI300 GPUs, achieving competitive performance with far fewer parameters.
Mistral AI debuts remote agents in Vibe and Mistral Medium 3.5, a 128B dense model scoring 77.6% on SWE-Bench Verified, marking a significant upgrade in its coding agent ecosystem.

Nvidia releases Nemotron 3 Nano Omni, an open multimodal model capable of processing text, image, video, and audio.

Poolside AI releases Laguna M.1 and Laguna XS.2, two agentic coding models achieving 72.5% and 68.2% on SWE-bench Verified, respectively.

Meet 'Talkie', a 13B-parameter language model that paints a nostalgic picture of the future based on texts from a bygone era.

New hardware that takes advantage of sparsity in AI models could significantly reduce their energy consumption and increase performance.
NVIDIA unveils Nemotron 3 Nano Omni, a cutting-edge AI model capable of processing and understanding long-context multimodal data, including documents, audio, and video.

Xiaomi releases open source AI large language models MiMo-V2.5 and MiMo-V2.5-Pro, which are highly efficient and affordable for agentic 'claw' tasks, available under the MIT License.

Chinese AI firm DeepSeek releases a preview of V4, its long-awaited new flagship model, which can process longer prompts and is more cost-effective than its predecessors.
DeepSeek releases V4 with a 1M-token context window, competitive benchmark numbers, and innovative architecture for efficient large context length support.