記事一覧に戻る

ML team documents critical compatibility issues when fine-tuning and deploying Google's Gemma-4 model

r/MachineLearning · 2026年4月18日

ML team documents critical compatibility issues when fine-tuning and deploying Google's Gemma-4 model

AI要約

  • PEFT library fails to recognize Gemma-4's custom ClippableLinear layers, requiring manual unwrapping before LoRA attachment
  • SFTTrainer from TRL silently breaks training by hardcoding use_cache=False, corrupting KV-sharing attention—fixed in transformers v5.5.2+
  • DeepSpeed ZeRO-3 produces incomplete LoRA adapters with zero-element tensors in half the layers, making fine-tuning ineffective
  • No mature runtime LoRA serving solutions exist yet, with vLLM experiencing significant latency issues during inference

関連記事

AIニュースを毎日お届け

200以上のソースから厳選したAIニュースを毎日無料でお届けします。

無料で始める