VaultGemma: The world's most capable differentially private LLM

| Source: Google DeepMind Blog

Tags: VaultGemma, differential privacy, language models, Google Research, scaling laws, AI privacy

VaultGemma is a new differentially private language model developed by Google Research, emphasizing privacy in AI. It introduces scaling laws that address the trade-offs between privacy, utility, and compute costs.

Details

Google Research has introduced VaultGemma, a language model designed with differential privacy at its core. This model, trained from scratch with 1 billion parameters, aims to balance privacy and performance. The accompanying research outlines new scaling laws for differentially private language models, highlighting the trade-offs between training stability, batch size, and computational costs. VaultGemma's weights will be available on platforms like Hugging Face and Kaggle, promoting further advancements in private AI development. The research emphasizes the importance of understanding the 'noise-batch ratio' in training, which influences how well the model learns under differential privacy constraints.