
LLM Agents and Evaluation: An Interview With Graham Neubig
Insights on LLM Agents, Evaluation, and Future Architectures from an Interview with Graham Neubig

LoRA: Low-Rank Adaptation for LLMs
Reduce LLM parameter counts and memory usage with LoRA: Low-Rank Adaptation for fast customization and fine-tuning of large language models
3 principles our VP of Product swears by
Empowering product teams with three key principles for making impactful decisions and building user-centric products at Duolingo

LoRA: Low-Rank Adaptation for LLMs
Efficient low-rank adaptation technique, LoRA, reduces complexity of fine-tuning large language models by approximating updates with low-rank matrices for significant memory and time savings.

How the New York Times Games Data Team Revamped Its Reporting
Enhancing data reporting for New York Times Games with revamped data architecture and dashboard suite

LoRA: Low-Rank Adaptation for LLMs
Efficiently fine-tune large language models using Low-Rank Adaptation (LoRA) to reduce trainable parameters by 10,000 times and GPU memory requirements by 3 times.

Gemma is now available on Google Cloud
Google Cloud introduces Gemma, a family of lightweight open models for AI development on Vertex AI and GKE.

Wrangle your alerts with open source Falco and the gcpaudit plugin
Enhance runtime security monitoring for Google Cloud services with Falco and gcpaudit plugin

Next generation Autocomplete is now available in Preview
Enhance user experience with new Autocomplete Preview featuring Address Validation, intuitive pricing, and expanded place types from the Places API

Orchestrate Vertex AI’s PaLM and Gemini APIs with Workflows
Orchestrate gen AI models with parallel steps using Workflows for efficient processing.

Digital exchanges achieving performance, scale, and resilience with Google Cloud
Optimizing digital exchanges for performance, scale, and resilience on Google Cloud

Advances in private training for production on-device language models
Exploring the latest advancements in private training for on-device language models, focusing on differential privacy guarantees and federated learning in Gboard production.