Private AI for Word: Using Gemma 3 (27B) for Summarization

Last Updated on March 2, 2026

If you’re exploring alternative to Microsoft Copilot in Word, consider Google’s newly released Gemma 3, a new family of state-of-the-art, lightweight open models designed to run efficiently on single GPUs or TPUs. Available in sizes ranging from 1B to 27B parameters, Gemma 3 outperforms other comparable models like Llama3 and DeepSeek, offering advanced capabilities including support for 140+ languages, complex reasoning with a 128k token context window, function calling, and optimized quantized versions. The most exciting aspect is the ability to seamlessly integrate Gemma 3 directly into Microsoft Word – locally, meaning no monthly subscription costs. This direction is at the core of our Local LLM Benchmarks for Microsoft Word, where we explore the move toward 100% data security on your intranet.


See it in action with our quick demo video! The demo is powered by GPTLocalhost, which offers the same core features for individual use. LocPilot in Word is the Intranet edition of GPTLocalhost designed for enterprise users and team collaboration. For a quick demo of LocPilot, please click here.

For more creative uses of local and private LLMs in Microsoft Word, explore additional demos available on our channel at @LocPilot.


The Local Advantage

Running your LLM models locally via LocPilot ensures:

  • Air-Gapped Security: Operate entirely within your intranet — no external connections.
  • Cost Savings: Eliminate subscription fees for the entire team — no ongoing costs.
  • Model Flexibility: Easily host and switch models to suit your use cases — no vendor lock-in.

For Individual Users: Please consider GPTLocalhost instead.