AI-assisted writing: LM Studio vs Microsoft Copilot in Word
Now that I’ve integrated Microsoft Word locally with several local LLM servers, it’s time to put its capabilities to the test. I’ll be benchmarking Microsoft Copilot in Word with using LM Studio in Word. In such integration, my local Word Add-in (GPTLocalhost) acts as a bridge between Microsoft Word and LM Studio. To evaluate its effectiveness, I’ll review the key features outlined in Holger’s story “Using Microsoft Copilot in Word”. Additionally, I’ll reflect on the benefits and drawbacks of my local first approach. In fact, I’m writing this story in Word right now with the assistance of the llama-3.2–3b-instruct model loaded in LM Studio.
There are three categories of use cases in Holger’s post:
(1) Draft a document
The first prompt in the post is “Write an essay about dad jokes” and I use the same prompt for a test. The following is the coherent output generated by LM Studio and shown in Microsoft Word. While subjective evaluation may be necessary depending on the context, the point here is to demonstrate that the llama-3.2–3b-instruct model is effective and can be integrated with Microsoft Word. Also, you’d have the flexibilities to experiment with different models or domain-specific models to optimize your writing.
The second example in the post is related to RAG (Retrieval-Augmented Generation):
You can reference up to three of your files to help guide Copilot’s drafting. Use the “Reference your content” button or type “/” followed by the file name in the compose box. This feature only accesses the files you select, ensuring your data remains private.
In general, such RAG features can be achieved by combining LLM Server and Vector Database. I plan to demonstrate this further by using AnythingLLM for vector database and LM Studio as its LLM provider in a future post. The main advantage in this local approach is that you can reference as many files as your hardware allows.
(2) Transform your text
For this use case, the example in Holger’s post is rewriting and editing. In terms of rewriting, I’ll use the text written in this story so far as the input and ask LM Studio to rewrite. The instruction given and its results are shown as follows.
Another example in Holger’s post is about converting text into a table. Unfortunately, this feature is not currently supported by the llama-3.2–3b-instruct model, which requires specific formatting and data structures. This limitation means that language models based solely on text cannot automatically convert text into tables without first being fine-tuned with training data tailored to Word document formats. That being said, there’s potential for enhancement in the future if the use case aligns with what users want.
(3) Chat about your document
For this use case, the example in the post is “Summarizing this doc”. Summarizing a document is a straightforward task for the llama-3.2–3b-instruct model because it can follow instructions. Summarization is a common use case, which is why my local Word Add-in provides “Summarize” as one of its default tasks. The “Summarize” task is defined as “You are a professional writer. Please summarize the following text:”. In the local Word Add-in, two other default tasks are also available: “Complete” for continuing writing and “Rewrite” for rephrasing the text. If you need a custom instruction, you can use “(Prompt)” to define it yourself and select the document in its entirety as input range by setting “Whole Document”. For demonstration purposes, I generated an article comparing the pros and cons of local LLMs versus cloud-based LLMs. Then I instructed the model to summarize the article, which is presented below.
These screenshots provide an additional layer of understanding of the key themes and implications of this story. If you want to learn more about how GPT-Localhost works, check out our demo page and give it a try.