Gemma has been introduced, and also introduced the way to chat with him using code.
This method of using code is more suitable for developers, with a relatively high threshold.
From a usability perspective, it is not very convenient to use.
Today, another more user-friendly way will be introduced, directly using visual software to communicate with Gemma.
The software to be used today is called lmstudio, a software that supports running numerous large models with one click.
https://lmstudio.ai (magic required)
Its latest version 0.2.16 already supports Gemma models, both 2 billion and 7 billion parameter models are supported.
This software is quite good, besides supporting chatting with a large number of open-source models, it can also start OpenAI API services.
It is user-friendly for both ordinary users and developers.
The software also supports multiple platforms, currently supporting Windows, Mac (Apple Silicon M series), and Linux systems.
In addition, the software seems to default to using CPU for inference, so the device threshold is greatly reduced.
The inference speed of small models is also very good.
After installing the software (old users remember to upgrade), you can see the Google's Gemma 2b Instruct card on the main interface. Find download on the card, click to start downloading the model.
The whole process is very simple, but internal network users may encounter problems with downloading.
After the download is complete, select the Gemma model at the top of the software, wait for the model to load, and then you can start the conversation.
Enter content after User at the bottom of the window to start chatting.
You can directly send Chinese, it can understand, but it will reply to you in English.