Mistral - isir/greta GitHub Wiki
The Mistral module can be used to dialog with Mistral through Greta
Installation
- make sure you have conda and python installed (usually you can install together with anaconda https://www.anaconda.com/)
To run the model locally
_ Warning : This requires an 6GB GPU_
-Install LmStudio : https://lmstudio.ai/ -Download model TheBloke/Mistral-7B-Instruct-v0.2-GGUF -Run local server with the following setting in LM Studio 👍 -port 1234
To run the model online
Warning : This is not free except for the two weeks where 5€ test are offered.
- Create an API key in https://console.mistral.ai/billing/
- In Common/Data/Mistral
- In Mistral.py replace “my_apy_key” by the API key created above
Use in Greta
- Add a Mistral Module to your Configuration
- Add a link between the Mistral module and the Behavior Planner
- Set up Port and Address (can stay the same)
- Check Enable
- Choose language and model (local or online)
- (Optional) Enter a system prompt if you want your agent to behave a certain way Ex : You are a museum guide who makes a lot of jokes. You always answer in rhymes
- Enter your request in the request panel
- Click Send
Tips
You can use a different model by modifying the model name in Mistral.py
An example of demo integration of this module is available at LLM DeepASR integration