Mistral - isir/greta GitHub Wiki

The Mistral module can be used to dialog with Mistral through Greta

Installation

To run the model locally

_ Warning : This requires an 6GB GPU_

-Install LmStudio : https://lmstudio.ai/ -Download model TheBloke/Mistral-7B-Instruct-v0.2-GGUF -Run local server with the following setting in LM Studio 👍 -port 1234

To run the model online

Warning : This is not free except for the two weeks where 5€ test are offered.

- Create an API key in https://console.mistral.ai/billing/
- In Common/Data/Mistral
- In Mistral.py replace “my_apy_key” by the API key created above

Use in Greta

  • Add a Mistral Module to your Configuration
  • Add a link between the Mistral module and the Behavior Planner
  • Set up Port and Address (can stay the same)
  • Check Enable
  • Choose language and model (local or online)
  • (Optional) Enter a system prompt if you want your agent to behave a certain way Ex : You are a museum guide who makes a lot of jokes. You always answer in rhymes
  • Enter your request in the request panel
  • Click Send

Tips

You can use a different model by modifying the model name in Mistral.py

An example of demo integration of this module is available at LLM DeepASR integration