Vistral là trợ lí Tiếng Việt nhiệt tình và trung thực.
This guide will help you run the chat application contained in the index.html file.
Original model: Vistral-7B-Chat - Towards a State-of-the-Art Large Language Model for Vietnamese.
Lightweight model: Vistral-7B-Chat - GGUF is now feasible to operate our Vistral-7B-Chat on a single computer even without a dedicated GPU.
Ensure you have the following:
- A modern web browser (Chrome, Firefox, Safari, etc.)
- A local web server (like Python's SimpleHTTPServer, Node's http-server, etc.). Or you can use Live Server feature from VSCode
- A laptop/PC with >8GB RAM
- Download model
Create a folder named models, then download ggml-vistral-7B-chat-q4_0.gguf from here https://huggingface.co/uonlp/Vistral-7B-Chat-gguf or detail https://huggingface.co/uonlp/Vistral-7B-Chat-gguf/blob/main/ggml-vistral-7B-chat-q4_0.gguf and put into the models folder
- Install llama_cpp Python
Follow the guide here to install llama_cpp Python https://github.com/abetlen/llama-cpp-python
## Ubuntu 20.04.1 LTS
python3 -V
Python 3.9.13
pip install llama-cpp-python
pip install uvicorn
pip install starlette
pip install fastapi
pip install pydantic_settings
pip install sse_starlette
pip install starlette_context
or
pip install -r requirements.txt- Run script python test
python3 test_llm.py- Run local server
Run the following script to run an API server locally. The server should run at port 8000
python3 -m llama_cpp.server --model "./models/ggml-vistral-7B-chat-q4_0.gguf" --n_gpu_layers 1 --n_ctx 4096
or
./server.shView API document of llama at http://localhost:8000/docs
- Start Web Server
python -m SimpleHTTPServer -p 8888
or
python3 -m http.server 8888If you're using Node's http-server, you can start it with the command:
http-server -p 8888Open your web browser and navigate to localhost on the port your server is running. For example, if your server is running on port 8888, you would navigate to http://localhost:8888/.
You should now see the chat interface in your browser. You can type messages into the input field and press "Send" to interact with the chatbot.
Please note that this is a simple setup meant for local development and testing. It is not suitable for a production environment.
This code is under the Apache License v2.