llmware
1
2from llmware.prompts import Prompt
3from llmware.models import ModelCatalog
4
5# Registered by default
6# --dragon models: dragon-mistral-7b-gguf | dragon-yi-6b-gguf | dragon-llama-7b-gguf
7# --The Bloke leading 7b chat models: llama2-chat | openhermes | zephyr | starling
8
9
10# example 1 - how to use a default gguf model in llmware
11def use_default_gguf_model():
12
13selected_gguf_model = "llmware/dragon-mistral-7b-gguf"
14prompter = Prompt().load_model(selected_gguf_model)
15
16response = prompter.prompt_main("How old am I?", context="I am 36 years old.")
17
18print("response: ", response)
19
20return response
21
22
23response = use_default_gguf_model()
24
25
26# example 2 - how to use any GGUF model from The Bloke on HuggingFace
27def register_gguf_model():
28
29prompter = Prompt()
30
31your_model_name = "my_gguf_model_1"
32hf_repo_name = "TheBloke/model_name"
33model_file = "abc.gguf"
34
35prompter.model_catalog.register_gguf_model(your_model_name,hf_repo_name, model_file, prompt_wrapper="open_chat")
36prompter.load_model(your_model_name)
37
38return 0
39
40
41# example 3 - how to use build-from-source custom/optimized llama.cpp
42def build_your_own_llama_cpp_lib():
43
44import os
45os.environ["GGUF_CUSTOM_LIB_PATH"] = "/path/to/your/custom/lib"
46
47return 0
48
49