Spaces:
Running
Running
as llm, use gpt-4o-mini
Browse files- src/rag.py +4 -3
src/rag.py
CHANGED
|
@@ -22,8 +22,9 @@ class RAG():
|
|
| 22 |
self.k = 3 # Number of relevant chunks to retrieve
|
| 23 |
|
| 24 |
# Constants
|
| 25 |
-
|
| 26 |
-
self.use_model = 'zephyr-7b-
|
|
|
|
| 27 |
# self.use_model = 'Mistral-Nemo-Base-2407'
|
| 28 |
|
| 29 |
# self.use_vectordb = 'chroma'
|
|
@@ -124,7 +125,7 @@ class RAG():
|
|
| 124 |
llm = ChatOpenAI(
|
| 125 |
model_name="gpt-4o-mini",
|
| 126 |
temperature=0)
|
| 127 |
-
elif self.use_model
|
| 128 |
provider = "hf-inference"
|
| 129 |
print(f'As llm, using HF-Endpint: {self.use_model} through provider: {provider}')
|
| 130 |
llm = HuggingFaceEndpoint(
|
|
|
|
| 22 |
self.k = 3 # Number of relevant chunks to retrieve
|
| 23 |
|
| 24 |
# Constants
|
| 25 |
+
self.use_model = 'gpt-4o-mini'
|
| 26 |
+
# self.use_model = 'zephyr-7b-alpha'
|
| 27 |
+
# self.use_model = 'zephyr-7b-beta'
|
| 28 |
# self.use_model = 'Mistral-Nemo-Base-2407'
|
| 29 |
|
| 30 |
# self.use_vectordb = 'chroma'
|
|
|
|
| 125 |
llm = ChatOpenAI(
|
| 126 |
model_name="gpt-4o-mini",
|
| 127 |
temperature=0)
|
| 128 |
+
elif self.use_model in ['zephyr-7b-alpha','zephyr-7b-beta'] :
|
| 129 |
provider = "hf-inference"
|
| 130 |
print(f'As llm, using HF-Endpint: {self.use_model} through provider: {provider}')
|
| 131 |
llm = HuggingFaceEndpoint(
|