inoid commited on
Commit
19a648e
·
1 Parent(s): a477476

Update local changes

Browse files
Files changed (1) hide show
  1. seminar_edition_ai.py +27 -3
seminar_edition_ai.py CHANGED
@@ -113,7 +113,7 @@ def predictProclamando(queryKey):
113
  ####
114
  #
115
  ####
116
- def predictFromInit( sermonTopic, llmModelList):
117
  global HISTORY_ANSWER
118
  keyStr = 'SERMON_TOPIC'
119
  templates = SermonGeminiPromptTemplate()
@@ -178,9 +178,11 @@ def predictFromInit( sermonTopic, llmModelList):
178
  ####
179
  #
180
  ####
181
- def predictQuestionBuild(sermonTopic):
 
182
  templates = SermonGeminiPromptTemplate()
183
  chain = updatePromptTemplate(
 
184
  templates.getSermonPromptTemplates()['BUILD_QUESTION'],
185
  ['SERMON_IDEA', 'context']
186
  )
@@ -217,15 +219,37 @@ def predictQuestionBuild(sermonTopic):
217
  ####
218
  #
219
  ####
220
- def predictDevotionBuild(sermonTopic):
221
  templates = SermonGeminiPromptTemplate()
 
222
 
223
  chain = updatePromptTemplate(
 
224
  templates.getSermonPromptTemplate()['BUILD_REFLECTIONS'],
225
  ['SERMON_IDEA', 'context']
226
  )
 
227
  global retriever
228
  global HISTORY_ANSWER
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
229
  answer = askQuestionEx(
230
  HISTORY_ANSWER,
231
  chain,
 
113
  ####
114
  #
115
  ####
116
+ def predictFromInit( sermonTopic, llmModelList = []):
117
  global HISTORY_ANSWER
118
  keyStr = 'SERMON_TOPIC'
119
  templates = SermonGeminiPromptTemplate()
 
178
  ####
179
  #
180
  ####
181
+ def predictQuestionBuild(sermonTopic, llmModelList = []):
182
+ llm = llmModelList[0] if len(llmModelList) > 0 else None
183
  templates = SermonGeminiPromptTemplate()
184
  chain = updatePromptTemplate(
185
+ llm,
186
  templates.getSermonPromptTemplates()['BUILD_QUESTION'],
187
  ['SERMON_IDEA', 'context']
188
  )
 
219
  ####
220
  #
221
  ####
222
+ def predictDevotionBuild(sermonTopic, llmModelList = []):
223
  templates = SermonGeminiPromptTemplate()
224
+ llm = llmModelList[0] if len(llmModelList) > 0 else None
225
 
226
  chain = updatePromptTemplate(
227
+ llm,
228
  templates.getSermonPromptTemplate()['BUILD_REFLECTIONS'],
229
  ['SERMON_IDEA', 'context']
230
  )
231
+
232
  global retriever
233
  global HISTORY_ANSWER
234
+ global embed_model
235
+
236
+ if embed_model == None:
237
+ llmBuilder = GeminiLLM()
238
+ embed_model = llmBuilder.getEmbeddingsModel()
239
+
240
+ if retriever == None:
241
+ doc = Document(page_content="text", metadata={"source": "local"})
242
+
243
+ vectorstore = Chroma.from_documents(
244
+ documents=[doc],
245
+ embedding=embed_model,
246
+ persist_directory="chroma_db_dir_sermon", # Local mode with in-memory storage only
247
+ collection_name="sermon_lab_ai"
248
+ )
249
+ retriever = vectorstore.as_retriever(
250
+ search_kwargs={"k": 3}
251
+ )
252
+
253
  answer = askQuestionEx(
254
  HISTORY_ANSWER,
255
  chain,