@@ -50,7 +50,7 @@ def _get_context(self, question: str) -> Union[str, list[str]]:
5050 return output
5151 except ValueError as ve :
5252 raise HTTPException (
53- status_code = 500 , detail = f"Error getting context: { str (ve )} "
53+ status_code = 404 , detail = f"Error getting context: { str (ve )} "
5454 )
5555 except Exception as e :
5656 logger .error (f"Unexpected error in _get_context for question '{ question } ': { str (e )} " , exc_info = True )
@@ -59,26 +59,31 @@ def _get_context(self, question: str) -> Union[str, list[str]]:
5959 )
6060
6161 def generate_llm_response (self , question : schemas .Question ) -> StreamingResponse :
62- context = self ._get_context (question .question )
62+ try :
63+ context = self ._get_context (question .question )
64+ except HTTPException as e :
65+ logger .error (f"No context found" , exc_info = True )
66+ context = ""
67+
6368 formatted_messages = ""
64- context_messages = ""
69+ # context_messages = ""
6570
6671 if question .messages :
6772 formatted_messages = "\n " .join (
6873 [f"{ msg .sender } : { msg .content } " for msg in question .messages ]
6974 )
70- context_messages = self ._get_context (formatted_messages )
75+ # context_messages = self._get_context(formatted_messages)
7176
7277 messages = [
7378 SystemMessage (self ._CHATBOT_INSTRUCTIONS ),
79+ SystemMessage (f"Conversazione precedente: { formatted_messages } " ),
7480 SystemMessage (
75- f"Contesto: { context } \n { context_messages } " ,
81+ f"Contesto: { context } " ,
7682 ),
77- SystemMessage (f"Conversazione precedente: { formatted_messages } " ),
7883 HumanMessage (f"Domanda a cui devi rispondere: { question } " ),
7984 ]
8085 print ()
81- print (f"PROMPT: { context } { context_messages } " )
86+ print (f"PROMPT: { context } " )
8287 print ()
8388 try :
8489 stream_response = self ._LLM ._model .astream (messages )
0 commit comments