Spaces:
Sleeping
Sleeping
-- Modified the system prompt to avoid responding with empty answer by bytedance/seed-oss-36b-instruct (nvidia) ML Model.
Browse files- llm_inference_service.py +2 -1
llm_inference_service.py
CHANGED
|
@@ -69,7 +69,8 @@ def extract_page_info_by_llm(user_query: str, scraped_markdown_content: str, mod
|
|
| 69 |
So always read the user query carefully and extract information accordingly.
|
| 70 |
|
| 71 |
If you do not find or know the answer, do not hallucinate, do not try to generate fake answers.
|
| 72 |
-
If no Context is given, simply state "No relevant information found to answer your question."
|
|
|
|
| 73 |
|
| 74 |
Context:
|
| 75 |
{context}
|
|
|
|
| 69 |
So always read the user query carefully and extract information accordingly.
|
| 70 |
|
| 71 |
If you do not find or know the answer, do not hallucinate, do not try to generate fake answers.
|
| 72 |
+
If no Context is given or you can't find or generate any relevant information to answer the question, simply state "No relevant information found to answer your question."
|
| 73 |
+
Please do not respond with empty string / answer.
|
| 74 |
|
| 75 |
Context:
|
| 76 |
{context}
|