Skip to content

Commit

Permalink
Merge pull request #646 from danielaskdd/Fix-broken-main
Browse files Browse the repository at this point in the history
Fix broken main
  • Loading branch information
LarFii authored Jan 25, 2025
2 parents 207c0e8 + 06673b6 commit e7e9073
Show file tree
Hide file tree
Showing 2 changed files with 17 additions and 8 deletions.
11 changes: 4 additions & 7 deletions lightrag/api/lightrag_server.py
Original file line number Diff line number Diff line change
Expand Up @@ -719,16 +719,13 @@ async def lifespan(app: FastAPI):

# Create working directory if it doesn't exist
Path(args.working_dir).mkdir(parents=True, exist_ok=True)
if args.llm_binding_host == "lollms" or args.embedding_binding == "lollms":
if args.llm_binding == "lollms" or args.embedding_binding == "lollms":
from lightrag.llm.lollms import lollms_model_complete, lollms_embed
if args.llm_binding_host == "ollama" or args.embedding_binding == "ollama":
if args.llm_binding == "ollama" or args.embedding_binding == "ollama":
from lightrag.llm.ollama import ollama_model_complete, ollama_embed
if args.llm_binding_host == "openai" or args.embedding_binding == "openai":
if args.llm_binding == "openai" or args.embedding_binding == "openai":
from lightrag.llm.openai import openai_complete_if_cache, openai_embed
if (
args.llm_binding_host == "azure_openai"
or args.embedding_binding == "azure_openai"
):
if args.llm_binding == "azure_openai" or args.embedding_binding == "azure_openai":
from lightrag.llm.azure_openai import (
azure_openai_complete_if_cache,
azure_openai_embed,
Expand Down
14 changes: 13 additions & 1 deletion lightrag/operate.py
Original file line number Diff line number Diff line change
Expand Up @@ -589,6 +589,9 @@ async def kg_query(
query, query_param, global_config, hashing_kv
)

logger.info(f"High-level keywords: {hl_keywords}")
logger.info(f"Low-level keywords: {ll_keywords}")

# Handle empty keywords
if hl_keywords == [] and ll_keywords == []:
logger.warning("low_level_keywords and high_level_keywords is empty")
Expand Down Expand Up @@ -1534,9 +1537,18 @@ async def naive_query(
if query_param.only_need_context:
return section

# Process conversation history
history_context = ""
if query_param.conversation_history:
history_context = get_conversation_turns(
query_param.conversation_history, query_param.history_turns
)

sys_prompt_temp = PROMPTS["naive_rag_response"]
sys_prompt = sys_prompt_temp.format(
content_data=section, response_type=query_param.response_type
content_data=section,
response_type=query_param.response_type,
history=history_context,
)

if query_param.only_need_prompt:
Expand Down

0 comments on commit e7e9073

Please sign in to comment.