Compare commits

...

3 Commits

Author SHA1 Message Date
Joe
0582442562 feat: update langsmith workflow trace metadata 2024-08-15 16:29:10 +08:00
Joe
77d923c951 fix: langfuse generation error 2024-08-15 12:34:45 +08:00
Joe
701c1c3a60 feat: ops trace add llm model 2024-08-14 16:20:22 +08:00
4 changed files with 14 additions and 5 deletions

View File

@@ -204,6 +204,7 @@ class LangFuseDataTrace(BaseTraceInstance):
node_generation_data = LangfuseGeneration(
name="llm",
trace_id=trace_id,
model=process_data.get("model_name"),
parent_observation_id=node_execution_id,
start_time=created_at,
end_time=finished_at,

View File

@@ -139,8 +139,7 @@ class LangSmithDataTrace(BaseTraceInstance):
json.loads(node_execution.execution_metadata) if node_execution.execution_metadata else {}
)
node_total_tokens = execution_metadata.get("total_tokens", 0)
metadata = json.loads(node_execution.execution_metadata) if node_execution.execution_metadata else {}
metadata = execution_metadata.copy()
metadata.update(
{
"workflow_run_id": trace_info.workflow_run_id,
@@ -156,6 +155,12 @@ class LangSmithDataTrace(BaseTraceInstance):
process_data = json.loads(node_execution.process_data) if node_execution.process_data else {}
if process_data and process_data.get("model_mode") == "chat":
run_type = LangSmithRunType.llm
metadata.update(
{
'ls_provider': process_data.get('model_provider', ''),
'ls_model_name': process_data.get('model_name', ''),
}
)
elif node_type == "knowledge-retrieval":
run_type = LangSmithRunType.retriever
else:

View File

@@ -105,7 +105,9 @@ class LLMNode(BaseNode):
'prompts': PromptMessageUtil.prompt_messages_to_prompt_for_saving(
model_mode=model_config.mode,
prompt_messages=prompt_messages
)
),
'model_provider': model_config.provider,
'model_name': model_config.model,
}
# handle invoke result

View File

@@ -89,10 +89,11 @@ class QuestionClassifierNode(LLMNode):
process_data = {
'model_mode': model_config.mode,
'prompts': PromptMessageUtil.prompt_messages_to_prompt_for_saving(
model_mode=model_config.mode,
prompt_messages=prompt_messages
model_mode=model_config.mode, prompt_messages=prompt_messages
),
'usage': jsonable_encoder(usage),
'model_provider': model_config.provider,
'model_name': model_config.model,
}
outputs = {
'class_name': category_name