Skip to content

Commit d404716

Browse files
committed
addressed comments
Signed-off-by: Tsai, Louie <louie.tsai@intel.com>
1 parent 99c6d45 commit d404716

File tree

1 file changed

+1
-2
lines changed

1 file changed

+1
-2
lines changed

tutorial/OpenTelemetry/deploy/ChatQnA.md

+1-2
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ A dashboard for monitoring CPU statistics is also available, offering comprehens
1313
## Telemetry Tracing with Jaeger on Gaudi
1414

1515
After ChatQnA processes a question, two traces should appear along the timeline.
16-
The trace for opea: ServiceOrchestrator.schedule runs on the CPU and includes seven spans, one of which represents the LLM host functions in general.
16+
The trace for opea: ServiceOrchestrator.schedule runs on the CPU and includes seven spans, one of which represents the LLM service running on CPU.
1717
For LLM functions executed on Gaudi, stream requests are displayed under opea: llm_generate_stream.
1818
This trace contains two spans: one for the first token and another for all subsequent tokens.
1919

@@ -44,7 +44,6 @@ highlighting the processing of stream requests and the associated spans for toke
4444
Clicking on the opea: llm_generate_stream trace will expand to reveal two spans along the timeline.
4545
The first span represents the execution time for the first token, which took 15.12 ms in this run.
4646
The second span captures the execution time for all subsequent tokens, taking 920 ms as shown in the diagram.
47-
These spans provide a detailed view of the token generation process and the performance of LLM functions on Gaudi.
4847

4948
![chatqna_gaudi_breakdown](../assets/jaeger_ui_opea_chatqna_req_breakdown_2.png)
5049

0 commit comments

Comments
 (0)