File tree 1 file changed +1
-2
lines changed
tutorial/OpenTelemetry/deploy
1 file changed +1
-2
lines changed Original file line number Diff line number Diff line change @@ -13,7 +13,7 @@ A dashboard for monitoring CPU statistics is also available, offering comprehens
13
13
## Telemetry Tracing with Jaeger on Gaudi
14
14
15
15
After ChatQnA processes a question, two traces should appear along the timeline.
16
- The trace for opea: ServiceOrchestrator.schedule runs on the CPU and includes seven spans, one of which represents the LLM host functions in general.
16
+ The trace for opea: ServiceOrchestrator.schedule runs on the CPU and includes seven spans, one of which represents the LLM service running on CPU.
17
17
For LLM functions executed on Gaudi, stream requests are displayed under opea: llm_generate_stream.
18
18
This trace contains two spans: one for the first token and another for all subsequent tokens.
19
19
@@ -44,7 +44,6 @@ highlighting the processing of stream requests and the associated spans for toke
44
44
Clicking on the opea: llm_generate_stream trace will expand to reveal two spans along the timeline.
45
45
The first span represents the execution time for the first token, which took 15.12 ms in this run.
46
46
The second span captures the execution time for all subsequent tokens, taking 920 ms as shown in the diagram.
47
- These spans provide a detailed view of the token generation process and the performance of LLM functions on Gaudi.
48
47
49
48
![ chatqna_gaudi_breakdown] ( ../assets/jaeger_ui_opea_chatqna_req_breakdown_2.png )
50
49
You can’t perform that action at this time.
0 commit comments