-
Home
- Troubleshooting
Debugging
Depending on the issue you are experiencing, you will have to either enable the debug mode for the Hardware Sentry Agent or for the OpenTelemetry Collector.
The Hardware Sentry Agent (core engine) is a Java process which launches the OpenTelemetry Collector and performs the hardware monitoring by connecting to each of the configured hosts and retrieving information about their hardware components. Enable its debug mode if there's no data available in your observability platform.
The OpenTelemetry Collector is in charge of pulling metrics, traces, and logs periodically and pushing them to the observability platform. Enable its debug mode, if data is missing (the monitoring coverage is incomplete).
Hardware Sentry Agent
To enable the debug mode of the core engine, edit the config/hws-config.yaml file and add the loggerLevel property:
loggerLevel: debug
hosts:
- host:
# [...]
Set loggerlevel to either all, trace, debug, info, warn, error, or fatal.
The debug output files are saved by default in the logs directory located under the hws directory:
-
On Windows, the output files are stored in the %LOCALAPPDATA%\hws folder of the account running the application:
- When the Local System account starts the Hardware Sentry Agent service, the output files are stored under C:\Windows\System32\config\systemprofile\AppData\Local\hws\logs.
- If a specific user starts the Hardware Sentry Agent service, the output files are stored under C:\Users\<username>\AppData\Local\hws\logs.
-
On Linux, the output files are stored in the installation directory: /opt/hws/logs.
To specify a different output directory, edit the hws-config.yaml file and add the outputDirectory parameter before the hosts section:
loggerLevel: debug
outputDirectory: C:\Users\<username>\AppData\Local\Temp\logs2021
hosts:
- host:
# [...]
Set loggerlevel to off to disable the debug mode.
OpenTelemetry Collector
To get more details about the Hardware Sentry operations (initialization, pipeline, termination), first set the log level to error, warn, or debug in the otel/otel-config.yaml file:
service:
telemetry:
logs:
level: debug
extensions: [health_check]
pipelines:
# [...]
Then, restart Hardware Sentry for these new settings to be considered.
Finally, check the logs/otelcol-<timestamp>.log file, where <timestamp> is the time at which the log was started.
Note: The logs/otelcol-<timestamp>.log file is reset each time the Collector is started. Previous logs are identified with the
<timestamp>value (ex:otelcol-2022-09-19-02-05-18.log). Hardware Sentry rotates the otelcol-<timestamp>.log file when it reaches a maximum size of 100MB and retains old log files for 2 days.
What to look for in otelcol-<timestamp>.log
First check that the Hardware Sentry Agent successfully launched the OpenTelemetry Collector:
[2022-09-19T14:05:18,459][INFO ][c.s.h.a.p.r.Executable] Started process command line: [C:\Program Files\hws\app\..\otel\otelcol-contrib, --config, "C:\ProgramData\hws\otel\otel-config.yaml", --feature-gates=pkg.translator.prometheus.NormalizeName]
[2022-09-19T14:05:19,363][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.363+0200 info service/telemetry.go:115 Setting up own telemetry...
[2022-09-19T14:05:19,363][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.363+0200 info service/telemetry.go:156 Serving Prometheus metrics {"address": "localhost:8888", "level": "basic"}
[2022-09-19T14:05:19,775][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.774+0200 info pipelines/pipelines.go:74 Starting exporters...
[2022-09-19T14:05:19,775][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.774+0200 info pipelines/pipelines.go:78 Exporter is starting... {"kind": "exporter", "data_type": "metrics", "name": "prometheusremotewrite/nb-prom"}
[2022-09-19T14:05:19,775][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.774+0200 info pipelines/pipelines.go:82 Exporter started. {"kind": "exporter", "data_type": "metrics", "name": "prometheusremotewrite/nb-prom"}
[2022-09-19T14:05:19,775][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.774+0200 info pipelines/pipelines.go:78 Exporter is starting... {"kind": "exporter", "data_type": "metrics", "name": "prometheus"}
[2022-09-19T14:05:19,776][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.775+0200 info pipelines/pipelines.go:82 Exporter started. {"kind": "exporter", "data_type": "metrics", "name": "prometheus"}
[2022-09-19T14:05:19,776][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.775+0200 info pipelines/pipelines.go:78 Exporter is starting... {"kind": "exporter", "data_type": "metrics", "name": "datadog/api"}
[2022-09-19T14:05:19,776][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.775+0200 info pipelines/pipelines.go:82 Exporter started. {"kind": "exporter", "data_type": "metrics", "name": "datadog/api"}
[2022-09-19T14:05:19,776][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.775+0200 info pipelines/pipelines.go:78 Exporter is starting... {"kind": "exporter", "data_type": "logs", "name": "logging"}
[2022-09-19T14:05:19,776][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.775+0200 info pipelines/pipelines.go:82 Exporter started. {"kind": "exporter", "data_type": "logs", "name": "logging"}
[2022-09-19T14:05:19,776][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.775+0200 info pipelines/pipelines.go:86 Starting processors...
[2022-09-19T14:05:19,776][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.775+0200 info pipelines/pipelines.go:90 Processor is starting... {"kind": "processor", "name": "metricstransform", "pipeline": "metrics"}
[2022-09-19T14:05:19,776][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.775+0200 info pipelines/pipelines.go:94 Processor started. {"kind": "processor", "name": "metricstransform", "pipeline": "metrics"}
[2022-09-19T14:05:19,776][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.775+0200 info pipelines/pipelines.go:90 Processor is starting... {"kind": "processor", "name": "resourcedetection", "pipeline": "metrics"}
[2022-09-19T14:05:19,776][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:19.775+0200 info internal/resourcedetection.go:136 began detecting resource information {"kind": "processor", "name": "resourcedetection", "pipeline": "metrics"}
[2022-09-19T14:05:20,834][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:20.834+0200 info internal/resourcedetection.go:150 detected resource information {"kind": "processor", "name": "resourcedetection", "pipeline": "metrics", "resource": {"host.name":"pc-nassim.internal.sentrysoftware.net","os.type":"windows"}}
[2022-09-19T14:05:20,835][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:20.834+0200 info pipelines/pipelines.go:94 Processor started. {"kind": "processor", "name": "resourcedetection", "pipeline": "metrics"}
[2022-09-19T14:05:20,835][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:20.834+0200 info pipelines/pipelines.go:90 Processor is starting... {"kind": "processor", "name": "batch", "pipeline": "metrics"}
[2022-09-19T14:05:20,835][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:20.834+0200 info pipelines/pipelines.go:94 Processor started. {"kind": "processor", "name": "batch", "pipeline": "metrics"}
[2022-09-19T14:05:20,835][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:20.834+0200 info pipelines/pipelines.go:90 Processor is starting... {"kind": "processor", "name": "memory_limiter", "pipeline": "metrics"}
[2022-09-19T14:05:20,835][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:20.834+0200 info pipelines/pipelines.go:94 Processor started. {"kind": "processor", "name": "memory_limiter", "pipeline": "metrics"}
[2022-09-19T14:05:20,835][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:20.834+0200 info pipelines/pipelines.go:90 Processor is starting... {"kind": "processor", "name": "resourcedetection", "pipeline": "logs"}
[2022-09-19T14:05:20,835][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:20.834+0200 info pipelines/pipelines.go:94 Processor started. {"kind": "processor", "name": "resourcedetection", "pipeline": "logs"}
[2022-09-19T14:05:20,835][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:20.834+0200 info pipelines/pipelines.go:90 Processor is starting... {"kind": "processor", "name": "batch", "pipeline": "logs"}
[2022-09-19T14:05:20,835][DEBUG][c.s.h.a.s.o.p.OtelCollectorExecutable] 2022-09-19T14:05:20.834+0200 info pipelines/pipelines.go:94 Processor started. {"kind": "processor", "name": "batch", "pipeline": "logs"}
Then check that the exporters and processors properly started.
Finally look for any connection issues or authentication failures to the configured observability platform(s) (Datadog, BMC Helix, Prometheus, Grafana, etc.).
Getting more details about the exported data
You can enable the logging exporter in the otel/otel-config.yaml file to check which metrics, labels, and values are sent by the Collector to the observability platforms and verify that the configured processors did not alter the collected data.
First, list the logging exporter under the exporters section and set verbosity to detailed:
exporters:
# [...]
logging:
verbosity: detailed
Then, declare the logging exporter in the pipeline:
service:
pipelines:
metrics:
receivers: # receivers
processors: # processors
exporters: [prometheusremotewrite/your-server,logging] # <-- added logging
Restart the Collector for the new settings to be considered.
The metric name, its labels and value are listed in the logs/otelcol-<timestamp>.log file.
Important: Disable the
loggingexporter when unused as its operation may affect the overall performance of the Collector and fill your file system.
Reducing the amount of information logged
To reduce the amount of information logged, you can configure the filter processor to only log metrics of specific hosts. In the example below, we configured the filter/keep1HostOnly processor to only log information about systems whose hostname contains my-server.big-corp.com:
processors:
filter/keep1HostOnly:
metrics:
include:
match_type: expr
expressions:
- Label("host.name") == "my-server.big-corp.com"
We then declared the filter/keep1HostOnly processor in the pipeline and restarted the Collector:
service:
pipelines:
metrics:
receivers: [otlp, prometheus/internal]
processors: [memory_limiter,batch,filter/keep1HostOnly] # <-- added filter
exporters: # exporters
Important: Remove the
filterprocessor from your pipeline once the troubleshooting is completed.