Dashboards reference
This document contains a complete reference on Sourcegraph's available dashboards, as well as details on how to interpret the panels and metrics.
To learn more about Sourcegraph's metrics and how to view these dashboards, see our metrics guide.
Frontend
Serves all end-user browser and API requests.
To see this dashboard, visit /-/debug/grafana/d/frontend/frontend
on your Sourcegraph instance.
Frontend: Search at a glance
frontend: 99th_percentile_search_request_duration
99th percentile successful search request duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100000
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le)(rate(src_search_streaming_latency_seconds_bucket{source="browser"}[5m])))
frontend: 90th_percentile_search_request_duration
90th percentile successful search request duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100001
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le)(rate(src_search_streaming_latency_seconds_bucket{source="browser"}[5m])))
frontend: hard_timeout_search_responses
Hard timeout search responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100010
on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(increase(src_graphql_search_response{status="timeout",source="browser",request_name!="CodeIntelSearch"}[5m])) + sum(increase(src_graphql_search_response{status="alert",alert_type="timed_out",source="browser",request_name!="CodeIntelSearch"}[5m]))) / sum(increase(src_graphql_search_response{source="browser",request_name!="CodeIntelSearch"}[5m])) * 100
frontend: hard_error_search_responses
Hard error search responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100011
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (status)(increase(src_graphql_search_response{status=~"error",source="browser",request_name!="CodeIntelSearch"}[5m])) / ignoring(status) group_left sum(increase(src_graphql_search_response{source="browser",request_name!="CodeIntelSearch"}[5m])) * 100
frontend: partial_timeout_search_responses
Partial timeout search responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100012
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (status)(increase(src_graphql_search_response{status="partial_timeout",source="browser",request_name!="CodeIntelSearch"}[5m])) / ignoring(status) group_left sum(increase(src_graphql_search_response{source="browser",request_name!="CodeIntelSearch"}[5m])) * 100
frontend: search_alert_user_suggestions
Search alert user suggestions shown every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100013
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (alert_type)(increase(src_graphql_search_response{status="alert",alert_type!~"timed_out|no_results__suggest_quotes",source="browser",request_name!="CodeIntelSearch"}[5m])) / ignoring(alert_type) group_left sum(increase(src_graphql_search_response{source="browser",request_name!="CodeIntelSearch"}[5m])) * 100
frontend: page_load_latency
90th percentile page load latency over all routes over 10m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100020
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.9, sum by(le) (rate(src_http_request_duration_seconds_bucket{route!="raw",route!="blob",route!~"graphql.*"}[10m])))
Frontend: Search-based code intelligence at a glance
frontend: 99th_percentile_search_codeintel_request_duration
99th percentile code-intel successful search request duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100100
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le)(rate(src_graphql_field_seconds_bucket{type="Search",field="results",error="false",source="browser",request_name="CodeIntelSearch"}[5m])))
frontend: 90th_percentile_search_codeintel_request_duration
90th percentile code-intel successful search request duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100101
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le)(rate(src_graphql_field_seconds_bucket{type="Search",field="results",error="false",source="browser",request_name="CodeIntelSearch"}[5m])))
frontend: hard_timeout_search_codeintel_responses
Hard timeout search code-intel responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100110
on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(increase(src_graphql_search_response{status="timeout",source="browser",request_name="CodeIntelSearch"}[5m])) + sum(increase(src_graphql_search_response{status="alert",alert_type="timed_out",source="browser",request_name="CodeIntelSearch"}[5m]))) / sum(increase(src_graphql_search_response{source="browser",request_name="CodeIntelSearch"}[5m])) * 100
frontend: hard_error_search_codeintel_responses
Hard error search code-intel responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100111
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (status)(increase(src_graphql_search_response{status=~"error",source="browser",request_name="CodeIntelSearch"}[5m])) / ignoring(status) group_left sum(increase(src_graphql_search_response{source="browser",request_name="CodeIntelSearch"}[5m])) * 100
frontend: partial_timeout_search_codeintel_responses
Partial timeout search code-intel responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100112
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (status)(increase(src_graphql_search_response{status="partial_timeout",source="browser",request_name="CodeIntelSearch"}[5m])) / ignoring(status) group_left sum(increase(src_graphql_search_response{status="partial_timeout",source="browser",request_name="CodeIntelSearch"}[5m])) * 100
frontend: search_codeintel_alert_user_suggestions
Search code-intel alert user suggestions shown every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100113
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (alert_type)(increase(src_graphql_search_response{status="alert",alert_type!~"timed_out",source="browser",request_name="CodeIntelSearch"}[5m])) / ignoring(alert_type) group_left sum(increase(src_graphql_search_response{source="browser",request_name="CodeIntelSearch"}[5m])) * 100
Frontend: Search GraphQL API usage at a glance
frontend: 99th_percentile_search_api_request_duration
99th percentile successful search API request duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100200
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le)(rate(src_graphql_field_seconds_bucket{type="Search",field="results",error="false",source="other"}[5m])))
frontend: 90th_percentile_search_api_request_duration
90th percentile successful search API request duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100201
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le)(rate(src_graphql_field_seconds_bucket{type="Search",field="results",error="false",source="other"}[5m])))
frontend: hard_error_search_api_responses
Hard error search API responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100210
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (status)(increase(src_graphql_search_response{status=~"error",source="other"}[5m])) / ignoring(status) group_left sum(increase(src_graphql_search_response{source="other"}[5m]))
frontend: partial_timeout_search_api_responses
Partial timeout search API responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100211
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_graphql_search_response{status="partial_timeout",source="other"}[5m])) / sum(increase(src_graphql_search_response{source="other"}[5m]))
frontend: search_api_alert_user_suggestions
Search API alert user suggestions shown every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100212
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (alert_type)(increase(src_graphql_search_response{status="alert",alert_type!~"timed_out|no_results__suggest_quotes",source="other"}[5m])) / ignoring(alert_type) group_left sum(increase(src_graphql_search_response{status="alert",source="other"}[5m]))
Frontend: Site configuration client update latency
frontend: frontend_site_configuration_duration_since_last_successful_update_by_instance
Duration since last successful site configuration update (by instance)
The duration since the configuration client used by the "frontend" service last successfully updated its site configuration. Long durations could indicate issues updating the site configuration.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100300
on your Sourcegraph instance.
Technical details
Query:
SHELLsrc_conf_client_time_since_last_successful_update_seconds{job=~`(sourcegraph-)?frontend`,instance=~`${internalInstance:regex}`}
frontend: frontend_site_configuration_duration_since_last_successful_update_by_instance
Maximum duration since last successful site configuration update (all "frontend" instances)
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100301
on your Sourcegraph instance.
Technical details
Query:
SHELLmax(max_over_time(src_conf_client_time_since_last_successful_update_seconds{job=~`(sourcegraph-)?frontend`,instance=~`${internalInstance:regex}`}[1m]))
Frontend: Codeintel: Precise code intelligence usage at a glance
frontend: codeintel_resolvers_total
Aggregate graphql operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100400
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_resolvers_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_resolvers_99th_percentile_duration
Aggregate successful graphql operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100401
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_resolvers_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_resolvers_errors_total
Aggregate graphql operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100402
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_resolvers_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_resolvers_error_rate
Aggregate graphql operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100403
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_resolvers_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_resolvers_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_resolvers_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_resolvers_total
Graphql operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100410
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_resolvers_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_resolvers_99th_percentile_duration
99th percentile successful graphql operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100411
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_resolvers_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_resolvers_errors_total
Graphql operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100412
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_resolvers_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_resolvers_error_rate
Graphql operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100413
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_resolvers_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_resolvers_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_resolvers_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Codeintel: Auto-index enqueuer
frontend: codeintel_autoindex_enqueuer_total
Aggregate enqueuer operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100500
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindex_enqueuer_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_autoindex_enqueuer_99th_percentile_duration
Aggregate successful enqueuer operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100501
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_autoindex_enqueuer_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_autoindex_enqueuer_errors_total
Aggregate enqueuer operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100502
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindex_enqueuer_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_autoindex_enqueuer_error_rate
Aggregate enqueuer operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100503
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_autoindex_enqueuer_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_autoindex_enqueuer_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_autoindex_enqueuer_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_autoindex_enqueuer_total
Enqueuer operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100510
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindex_enqueuer_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_autoindex_enqueuer_99th_percentile_duration
99th percentile successful enqueuer operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100511
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_autoindex_enqueuer_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_autoindex_enqueuer_errors_total
Enqueuer operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100512
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindex_enqueuer_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_autoindex_enqueuer_error_rate
Enqueuer operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100513
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_autoindex_enqueuer_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_autoindex_enqueuer_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_autoindex_enqueuer_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Codeintel: dbstore stats
frontend: codeintel_uploads_store_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100600
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_store_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100601
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploads_store_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_store_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100602
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_store_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100603
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_uploads_store_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_uploads_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_uploads_store_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100610
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_store_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100611
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_store_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_uploads_store_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100612
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_store_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100613
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_store_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Workerutil: lsif_indexes dbworker/store stats
frontend: workerutil_dbworker_store_codeintel_index_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100700
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_codeintel_index_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: workerutil_dbworker_store_codeintel_index_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100701
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_workerutil_dbworker_store_codeintel_index_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: workerutil_dbworker_store_codeintel_index_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100702
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_codeintel_index_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: workerutil_dbworker_store_codeintel_index_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100703
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_workerutil_dbworker_store_codeintel_index_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_workerutil_dbworker_store_codeintel_index_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_workerutil_dbworker_store_codeintel_index_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Codeintel: lsifstore stats
frontend: codeintel_uploads_lsifstore_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100800
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_lsifstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_lsifstore_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100801
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploads_lsifstore_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_lsifstore_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100802
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_lsifstore_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100803
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_uploads_lsifstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_uploads_lsifstore_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100810
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_lsifstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_lsifstore_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100811
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploads_lsifstore_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_uploads_lsifstore_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100812
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploads_lsifstore_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100813
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_uploads_lsifstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_uploads_lsifstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Codeintel: gitserver client
frontend: codeintel_gitserver_total
Aggregate client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100900
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_gitserver_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_gitserver_99th_percentile_duration
Aggregate successful client operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100901
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_gitserver_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_gitserver_errors_total
Aggregate client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100902
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_gitserver_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_gitserver_error_rate
Aggregate client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100903
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_gitserver_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_gitserver_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_gitserver_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_gitserver_total
Client operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100910
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_gitserver_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_gitserver_99th_percentile_duration
99th percentile successful client operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100911
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_gitserver_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_gitserver_errors_total
Client operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100912
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_gitserver_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_gitserver_error_rate
Client operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=100913
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_gitserver_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_gitserver_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_gitserver_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Codeintel: uploadstore stats
frontend: codeintel_uploadstore_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101000
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploadstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploadstore_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101001
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_uploadstore_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploadstore_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101002
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploadstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploadstore_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101003
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_uploadstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_uploadstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_uploadstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_uploadstore_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101010
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploadstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploadstore_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101011
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_uploadstore_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_uploadstore_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101012
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploadstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_uploadstore_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101013
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_uploadstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_uploadstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_uploadstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Codeintel: dependencies service stats
frontend: codeintel_dependencies_total
Aggregate service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101100
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_dependencies_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_99th_percentile_duration
Aggregate successful service operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101101
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_dependencies_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_errors_total
Aggregate service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101102
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_dependencies_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_error_rate
Aggregate service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101103
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_dependencies_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_dependencies_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_dependencies_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_dependencies_total
Service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101110
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_dependencies_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_99th_percentile_duration
99th percentile successful service operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101111
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_dependencies_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_dependencies_errors_total
Service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101112
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_dependencies_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_error_rate
Service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101113
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_dependencies_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_dependencies_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_dependencies_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Codeintel: dependencies service store stats
frontend: codeintel_dependencies_background_total
Aggregate service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101200
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_dependencies_background_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_background_99th_percentile_duration
Aggregate successful service operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101201
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_dependencies_background_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_background_errors_total
Aggregate service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101202
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_dependencies_background_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_background_error_rate
Aggregate service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101203
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_dependencies_background_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_dependencies_background_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_dependencies_background_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_dependencies_background_total
Service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101210
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_dependencies_background_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_background_99th_percentile_duration
99th percentile successful service operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101211
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_dependencies_background_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_dependencies_background_errors_total
Service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101212
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_dependencies_background_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_background_error_rate
Service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101213
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_dependencies_background_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_dependencies_background_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_dependencies_background_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Codeintel: dependencies service background stats
frontend: codeintel_dependencies_background_total
Aggregate service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101300
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_dependencies_background_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_background_99th_percentile_duration
Aggregate successful service operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101301
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_dependencies_background_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_background_errors_total
Aggregate service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101302
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_dependencies_background_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_background_error_rate
Aggregate service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101303
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_dependencies_background_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_dependencies_background_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_dependencies_background_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_dependencies_background_total
Service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101310
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_dependencies_background_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_background_99th_percentile_duration
99th percentile successful service operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101311
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_dependencies_background_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_dependencies_background_errors_total
Service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101312
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_dependencies_background_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_dependencies_background_error_rate
Service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101313
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_dependencies_background_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_dependencies_background_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_dependencies_background_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Codeintel: lockfiles service stats
frontend: codeintel_lockfiles_total
Aggregate service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101400
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_lockfiles_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_lockfiles_99th_percentile_duration
Aggregate successful service operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101401
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_lockfiles_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_lockfiles_errors_total
Aggregate service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101402
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_lockfiles_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_lockfiles_error_rate
Aggregate service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101403
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_lockfiles_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_codeintel_lockfiles_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_codeintel_lockfiles_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: codeintel_lockfiles_total
Service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101410
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_lockfiles_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_lockfiles_99th_percentile_duration
99th percentile successful service operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101411
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_lockfiles_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: codeintel_lockfiles_errors_total
Service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101412
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_lockfiles_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: codeintel_lockfiles_error_rate
Service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101413
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_lockfiles_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_codeintel_lockfiles_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_codeintel_lockfiles_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Gitserver: Gitserver Client
frontend: gitserver_client_total
Aggregate graphql operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101500
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_99th_percentile_duration
Aggregate successful graphql operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101501
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_errors_total
Aggregate graphql operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101502
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_error_rate
Aggregate graphql operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101503
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_gitserver_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: gitserver_client_total
Graphql operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101510
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_99th_percentile_duration
99th percentile successful graphql operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101511
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op,scope)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: gitserver_client_errors_total
Graphql operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101512
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: gitserver_client_error_rate
Graphql operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101513
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op,scope)(increase(src_gitserver_client_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Batches: dbstore stats
frontend: batches_dbstore_total
Aggregate store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101600
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_dbstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_dbstore_99th_percentile_duration
Aggregate successful store operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101601
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_batches_dbstore_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_dbstore_errors_total
Aggregate store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101602
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_dbstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_dbstore_error_rate
Aggregate store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101603
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_dbstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_batches_dbstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_batches_dbstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: batches_dbstore_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101610
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_dbstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_dbstore_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101611
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_batches_dbstore_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: batches_dbstore_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101612
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_dbstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_dbstore_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101613
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_dbstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_batches_dbstore_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_batches_dbstore_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Batches: service stats
frontend: batches_service_total
Aggregate service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101700
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_service_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_service_99th_percentile_duration
Aggregate successful service operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101701
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_batches_service_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_service_errors_total
Aggregate service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101702
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_service_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_service_error_rate
Aggregate service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101703
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_service_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_batches_service_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_batches_service_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: batches_service_total
Service operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101710
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_service_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_service_99th_percentile_duration
99th percentile successful service operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101711
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_batches_service_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: batches_service_errors_total
Service operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101712
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_service_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_service_error_rate
Service operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101713
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_service_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_batches_service_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_batches_service_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Batches: Workspace execution dbstore
frontend: workerutil_dbworker_store_batch_spec_workspace_execution_worker_store_total
Store operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101800
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_workerutil_dbworker_store_batch_spec_workspace_execution_worker_store_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: workerutil_dbworker_store_batch_spec_workspace_execution_worker_store_99th_percentile_duration
99th percentile successful store operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101801
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_workerutil_dbworker_store_batch_spec_workspace_execution_worker_store_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: workerutil_dbworker_store_batch_spec_workspace_execution_worker_store_errors_total
Store operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101802
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_workerutil_dbworker_store_batch_spec_workspace_execution_worker_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: workerutil_dbworker_store_batch_spec_workspace_execution_worker_store_error_rate
Store operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101803
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_workerutil_dbworker_store_batch_spec_workspace_execution_worker_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_workerutil_dbworker_store_batch_spec_workspace_execution_worker_store_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_workerutil_dbworker_store_batch_spec_workspace_execution_worker_store_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Batches: HTTP API File Handler
frontend: batches_httpapi_total
Aggregate http handler operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101900
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_httpapi_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_httpapi_99th_percentile_duration
Aggregate successful http handler operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101901
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_batches_httpapi_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_httpapi_errors_total
Aggregate http handler operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101902
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_httpapi_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_httpapi_error_rate
Aggregate http handler operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101903
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_batches_httpapi_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_batches_httpapi_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_batches_httpapi_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: batches_httpapi_total
Http handler operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101910
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_httpapi_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_httpapi_99th_percentile_duration
99th percentile successful http handler operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101911
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_batches_httpapi_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: batches_httpapi_errors_total
Http handler operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101912
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_httpapi_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: batches_httpapi_error_rate
Http handler operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=101913
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_batches_httpapi_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op)(increase(src_batches_httpapi_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op)(increase(src_batches_httpapi_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Out-of-band migrations: up migration invocation (one batch processed)
frontend: oobmigration_total
Migration handler operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102000
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_oobmigration_total{op="up",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: oobmigration_99th_percentile_duration
Aggregate successful migration handler operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102001
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_oobmigration_duration_seconds_bucket{op="up",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: oobmigration_errors_total
Migration handler operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102002
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_oobmigration_errors_total{op="up",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: oobmigration_error_rate
Migration handler operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102003
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_oobmigration_errors_total{op="up",job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_oobmigration_total{op="up",job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_oobmigration_errors_total{op="up",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Out-of-band migrations: down migration invocation (one batch processed)
frontend: oobmigration_total
Migration handler operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102100
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_oobmigration_total{op="down",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: oobmigration_99th_percentile_duration
Aggregate successful migration handler operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102101
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_oobmigration_duration_seconds_bucket{op="down",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: oobmigration_errors_total
Migration handler operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102102
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_oobmigration_errors_total{op="down",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: oobmigration_error_rate
Migration handler operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102103
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_oobmigration_errors_total{op="down",job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_oobmigration_total{op="down",job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_oobmigration_errors_total{op="down",job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Frontend: Zoekt Configuration GRPC server metrics
frontend: zoekt_configuration_grpc_request_rate_all_methods
Request rate across all methods over 2m
The number of gRPC requests received per second across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102200
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m]))
frontend: zoekt_configuration_grpc_request_rate_per_method
Request rate per-method over 2m
The number of gRPC requests received per second broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102201
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])) by (grpc_method)
frontend: zoekt_configuration_error_percentage_all_methods
Error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102210
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_code!="OK",instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m]))) / (sum(rate(grpc_server_handled_total{instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m]))) ))
frontend: zoekt_configuration_grpc_error_percentage_per_method
Error percentage per-method over 2m
The percentage of gRPC requests that fail per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102211
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_method=~`${zoekt_configuration_method:regex}`,grpc_code!="OK",instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])) by (grpc_method)) / (sum(rate(grpc_server_handled_total{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])) by (grpc_method)) ))
frontend: zoekt_configuration_p99_response_time_per_method
99th percentile response time per method over 2m
The 99th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102220
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p90_response_time_per_method
90th percentile response time per method over 2m
The 90th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102221
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p75_response_time_per_method
75th percentile response time per method over 2m
The 75th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102222
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p99_9_response_size_per_method
99.9th percentile total response size per method over 2m
The 99.9th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102230
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p90_response_size_per_method
90th percentile total response size per method over 2m
The 90th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102231
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p75_response_size_per_method
75th percentile total response size per method over 2m
The 75th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102232
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p99_9_invididual_sent_message_size_per_method
99.9th percentile individual sent message size per method over 2m
The 99.9th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102240
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p90_invididual_sent_message_size_per_method
90th percentile individual sent message size per method over 2m
The 90th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102241
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_p75_invididual_sent_message_size_per_method
75th percentile individual sent message size per method over 2m
The 75th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102242
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))
frontend: zoekt_configuration_grpc_response_stream_message_count_per_method
Average streaming response message count per-method over 2m
The average number of response messages sent during a streaming RPC method, broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102250
on your Sourcegraph instance.
Technical details
Query:
SHELL((sum(rate(grpc_server_msg_sent_total{grpc_type="server_stream",instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])) by (grpc_method))/(sum(rate(grpc_server_started_total{grpc_type="server_stream",instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])) by (grpc_method)))
frontend: zoekt_configuration_grpc_all_codes_per_method
Response codes rate per-method over 2m
The rate of all generated gRPC response codes per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102260
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_handled_total{grpc_method=~`${zoekt_configuration_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])) by (grpc_method, grpc_code)
Frontend: Zoekt Configuration GRPC "internal error" metrics
frontend: zoekt_configuration_grpc_clients_error_percentage_all_methods
Client baseline error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods (regardless of whether or not there was an internal error), aggregated across all "zoekt_configuration" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102300
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_code!="OK"}[2m])))) / ((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))))))
frontend: zoekt_configuration_grpc_clients_error_percentage_per_method
Client baseline error percentage per-method over 2m
The percentage of gRPC requests that fail per method (regardless of whether or not there was an internal error), aggregated across all "zoekt_configuration" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102301
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_method=~"${zoekt_configuration_method:regex}",grpc_code!="OK"}[2m])) by (grpc_method))) / ((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_method=~"${zoekt_configuration_method:regex}"}[2m])) by (grpc_method))))))
frontend: zoekt_configuration_grpc_clients_all_codes_per_method
Client baseline response codes rate per-method over 2m
The rate of all generated gRPC response codes per method (regardless of whether or not there was an internal error), aggregated across all "zoekt_configuration" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102302
on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_method=~"${zoekt_configuration_method:regex}"}[2m])) by (grpc_method, grpc_code))
frontend: zoekt_configuration_grpc_clients_internal_error_percentage_all_methods
Client-observed gRPC internal error percentage across all methods over 2m
The percentage of gRPC requests that appear to fail due to gRPC internal errors across all methods, aggregated across all "zoekt_configuration" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "zoekt_configuration" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error
) as opposed to normal
application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:
, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102310
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_code!="OK",is_internal_error="true"}[2m])))) / ((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))))))
frontend: zoekt_configuration_grpc_clients_internal_error_percentage_per_method
Client-observed gRPC internal error percentage per-method over 2m
The percentage of gRPC requests that appear to fail to due to gRPC internal errors per method, aggregated across all "zoekt_configuration" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "zoekt_configuration" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error
) as opposed to normal
application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:
, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102311
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_method=~"${zoekt_configuration_method:regex}",grpc_code!="OK",is_internal_error="true"}[2m])) by (grpc_method))) / ((sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_method=~"${zoekt_configuration_method:regex}"}[2m])) by (grpc_method))))))
frontend: zoekt_configuration_grpc_clients_internal_error_all_codes_per_method
Client-observed gRPC internal error response code rate per-method over 2m
The rate of gRPC internal-error response codes per method, aggregated across all "zoekt_configuration" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "zoekt_configuration" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error
) as opposed to normal
application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:
, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102312
on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(grpc_method_status{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",is_internal_error="true",grpc_method=~"${zoekt_configuration_method:regex}"}[2m])) by (grpc_method, grpc_code))
Frontend: Zoekt Configuration GRPC retry metrics
frontend: zoekt_configuration_grpc_clients_retry_percentage_across_all_methods
Client retry percentage across all methods over 2m
The percentage of gRPC requests that were retried across all methods, aggregated across all "zoekt_configuration" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102400
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",is_retried="true"}[2m])))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService"}[2m])))))))
frontend: zoekt_configuration_grpc_clients_retry_percentage_per_method
Client retry percentage per-method over 2m
The percentage of gRPC requests that were retried aggregated across all "zoekt_configuration" clients, broken out per method.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102401
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",is_retried="true",grpc_method=~"${zoekt_configuration_method:regex}"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_method=~"${zoekt_configuration_method:regex}"}[2m])) by (grpc_method))))))
frontend: zoekt_configuration_grpc_clients_retry_count_per_method
Client retry count per-method over 2m
The count of gRPC requests that were retried aggregated across all "zoekt_configuration" clients, broken out per method
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102402
on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"sourcegraph.zoekt.configuration.v1.ZoektConfigurationService",grpc_method=~"${zoekt_configuration_method:regex}",is_retried="true"}[2m])) by (grpc_method))
Frontend: Internal Api GRPC server metrics
frontend: internal_api_grpc_request_rate_all_methods
Request rate across all methods over 2m
The number of gRPC requests received per second across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102500
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m]))
frontend: internal_api_grpc_request_rate_per_method
Request rate per-method over 2m
The number of gRPC requests received per second broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102501
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])) by (grpc_method)
frontend: internal_api_error_percentage_all_methods
Error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102510
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_code!="OK",instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m]))) / (sum(rate(grpc_server_handled_total{instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m]))) ))
frontend: internal_api_grpc_error_percentage_per_method
Error percentage per-method over 2m
The percentage of gRPC requests that fail per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102511
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_method=~`${internal_api_method:regex}`,grpc_code!="OK",instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])) by (grpc_method)) / (sum(rate(grpc_server_handled_total{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])) by (grpc_method)) ))
frontend: internal_api_p99_response_time_per_method
99th percentile response time per method over 2m
The 99th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102520
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p90_response_time_per_method
90th percentile response time per method over 2m
The 90th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102521
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p75_response_time_per_method
75th percentile response time per method over 2m
The 75th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102522
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p99_9_response_size_per_method
99.9th percentile total response size per method over 2m
The 99.9th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102530
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p90_response_size_per_method
90th percentile total response size per method over 2m
The 90th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102531
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p75_response_size_per_method
75th percentile total response size per method over 2m
The 75th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102532
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p99_9_invididual_sent_message_size_per_method
99.9th percentile individual sent message size per method over 2m
The 99.9th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102540
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p90_invididual_sent_message_size_per_method
90th percentile individual sent message size per method over 2m
The 90th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102541
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_p75_invididual_sent_message_size_per_method
75th percentile individual sent message size per method over 2m
The 75th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102542
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))
frontend: internal_api_grpc_response_stream_message_count_per_method
Average streaming response message count per-method over 2m
The average number of response messages sent during a streaming RPC method, broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102550
on your Sourcegraph instance.
Technical details
Query:
SHELL((sum(rate(grpc_server_msg_sent_total{grpc_type="server_stream",instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])) by (grpc_method))/(sum(rate(grpc_server_started_total{grpc_type="server_stream",instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])) by (grpc_method)))
frontend: internal_api_grpc_all_codes_per_method
Response codes rate per-method over 2m
The rate of all generated gRPC response codes per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102560
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_handled_total{grpc_method=~`${internal_api_method:regex}`,instance=~`${internalInstance:regex}`,grpc_service=~"api.internalapi.v1.ConfigService"}[2m])) by (grpc_method, grpc_code)
Frontend: Internal Api GRPC "internal error" metrics
frontend: internal_api_grpc_clients_error_percentage_all_methods
Client baseline error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods (regardless of whether or not there was an internal error), aggregated across all "internal_api" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102600
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",grpc_code!="OK"}[2m])))) / ((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))))))
frontend: internal_api_grpc_clients_error_percentage_per_method
Client baseline error percentage per-method over 2m
The percentage of gRPC requests that fail per method (regardless of whether or not there was an internal error), aggregated across all "internal_api" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102601
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",grpc_method=~"${internal_api_method:regex}",grpc_code!="OK"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",grpc_method=~"${internal_api_method:regex}"}[2m])) by (grpc_method))))))
frontend: internal_api_grpc_clients_all_codes_per_method
Client baseline response codes rate per-method over 2m
The rate of all generated gRPC response codes per method (regardless of whether or not there was an internal error), aggregated across all "internal_api" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102602
on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",grpc_method=~"${internal_api_method:regex}"}[2m])) by (grpc_method, grpc_code))
frontend: internal_api_grpc_clients_internal_error_percentage_all_methods
Client-observed gRPC internal error percentage across all methods over 2m
The percentage of gRPC requests that appear to fail due to gRPC internal errors across all methods, aggregated across all "internal_api" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "internal_api" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error
) as opposed to normal
application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:
, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102610
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",grpc_code!="OK",is_internal_error="true"}[2m])))) / ((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))))))
frontend: internal_api_grpc_clients_internal_error_percentage_per_method
Client-observed gRPC internal error percentage per-method over 2m
The percentage of gRPC requests that appear to fail to due to gRPC internal errors per method, aggregated across all "internal_api" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "internal_api" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error
) as opposed to normal
application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:
, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102611
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",grpc_method=~"${internal_api_method:regex}",grpc_code!="OK",is_internal_error="true"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",grpc_method=~"${internal_api_method:regex}"}[2m])) by (grpc_method))))))
frontend: internal_api_grpc_clients_internal_error_all_codes_per_method
Client-observed gRPC internal error response code rate per-method over 2m
The rate of gRPC internal-error response codes per method, aggregated across all "internal_api" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "internal_api" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error
) as opposed to normal
application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:
, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102612
on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_method_status{grpc_service=~"api.internalapi.v1.ConfigService",is_internal_error="true",grpc_method=~"${internal_api_method:regex}"}[2m])) by (grpc_method, grpc_code))
Frontend: Internal Api GRPC retry metrics
frontend: internal_api_grpc_clients_retry_percentage_across_all_methods
Client retry percentage across all methods over 2m
The percentage of gRPC requests that were retried across all methods, aggregated across all "internal_api" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102700
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"api.internalapi.v1.ConfigService",is_retried="true"}[2m])))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"api.internalapi.v1.ConfigService"}[2m])))))))
frontend: internal_api_grpc_clients_retry_percentage_per_method
Client retry percentage per-method over 2m
The percentage of gRPC requests that were retried aggregated across all "internal_api" clients, broken out per method.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102701
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"api.internalapi.v1.ConfigService",is_retried="true",grpc_method=~"${internal_api_method:regex}"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"api.internalapi.v1.ConfigService",grpc_method=~"${internal_api_method:regex}"}[2m])) by (grpc_method))))))
frontend: internal_api_grpc_clients_retry_count_per_method
Client retry count per-method over 2m
The count of gRPC requests that were retried aggregated across all "internal_api" clients, broken out per method
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102702
on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"api.internalapi.v1.ConfigService",grpc_method=~"${internal_api_method:regex}",is_retried="true"}[2m])) by (grpc_method))
Frontend: Internal service requests
frontend: internal_indexed_search_error_responses
Internal indexed search error responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102800
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(code) (increase(src_zoekt_request_duration_seconds_count{code!~"2.."}[5m])) / ignoring(code) group_left sum(increase(src_zoekt_request_duration_seconds_count[5m])) * 100
frontend: internal_unindexed_search_error_responses
Internal unindexed search error responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102801
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(code) (increase(searcher_service_request_total{code!~"2.."}[5m])) / ignoring(code) group_left sum(increase(searcher_service_request_total[5m])) * 100
frontend: 99th_percentile_gitserver_duration
99th percentile successful gitserver query duration over 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102810
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,category)(rate(src_gitserver_request_duration_seconds_bucket{job=~"(sourcegraph-)?frontend"}[5m])))
frontend: gitserver_error_responses
Gitserver error responses every 5m
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102811
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (category)(increase(src_gitserver_request_duration_seconds_count{job=~"(sourcegraph-)?frontend",code!~"2.."}[5m])) / ignoring(code) group_left sum by (category)(increase(src_gitserver_request_duration_seconds_count{job=~"(sourcegraph-)?frontend"}[5m])) * 100
frontend: observability_test_alert_warning
Warning test alert metric
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102820
on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(owner) (observability_test_metric_warning)
frontend: observability_test_alert_critical
Critical test alert metric
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102821
on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(owner) (observability_test_metric_critical)
Frontend: Authentication API requests
frontend: sign_in_rate
Rate of API requests to sign-in
Rate (QPS) of requests to sign-in
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102900
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(irate(src_http_request_duration_seconds_count{route="sign-in",method="post"}[5m]))
frontend: sign_in_latency_p99
99 percentile of sign-in latency
99% percentile of sign-in latency
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102901
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_http_request_duration_seconds_bucket{route="sign-in",method="post"}[5m])) by (le))
frontend: sign_in_error_rate
Percentage of sign-in requests by http code
Percentage of sign-in requests grouped by http code
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102902
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (code)(irate(src_http_request_duration_seconds_count{route="sign-in",method="post"}[5m]))/ ignoring (code) group_left sum(irate(src_http_request_duration_seconds_count{route="sign-in",method="post"}[5m]))*100
frontend: sign_up_rate
Rate of API requests to sign-up
Rate (QPS) of requests to sign-up
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102910
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(irate(src_http_request_duration_seconds_count{route="sign-up",method="post"}[5m]))
frontend: sign_up_latency_p99
99 percentile of sign-up latency
99% percentile of sign-up latency
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102911
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_http_request_duration_seconds_bucket{route="sign-up",method="post"}[5m])) by (le))
frontend: sign_up_code_percentage
Percentage of sign-up requests by http code
Percentage of sign-up requests grouped by http code
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102912
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (code)(irate(src_http_request_duration_seconds_count{route="sign-up",method="post"}[5m]))/ ignoring (code) group_left sum(irate(src_http_request_duration_seconds_count{route="sign-out"}[5m]))*100
frontend: sign_out_rate
Rate of API requests to sign-out
Rate (QPS) of requests to sign-out
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102920
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(irate(src_http_request_duration_seconds_count{route="sign-out"}[5m]))
frontend: sign_out_latency_p99
99 percentile of sign-out latency
99% percentile of sign-out latency
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102921
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum(rate(src_http_request_duration_seconds_bucket{route="sign-out"}[5m])) by (le))
frontend: sign_out_error_rate
Percentage of sign-out requests that return non-303 http code
Percentage of sign-out requests grouped by http code
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102922
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (code)(irate(src_http_request_duration_seconds_count{route="sign-out"}[5m]))/ ignoring (code) group_left sum(irate(src_http_request_duration_seconds_count{route="sign-out"}[5m]))*100
frontend: account_failed_sign_in_attempts
Rate of failed sign-in attempts
Failed sign-in attempts per minute
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102930
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_frontend_account_failed_sign_in_attempts_total[1m]))
frontend: account_lockouts
Rate of account lockouts
Account lockouts per minute
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=102931
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_frontend_account_lockouts_total[1m]))
Frontend: External HTTP Request Rate
frontend: external_http_request_rate_by_host
Rate of external HTTP requests by host over 1m
Shows the rate of external HTTP requests made by Sourcegraph to other services, broken down by host.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103000
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (host) (rate(src_http_client_external_request_count{host=~`${httpRequestHost:regex}`}[1m]))
frontend: external_http_request_rate_by_host_by_code
Rate of external HTTP requests by host and response code over 1m
Shows the rate of external HTTP requests made by Sourcegraph to other services, broken down by host and response code.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103010
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (host, status_code) (rate(src_http_client_external_request_count{host=~`${httpRequestHost:regex}`}[1m]))
Frontend: Cody API requests
frontend: cody_api_rate
Rate of API requests to cody endpoints (excluding GraphQL)
Rate (QPS) of requests to cody related endpoints. completions.stream is for the conversational endpoints. completions.code is for the code auto-complete endpoints.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103100
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (route, code)(irate(src_http_request_duration_seconds_count{route=~"^completions.*"}[5m]))
Frontend: Cloud KMS and cache
frontend: cloudkms_cryptographic_requests
Cryptographic requests to Cloud KMS every 1m
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103200
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_cloudkms_cryptographic_total[1m]))
frontend: encryption_cache_hit_ratio
Average encryption cache hit ratio per workload
- Encryption cache hit ratio (hits/(hits+misses)) - minimum across all instances of a workload.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103201
on your Sourcegraph instance.
Technical details
Query:
SHELLmin by (kubernetes_name) (src_encryption_cache_hit_total/(src_encryption_cache_hit_total+src_encryption_cache_miss_total))
frontend: encryption_cache_evictions
Rate of encryption cache evictions - sum across all instances of a given workload
- Rate of encryption cache evictions (caused by cache exceeding its maximum size) - sum across all instances of a workload
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103202
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (kubernetes_name) (irate(src_encryption_cache_eviction_total[5m]))
Frontend: Database connections
frontend: max_open_conns
Maximum open
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103300
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_max_open{app_name="frontend"})
frontend: open_conns
Established
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103301
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_open{app_name="frontend"})
frontend: in_use
Used
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103310
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_in_use{app_name="frontend"})
frontend: idle
Idle
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103311
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (src_pgsql_conns_idle{app_name="frontend"})
frontend: mean_blocked_seconds_per_conn_request
Mean blocked seconds per conn request
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103320
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_blocked_seconds{app_name="frontend"}[5m])) / sum by (app_name, db_name) (increase(src_pgsql_conns_waited_for{app_name="frontend"}[5m]))
frontend: closed_max_idle
Closed by SetMaxIdleConns
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103330
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_idle{app_name="frontend"}[5m]))
frontend: closed_max_lifetime
Closed by SetConnMaxLifetime
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103331
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_lifetime{app_name="frontend"}[5m]))
frontend: closed_max_idle_time
Closed by SetConnMaxIdleTime
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103332
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (app_name, db_name) (increase(src_pgsql_conns_closed_max_idle_time{app_name="frontend"}[5m]))
Frontend: Container monitoring (not available on server)
frontend: container_missing
Container missing
This value is the number of times a container has not been seen for more than one minute. If you observe this value change independent of deployment events (such as an upgrade), it could indicate pods are being OOM killed or terminated for some other reasons.
- Kubernetes:
- Determine if the pod was OOM killed using
kubectl describe pod (frontend\|sourcegraph-frontend)
(look forOOMKilled: true
) and, if so, consider increasing the memory limit in the relevantDeployment.yaml
. - Check the logs before the container restarted to see if there are
panic:
messages or similar usingkubectl logs -p (frontend\|sourcegraph-frontend)
.
- Determine if the pod was OOM killed using
- Docker Compose:
- Determine if the pod was OOM killed using
docker inspect -f '\{\{json .State\}\}' (frontend\|sourcegraph-frontend)
(look for"OOMKilled":true
) and, if so, consider increasing the memory limit of the (frontend|sourcegraph-frontend) container indocker-compose.yml
. - Check the logs before the container restarted to see if there are
panic:
messages or similar usingdocker logs (frontend\|sourcegraph-frontend)
(note this will include logs from the previous and currently running container).
- Determine if the pod was OOM killed using
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103400
on your Sourcegraph instance.
Technical details
Query:
SHELLcount by(name) ((time() - container_last_seen{name=~"^(frontend|sourcegraph-frontend).*"}) > 60)
frontend: container_cpu_usage
Container cpu usage total (1m average) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103401
on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_cpu_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend).*"}
frontend: container_memory_usage
Container memory usage by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103402
on your Sourcegraph instance.
Technical details
Query:
SHELLcadvisor_container_memory_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend).*"}
frontend: fs_io_operations
Filesystem reads and writes rate by instance over 1h
This value indicates the number of filesystem read and write operations by containers of this service. When extremely high, this can indicate a resource usage problem, or can cause problems with the service itself, especially if high values or spikes correlate with {{CONTAINER_NAME}} issues.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103403
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(name) (rate(container_fs_reads_total{name=~"^(frontend|sourcegraph-frontend).*"}[1h]) + rate(container_fs_writes_total{name=~"^(frontend|sourcegraph-frontend).*"}[1h]))
Frontend: Provisioning indicators (not available on server)
frontend: provisioning_container_cpu_usage_long_term
Container cpu usage total (90th percentile over 1d) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103500
on your Sourcegraph instance.
Technical details
Query:
SHELLquantile_over_time(0.9, cadvisor_container_cpu_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend).*"}[1d])
frontend: provisioning_container_memory_usage_long_term
Container memory usage (1d maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103501
on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend).*"}[1d])
frontend: provisioning_container_cpu_usage_short_term
Container cpu usage total (5m maximum) across all cores by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103510
on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_cpu_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend).*"}[5m])
frontend: provisioning_container_memory_usage_short_term
Container memory usage (5m maximum) by instance
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103511
on your Sourcegraph instance.
Technical details
Query:
SHELLmax_over_time(cadvisor_container_memory_usage_percentage_total{name=~"^(frontend|sourcegraph-frontend).*"}[5m])
frontend: container_oomkill_events_total
Container OOMKILL events total by instance
This value indicates the total number of times the container main process or child processes were terminated by OOM killer. When it occurs frequently, it is an indicator of underprovisioning.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103512
on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (name) (container_oom_events_total{name=~"^(frontend|sourcegraph-frontend).*"})
Frontend: Golang runtime monitoring
frontend: go_goroutines
Maximum active goroutines
A high value here indicates a possible goroutine leak.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103600
on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_goroutines{job=~".*(frontend|sourcegraph-frontend)"})
frontend: go_gc_duration_seconds
Maximum go garbage collection duration
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103601
on your Sourcegraph instance.
Technical details
Query:
SHELLmax by(instance) (go_gc_duration_seconds{job=~".*(frontend|sourcegraph-frontend)"})
Frontend: Kubernetes monitoring (only available on Kubernetes)
frontend: pods_available_percentage
Percentage pods available
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103700
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by(app) (up{app=~".*(frontend|sourcegraph-frontend)"}) / count by (app) (up{app=~".*(frontend|sourcegraph-frontend)"}) * 100
Frontend: Search: Ranking
frontend: total_search_clicks
Total number of search clicks over 6h
The total number of search clicks across all search types over a 6 hour window.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103800
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (ranked) (increase(src_search_ranking_result_clicked_count[6h]))
frontend: percent_clicks_on_top_search_result
Percent of clicks on top search result over 6h
The percent of clicks that were on the top search result, excluding searches with very few results (3 or fewer).
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103801
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (ranked) (increase(src_search_ranking_result_clicked_bucket{le="1",resultsLength=">3"}[6h])) / sum by (ranked) (increase(src_search_ranking_result_clicked_count[6h])) * 100
frontend: percent_clicks_on_top_3_search_results
Percent of clicks on top 3 search results over 6h
The percent of clicks that were on the first 3 search results, excluding searches with very few results (3 or fewer).
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103802
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (ranked) (increase(src_search_ranking_result_clicked_bucket{le="3",resultsLength=">3"}[6h])) / sum by (ranked) (increase(src_search_ranking_result_clicked_count[6h])) * 100
frontend: distribution_of_clicked_search_result_type_over_6h_in_percent
Distribution of clicked search result type over 6h
The distribution of clicked search results by result type. At every point in time, the values should sum to 100.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103810
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_search_ranking_result_clicked_count{type="repo"}[6h])) / sum(increase(src_search_ranking_result_clicked_count[6h])) * 100
frontend: percent_zoekt_searches_hitting_flush_limit
Percent of zoekt searches that hit the flush time limit
The percent of Zoekt searches that hit the flush time limit. These searches don`t visit all matches, so they could be missing relevant results, or be non-deterministic.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103811
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(zoekt_final_aggregate_size_count{reason="timer_expired"}[1d])) / sum(increase(zoekt_final_aggregate_size_count[1d])) * 100
Frontend: Email delivery
frontend: email_delivery_failures
Email delivery failure rate over 30 minutes
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103900
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_email_send{success="false"}[30m])) / sum(increase(src_email_send[30m])) * 100
frontend: email_deliveries_total
Total emails successfully delivered every 30 minutes
Total emails successfully delivered.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103910
on your Sourcegraph instance.
Technical details
Query:
SHELLsum (increase(src_email_send{success="true"}[30m]))
frontend: email_deliveries_by_source
Emails successfully delivered every 30 minutes by source
Emails successfully delivered by source, i.e. product feature.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=103911
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (email_source) (increase(src_email_send{success="true"}[30m]))
Frontend: Sentinel queries (only on sourcegraph.com)
frontend: mean_successful_sentinel_duration_over_2h
Mean successful sentinel search duration over 2h
Mean search duration for all successful sentinel queries
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104000
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_search_response_latency_seconds_sum{source=~`searchblitz.*`, status=`success`}[2h])) / sum(rate(src_search_response_latency_seconds_count{source=~`searchblitz.*`, status=`success`}[2h]))
frontend: mean_sentinel_stream_latency_over_2h
Mean successful sentinel stream latency over 2h
Mean time to first result for all successful streaming sentinel queries
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104001
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_search_streaming_latency_seconds_sum{source=~"searchblitz.*"}[2h])) / sum(rate(src_search_streaming_latency_seconds_count{source=~"searchblitz.*"}[2h]))
frontend: 90th_percentile_successful_sentinel_duration_over_2h
90th percentile successful sentinel search duration over 2h
90th percentile search duration for all successful sentinel queries
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104010
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le)(label_replace(rate(src_search_response_latency_seconds_bucket{source=~"searchblitz.*", status="success"}[2h]), "source", "$1", "source", "searchblitz_(.*)")))
frontend: 90th_percentile_sentinel_stream_latency_over_2h
90th percentile successful sentinel stream latency over 2h
90th percentile time to first result for all successful streaming sentinel queries
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104011
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le)(label_replace(rate(src_search_streaming_latency_seconds_bucket{source=~"searchblitz.*"}[2h]), "source", "$1", "source", "searchblitz_(.*)")))
frontend: mean_successful_sentinel_duration_by_query
Mean successful sentinel search duration by query
Mean search duration for successful sentinel queries, broken down by query. Useful for debugging whether a slowdown is limited to a specific type of query.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104020
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_search_response_latency_seconds_sum{source=~"searchblitz.*", status="success"}[$sentinel_sampling_duration])) by (source) / sum(rate(src_search_response_latency_seconds_count{source=~"searchblitz.*", status="success"}[$sentinel_sampling_duration])) by (source)
frontend: mean_sentinel_stream_latency_by_query
Mean successful sentinel stream latency by query
Mean time to first result for successful streaming sentinel queries, broken down by query. Useful for debugging whether a slowdown is limited to a specific type of query.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104021
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_search_streaming_latency_seconds_sum{source=~"searchblitz.*"}[$sentinel_sampling_duration])) by (source) / sum(rate(src_search_streaming_latency_seconds_count{source=~"searchblitz.*"}[$sentinel_sampling_duration])) by (source)
frontend: 90th_percentile_successful_sentinel_duration_by_query
90th percentile successful sentinel search duration by query
90th percentile search duration for successful sentinel queries, broken down by query. Useful for debugging whether a slowdown is limited to a specific type of query.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104030
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum(rate(src_search_response_latency_seconds_bucket{source=~"searchblitz.*", status="success"}[$sentinel_sampling_duration])) by (le, source))
frontend: 90th_percentile_successful_stream_latency_by_query
90th percentile successful sentinel stream latency by query
90th percentile time to first result for successful streaming sentinel queries, broken down by query. Useful for debugging whether a slowdown is limited to a specific type of query.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104031
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum(rate(src_search_streaming_latency_seconds_bucket{source=~"searchblitz.*"}[$sentinel_sampling_duration])) by (le, source))
frontend: 90th_percentile_unsuccessful_duration_by_query
90th percentile unsuccessful sentinel search duration by query
90th percentile search duration of unsuccessful sentinel queries (by error or timeout), broken down by query. Useful for debugging how the performance of failed requests affect UX.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104040
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum(rate(src_search_response_latency_seconds_bucket{source=~`searchblitz.*`, status!=`success`}[$sentinel_sampling_duration])) by (le, source))
frontend: 75th_percentile_successful_sentinel_duration_by_query
75th percentile successful sentinel search duration by query
75th percentile search duration of successful sentinel queries, broken down by query. Useful for debugging whether a slowdown is limited to a specific type of query.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104050
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_search_response_latency_seconds_bucket{source=~"searchblitz.*", status="success"}[$sentinel_sampling_duration])) by (le, source))
frontend: 75th_percentile_successful_stream_latency_by_query
75th percentile successful sentinel stream latency by query
75th percentile time to first result for successful streaming sentinel queries, broken down by query. Useful for debugging whether a slowdown is limited to a specific type of query.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104051
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_search_streaming_latency_seconds_bucket{source=~"searchblitz.*"}[$sentinel_sampling_duration])) by (le, source))
frontend: 75th_percentile_unsuccessful_duration_by_query
75th percentile unsuccessful sentinel search duration by query
75th percentile search duration of unsuccessful sentinel queries (by error or timeout), broken down by query. Useful for debugging how the performance of failed requests affect UX.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104060
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum(rate(src_search_response_latency_seconds_bucket{source=~`searchblitz.*`, status!=`success`}[$sentinel_sampling_duration])) by (le, source))
frontend: unsuccessful_status_rate
Unsuccessful status rate
The rate of unsuccessful sentinel queries, broken down by failure type.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104070
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_graphql_search_response{source=~"searchblitz.*", status!="success"}[$sentinel_sampling_duration])) by (status)
Frontend: Incoming webhooks
frontend: p95_time_to_handle_incoming_webhooks
P95 time to handle incoming webhooks
p95 response time to incoming webhook requests from code hosts.
Increases in response time can point to too much load on the database to keep up with the incoming requests.
See this documentation page for more details on webhook requests: (https://sourcegraph.com/docs/admin/config/webhooks/incoming)
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104100
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum (rate(src_http_request_duration_seconds_bucket{route=~"webhooks|github.webhooks|gitlab.webhooks|bitbucketServer.webhooks|bitbucketCloud.webhooks"}[5m])) by (le, route))
Frontend: Search aggregations: proactive and expanded search aggregations
frontend: insights_aggregations_total
Aggregate search aggregations operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104200
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_insights_aggregations_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: insights_aggregations_99th_percentile_duration
Aggregate successful search aggregations operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104201
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_insights_aggregations_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: insights_aggregations_errors_total
Aggregate search aggregations operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104202
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_insights_aggregations_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: insights_aggregations_error_rate
Aggregate search aggregations operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104203
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_insights_aggregations_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum(increase(src_insights_aggregations_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum(increase(src_insights_aggregations_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
frontend: insights_aggregations_total
Search aggregations operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104210
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,extended_mode)(increase(src_insights_aggregations_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: insights_aggregations_99th_percentile_duration
99th percentile successful search aggregations operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104211
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op,extended_mode)(rate(src_insights_aggregations_duration_seconds_bucket{job=~"^(frontend|sourcegraph-frontend).*"}[5m])))
frontend: insights_aggregations_errors_total
Search aggregations operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104212
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,extended_mode)(increase(src_insights_aggregations_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))
frontend: insights_aggregations_error_rate
Search aggregations operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/frontend/frontend?viewPanel=104213
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,extended_mode)(increase(src_insights_aggregations_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) / (sum by (op,extended_mode)(increase(src_insights_aggregations_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m])) + sum by (op,extended_mode)(increase(src_insights_aggregations_errors_total{job=~"^(frontend|sourcegraph-frontend).*"}[5m]))) * 100
Git Server
Stores, manages, and operates Git repositories.
To see this dashboard, visit /-/debug/grafana/d/gitserver/gitserver
on your Sourcegraph instance.
gitserver: go_routines
Go routines
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100000
on your Sourcegraph instance.
Technical details
Query:
SHELLgo_goroutines{app="gitserver", instance=~`${shard:regex}`}
gitserver: cpu_throttling_time
Container CPU throttling time %
- A high value indicates that the container is spending too much time waiting for CPU cycles.
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100010
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) ((rate(container_cpu_cfs_throttled_periods_total{container_label_io_kubernetes_container_name="gitserver", container_label_io_kubernetes_pod_name=~`${shard:regex}`}[5m]) / rate(container_cpu_cfs_periods_total{container_label_io_kubernetes_container_name="gitserver", container_label_io_kubernetes_pod_name=~`${shard:regex}`}[5m])) * 100)
gitserver: cpu_usage_seconds
Cpu usage seconds
- This value should not exceed 75% of the CPU limit over a longer period of time.
-
We cannot alert on this as we don`t know the resource allocation.
-
If this value is high for a longer time, consider increasing the CPU limit for the container.
-
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100011
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (container_label_io_kubernetes_pod_name) (rate(container_cpu_usage_seconds_total{container_label_io_kubernetes_container_name="gitserver", container_label_io_kubernetes_pod_name=~`${shard:regex}`}[5m]))
gitserver: disk_space_remaining
Disk space remaining
Indicates disk space remaining for each gitserver instance, which is used to determine when to start evicting least-used repository clones from disk (default 10%, configured by SRC_REPOS_DESIRED_PERCENT_FREE
).
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100020
on your Sourcegraph instance.
Technical details
Query:
SHELL(src_gitserver_disk_space_available{instance=~`${shard:regex}`} / src_gitserver_disk_space_total{instance=~`${shard:regex}`}) * 100
gitserver: running_git_commands
Git commands running on each gitserver instance
A high value signals load.
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100030
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (instance, cmd) (src_gitserver_exec_running{instance=~`${shard:regex}`})
gitserver: git_commands_received
Rate of git commands received
per second rate per command
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100031
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (cmd) (rate(src_gitserver_exec_duration_seconds_count{instance=~`${shard:regex}`}[5m]))
gitserver: echo_command_duration_test
Echo test command duration
Refer to the alerts reference for 2 alerts related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100040
on your Sourcegraph instance.
Technical details
Query:
SHELLmax(src_gitserver_echo_duration_seconds)
gitserver: repo_corrupted
Number of times a repo corruption has been identified
A non-null value here indicates that a problem has been detected with the gitserver repository storage. Repository corruptions are never expected. This is a real issue. Gitserver should try to recover from them by recloning repositories, but this may take a while depending on repo size.
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100041
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_repo_corrupted[5m]))
gitserver: repository_clone_queue_size
Repository clone queue size
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100050
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_gitserver_clone_queue)
gitserver: src_gitserver_repo_count
Number of repositories on gitserver
This metric is only for informational purposes. It indicates the total number of repositories on gitserver.
It does not indicate any problems with the instance.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100051
on your Sourcegraph instance.
Technical details
Query:
SHELLsrc_gitserver_repo_count
Git Server: Gitservice for internal cloning
gitserver: aggregate_gitservice_request_duration
95th percentile gitservice request duration aggregate
A high value means any internal service trying to clone a repo from gitserver is slowed down.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100100
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_gitserver_gitservice_duration_seconds_bucket{type=`gitserver`, error=`false`}[5m])) by (le))
gitserver: gitservice_request_duration
95th percentile gitservice request duration per shard
A high value means any internal service trying to clone a repo from gitserver is slowed down.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100101
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_gitserver_gitservice_duration_seconds_bucket{type=`gitserver`, error=`false`, instance=~`${shard:regex}`}[5m])) by (le, instance))
gitserver: aggregate_gitservice_error_request_duration
95th percentile gitservice error request duration aggregate
95th percentile gitservice error request duration aggregate
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100110
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_gitserver_gitservice_duration_seconds_bucket{type=`gitserver`, error=`true`}[5m])) by (le))
gitserver: gitservice_request_duration
95th percentile gitservice error request duration per shard
95th percentile gitservice error request duration per shard
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100111
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_gitserver_gitservice_duration_seconds_bucket{type=`gitserver`, error=`true`, instance=~`${shard:regex}`}[5m])) by (le, instance))
gitserver: aggregate_gitservice_request_rate
Aggregate gitservice request rate
Aggregate gitservice request rate
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100120
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_gitservice_duration_seconds_count{type=`gitserver`, error=`false`}[5m]))
gitserver: gitservice_request_rate
Gitservice request rate per shard
Per shard gitservice request rate
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100121
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_gitservice_duration_seconds_count{type=`gitserver`, error=`false`, instance=~`${shard:regex}`}[5m]))
gitserver: aggregate_gitservice_request_error_rate
Aggregate gitservice request error rate
Aggregate gitservice request error rate
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100130
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_gitservice_duration_seconds_count{type=`gitserver`, error=`true`}[5m]))
gitserver: gitservice_request_error_rate
Gitservice request error rate per shard
Per shard gitservice request error rate
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100131
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(src_gitserver_gitservice_duration_seconds_count{type=`gitserver`, error=`true`, instance=~`${shard:regex}`}[5m]))
gitserver: aggregate_gitservice_requests_running
Aggregate gitservice requests running
Aggregate gitservice requests running
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100140
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_gitserver_gitservice_running{type=`gitserver`})
gitserver: gitservice_requests_running
Gitservice requests running per shard
Per shard gitservice requests running
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100141
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(src_gitserver_gitservice_running{type=`gitserver`, instance=~`${shard:regex}`}) by (instance)
Git Server: Gitserver cleanup jobs
gitserver: janitor_running
Janitor process is running
1, if the janitor process is currently running
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100200
on your Sourcegraph instance.
Technical details
Query:
SHELLmax by (instance) (src_gitserver_janitor_running{instance=~`${shard:regex}`})
gitserver: janitor_job_duration
95th percentile job run duration
95th percentile job run duration
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100210
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.95, sum(rate(src_gitserver_janitor_job_duration_seconds_bucket{instance=~`${shard:regex}`}[5m])) by (le, job_name))
gitserver: janitor_job_failures
Failures over 5m (by job)
the rate of failures over 5m (by job)
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100220
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (job_name) (rate(src_gitserver_janitor_job_duration_seconds_count{instance=~`${shard:regex}`,success="false"}[5m]))
gitserver: repos_removed
Repositories removed due to disk pressure
Repositories removed due to disk pressure
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100230
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (instance) (rate(src_gitserver_repos_removed_disk_pressure{instance=~`${shard:regex}`}[5m]))
gitserver: non_existent_repos_removed
Repositories removed because they are not defined in the DB
Repositoriess removed because they are not defined in the DB
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100240
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (instance) (increase(src_gitserver_non_existing_repos_removed[5m]))
gitserver: sg_maintenance_reason
Successful sg maintenance jobs over 1h (by reason)
the rate of successful sg maintenance jobs and the reason why they were triggered
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100250
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (reason) (rate(src_gitserver_maintenance_status{success="true"}[1h]))
gitserver: git_prune_skipped
Successful git prune jobs over 1h
the rate of successful git prune jobs over 1h and whether they were skipped
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100260
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (skipped) (rate(src_gitserver_prune_status{success="true"}[1h]))
Git Server: Search
gitserver: search_latency
Mean time until first result is sent
Mean latency (time to first result) of gitserver search requests
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100300
on your Sourcegraph instance.
Technical details
Query:
SHELLrate(src_gitserver_search_latency_seconds_sum[5m]) / rate(src_gitserver_search_latency_seconds_count[5m])
gitserver: search_duration
Mean search duration
Mean duration of gitserver search requests
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100301
on your Sourcegraph instance.
Technical details
Query:
SHELLrate(src_gitserver_search_duration_seconds_sum[5m]) / rate(src_gitserver_search_duration_seconds_count[5m])
gitserver: search_rate
Rate of searches run by pod
The rate of searches executed on gitserver by pod
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100310
on your Sourcegraph instance.
Technical details
Query:
SHELLrate(src_gitserver_search_latency_seconds_count{instance=~`${shard:regex}`}[5m])
gitserver: running_searches
Number of searches currently running by pod
The number of searches currently executing on gitserver by pod
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100311
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (instance) (src_gitserver_search_running{instance=~`${shard:regex}`})
Git Server: Gitserver: Gitserver Backend
gitserver: concurrent_backend_operations
Number of concurrently running backend operations
The number of requests that are currently being handled by gitserver backend layer, at the point in time of scraping.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100400
on your Sourcegraph instance.
Technical details
Query:
SHELLsrc_gitserver_backend_concurrent_operations
gitserver: gitserver_backend_total
Aggregate operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100410
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_backend_total{job=~"^gitserver.*"}[5m]))
gitserver: gitserver_backend_99th_percentile_duration
Aggregate successful operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100411
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_gitserver_backend_duration_seconds_bucket{job=~"^gitserver.*"}[5m]))
gitserver: gitserver_backend_errors_total
Aggregate operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100412
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_backend_errors_total{job=~"^gitserver.*"}[5m]))
gitserver: gitserver_backend_error_rate
Aggregate operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100413
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_backend_errors_total{job=~"^gitserver.*"}[5m])) / (sum(increase(src_gitserver_backend_total{job=~"^gitserver.*"}[5m])) + sum(increase(src_gitserver_backend_errors_total{job=~"^gitserver.*"}[5m]))) * 100
gitserver: gitserver_backend_total
operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100420
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_backend_total{job=~"^gitserver.*"}[5m]))
gitserver: gitserver_backend_99th_percentile_duration
99th percentile successful operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100421
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_gitserver_backend_duration_seconds_bucket{job=~"^gitserver.*"}[5m])))
gitserver: gitserver_backend_errors_total
operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100422
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_backend_errors_total{job=~"^gitserver.*"}[5m]))
gitserver: gitserver_backend_error_rate
operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100423
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_gitserver_backend_errors_total{job=~"^gitserver.*"}[5m])) / (sum by (op)(increase(src_gitserver_backend_total{job=~"^gitserver.*"}[5m])) + sum by (op)(increase(src_gitserver_backend_errors_total{job=~"^gitserver.*"}[5m]))) * 100
Git Server: Gitserver: Gitserver Client
gitserver: gitserver_client_total
Aggregate graphql operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100500
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_total{job=~"^*.*"}[5m]))
gitserver: gitserver_client_99th_percentile_duration
Aggregate successful graphql operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100501
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^*.*"}[5m]))
gitserver: gitserver_client_errors_total
Aggregate graphql operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100502
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^*.*"}[5m]))
gitserver: gitserver_client_error_rate
Aggregate graphql operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100503
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_gitserver_client_errors_total{job=~"^*.*"}[5m])) / (sum(increase(src_gitserver_client_total{job=~"^*.*"}[5m])) + sum(increase(src_gitserver_client_errors_total{job=~"^*.*"}[5m]))) * 100
gitserver: gitserver_client_total
Graphql operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100510
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_total{job=~"^*.*"}[5m]))
gitserver: gitserver_client_99th_percentile_duration
99th percentile successful graphql operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100511
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op,scope)(rate(src_gitserver_client_duration_seconds_bucket{job=~"^*.*"}[5m])))
gitserver: gitserver_client_errors_total
Graphql operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100512
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^*.*"}[5m]))
gitserver: gitserver_client_error_rate
Graphql operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100513
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^*.*"}[5m])) / (sum by (op,scope)(increase(src_gitserver_client_total{job=~"^*.*"}[5m])) + sum by (op,scope)(increase(src_gitserver_client_errors_total{job=~"^*.*"}[5m]))) * 100
Git Server: Repos disk I/O metrics
gitserver: repos_disk_reads_sec
Read request rate over 1m (per instance)
The number of read requests that were issued to the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100600
on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_completed_total{instance=~`node-exporter.*`}[1m])))))
gitserver: repos_disk_writes_sec
Write request rate over 1m (per instance)
The number of write requests that were issued to the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100601
on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_completed_total{instance=~`node-exporter.*`}[1m])))))
gitserver: repos_disk_read_throughput
Read throughput over 1m (per instance)
The amount of data that was read from the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100610
on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_read_bytes_total{instance=~`node-exporter.*`}[1m])))))
gitserver: repos_disk_write_throughput
Write throughput over 1m (per instance)
The amount of data that was written to the device per second.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100611
on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_written_bytes_total{instance=~`node-exporter.*`}[1m])))))
gitserver: repos_disk_read_duration
Average read duration over 1m (per instance)
The average time for read requests issued to the device to be served. This includes the time spent by the requests in queue and the time spent servicing them.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100620
on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_read_time_seconds_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_completed_total{instance=~`node-exporter.*`}[1m])))))))
gitserver: repos_disk_write_duration
Average write duration over 1m (per instance)
The average time for write requests issued to the device to be served. This includes the time spent by the requests in queue and the time spent servicing them.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100621
on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_write_time_seconds_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_completed_total{instance=~`node-exporter.*`}[1m])))))))
gitserver: repos_disk_read_request_size
Average read request size over 1m (per instance)
The average size of read requests that were issued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100630
on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_read_bytes_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_completed_total{instance=~`node-exporter.*`}[1m])))))))
gitserver: repos_disk_write_request_size)
Average write request size over 1m (per instance)
The average size of write requests that were issued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100631
on your Sourcegraph instance.
Technical details
Query:
SHELL(((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_written_bytes_total{instance=~`node-exporter.*`}[1m])))))) / ((max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_completed_total{instance=~`node-exporter.*`}[1m])))))))
gitserver: repos_disk_reads_merged_sec
Merged read request rate over 1m (per instance)
The number of read requests merged per second that were queued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100640
on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_reads_merged_total{instance=~`node-exporter.*`}[1m])))))
gitserver: repos_disk_writes_merged_sec
Merged writes request rate over 1m (per instance)
The number of write requests merged per second that were queued to the device.
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100641
on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_writes_merged_total{instance=~`node-exporter.*`}[1m])))))
gitserver: repos_disk_average_queue_size
Average queue size over 1m (per instance)
The number of I/O operations that were being queued or being serviced. See https://blog.actorsfit.com/a?ID=00200-428fa2ac-e338-4540-848c-af9a3eb1ebd2 for background (avgqu-sz).
Note: Disk statistics are per device, not per service. In certain environments (such as common docker-compose setups), gitserver could be one of many services using this disk. These statistics are best interpreted as the load experienced by the device gitserver is using, not the load gitserver is solely responsible for causing.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100650
on your Sourcegraph instance.
Technical details
Query:
SHELL(max by (instance) (gitserver_mount_point_info{mount_name="reposDir",instance=~`${shard:regex}`} * on (device, nodename) group_left() (max by (device, nodename) (rate(node_disk_io_time_weighted_seconds_total{instance=~`node-exporter.*`}[1m])))))
Git Server: Gitserver GRPC server metrics
gitserver: gitserver_grpc_request_rate_all_methods
Request rate across all methods over 2m
The number of gRPC requests received per second across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100700
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m]))
gitserver: gitserver_grpc_request_rate_per_method
Request rate per-method over 2m
The number of gRPC requests received per second broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100701
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_started_total{grpc_method=~`${gitserver_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])) by (grpc_method)
gitserver: gitserver_error_percentage_all_methods
Error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100710
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_code!="OK",instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m]))) / (sum(rate(grpc_server_handled_total{instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m]))) ))
gitserver: gitserver_grpc_error_percentage_per_method
Error percentage per-method over 2m
The percentage of gRPC requests that fail per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100711
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ( (sum(rate(grpc_server_handled_total{grpc_method=~`${gitserver_method:regex}`,grpc_code!="OK",instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])) by (grpc_method)) / (sum(rate(grpc_server_handled_total{grpc_method=~`${gitserver_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])) by (grpc_method)) ))
gitserver: gitserver_p99_response_time_per_method
99th percentile response time per method over 2m
The 99th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100720
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${gitserver_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: gitserver_p90_response_time_per_method
90th percentile response time per method over 2m
The 90th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100721
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${gitserver_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: gitserver_p75_response_time_per_method
75th percentile response time per method over 2m
The 75th percentile response time per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100722
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(grpc_server_handling_seconds_bucket{grpc_method=~`${gitserver_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: gitserver_p99_9_response_size_per_method
99.9th percentile total response size per method over 2m
The 99.9th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100730
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${gitserver_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: gitserver_p90_response_size_per_method
90th percentile total response size per method over 2m
The 90th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100731
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${gitserver_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: gitserver_p75_response_size_per_method
75th percentile total response size per method over 2m
The 75th percentile total per-RPC response size per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100732
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_bytes_per_rpc_bucket{grpc_method=~`${gitserver_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: gitserver_p99_9_invididual_sent_message_size_per_method
99.9th percentile individual sent message size per method over 2m
The 99.9th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100740
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.999, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${gitserver_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: gitserver_p90_invididual_sent_message_size_per_method
90th percentile individual sent message size per method over 2m
The 90th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100741
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.90, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${gitserver_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: gitserver_p75_invididual_sent_message_size_per_method
75th percentile individual sent message size per method over 2m
The 75th percentile size of every individual protocol buffer size sent by the service per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100742
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.75, sum by (le, name, grpc_method)(rate(src_grpc_server_sent_individual_message_size_bytes_per_rpc_bucket{grpc_method=~`${gitserver_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])))
gitserver: gitserver_grpc_response_stream_message_count_per_method
Average streaming response message count per-method over 2m
The average number of response messages sent during a streaming RPC method, broken out per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100750
on your Sourcegraph instance.
Technical details
Query:
SHELL((sum(rate(grpc_server_msg_sent_total{grpc_type="server_stream",instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])) by (grpc_method))/(sum(rate(grpc_server_started_total{grpc_type="server_stream",instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])) by (grpc_method)))
gitserver: gitserver_grpc_all_codes_per_method
Response codes rate per-method over 2m
The rate of all generated gRPC response codes per method, aggregated across all instances.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100760
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(rate(grpc_server_handled_total{grpc_method=~`${gitserver_method:regex}`,instance=~`${shard:regex}`,grpc_service=~"gitserver.v1.GitserverService"}[2m])) by (grpc_method, grpc_code)
Git Server: Gitserver GRPC "internal error" metrics
gitserver: gitserver_grpc_clients_error_percentage_all_methods
Client baseline error percentage across all methods over 2m
The percentage of gRPC requests that fail across all methods (regardless of whether or not there was an internal error), aggregated across all "gitserver" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100800
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",grpc_code!="OK"}[2m])))) / ((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService"}[2m])))))))
gitserver: gitserver_grpc_clients_error_percentage_per_method
Client baseline error percentage per-method over 2m
The percentage of gRPC requests that fail per method (regardless of whether or not there was an internal error), aggregated across all "gitserver" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100801
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",grpc_method=~"${gitserver_method:regex}",grpc_code!="OK"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",grpc_method=~"${gitserver_method:regex}"}[2m])) by (grpc_method))))))
gitserver: gitserver_grpc_clients_all_codes_per_method
Client baseline response codes rate per-method over 2m
The rate of all generated gRPC response codes per method (regardless of whether or not there was an internal error), aggregated across all "gitserver" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100802
on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",grpc_method=~"${gitserver_method:regex}"}[2m])) by (grpc_method, grpc_code))
gitserver: gitserver_grpc_clients_internal_error_percentage_all_methods
Client-observed gRPC internal error percentage across all methods over 2m
The percentage of gRPC requests that appear to fail due to gRPC internal errors across all methods, aggregated across all "gitserver" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "gitserver" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error
) as opposed to normal
application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:
, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100810
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",grpc_code!="OK",is_internal_error="true"}[2m])))) / ((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService"}[2m])))))))
gitserver: gitserver_grpc_clients_internal_error_percentage_per_method
Client-observed gRPC internal error percentage per-method over 2m
The percentage of gRPC requests that appear to fail to due to gRPC internal errors per method, aggregated across all "gitserver" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "gitserver" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error
) as opposed to normal
application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:
, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100811
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",grpc_method=~"${gitserver_method:regex}",grpc_code!="OK",is_internal_error="true"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",grpc_method=~"${gitserver_method:regex}"}[2m])) by (grpc_method))))))
gitserver: gitserver_grpc_clients_internal_error_all_codes_per_method
Client-observed gRPC internal error response code rate per-method over 2m
The rate of gRPC internal-error response codes per method, aggregated across all "gitserver" clients.
Note: Internal errors are ones that appear to originate from the https://github.com/grpc/grpc-go library itself, rather than from any user-written application code. These errors can be caused by a variety of issues, and can originate from either the code-generated "gitserver" gRPC client or gRPC server. These errors might be solvable by adjusting the gRPC configuration, or they might indicate a bug from Sourcegraph`s use of gRPC.
When debugging, knowing that a particular error comes from the grpc-go library itself (an internal error
) as opposed to normal
application code can be helpful when trying to fix it.
Note: Internal errors are detected via a very coarse heuristic (seeing if the error starts with grpc:
, etc.). Because of this, it`s possible that some gRPC-specific issues might not be categorized as internal errors.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100812
on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_method_status{grpc_service=~"gitserver.v1.GitserverService",is_internal_error="true",grpc_method=~"${gitserver_method:regex}"}[2m])) by (grpc_method, grpc_code))
Git Server: Gitserver GRPC retry metrics
gitserver: gitserver_grpc_clients_retry_percentage_across_all_methods
Client retry percentage across all methods over 2m
The percentage of gRPC requests that were retried across all methods, aggregated across all "gitserver" clients.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100900
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverService",is_retried="true"}[2m])))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverService"}[2m])))))))
gitserver: gitserver_grpc_clients_retry_percentage_per_method
Client retry percentage per-method over 2m
The percentage of gRPC requests that were retried aggregated across all "gitserver" clients, broken out per method.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100901
on your Sourcegraph instance.
Technical details
Query:
SHELL(100.0 * ((((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverService",is_retried="true",grpc_method=~"${gitserver_method:regex}"}[2m])) by (grpc_method))) / ((sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverService",grpc_method=~"${gitserver_method:regex}"}[2m])) by (grpc_method))))))
gitserver: gitserver_grpc_clients_retry_count_per_method
Client retry count per-method over 2m
The count of gRPC requests that were retried aggregated across all "gitserver" clients, broken out per method
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=100902
on your Sourcegraph instance.
Technical details
Query:
SHELL(sum(rate(src_grpc_client_retry_attempts_total{grpc_service=~"gitserver.v1.GitserverService",grpc_method=~"${gitserver_method:regex}",is_retried="true"}[2m])) by (grpc_method))
Git Server: Site configuration client update latency
gitserver: gitserver_site_configuration_duration_since_last_successful_update_by_instance
Duration since last successful site configuration update (by instance)
The duration since the configuration client used by the "gitserver" service last successfully updated its site configuration. Long durations could indicate issues updating the site configuration.
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101000
on your Sourcegraph instance.
Technical details
Query:
SHELLsrc_conf_client_time_since_last_successful_update_seconds{job=~`.*gitserver`,instance=~`${shard:regex}`}
gitserver: gitserver_site_configuration_duration_since_last_successful_update_by_instance
Maximum duration since last successful site configuration update (all "gitserver" instances)
Refer to the alerts reference for 1 alert related to this panel.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101001
on your Sourcegraph instance.
Technical details
Query:
SHELLmax(max_over_time(src_conf_client_time_since_last_successful_update_seconds{job=~`.*gitserver`,instance=~`${shard:regex}`}[1m]))
Git Server: Codeintel: Coursier invocation stats
gitserver: codeintel_coursier_total
Aggregate invocations operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101100
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_coursier_total{op!="RunCommand",job=~"^gitserver.*"}[5m]))
gitserver: codeintel_coursier_99th_percentile_duration
Aggregate successful invocations operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101101
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_coursier_duration_seconds_bucket{op!="RunCommand",job=~"^gitserver.*"}[5m]))
gitserver: codeintel_coursier_errors_total
Aggregate invocations operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101102
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_coursier_errors_total{op!="RunCommand",job=~"^gitserver.*"}[5m]))
gitserver: codeintel_coursier_error_rate
Aggregate invocations operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101103
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_coursier_errors_total{op!="RunCommand",job=~"^gitserver.*"}[5m])) / (sum(increase(src_codeintel_coursier_total{op!="RunCommand",job=~"^gitserver.*"}[5m])) + sum(increase(src_codeintel_coursier_errors_total{op!="RunCommand",job=~"^gitserver.*"}[5m]))) * 100
gitserver: codeintel_coursier_total
Invocations operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101110
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_coursier_total{op!="RunCommand",job=~"^gitserver.*"}[5m]))
gitserver: codeintel_coursier_99th_percentile_duration
99th percentile successful invocations operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101111
on your Sourcegraph instance.
Technical details
Query:
SHELLhistogram_quantile(0.99, sum by (le,op)(rate(src_codeintel_coursier_duration_seconds_bucket{op!="RunCommand",job=~"^gitserver.*"}[5m])))
gitserver: codeintel_coursier_errors_total
Invocations operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101112
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_coursier_errors_total{op!="RunCommand",job=~"^gitserver.*"}[5m]))
gitserver: codeintel_coursier_error_rate
Invocations operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101113
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_coursier_errors_total{op!="RunCommand",job=~"^gitserver.*"}[5m])) / (sum by (op)(increase(src_codeintel_coursier_total{op!="RunCommand",job=~"^gitserver.*"}[5m])) + sum by (op)(increase(src_codeintel_coursier_errors_total{op!="RunCommand",job=~"^gitserver.*"}[5m]))) * 100
Git Server: Codeintel: npm invocation stats
gitserver: codeintel_npm_total
Aggregate invocations operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101200
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_npm_total{op!="RunCommand",job=~"^gitserver.*"}[5m]))
gitserver: codeintel_npm_99th_percentile_duration
Aggregate successful invocations operation duration distribution over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101201
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (le)(rate(src_codeintel_npm_duration_seconds_bucket{op!="RunCommand",job=~"^gitserver.*"}[5m]))
gitserver: codeintel_npm_errors_total
Aggregate invocations operation errors every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101202
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_npm_errors_total{op!="RunCommand",job=~"^gitserver.*"}[5m]))
gitserver: codeintel_npm_error_rate
Aggregate invocations operation error rate over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101203
on your Sourcegraph instance.
Technical details
Query:
SHELLsum(increase(src_codeintel_npm_errors_total{op!="RunCommand",job=~"^gitserver.*"}[5m])) / (sum(increase(src_codeintel_npm_total{op!="RunCommand",job=~"^gitserver.*"}[5m])) + sum(increase(src_codeintel_npm_errors_total{op!="RunCommand",job=~"^gitserver.*"}[5m]))) * 100
gitserver: codeintel_npm_total
Invocations operations every 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101210
on your Sourcegraph instance.
Technical details
Query:
SHELLsum by (op)(increase(src_codeintel_npm_total{op!="RunCommand",job=~"^gitserver.*"}[5m]))
gitserver: codeintel_npm_99th_percentile_duration
99th percentile successful invocations operation duration over 5m
This panel has no related alerts.
To see this panel, visit /-/debug/grafana/d/gitserver/gitserver?viewPanel=101211
on your Sourcegraph instance.
Technical details
Query:
SHELL