Developing the code insights backend
State of the backend
- Supports running search-based insights over all indexable repositories on the Sourcegraph installation.
- Is backed by a separate Postgres instance. See the database section below for more information.
- Optimizes unnecessary search queries by using an index of commits to query only for time periods that have had at least one commit.
- Supports regexp based drilldown on repository name.
- Provides permissions restrictions by filtering of repositories that are not visible to the user at query time.
- Does not yet support synchronous insight creation through an API. Read more below in the Insight Metadata section.
The current version of the backend is an MVP to achieve beta status to unblock the feature request of "running an insight over all my repos".
Architecture
The following architecture diagram shows how the backend fits into the two Sourcegraph services "frontend" (the Sourcegraph monolithic service) and "worker" (the Sourcegraph "background-worker" service), click to expand:
Deployment Status
Code Insights backend is currently disabled on sourcegraph.com
until solutions can be built to address the large indexed repo count.
Feature Flags
Code Insights ships with an "escape hatch" feature flag that will completely disable the dependency on the Code Insights DB (named codeinsights-db
). This feature flag is implemented as an environment variable that if set true DISABLE_CODE_INSIGHTS=true
will disable the dependency and will not start the Code Insights background workers or GraphQL resolvers. This variable must be set on both the worker
and frontend
services to remove the dependency. If the flag is not set on both services, the codeinsights-db
dependency will be required.
Implementation of this environment variable can be found in the frontend
and worker
services.
This flag should be used judiciously and should generally be considered a last resort for Sourcegraph installations that need to disable Code Insights or remove the database dependency.
With version 3.31 this flag has moved from the repo-updater
service to the worker
service.
Sourcegraph Setting
Code Insights is currently enabled by default on customer instances 3.32 and later, but can be disabled from appearing in the UI by setting this flag to false.
"experimentalFeatures": { "codeInsights": true },
Database
Historically, Code Insights used a TimescaleDB database running on the OSS license. The original intention was to use some of the timeseries query features, as well as the hypertable. Many of these are behind a proprietary license that would have required non-trivial work to bundle with Sourcegraph.
As of Sourcegraph 3.38, Code Insights no longer uses TimescaleDB and has moved to a standard vanilla Postgres image. The Code Insights database is still separate from the main Sourcegraph database.
Insight Metadata
Historically, insights ran entirely within the Sourcegraph extensions API on the browser. These insights are limited to small sets of manually defined repositories since they execute in real time on page load with no persistence of the timeseries data. Sourcegraph extensions have access to settings (user / org / global) , so the original storage location for extension based insight metadata (query string, labels, title, etc) was settings.
During beta, insight metadata were stored in Sourcegraph settings files, and periodically synced to the backend database.
As of 3.35, Code Insights data is stored entirely in the codeinsights-db
database, and exposed through a GraphQL API. Settings are deprecated as a storage
option, although the text in the settings will persist unless deleted. In this release Code Insights shipped an out of band migration that automatically migrates
all data from settings to the database for the last time. 3.35 also disabled the previously running sync jobs by default, which can be re-enabled using an environment variable feature flag ENABLE_CODE_INSIGHTS_SETTINGS_STORAGE
on the worker
and frontend
services. This flag is not meant to be used and is only provided as a last resort option for any users unable to use Code Insights.
Life of an insight
Unique ID
An Insight View is defined to have a globally unique referencable ID. Each ID is generated when the view is created.
A note about data series
Currently, data series are can be defined with a repository scope that will ultimately define how the series is generated. Any series that is provided a repository scope will be executed just-in-time, whereas any series missing a repository scope will be assumed to be global and will be recorded in the background. This is an area of improvement in , where we will likely integrate with other "repository group" objects such as Search Contexts.
Data series are uniquely identified by a randomly generated unique ID. Data series are also identified by a compound key that is used to preserve data series that have already been calculated. This will effectively share this data series among all users if the compound key matches.
Data series are defined with a recording interval that will define the frequency of samples that are taken for the series.
Data series are also given a field that describes how the series can be populated, called generation_method.
These generation_method
types will allow the insights backend to select different behaviors depending on the series definition, for example, to execute a compute
query instead of a standard search.
A note about capture group insight series
A standard search series will execute Sourcegraph searches, and tabulate the count based on the number of matches in the response. A highly requested feature from our customers
was to be able to derive the series themselves from the results; that is to say a result of (result: 1.17, count: 5) (result: 1.13, count: 3)
would generate two individual time series,
one for each unique result.
We support this behavior by tabulating the results of a compute
search, and dynamically modifying the series that are returned. These series can be calculated both globally, and
just-in-time.
(2) The insight enqueuer (indexed recorder) detects the new insight
The insight enqueuer is a background goroutine running in the worker
service of Sourcegraph (code), which runs all background goroutines for Sourcegraph - so long as DISABLE_CODE_INSIGHTS=true
is not set on the worker
container/process.
Its job is to periodically schedule a recording of 'current' values for Insights by enqueuing a recording using a global query. This only requires a single query per insight regardless of the number of repositories,
and will return results for all the matched repositories. Each repository will still be recorded individually. These queries are placed on the same queue as historical queries (insights_query_runner_jobs
) and can
be identified by the lack of a revision and repo filter on the query string.
For example, insights
might be an global recording, where insights repo:^codehost\.com/myorg/somerepo@ref$
would be a historical recording for a specific repo / revision.
You can find these search queries for queued jobs on the (primary postgres) table insights_query_runner_jobs.search_query
Insight recordings are scheduled using the database field (codeinsights-db) insight_series.next_recording_after
, and will only be taken if the field time is less than the execution time of the job.
Recordings are scheduled to occur one interval (per series definition) following the execution time. For example, if a recording was taken at 2021-08-27T15:29:00.000Z
with an interval definition of 1 day, the next recording will be scheduled for 2021-08-28T15:29:00.000Z
. The first recording after insight creation will occur on the same interval.
Note: There is a field (codeinsights-db) insight_series.recording_interval_days
that was intended to provide some configurable value to this recording interval. We have limited
product validation with respect to time intervals and the granularity of recordings, so beta has launched with fixed first-of-month
scheduling.
This will be an area of development throughout and into .
(3) The historical data enqueuer (historical recorder) gets to work
If we only record data starting when the series were created, it would take months or longer for users to get any value out of backend insights. This introduces the need for us to backfill data by running search queries that answer "how many results existed in the past?" so we can populate historical data.
Similar to the insight enqueuer, the historical insight enqueuer is a background goroutine (code) which locates and enqueues work to populate historical data points.
The most naive implementation of backfilling is as follows:
For each relevant repository: For each relevant time point: Execute a search at the most recent revision
Naively implemented, the historical backfiller would take a long time on any reasonably sized Sourcegraph installation. As an optimization, the backfiller will only query for data frames that have recorded changes in each repository. This is accomplished by looking at an index of commits and determining if that frame is eligible for removal. Read more below
There is a rate limit associated with analyzing historical data frames. This limit can be configured using the site setting
insights.historical.worker.rateLimit
. As a rule of thumb, this limit should be set as high as possible without performance
impact to gitserver
. A likely safe starting point on most Sourcegraph installations is insights.historical.worker.rateLimit=20
.
Backfill compression
Read more about the backfilling compression in the proposal RFC 392
We maintain an index of commits (table commit_index
in codeinsights-db) that are used to filter out repositories that do not need a search query. This index is
periodically refreshed
with changes since its previous refresh. Metadata for each repositories refresh is tracked in a table commit_index_metadata
.
To avoid race conditions with the index, data frames are only filtered
out if the commit_index_metadata.last_updated_at
is greater than the data point we are attempting to compress. Additionally, if the index does not contain enough history (the commit
falls before the oldest commit in the index), all of the data frames prior to the history will be uncompressed.
Currently, we only generate 12 months of history for this commit index to keep it reasonably sized. We do not currently do any pruning, but is an area that will need development in the future.
Limiting to a scope of repositories
Naturally, some insights will not need or want to execute over all repositories and would prefer to execute over a subset to generate faster. As a trade off to reach beta we made the decision that all insights will execute over all repositories. The primary justification was that the most significant blocker for beta was the ability to run over all insights, and therefore unlocking that capability also unlocks the capability for users that want to run over a subset, they will just need to wait longer.
This is non-trivial problem to solve, and raises many questions:
- How do we represent these sets? Do we list each repository out for each insight? This could result in a very large cardinality and grow the database substantially.
- What happens if users change the set of repositories after we have already backfilled?
- What does the architecture of this look like internally? How do we balance the priority of backfilling other larger insights with much smaller ones?
Additionally, Search Contexts now support query prefixes (for example a repo
prefix repo:github\./com/sourcegraph/.*
) which are a popular and highly requested feature we would like to integrate
into the Code Insights repository scope.
This will be an area of improvement in .
Detecting if an insight is complete
Given the large possible cardinality of required queries to backfill an insight, it is clear this process can take some time. Through dogfooding we have found on a Sourcegraph installation with ~36,000 repositories, we can expect to backfill an average insight in 20-30 minutes. The actual benchmarks of how long this will take vary greatly depending on the commit patterns and size of the Installation.
One important piece of information that needs to be surfaced to users is the answer to the question is my insight still processing?
This is a non-trivial question to answer:
- Work is processed asynchronously, so querying the state of the queue is necessary
- Iterating many thousands of repositories can result in some transient errors causing individual repositories to fail, and ultimately not be included in the queue issue
As a temporary measure to try and answer this question with some degree of accuracy, the historical backfiller applies the following semantic:
Flag an insight as completed backfill
if the insight was able to complete one full iteration of all repositories without any hard
errors (such as low level DB errors, etc).
This flag is represented as the database field insight_series.backfill_queued_at
and is set at the end of the complete repository iteration.
This semantic does not fully capture all possible states. For example, if a repository encounters a soft
error (unable to fetch git metadata, for example)
it will be skipped and ultimately not populate in the data series. Improving this is an area of design and work in .
(4) The queryrunner worker gets work and runs the search query
The queryrunner (code) is a background goroutine running in the worker
service of Sourcegraph (code), it is responsible for:
- Dequeueing search queries that have been queued by the either the current, snapshot, or historical recorder. Queries are stored with a
priority
field that dequeues queries in ascending priority order (0 is higher priority than 100). - Executing a search against Sourcegraph with the provided query. These queries are executed against the
internal
API, meaning they are unauthorized and can see all results. This allows us to build global results and filter based on user permissions at query time. - Flagging any error states (such as limitHit, meaning there was some reason the search did not return all possible results) as a
dirty query
. These queries are stored in a tableinsight_dirty_queries
that allow us to surface some information to the end user about the data series. Not all error states are currently collected here, and this will be an area of work for . - Aggregating the search results, per repository (and in the near-future, per unique match to support capture groups) and storing them in the
series_points
table.
The queue is managed by a common executor called Worker
(note: the naming collision with the worker
service is confusing, but they are not the same).
Read more about Worker
and how it works in this search notebook.
These queries can be executed concurrently by using the site setting insights.query.worker.concurrency
and providing
the desired concurrency factor. With insights.query.worker.concurrency=1
queries will be executed in serial.
There is a rate limit associated with the query worker. This limit is shared across all concurrent handlers and can be configured
using the site setting insights.query.worker.rateLimit
. This value to set will depend on the size and scale of the Sourcegraph
installations Searcher
service.
(5) Query-time and rendering!
The webapp frontend invokes a GraphQL API which is served by the Sourcegraph frontend
monolith backend service in order to query information about backend insights. (code)
- A GraphQL resolver
insightViewResolver
returns all the distinct data series in a single insight (UI panel) (code) - A resolver is selected depending on the type of series, and whether or not dynamic search results need to be expanded.
- A GraphQL resolver ultimately provides data points for a single series of data (code)
- The series points resolver merely queries the insights store for the data points it needs, and the store itself merely runs SQL queries against the database to get the datapoints (code)
Note: There are other better developer docs which explain the general reasoning for why we have a "store" abstraction. Insights usage of it is pretty minimal, we mostly follow it to separate SQL operations from GraphQL resolver code and to remain consistent with the rest of Sourcegraph's architecture.
Once the web client gets data points back, it renders them! For more information, please contact an @codeinsights frontend engineer.
User Permissions
We made the decision to generate data series for all repositories and restrict the information returned to the user at query time. There were a few driving factors behind this decision:
- We have split feedback between customers that want to share insights globally without regard for permissions, and other customers that want strict permissions mapped to repository visibility. In order to possibly support both (or either), we gain the most flexibility by performing query time limitations.
- We can reuse pre-calculated data series across multiple users if they provide the same query to generate an insight. This not only reduces the storage overhead, but makes the user experience substantially better if the data series is already calculated.
Given the large possible cardinality of the visible repository set, it is not practical to select all repos a user has access to at query time. Additionally, this data does not live in the same database as the timeseries data, requiring some network traversal.
User permissions are currently implemented by negating the set of repos a user does not have access to. This is based on the assumption that most users of Sourcegraph have access to most repositories. This is a fairly highly validated assumption, and matches the premise of Sourcegraph to begin with (that you can search across all repos). This may not be suitable for Sourcegraph installations with highly controlled repository permissions, and may need revisiting.
Storage Format
The code insights time series are currently stored entirely within Postgres.
As a design, insight data is stored as a full vector of match results per unique time point. This means that for some time T
, all of the unique timeseries that fall under
one insight series can be aggregated to form the total result. Given that the processing system will execute every query at-least once, the possiblity of duplicates
exist within a unique timeseries. A simple deduplication is performed at query time.
Read more about the history of this format.
Running Locally
Using sg
, run the enterprise-codeinsights
to run everything needed for code insights.
sg start enterprise-codeinsights
Insights can then be created either via the locally running webapp, or created via the GraphQL API.
If you've created an insight that needs to generate series data on the backend, be aware of the time interval at which these series will be picked up for backfilling. You may want to restart the service so that the new series will be picked up right away for processing.
Unit Tests
The codeinsights-db
must be running in order for tests against any of the insight related stores
to work correctly, as these interact with the database. You can run the following command to start the codeinsights-db
:
sg run codeinsights-db
Debugging
This being a pretty complex, high cardinality, and slow-moving system - debugging can be tricky.
In this section, I'll cover useful tips I have for debugging the system when developing it or otherwise using it.
Accessing the Code Insights database
Dev and docker compose deployments
docker exec -it codeinsights-db psql -U postgres
Kubernetes deployments
kubectl exec -it deployment/codeinsights-db -- psql -U postgres
- If trying to access Sourcegraph.com's DB:
kubectl -n prod exec -it deployment/codeinsights-db -- psql -U postgres
- If trying to access k8s.sgdev.org's DB:
kubectl -n dogfood-k8s exec -it deployment/codeinsights-db -- psql -U postgres
Finding logs
Since insights runs inside of the frontend
and worker
containers/processes, it can be difficult to locate the relevant logs. Best way to do it is to grep for insights
.
The frontend
will contain logs about e.g. the GraphQL resolvers and Postgres migrations being ran, while worker
will have the vast majority of logs coming from the insights background workers.
Docker compose deployments
docker logs sourcegraph-frontend-0 | grep insights
and
docker logs worker | grep insights
Inspecting the Code Insights database
Read the initial schema migration which contains all of the tables we create in Postgres and describes them in detail. This will explain the general layout of the database schema, etc.
The most important table in the insights database is series_points
, that's where the actual data is stored.
Querying data
SELECT * FROM series_points ORDER BY time DESC LIMIT 100;
Query data, filtering by repo and returning metadata
SELECT * FROM series_points JOIN metadata ON metadata.id = metadata_id WHERE repo_name_id IN ( SELECT id FROM repo_names WHERE name ~ '.*-renamed' ) ORDER BY time DESC LIMIT 100;
(note: we don't actually use metadata currently, so it's always empty.)
{"hello": "world"}
Query data, filter by metadata containing SELECT * FROM series_points JOIN metadata ON metadata.id = metadata_id WHERE metadata @> '{"hello": "world"}' ORDER BY time DESC LIMIT 100;
(note: we don't actually use metadata currently, so it's always empty.)
Query data, filter by metadata containing Go languages
SELECT * FROM series_points JOIN metadata ON metadata.id = metadata_id WHERE metadata @> '{"languages": ["Go"]}' ORDER BY time DESC LIMIT 100;
(note: we don't actually use metadata currently, so it's always empty. The above gives you some ideas for how we intended to use it.)
See https://www.postgresql.org/docs/9.6/functions-json.html for more metadata jsonb
operator possibilities. Only ?
, ?&
, ?|
, and @>
operators are indexed (gin index)
Query data the way we do for the frontend, but for every series
SELECT sub.series_id, sub.interval_time, SUM(sub.value) AS value, sub.metadata FROM ( SELECT sp.repo_name_id, sp.series_id, sp.time AS interval_time, MAX(value) AS value, NULL AS metadata FROM series_points sp JOIN repo_names rn ON sp.repo_name_id = rn.id GROUP BY sp.series_id, interval_time, sp.repo_name_id ORDER BY sp.series_id, interval_time, sp.repo_name_id DESC ) sub GROUP BY sub.series_id, sub.interval_time, sub.metadata ORDER BY sub.series_id, sub.interval_time DESC
Inserting data
Upserting repository names
The repo_names
table contains a mapping of repository names to small numeric identifiers. You can upsert one into the database using e.g.:
WITH e AS( INSERT INTO repo_names(name) VALUES ('github.com/gorilla/mux-original') ON CONFLICT DO NOTHING RETURNING id ) SELECT * FROM e UNION SELECT id FROM repo_names WHERE name='github.com/gorilla/mux-original';
Upserting event metadata
Similar to repo_names
, there is a separate metadata
table which stores unique metadata jsonb payloads and maps them to small numeric identifiers. You can upsert metadata using e.g.:
WITH e AS( INSERT INTO metadata(metadata) VALUES ('{"hello": "world", "languages": ["Go", "Python", "Java"]}') ON CONFLICT DO NOTHING RETURNING id ) SELECT * FROM e UNION SELECT id FROM metadata WHERE metadata='{"hello": "world", "languages": ["Go", "Python", "Java"]}';
Inserting a data point
You can insert a data point using e.g.:
INSERT INTO series_points( series_id, time, value, metadata_id, repo_id, repo_name_id, original_repo_name_id ) VALUES( "my unique test series ID", now(), 0.5, (SELECT id FROM metadata WHERE metadata = '{"hello": "world", "languages": ["Go", "Python", "Java"]}'), 2, (SELECT id FROM repo_names WHERE name = 'github.com/gorilla/mux-renamed'), (SELECT id FROM repo_names WHERE name = 'github.com/gorilla/mux-original') );
You can omit all of the *repo*
fields (nullable) if you want to store a data point describing a global (associated with no repository) series of data.
Creating DB migrations
migrations/codeinsights
in the root of this repository contains the migrations for the Code Insights database, they are executed when the frontend starts up (as is the same with e.g. codeintel DB migrations.)
Currently, the migration process blocks frontend
and worker
startup - which is one issue we will need to solve.