Optimizing a code intelligence commit graph (Part 1)
Sourcegraph's Code Intelligence team builds tools and services that provide contextual information around code. These enable users to perform fast, comprehensive, and accurate code navigation, and to surface dependency relationships across projects, repositories, and languages. In this post I'll dive into how Sourcegraph can resolve code intelligence queries using data from older commits when data on the requested commit is not yet available. In Part 2, we'll cover how additional scalability concerns presented themselves and how we tackled them.
Since the first version of Sourcegraph, precise code navigation has been a first-order concern. Its ability to provide compiler-accurate code navigation in a web-based interface is a superpower for our users.
The journey to our current implementation began in February 2019 when we shifted our efforts from running Language Servers alongside Sourcegraph to pre-indexing source code via the Language Server Index Format (LSIF) and uploading it to Sourcegraph. This change introduced a new requirement of the user: they are now responsible for producing and uploading the LSIF index.
The method to produce an LSIF index is highly variable and depends on many factors. If a repository is not too large, a user may wish to create a new index on every commit in their CI. If a repository is very large (the monorepo case), or an organization has a large number of repositories, it may be better to instead index code periodically and upload a refreshed index. Google uses such a scheduled job to refresh their monorepo's index twice a day.
Regardless of the indexing method, it is always possible for a user to find themselves browsing a commit for which Sourcegraph has not received an index. This can happen due to a slow job in a continuous integration server, a backed-up indexing processing queue, or the commit may simply be skipped in the case of a periodic refresh.
This is a glaring hole in the feature. The majority of users are exploring code on the tip of a branch, which is the least likely commit to have an index immediately after it's been pushed to the code host.
In order to plug this hole, we determine the set of nearby commits for which Sourcegraph has received an index, query these indexes on behalf of the requested commit, then adjust the resulting locations (file paths and ranges within a document) using the Git diff between the commits as a guide. This enables Sourcegraph to respond with precise results to requests on commits missing an index.
Tracking cross-commit index visibility
The first step in this process is to track how commits of a repository relate to one another. Unfortunately, the service providing the code intelligence features was separated (by design) from the rest of the product. We had only recently gained access to the Sourcegraph PostgreSQL database used by the rest of the application, and no other team was tracking commit information. The source of truth for that data was another service called gitserver, which required both an RPC call and a subprocess to access.
Our initial stab at this problem was to introduce 2 new tables to PostgreSQL: commits
and lsif_data_markers
.
The commits
table stores data similar to a flattened version of the output from git log --all --pretty='%H %P'
(a commit followed by a list of its parents), for each repository. Example values for this table are shown below to aid our running example. This table would generally store the full 40-character commit ID—we are abbreviating them for brevity here.
repository | commit | parent_commit |
---|---|---|
github.com/sourcegraph/sample | a36064 | f4fb06 |
github.com/sourcegraph/sample | f4fb06 | 4c8d9d |
github.com/sourcegraph/sample | 313082 | 4c8d9d |
github.com/sourcegraph/sample | 6a06fc | 313082 |
github.com/sourcegraph/sample | 4c8d9d | d67b8d |
github.com/sourcegraph/sample | d67b8d | 323e23 |
github.com/sourcegraph/sample | 323e23 |
This table is synchronized with the source of truth in gitserver whenever we receive a request for a commit that we didn't know about previously. A row present in the lsif_data_markers
table denotes that an index was uploaded for a particular commit. Example values for this table are also shown below.
repository | commit |
---|---|
github.com/sourcegraph/sample | f4fb06 |
github.com/sourcegraph/sample | d67b8d |
github.com/sourcegraph/sample | 0eed16 |
Together, these 2 tables enable us to recursively search up and down the commit graph starting at a particular requested commit, and stop the search once we reach a commit that has index data. The following recursive common table expression does just that.
This query maintains a worklist (the query-local lineage
table) seeded by the requested commit, and grows by finding untracked commits that are either a parent (the ancestor A
direction) or a child (the descendant D
direction) of a commit already in the worklist and inserting it. The table is implicitly ordered by its insertions, so the final select returns the commit with index data that also has the smallest commit distance to the requested commit. The WHERE
clause inside of the CTE is there to limit the size of the working table, which can be pathologically large in the case of a large commit graph and no uploads visible from a given commit.
Example
The values of the tables above represent the following hypothetical commit graph, where a36064
is the head of the main
branch, 6106fc
is the head of the feature branch feat/x
, and the commits with uploads (f4fb06
and d67b8d
) are drawn in blue.
Running the query above from the commit 313082
produces the following CTE results over 3 iterations before halting, and ultimately returns d67b8d
as the nearest commit visible to the target query.
iteration | commit | parent_commit | distance | direction |
---|---|---|---|---|
1 | 313082 | 4c8d9d | 0 | A |
1 | 313082 | 4c8d9d | 0 | D |
2 | 4c8d9d | d67b8d | 1 | A |
2 | 6a06fc | 313082 | 1 | D |
3 | d67b8d | 323e23 | 2 | A |
Of particular note is that we don't visit sibling commits: once we reach a commit by travelling in the ancestor direction, we can't suddenly flip direction. Maintaining the direction heading during traversal ensures that the algorithm will eventually terminate.
Performance improvements
Unfortunately, our first stab at an implementation had a number of rather disappointing performance characteristics, as first stabs typically do. The query above is basically a SQL translation of an imperative graph-walking algorithm. Thinking of the query in these terms makes it easier to see that the runtime cost of the query will increase proportionally to the distance travelled through the graph.
The following query plan shows an execution trace that visited around 10 commits (the resulting commit was 5 commits away, and we search in both directions). This query takes about 40ms, which is already a high cost for a query that runs every time a user hovers over an identifier.
The following query plan shows an execution trace that visited around 100 commits (the resulting commit was 44 commits away). This query takes about 330ms, which is well over the noticeable latency threshold.
Adding additional indexes to the commits
table helped a bit, but did not fundamentally change the performance characteristics of the query. An even larger pathology was discovered in repositories with a large number of merge commits. In order to understand the performance issue, it's important first to understand how the recursive query evaluation works in the case of duplicates, which was initially unintuitive to us. Paraphrasing the PostgreSQL documentation, recursive queries are evaluated with the following steps (emphasis ours):
- Evaluate the non-recursive term and discard duplicate rows
- Insert rows into result set as well as a temporary working table
- While the working table is not empty:
- Evaluate the recursive term by substituting the recursive self-reference with the working table and discard duplicate rows
- Insert rows into result set as well as a new temporary intermediate table
- Replace the working table with the intermediate table and empty the intermediate table
A row is a duplicate of another row (from PostgreSQL's point of view) if they both contain the same set of values. However, from our point of view, a row is a duplicate of another row if only their commit values match. After all, we're running a breadth-first search over a graph and by the time we've seen a commit for the second time, we've already seen it via the shortest path. This mismatch in expectations don't cost us correctness, but it does cause performance problems and the pain that comes with it.
Example
The following hypothetical commit graph contains a number of feature branches that are eventually merged back into mainline, unlike our previous example where all commits had at most one parent.
Running the query above from the commit 703e33
produces the following CTE results over the first 4 iterations.
iteration | commit | parent_commit | distance | direction |
---|---|---|---|---|
1 | 703e33 | 6307e6 | 0 | A |
1 | 703e33 | 6307e6 | 0 | D |
1 | 703e33 | 5a24e4 | 0 | A |
1 | 703e33 | 5a24e4 | 0 | D |
2 | 5a24e4 | 3d2f27 | 1 | A |
2 | 6307e6 | 4a848f | 1 | A |
3 | 3d2f27 | 09210f | 2 | A |
3 | 3d2f27 | 2190d3 | 2 | A |
3 | 4a848f | 3d2f27 | 2 | A |
4 | 09210f | 3 | A | |
4 | 2190d3 | 1f64f9 | 3 | A |
4 | 3d2f27 | 09210f | 3 | A |
4 | 3d2f27 | 2190d3 | 3 | A |
Notice that there are 2 ways to get from commit 7033ee
to commit 3d2f27
, therefore the entries for 3d2f27
are duplicated in the CTE results (and therefore the working table). Also notice that the number of new rows per iteration is growing as the iteration count rises in such graphs. For some configuration of input, this query is quadratic instead of linear.
What we wanted to happen was for the "duplicate rows" not to be inserted into the working table at all. Unfortunately, each record in this example is distinct due to the differing path lengths—a detail we glossed over when designing this query in the first place. Another classic case of the computer doing what you told it to do instead of what you wanted it to do. 🙄
Our first attempt to optimize this query directly tackled the problem of duplicate rows in the worktable, as shown in the above example. This change simply removes the distance
column from the lineage table expression. The "duplicates" that we now throw out are records for commits that have already been seen via a shorter path. This required that we move the limiting condition from table expression into the select, which changes the behavior very slightly (it now limits by working set size, not by distance, which was an acceptable trade-off for the performance increase).
Additional efforts to optimize this query were highly successful. The following chart compares the query latency of the original query (quadratic, blue) and the optimized query (fast linear, green), and we've very clearly removed the term that was creating the quadratic behavior.
Looking back at the query plan above, the we can now determine that the culprit drastically affecting performance is the index scan within the nested loop. Ignore the sequential scan block here, which is usually suspicious but happens to be a red herring in this case. The sequential scan happens in favor of an index because of the small size of the lsif_data_markers
dataset. When the table becomes larger, it is replaced with an efficient index scan.
But index scans are supposed to be fast! Well, they are faster than a sequential scan, but may not be as fast as an index scan that uses a different index, or uses an index in a slightly different way. This particular scan fetches rows from the commits table using only c.repository = l.repository
as the index condition. This pulls back literally the entire commit graph for the repository and each row is then filtered based on the remaining conditions. There are two multicolumn indexes that could conceivably be used here: one on (repository
, commit
) and one on (repository
, parent_commit
). However, the following filter conditions would require the use of both indexes, which PostgreSQL seems unable to do, or unwilling to do due to an inaccurate query cost estimate.
To reduce the cost of this index scan, we rewrote the recursive term within the table expression above (the second term of the UNION
). Instead of using a SELECT
with an OR
condition (which had trouble using our existing set of indexes), we use UNION
to combine the results matching each side of the OR
clause.
The query plan for this new query, shown below, seems more complex at first glance. However, this query is drastically more efficient. The same input that required 330ms to evaluate now takes under 1ms to evaluate.
Even though we're now executing a greater number of steps, each step can be evaluated efficiently. The old, inefficient index scan has been broken into 2 different index scans: one that traverses the ancestor direction, and another that traverses the descendant direction. The simple conditionals in each query can be evaluated without an index filter by each of the multicolumn indexes above. Instead of pulling back the entire commit graph on each iteration of the table expression, we pull exactly the set of rows that need to be added to the working table.
This change makes the query efficient enough that we no longer have to worry about caching results, as running it on every request only contributes a negligible amount of time. Thus solving the problem once and for all!
Well, it solved the problem once and for all at a particular scale.
Lessons learned
One major take away for us was the concrete reinforcement that databases are a hugely deep subject, but are not magic. Building and maintaining an accurate mental model of how your tools work seems to be a necessary skill in software engineering, and acting with poor mental models can lead to disastrous performance results (at best).
Another major takeaway is that the cliché holds: computers do what you tell them to and nothing more. When debugging correctness and performance issues, it's good to shed your previous mental model and re-assess its correctness as you dive deeper into your stack.
Check out Part 2, in which we tackle additional scalability challenges.