Our changelog, announcements, dev posts, and anything else we think you'll find interesting.
The difference between “I can figure this out in five minutes with a Sourcegraph search” versus “I can ask someone on another team in another timezone, and if they don’t know, follow up with a third person in a third timezone” is often the difference between taking care of something versus letting that task die on a todo list backlog.
Sourcegraph is a powerful developer tool used by engineers at top companies, from Uber and Lyft to Amazon and SoFi. But as we've dogfooded our own product, we’ve also discovered it’s a powerful timesaver and cheat sheet for teams across an organization, such as Product Management, Business Operations, Analytics, and Marketing.
Sourcegraph saves our Product Managers and Analytics/Business Operations teammates 20-60 minutes a few times a week, and provides definitive answers to a surprising range of questions.
In this post, we'll be using a feature called Notebooks, which debuted in our 3.36 release. Notebooks enable Sourcegraph users to document and share Sourcegraph searches and in this post, we'll use them to share the precise searches our team members used. Hit
Run all blocks to perform the search and see the results.
To go to the sections most relevant to you, use the table of contents below.
- Product Managers
- What actions do our metrics track?
- I’m mid-conversation with a customer. Is another team’s feature affecting our feature?
- What's the Sourcegraph standard for error message syntax?
- Situation: Determine the different [blank]@sourcegraph emails we direct users to across our entire user-facing codebase
- Should we use “prototype” or “experimental” for early-stage product features?
- Analytics and Business Operations
- How has the Sourcegraph product data architecture changed over time?
- How can I monitor our self-hosted telemetry infrastructure to get ahead of potential bugs?
- Which Google Analytics tags are used on our various web properties?
- Where do fields in HubSpot forms come from and how should we implement new ones?
This section is written by Joel Kwartler, the Product Manager for Code Insights. He’s a brilliant, genuinely kind human being who’s being played by Brad Pitt in an upcoming HBO biopic.
As a product manager at a fast-growing startup, I frequently need to know how our features interact, which analytics metrics other teammates have already built, and which product team processes or standards exist versus those that haven’t been built out yet. I use Sourcegraph to help me quickly answer these types of questions.
What actions do our metrics track?
I’d just joined Sourcegraph in September 2020 as the first PM for the extensibility team. I went through our Business Intelligence platform but didn’t know what “ViewRegistryExtensionManage” meant or how it was different than “ViewRegistryExtensionContributions.” I turned to Sourcegraph.
Search: First, I ran this search, got no results, ran this search, got results, and then ran this search, which gave me a quick reference for the pages all the events fired on. From there, I could glance through the strings and comments on the page to map them back to states in our app.
Impact: I didn’t need to ask a teammate or pull the repo(s) and search it locally. The latter means I also didn’t have to clutter my desktop with yet another open application window.
Time Saved: 10 minutes (20 if we had multiple repos) and I didn’t need another open desktop application.
Follow-up situation: We can quickly see the names of all of our view analytics events and be able to go to what page they’re triggered on with this search.
How big a problem are unversioned docs links, and how do I stop people from using them?
We built a system to version our docs links, but it turns out many people weren't aware of it and were using literal links to docs. (This is bad because it might take someone from an older version to the latest content, which may not be relevant to their version.)
Search: First I ran this search and found dozens of unversioned docs links.
- Created a code monitor to send off a slack message to our frontend channel whenever someone added another unversioned (bad) link, which looked like this:
- Migrated a bunch of links to the correct format, tracked them with
to="\/help file:client/web/src/enterprise, and created a code insight to monitor that migration and ensure it never became a problem again, which looked like this:
Impact: Users don't get lost or confused by landing on the wrong version of a docs page.
Time Saved: 10 minutes I would have spent manually pulling and checking the codebase the first time, plus the many days saved over preventing regression with more difficult or manual automation compared to code monitors and insights.
I’m mid-conversation with a customer. Is another team’s feature affecting our feature?
A customer was seeing incomplete result counts for the search results returned for a Code Insights feature, making it impossible to reason about their code based on these incomplete results.
Search: A term like “limit 5000” is likely to bring up many hits throughout our codebase, so I used a diff search with a timeframe to scope my search to the time the behavior started appearing.
I quickly found our search team had changed a limit that affected one of my team’s features. I was able to understand who made the change and why they made it in a few seconds.
Impact: We could immediately respond to the customer and confirm the behavior as well as provide more context for the estimated time to fix.
Time Saved: 18 hours waiting for a teammate to answer my question—hours that were especially important given we were actively working with a customer.
What's the Sourcegraph standard for error message syntax?
I needed to finish writing the error messages that display to users for a major new feature. We have error message guidelines in our handbook, but it’s easier to quickly learn something with an example.
Search: This search quickly gave me numerous in-product examples.
Impact: I was able to produce more unified product messaging.
Time Saved: I saved 10 minutes I would have spent clicking through the product and avoided causing errors I would have otherwise needed to cause so I could find live examples.
Situation: Determine the different [blank]@sourcegraph emails we direct users to across our entire user-facing codebase
I needed to add a contact email to some new documentation, but wasn’t sure what email addresses we’d already set up and which would be the best to reuse.
Search: This search showed me what email addresses were already in use.
In fact, I can also make a code insight that will automatically pull all email variations into a usage graph and show me trends over time for the pattern
mailto:(\email@example.com) by email handle:
Impact: We learned we have a “support@sourcegraph” email that was better than the “feedback@” email I was about to use.
Time Saved: We saved hours to days of waiting for an answer. The alternative solution would have involved searching every single repository or asking every long-tenured teammate to list what emails they know are used.
Follow-up situation: What anchor text do we generally use for the email link, or do we display the direct email? Search:
Should we use “prototype” or “experimental” for early-stage product features?
Feature classifications are important because they help us set user expectations, classify support SLAs and ownership, and communicate internally about the maturity of a feature. In choosing what to call our pre-beta feature level, we wanted to quickly see what teammates had already gravitated towards.
Search one returns 2.0k results.
Search two returns 452 results.
Impact: In 60 seconds, we determined we used “experimental” fives times as often as “prototype.” Given that already existing, informal tendency, we codified “experimental” in our handbook as the official Sourcegraph label.
Analytics and Business Operations
This section was written by Eric-Brody Moore, a member of the Business Operations team. HBO only had room for one Sourcegrapher biopic and they unfortunately chose Joel despite Eric also being a brilliant and kind human being.
How has the Sourcegraph product data architecture changed over time?
The architecture of our internally facing data is ever changing, so it’s critical we’re able to keep up with it so our data models are up-to-date and accurate.
This search shows how our PostgreSQL database structure changes over time.
Impact: These searches put the information that analysts need in their hands so that they can understand our backend data structures when they create reports. We can click through the PRs that make these changes to understand the full context behind why a change was made and what it means.
Time saved: In most cases, these searches remove the need to talk to the teams who maintain our database schemas, a process that can take days, especially in an asynchronous environment.
How can I monitor our self-hosted telemetry infrastructure to get ahead of potential bugs?
Our data warehouse schemas need to mirror schemas in our telemetry code. If it doesn’t, it isn’t able to receive data coming from self-hosted instances. This gap can lead to data losses, which have ramifications on our ability to inform decisions with data.
Search: This search checks for schema mirroring.
Impact: A reality for most data teams is that telemetry sometimes breaks. At Sourcegraph, it used to take days to realize something was broken. It then took hours to search through logs to determine which change was made to our telemetry and how to fix it. Now, Sourcegraph code monitoring immediately alerts us when a change goes in and what it was, so we can minimize data loss.
Time saved: Code monitoring saves us the 30 minutes required to get up to speed on what is leading to broken data pipelines, and saves us even more time and resources we’d otherwise spend minimizing data loss (and explaining to the company why data was lost and what the consequences of that loss will be).
Which Google Analytics tags are used on our various web properties?
“Which Google Analytics account should I even be using?” was a frequent question over the first few months we had a marketing team.
Search: This search finds which Google Analytics tags were in use on our site.
Impact: Our Google Analytics had inconsistent data from historical changes in our web properties and nobody quite knew which web property to use within Google Analytics because they all had different numbers. I did a diff search for which tags were actually on our site and the history of changes and the search above came back with results. I was able to see which tag was in production, when it was added, and who to talk to about the history of our Google Tag Manager (spoiler: it was our CEO Quinn Slack – I decided not to bother him with it and just move forward!)
Time saved: Exploring Google Analytics with Sourcegraph saved us a couple hours we’d have otherwise spent tracking down who might have historical knowledge of our Google Tag Manager configuration. It likely saved us even more time too, since we needed to investigate past changes in Google Analytics.
Where do fields in HubSpot forms come from and how should we implement new ones?
We use HubSpot forms for all of our opt-in form submissions. These forms are coded into our web properties by the software engineering teams, so it’s essential that the communication about what is being collected is seamless between the marketing, data, and engineering teams.
We used this search to start the process, then:
This file to figure out which fields we’re sending back and the data type.
We could then send this to a dev to understand how we implement forms and who to talk to if needed.
We then gave this link example to developers for understanding how to implement a new in-product survey.
We were wondering if
in HubSpot was populating from our code or HubSpot workflow, so I searched this
Impact: I was able to give our software engineers a clear requirements document on which fields they should pull from when implementing a new HubSpot field. I was also able to include best practices for adding HubSpot forms in our code based on previous PRs we had committed.
Time saved: I saved a few async hours going back and forth with a software engineer to answer a question from our marketing team, and saved any time I would’ve gone back and forth with our software engineering team on building new forms into product.
If you're interested in trying Sourcegraph search for yourself, you can sign up for an account and get started searching your codebase and the OSS universe in seconds. If you're interested in Notebooks, you can create a Sourcegraph account and start documenting your searches today.