Analytics Instrumentation Group
Vision
The Analytics Instrumentation Group is part of the Analytics section. Our group focuses on providing GitLab’s team with data-driven product insights to build a better GitLab. To do this, we build data collection and analytics tools within the GitLab product in a privacy-focused manner. Insights generated from Analytics Instrumentation enable us to identify the best places to invest people and resources, what product categories mature faster, where our user experience can be improved, and how product changes impact the business. You can learn more about what we’re building next on the Analytics Instrumentation Direction page.
How we work
- We work in accordance with our GitLab values.
- We work transparently with nearly everything public.
- We get a chance to work on the things we want to work on.
- We have a bias for action.
- We make data-driven decisions.
- Everyone can contribute to our work.
How to reach us
If you have any questions start by @ mentioning the product manager for the Analytics Instrumentation Group or by creating an issue in our issue board.
Responsibilities
Internal Events
We provide the tooling to allow for instrumentation of product usage events across GitLab’s services through a system we call internal events which uses the Snowplow event collection infrastructure under the hood. A list of all instrumented events can be found in our metrics dictionary. It is the responsibility of each engineering group to create and maintain their own events as those are domain specific, but Analytics Instrumentation is always willing to help on those issues be it via pair programming or our office hours.
We are actively monitoring the overall health of internal events via Monte Carlo alerts and Tableau dashboards.
Service Ping Metrics
We’re responsible to deliver a reliable Service Ping that runs every week on SaaS and Self Managed instances. Our responsiblity is tooling and automations for metric collections to set the company up for success to deliver Service Ping data to our data warehouse. It is the responsibility of each engineering group to create and maintain their own metrics as those are domain specific, but Analytics Instrumentation is always willing to help on those issues be it via pair programming or our office hours. A list of all metrics can be found on metrics.gitlab.com.
For questions related to a specific metric, its definition and/or implementation, please contact the Product Manager of the group which owns the metric. You can find information about the metric including its data category, whether it is considered an xMAU metric, its group designation and more in the metric dictionary. We are actively monitoring the overall health of Service Ping via Monte Carlo alerts and Tableau dashboards.
Projects
These are important projects that we maintain besides our involvement in the main GitLab codebase.
- Version App
- Metrics Dictionary and Metrics Dictionary from Growth group.
- Snowplow Pseudonymization
- Iglu Schema repository for GitLab
Team members
The following people are permanent members of the Analytics Instrumentation Group:
Name | Role |
---|---|
Ankit Panchal | Senior Frontend Engineer, Analytics:Analytics Instrumentation |
Greg Myers | Security Engineer, Application Security, Package (Package Registry, Container Registry), US Public Sector Services, Gitaly Cluster, Analytics (Analytics Instrumentation, Product Analytics), AI Working Group |
Jonas Larsen | Senior Backend Engineer, Analytics:Analytics Instrumentation |
Michał Wielich | Backend Engineer, Analytics:Analytics Instrumentation |
Niko Belokolodov | Senior Backend Engineer, Analytics:Analytics Instrumentation |
Nwanna Isong | Backend Engineer, Analytics:Analytics Instrumentation |
Sarah Yasonik | Senior Backend Engineer, Analytics:Analytics Instrumentation |
Tanuja Jayarama Raju | Product Manager, Analytics:Analytics Instrumentation |
Project management process
Our team uses a hybrid of Scrum for our project management process. This process follows GitLab’s monthly milestone release cycle.
- We only work off of issue boards which act as our single source of truth.
- We continuously progress issues to the next workflow stage.
- We work on both product and engineering initiatives.
- We prioritize and estimate all issues we work on.
- We do monthly milestone planning to prepare for our upcoming milestone.
- We do weekly check-ins to share updates within the team
Check-In
We do a weekly automated check-in within a separate Slack channel. We feel that a weekly cadence is enough to keep everyone up to date about the most important developments within the team. A bot asks every team member autoamtically the following questions at the beginning of the week and posts them to the channel
- What Victories did you have in the last week?: This should be the most important achievements from the last week. They can also be personal achievements.
- What Impediments did you have in the last week? Which impediments do you anticipate this week?: This is important information for other stakeholders, to understand where a team member might need support.
- Which Priorities do you have for the current week?: This should be the most important things you want to achieve during the week.
Retrospective
We believe that continuously evolving and improving the way we work is crucial to delivering good outcomes. Traditionally, retrospectives are used to facilitate this improvement. Due to our geographical distribution we cannot run typical sync retrospectives that include everyone. We’ve also found that completely async retrospectives only see limited participation and discussion.
We’re currently trialing a mixed retrospective format, we’re team members pair on filling in the retro issue. The process is as follows:
- Every milestone the async-retrospective project automatically creates an issue in the analytics-instrumentation retro project which contains a list of retro questions.
- Every 4th week a donut Slackbot will pair up all team members in random groups of two.
- Every group of two should pair on filling out the retro issue for the current milestone together within that week:
- This should happen ideally through a sync meeting or alternatively via Slack.
- Focus on the what happened the last two weeks, since the last pairing.
- Pairing should involve taking the time to think through the questions, talking about the answers with your pair and coming up with action items based on the identified problems.
- Expect an overall time commitment of 30 minutes to an hour for the bi-weekly pairing.
- Every 2nd week we have a 30-minute time blocker event, that is meant for everyone to take the time to answer Retro discussions.
Workflow
Our team use the following workflow stages defined in the Product Development Flow:
Validation stage
Label | Usage |
---|---|
~"workflow::validation backlog" |
Applied by the Product Manager for incoming issues that have not been refined or prioritized. |
~"workflow::problem validation" |
Applied by the Product Manager for issues where the PM is developing a thorough understanding of the problem |
~"workflow::design" |
Applied by the Product Manager or Designer (or Analytics Instrumentation Engineer) to ideate and propose solutions. The proposed solutions should be reviewed by engineering to ensure technical feasibility. |
~"workflow::solution validation" |
Applied by the Product Manager or Designer (or Analytics Instrumentation Engineer) to validate a proposed solution through user interviews or usability testing. |
Build stage
Label | Usage | |
---|---|---|
~"workflow::planning breakdown" |
Applied by the Product Manager for Engineers to begin breaking down issues and adding estimates. | |
~"workflow::ready for development" |
Applied by either Engineering or Product Manager after an issue has been broken down and scheduled for development. | |
~"workflow::in dev" |
Applied by the Engineer after work (including documentation) has begun on the issue. An MR is typically linked to the issue at some point throughout this stage. | |
~"workflow::in review" |
Applied by the Engineer indicating that all MRs required to close an issue are in review. | |
~"workflow::verification" |
Applied by the Engineer after the MRs in the issue have been merged, this label is applied signaling the issue needs to be verified in staging or production. | |
~"workflow::complete" |
Applied by the Engineer after all MRs have merged and the issue has been verified. At this step, the issue should also be closed. | |
~"workflow::blocked" |
Applied by any team member if at any time during development the issue is blocked. For example: technical issue, open question to PM or PD, cross-group dependency. |
OKRs & Epics
Analytics Instrumentation uses OKRs to plan quarterly goals and direct prioritization. To make key results present in daily prioritization we create an epic for every objective or key result, like in the FY25Q2 OKRs. Our long term roadmap is written down on our direction page.
Issue boards
We use issue boards to track issue progress on a daily basis. Issue boards are our single source of truth for the status of our work. Issue boards should be viewed at the highest group level for visibility into all nested projects in a group.
- Analytics Instrumentation Issue Board - Current Milestone
- Analytics Instrumentation Issue Board - by milestone
Picking something to work on
Engineers can find and open the board for the current milestone. Engineers should start at the top of the “workflow::ready for development” column and pick the first available, non-assigned issue. When picking an issue, the engineer should assign themselves as a signal that they are taking ownership of the issue and move them to “workflow::in development” to signal the start of development.
If the next available issue is not a viable candidate (due to amount of capacity vs. issue weight, complexity, knowledge domain, etc.) the engineer may choose to skip an issue and pick the next issue in order of priority.
If all work within a milestone is picked, engineers are free to choose what to work on. Acceptable options include:
- Post in Slack channel to see if any engineers would like help/pair on something they are working on
- Reach out to EM to pull something from the next milestone into the current one
- Create/work on tech debt issue
- Work on a passion issue
- Other (study, research, learning)
Prioritization
We prioritize our product roadmap in the Issue Board by Milestone. Issues appear on each list in order of priority and prioritization of our product roadmap is determined by our product managers.
60% of our development time is spent on issues priotized by product management and the remaining 40% on issues prioritized by engineering as described in our engineering initiatives.
Milestone Planning and Timeline
Our team mostly follows the Product Development Timeline as our group is dependent on the GitLab self-managed release cycle.
The specific application of this timeline to the Analytics Instrumentation Milestone planning process is summarized below.
Overview
With the monthly release date being the third Thursday of the release month, the code cut-off is the Friday prior. We orient our planning around the code-cut off as the last day of the milestone
Phase | Time |
---|---|
Planning & Breakdown Phase | 2 weeks prior to code cut-off in month N |
Development Phase | Monday after code cut-off in Month N to code cut-off in Month N+1 |
1. Planning & Breakdown Phase
-
Timeline: 2 weeks prior to code cut-off in month N
-
Tasks:
-
Initial Planning
- PM: Milestone planning issue gets created (example).
- PM: Adds overall objective and theme for the milestone to the planning issue
- PM: Adds and prioritizes issues from product perspective in the planning issue.
- EM: Adds and prioritizes issues from engineering perspective as well as high priority bugs in the planning issue.
- PM/EM: Ensures all issues in the planning issue are assigned to the correct milestone and all other issues, not in the planning issue are removed from the milestone.
-
Breakdown and weighing
- EM: Removes weight from issues that have been weighed more than 6 months ago to make sure weights consider recent changes.
- EM: Moves all unweighed issues to ~“workflow::planning breakdown” stage and tags engineers for estimation.
- Engineers: Add a solution proposal if none is present yet.
- Engineers: Add their estimation, ask clarifying questions, link potential blockers, and break down the issues further if needed.
- EM: Moves estimated and broken down issues into ~“workflow::ready for development” stage.
- Engineers: Indicate their availability for the next milestone in our capacity planning sheet.
-
Final Planning
- EM: Calculates estimated capacity.
- EM: Adds issues that probably need to be carried over from the current milestone to the planning issue.
- EM & PM: Add or remove issues based on weights and capacity.
- EM: Prioritizes issues in the Issue Board by Milestone based on the planning issue.
2. Development Phase
-
Timeline:
-
Tasks:
- Engineers: Work on the issues in the milestone based on the outlined priority:
- Engineers assign themselves to issues based on interest/experience.
- If no more issues are available in the milestone, they first check if they can take over or help with problems in the milestone assigned to another engineer. Otherwise, they inform the EM, who pulls in issues from the next milestone.
- PM/EM: Present the plan for the milestone in the first Analytics Instrumentation sync of the new milestone focusing on the overall objective and themes.
Milestone Capacity
Our milestone capacity tells us how many issue weights we can expect to complete in a given milestone. To estimate this we calculate the average daily weight completed by an engineer per day across the previous three milestones. This is multiplied with the actual working days available to us in a given milestone.
Previous Three Milestones:
- Total weights completed: 140 weights
- Available work days: 60 days * 5 engineers = 300 days
- Actual work days: 300 days - 20 days off = 280 days
- Average weight per engineer/day: 140 weights / 280 days = 0,5 weights/day
Next Milestone:
- Available work days: 20 days * 5 engineers = 100 days
- Actual work days: 100 days - 10 days off = 90 actual days
- Maximum capacity: 90 days * 0,5 weights/day = 45 weights
In this example, the next milestone’s capacity is 0,5 weights for the whole team. Keep in mind that neither estimations nor this calculation are an exact science. The capacity planning is supposed to help the EM and PM set realistic expectations around deliverables inside and outside time. We do not expect to hit the exact amount of predicted weights.
Milestone Commitment
A milestone commitment is a list of issues our team aims to complete in the milestone. The product team follows our GitLab principle of planning ambitiously and therefore expect that we won’t always be able to deliver everything that we wanted in every milestone. After issues are broken down, estimated, and prioritized, the product manager will apply the ~Deliverable
label to applicable issues. Issues marked with the ~Deliverable
label represent the commitment we are intending to ship in that milestone.
Due dates
To properly set expectations for product managers and other stakeholders, our team may decide to add a due date onto an issue. Due dates are not meant to pressure our team but are instead used to communicate an expected delivery date.
We may also use due dates as a way to timebox our iterations. Instead of spending a month on shipping a feature, we may set a due date of a week to force ourselves to come up with a smaller iteration.
Estimation
We estimate issues async and aim to provide an initial estimate (weight) for all issues scheduled for an upcoming milestone.
We require a minimum of two estimations for weighing an issue. Exceptions can be made for issues that come up during the milestone or if only a single engineer with the required specialty is available. We consider reacting with a ➕ emoji to the estimation as agreeing with it (and thus contributing to the minimal count of estimations). If both estimations agree, the engineer who did the second estimation should add the agreed-upon weight to the issue. If there is disagreement, the second engineer should @-mention the first one to resolve the conflict.
Estimating includes adding a “Proposed Solution” to the issue if none is documented yet or the estimation brings up a different one than originally documented. Spikes are exempted from this as discovering solutions is their main point and we default spike issues to a weight of 8.
In planning and estimation, we value velocity over predictability. The main goal of our planning and estimation is to focus on the MVC, uncover blind spots, and help us achieve a baseline level of predictability without over-optimizing. We aim for 70% predictability instead of 90%.
If an issue has many unknowns where it’s unclear if it’s a 1 or a 5, we will be cautious and estimate high (5).
If an initial estimate needs to be adjusted, we revise the estimate immediately and inform the Product Manager. The Product Manager and team will decide if a milestone commitment needs to be changed.
Issues estimation examples
Weight | Definition | Example (Engineering) |
---|---|---|
1 | The simplest possible change. We are confident there will be no side effects. | Add missing metric definition for “counts_monthly.promoted_issues”, Add instrumentation classes for license standard metrics, Update Registration Features text |
2 | A simple change (minimal code changes), where we understand all of the requirements. | VersionApp: Add indexed on other tables that are exported, Set values for StandardContext in Frontend |
3 | A simple change, but the code footprint is bigger (e.g. lots of different files, or tests effected). The requirements are clear. | Update Registration Features CTA for repository size limit, More paid features available to free users |
5 | A more complex change that will impact multiple areas of the codebase, there may also be some refactoring involved. Requirements are understood but you feel there are likely to be some gaps along the way. | Spike Service Ping health dashboard, Remove deprecated metric status |
8 | A complex change, that will involve much of the codebase or will require lots of input from others to determine the requirements. | Dispatch Snowplow events from their event definitions, Add metrics yml files for usage data metrics definition |
13 | A significant change that may have dependencies (other teams or third-parties) and we likely still don’t understand all of the requirements. It’s unlikely we would commit to this in a milestone, and the preference would be to further clarify requirements and/or break in to smaller Issues. | Create Snowplow monitoring framework, Enable batch counting for some individual queries |
? | For issues where don’t know how to estimate |
Estimation template
The following is a guiding mental framework for engineers to consider when contributing to estimates on issues.
### Refinement / Weighing
**Ready for Development**: Yes/No
<!--
Yes/No
Is this issue sufficiently small enough, or could it be broken into smaller issues? If so, recommend how the issue could be broken up.
Is the issue clear and easy to understand?
-->
**Weight**: X
**Reasoning**:
<!--
Add some initial thoughts on how you might break down this issue. A bulleted list is fine.
This will likely require the code changes similar to the following:
- replace the hexdriver with a sonic screwdriver
- rewrite backups to magnetic tape
- send up semaphore flags to warn others
Links to previous example. Discussions on prior art. Notice examples of the simplicity/complexity in the proposed designs.
-->
**Iteration MR/Issues Count**: Y
<!--
Are there any opportunities to split the issue into smaller issues?
- 1 MR to update the driver worker
- 1 MR to update docs regarding mag tape backups
Let me draw your attention to potential caveats.
-->
**Documentation required**: Y/N
<!--
- Do we need to add or change documentation for the issue?
-->
Epic, issue and MR creation
We aim to create issues in the same project as where the future merge request will live and we aim to create epics at the topmost-level group that makes the most sense for its collection of child epics and issues.
For example, if an experiment is being run in the CustomersDot, the epic should be created in the gitlab-org
group, and the issue should be created in the gitlab-org/customers-gitlab-com
project.
We focus on creating the issue in the right project to avoid having to close and move it later in the development process.
When creating an issue, use the linked template and follow its instructions.
In case the issue is not created for the GitLab project, copy the template’s content into the appropriate project.
Ratio of issues to MRs
The ratio of issues to MRs is at the responsible engineer’s discretion. MRs should follow the MVC principle. If it is evident in advance that an issue will require more than 2 MRs we should evaluate whether we can split the issue further to document the split of the work more clearly.
Merge request labels
MR labels should mirror issue labels (which is automatically done when created from an issue):
Required labels
- Section:
~section::analytics
- Group:
~group::analytics instrumentation
- Type:
~"type::bug"
,~"type::feature"
,~"type::tooling"
,~"type::maintenance"
Milestones
We tag each issue and MR with the planned milestone or the milestone at time of completion.
Team Meetings
Our group holds synchronous meetings to gain additional clarity and alignment on our async discussions. We aim to record all of our meetings as our team members are spread across several timezones and often cannot attend at the scheduled time.
Meeting rules
- Agenda items should be filled in 6 hours before meetings otherwise it’s possible to cancel the meeting.
- It’s fine to add agenda items during the meeting as things come up in sync meetings we might not have thought about beforehand.
- Meetings start :30 seconds after start time
- Whoever has the first agenda item starts the meeting.
- Whoever has the last agenda item ends the meeting.
- Meetings end early or on time.
- Any missed agenda items are bumped to the next meeting.
Our meetings
- Analytics Instrumentation Sync: an optional weekly meeting for the Analytics Instrumentation team to discuss any topics they please. At the end of the sync we shut off the recording and continue the call as a social chat.
Development metrics
Error budget
We maintain UsageData API endpoints under the service_ping
feature to track events, and because of this we must monitor our budget spend.
To investigate budget spend, see the overview and details Grafana dashboards for Analytics Instrumentation. You can also check requests contributing to spending the budget in Kibana by filtering by the service_ping
feature. An example Kibana view can be found here.
Note that the budget spend is calculated proportionally by requests failing apdex or failing with an error, and not by how much the target is exceeded. For example, if we had an endpoint with a set goal of 1s request duration, then bringing the request duration from 10s to 5s would not improve the budget.
Incidents
Within Analytics Instrumentation we have an incident process that’s separate to GitLab’s main incident management process, because most incidents that Analytics Instrumentation faces are not customer facing and can be resolved more effectively within the team. The process below outlines the different stages of the Analytics Instrumentation incident detection and resolution process and the steps to be taken by the corresponding Directly Responsible Individuals (DRIs). Please reach out to the Analytics Instrumentation Group EM/PM for any recommendations to changes in the process.
Incident Definition
We define incidents as a deviation from the intended process that significantly disrupts the reporting of metrics to the point that immediate action is required. These reasons should lead to the creation of an incident:
- Any SEV-1 Monte Carlo alert posted into #g_analytics_instrumentation_alerts that’s not directly associated with an exisiting incident.
- Any disruption of event collection on our main Snowplow event collection infrastructure.
- Any other suspected loss or delay of analytics data that might affect metrics with a
performance_indicator_type
and could be originating in the Analytics Instrumentation domain.
Examples of incidents:
When choosing whether to declare an incident:
- Rather err on the side of declaring an incident even if you run the risk of it not being one. We’d rather close an incident as false positive than miss out on one.
- If there’s an existing incident issue with the data team still follow our process and link the data team’s issue in ours.
Incident Creation
(DRI: The team/individual detecting the issue or first team member to see the alert)
- Create an issue and fill all necessary information using the Analytics Instrumentation Incident Template.
- Add appropriate label using the below guideline
~"Analytics Instrumentation::Incident-High Severity"
for impending loss of data for many metrics or moderate to severe loss in business critical metrics that have a performance_indicator_type value.~"Analytics Instrumentation::Incident-Medium Severity"
for data delay.- Add an appropriate bug severity as per the definition
- For cases when there is minimal impact on data and manual steps or correction is needed, please raise a bug rather than an incident.
- Assign the issue to Analytics Instrumentation Group PM and EM.
- Post in the #g_analyze_analytics_instrumentation slack channel and tag Analytics Instrumentation Group PM and EM.
- Notify these slack channels #analytics-section, #data-rd-analytics, #data with link to the issue.
- Depending on your own experience either take on the role of resolution DRI, or actively tag EM and engineers in slack to find DRI for incident.
Incident Resolution
(DRI: To be identified by EM of the Analytics Instrumentation group)
- DRI to work on resolving the issue as quickly as possible. The first priority is to find a fix, even if that is a temporary one, before working on a long term resolution. Our monitoring and troubleshooting guide can be helpful here.
- EM to review severities assigned by detection DRI.
- In case of a ~“Analytics Instrumentation::Incident-High Severity” issue:
- DRI to create a temporary channel for the incident in Slack and invite the whole group including PM and relevant stakeholders based on the incident.
- EM (or DRI if EM not available) to announce a feature change lock specific to the analytics instrumentation group.
- All group members concentrate on finding a fix for the issue.
- DRI to post an update in the channel about the current status at least twice per day.
- In case of a ~“Analytics Instrumentation::Incident-Medium Severity” issue:
- DRI to create a slack thread within #g_analyze_analytics_instrumentation for coordination around the incident.
- DRI to post daily updates in the channel and issue about the current status of the incident.
- DRI to update the incident issue with an expected resolution time as soon as one is clear.
- PM/EM, in coordination with the individual/team reporting the incident, to close the issue after verifying that the fix is working.
- If a patch release is necessary:
- DRI to create a merge request for a patch release if required and link the merge request to the main issue
- DRI to announce in the main issue when the Patch release is completed
Incident Notification
(DRI: The PM of the Analytics Instrumentation group)
- Notify these slack channels #g_analyze_product_analytics, #data_rd_fusion, #data with link to the issue.
- Inform Monitor stage Engineering & Product GPM.
- Update aforementioned slack channels and individuals on resolution time, changes to resolution times, and when incident is resolved.
- Ensure the incident and status is reflected in the next monthly state of data issue.
Out of office coverage process
An OOO coverage process helps reduce the mental load of “remembering all the things” while preparing for being away from work. This process allows us to organize the tasks we need to complete before time off and make the team successful.
Open a new issue in the Analytics Instrumentation project with the out_of_office_coverage_template
.
Onboarding
All new team members to the Analytics Instrumentation teams are provided an onboarding issue to help ramp up on our analytics tooling. New team member members should create their own onboarding issue in the gitlab-org/analytics-section/analytics-instrumentation/internal project using the engineer_onboarding
template.
Quick Links
Resource | Description |
---|---|
Internal Analytics Docs | Docs for instrumenting internal analytics at GitLab |
Analytics Instrumentation Monitoring and Troubleshooting | Information around Troubleshooting Analytics Instrumentation infrastructure |
Analytics Instrumentation Infrastructure | Information about the infrastructure we run |
Service Ping Guide | An implementation guide for Service Ping |
Privacy Policy | Our privacy policy outlining what data we collect and how we handle it |
Analytics Instrumentation Direction | The roadmap for Analytics Instrumentation at GitLab |
Internal Analytics Infrastructure
2eb0e162
)