AI Model Validation Group
π Mission
The AI Model Validation team mission is to support and improve the integrity, reliability, and effectiveness of Generative AI solutions through evaluation, validation, and research science processes for GitLab Duo Features. We offer a centralized evaluation framework that promotes data-driven decision-making and pragmatic refinement of AI features. We further explore methods and techniques on forward-deployed research science in the Gen AI space.
Direction
Here is the group direction page. We have two categories under this group: Category AI Evaluation and Category AI Research. This group is part of the AI-Powered stage.
Central Evaluation Framework (CEF)
Model validation is the primary maintainer of our Central Evaluation Framework (CEF). This tool supports the entire end-to-end process of AI feature creation, from selecting the appropriate model for a use case to evaluating the AI featuresβ output. AI Validation works in concert with other types of evaluation, such as SET Quality testing and diagnostic testing, but is specifically focused on the interaction with Generative AI.
Read about how we How we validate and test AI models at scale on the GitLab blog.
π Team members
The team is composed of ML engineers focused on ML science and MLOps backend, and they are permanent members of the AI Model Validation Group.
Who | Role |
---|---|
Hongtao Yang | ML Engineer |
Andras Herczeg | Senior ML Engineer |
Stephan Rayner | Senior ML Engineer |
Tan Le | Senior ML Engineer |
Susie Bitters | Senior Product Manager |
βοΈ How to contact us
- Tag a team member in a merge request or issue
- Slack Channel:
#g_ai_model_validation
How we work
Estimation
The estimation and planning process is managed primarily asynchronously; however, synchronous calls may be scheduled if necessary to clarify blockers or ensure alignment.
A week before the milestone is finalized, the team is provided with a list of issues to evaluate. Throughout that week, team members engage in discussions to assign estimates or weights to tasks. Any issues that are too large or lack team consensus are either reworked, removed from the milestone, or scheduled for further investigation as a spike.
The task estimation and planning process follows a structured workflow:
- The Product Manager (PM) maintains a comprehensive list of all issues relevant to the product’s progress.
- The Engineering Manager (EM) works with the Product Manager (PM) to select the subset of issues that the team will evaluate for the upcoming release, ensuring the list is manageable and aligned with team priorities.
- Domain experts within the team may be consulted to clarify ambiguous issues before estimation, ensuring that tasks are well understood before being assigned weights.
By the end of the estimation process, every issue designated as a deliverable for the upcoming release must meet the following criteria:
- Each issue must have a weight or be time-boxed.
- There must be a clear description of one or more actionable tasks that make up the issue.
- The issue must include a definition of done to guide the development process and ensure clarity around when a task is considered complete.
Weighting issues
We use weights to forecast the complexity of each given issue aimed at being scheduled into a given milestone. These weights help us ensure that the amount of scheduled work in a cycle is reasonable, both for the team as a whole and for each individual. We understand weights are mere forecasts and we accept the uncertainty that comes with this.
Before each milestone, the team sets weights on all issues currently aimed at the next milestone by Product and triaging processes, along with updating the description with the proposed solution if agreed upon. This exercise may require multiple team members to collaborate on some issues in order to properly set a weight and agree on a solution. The weights we use are:
Weight | Description |
---|---|
1: Trivial | The solution is understood, no extra investigation is required, and anyone can pick this up. This task should take no more than 1-2 hours. |
3: Medium | The solution is understood, but some extra investigation or effort will be required to realize the solution. One engineer should be able to finish 2-3 of these tasks in a week. |
5: Large | The solution is outlined, but there are unknowns with the workβthese issues can be scheduled but should be broken down first. Likely a major piece of work, potentially worked on by two engineers. |
8: Unknown | There are many unknowns with the proposed solution or how to implement it. These issues will not be scheduled and instead should be broken down, or a spike should be scheduled to investigate further. |
Retrospectives
We conduct monthly async retrospectives which are located here.
Customer outcomes we are driving for GitLab Duo
The customer outcomes we are focused on can be divided into themes below:
Benchmark for Quality and Performance Metric of Foundational Model and Feature
We first assess the models and the feature on a large-scale dataset to understand and benchmark quality metrics from a set of metrics. We provide dashboards for diagnostic purposes as well as a continuous daily run dashboard so we can track how the features are performing based on the benchmark.
Evaluation as a Tool for Software Engineers to Experiment as They Iterate and Build AI Features
After the initial assessment, we have a dynamic dataset pull from scheduled runs so feature teams can run the datasets with every code and prompt change via CLI. This helps them understand how changes in code/prompt/system can impact quality based on the variance between control (before change) and test (after change) code on a primary metric of choice.
Documentation as We Use New Ways and Processes for GenAI Evaluation
We are further iterating and documenting an evaluation-centric way of building GenAI features. This is mainly for the internal team, and the epic to track this can be found here.
Our current customers include GitLab AI-powered Duo feature teams:
The immediate customers include:
- AI Powered: Duo-Chat team
- Create: Code Creation team
- Govern: Threat Insights
AI Vulnerability Management
team - Root Cause Analysis
- RAG Evaluation
- Issue Summarization
- AI Powered: Group Custom Models
π§ͺ Top FY25 Priorities
Data-driven evaluated AI solutions with every code change. We encompass two categories: AI Evaluation and AI Research. Our goal is to empower each team building AI features to confidently deliver meaningful and relevant features for GitLab customers. As a long-term initiative, we aim to expand our Centralized Evaluation Framework to assess various models, AI features, and components based on quality, cost, and latency. The primary decision factors for AI content quality are:
- Is it honest? (consistent with facts)
- Is it harmless? (does not include content that might offend and harm)
- Is it helpful? (accomplishing the end goal of the user)
We also aim for AI Engineers to leverage the Centralized Framework for experimenting and expanding from prompt engineering, RAG, Agent, to model tuning. This can be achieved through the Framework’s API for the Prompt Library, recognizing that every code change significantly impacts the input and output of LLMs.
Further, there are novel research topics, and we would love for GitLab to be represented in the AI research community by publishing our approaches on evaluation.
π Prompt Library (Data)
We create large libraries (prompts as data) that serve as a proxy to production. We do this by understanding the various complexities of tasks and methods, allowing us to holistically evaluate a set of data beyond a few tests and more as a performance in production. We do this with a combination of industry benchmarks and customized dataset for various tasks. The current tasks we have included or are planning to include in the prompt library are as follows:
- Code Completion
- Code Generation
- Code Explantation
- Issue/Epic Question Answering
- GitLab Documentation Question Answering Dataset
/slash
Commands:/explain
,/test
, and/refactor
(In Progress)- Vulnerability Explanation/ Resolve (In Progress)
- Root Cause Analysis (In Progress)
- Feature Summarization (To be added)
We further are planning to build customized workflow dataset particularly for System ( RAG , Agent) and Contextual Evaluation ( text follow up questions)
π Metrics
There are a few different metrics that we use to assess. If we have established ground truth, we conduct an assessment with similarity and cross-similarity scores. If ground truth is not established, we use Consensus Filtering as an LLM-based evaluator through an Independent LLM Judge and a Collective LLM Judge. We are always iterating and evolving our metric pipeline.
Similarity Score
This metric evaluates the degree of similarity between an answer generated by a point solution and those produced by other LLMs, such as Claude-2, Text-Bison, and GPT-4, in response to the same question or to ground truth.
Independent LLM Judge
This metric involves soliciting evaluations from LLM Judges to assess the quality of answers provided given a specific question and context. Judges are tasked with assigning scores based on three key aspects: correctness, comprehensiveness, and readability. To enhance the credibility of these scores, multiple LLMs can participate as judges. For instance, if three judges unanimously agree that an answer is subpar, we can confidently conclude its quality.
Collective LLM Judge
This metric operates similarly to the “LLM Judge” metric but consolidates all answers generated by each answering model into a single prompt. Judges are then tasked with comparing these consolidated responses and assigning scores accordingly.
Performance Metrics (To be Added)
In addition to the similarity and consensus-based metrics, we also will track several performance metrics to evaluate the overall effectiveness of our AI models and features:
-
Latency: We measure the time it takes for a model to generate a response, ensuring that the latency is within acceptable limits for a seamless user experience.
-
Requests per Second (Concurrency): We monitor the number of requests our models can handle per second, allowing us to understand the scalability and capacity of our AI infrastructure.
-
Tokens per Second: We monitor the counts of the tokens rendered per second during LLM response streaming. This helps assesses the speed and efficiency of the LLM in generating and streaming responses, which is critical for user experience in real-time applications.
By continuously monitoring these performance metrics, we can make data-driven decisions to optimize the performance, reliability, and user experience of our AI-powered solutions through our synthetic CEF as proxy to production.
π¦ Team Processes
We have a globally distributed team spanning across EMEA, AMER, and APAC. We hold two synchronous sessions weekly, emphasizing the team’s preferences on the schedule and periodically changing the time based on these preferences. We have meetings dedicated to milestone planning, as well as engineering discussions and ideation.
π Regular Team Meetings
Team Meetings
-
Weekly Team Sync
- When: Every Wednesday, 21:00 GMT
- What: This meeting is dedicated to working on the vision and roadmap. The Engineering Manager and Product Manager ideate, discuss, and assign work as needed for the entire team.
-
Weekly Engineering Sync
- When: Every Tuesday, 21:00 GMT
- What: This meeting is dedicated to the engineering team for the purpose of syncing up on progress, discussing technical challenges, and planning the upcoming week and milestones. This is also to ideate on future milestones and building validation as a product.
-
Quarterly Creative Destruction Labs
- When: Once in 10 weeks
- What: This is a 48-hour working session, comprised of both synchronous and asynchronous activities, where the team comes together under the AI research category as part of a lab. The goal is to take a topic, deconstruct the old approach, and rebuild it in a new way to rapidly iterate toward the product roadmap and vision.
π Shared Calendars
- AI-Powered Stage Calendar (Calendar ID:
c_n5pdr2i2i5bjhs8aopahcjtn84@group.calendar.google.com
)
π Weekly and Quarterly Updates
Each week, the team publishes a report on the progress made during the past week and outlines the focus for the upcoming week. The report also includes a GenAI reading list to ensure that the engineers stay up-to-date in the ever-changing GenAI space here.
We also publish a quarterly report that summarizes how we perform in reference to OKRs, highlights our achievements, celebrates milestones, identifies opportunities, and shares learnings here.
πΉ GitLab Playlist
We conduct regular walkthroughs as we add data, metrics, and evaluation workflows. GitLab AI Model Validation Playlist includes a list of these walkthroughs. Some videos published might be for internal purposes only.
π― Current OKR
Our current OKR can be viewed here (GitLab internal).
π Epics and Themes
We have two major epics that can be subdivided into further sub-epics and issues. The themes are based on the Category AI Evaluation and Category AI Research as below.
π How to work with us?
We have issue templates for requesting a new model evaluation or for evaluating a feature (Internal Only). Below are the request templates that can be used.
- If a feature team would like a model to be evaluated for a certain task, here is the request template: Model Request
- If a feature team would like to evaluate a certain use case, here is the request template: Use-Case Request
- If a feature team finds bugs that impact their ability to use the framework, here is a template to request a fix in a new issue. Additionally encourage to post on #g_ai_model_validation
Further, we iterate and act more quickly with feedback and here is the best place to provide feedback.
π Dashboards and Additional Resources (Internal Only)
- GitLab Duo-Chat Evaluation Dashboard
- GitLab Code Completion Foundational Model Dashboard
- GitLab Code Suggestion Dashboard
- GitLab AI Testing and Evaluation Framework
- How to use CEF for Duo-chat Experimentation
- How to use CEF for Code Suggestion Experimentation
- Prompt Template for Models and Tasks Evaluated
- General Guidance of A/B Testing using CEF
π Additional Resources
- Current Milestone Commitments (don’t forget to filter by current milestone!)
- GitLab Validation Metrics
- GitLab Evaluation Procedure
- Blog: How we validate and test at scale
Required labels
- Group:
~group::ai model validation
b38578d9
)