Using Feast
Feast is an open source feature store aimed to improve the process of obtaining, documenting and deploying features for data science (DS) models. In the DS GitLab team we are currently focused on using the Offline Store (used for batch non-realtime predictions). The repository for Feast can be found HERE
Note: Since Feast feature transformations capabilities are not yet mature enough and could not match the performance capabilities we need, we have opted for creating a separate fork of Feast where we optimize feature transformations based on our particular use case. You can read more about the fork, HERE.
The most important concepts to understand the Feast Offline Store are:
- Data Sources: The building blocks that contain the features that will be served. These can be specified as either a physical data source (CSV, parquet file, …) or a SQL query. Since Feast is currently lacking a feature transformation engine, we define any feature transformations as SQL queries.
- Feature Views: Built with data sources and metadata of fields such as descriptions, tags, field type. Feature views are used when creating a new model, can be fetched through Feast.
- Feature Services: Combination of fields from feature views that are needed to make inferences with the trained Machine Learning models. Feature services are used at the end, when we know which features are needed in the final model.
graph LR A[Data Sources] --> B[Feature Views] A ~~~|"Data from Snowflake transformed through SQL"| A B ~~~|"Data ready to be fetched through Feast"| B C ~~~|"Group of feature views that define a ML model (e.g. PtC, PtE, ...)"| C B --> C[Feature Services]
Advantages of using a feature store
- Reusable Feature Pipelines: Reduces time to build predictive models by enabling the creation of feature pipelines that can be shared and reused across different teams and projects.
- Consistency in Feature Engineering: Ensures uniformity of features during both training and production stages, as features are engineered once and used in multiple contexts; reduces SQL redundancies
- Point-in-Time Correctness: Automates the maintenance of historical data which helps with retrieving correct values at any specific point in the past, crucial for predictive accuracy.
- User Interface (UI): Facilitate easier discovery and management of features, reducing time to model creation.
- Model Feature Versioning: Implements version control for features, aiding in tracking changes and maintaining consistency across model iterations.
- Online Feature Store: Supports real-time predictions with an online store, potentially valuable for future projects requiring immediate data processing.
Setup
Feast uses a registry that acts as a central catalog which stores the feature definitions and the necessary metadata. At GitLab we store the registry in Snowflake.
There are two ways of interacting with the Feature Store:
- Access already created feature views and services
- Creating new feature views / services locally and testing them in your own unique database/Feast registry before asking for it to be merged to the production registry
For both of these workflows you need a read access to the FEATURE_STORE.FEAST
schema in Snowflake, which can be granted through Access Request.
To connect to the feature store we use the following command in python / jupyter notebook:
from feast import FeatureStore
store = FeatureStore(fs_yaml_file=[credential_yaml_file])
Here [credential_yaml_file]
indicates the YAML file with the necessary credentials to connect to the feature store. There are currently 3 credential files that can be chosen (these can be found in the Feast repository):
- production_feature_store.yaml: This one connects to the production registry stored in the Snowflake database.schema
FEATURE_STORE.FEAST
. This YAML is used by the models ran in production and/or trained by CI. - production_local_feature_store.yaml: Similar to above, instead it uses your Snowflake credentials instead of production ones.
- staging_local_feature_store.yaml: This one connects to your personal Snowflake database and uses the registry stored there. For example,
{YOUR_USER}_PROD.FEAST
. This yaml file is only needed if you intend to follow the second workflow.
Instruction
If you only want to interact with the Feature Store as described in the first option (fetching already made feature views / services), then follow only the first 3 steps. If you also want to create new and/or modify feature views and services, complete all of them:
-
Set up the environment variables in your
~/.zshrc
file. These are the variables that are referenced in the[credential_yaml_files]
:# Feast Development variables export FEAST_DEV_SNOWFLAKE_ACCOUNT="GITLAB" export FEAST_DEV_SNOWFLAKE_USER="{USER}@gitlab.com" export FEAST_DEV_SNOWFLAKE_AUTHENTICATOR="externalbrowser" export FEAST_DEV_SNOWFLAKE_ROLE="{USER}" export FEAST_DEV_SNOWFLAKE_WAREHOUSE="DEV_XS" export FEAST_DEV_SNOWFLAKE_SCHEMA="FEAST" export FEAST_DEV_SNOWFLAKE_DATABASE="{USER}_PROD" export FEAST_PROD_SNOWFLAKE_DATABASE="FEATURE_STORE" export FEAST_PROD_SNOWFLAKE_SCHEMA="FEAST" export FEAST_USAGE=False
Where user correspond to your snowflake user. For example, JPEGUERO. Make sure {USER} is in uppercase.
-
Run the command:
source ~/.zshrc
-
In Snowflake, run the following command:
CREATE OR REPLACE SCHEMA {USER}_PROD.FEAST;
. For example:CREATE OR REPLACE SCHEMA JPEGUERO_PROD.FEAST;
-
To run the repository locally install pipenv.
-
Clone the repository into the
~/repos
directory and cd into it. -
Finally, run the following commands:
mkdir .venv pipenv install make run-feast-local
These commands set up the python virtual environment. If you would also like to run the sample workbook make sure to also install the dev dependencies:
pipenv install --dev
make run-feast-local
will run feast apply
which creates the necessary infrastructure to deploy the offline store by the feature definitions in the feature_repo directory.
After this, the model_feature_service_manager is ran. This creates feature services out of YAML file definitions. These definitions needs to be stored in the feature_repo/models directory to be picked up.
Running the UI locally
Currently we have the UI running in a VM that can be accessed by requesting permission to the Data Engineering team. In case you would like to run the UI locally, you can use one of these two commands:
make ui-local
: This runs the UI locally against the production database / production registry. After running the command you can access it through the followinghttps://0.0.0.0:8889
.make ui-staging
: This runs the UI against your personal development database instead of the production database (for example, when creating a MR to introduce new features and you want to test locally).
Project Structure
├── feature_repo
│ └── product <- Feature definitions for product data.
| | └── queries <- Helper SQL queries to be used in the product data sources definitions.
│ └── marketing <- Intermediate data that has been transformed.
| | └── queries <- Helper SQL queries to be used in the marketing data sources definitions.
│ └── ... <- Same as above.
| | └── queries <- Same as above.
| |
| └── models <- Contains YAML files with definitions for creation of feature services.
| |
│ ├── entities.py <- [Entity](https://docs.feast.dev/getting-started/concepts/entity) definitions for the feature store.
| |
│ └── helpers.py <- Helper functions to be used when building the feature definitions.
│ |
| └── production_feature_store.yaml <- Feast yaml file that contains credentials to sign in to database. To be used when deploying changes to production and reading features for production models.
| └── production_local_feature_store.yaml <- Feast yaml file that contains credentials to sign in to database. To be used when querying the feature store locally.
| └── staging_local_feature_store.yaml <- Feast yaml file that contains credentials to sign in to database. To be used when testing changes to the feature store locally.
│
├── model_feature_service_manager <- Module to generate feature services from a YAML file. Used to create feature services for DS models (PtC, PtE, ...)
│
├── create_views <- Module that generates views out of the features defined in Feast. Helpful for taking advantage of MonteCarlo table lineage.
Currently (V1) feature repository is split in different functional areas such as product data, marketing, sales, … Each one of those directories contains the necessary SQL queries that are used to create the data sources and the python files that create the feature views and annotate the fields with metadata.
actionable_insight
is one of the metadata fields that describes the feature, which is used to populate insights about scores. To learn how actionable insight features are used with a feature service, review feature_repo.helpers.get_actionable_insight_descriptions.
Workflow
Building ML models with Feast
Workflow summary:
-
Go to the Feast UI and find the features views that you want to use to train your model. Example:
product_stage_by_account_5_period_unit
andproduct_stage_change_metrics_1_period_unit
-
Copy the
production_feature_store.yaml
,production_local_feature_store.yaml
andstaging_local_feature_store.yaml
to your ML project repository root. -
Access feature views using the function get_fields_from_feature_views
-
Note that our ML model might only need a fraction of the fields in the feature views for inference, therefore you can create feature services. To do that, in the model directory create a YAML file that defines the name of the feature service and the name of the feature views used together with the required fields. An example can be seen here for the PtC V4 model.
- An utility function that is helpful to create such YAMLs is create_yaml_for_model_feature_service.
-
Create a MR in this project and add the feature service yaml to the models directory. Test that the feature service contains all the features that you need.
- Feast can be build in a personal database by following the above Setup section. It allows to do local testing before pushing the changes to production. Make sure to use the
staging_local_feature_store.yaml
when fetching the feature services when working with the MR workflow.
- Feast can be build in a personal database by following the above Setup section. It allows to do local testing before pushing the changes to production. Make sure to use the
-
Push the changes to the remote branch in the repository. Wait for the CI jobs
clone-image
(clones the Docker image from production so Feast can run) andclone-feature-store-db
(clones theFEATURE_STORE.FEAST
schema from production to a dedicated MR database) to pass. -
Test the changes introduced to the feature store by running the CI job
test-on-mr
. This job runs the commandfeast apply
on the MR feature store database testing that the feature store can run after the changes introduced. -
After merging the MR and the feature service is now in production, connect to the feature store using the
production_local_feature_store.yaml
credential file orproduction_feature_store.yaml
when running on CI. Going back to our python code, this would look like:from feast import FeatureStore if is_local_development: store = FeatureStore(fs_yaml_file='path/to/production_local_feature_store.yaml') else: store = FeatureStore(fs_yaml_file='path/to/production_feature_store.yaml')
Sample jupyter notebook can be found HERE.
Also, make sure to check out the current implementation of PtC, specifically create_prod_models.ipynb and scoring_code.ipynb, as it is a model fully created and moved to production CI using Feast.
From Data Sources to serving features
Per diagram at the top of the page, there are three stages to serve features in Feast:
- Data Sources
- Feature Views
- Feature Services
Defining Data Sources and feature transformations
There are two ways of accomplishing this
Using scaffold queries (old method - less efficient for bigger tables)
To address lacking transformations in Feast, following types of SQL queries are used in the GitLab Feast repository:
- Simple aggregations such as SUM, AVG (with optional grouping by account id, opportunity id, etc)
- Rolling / window aggregations - summing / averaging a certain value over a rolling window of past months / weeks / days. Example can be found in product usage period usage query, to create this type of aggregation:
a. Create a “scaffold” CTE, this is, for every distinct account in the base table that the data source is based on we create an entry for the aggregation date we intend to use. In this case this means take all the entities together with all the snapshot_months.
b. Join the scaffold back to the base CTE, we join on the main entity_id AND on the desired interval of aggregation date. In this case, we wanted a rolling window of 5 months, and this was accomplished with:
base.snapshot_month BETWEEN ADD_MONTHS(scaffold.snapshot_month, -5) AND scaffold.snapshot_month
. - Period over period transformation. A good example is product usage change metrics query. The pattern here is to define a base CTE where we have the table at the wanted aggregation period, then we create a period_2 CTE which will be the data in the past. In this example we wanted to shift the data by 1 month in the past and this is accomplished with
DATEADD('month', 1, snapshot_month) AS snapshot_month_2
in the period_2 CTE.
The data source table in Feast should have following format:
Entity_id (e.g. CRM_ACCOUNT_ID) | event_timestamp | metric (e.g. product_usage_unique_active_users) |
---|---|---|
xxxx | 2023-01-01 | 44 |
xxxx | 2023-02-01 | 49 |
The main entity_id
in the data source is unique when combined with the event_timestamp
. The metric field is the result of the SQL transformation. In case of product usage period usage query this would be a CRM Account per event_timestamp (here: monthly aggregated data), and the metric would be something like the window average of unique_active_users
in the period_unit
of aggregation. If period_unit = 5
we would have that the record of 2023-01-01 is averaging unique_active_users
from 2022-09-01 to 2023-01-01, the record of 2023-02-01 averages from 2022-10-01 to 2023-03-02 etc
Using jinja templating
As explained before, we are using a fork of Feast to add additional features to Feast. Namely, adding the use of Jinja templating to speed up feature transformation, specifically transformations involving rolling / window aggregations.
In the previous approach to creating feature transformations that need to consider these window aggregations we would have a scaffold query like this:
WITH base AS (
SELECT *
FROM prod.workspace_data_science.monthly_stage_usage_by_account
), dim_date AS (
SELECT first_day_of_month AS snapshot_month
FROM prod.common.dim_date
WHERE date_actual <= CURRENT_DATE
AND date_actual >= '2021-02-01'::DATE
), scaffold AS (
SELECT DISTINCT
base.dim_crm_account_id,
dim_date.snapshot_month
FROM base
CROSS JOIN dim_date
)
SELECT
a.dim_crm_account_id,
a.snapshot_month AS product_usage_date,
--number of all time features used
SUM(b.stage_create_alltime_features) AS stage_create_alltime_features_cnt,
FROM scaffold a
LEFT JOIN base b
ON a.dim_crm_account_id = b.dim_crm_account_id
AND b.snapshot_month BETWEEN ADD_MONTHS(a.snapshot_month, -{period_unit}) AND a.snapshot_month
GROUP BY 1, 2, 3
to:
-- USE_TEMPLATE_WORKFLOW
WITH base AS (
SELECT *
FROM prod.workspace_data_science.monthly_stage_usage_by_account
{% if validation %}
LIMIT 100
{% endif %}
)
SELECT
a.dim_crm_account_id,
{% if get_historical_features %} b."entity_timestamp"::DATE {% endif %}
{% if validation %} '2024-05-01'::DATE {% endif %} -- just for validation, hardcore a date
AS snapshot_month,
--number of all time features used
SUM(a.stage_create_alltime_features) AS stage_create_alltime_features_cnt
FROM base a
{% if get_historical_features %}
INNER JOIN "entity_dataframe" b
ON a.dim_crm_account_id = b.dim_crm_account_id
AND a.snapshot_month BETWEEN DATE_TRUNC('month', ADD_MONTHS(b."entity_timestamp"::DATE, -6)) AND DATE_TRUNC('month', ADD_MONTHS(b."entity_timestamp"::DATE, -1))
{% endif %}
GROUP BY 1, 2
Notice how in the first query we need to create a scaffold of all combinations of the entity key (dim_crm_account_id).
In the second query, we can access the “entity_dataframe” (a list of accounts/namespaces/… that we want to retrieve the features for) using the jinja variable get_historical_features
. This part of the query will only be executed during feature retrieval (when using the feature store to train or score a model). This way, we only calculate the transformation at feature retrieval time for the entities (for example, customer_ids) and timeframe we care about, speeding the query process.
We also introduce the validation
variable. This is because when running feast apply, Feast needs to validate that the query can be built. With this we can add logic to make sure that the date key (snapshot_month) is present at validation time (as there is no entity_dataframe
at validation time). Also, it lets us select only a limited amount of rows in the base CTE to make the validation query even faster.
To let feast know it needs to use this jinja workflow you need to add a comment at the top of your SQL query, -- USE_TEMPLATE_WORKFLOW
.
You can read more about this in the Feast fork.
Parameters
Feast allows to create reusable queries that can be modified with parameters. It is useful when one model may require rolling 5 months for the product usage data, while the other - 3 months of the same product usage metrics.
- period_unit: Queries with
period_unit
result in columns that indicate window of time (usually in months). When creating the feature view, the period_unit is substituted with provided value. For example, for product usage period usage query a feature view namedproduct_usage_metrics_5_period_unit
was created by setting the{period_unit}
value to 5 months, as seen here - TODO - not implemented yet - entity: To address various levels of aggregation, variable
entity
can be used, which allows to extract specific granularity from the table. For example, when working with product-related table that is unique on both account and namespace level - entity will allow to refer to either of those granularities.
Connecting Feast to MonteCarlo (MC)
In the efforts the rest of the data teams more aware of how the DS team is using the source data to create machine learning (ML) models we started a process to make MC aware of the data that feast creates to feed into ML models. The benefits being:
- More integration of the DS team with the whole set of tools used by the rest of the team
- Easier for AEs and DEs to understand how changes in a data model might affect a downstream ML model that it is powered from that data
- Help the DS team be aware whenever there is an upstream fail in one of the tables that are critical to the DS model
Once MC is aware of the feature definitions that Feast has created, we can leverage MC table lineage tracking and notification system to accomplish the above.
To make this connection possible we have created a script in the Feast repository, create_views. This works by taking all the feature definitions stored in Feast (both Feature views and services) and creating views out of them in the PROD.LINEAGE_TRACKING
schema. Since all these feature definitions reference snowflake tables, MC can then use its lineage tracking to achieve the above.
This script is automated and runs in Gitlab CI every time a new branch is merged into the main
branch.
An example of part of the PtC table lineage can be seen here:
Future improvements
- Implement
feast apply
to run transformations for only those feature views or services that were updated (especially when testing locally) - Move helper functions to the
gitlabdata
python package. - Create authentication functions to replace the credential yaml files and avoid copy-pasting between repositories. This functionality should be added to
gitlabdata
python package.
ac0e3d5e
)