Teradata Package for Python Function Reference on VantageCloud Lake - build_time_series - Teradata Package for Python - Look here for syntax, methods and examples for the functions included in the Teradata Package for Python.
Teradata® Package for Python Function Reference on VantageCloud Lake
- Deployment
- VantageCloud
- Edition
- Lake
- Product
- Teradata Package for Python
- Release Number
- 20.00.00.08
- Published
- November 2025
- ft:locale
- en-US
- ft:lastEdition
- 2025-12-05
- dita:id
- TeradataPython_FxRef_Lake_2000
- Product Category
- Teradata Vantage
- teradataml.store.feature_store.models.DatasetCatalog.build_time_series = build_time_series(self, entity, selected_features, view_name, description=None, include_historic_records=False)
- DESCRIPTION:
Builds the dataset with start time and end time for feature values available in
feature catalog. Once dataset is created, user can create a teradataml DataFrame
on the dataset.
PARAMETERS:
entity:
Required Argument.
Specifies the name of the Entity of Object of Entity
to be included in the dataset.
Types: str or Entity.
selected_features:
Required Argument.
Specifies the names of Features and the corresponding feature version
to be included in the dataset.
Notes:
* Key is the name of the feature and value is the version of the
feature.
* Look at FeatureCatalog.list_feature_versions() to get the list of
features and their versions.
Types: dict
view_name:
Required Argument.
Specifies the name of the view to be named for dataset.
Types: str.
description:
Optional Argument.
Specifies the description for the dataset.
Types: str.
include_historic_records:
Optional Argument.
Specifies whether to include historic data in the dataset.
Default Value: False.
Types: bool.
RETURNS:
Teradataml DataFrame.
RAISES:
TeradataMlException
EXAMPLES:
# Ingest sales data to feature catalog configured for repo 'vfs_v1'.
>>> from teradataml import load_example_data, FeatureProcess
>>> load_example_data('dataframe', 'sales')
>>> df = DataFrame("sales")
>>> df
Feb Jan Mar Apr datetime
accounts
Red Inc 200.0 150.0 140.0 NaN 04/01/2017
Blue Inc 90.0 50.0 95.0 101.0 04/01/2017
Alpha Co 210.0 200.0 215.0 250.0 04/01/2017
Orange Inc 210.0 NaN NaN 250.0 04/01/2017
Yellow Inc 90.0 NaN NaN NaN 04/01/2017
Jones LLC 200.0 150.0 140.0 180.0 04/01/2017
# Create a FeatureStore.
>>> from teradataml import FeatureStore
>>> fs = FeatureStore(repo='vfs_v1', data_domain='sales')
Repo vfs does not exist. Run FeatureStore.setup() to create the repo and setup FeatureStore.
>>> fs.setup()
True
# Initiate FeatureProcess to ingest features.
>>> fp = FeatureProcess(repo='vfs_v1', data_domain='sales', object=df, entity='accounts', features=['Jan', 'Feb', 'Mar', 'Apr'])
# Run the feature process.
>>> fp.run()
Process 'a9f29a4e-3f75-11f0-b43b-f020ff57c62c' started.
Process 'a9f29a4e-3f75-11f0-b43b-f020ff57c62c' completed.
# Example 1: Build dataset with features 'Jan', 'Feb' from repo 'vfs_v1' and sales data domain.
# Name the dataset as 'ds_jan_feb'.
>>> from teradataml import DatasetCatalog
>>> dc = DatasetCatalog(repo='vfs_v1', data_domain='sales')
>>> dataset = dc.build_time_series(entity='accounts',
... selected_features = {
... 'Jan': 'a9f29a4e-3f75-11f0-b43b-f020ff57c62c',
... 'Feb': 'a9f29a4e-3f75-11f0-b43b-f020ff57c62c'},
... view_name='ds_jan_feb',
... description='Dataset with Jan and Feb features')
>>> dataset
accounts Jan Jan_start_time Jan_end_time Feb Feb_start_time Feb_end_time
0 Blue Inc 50.0 2025-06-20 12:17:14.040000+00: 9999-12-31 23:59:59.999999+00: 90.0 2025-06-20 12:17:14.040000+00: 9999-12-31 23:59:59.999999+00:
1 Red Inc 150.0 2025-06-20 12:17:14.040000+00: 9999-12-31 23:59:59.999999+00: 200.0 2025-06-20 12:17:14.040000+00: 9999-12-31 23:59:59.999999+00:
2 Yellow Inc NaN 2025-06-20 12:17:14.040000+00: 9999-12-31 23:59:59.999999+00: 90.0 2025-06-20 12:17:14.040000+00: 9999-12-31 23:59:59.999999+00:
3 Alpha Co 200.0 2025-06-20 12:17:14.040000+00: 9999-12-31 23:59:59.999999+00: 210.0 2025-06-20 12:17:14.040000+00: 9999-12-31 23:59:59.999999+00:
4 Jones LLC 150.0 2025-06-20 12:17:14.040000+00: 9999-12-31 23:59:59.999999+00: 200.0 2025-06-20 12:17:14.040000+00: 9999-12-31 23:59:59.999999+00:
5 Orange Inc NaN 2025-06-20 12:17:14.040000+00: 9999-12-31 23:59:59.999999+00: 210.0 2025-06-20 12:17:14.040000+00: 9999-12-31 23:59:59.999999+00:
# Example 2: Build dataset with features 'f_int', 'f_float' from repo 'vfs_v1' and 'sales' data domain.
# Show the latest and history of the data.
>>> import time
>>> from datetime import datetime as dt, date as d
# Retrieve the record where accounts == 'Blue Inc'.
>>> df_test = df[df['accounts'] == 'Blue Inc']
>>> df_test
Feb Jan Mar Apr datetime
accounts
Blue Inc 90.0 50.0 95.0 101.0 04/01/2017
# Writes record stored in a teradataml DataFrame to Teradata Vantage.
>>> df_test.to_sql('sales_test', if_exists='replace')
>>> test_df = DataFrame('sales_test')
>>> test_df
accounts Feb Jan Mar Apr datetime
0 Blue Inc 90.0 50 95 101 17/01/04
>>> # Create a feature process.
>>> fp = FeatureProcess(repo=repo,
... data_domain=data_domain,
... object=test_df,
... entity='accounts',
... features=['Jan', 'Feb'])
>>> # Run the feature process
>>> fp.run()
Process '6cb49b4b-79d4-11f0-8c5e-b0dcef8381ea' started.
Process '6cb49b4b-79d4-11f0-8c5e-b0dcef8381ea' completed.
True
>>> # Running the same process more than once to demonstrate how user can
>>> # retrieve specific version of Features using argument 'as_of'.
>>> # Wait for 20 seconds. Then update the data. Then run again.
>>> time.sleep(20)
>>> execute_sql("update sales_test set Jan = Jan * 10, Feb = Feb * 10")
TeradataCursor uRowsHandle=269 bClosed=False
>>> # Run the feature process again.
>>> fp.run()
Process '6cb49b4b-79d4-11f0-8c5e-b0dcef8381ea' started.
Process '6cb49b4b-79d4-11f0-8c5e-b0dcef8381ea' completed.
True
>>> # Then again wait for 20 seconds. Then update the data. Then run again.
>>> time.sleep(20)
>>> execute_sql("update sales_test set Jan = Jan * 10, Feb = Feb * 10")
TeradataCursor uRowsHandle=397 bClosed=False
>>> # Run the feature process again.
>>> fp.run()
Process '6cb49b4b-79d4-11f0-8c5e-b0dcef8381ea' started.
Process '6cb49b4b-79d4-11f0-8c5e-b0dcef8381ea' completed.
True
>>> dc = DatasetCatalog(repo='vfs_v1', data_domain='sales')
>>> exclude_history = dc.build_time_series(entity='accounts',
... selected_features={'Feb': fp.process_id,
... 'Jan': fp.process_id},
... view_name='exclude_history',
... include_historic_records=False)
>>> exclude_history
accounts Feb Feb_start_time Feb_end_time Jan Jan_start_time Jan_end_time
0 Blue Inc 9000.0 2025-08-15 13:24:58.140000+00: 9999-12-31 23:59:59.999999+00: 5000 2025-08-15 13:24:58.140000+00: 9999-12-31 23:59:59.999999+00:
>>> dc = DatasetCatalog(repo='vfs_v1', data_domain='sales')
>>> include_history = dc.build_time_series(entity='accounts',
... selected_features={'Feb': fp.process_id,
... 'Jan': fp.process_id},
... view_name='include_history',
... include_historic_records=True)
>>> include_history
accounts Feb Feb_start_time Feb_end_time Jan Jan_start_time Jan_end_time
0 Blue Inc 9000.0 2025-08-15 13:24:58.140000+00: 9999-12-31 23:59:59.999999+00: 5000 2025-08-15 13:24:58.140000+00: 9999-12-31 23:59:59.999999+00:
1 Blue Inc 90.0 2025-08-15 13:23:41.780000+00: 2025-08-15 13:24:31.320000+00: 50 2025-08-15 13:23:41.780000+00: 2025-08-15 13:24:31.320000+00:
2 Blue Inc 90.0 2025-08-15 13:23:41.780000+00: 2025-08-15 13:24:31.320000+00: 5000 2025-08-15 13:24:58.140000+00: 9999-12-31 23:59:59.999999+00:
3 Blue Inc 900.0 2025-08-15 13:24:31.320000+00: 2025-08-15 13:24:58.140000+00: 500 2025-08-15 13:24:31.320000+00: 2025-08-15 13:24:58.140000+00:
4 Blue Inc 900.0 2025-08-15 13:24:31.320000+00: 2025-08-15 13:24:58.140000+00: 5000 2025-08-15 13:24:58.140000+00: 9999-12-31 23:59:59.999999+00:
5 Blue Inc 900.0 2025-08-15 13:24:31.320000+00: 2025-08-15 13:24:58.140000+00: 50 2025-08-15 13:23:41.780000+00: 2025-08-15 13:24:31.320000+00:
6 Blue Inc 90.0 2025-08-15 13:23:41.780000+00: 2025-08-15 13:24:31.320000+00: 500 2025-08-15 13:24:31.320000+00: 2025-08-15 13:24:58.140000+00:
7 Blue Inc 9000.0 2025-08-15 13:24:58.140000+00: 9999-12-31 23:59:59.999999+00: 50 2025-08-15 13:23:41.780000+00: 2025-08-15 13:24:31.320000+00:
8 Blue Inc 9000.0 2025-08-15 13:24:58.140000+00: 9999-12-31 23:59:59.999999+00: 500 2025-08-15 13:24:31.320000+00: 2025-08-15 13:24:58.140000+00: