Teradata Package for Python Function Reference on VantageCloud Lake - build_time_series - Teradata Package for Python - Look here for syntax, methods and examples for the functions included in the Teradata Package for Python.

Teradata® Package for Python Function Reference on VantageCloud Lake

Deployment
VantageCloud
Edition
Lake
Product
Teradata Package for Python
Release Number
20.00.00.08
Published
November 2025
ft:locale
en-US
ft:lastEdition
2025-12-05
dita:id
TeradataPython_FxRef_Lake_2000
Product Category
Teradata Vantage
teradataml.store.feature_store.models.DatasetCatalog.build_time_series = build_time_series(self, entity, selected_features, view_name, description=None, include_historic_records=False)
DESCRIPTION:
    Builds the dataset with start time and end time for feature values available in
    feature catalog. Once dataset is created, user can create a teradataml DataFrame
    on the dataset.
 
PARAMETERS:
    entity:
        Required Argument.
        Specifies the name of the Entity of Object of Entity
        to be included in the dataset.
        Types: str or Entity.
 
    selected_features:
        Required Argument.
        Specifies the names of Features and the corresponding feature version
        to be included in the dataset.
        Notes:
             * Key is the name of the feature and value is the version of the
               feature.
             * Look at FeatureCatalog.list_feature_versions() to get the list of
               features and their versions.
        Types: dict
 
    view_name:
        Required Argument.
        Specifies the name of the view to be named for dataset.
        Types: str.
 
    description:
        Optional Argument.
        Specifies the description for the dataset.
        Types: str.
 
    include_historic_records:
        Optional Argument.
        Specifies whether to include historic data in the dataset.
        Default Value: False.
        Types: bool.
 
RETURNS:
    Teradataml DataFrame.
 
RAISES:
    TeradataMlException
 
EXAMPLES:
    # Ingest sales data to feature catalog configured for repo 'vfs_v1'.
    >>> from teradataml import load_example_data, FeatureProcess
    >>> load_example_data('dataframe', 'sales')
    >>> df = DataFrame("sales")
    >>> df
                  Feb    Jan    Mar    Apr    datetime
    accounts
    Red Inc     200.0  150.0  140.0    NaN  04/01/2017
    Blue Inc     90.0   50.0   95.0  101.0  04/01/2017
    Alpha Co    210.0  200.0  215.0  250.0  04/01/2017
    Orange Inc  210.0    NaN    NaN  250.0  04/01/2017
    Yellow Inc   90.0    NaN    NaN    NaN  04/01/2017
    Jones LLC   200.0  150.0  140.0  180.0  04/01/2017
 
    # Create a FeatureStore.
    >>> from teradataml import FeatureStore
    >>> fs = FeatureStore(repo='vfs_v1', data_domain='sales')
    Repo vfs does not exist. Run FeatureStore.setup() to create the repo and setup FeatureStore.
    >>> fs.setup()
    True
 
    # Initiate FeatureProcess to ingest features.
    >>> fp = FeatureProcess(repo='vfs_v1', data_domain='sales', object=df, entity='accounts', features=['Jan', 'Feb', 'Mar', 'Apr'])
    # Run the feature process.
    >>> fp.run()
    Process 'a9f29a4e-3f75-11f0-b43b-f020ff57c62c' started.
    Process 'a9f29a4e-3f75-11f0-b43b-f020ff57c62c' completed.
 
    # Example 1: Build dataset with features 'Jan', 'Feb' from repo 'vfs_v1' and sales data domain.
    #            Name the dataset as 'ds_jan_feb'.
    >>> from teradataml import DatasetCatalog
    >>> dc = DatasetCatalog(repo='vfs_v1', data_domain='sales')
    >>> dataset = dc.build_time_series(entity='accounts',
    ...                                selected_features = {
    ...                                    'Jan': 'a9f29a4e-3f75-11f0-b43b-f020ff57c62c',
    ...                                    'Feb': 'a9f29a4e-3f75-11f0-b43b-f020ff57c62c'},
    ...                                view_name='ds_jan_feb',
    ...                                description='Dataset with Jan and Feb features')
    >>> dataset
         accounts    Jan                  Jan_start_time                    Jan_end_time    Feb                  Feb_start_time                    Feb_end_time
    0    Blue Inc   50.0  2025-06-20 12:17:14.040000+00:  9999-12-31 23:59:59.999999+00:   90.0  2025-06-20 12:17:14.040000+00:  9999-12-31 23:59:59.999999+00:
    1     Red Inc  150.0  2025-06-20 12:17:14.040000+00:  9999-12-31 23:59:59.999999+00:  200.0  2025-06-20 12:17:14.040000+00:  9999-12-31 23:59:59.999999+00:
    2  Yellow Inc    NaN  2025-06-20 12:17:14.040000+00:  9999-12-31 23:59:59.999999+00:   90.0  2025-06-20 12:17:14.040000+00:  9999-12-31 23:59:59.999999+00:
    3    Alpha Co  200.0  2025-06-20 12:17:14.040000+00:  9999-12-31 23:59:59.999999+00:  210.0  2025-06-20 12:17:14.040000+00:  9999-12-31 23:59:59.999999+00:
    4   Jones LLC  150.0  2025-06-20 12:17:14.040000+00:  9999-12-31 23:59:59.999999+00:  200.0  2025-06-20 12:17:14.040000+00:  9999-12-31 23:59:59.999999+00:
    5  Orange Inc    NaN  2025-06-20 12:17:14.040000+00:  9999-12-31 23:59:59.999999+00:  210.0  2025-06-20 12:17:14.040000+00:  9999-12-31 23:59:59.999999+00:
 
    # Example 2: Build dataset with features 'f_int', 'f_float' from repo 'vfs_v1' and 'sales' data domain.
    #            Show the latest and history of the data.
    >>> import time
    >>> from datetime import datetime as dt, date as d
 
    # Retrieve the record where accounts == 'Blue Inc'.
    >>> df_test = df[df['accounts'] == 'Blue Inc']
    >>> df_test
                  Feb    Jan    Mar    Apr    datetime
    accounts
    Blue Inc     90.0   50.0   95.0  101.0  04/01/2017
 
    # Writes record stored in a teradataml DataFrame to Teradata Vantage.
    >>> df_test.to_sql('sales_test', if_exists='replace')
    >>> test_df = DataFrame('sales_test')
    >>> test_df
       accounts   Feb  Jan  Mar  Apr  datetime
    0  Blue Inc  90.0   50   95  101  17/01/04
 
    >>> # Create a feature process.
    >>> fp = FeatureProcess(repo=repo,
    ...                     data_domain=data_domain,
    ...                     object=test_df,
    ...                     entity='accounts',
    ...                     features=['Jan', 'Feb'])
 
    >>> # Run the feature process
    >>> fp.run()
    Process '6cb49b4b-79d4-11f0-8c5e-b0dcef8381ea' started.
    Process '6cb49b4b-79d4-11f0-8c5e-b0dcef8381ea' completed.
    True
 
    >>> # Running the same process more than once to demonstrate how user can
    >>> # retrieve specific version of Features using argument 'as_of'.
    >>> # Wait for 20 seconds. Then update the data. Then run again.
    >>> time.sleep(20)
    >>> execute_sql("update sales_test set Jan = Jan * 10, Feb = Feb * 10")
    TeradataCursor uRowsHandle=269 bClosed=False
 
    >>> # Run the feature process again.
    >>> fp.run()
    Process '6cb49b4b-79d4-11f0-8c5e-b0dcef8381ea' started.
    Process '6cb49b4b-79d4-11f0-8c5e-b0dcef8381ea' completed.
    True
 
    >>> # Then again wait for 20 seconds. Then update the data. Then run again.
    >>> time.sleep(20)
    >>> execute_sql("update sales_test set Jan = Jan * 10, Feb = Feb * 10")
    TeradataCursor uRowsHandle=397 bClosed=False
 
    >>> # Run the feature process again.
    >>> fp.run()
    Process '6cb49b4b-79d4-11f0-8c5e-b0dcef8381ea' started.
    Process '6cb49b4b-79d4-11f0-8c5e-b0dcef8381ea' completed.
    True
 
    >>> dc = DatasetCatalog(repo='vfs_v1', data_domain='sales')
    >>> exclude_history = dc.build_time_series(entity='accounts',
    ...                                        selected_features={'Feb': fp.process_id,
    ...                                                           'Jan': fp.process_id},
    ...                                        view_name='exclude_history',
    ...                                        include_historic_records=False)
    >>> exclude_history
       accounts     Feb                  Feb_start_time                    Feb_end_time   Jan                  Jan_start_time                    Jan_end_time
    0  Blue Inc  9000.0  2025-08-15 13:24:58.140000+00:  9999-12-31 23:59:59.999999+00:  5000  2025-08-15 13:24:58.140000+00:  9999-12-31 23:59:59.999999+00:
 
    >>> dc = DatasetCatalog(repo='vfs_v1', data_domain='sales')
    >>> include_history = dc.build_time_series(entity='accounts',
    ...                                        selected_features={'Feb': fp.process_id,
    ...                                                           'Jan': fp.process_id},
    ...                                        view_name='include_history',
    ...                                        include_historic_records=True)
    >>> include_history
       accounts     Feb                  Feb_start_time                    Feb_end_time   Jan                  Jan_start_time                    Jan_end_time
    0  Blue Inc  9000.0  2025-08-15 13:24:58.140000+00:  9999-12-31 23:59:59.999999+00:  5000  2025-08-15 13:24:58.140000+00:  9999-12-31 23:59:59.999999+00:
    1  Blue Inc    90.0  2025-08-15 13:23:41.780000+00:  2025-08-15 13:24:31.320000+00:    50  2025-08-15 13:23:41.780000+00:  2025-08-15 13:24:31.320000+00:
    2  Blue Inc    90.0  2025-08-15 13:23:41.780000+00:  2025-08-15 13:24:31.320000+00:  5000  2025-08-15 13:24:58.140000+00:  9999-12-31 23:59:59.999999+00:
    3  Blue Inc   900.0  2025-08-15 13:24:31.320000+00:  2025-08-15 13:24:58.140000+00:   500  2025-08-15 13:24:31.320000+00:  2025-08-15 13:24:58.140000+00:
    4  Blue Inc   900.0  2025-08-15 13:24:31.320000+00:  2025-08-15 13:24:58.140000+00:  5000  2025-08-15 13:24:58.140000+00:  9999-12-31 23:59:59.999999+00:
    5  Blue Inc   900.0  2025-08-15 13:24:31.320000+00:  2025-08-15 13:24:58.140000+00:    50  2025-08-15 13:23:41.780000+00:  2025-08-15 13:24:31.320000+00:
    6  Blue Inc    90.0  2025-08-15 13:23:41.780000+00:  2025-08-15 13:24:31.320000+00:   500  2025-08-15 13:24:31.320000+00:  2025-08-15 13:24:58.140000+00:
    7  Blue Inc  9000.0  2025-08-15 13:24:58.140000+00:  9999-12-31 23:59:59.999999+00:    50  2025-08-15 13:23:41.780000+00:  2025-08-15 13:24:31.320000+00:
    8  Blue Inc  9000.0  2025-08-15 13:24:58.140000+00:  9999-12-31 23:59:59.999999+00:   500  2025-08-15 13:24:31.320000+00:  2025-08-15 13:24:58.140000+00: