Teradata Package for Python Function Reference on VantageCloud Lake - archive_datasets - Teradata Package for Python - Look here for syntax, methods and examples for the functions included in the Teradata Package for Python.

Teradata® Package for Python Function Reference on VantageCloud Lake

Deployment
VantageCloud
Edition
Lake
Product
Teradata Package for Python
Release Number
20.00.00.08
Published
November 2025
ft:locale
en-US
ft:lastEdition
2025-12-05
dita:id
TeradataPython_FxRef_Lake_2000
Product Category
Teradata Vantage
teradataml.store.feature_store.models.DatasetCatalog.archive_datasets = archive_datasets(self, id)
DESCRIPTION:
    Archives the dataset from the dataset catalog.
 
PARAMETERS:
    id:
        Required Argument.
        Specifies id(s) of the dataset(s) to be archived from dataset catalog.
        Note:
            * Duplicate ids are processed only once.
        Types: str or list of str.
 
RETURNS:
    bool
 
RAISES:
    TeradataMlException
 
EXAMPLES:
    # Upload features first to create a dataset.
    >>> from teradataml import load_example_data, FeatureProcess
    >>> load_example_data('dataframe', 'sales')
    >>> df = DataFrame("sales")
 
    # Create a FeatureStore.
    >>> fs = FeatureStore(repo='vfs_v1', data_domain='sales')
    Repo vfs_test does not exist. Run FeatureStore.setup() to create the repo and setup FeatureStore.
    >>> fs.setup()
    True
 
    # Run the feature process to ingest features.
    >>> fp = FeatureProcess(repo='vfs_v1', data_domain='sales', object=df, entity='accounts',
    ...                     features=['Jan', 'Feb', 'Mar', 'Apr'])
    >>> fp.run()
    Process '3acf5632-5d73-11f0-99c5-a30631e77953' started.
    Process '3acf5632-5d73-11f0-99c5-a30631e77953' completed.
    True
 
    # Build dataset.
    >>> from teradataml import DatasetCatalog
    >>> dc = DatasetCatalog(repo='vfs_v1', data_domain='sales')
    >>> dataset = dc.build_dataset(entity='accounts',
    ...                            selected_features = {
    ...                                        'Jan': fp.process_id,
    ...                                        'Feb': fp.process_id},
    ...                            view_name='ds_jan_feb',
    ...                            description='Dataset with Jan and Feb features')
 
    # List datasets.
    >>> dc.list_datasets()
                                         data_domain          name entity_name                            description                      valid_start                        valid_end
    id
    851a651a-68a3-4eb6-b606-df2617089068       sales  ds_jan_feb_1    accounts      Dataset with Jan and Feb features    2025-07-10 09:50:25.527852+00:  9999-12-31 23:59:59.999999+00:
 
    # Example: Archive dataset.
    >>> dc.archive_datasets('851a651a-68a3-4eb6-b606-df2617089068')
    Dataset id(s) '851a651a-68a3-4eb6-b606-df2617089068' is/are archived from the dataset catalog.
    True
 
    # List datasets.
    >>> dc.list_datasets()
                                         data_domain          name entity_name                            description                      valid_start                       valid_end
    id
    851a651a-68a3-4eb6-b606-df2617089068       sales  ds_jan_feb_1    accounts      Dataset with Jan and Feb features   2025-07-10 09:50:25.527852+00:  2025-07-10 09:55:02.830000+00: