Page tree
Skip to end of metadata
Go to start of metadata

This article describes how to import data via SQL query using the Database extractor ETL.

This topic contains the following sections:

Integration steps

To integrate TrueSight Capacity Optimization with the CSV database extractor, perform the following task:

  1. Navigate to Administration >  ETL & SYSTEM TASKS > ETL tasks.
  2. In the ETL tasks page, click Add > Add ETL under the Last run tab.
  3. In the Add ETL page, set values for the following properties under each expandable tab.

    Note

    Basic properties are displayed by default in the Add ETL page. These are the most common properties that you can set for an ETL, and it is acceptable to leave the default selections for each as is.

    Basic properties

    Property Description
    Run configuration
    ETL module Select Generic - Database extractor.
    ETL task name Default name is already filled out for you.
    Run configuration name Default name is already filled out for you.
    Deploy status Select Production.
    Description (Optional) Enter a brief description.
    Log level Select how detailed you want the log to be:
    • 1 - Light: Add bare minimum activity logs to the log file.
    • 5 - Medium: Add medium-detailed activity logs to the log file.
    • 10 - Verbose: Add detailed activity logs to the log file.
    Execute in simulation mode Select Yes.
    When set to Yes, the ETL will not store actual data into the data warehouse. This option is useful while testing a new ETL task.
    Module selection

    Ensure that the Based on datasource option is selected.

    Note

    If you select Based on Open ETL template, TrueSight Capacity Optimization is integrated with a Generic extractor based on the selected Open ETL template. For more information, see  Generic ETL based on a template

    Module description A link that points you to technical documentation for this ETL.
    Datasets
    1. Click Edit.
    2. Select one (click) or more (shift+click) datasets that you want to include from Available datasets and click >> to move them to Selected datasets.
    3. Click Apply.
    Entitycatalog
    Sharing status Select any one:
    • Shared entity catalog: Select this option if, for the same entities, data is coming from multiple sources. For example, BPA ETL.
      • Sharing with Entity Catalog : Select an entity catalog from the drop-down list.
    • Private entity catalog: Select this option if, for the same entity, data is coming from a single source.
    Object relationships
    After import

    Specify the domain where you want to add the entities created by the ETL. You can select an existing domain or create a new one.

    Select any one of the following options:

    • leave all new entities in 'Newly Discovered'.
    • move all new entities in a new Domain.
      • New domain: Create a new domain. Specify the following properties under New domain:
        • Parent: Select a parent domain for your new domain from the domain selector control.
        • Name: Specify a name for your new domain.
    • move all new entities in an existing Domain 
      • Domain: Select an existing domain from the domain selector control.
    SQL query
    SQL query This tab is only available when a custom database query has to be built in order to get the data from a generic source. The form allows you to configure the following properties. A text area in which you can type the SQL extraction query.
    Lastcounter column Specify which is the lastcounter column that will be used to identify new data; generally this is set as the timestamp (TS).
    SQL parameters number You can add up to 10 custom parameters to the query; select how many you need from the drop-down menu.
    Connection parameters
    Database type

    Select the database type:

    • Oracle
    • Microsoft SQL Server/Sybase
    • Other database .
    Database user

    Type in the database user name.

    Database password Enter the database password.
    Database TNS name Type in the Database TNS name.
    Data source name (Applies to  Microsoft SQL Server/Sybase ) Type in the data source name.
    Database name  (Applies to  Microsoft SQL Server/Sybase ) Type in the database name.
    Connection URL (Applies to  Other database ) Enter the connection URL.
    ETL task properties
    Task group Select a task group to classify this ETL into.
    Running on scheduler Select the scheduler you want to run the ETL on.
    Maximum execution time before warning The number of hours, minutes or days to to execute the ETL for before generating warnings, if any.
    Frequency Select the frequency of ETL execution. Available options are:
    • Predefined: Select a Predefined frequency from Each Day, Each Week or Each Month.
    • Custom: Enter a Custom frequency (time interval) as the number of minutes, hours, days or weeks to run the ETL in.
    Start timestamp: hour\minute (Applies to Predefined frequency) The HH:MM start timestamp to add to the ETL execution running on a Predefined frequency.
    Custom start timestamp Select a YYYY-MM-DD HH:MM timestamp to add to the ETL execution running on a Custom frequency.

    Note

    To view or configure Advanced properties, click Advanced. You do not need to set or modify these properties unless you want to change the way the ETL works. These properties are for advanced users and scenarios only.

    Advanced properties

    Property Description
    Run configuration
    Default locale Enter the  default   locale  information
    Default last counter Set the default last counter in dd/mm/yyy hh:mm:ss format.
    Collection level 
    Metric profile selection

    Select any one:

    • Use Global metric profile: Select this to use an out-of-the-box global profile, that is available on the Adding and modifying metric profiles page. By default, all ETL modules use this profile.
    • Select a custom metric profile: Any metric profiles you add in the Add metric profile page (Administration  >  DATAWAREHOUSE  >  Metric profiles).

    For more information, see Adding and modifying metric profiles.

    Levels up to

    The metric level defines the amount of metric imported into the data warehouse. If you increase the level, additional load is added to the data warehouse while decreasing the metric level reduces the number of imported metrics.

    Choose the metric level to apply on selected metrics:

    • [1] Essential
    • [2] Basic
    • [3] Standard
    • [4] Extended

    For more information, see Aging Class mapping.

    Metric

    Metric list  <for selected dataset>

    Note: This property is activated only after you select Datasets under theRun configuration tab.

    1. Click Edit.
    2. Select one (click) or more (shift+click) metrics that you want to include from Available items and click  >>  to move them to Selected items.
    3. Click Apply.
    Format customization
    Timestamp format Specify a format that should be used by the ETL if the user tables or CSV files use an unsupported format. <YYYY-MM-DD HH:MM:SS> is the supported format. Depending on the type of ETL used, you might need to specify a custom format.
    Percentage format
    • 0 to 1:
    • 0 to 100:
    Additional properties
    List of properties
    1. Click Add.
    2. Add an additional property in the etl.additional.prop.n box.
    3. Click Apply.
      Repeat this task to add more properties.

    To initiate a new database connection when each query is run, add the following property:

    extract.sql.disconnect=true

    Loader configuration
    Empty dataset behavior Choose one of the following actions if the loader encounters an empty dataset:
    • Abort: Abort the loader.
    • Ignore: Ignore the empty dataset and continue parsing.
    ETL log file name Name of the file that contains the ETL execution log; the default value is: %BASE/log/%AYEAR%AMONTH%ADAY%AHOUR%MINUTE%TASKID
    Maximum number of rows for CSV output A number which limits the size of the output files.
    CSV loader output file name Name of the file generated by the CSV loader; the default value is: %BASE/output/%DSNAME%AYEAR%AMONTH%ADAY%AHOUR%ZPROG%DSID%TASKID.
    BCO loader output file name Name of the file generated by the TrueSight Capacity Optimization loader; the default value is: %BASE/output/%DSNAME%AYEAR%AMONTH%ADAY%AHOUR%ZPROG%DSID%TASKID.
    Detail mode Select the level of detail:
    • Standard: Data will be stored on the database in different tables at the following time granularities: Detail (configurable, by default: 5 minutes), Hourly, Daily, Monthly.
    • Raw also: Data will be stored on the database in different tables at the following time granularities: Raw (as available from the original data source), Detail (configurable, by default: 5 minutes), Hourly, Daily, Monthly.
    • Raw only: Data will be stored on the database in a table only at Raw granularity (as available from the original data source).

    For more information on granularities, see Accessing data using public views and Sizing and scalability considerations.

    Reduce priority
    • Normal:
    • High:
    Remove domain suffix from datasource name (Only for systems) If set to True, the domain name is removed from the data source name. For example, server.domain.com will be saved as server.
    Leave domain suffix to system name (Only for systems) If set to True, the domain name is maintained in the system name. For example: server.domain.com will be saved as such.
    Update grouping object definition If set to True, the ETL will be allowed to update the grouping object definition for a metric loaded by an ETL.
    Skip entity creation (Only for ETL tasks sharing lookup with other tasks) If set to True, this ETL does not create an entity, and discards data from its data source for entities not found in TrueSight Capacity Optimization. It uses one of the other ETLs that share lookup to create the new entity.
    Scheduling options
    Hour mask Specify a value to execute the task only during particular hours within the day. For example, 0 – 23 or 1,3,5 – 12.
    Day of week mask Select the days so that the task can be executed only during the selected days of the week. To avoid setting this filter, do not select any option for this field.
    Day of month mask Specify a value to execute the task only during particular days within a month. For example, 5, 9, 18, 27 – 31.
    Apply mask validation By default this property is set to True. Set it to False if you want to disable the preceding Scheduling options that you specified. Setting it to False is useful if you want to temporarily turn off the mask validation without removing any values.
    Execute after time Specify a value in the hours:minutes format (for example, 05:00 or 16:00) to wait before the task must be executed. This means that once the task is scheduled, the task execution starts only after the specified time passes.
    Enqueueable Select one of the following options:
    • False (Default): While a particular task is already running, if the next execution command arises – it is ignored.
    • True: While a particular task is already running, if the next execution command arises – it is placed in a queue and is executed as soon as the current execution ends.
  4. Click Save.
    You return to the Last run tab under the ETL tasks page.
  5. Validate the results in simulation mode: In the ETL tasks table under ETL tasks > Last run, locate your ETL (ETL task name), click  to run the ETL.
    After you run the ETL, the Last exit column in the ETL tasks table will display one of the following values:
    • OK: The ETL executed without any error in simulation mode.
    • WARNING: The ETL execution returned some warnings in simulation mode. Check the ETL log.
    • ERROR: The ETL execution returned errors and was unsuccessful. Edit the active Run configuration and try again.
  6. Switch the ETL to production mode: To do this, perform the following task:
    1. In the ETL tasks table under ETL tasks > Last run, click the ETL under the Name column.
    2. In the Run configurations table in the ETL details page, click to edit the active run configuration.
    3. In the Edit runconfiguration page, navigate to the Run configuration expandable tab and set Execute in simulation mode to No.
    4. Click Save.
  7. Locate the ETL in the ETL tasks table and click  to Run it, or schedule an ETL run.
    After you run the ETL, or schedule the ETL for a run, it will extract the data form the source and transfer it to the TrueSight Capacity Optimization database.

Setup database connectivity

To setup database connectivity, refer to Configuring database connection using Perl or Java ETLs

Preparing the extraction query

The query should be formatted using the format supported by the database extractor.
Here is a list of fields that the query must provide. Some fields changes based on the selected dataset.

Column

Description

TS

Sample timestamp - recommended format is YYYY-MM-DD HH24:MI:SS (for example, 2012-03-28 12:30:00).

DURATION

Duration of the sample (in seconds).

OBJNM

TrueSight Capacity Optimization metric to import (for example, CPU_UTIL).

VALUE

Value of the sample.

SUBOBJNM

Metric subobject name (if required by the metric, not mandatory).

Depending on the selected dataset, the entity identifier fields are:

Dataset

Column

Description

SYSDAT

DS_SYSNM

System lookup identifier

SYSDAT

SYSNM

System name

Dataset

Column

Description

WKLDAT

DS_WKLDNM

Business driver lookup identifier

WKLDAT

WKLDNM

Business driver name

Dataset

Column

Description

APPDAT

DS_APPNM

Domain lookup identifier

APPDAT

APPNM

Domain name

Here is a list of additional fields supported in the extraction query:

Column

Description

MINVALUE

Minimum value of the sample.

MAXVALUE

Maximum value of the sample.

DS_SUBOBJNM

Datasource subobject identifier.

WEIGHT

Weight of the sample.

STRONGLOOKUPFIELDS

List of strong lookup fields.

WEAKLOOKUPFIELDS

List of weak lookup fields.

To import the minimum (maximum) value, you can add a new column called MINVALUE (MAXVALUE). The already existing value column must be populated, e.g. with the average value.

The extraction query is as follows (after selecting SYSDAT):

SELECT myTsCol AS TS, 3600 AS DURATION, 'sys1' AS ds_sysnm,
       'sys1' AS sysnm, 'CPU_UTIL' AS objnm, 'GLOBAL' AS subobjnm, myValue AS VALUE,
       myMinValue AS MINVALUE, myMaxVAlue AS maxvalue
  FROM TABLE

Query lastcounter is set as default on TS column.

You can use the :PARL placeholder parameter one or multiple times in a query; the parameter is replaced by the last saved lastcounter value.

#Example 1
SELECT TS,DURATION,DS_SYSNM,SYSNM,OBJNM,VALUE FROM TABLE WHERE TS > :PARL

#Example 2
SELECT TS,DURATION,DS_SYSNM,SYSNM,OBJNM,VALUE FROM TABLE WHERE TS > :PARL OR ACTIVITYDATE > :PARL

Related topics

Using ETL datasets

Generic - CSV file parser

Developing custom ETLs

Dataset reference for ETL tasks

Horizontal and Vertical datasets

Viewing datasets and metrics by dataset and ETL module

Determining how to aggregate data to a given resolution using SQL 

1 Comment

  1.