This documentation is for an older version of HVR.
Actions
Actions in HVR allows you to define the behavior of replication. Every action has a collection of parameters that provide a finer control of the behavior. To initiate data replication, at least two actions Capture and Integrate must be defined on source and target locations accordingly.
AdaptDDL
| Parameter | Argument | Description |
|---|---|---|
| /AddTablePattern | patt | Add new tables to channel if they match. |
| /IgnoreTablePattern | patt | Ignore new tables which match pattern. |
| /CaptureSchema | db_schema | Database schema for matching tables. |
| /IntegrateSchema | db_schema | Generate /Schema for target location(s). |
| /RefreshOptions | refr_opts | Configure options for adapt's refresh of target. |
| /OnDropTable | pol | Behavior when source table dropped. Default: from channel only. |
| /KeepExistingStructure | Preserve old columns in target, and do not reduce data types sizes. | |
| /KeepOldRows | Preserve old rows in target during recreate. |
AgentPlugin
| Parameter | Argument | Description |
|---|---|---|
| /Command | path | Call OS command during replication jobs. |
| /DbProc | dbproc | Call database procedure dbproc during replication jobs. |
| /UserArgument | str | Pass argument str to each agent execution. |
| /ExecOnHub | Execute agent on hub instead of location's machine. | |
| /Order | int | Specify order of agent execution. |
| /Path | dir | Search directory dir for agent. |
| /Context | context | Action only applies if Refresh/Compare context matches. |
Capture
| Parameter | Argument | Description |
|---|---|---|
| /IgnoreSessionName | sess_name | Capture changes directly from DBMS logging system. |
| /Coalesce | Coalesce consecutive changes on the same row into a single change. | |
| /NoBeforeUpdate | Only capture the new values for updated rows. | |
| /NoTruncate | Do not capture truncate table statements. | |
| /SupplementalLogging | action | Mechanism used to enable supplemental logging for SQL Server tables. |
| /LogReadMethod | method | Method of reading SQL Server's transaction log. |
| /LogTruncate | action | Specify who advances transaction log truncation point. |
| /AugmentIncomplete | col_type | Capture job must select for column values. Can be NONE, LOB or ALL. |
| /ArchiveLogPath | dir | Read archives from an alternative directory. |
| /ArchiveLogFormat | format | Format of base filename of archive files in directory /ArchiveLogPath. |
| /ArchiveLogOnly | Capture data from archives only. Do not read from online redos. | |
| /XLogDirectory | dir | Directory containing current PostgreSQL xlog files. |
| /LogJournal | schema.journal | Specify DB2-for-i journal. |
| /LogJournalSysSeq | Capture from journal using *SYSSEQ. | |
| /CheckpointFrequency | secs | Checkpointing frequency in seconds for long running transactions, so the capture job can recover quickly when it restarts. |
| /CheckpointStorage | STOR | Storage location of capture checkpoint files for quick capture recovery. |
| /CheckpointRetention | period | Retains capture checkpoint files up to the specified period. |
| /TriggerBased | Capture changes through generated DMBS triggers. | |
| /QuickToggle | Avoid shared lock on toggle table. | |
| /ToggleFrequency | secs | Sleep between toggles instead of waiting for database alert (in seconds). |
| /KeyOnlyCaptureTable | Only keep keys in capture table; outer join others later. | |
| /IgnoreCondition | sql_expr | Ignore changes that satisfy expression. |
| /IgnoreUpdateCondition | sql_expr | Ignore update changes that satisfy expression. |
| /HashBuckets | int | Hash structure to improve parallelism of captured tables. |
| /HashKey | col_list | Hash capture table on specific key columns. |
| /DeleteAfterCapture | Delete file after capture, instead of capturing recently changed files. | |
| /Pattern | pattern | Only capture files whose names match pattern. |
| /IgnorePattern | pattern | Ignore files whose names match pattern. |
| /IgnoreUnterminated | pattern | Ignore files whose last line does not match pattern. |
| /IgnoreSizeChanges | Changes in file size during capture is not considered an error. | |
| /AccessDelay | secs | Delay read for secs seconds to ensure writing is complete. |
| /UseDirectoryTime | Check timestamp of parent dir, as Windows move doesn't change mod-time. |
CollisionDetect
| Parameter | Argument | Description |
|---|---|---|
| /TreatCollisionAsError | Do not resolve collisions automatically. | |
| /TimestampColumn | col_name | Exploit timestamp column col_name for collision detection. |
| /AutoHistoryPurge | Delete history table row when no longer needed for collision detection. | |
| /DetectDuringRefresh | colname | During row–wise refresh, discard updates if target timestamp is newer. |
| /Context | context | Action only applies if Refresh/Compare context matches. |
ColumnProperties
| Parameter | Argument | Description |
|---|---|---|
| /Name | col_name | Name of column in hvr_column catalog. |
| /DatatypeMatch | data_type | Data type used for matching instead of /Name. |
| /BaseName | col_name | Database column name differs from hvr_column catalog. |
| /Extra | Column exists in base table but not in hvr_column catalog. | |
| /Absent | Column does not exist in base table. | |
| /CaptureExpression | sql_expr | SQL expression for column value when capturing or reading. |
| /CaptureExpressionType | Type of mechanism used by HVR capture, refresh and compare job to evaluate value in parameter /CaptureExpression. | |
| /IntegrateExpression | sql_expr | SQL expression for column value when integrating. |
| /ExpressionScope | expr_scope | Operation scope for expressions, e.g. INSERT, DELETE or UPDATE_AFTER. |
| /CaptureFromRowId | Capture values from table's DBMS row-id. | |
| /TrimDatatype | int | Reduce width of data type when selecting or capturing changes. |
| /Key | Add column to table's replication key. | |
| /SurrogateKey | Use column instead of the regular key during replication. | |
| /DistributionKey | Distribution key column. | |
| /PartitionKeyOrder | int | Define the column as a partition key and set partitioning order for the column. |
| /SoftDelete | Convert deletes to update of this column to 1. Value 0 means not deleted. | |
| /TimeKey | Convert all changes to inserts, using this column for time dimension. | |
| /IgnoreDuringCompare | Ignore values in column during compare and refresh. | |
| /Datatype | data_type | Data type in database if it differs from hvr_column catalog. |
| /Length | int | String length in db if it differs from length in catalog. |
| /Precision | int | Precision in db if it differs from precision in catalog. |
| /Scale | int | Integer scale in db if it differs from scale in catalog. |
| /Nullable | Nullability in db if it differs from nullability in catalog. | |
| /Identity | Column has SQL Server identity attribute. | |
| /Context | ctx | Ignore action unless refresh/compare context ctx is enabled. |
DbObjectGeneration
| Parameter | Argument | Description |
|---|---|---|
| /NoCaptureInsertTrigger | Inhibit generation of capture insert trigger. | |
| /NoCaptureUpdateTrigger | Inhibit generation of capture update trigger. | |
| /NoCaptureDeleteTrigger | Inhibit generation of capture delete trigger. | |
| /NoCaptureDbProc | Inhibit generation of capture database procedures. | |
| /NoCaptureTable | Inhibit generation of capture tables. | |
| /NoIntegrateDbProc | Inhibit generation of integrate database procedures. | |
| /IncludeSqlFile | file | Search directory for include SQL file. |
| /IncludeSqlDirectory | dir | Search directory for include SQL file. |
| /BurstTableStorage | Storage for integrate burst table creation statement. | |
| /RefreshTableStorage | Storage for base table creation statement during refresh. | |
| /CaptureTableCreateClause | sql_expr | Clause for trigger-based capture table creation statement. |
| /StateTableCreateClause | sql_expr | Clause for state table creation statement. |
| /BurstTableCreateClause | sql_expr | Clause for integrate burst table creation statement. |
| /FailTableCreateClause | sql_expr | Clause for fail table creation statement. |
| /HistoryTableCreateClause | sql_expr | Clause for history table creation statement. |
| /RefreshTableCreateClause | sql_expr | Clause for base table creation statement during refresh. |
| /RefreshTableGrant | Executes a grant statement on the base table created during HVR Refresh. |
DbSequence
| Parameter | Argument | Description |
|---|---|---|
| /CaptureOnly | Only capture DB sequences, do not integrate them. | |
| /IntegrateOnly | Only integrate DB sequences, do not capture them. | |
| /Name | seq_name | Name of database sequence in HVR catalogs. |
| /Schema | db_schema | Schema which owns DB sequence. |
| /BaseName | seq_name | Name of sequence in DB if it differs from name in HVR. |
Environment
| Parameter | Argument | Description |
|---|---|---|
| /Name | name | Name of environment variable. |
| /Value | value | Value of environment variable. |
FileFormat
| Parameter | Argument | Description |
|---|---|---|
| /Xml | Transform rows form/into xml-files. | |
| /Csv | Transforms rows from/into csv files. | |
| /Avro | Transforms rows into Apache AVRO format. Integrate only. | |
| /JSON | Transforms rows into JSON format. The content of the file depends on the value for parameter /JsonMode. This parameter only has an effect on the integrate location. | |
| /Parquet | Read and write files as Parquet format. | |
| /Compact | Write compact XML tags like <r>&<c> instead of <row>&<column>. | |
| /Compress | algorithm | Compress/uncompress while writing/reading. algorithm is GZIP or LZ4. |
| /Encoding | encoding | Encoding of file. |
| /HeaderLine | First line of file contains column names. | |
| /FieldSeparator | str_esc | Field separator. Defaults to comma (,). Examples:, \\x1f or \\t |
| /LineSeparator | str_esc | Line separator. Defaults to newline (\\n). Examples: ;\\n or \r\\n |
| /QuoteCharacter | str_esc | Character to quote a field with, if the fields contains separators. Defaults to quote (\"). |
| /EscapeCharacter | str_esc | Character to escape the quote character with. Defaults to quote (\"). |
| /FileTerminator | str_esc | File termination at end-of-file. Example: EOF or \xff |
| /NullRepresentation | esc_str | String representation for columns with NULL value. |
| /AvroCompression | codec | Avro compression codec. Value should be Deflate. |
| /AvroVersion | version | Version of Apache AVRO format. Possible values are v1_6, v1_7 and v1_8 (the default). |
| /JsonMode | mode | Style used to write row into JSON format. |
| /PageSize | Parquet page size in bytes. | |
| /RowGroupThreshold | Maximum row group size in bytes for Parquet. | |
| /ParquetVersion | version | Category of data types to represent complex data into Parquet format. |
| /ConvertNewlinesTo | style | Write files with UNIX or DOS style newlines. |
| /CaptureConverter | path | Run files through converter before reading. |
| /CaptureConverterArguments | userarg | Arguments to the capture converter. |
| /IntegrateConverter | path | Run files through converter after writing. |
| /IntegrateConverterArguments | userarg | Arguments to the integrate converter program. |
| /Context | context | Action only applies if Refresh/Compare context matches. |
Integrate
| Parameter | Argument | Description |
|---|---|---|
| /Burst | Resort changes, load into staging table and apply with set-wise SQL. | |
| /BurstCommitFrequency | freq | Frequency of commits. Values STATEMENT, TABLE or CYCLE. |
| /Coalesce | Coalesce consecutive changes on the same row into a single change. | |
| /ReorderRows | mode | Control order in which changes are written to files. Values NONE, BATCH_BY_TABLE, ORDER_BY_TABLE or SORT_COALESCE. |
| /Resilient | mode | Resilient integrate for inserts, updates and deletes. Values WARNING or SILENT. |
| /OnErrorSaveFailed | Write failed row to fail table. | |
| /DbProc | Apply changes by calling integrate database procedures. | |
| /TxBundleSize | int | Bundle small transactions for improved performance. |
| /TxSplitLimit | int | Split very large transactions to limit resource usage. |
| /NoTriggerFiring | Enable/Disable triggering of database rules. | |
| /SessionName | sess_name | Integrate changes with special session name sess_name. |
| /Topic | expression | Name of the Kafka topic. You can use strings/text or expressions as Kafka topic name. |
| /MessageBundling | mode | Number of messages written into single Kafka message. Kafka message contains one row by default. |
| /MessageBundlingThreshold | int | The threshold for bundling rows in a Kafka message. The default value is 800,000 bytes. |
| /MessageKey | expression | Expression to generate user defined key in a Kafka message. |
| /RenameExpression | expression | Expression to name new files, containing brace substitutions. |
| /ComparePattern | patt | Perform direct file compare. |
| /ErrorOnOverwrite | Error if a new file has same name as an existing file. | |
| /MaxFileSize | size | Limit each XML file to size bytes. |
| /Verbose | Report name of each file integrated. | |
| /TableName | apitab | API name of table to upload attachments into. |
| /KeyName | apikey | API name of attachment table's key column. |
| /CycleByteLimit | int | Max amount of routed data (compressed) to process per integrate cycle. |
| /JournalRouterFiles | Move processed router files to journal directory on hub. | |
| /JournalBurstTable | Keep track of changes in the burst table during. | |
| /Delay | N | Delay integration of changes for N seconds. |
| /Context | ctx | Action only applies if Refresh/Compare context matches. |
LocationProperties
| Parameter | Argument | Description |
|---|---|---|
| /SslRemoteCertificate | file | Enable SSL encryption to remote location; verify location with certificate. |
| /SslLocalCertificateKeyPair | path | Enable SSL encryption to remote location; identify with certificate/key. |
| /ThrottleKbytes | kbytes | Restrain net bandwidth into packets of kbytes bytes. |
| /ThrottleMillisecs | msecs | Restrain net bandwidth by msecs second(s) wait between packets. |
| /Proxy | proxy | Proxy server URL for FTP, SFTP, WebDAV or Salesforce locations. |
| /Order | N | Specify order of hub->loc proxy chain. |
| /StateDirectory | path | Directory for file location state files. Defaults to <top>/_hvr_state. |
| /IntermediateDirectory | dir | Directory for storing 'intermediate files' that are generated during compare. |
| /CaseSensitiveNames | DBMS table and columns names are treated case sensitive by HVR. | |
| /StagingDirectoryHvr | URL | Directory for bulk load staging files. |
| /StagingDirectoryDb | URL | Location for the bulk load staging files visible from the Database. |
| /StagingDirectoryCredentials | credentials | Credentials to be used for S3 authentication during RedShift bulk load. |
| /S3Encryption | keyinfo | Key information to be used for S3 client side encryption. |
| /BucketsCount | Number of buckets to be specified while creating a table in Hive ACID. | |
| /BulkAPI | Use Salesforce Bulk API (instead of the SOAP interface). | |
| /SerialMode | Force serial mode instead of parallel processing for Bulk API. | |
| /CloudLicense | Location runs on cloud node with on-demand licensing, for example in Amazon or Azure Marketplace. |
Restrict
| Parameter | Argument | Description |
|---|---|---|
| /CaptureCondition | sql_expr | Restrict during capture. |
| /IntegrateCondition | sql_expr | Restrict during integration. |
| /RefreshCondition | sql_expr | Restrict during refresh and compare. |
| /CompareCondition | sql_expr | Restrict during compare. |
| /HorizColumn | col_name | Horizontal partition table based on value in col_name. |
| /HorizLookupTable | tbl_name | Join partition column with horizontal lookup table. |
| /DynamicHorizLookup | Changes to lookup table also trigger replication. | |
| /AddressTo | addr | Only send changes to locations specified by address. |
| /AddressSubscribe | addr | Get copy of any changes sent to matching address. |
| /SelectDistinct | Filter duplicate records during refresh/compare. | |
| /Context | ctx | Action only applies if Refresh/Compare context matches. |
Scheduling
| Parameter | Argument | Description |
|---|---|---|
| /CaptureStartTimes | times | Trigger capture job at specific times, rather than continuous cycling. |
| /CaptureOnceOnStart | Capture job runs for one cycle after trigger. | |
| /IntegrateStartAfterCapture | Trigger integrate job only after capture job routes new data. | |
| /IntegrateStartTimes | times | Trigger integrate job at specific times, rather than continuous cycling. |
| /IntegrateOnceOnStart | Integrate job runs for one cycle after trigger. | |
| /RefreshStartTimes | times | Trigger refresh job at specific times. |
| /CompareStartTimes | crono | Trigger compare job at specific times. |
| /StatsMetrics | Set of metrics that hvrstats job must gather | |
| /StatsGranularity | Time granularity based on which the hvrstats job gathers metrics | |
| /StatsHistory | size | Size of history maintained by hvrstats job, before it purges its own rows. |
| /StatsCycleDelay | Time delay between metric gather cycle. |
TableProperties
| Parameter | Argument | Description |
|---|---|---|
| /BaseName | tbl_name | Name of table in database differs from name in catalogs. |
| /Absent | Exclude table (which is available in the channel) from being replicated/integrated into target. | |
| /DuplicateRows | Table has duplicate rows and no unique key. | |
| /Schema | schema | Database schema which owns table. |
| /IgnoreCoerceError | Coerce illegal/big/small values to empty/max/min. | |
| /CoerceErrorPolicy | Defines a policy to handle type coercion error. | |
| /CoerceErrorType | Defines which types of coercion errors are affected by /CoerceErrorPolicy. | |
| /TrimWhiteSpace | Remove trailing whitespace from varchar. | |
| /TrimTime | policy | Trim time when converting from Oracle and SQL Server date. |
| /MapEmptyStringToSpace | Convert between empty varchar and Oracle varchar space. | |
| /MapEmptyDateToConstant | date | Convert between constant date (dd/mm/yyyy) and Ingres empty date. |
| /CreateUnicodeDatatypes | On table creation use Unicode data types, e.g. map varchar to nvarchar. | |
| /DistributionKeyLimit | int | Maximum number of columns in the implicit distribution key. |
| /DistributionKeyAvoidPattern | patt | Avoid putting given columns in the implicit distribution key. |
| /CharacterMapping | rules | Specify the replacement rules for unsupported characters. |
| /MapBinary | policy | Specify how binary data is represented on the target side. |
| /MissingRepresentationString | str | Inserts value str into the string data type column(s) if value is missing/empty in the respective column(s) during integration. |
| /MissingRepresentationNumeric | str | Inserts value str into the numeric data type column(s) if value is missing/empty in the respective column(s) during integration. |
| /MissingRepresentationDate | str | Inserts value str into the date data type column(s) if value is missing/empty in the respective column(s) during integration. |
Transform
| Parameter | Argument | Description |
|---|---|---|
| /Command | path | Path to script or executable performing custom transformation. |
| /CommandArguments | userarg | Value(s) of parameter(s) for transform (space separated). |
| /SapAugment | Capture job selecting for de-clustering of multi-row SAP cluster tables. | |
| /SapXForm | Invoke SAP transformation for SAP pool and cluster tables. | |
| /UnpackTables | Transform will map *_pack tables into *_*_unpack* tables. | |
| /ExecOnHub | Execute transform on hub instead of location's machine. | |
| /Parallel | n | Distribute rows to multiple transformation processes. |
| /Context | context | Action only applies if Refresh/Compare context matches. |