Package | Description |
---|---|
oracle.odi.mapping.generation | |
oracle.odi.mapping.generation.spark |
Modifier and Type | Method and Description |
---|---|
java.lang.String |
Table.getGeneratedName(java.lang.String suffix)
Get the generated table name for this table, with an extra suffix appended.
|
SqlQuery |
SqlInsertStatement.getQuery()
Get the source QUERY (SqlQuery) object for this insert statement.
|
Modifier and Type | Method and Description |
---|---|
AbstractSyntaxTree |
SparkScript.cloneAST()
Clone the current script as the new AST object.
|
java.util.List |
SparkCommand.convertExprs(oracle.odi.domain.mapping.MapConnectorPoint scopingInputPoint,
java.util.List exprs,
boolean referenceOutputFields)
Convert list of mapping expressions into a list of expression texts.
|
java.util.List |
SparkCommand.convertExprsDF(oracle.odi.domain.mapping.MapConnectorPoint scopingInputPoint,
java.util.List exprs,
boolean referenceOutputFields,
boolean qualified)
Convert list of mapping expressions into a list of expression texts, specific to DataFrames.
|
java.lang.String |
SparkScript.customImports()
Return String value for the property spark.customPythonImports
|
java.util.List |
SparkSQLStoreCmd.generateRDDTypes()
Generate list of RDD types.
|
java.util.List |
SparkSQLStoreCmd.generateTargetColumnNames()
Generate the names for target columns.
|
java.util.List |
SparkStoreCmd.generateTargetColumnNames()
Generate the names for target columns.
|
java.lang.String |
SparkSQLStoreCmd.generateTargetColumnsDF()
Generate the target columns list.
|
java.lang.String |
SparkStoreCmd.generateTargetColumnsDF()
Generate the target columns list.
|
java.lang.String |
SparkSQLStoreCmd.generateTargetDefinition()
Generate the target definition.
|
java.lang.String |
SparkStoreCmd.generateTargetDefinition()
Generate the target definition.
|
java.lang.String |
SparkSQLStoreCmd.generateTargetDefinitionDF()
Generate the target definition for dataframes.
|
java.lang.String |
SparkStoreCmd.generateTargetDefinitionDF()
Generate the target definition for dataframes.
|
java.lang.String |
SparkHiveStoreCmd.generateTargetDefinitionHive()
Get the generated target definition.
|
java.util.List |
SparkSQLStoreCmd.generateTargetTypes()
Generate list of target types.
|
java.util.List |
SparkStoreCmd.generateTargetTypes()
Generate list of target types.
|
java.util.List |
SparkGroupCmd.getAllAggFunctions()
Gets names of all the aggregate functions.
|
java.util.List |
SparkGroupCmd.getAllAggrAliases()
Gets the list of aliases for aggregation expressions.
|
java.util.List |
SparkGroupCmd.getAllAggrExprs()
Gets the list of aggregation expressions.
|
java.util.List |
SparkSQLCmd.getAPNodeAttributes()
Gets list of AP node attributes.
|
java.lang.String |
SparkStoreCmd.getBootstrapServers()
Get the bootstrap servers.
|
java.util.List |
SparkCommand.getChildren() |
java.util.List |
SparkScript.getChildren() |
SparkCommand |
SparkScript.getCommand()
Get the very last command in this this spark script.
|
int |
SparkCommand.getCommandIndex()
Get the index of this command.
|
java.util.List |
SparkScript.getCommands()
Get list of commands for this spark script.
|
SparkCommandType |
SparkScript.getCommandType()
Get the type of the last command for spark script
|
java.util.List |
SparkGroupCmd.getComplexAggrAliases()
Gets the list of names for complex aggregation expressions.
|
java.util.List |
SparkGroupCmd.getComplexAggrExprs()
Gets the list of complex aggregation expressions.
|
oracle.odi.domain.mapping.expression.MapExpression |
SparkOrderByCmd.getCondition()
Get the condition of this command.
|
oracle.odi.domain.mapping.physical.MapPhysicalDesign |
SparkScript.getCurrentPhysicalDesign()
Get the physical design for the physical node associated with this script.
|
java.lang.String |
SparkLoadCmd.getDateFormat()
Get Date format property
|
java.lang.String |
SparkScript.getDebugMode()
Get the debug mode for this script.
|
java.util.Set |
SparkScript.getDestinationTargets()
Get the set of destination target names for this script.
|
java.util.List |
SparkGroupCmd.getDirectMapAliases()
Gets the list of names for directly mapped attributes.
|
java.util.List |
SparkGroupCmd.getDirectMapExprs()
Gets the list of expressions on directly mapped attributes.
|
java.lang.String |
SparkStoreCmd.getDirectory()
Get the directory name.
|
java.lang.String |
SparkScript.getEkmOptionValue(java.lang.String key)
Returns an EKM option value.
|
java.lang.String |
SparkScript.getEkmOptionValue(java.lang.String key,
java.lang.String defaultValue)
Returns an EKM option value.
|
java.util.List |
SparkScript.getExpressions() |
java.lang.String |
SparkCommand.getExprText(oracle.odi.domain.mapping.MapConnectorPoint scopingInputPoint,
oracle.odi.domain.mapping.expression.MapExpression expr)
Return the expression text according to a given map expression.
|
java.lang.String |
SparkScript.getExprText(oracle.odi.domain.mapping.MapConnectorPoint scopingInputPoint,
oracle.odi.domain.mapping.expression.MapExpression expr,
boolean referenceOutputFields)
Return the spark script expression text according to a given map expression.
|
java.lang.String |
SparkCommand.getExprTextDF(oracle.odi.domain.mapping.MapConnectorPoint scopingInputPoint,
oracle.odi.domain.mapping.expression.MapExpression expr)
Return the expression text according to a given map expression, specific to DataFrames.
|
java.lang.String |
SparkScript.getExprTextDF(oracle.odi.domain.mapping.MapConnectorPoint scopingInputPoint,
oracle.odi.domain.mapping.expression.MapExpression expr,
boolean referenceOutputFields,
boolean qualified)
Return the spark script expression text according to a given map expression, specific to DataFrames.
|
java.lang.String |
SparkScript.getExprTextDF(oracle.odi.domain.mapping.MapConnectorPoint scopingInputPoint,
oracle.odi.domain.mapping.expression.MapExpression expr,
boolean referenceOutputFields,
boolean qualified,
java.lang.String qualifier)
Return the spark script expression text according to a given map expression, specific to DataFrames.
|
java.lang.String |
SparkLoadCmd.getFieldSep()
Get the file separator.
|
java.lang.String |
SparkStoreCmd.getFieldSep()
Get the field separator.
|
java.lang.String |
SparkLoadCmd.getFileFormat()
Get the file format.
|
java.lang.String |
SparkLoadCmd.getFileLoc()
Gets the file location.
|
oracle.odi.domain.mapping.expression.MapExpression |
SparkFilterCmd.getFilterCond()
Get the filter condition.
|
java.lang.String |
SparkCommand.getFlexField(java.lang.String flexFieldName,
java.lang.String defaultFieldValue)
Get the flexfield value by using the flexField name.
|
java.util.List |
SparkStoreCmd.getFullTgtAttrs()
Get full target attributes.
|
boolean |
SparkLoadCmd.getGenerateInputSchemaDefinition()
Gets a flag indicating if the input schema definition should be generated in the generated code.
|
java.util.List |
SparkGroupCmd.getGroupByAliases()
Gets the names of group by fields.
|
java.util.List |
SparkPivotCmd.getGroupByAliases()
Gets the names of fields used for group by.
|
java.util.List |
SparkUnpivotCmd.getGroupByAliases()
Get group by aliases.
|
java.util.List |
SparkGroupCmd.getGroupByExprs()
Gets the list of expressions for "group by" fields
|
java.util.List |
SparkPivotCmd.getGroupByExprs()
Gets group by expressions.
|
java.util.List |
SparkUnpivotCmd.getGroupByExprs()
Get group by expressions.
|
java.util.List |
SparkPivotCmd.getGroupByIndexes()
Get list of indexes for group by.
|
oracle.odi.domain.mapping.expression.MapExpression |
SparkGroupCmd.getHavingCond()
Gets the having condition.
|
java.util.List |
SparkSplitCmd.getIfCache()
Gets the list of ifCache settings under if-condition control.
|
java.util.List |
SparkSplitCmd.getIfCacheLevel()
Gets the list of ifCacheLevel settings under if-condition control.
|
java.util.List |
SparkSplitCmd.getIfIsMultiColumnKey()
Gets the list of ifIsMultiColumnKey settings under if-condition control.
|
java.util.List |
SparkSplitCmd.getIfLevelParallelism()
Gets the list of ifLevelParallelism settings under if-condition control.
|
java.util.List |
SparkSplitCmd.getIfPartitionKeyFunc()
Gets the list of ifPartitionKeyFunc settings under if-condition control.
|
java.util.List |
SparkSplitCmd.getIfPartitionSortType()
Gets the list of ifPartitionSortType settings under if-condition control.
|
java.util.List |
SparkSplitCmd.getIfRepartitionFunc()
Gets the list of ifRepartitionFunc settings under if-condition control.
|
java.util.List |
SparkSplitCmd.getIfRepartitionValue()
Gets the list of ifRepartitionValue settings under if-condition control.
|
java.util.List |
SparkSplitCmd.getIfSortPartitions()
Gets the list of ifSortPartitions settings under if-condition control.
|
java.util.List |
SparkSplitCmd.getIfTargetAliases()
Gets the list of target names under if-condition control.
|
java.util.List |
SparkSplitCmd.getIfUseSQLExpressions()
Gets the list of ifUseSQLExpressions settings under if-condition control.
|
boolean |
SparkSQLStoreCmd.getInferSchema()
Return true if the schema for the target dataframe is inferred from RDD data.
|
boolean |
SparkLoadCmd.getInferSourceShema()
Get infer source schema property
|
oracle.odi.domain.mapping.MapConnectorPoint |
SparkScript.getInputConnectorPoint(SparkSchema inputSchema)
Get the input connector point according to the spark input schema.
|
SparkField |
SparkScript.getInputField(oracle.odi.domain.mapping.MapAttribute attr)
Get the input spark field according to a specific map attribute.
|
java.lang.String |
SparkLoadCmd.getInputSchemaDefinition()
Gets the valid definition of input schema .
|
java.util.List |
SparkScript.getInputSchemas()
Get all the input spark schemas for this script.
|
java.util.List |
SparkScript.getInputScripts()
Get a list of scripts as the input of current one.
|
java.util.List |
SparkScript.getInputSqlQueries()
Get the Input sql queries.
|
java.util.List |
SparkLoadCmd.getInputStructTypeSchemaDefinition()
Gets the valid definition of input schema defined by the specified structure
type, for the data frame that was read.
|
java.util.List |
SparkLoadCmd.getInputStructTypeSchemaDefinitionDF()
Gets the valid definition of input schema defined by the specified structure
type, for the data frame that was read.
|
java.lang.String |
SparkSQLCmd.getJdbcDriver()
Get the name of JDBC driver.
|
java.lang.String |
SparkSQLStoreCmd.getJdbcDriver()
Get the name of JDBC driver.
|
java.lang.String |
SparkSQLCmd.getJdbcUrl()
Get the JDBC URL.
|
java.lang.String |
SparkSQLStoreCmd.getJdbcUrl()
Get the JDBC URL.
|
oracle.odi.domain.mapping.expression.MapExpression |
SparkJoinCmd.getJoinCond()
Get join condition.
|
JoinTable.JoinType |
SparkJoinCmd.getJoinType()
Get the join type for this command
|
oracle.odi.domain.mapping.physical.MapPhysicalNode |
SparkLoadCmd.getKMNode()
Get then map physical node this command is corresponding to.
|
oracle.odi.domain.mapping.physical.MapPhysicalNode |
SparkSQLCmd.getKMNode()
Get the physical node referenced by this command
|
oracle.odi.domain.mapping.physical.MapPhysicalNode |
SparkStoreCmd.getKMNode()
Get the physical node which this command is corresponding to.
|
java.lang.String |
SparkLoadCmd.getLoadFuncArgs()
Gets the arguments to load function.
|
java.lang.String |
SparkLoadCmd.getLoadFunction()
Gets the load function
|
java.lang.String |
SparkLoadCmd.getLoadOptions()
Get the load options
|
java.lang.String |
SparkKeyBasedLookupCmd.getLookupCmdText()
Get the text for lookup command.
|
java.util.List |
SparkLookupCmd.getLookupKeyExpressions()
Get key expressions for lookup.
|
java.lang.String |
SparkKeyBasedLookupCmd.getLookupSourceLoadingCmdText()
Get the text for the source loading commands used for lookup.
|
long |
SparkSQLCmd.getLowerBound()
Get lower bound of partitions.
|
oracle.odi.domain.mapping.expression.MapExpression |
SparkGroupCmd.getManualGroupBy()
Gets the manual gourp by clause.
|
java.lang.String |
SparkScript.getMapName()
Get the text for the fileName.
|
java.lang.String |
SparkScript.getMapNameLong()
Get the text for the fileNameLong.
|
oracle.odi.domain.mapping.Mapping |
SparkScript.getMapping()
Get the corresponding mapping for this script.
|
java.util.HashMap |
SparkCommand.getModulePaths()
Gets import module paths; Each context has a corresponding module path.
|
java.util.List |
SparkHiveStoreCmd.getNonPartitionColumns()
Get partition columns.
|
int |
SparkSQLCmd.getNumPartitions()
Get the number of the partitions.
|
java.util.List |
SparkSetCmd.getOperations()
Gets the list of set operations.
|
java.lang.String |
SparkCommand.getOptionValue(java.lang.String optionName)
Get the option value from a physical node by using option name
|
java.util.List |
SparkOrderByCmd.getOrderByExprs()
Gets the list of order by expressions.
|
java.util.List |
SparkOrderByCmd.getOrderByTypes()
Gets the order by types.
|
oracle.odi.domain.mapping.MapConnectorPoint |
SparkScript.getOutputConnectorPoint()
Get the output connector point for the physical map node which this script reference.
|
SparkSchema |
SparkScript.getOutputSchema()
Get the output spark schema which is calculated from the last command.
|
java.lang.String |
SparkSQLCmd.getOutputSchemaConnectedFields()
Gets list of connected fields for output schema.
|
java.lang.String |
SparkSQLCmd.getOutputSchemaDefinition()
Gets the definition for output schema
|
java.lang.String |
SparkSQLCmd.getOutputSchemaDefinitionRow()
Gets the definition of output schema for sql format
|
java.util.List |
SparkLoadCmd.getOutputSchemaFieldList()
Generate the names for source columns.
|
java.lang.String |
SparkSQLCmd.getOutputSchemaFields()
Gets list of fields for output schema.
|
java.lang.String |
SparkLoadCmd.getOutputSchemaFieldsDF()
Gets list of fields for output schema, specific to DataFrames.
|
java.lang.String |
SparkSQLCmd.getOutputSchemaFieldsDF()
Gets list of fields for output schema, specific to DataFrames.
|
SqlQuery |
SparkScript.getOutputSqlQuery()
Get the Output SqlQuery.
|
boolean |
SparkSQLStoreCmd.getOverwriteTargetTable()
Get flag OverwriteTargetTable.
|
java.util.List |
SparkSplitCmd.getOWCache()
Gets the otherwise target cache setting.
|
java.util.List |
SparkSplitCmd.getOWCacheLevel()
Gets the otherwise target owCacheLevel setting.
|
java.util.List |
SparkSplitCmd.getOWIsMultiColumnKey()
Gets the list of owIsMultiColumnKey settings under otherwise control.
|
java.util.List |
SparkSplitCmd.getOWLevelParallelism()
Gets the list of owLevelParallelism settings under otherwise control.
|
java.util.List |
SparkSplitCmd.getOWPartitionKeyFunc()
Gets the list of owPartitionKeyFunc settings under otherwise control.
|
java.util.List |
SparkSplitCmd.getOWPartitionSortType()
Gets the list of owPartitionSortType settings under otherwise control.
|
java.util.List |
SparkSplitCmd.getOWRepartitionFunc()
Gets the list of owRepartitionFunc settings under otherwise control.
|
java.util.List |
SparkSplitCmd.getOWRepartitionValue()
Gets the list of owRepartitionValue settings under otherwise control.
|
java.util.List |
SparkSplitCmd.getOWSortPartitions()
Gets the list of owSortPartitions settings under otherwise control.
|
java.util.List |
SparkSplitCmd.getOWTargetAlias()
Gets the otherwise target name.
|
java.util.List |
SparkSplitCmd.getOWUseSQLExpressions()
Gets the list of owUseSQLExpressions settings under otherwise control.
|
SparkScript |
SparkCommand.getParentScript()
Get the parent script.
|
java.lang.String |
SparkSQLCmd.getPartitionColumn()
Get the partition column.
|
java.util.List |
SparkHiveStoreCmd.getPartitionColumns()
Get partition columns.
|
java.lang.String |
SparkSQLCmd.getPartitionPredicates()
Get the partition predicates
|
oracle.odi.domain.mapping.physical.MapPhysicalNode |
SparkScript.getPhysicalNode()
Get the physical node that the last command is corresponding to.
|
java.util.List |
SparkFlattenCmd.getPrefixedAttrNamesDF(java.util.List attributes) |
java.util.List |
SparkDistinctCmd.getPrefixedTgtAttrNamesDF() |
java.util.List |
SparkExpressionCmd.getPrefixedTgtAttrNamesDF() |
java.util.List |
SparkSetCmd.getPrefixedTgtAttrNamesDF() |
java.util.List |
SparkGroupCmd.getPrefixedTgtAttrNamesDF(java.util.List p_attrubuteNames) |
java.lang.String |
SparkLoadCmd.getRowSep()
Get the row separator.
|
java.lang.String |
SparkStoreCmd.getRowSep()
Get the row separator.
|
int |
SparkSQLStoreCmd.getSamplingRatio()
Get the sampling ratio.
|
java.lang.String[] |
SparkCommand.getScriptLevelCommonContext()
Get the common context in the script level.
|
java.lang.String |
SparkSQLCmd.getServerName()
Get the name of the server.
|
java.lang.String |
SparkSQLStoreCmd.getServerName()
Get the server name.
|
SparkScript |
SparkScript.getSharedInputScript()
Get shared input script.
|
java.util.List |
SparkGroupCmd.getSimpleAggrAliases()
Gets the list of names for simple aggregate expressions.
|
java.util.List |
SparkGroupCmd.getSimpleAggrEpxrs()
Gets the list of simple aggregate expressions.
|
java.util.List |
SparkGroupCmd.getSimpleAggrFuncs()
Gets the list of aggregate functions on simple aggr exprs.
|
java.lang.String |
SparkCommand.getSourceAlias()
Get the the source alias.
|
java.util.List |
SparkCommand.getSourceAliases()
Gets the source aliases of this BigData command.
|
java.lang.String |
SparkCommand.getSparkContextName() |
java.lang.String |
SparkHqlCmd.getSparkContextName()
Get the Hive context name.
|
java.lang.String |
SparkSQLCmd.getSparkContextName()
Gets the context name for Spark.
|
java.lang.String |
SparkCommand.getSparkContextType() |
java.lang.String |
SparkHqlCmd.getSparkContextType()
Get the Hive context type.
|
java.lang.String |
SparkSQLCmd.getSparkContextType()
Gets the context type for Spark.
|
java.lang.String |
SparkScript.getSparkDataServerProperty(java.lang.String key)
Returns a ?-level codegen expression expanding to the respective Spark DataServer property.
|
java.lang.String |
SparkScript.getSparkDataServerProperty(java.lang.String key,
java.lang.String defaultValue)
Returns a ?-level codegen expression expanding to the respective Spark DataServer property.
|
oracle.odi.domain.mapping.physical.ExecutionUnit |
SparkScript.getSparkExecutionUnit()
Get Spark Execution Unit associated with this script.
|
java.lang.String |
SparkScript.getSparkLSchemaName()
Get logical schema name of Spark Execution Unit associated with this script.
|
java.util.List |
SparkSplitCmd.getSplitConditions_DF()
Gets the split conditions.
|
java.util.List |
SparkSplitCmd.getSplitConditions()
Gets the split conditions.
|
SqlQuery |
SparkSQLCmd.getSQLStatement()
Get the Sql Statement for this command.
|
java.lang.String |
SparkSQLCmd.getSrcSchemaName()
Get the name of source schema.
|
int |
SparkGroupCmd.getStateRetention()
Gets the retention period for the spark command.
|
java.lang.String |
SparkStoreCmd.getStoreFuncArgs()
Get the arguments of store functions.
|
java.lang.String |
SparkStoreCmd.getStoreFunction()
Get the name for the store function.
|
java.lang.String |
SparkStoreCmd.getStoreOptions()
Get the store options.
|
java.lang.String |
SparkCommand.getStrctType(java.lang.String builtinType)
Get the Spark valid Struct DataTypes
|
boolean |
SparkCommand.getStreamEnableWindowing()
Get the enable windowing flag
|
java.lang.String |
SparkScript.getStreamingBatchDuration()
Return integer value for the property spark.batchDuration.
|
java.lang.Integer |
SparkCommand.getStreamSlideInterval()
Get the steam slide interval, it's used only for the stream mode.
|
java.lang.Integer |
SparkCommand.getStreamWindowLength()
Get the length of the steam window.
|
java.util.List |
SparkKeyBasedLookupCmd.getSubCommands()
Get list of sub-commands.
|
java.lang.String |
SparkPivotCmd.getSubGrpAggrFunc()
Gets the aggregate function on subgroup.
|
java.lang.String |
SparkUnpivotCmd.getSubGrpAlias()
Get sub-group alias.
|
java.util.List |
SparkPivotCmd.getSubGrpAliases()
Gets the names for subgroup output fields.
|
java.util.List |
SparkPivotCmd.getSubGrpExprs()
Gets the expressions for computing output values for fields in subgroup.
|
java.lang.String |
SparkPivotCmd.getSubGrpMatchAlias()
Gets the name for calculating the value from a group so that the group can
be further partitioned into subgroups
|
oracle.odi.domain.mapping.expression.MapExpression |
SparkPivotCmd.getSubGrpMatchExpr()
Gets the expression for calculating the value from a group so that the
group can be further partitioned into subgroups.
|
java.util.List |
SparkPivotCmd.getSubGrpValues()
Gets the compare values for subgroups.
|
java.util.List |
SparkUnpivotCmd.getSubGrpValues()
Get list of sub-group values.
|
java.lang.String |
SparkCommand.getTargetAlias()
Get the target alias for this command.
|
java.util.List |
SparkDistinctCmd.getTargetAttrExprs()
Gets the list of expressions on target attributes.
|
java.util.List |
SparkDistinctCmd.getTargetAttrNames()
Gets the names of target attributes.
|
java.lang.String |
SparkStoreCmd.getTargetSchema()
Get the target schema.
|
java.lang.String |
SparkSQLStoreCmd.getTargetTableName()
Get name of target table.
|
java.lang.String |
SparkSQLStoreCmd.getTechno()
Get the technology.
|
java.lang.String |
SparkLoadCmd.getTechnology()
Get the technology type.
|
java.lang.String |
SparkCommand.getText()
Return the text format of the spark command.
|
java.lang.String |
SparkScript.getText()
Get the text for this script.
|
java.util.List |
SparkExpressionCmd.getTgtAttrExprs()
Gets the list of expressions on target attributes.
|
java.util.List |
SparkInputSigCmd.getTgtAttrExprs()
Get list of expressions for target attributes.
|
java.util.List |
SparkOutputSigCmd.getTgtAttrExprs()
Get list of target expressions.
|
java.util.List |
SparkStoreCmd.getTgtAttrExprs()
Get the expressions for target attributes.
|
java.util.List |
SparkSetCmd.getTgtAttrExprsList()
Gets the list of target output expressions.
|
java.util.List |
SparkExpressionCmd.getTgtAttrNames()
Gets the names of target attributes.
|
java.util.List |
SparkInputSigCmd.getTgtAttrNames()
Get names of the target attributes.
|
java.util.List |
SparkOutputSigCmd.getTgtAttrNames()
Get the names of target attributes.
|
java.util.List |
SparkSetCmd.getTgtAttrNames()
Gets the list of target output field names.
|
java.util.List |
SparkStoreCmd.getTgtAttrNames()
Get the names of target attributes.
|
java.util.List |
SparkStoreCmd.getTgtDataTypes()
Get the list of target data types.
|
java.lang.String |
SparkStartStreamingCmd.getTimeout()
Gets the timeout configured for this command, using default time unit.
|
java.util.List |
SparkUnpivotCmd.getTransformAliases()
Get transform aliases.
|
java.util.List |
SparkUnpivotCmd.getTransformExprsList()
Get the list of transform expressions
|
java.lang.String |
SparkCommand.getType()
Get the value of SparkCommandType
|
java.lang.String |
SparkScript.getType()
Get the name of AST Object as the type.
|
SparkCommandType |
SparkDistinctCmd.getTypeEnum()
Get the type of spark command
|
SparkCommandType |
SparkExpressionCmd.getTypeEnum()
Get the type of spark command
|
SparkCommandType |
SparkFilterCmd.getTypeEnum()
Get the type of this spark command.
|
SparkCommandType |
SparkFlattenCmd.getTypeEnum()
Get the type of this spark command.
|
SparkCommandType |
SparkGroupCmd.getTypeEnum()
Get the type of this spark command.
|
SparkCommandType |
SparkInputSigCmd.getTypeEnum()
Get the type of this spark command.
|
SparkCommandType |
SparkJoinCmd.getTypeEnum()
Get the type of spark command
|
SparkCommandType |
SparkKeyBasedLookupCmd.getTypeEnum()
Get the type of spark command
|
SparkCommandType |
SparkLoadCmd.getTypeEnum()
Get the type of spark command
|
SparkCommandType |
SparkLookupCmd.getTypeEnum()
Get the type of spark command
|
SparkCommandType |
SparkOrderByCmd.getTypeEnum()
Get the type of spark command
|
SparkCommandType |
SparkOutputSigCmd.getTypeEnum()
Get the type of spark command
|
SparkCommandType |
SparkPivotCmd.getTypeEnum()
Get the type of spark command
|
SparkCommandType |
SparkSQLCmd.getTypeEnum()
Get the type of spark command
|
SparkCommandType |
SparkSQLStoreCmd.getTypeEnum()
Get the type of spark command
|
SparkCommandType |
SparkSetCmd.getTypeEnum()
Get the type of spark command
|
SparkCommandType |
SparkSplitCmd.getTypeEnum()
Get the type of spark command
|
SparkCommandType |
SparkStartStreamingCmd.getTypeEnum()
Get the type of spark command
|
SparkCommandType |
SparkStoreCmd.getTypeEnum()
Get the type of spark command
|
long |
SparkSQLCmd.getUpperBound()
Get upper bound of partitions.
|
java.lang.String |
SparkSQLCmd.getUser()
Get the user name used in JDBC URL.
|
java.lang.String |
SparkSQLStoreCmd.getUser()
Get the user name for JDCB connection.
|
boolean |
SparkScript.isCheckpointEnabled()
Return boolean value for the property spark.checkpointing
|
boolean |
SparkFlattenCmd.isCollection()
Return true if the complex type is wrapped by a collection such as an array
|
boolean |
SparkPivotCmd.isDecorator()
Return true if the corresponding component is a decorator.
|
boolean |
SparkSQLCmd.isFilteringAllowed()
Return true if filtering is allowed.
|
boolean |
SparkStoreCmd.isFlattenInput()
Return true if flatten input is used.
|
boolean |
SparkStoreCmd.isHasScaledField()
Return true if scaled field is used.
|
boolean |
SparkUnpivotCmd.isIncludingNulls()
Return true if including nulls is allowed.
|
boolean |
SparkLookupCmd.isKeyBasedQueryAttached()
Return true if the key based query is attached.
|
boolean |
SparkLoadCmd.isKeySupport()
Return true if key is supported.
|
boolean |
SparkLoadCmd.isLoadFile()
Return true if loading is for file.
|
java.lang.Boolean |
SparkJoinCmd.isMapSide()
Return true if it's on the map side.
|
java.lang.Boolean |
SparkLookupCmd.isMapSide()
Return true if it's on the map side.
|
boolean |
SparkSQLStoreCmd.isMatchSchemaToTargetTypes()
Return true if infer schema is false and the schema applied to the target dataframe is matching target datastore datatypes.
|
static boolean |
SparkScript.isNodeInStreamingMode(oracle.odi.domain.mapping.physical.MapPhysicalNode node)
Return true if streaming mode is being used for the associated mapping node.
|
boolean |
SparkScript.isPhysicalDesignInStreamingMode()
Return true if streaming mode is switched on for the mapping.
|
boolean |
SparkSQLStoreCmd.isQuotedColumns()
Return true target table columns need to be quoted.
|
boolean |
SparkStoreCmd.isQuotedColumns()
Return true target table columns need to be quoted.
|
boolean |
SparkLookupCmd.isReturnDefaultIfNthMissing()
For a single match with Nth row matching strategy : if Nth row > 1 true means we return default data, false means we return the matching row
|
boolean |
SparkPivotCmd.isSelector()
Return true if the corresponding component is a selector.
|
boolean |
SparkScript.isSelector()
Return True if the last command on this script is based on a selector component.
|
boolean |
SparkLookupCmd.isSimpleEqualLookup()
Return true if a simple equal condition is used for lookup.
|
boolean |
SparkGroupCmd.isStatefulEnabled()
Return true if stateful is enabled
|
boolean |
SparkStoreCmd.isStoreKeys()
Return true of keys are needed to store.
|
boolean |
SparkStoreCmd.isStoreToFile()
Return true if this command is used to store to a file.
|
boolean |
SparkCommand.isStreamingMode()
Return true if the corresponding script is on streaming mode.
|
boolean |
SparkFlattenCmd.isStructure()
Return true if the complex type is a structure
|
boolean |
SparkGroupCmd.isUpdatesOnlyEnabled()
Return true if updates only is enabled.
|
boolean |
SparkSQLCmd.needFormatOutput()
Return true if output format is needed.
|
boolean |
SparkSQLCmd.needPartitionColumn()
Return true if partition column is needed.
|
boolean |
SparkSQLCmd.needPartitionPredicates()
Return true if partition predicates are needed.
|
void |
SparkLookupCmd.setReturnDefaultIfNthMissing(boolean pReturnDefaultIfNthMissing) |
java.lang.String |
SparkCommand.toString()
Return basic information about SparkCommand with the description of its contents
|
java.lang.String |
SparkScript.toString()
Return basic information about SparkScript with the description of its contents
|
boolean |
SparkScript.useDataFrames()
Return boolean value for the property spark.useDataFrames
|
boolean |
SparkSQLCmd.useQueryAlias()
Return true if source query alias is needed.
|
boolean |
SparkCommand.useSQLExpressions()
Use SQL expressions.
|