Package | Description |
---|---|
com.huawei.dli.sdk | |
com.huawei.dli.sdk.authentication | |
com.huawei.dli.sdk.common | |
com.huawei.dli.sdk.common.type |
Modifier and Type | Method and Description |
---|---|
protected void |
BaseJob.apiExceptionRetryManage(int statusCode,
String serviceCode,
String errorMsg,
int currentTimes,
boolean mustRetry) |
void |
BatchJob.asyncSubmit() |
void |
ImportJob.asyncSubmit() |
void |
SQLJob.asyncSubmit() |
abstract void |
SparkJob.asyncSubmit()
Submit a Spark job in async.
|
com.huawei.dli.restapi.model.GlobalResponse |
DLIClient.authorizeBucket(com.huawei.dli.restapi.model.ObsBuckets obsBuckets)
用户主动授权起OBS桶的操作权限给DLI服务, 用于保存用户作业的checkpoint、作业的运行日志等
|
void |
UploadJob.beginCommit() |
void |
UploadJob.beginCommit(String transactionId) |
void |
SQLJob.cancelExportJob()
If exportResult is RUNNING or LAUNCHING,you can cancel it.
|
void |
Job.cancelJob()
Cancel RUNNING or LAUNCHING status job.
|
void |
DLIClient.cancelJob(String jobId)
Cancel LAUNCHING,RUNNING status job,
and the job status changes to CANCELLED.
|
protected boolean |
Job.checkJobFinish(long startTime,
JobResultInfo jobResultInfo) |
protected boolean |
Job.checkJobFinish(long startTime,
JobResultInfo jobResultInfo,
Long waitTime) |
protected void |
SparkJob.checkJobRunningState() |
protected void |
Job.checkJobRunningStatus(String jobId)
Check job running status by jobId after submit(),
when job status is FINISHED,FAILED,CANCELLED,mean job is over.
|
com.huawei.dli.restapi.model.ResponseJobInfo |
DLIClient.checkSQL(String sql)
Check the syntax of SQL, if syntax is wrong, it will throw exception.
|
com.huawei.dli.restapi.model.ResponseJobInfo |
DLIClient.checkSQL(String sql,
String currentdb)
Check the syntax of SQL, if syntax is wrong, it will throw exception.
|
void |
ObsWriter.close() |
void |
AvroObsWriter.close() |
void |
Writer.close() |
Cluster |
DLIClient.createCluster(String clusterName,
int cuCount,
String description)
Create a cluster.
|
Database |
DLIClient.createDatabase(String databaseName) |
Database |
DLIClient.createDatabase(String databaseName,
String description)
Create a database,must sure databaseName is not exist,
and there are enough quotas.
|
Database |
DLIClient.createDatabase(String databaseName,
String description,
String enterpriseProjectId) |
Table |
Database.createDLITable(String tableName,
String description,
List<Column> columns,
List<String> sortColumns)
Create DLI table.
|
com.huawei.dli.restapi.model.JobStatusResponse |
DLIClient.createFlinkJarJob(com.huawei.dli.restapi.model.CreateFlinkJarJobRequest body)
创建一个用户自定义Flink作业
用户自定义作业目前支持jar格式, 运行在独享集群中
|
com.huawei.dli.restapi.model.FlinkJobTemplateCreateResponse |
DLIClient.createFlinkJobTemplate(com.huawei.dli.restapi.model.CreateFlinkJobTemplateRequest body)
创建作业模板
Create job template on Data lake insight
|
Table |
Database.createObsTable(String tableName,
String description,
List<Column> columns,
StorageType storageType,
String dataPath,
CsvFormatInfo csvFormatInfo)
Create OBS table.
|
Queue |
DLIClient.createQueue(String queueName,
int cuCount,
ChargingMode chargingMode,
String description)
Create a queue.
|
Reader |
DownloadJob.createReader() |
List<Reader> |
DownloadJob.createReaders(int expectedMaxConcurrent)
Get readers to read data in concurrent, the reader num is calculate by DLI with the expectedMaxConcurrent,
that is the reader num is less or equal expectedMaxConcurrent if the data size is small.
|
Table |
Database.createTable(String tableName,
String dataLocation,
String description,
List<Column> columns,
StorageType storageType,
String dataPath,
CsvFormatInfo csvFormatInfo,
List<String> sortColumns)
Create DLI/OBS table
|
com.huawei.dli.restapi.model.CreateTsdbResp |
DLIClient.createTsdb(com.huawei.dli.restapi.model.CreateTsdbReq createTsdbReq,
String projectId)
创建时序数据库配置信息
|
Writer |
UploadJob.createWriter() |
com.huawei.dli.restapi.model.MessageInfo |
BatchJob.delBatch()
delete batchJob (recommend)
|
com.huawei.dli.restapi.model.MessageInfo |
BatchJob.delBatch(String batchId)
delete batchJob by batchId (deprecated)
|
com.huawei.dli.restapi.model.MessageInfo |
DLIClient.delBatchJob(String batchId)
Delete batchJob by batchId
|
void |
Cluster.deleteCluster()
Delete this cluster.
|
void |
Database.deleteDatabase()
Delete the current database.
|
List<com.huawei.dli.restapi.model.GlobalBatchResponse> |
DLIClient.deleteFlinkJobInBatch(com.huawei.dli.restapi.model.DeleteJobInBatchRequest body)
批量删除作业
|
com.huawei.dli.restapi.model.FlinkJobTemplateDeleteResponse |
DLIClient.deleteFlinkJobTemplate(Long templateId)
删除作业模板
当前模板被引用也允许删除模板
|
com.huawei.dli.restapi.model.GlobalResponse |
DLIClient.deleteJob(Long jobId)
删除任何状态的作业
|
void |
Queue.deleteQueue()
Delete this queue.
|
void |
DLIClient.deleteResource(String resourceName)
Delete package resources by name.
|
void |
DLIClient.deleteResource(String resourceName,
String group)
Delete package resources by name and group.
|
void |
Table.deleteTable()
Delete current table.
|
com.huawei.dli.restapi.model.GlobalResponse |
DLIClient.deleteTsdb(String projectId,
String tsdbId)
删除时序数据库配置信息
删除时序数据库配置信息
|
protected <T> com.huawei.dli.restapi.ApiResponse<T> |
BaseJob.executeCall(IHttpExecAdpter<T> exec,
Class<T> resDtoClass,
Function<T,com.huawei.dli.restapi.model.GlobalErrorResponse> errorFetcher,
Function<T,String> transactionConflictManage) |
<T,V> T |
BaseJob.executeWithRetry(IHttpExecAdpter<T> exec,
Class<T> resDtoClass,
Function<T,com.huawei.dli.restapi.model.GlobalErrorResponse> errorFetcher,
Function<T,String> transactionConflictManage) |
Reader |
SQLJob.exportResult()
Export QUEUE type job result with a iterator
|
Reader |
SQLJob.exportResult(Integer limit) |
void |
SQLJob.exportResult(String dataPath,
StorageType storageType,
CompressType compressType,
ExportMode exportMode,
String queueName)
Export QUEUE type job result.
|
void |
SQLJob.exportResult(String dataPath,
StorageType storageType,
CompressType compressType,
ExportMode exportMode,
String queueName,
boolean withColumnHeader) |
void |
SQLJob.exportResult(String dataPath,
StorageType storageType,
CompressType compressType,
ExportMode exportMode,
String queueName,
boolean withColumnHeader,
Integer limitNum) |
void |
SQLJob.exportResult(String dataPath,
StorageType storageType,
CompressType compressType,
ExportMode exportMode,
String queueName,
boolean withColumnHeader,
Integer limitNum,
String encodingType) |
void |
ObsWriter.flush() |
void |
AvroObsWriter.flush() |
void |
Writer.flush() |
String |
DLIClient.getAuthToken() |
Cluster |
DLIClient.getCluster(String clusterName)
Get a Cluster object by queueName.
|
JobStatus |
UploadJob.getCommitStatus()
Deprecated.
|
Database |
DLIClient.getDatabase(String databaseName) |
Queue |
DLIClient.getDefaultQueue()
Get default queue,
get a Queue object and queueName is "default"
|
String |
DownloadJob.getDownloadStatus() |
List<String> |
BatchJob.getDriverLog(int from,
int size,
int index) |
abstract List<String> |
SparkJob.getDriverLog(int from,
int size,
int index)
Get a Spark job driver log when job is finished.
|
List<String> |
BatchJob.getDriverLogByJobId(String jobId,
int from,
int size,
int index) |
abstract List<String> |
SparkJob.getDriverLogByJobId(String jobId,
int from,
int size,
int index) |
JobStatus |
SQLJob.getExportJobStatus()
If execute exportResult,you can get exportJob status.
|
com.huawei.dli.restapi.model.FlinkJobApigSinksResponse |
DLIClient.getFlinkApigSinks(Long jobId)
查询作业apig网关服务访问地址
The apig sinks is json format.
|
com.huawei.dli.restapi.model.GetFlinkJobDetailResponse |
DLIClient.getFlinkJobDetail(Long jobId)
查询作业详情
Get job detail information.
|
com.huawei.dli.restapi.model.FlinkJobExecutePlanResponse |
DLIClient.getFlinkJobExecuteGraph(Long jobId)
查询作业执行计划图
The execution graph is json format.
|
com.huawei.dli.restapi.model.QueryFlinkJobListResponse |
DLIClient.getFlinkJobs(String name,
String status,
String queueName,
Boolean showDetail,
Long offset,
Integer limit,
String order,
Long rootJobId,
String username,
String jobType,
String showJobType)
查询作业列表
作业列表查询, 支持以下参数: name, status, show_detail, offset, limit, order, root_job_id.
|
com.huawei.dli.restapi.model.QueryFlinkJobMetricsResponse |
DLIClient.getFlinkJobsMetrics(com.huawei.dli.restapi.model.GetFlinkJobsMetricsBody body)
查询作业监控信息
查询作业监控信息, 支持同时查询多个作业监控信息
|
com.huawei.dli.restapi.model.FlinkJobTemplateListResponse |
DLIClient.getFlinkJobTemplates(Long offset,
Integer limit,
String order)
查询作业模板列表
|
String |
UploadJob.getInternalJobId() |
String |
DownloadJob.getInternalJobId() |
JobProgressResultInfo |
DLIClient.getJobProgress(String jobId)
Obtains the job execution progress based on jobId.
|
List<Row> |
SQLJob.getJobResult()
If job type is "QUEUE",you can get this job result.
|
List<Row> |
SQLJob.getJobResult(Integer pageSize,
Integer currentPage)
If job type is "QUEUE",you can get this job result.
|
List<String> |
BatchJob.getLog(int from,
int size) |
abstract List<String> |
SparkJob.getLog(int from,
int size)
Get a Spark job client log when job is finished.
|
List<String> |
BatchJob.getLogByJobId(String jobId,
int from,
int size) |
abstract List<String> |
SparkJob.getLogByJobId(String jobId,
int from,
int size) |
PartitionSchema |
Table.getPartitionDetail(Integer offset,
Integer limit)
Get describes the partition information for the table in specified offset and limit.
|
PartitionSchema |
Table.getPartitions()
Get describes the partition information for the specified table.
|
Queue |
DLIClient.getQueue(String queueName)
Get a Queue object by queueName.Must be sure that the name exists,
otherwise throw exception:"There is no Queue named"
|
PackageResource |
DLIClient.getResource(String resourceName)
Get a PackageResource object by resourceName.
|
PackageResource |
DLIClient.getResource(String resourceName,
String group)
Get a PackageResource object by resourceName.
|
long |
DownloadJob.getResultCount() |
TableSchema |
SQLJob.getResultSchema() |
int |
ResultSet.getRowCount()
Get row count of the ResultSet
|
int |
SQLJob.getRowCount() |
TableSchema |
ResultSet.getSchema()
Get the Schema of the ResultSet
|
SparkJobStatus |
BatchJob.getStatus() |
JobStatus |
Job.getStatus()
Get the latest job status
|
abstract SparkJobStatus |
SparkJob.getStatus()
Get a Spark job status, it may be starting, running, dead or success.
|
protected JobResultInfo |
Job.getStatusResult(String jobId)
Get this job resultInfo: detail,statement,status...
|
protected JobResultInfo |
SQLJob.getStatusResult(String jobId) |
Table |
Database.getTable(String tableName)
Get the table in the current database by table name.
|
TableSchema |
Table.getTableDetail()
Get describes the metadata information for the specified table.
|
TableSchema |
Database.getTableDetail(String tableName) |
com.huawei.dli.restapi.model.GetTsdbByIdResp |
DLIClient.getTsdbById(String projectId,
Long tsdbId)
查询时序数据库配置信息详情
根据时序数据库ID,查询时序数据库详情
|
com.huawei.dli.restapi.model.GetTsdbListResp |
DLIClient.getTsdbList(String projectId,
Integer offset,
Integer limit)
查询时序数据库配置信息列表
|
long |
UploadJob.getUploadResultCount() |
String |
UploadJob.getUploadStatus() |
List<SparkJobResultInfo> |
DLIClient.listAllBatchJobs(BatchListFilter batchListFilter)
GET /v2.0/{project_id}/batches
Lists the currently user-owned Spark batch jobs.
|
List<SparkJobResultInfo> |
DLIClient.listAllBatchJobs(String clusterName,
int from,
int size)
Lists the currently user-owned Spark batch jobs.
|
List<Cluster> |
DLIClient.listAllClusters()
Lists the currently user-owned Clusters.
|
List<Database> |
DLIClient.listAllDatabases()
List the currently user-owned all databases .
|
List<JobResultInfo> |
DLIClient.listAllJobs()
Get all JobsResultInfo list below the current project.
|
List<JobResultInfo> |
DLIClient.listAllJobs(Integer pageSize,
Integer currentPage,
JobType jobType)
Get all JobsResultInfo list below the current project.
|
List<JobResultInfo> |
DLIClient.listAllJobs(Integer pageSize,
Integer currentPage,
Long start,
Long end,
JobType jobType)
Get all JobsResultInfo list below the current project.
|
List<JobResultInfo> |
DLIClient.listAllJobs(JobFilter jobFilter)
Get all JobsResultInfo list below the current project.
|
List<JobResultInfo> |
DLIClient.listAllJobs(JobType jobType)
Get all JobsResultInfo list below the current project.
|
List<JobResultInfo> |
DLIClient.listAllJobs(Long start,
Long end,
JobType jobType)
Get all JobsResultInfo list below the current project.
|
List<Queue> |
DLIClient.listAllQueues()
Lists the currently user-owned Queues.
|
Resources |
DLIClient.listAllResources()
Lists the currently user-owned Resources.
|
List<JobResultInfo> |
DLIClient.listAllSQLJobs(JobFilter jobFilter)
Query all the information of the recently executed sql job submitted in the edit box under the current project.
|
List<Table> |
Database.listAllTables()
List all tables in the current database.
|
List<Table> |
Database.listAllTables(boolean withDetail)
List all tables in the current database.
|
List<Table> |
Database.listAllTables(String keyword)
List all tables in the current database.
|
List<Table> |
Database.listAllTables(String keyword,
boolean withDetail)
List all tables in the current database.
|
List<Table> |
Database.listAllTables(String keyword,
boolean withDetail,
TableType type)
List all tables in the current database.
|
List<Table> |
Database.listAllTables(TableType type)
List all tables in the current database.
|
List<JobResultInfo> |
DLIClient.listJobsByPage(JobFilter jobFilter)
Get specify page and pageSize JobsResultInfo list below the current project.
|
List<JobResultInfo> |
DLIClient.listSQLJobs()
Query the information of the recently executed job submitted in the edit box under the current project.
|
List<JobResultInfo> |
DLIClient.listSQLJobs(Integer pageSize,
Integer currentPage)
Query the information of the recently executed job submitted in the edit box under the current project.
|
List<JobResultInfo> |
DLIClient.listSQLJobsByPage(JobFilter jobFilter)
Query the information of the recently executed job submitted in the edit box under the current project.
|
PageTables |
Database.listTableByPage(String keyword,
boolean withDetail,
Integer pageSize,
Integer currentPage,
TableType type)
List tables by page in the current database.
|
boolean |
ResultSet.next()
Moves the cursor forward one row from its current position.
|
protected void |
TunnelJob.prepare(Queue queue,
String dbName,
String tblName,
String opType) |
protected void |
TunnelJob.prepare(Queue queue,
TableSchema tblSchema) |
List<Row> |
SQLJob.previewJobResult()
If job type is "QUEUE",you can get this job result
Just get top 1000
|
JobResultInfo |
DLIClient.queryJobResultInfo(String jobId)
Queue the JobResultInfo by job id.
|
com.huawei.dli.restapi.model.QueueActionResponse |
DLIClient.queueAction(String queueName,
com.huawei.dli.restapi.model.QueueActionRequest body)
队列操作
|
Row |
ObsReader.read() |
Row |
Reader.read() |
Row |
JsonObsReader.read() |
Row |
AvroObsReader.read() |
List<com.huawei.dli.restapi.model.GlobalBatchResponse> |
DLIClient.runFlinkJob(com.huawei.dli.restapi.model.RunFlinkJobRequest body)
批量运行作业
触发运行作业
|
com.huawei.dli.restapi.model.PartitionResult |
DLIClient.showPartitions(String databaseName,
String tableName)
Get describes the partition information for the specified table
|
com.huawei.dli.restapi.model.PartitionResult |
DLIClient.showPartitions(String databaseName,
String tableName,
Integer offset,
Integer limit)
Get describes the partition information for the table in specified offset and limit.
|
List<com.huawei.dli.restapi.model.GlobalBatchResponse> |
DLIClient.stopFlinkJob(com.huawei.dli.restapi.model.StopFlinkJobRequest body)
批量停止作业
|
com.huawei.dli.restapi.model.GlobalBatchResponse |
DLIClient.stopSingleFlinkJob(Long jobId,
Boolean isResumeSavepoint)
取消flink自定义作业
|
void |
BatchJob.submit() |
void |
ImportJob.submit()
Submit import job
|
abstract void |
Job.submit() |
void |
SQLJob.submit()
Submit SQL job
|
abstract void |
SparkJob.submit()
Submit a Spark job.
|
void |
ExportJob.submit()
Submit export job.
|
com.huawei.dli.restapi.model.JobStatusResponse |
DLIClient.submitFlinkSqlJob(com.huawei.dli.restapi.model.SubmitFlinkSqlJobRequest body)
提交流式SQL作业到DLI服务
通过POST方式, 提交流式SQL作业, 请求体为JSON格式
|
ResultSet |
SQLJob.submitQuery()
Submit a SQL job, which returns a
ResultSet object. |
com.huawei.dli.restapi.model.JobUpdateResponse |
DLIClient.updateFlinkJarJob(com.huawei.dli.restapi.model.UpdateFlinkJarJobRequest body,
Long jobId)
更新用户自定义Flink作业
目前仅支持Jar格式, 运行在独享集群中
|
com.huawei.dli.restapi.model.GlobalResponse |
DLIClient.updateFlinkJobTemplate(com.huawei.dli.restapi.model.UpdateFlinkJobTemplateRequest body,
Long templateId)
更新作业模板
Update job template on Data lake insight
|
com.huawei.dli.restapi.model.JobUpdateResponse |
DLIClient.updateFlinkSqlJob(com.huawei.dli.restapi.model.UpdateFlinkSqlJobRequest body,
Long jobId)
更新流式SQL作业
流式SQL的语法扩展了Apache Flink SQL, 具体详情请参考DLI官方文档
|
void |
DLIClient.updateResourceOrGroupOwner(com.huawei.dli.restapi.model.UpdateResourceOwner updateResourceOwner)
Update resource or group owner
|
com.huawei.dli.restapi.model.GlobalResponse |
DLIClient.updateTsdb(String projectId,
String tsdbId,
com.huawei.dli.restapi.model.UpdateTsdbReq updateTsdbReq)
修改时序数据库配置信息
修改时序数据库配置信息
|
List<PackageResource> |
DLIClient.uploadResources(String kind,
String[] paths,
String description)
Upload resources sync.
|
List<PackageResource> |
DLIClient.uploadResources(String kind,
String group,
String[] paths,
String description)
Upload resources sync.
|
List<String> |
DLIClient.uploadResourcesAsync(String kind,
String[] paths,
String description)
Upload resources async.
|
List<String> |
DLIClient.uploadResourcesAsync(String kind,
String group,
String[] paths,
String description)
Upload resources async.
|
void |
ObsWriter.write(Row row) |
void |
AvroObsWriter.write(Row row) |
void |
Writer.write(Row row) |
Constructor and Description |
---|
DownloadJob(Queue queue,
DownloadJobParam jobParam)
Create a download job to download data from DLI.
|
DownloadJob(Queue queue,
SQLJob job)
Create a download job to download query result
|
DownloadJob(Queue queue,
SQLJob job,
Integer limit)
Create a download job to download query result
|
DownloadJob(Queue queue,
String dbName,
String tblName)
Create a download job to download data from DLI.
|
DownloadJob(Queue queue,
String dbName,
String tblName,
BadRecordPolicy badRecordPolicy)
Deprecated.
|
DownloadJob(Queue queue,
String dbName,
String tblName,
List<String> partitionSpecs)
Create a download job to download data from DLI.
|
DownloadJob(Queue queue,
String dbName,
String tblName,
PartitionSpec partitionSpec)
Create a download job to download data from DLI.
|
DownloadJob(Queue queue,
String dbName,
String tblName,
PartitionSpec partitionSpec,
BadRecordPolicy badRecordPolicy)
Deprecated.
|
DownloadJob(Queue queue,
String dbName,
String tblName,
String partitionSpec) |
DownloadJob(Queue queue,
String dbName,
String tblName,
String partitionSpec,
BadRecordPolicy badRecordPolicy)
Deprecated.
|
UploadJob(Queue queue,
String dbName,
String tableName)
Deprecated.
|
UploadJob(Queue queue,
String dbName,
String tableName,
Boolean isOverwrite)
Create a upload job to upload data to DLI.
|
UploadJob(Queue queue,
String dbName,
String tableName,
PartitionSpec partitionSpec,
Boolean isOverwrite)
Create a upload job to upload data to DLI.
|
UploadJob(Queue queue,
String dbName,
String tableName,
String partitionSpec,
Boolean isOverwrite) |
Modifier and Type | Method and Description |
---|---|
String |
Authentication.getAuthToken() |
Modifier and Type | Method and Description |
---|---|
boolean |
TableSchema.containsColumn(String name) |
com.huawei.dli.restapi.ApiResponse<V> |
IHttpExecAdpter.exec() |
DliArray |
Row.getArray(int columnIdx) |
BigDecimal |
Row.getBigDecimal(int columnIdx) |
BigInteger |
Row.getBigInteger(int columnIdx) |
Boolean |
Row.getBoolean(int columnIdx) |
Column |
TableSchema.getColumn(String name) |
int |
TableSchema.getColumnIdx(String name) |
String |
Row.getCsvString() |
Date |
Row.getDate(int columnIdx) |
Double |
Row.getDouble(int columnIdx) |
Float |
Row.getFloat(int columnIdx) |
Integer |
Row.getInt(int columnIdx) |
Long |
Row.getLong(int columnIdx) |
DliMap |
Row.getMap(int columnIdx) |
String |
Row.getRowString() |
Short |
Row.getShort(int columnIdx) |
String |
Row.getString(int columnIdx) |
DliStruct |
Row.getStruct(int columnIdx) |
Timestamp |
Row.getTimestamp(int columnIdx) |
Byte |
Row.getTinyint(int columnIdx) |
void |
Row.setArray(int index,
DliArray data) |
void |
Row.setBigDecimal(int index,
BigDecimal data) |
void |
Row.setBigInteger(int index,
BigInteger data) |
void |
Row.setBoolean(int index,
Boolean data) |
void |
Row.setDate(int index,
Date data) |
void |
Row.setDouble(int index,
Double data) |
void |
Row.setFloat(int index,
Float data) |
void |
Row.setInt(int index,
Integer data) |
void |
Row.setLong(int index,
Long data) |
void |
Row.setMap(int index,
DliMap data) |
void |
Row.setShort(int index,
Short data) |
void |
Row.setString(int index,
String data) |
void |
Row.setStruct(int index,
DliStruct data) |
void |
Row.setTimestamp(int index,
Timestamp data) |
void |
Row.setTinyint(int index,
Byte tinyintData) |
Modifier and Type | Method and Description |
---|---|
List<Object> |
DliArray.getElementValues()
Get the list of the elements in the Array
|
List<Object> |
SimpleArray.getElementValues() |
int |
DliStruct.getFieldCount()
Get the field number of this struct.
|
int |
SimpleStruct.getFieldCount() |
Object |
DliStruct.getFieldValue(int index)
Get the field value by the index of the struct
|
Object |
SimpleStruct.getFieldValue(int index) |
Object |
DliStruct.getFieldValue(String fieldName)
Get the field value by the field name of the struct
|
Object |
SimpleStruct.getFieldValue(String fieldName) |
List<Object> |
DliStruct.getFieldValues()
Get the List of field values of the struct
|
List<Object> |
SimpleStruct.getFieldValues() |
Map |
SimpleMap.getValues() |
Map |
DliMap.getValues()
Get the key and value of the Map
|
static Object |
TypeUtils.transform(String data,
DataTypeInfo valueTypeInfo)
Transform the data value from string to real data type in java
|
Copyright © 2023. All rights reserved.