Skip to content

Commit 6c95381

Browse files
authored
Merge pull request #1129 from WeDataSphere/master
Update branch-1.2.1
2 parents 225c39d + bb85d7f commit 6c95381

File tree

398 files changed

+17831
-1139
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

398 files changed

+17831
-1139
lines changed

.gitignore

+3
Original file line numberDiff line numberDiff line change
@@ -101,6 +101,9 @@ dss-apps/dss-data-governance/dss-data-warehouse-dao/target
101101
dss-apps/dss-data-governance/dss-data-warehouse-service/target
102102
dss-apps/dss-data-governance/dss-data-warehouse-server/target
103103

104+
#dss-git
105+
dss-git/dss-git-common/target
106+
dss-git/dss-git-server/target
104107

105108
# plugins
106109
plugins/azkaban/linkis-jobtype/target

assembly/dss-package/pom.xml

+1-1
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
<parent>
2222
<artifactId>dss</artifactId>
2323
<groupId>com.webank.wedatasphere.dss</groupId>
24-
<version>1.1.0.20-SNAPSHOT</version>
24+
<version>1.5.0-SNAPSHOT</version>
2525
<relativePath>../../pom.xml</relativePath>
2626
</parent>
2727
<modelVersion>4.0.0</modelVersion>

assembly/pom.xml

+1-1
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@
2222
<parent>
2323
<artifactId>dss</artifactId>
2424
<groupId>com.webank.wedatasphere.dss</groupId>
25-
<version>1.1.0.20-SNAPSHOT</version>
25+
<version>1.5.0-SNAPSHOT</version>
2626
<relativePath>../pom.xml</relativePath>
2727
</parent>
2828
<packaging>pom</packaging>

conf/dss-framework-orchestrator-server.properties

-1
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,6 @@ wds.linkis.server.mybatis.typeAliasesPackage=com.webank.wedatasphere.dss.server.
3232

3333
wds.linkis.server.mybatis.BasePackage=com.webank.wedatasphere.dss.framework.appconn.dao,com.webank.wedatasphere.dss.orchestrator.core.dao,com.webank.wedatasphere.dss.server.dao,com.webank.wedatasphere.dss.application.dao,com.webank.wedatasphere.dss.workspace.mapper,com.webank.wedatasphere.dss.workspace.common.dao,com.webank.wedatasphere.dss.workspace.common.dao,com.webank.wedatasphere.dss.orchestrator.db.dao,com.webank.wedatasphere.dss.workflow.dao,com.webank.wedatasphere.dss.framework.appconn.dao,com.webank.wedatasphere.dss.flow.execution.entrance.dao
3434

35-
wds.dss.server.scheduling.clear.cs.cron=0 0 3 * * ?
3635

3736
wds.dss.publish.max.remain.version=3
3837

db/dss_ddl.sql

+33
Original file line numberDiff line numberDiff line change
@@ -51,6 +51,7 @@ CREATE TABLE `dss_orchestrator_info` (
5151
`orchestrator_level` varchar(32) DEFAULT NULL COMMENT '工作流级别',
5252
`update_user` varchar(100) DEFAULT NULL COMMENT '更新人',
5353
`update_time` datetime DEFAULT CURRENT_TIMESTAMP COMMENT '更新时间',
54+
`status` VARCHAR(64),
5455
PRIMARY KEY (`id`) USING BTREE,
5556
UNIQUE KEY `unique_idx_uuid` (`uuid`)
5657
) ENGINE=InnoDB AUTO_INCREMENT=326 DEFAULT CHARSET=utf8mb4 ROW_FORMAT=COMPACT;
@@ -70,6 +71,7 @@ CREATE TABLE `dss_orchestrator_version_info` (
7071
`content` varchar(255) DEFAULT NULL,
7172
`context_id` varchar(200) DEFAULT NULL COMMENT '上下文ID',
7273
`valid_flag` INT(1) DEFAULT '1' COMMENT '版本有效标示,0:无效;1:有效',
74+
`commit_id` varchar(64),
7375
PRIMARY KEY (`id`) USING BTREE
7476
) ENGINE=InnoDB AUTO_INCREMENT=422 DEFAULT CHARSET=utf8mb4 ROW_FORMAT=COMPACT;
7577

@@ -116,6 +118,7 @@ CREATE TABLE `dss_project` (
116118
`dev_process` varchar(200) COLLATE utf8_bin DEFAULT NULL COMMENT '开发流程,多个以英文逗号分隔,取得的值是dss_workspace_dictionary中的dic_key(parent_key=p_develop_process)',
117119
`orchestrator_mode` varchar(200) COLLATE utf8_bin DEFAULT NULL COMMENT '编排模式,多个以英文逗号分隔,取得的值是dss_workspace_dictionary中的dic_key(parent_key=p_arrangement_mode或下面一级)',
118120
`visible` tinyint(4) DEFAULT '1' COMMENT '0:已删除;1:未删除(默认)',
121+
`associate_git` TINYINT DEFAULT '0' COMMENT '0:未接入git,1:已接入git',
119122
PRIMARY KEY (`id`)
120123
) ENGINE=InnoDB AUTO_INCREMENT=313 DEFAULT CHARSET=utf8 COLLATE=utf8_bin ROW_FORMAT=COMPACT;
121124

@@ -639,3 +642,33 @@ key `idx_limit_name` (`limit_name`)
639642
) ENGINE = InnoDB
640643
DEFAULT CHARSET = utf8mb4 COLLATE=utf8mb4_bin COMMENT ='dss用户限制表';
641644

645+
DROP TABLE IF EXISTS `dss_workspace_associate_git`;
646+
CREATE TABLE `dss_workspace_associate_git` (
647+
`id` bigint(20) NOT NULL AUTO_INCREMENT,
648+
`workspace_id` bigint(20) DEFAULT NULL,
649+
`git_user` varchar(64) DEFAULT NULL COMMENT 'git登录用户名',
650+
`git_password` VARCHAR(255) DEFAULT NULL COMMENT 'git登录密码,用于跳转',
651+
`git_token` varchar(255) COMMENT '用户配置的git token',
652+
`git_url` varchar(255),
653+
`create_time` datetime DEFAULT NULL,
654+
`update_time` datetime DEFAULT NULL,
655+
`create_by` varchar(128) DEFAULT NULL,
656+
`update_by` varchar(128) DEFAULT NULL,
657+
`type` varchar(32) DEFAULT NULL,
658+
`git_user_id` varchar(32) DEFAULT NULL,
659+
PRIMARY KEY (`id`)
660+
) ENGINE=InnoDB DEFAULT CHARSET=utf8mb4 COLLATE=utf8mb4_bin COMMENT='工作空间绑定的git信息';
661+
662+
663+
DROP TABLE IF EXISTS `dss_orchestrator_submit_job_info`;
664+
CREATE TABLE `dss_orchestrator_submit_job_info` (
665+
`id` bigint(20) NOT NULL AUTO_INCREMENT COMMENT '主键',
666+
`orchestrator_id` bigint(20) NOT NULL,
667+
`create_time` timestamp NOT NULL DEFAULT CURRENT_TIMESTAMP COMMENT '创建时间',
668+
`update_time` timestamp NOT NULL DEFAULT CURRENT_TIMESTAMP COMMENT '修改时间',
669+
`instance_name` varchar(128) DEFAULT NULL COMMENT '提交任务的实例',
670+
`status` varchar(128) DEFAULT NULL COMMENT '提交任务状态',
671+
`error_msg` varchar(2048) DEFAULT NULL COMMENT '提交任务异常信息',
672+
PRIMARY KEY (`id`)
673+
) ENGINE=InnoDB DEFAULT CHARSET=utf8mb4 COLLATE=utf8mb4_bin COMMENT='dss_orchestrator_submit_job_info表';
674+

dss-appconn/appconns/dss-datachecker-appconn/pom.xml

+2-2
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
<parent>
2222
<artifactId>dss</artifactId>
2323
<groupId>com.webank.wedatasphere.dss</groupId>
24-
<version>1.1.0.20-SNAPSHOT</version>
24+
<version>1.5.0-SNAPSHOT</version>
2525
<relativePath>../../../pom.xml</relativePath>
2626
</parent>
2727
<modelVersion>4.0.0</modelVersion>
@@ -54,7 +54,7 @@
5454
<dependency>
5555
<groupId>com.webank.wedatasphere.dss</groupId>
5656
<artifactId>dss-origin-sso-integration-standard</artifactId>
57-
<version>1.1.0.20-SNAPSHOT</version>
57+
<version>1.5.0-SNAPSHOT</version>
5858
<exclusions>
5959
<exclusion>
6060
<groupId>org.apache.linkis</groupId>

dss-appconn/appconns/dss-datachecker-appconn/src/main/java/com/webank/wedatasphere/dss/appconn/datachecker/DataChecker.java

+1-1
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ public DataChecker(Properties p, DataCheckerExecutionAction action) {
4848
maxWaitTime = Long.valueOf(p.getProperty(DataChecker.WAIT_TIME, "1")) * 3600 * 1000;
4949
//test over time
5050
// maxWaitTime = Long.valueOf(p.getProperty(DataChecker.WAIT_TIME, "1")) * 120 * 1000;
51-
queryFrequency = Integer.valueOf(p.getProperty(DataChecker.QUERY_FREQUENCY, "30000"));
51+
queryFrequency = Integer.valueOf(p.getProperty(DataChecker.QUERY_FREQUENCY, "60000"));
5252

5353
}
5454

dss-appconn/appconns/dss-datachecker-appconn/src/main/java/com/webank/wedatasphere/dss/appconn/datachecker/connector/DataCheckerDao.java

+3-60
Original file line numberDiff line numberDiff line change
@@ -54,13 +54,6 @@ public class DataCheckerDao {
5454
private static final String SQL_SOURCE_TYPE_JOB_PARTITION =
5555
"SELECT * FROM DBS d JOIN TBLS t ON t.DB_ID = d.DB_ID JOIN PARTITIONS p ON p.TBL_ID = t.TBL_ID WHERE d.NAME=? AND t.TBL_NAME=? AND p.PART_NAME=?";
5656

57-
private static final String SQL_SOURCE_TYPE_BDP =
58-
"SELECT * FROM desktop_bdapimport WHERE bdap_db_name = ? AND bdap_table_name = ? AND target_partition_name = ? AND status = '1';";
59-
60-
private static final String SQL_SOURCE_TYPE_BDP_WITH_TIME_CONDITION =
61-
"SELECT * FROM desktop_bdapimport WHERE bdap_db_name = ? AND bdap_table_name = ? AND target_partition_name = ? " +
62-
"AND (UNIX_TIMESTAMP() - UNIX_TIMESTAMP(STR_TO_DATE(modify_time, '%Y-%m-%d %H:%i:%s'))) <= ? AND status = '1';";
63-
6457
private static final String SQL_DOPS_CHECK_TABLE =
6558
"SELECT * FROM dops_clean_task_list WHERE db_name = ? AND tb_name = ? AND part_name is null AND task_state NOT IN (10,13) order by order_id desc limit 1";
6659
private static final String SQL_DOPS_CHECK_PARTITION =
@@ -72,7 +65,6 @@ public class DataCheckerDao {
7265
private static final String MASK_SOURCE_TYPE = "maskdb";
7366

7467
private static DataSource jobDS;
75-
private static DataSource bdpDS;
7668

7769
private static DataSource dopsDS;
7870
private static volatile DataCheckerDao instance;
@@ -96,13 +88,6 @@ public boolean validateTableStatusFunction(Properties props, Logger log, DataChe
9688
return false;
9789
}
9890
}
99-
if (bdpDS == null) {
100-
bdpDS = DataDruidFactory.getBDPInstance(props, log);
101-
if (bdpDS == null) {
102-
log.warn("Error getting job Druid DataSource instance");
103-
return false;
104-
}
105-
}
10691
boolean systemCheck = Boolean.valueOf(props.getProperty(DataChecker.QUALITIS_SWITCH));
10792
if (systemCheck && dopsDS == null) {
10893
dopsDS = DataDruidFactory.getDopsInstance(props, log);//通过alibaba的druid数据库连接池获取JOB数据库连接
@@ -122,7 +107,7 @@ public boolean validateTableStatusFunction(Properties props, Logger log, DataChe
122107
}
123108
log.info("(DataChecker info) database table partition info : " + dataCheckerInfo);
124109
long waitTime = Long.valueOf(props.getProperty(DataChecker.WAIT_TIME, "1")) * 3600 * 1000;
125-
int queryFrequency = Integer.valueOf(props.getProperty(DataChecker.QUERY_FREQUENCY, "30000"));
110+
int queryFrequency = Integer.valueOf(props.getProperty(DataChecker.QUERY_FREQUENCY, "60000"));
126111
// String timeScape = props.getProperty(DataChecker.TIME_SCAPE, "NULL");
127112
log.info("(DataChecker info) wait time : " + waitTime);
128113
log.info("(DataChecker info) query frequency : " + queryFrequency);
@@ -134,13 +119,12 @@ public boolean validateTableStatusFunction(Properties props, Logger log, DataChe
134119
});
135120
QualitisUtil qualitisUtil = new QualitisUtil(props);
136121
try (Connection jobConn = jobDS.getConnection();
137-
Connection bdpConn = bdpDS.getConnection();
138122
Connection dopsConn = dopsDS != null ? dopsDS.getConnection() : null) {
139123
List<Boolean> allCheckRes = dataObjectList
140124
.parallelStream()
141125
.map(proObjectMap -> {
142126
log.info("Begin to Check dataObject:" + proObjectMap.entrySet().toString());
143-
boolean checkRes = getDataCheckResult(proObjectMap, jobConn, bdpConn, dopsConn, props, log,action,qualitisUtil);
127+
boolean checkRes = getDataCheckResult(proObjectMap, jobConn, dopsConn, props, log,action,qualitisUtil);
144128
if (null != action.getExecutionRequestRefContext()) {
145129
if (checkRes) {
146130
action.getExecutionRequestRefContext().appendLog("Database table partition info : " + proObjectMap.get(DataChecker.DATA_OBJECT) + " has arrived");
@@ -178,7 +162,6 @@ public boolean validateTableStatusFunction(Properties props, Logger log, DataChe
178162

179163
private boolean getDataCheckResult(Map<String, String> proObjectMap,
180164
Connection jobConn,
181-
Connection bdpConn,
182165
Connection dopsConn,
183166
Properties props,
184167
Logger log,
@@ -231,7 +214,7 @@ private boolean getDataCheckResult(Map<String, String> proObjectMap,
231214
}
232215
log.info("start to check maskis");
233216
proObjectMap.put(DataChecker.SOURCE_TYPE, MASK_SOURCE_TYPE);
234-
normalCheck= (getBdpTotalCount(dataObject, bdpConn, log, props) > 0 || "success".equals(fetchMaskCode(dataObject, log, props).get("maskStatus")));
217+
normalCheck= "success".equals(fetchMaskCode(dataObject, log, props).get("maskStatus"));
235218
if (null != action.getExecutionRequestRefContext()){
236219
action.getExecutionRequestRefContext().appendLog(dataObjectStr+" check maskis end,check result:"+normalCheck);
237220
}
@@ -316,25 +299,6 @@ private PreparedStatement getJobStatement(Connection conn, CheckDataObject dataO
316299
}
317300
}
318301

319-
/**
320-
* 构造查询maskis的查询
321-
*/
322-
private PreparedStatement getBdpStatement(Connection conn, CheckDataObject dataObject, String timeScape) throws SQLException {
323-
PreparedStatement pstmt = null;
324-
if (timeScape.equals("NULL")) {
325-
pstmt = conn.prepareCall(SQL_SOURCE_TYPE_BDP);
326-
} else {
327-
pstmt = conn.prepareCall(SQL_SOURCE_TYPE_BDP_WITH_TIME_CONDITION);
328-
pstmt.setInt(4, Integer.valueOf(timeScape) * 3600);
329-
}
330-
if (dataObject.getPartitionName() == null) {
331-
dataObject.setPartitionName("");
332-
}
333-
pstmt.setString(1, dataObject.getDbName());
334-
pstmt.setString(2, dataObject.getTableName());
335-
pstmt.setString(3, dataObject.getPartitionName());
336-
return pstmt;
337-
}
338302

339303
/**
340304
* 构造查询dops库的查询
@@ -414,27 +378,6 @@ private long getJobTotalCount(CheckDataObject dataObject, Connection conn, Logge
414378
}
415379
}
416380

417-
/**
418-
* 查mask db
419-
*/
420-
private long getBdpTotalCount(CheckDataObject dataObject, Connection conn, Logger log, Properties props) {
421-
String timeScape = props.getOrDefault(DataChecker.TIME_SCAPE, "NULL").toString();
422-
log.info("-------------------------------------- search bdp data ");
423-
log.info("-------------------------------------- dataObject: " + dataObject.toString());
424-
try (PreparedStatement pstmt = getBdpStatement(conn, dataObject, timeScape)) {
425-
ResultSet rs = pstmt.executeQuery();
426-
long ret = 0L;
427-
while (rs.next()) {
428-
ret ++;
429-
}
430-
// long ret=rs.last() ? rs.getRow() : 0;
431-
log.info("-------------------------------------- bdp data result:"+ret);
432-
return ret;
433-
} catch (SQLException e) {
434-
log.error("fetch data from bdp error", e);
435-
return 0;
436-
}
437-
}
438381

439382
/**
440383
* - 返回0表示未找到任何记录 ;

dss-appconn/appconns/dss-datachecker-appconn/src/main/java/com/webank/wedatasphere/dss/appconn/datachecker/connector/DataDruidFactory.java

+4-1
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,9 @@
2525

2626
public class DataDruidFactory {
2727
private static volatile DruidDataSource jobInstance;
28-
private static volatile DruidDataSource bdpInstance;
28+
29+
private static volatile DruidDataSource dopsInstance;
30+
2931
private static volatile DruidDataSource msgInstance;
3032

3133
public static DruidDataSource getJobInstance(Properties props, Logger log) {
@@ -42,6 +44,7 @@ public static DruidDataSource getJobInstance(Properties props, Logger log) {
4244
}
4345
return jobInstance;
4446
}
47+
4548
public static DruidDataSource getBDPInstance(Properties props, Logger log) {
4649
if (bdpInstance == null ) {
4750
synchronized (DataDruidFactory.class) {

dss-appconn/appconns/dss-dolphinscheduler-appconn/pom.xml

+2-3
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66
<parent>
77
<artifactId>dss</artifactId>
88
<groupId>com.webank.wedatasphere.dss</groupId>
9-
<version>1.1.0.20-SNAPSHOT</version>
9+
<version>1.5.0-SNAPSHOT</version>
1010
<relativePath>../../../pom.xml</relativePath>
1111
</parent>
1212
<modelVersion>4.0.0</modelVersion>
@@ -18,7 +18,6 @@
1818
<groupId>com.webank.wedatasphere.dss</groupId>
1919
<artifactId>dss-scheduler-appconn</artifactId>
2020
<version>${dss.version}</version>
21-
2221
<scope>provided</scope>
2322
</dependency>
2423

@@ -81,7 +80,7 @@
8180
<dependency>
8281
<groupId>com.google.guava</groupId>
8382
<artifactId>guava</artifactId>
84-
<version>28.2-android</version>
83+
<version>33.1.0-jre</version>
8584
</dependency>
8685

8786

dss-appconn/appconns/dss-eventchecker-appconn/pom.xml

+1-1
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
<parent>
2222
<artifactId>dss</artifactId>
2323
<groupId>com.webank.wedatasphere.dss</groupId>
24-
<version>1.1.0.20-SNAPSHOT</version>
24+
<version>1.5.0-SNAPSHOT</version>
2525
<relativePath>../../../pom.xml</relativePath>
2626
</parent>
2727
<modelVersion>4.0.0</modelVersion>

dss-appconn/appconns/dss-eventchecker-appconn/src/main/java/com/webank/wedatasphere/dss/appconn/eventchecker/entity/EventChecker.java

+3-3
Original file line numberDiff line numberDiff line change
@@ -67,10 +67,10 @@ public EventChecker(Properties p, EventCheckerExecutionAction action) {
6767
String waitTime = p.getProperty(EventChecker.WAIT_TIME, "1");
6868
Double doubleWaitTime = Double.valueOf(waitTime) * 3600 * 1000;
6969
maxWaitTime = Long.valueOf(doubleWaitTime.longValue());
70-
String query_frequency = p.getProperty(EventChecker.QUERY_FREQUENCY, "30000");
70+
String query_frequency = p.getProperty(EventChecker.QUERY_FREQUENCY, "60000");
7171
queryFrequency = Integer.valueOf(query_frequency);
72-
if(queryFrequency <10000){
73-
queryFrequency = 10000;
72+
if(queryFrequency <60000){
73+
queryFrequency = 60000;
7474
}
7575
}
7676

dss-appconn/appconns/dss-eventchecker-appconn/src/main/java/com/webank/wedatasphere/dss/appconn/eventchecker/service/AbstractEventCheck.java

+1-1
Original file line numberDiff line numberDiff line change
@@ -70,7 +70,7 @@ void initECParams(Properties props){
7070
runDate = props.getProperty("run_date");
7171
userTime = props.getProperty(EventChecker.USER_TIME);
7272
waitTime = props.getProperty(EventChecker.WAIT_TIME, "1");
73-
query_frequency = props.getProperty(EventChecker.QUERY_FREQUENCY, "30000");
73+
query_frequency = props.getProperty(EventChecker.QUERY_FREQUENCY, "60000");
7474
afterSend = props.getProperty(EventChecker.AFTERSEND);
7575
}
7676

dss-appconn/appconns/dss-schedulis-appconn/pom.xml

+1-1
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
<parent>
2222
<artifactId>dss</artifactId>
2323
<groupId>com.webank.wedatasphere.dss</groupId>
24-
<version>1.1.0.20-SNAPSHOT</version>
24+
<version>1.5.0-SNAPSHOT</version>
2525
<relativePath>../../../pom.xml</relativePath>
2626
</parent>
2727
<modelVersion>4.0.0</modelVersion>

dss-appconn/appconns/dss-schedulis-appconn/src/main/java/com/webank/wedatasphere/dss/appconn/schedulis/constant/AzkabanConstant.java

+1
Original file line numberDiff line numberDiff line change
@@ -29,5 +29,6 @@ public class AzkabanConstant {
2929
public final static String FLOW_CONTEXT_ID = "wds.linkis.flow.contextID=";
3030
public final static String LINKIS_VERSION = "linkis.version";
3131
public final static String JOB_COMMENT = "comment";
32+
public final static String AUTO_DISABLED = "auto.disabled";
3233

3334
}

dss-appconn/appconns/dss-schedulis-appconn/src/main/java/com/webank/wedatasphere/dss/appconn/schedulis/linkisjob/LinkisJob.java

+10
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,8 @@ public class LinkisJob {
2323
private String type;
2424
private String linkistype;
2525
private String proxyUser;
26+
27+
private String autoDisabled;
2628
private String dependencies;
2729
private Map<String, String> conf;
2830
private String command;
@@ -60,6 +62,14 @@ public void setProxyUser(String proxyUser) {
6062
this.proxyUser = proxyUser;
6163
}
6264

65+
public String getAutoDisabled() {
66+
return autoDisabled;
67+
}
68+
69+
public void setAutoDisabled(String autoDisabled) {
70+
this.autoDisabled = autoDisabled;
71+
}
72+
6373
public String getDependencies() {
6474
return dependencies;
6575
}

dss-appconn/appconns/dss-schedulis-appconn/src/main/java/com/webank/wedatasphere/dss/appconn/schedulis/linkisjob/LinkisJobConverter.java

+3-1
Original file line numberDiff line numberDiff line change
@@ -70,6 +70,7 @@ private String convertJobToString(LinkisJob job){
7070
map.put(WorkflowConstant.PROXY_USER,job.getProxyUser());
7171
map.put(AzkabanConstant.JOB_COMMAND,job.getCommand());
7272
map.put(AzkabanConstant.JOB_COMMENT,job.getComment());
73+
map.put(AzkabanConstant.AUTO_DISABLED,job.getAutoDisabled());
7374
Map<String, Object> labels = new HashMap<>(1);
7475
labels.put("route", SchedulerConf.JOB_LABEL.getValue());
7576
map.put(AzkabanConstant.JOB_LABELS, DSSCommonUtils.COMMON_GSON.toJson(labels));
@@ -114,7 +115,8 @@ private void convertConfiguration(WorkflowNode workflowNode, LinkisJob job){
114115
configuration.forEach((k,v)-> {
115116
if(null!=v) {
116117
v.forEach((k2, v2) -> {
117-
if(null !=v2) {job.getConf().put(confprefix + k + "." + k2, v2.toString());}
118+
if(AzkabanConstant.AUTO_DISABLED.equals(k2) && null !=v2){job.setAutoDisabled(v2.toString());}
119+
else if(null !=v2) {job.getConf().put(confprefix + k + "." + k2, v2.toString());}
118120
});
119121
}
120122
});

dss-appconn/appconns/dss-schedulis-appconn/src/main/java/com/webank/wedatasphere/dss/appconn/schedulis/operation/SchedulisProjectSearchOperation.java

+1
Original file line numberDiff line numberDiff line change
@@ -28,6 +28,7 @@ public ProjectResponseRef searchProject(RefProjectContentRequestRef.RefProjectCo
2828
params.put("project", requestRef.getProjectName());
2929
params.put("ajax", "fetchprojectflows");
3030
try {
31+
logger.info("request url from Schedulis is: {}.", queryUrl);
3132
String responseBody = SchedulisHttpUtils.getHttpGetResult(queryUrl, params, ssoRequestOperation, requestRef.getWorkspace());
3233
logger.info("responseBody from Schedulis is: {}.", responseBody);
3334
Map<String,Object> map = DSSCommonUtils.COMMON_GSON.fromJson(responseBody, new TypeToken<Map<String,Object>>(){}.getType());

0 commit comments

Comments
 (0)