Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,7 @@ The platform is based on plug-in design, and the following modules support user-

- **Data Source**: `MySQL`, `Impala`, `StarRocks`, `Doris`, `Presto`, `Trino`, `ClickHouse`, `PostgreSQL` are already supported
- **Check Rules**: 27 check rules such as built-in null value check, non-null check, enumeration check, etc.
- **Job Execution Engine**: Two execution engines `Spark` and `Local` have been supported. The `Spark` engine currently only supports the `Spark2.4` version, and the `Local` engine is a local execution engine developed based on `JDBC`, without relying on other execution engines.
- **Job Execution Engine**: Two execution engines `Spark` and `Local` have been supported. The `Spark` engine currently supports the `Spark3.4.x` version, and the `Local` engine is a local execution engine developed based on `JDBC`, without relying on other execution engines.
- **Alert Channel**: Supported **Email**
- **Error Data Storage**: `MySQL` and **local files** are already supported (only `Local` execution engine is supported)
- **Registry**: Already supports `MySQL`, `PostgreSQL` and `ZooKeeper`
Expand Down
2 changes: 1 addition & 1 deletion README.zh-CN.md
Original file line number Diff line number Diff line change
Expand Up @@ -72,7 +72,7 @@ $ mvn clean package -Prelease -DskipTests

- **数据源**:已支持 `MySQL`、`Impala`、`StarRocks`、`Doris`、`Presto`、`Trino`、`ClickHouse`、`PostgreSQL`
- **检查规则**:内置空值检查、非空检查、枚举检查等27个检查规则
- **作业执行引擎**:已支持`Spark`和`Local`两种执行引擎。`Spark `引擎目前仅支持`Spark2.4`版本,`Local` 引擎则是基于`JDBC`开发的本地执行引擎,无需依赖其他执行引擎。
- **作业执行引擎**:已支持`Spark`和`Local`两种执行引擎。`Spark `引擎目前支持`Spark3.4.x`版本,`Local` 引擎则是基于`JDBC`开发的本地执行引擎,无需依赖其他执行引擎。
- **告警通道**:已支持**邮件**
- **错误数据存储**:已支持 `MySQL` 和 **本地文件**(仅支持`Local`执行引擎)
- **注册中心**:已支持 `MySQL`、`PostgreSQL` 和 `ZooKeeper`
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -19,11 +19,9 @@
public enum SparkVersion {

/**
* 0 SPARK1
* 1 SPARK2
* 0 SPARK3
*/
SPARK2(0, "SPARK2"),
SPARK3(1, "SPARK3");
SPARK3(0, "SPARK3");

SparkVersion(int code, String description) {
this.code = code;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -34,9 +34,9 @@
</modules>

<properties>
<scala.binary.version>2.11</scala.binary.version>
<spark.version>2.4.0</spark.version>
<jackson.version>2.9.0</jackson.version>
<scala.binary.version>2.12</scala.binary.version>
<spark.version>3.4.4</spark.version>
<jackson.version>2.14.2</jackson.version>
<spark.scope>provided</spark.scope>
</properties>

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -32,9 +32,9 @@
<artifactId>datavines-engine-spark-api</artifactId>

<properties>
<scala.binary.version>2.11</scala.binary.version>
<spark.version>2.4.0</spark.version>
<jackson.version>2.9.0</jackson.version>
<scala.binary.version>2.12</scala.binary.version>
<spark.version>3.4.4</spark.version>
<jackson.version>2.14.2</jackson.version>
<scope>provided</scope>
</properties>

Expand All @@ -43,6 +43,7 @@
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-streaming_${scala.binary.version}</artifactId>
<version>${spark.version}</version>
<scope>${scope}</scope>
</dependency>

Expand All @@ -53,7 +54,7 @@
<scope>${scope}</scope>
<exclusions>
<exclusion>
<artifactId>jackson-module-scala_2.11</artifactId>
<artifactId>jackson-module-scala_${scala.binary.version}</artifactId>
<groupId>com.fasterxml.jackson.module</groupId>
</exclusion>
</exclusions>
Expand All @@ -76,24 +77,6 @@
<groupId>org.apache.spark</groupId>
<artifactId>spark-hive_${scala.binary.version}</artifactId>
<version>${spark.version}</version>
<exclusions>
<exclusion>
<groupId>commons-httpclient</groupId>
<artifactId>commons-httpclient</artifactId>
</exclusion>
<exclusion>
<groupId>org.apache.httpcomponents</groupId>
<artifactId>httpclient</artifactId>
</exclusion>
<exclusion>
<artifactId>jackson-core-asl</artifactId>
<groupId>org.codehaus.jackson</groupId>
</exclusion>
<exclusion>
<artifactId>jackson-mapper-asl</artifactId>
<groupId>org.codehaus.jackson</groupId>
</exclusion>
</exclusions>
<scope>${scope}</scope>
</dependency>

Expand All @@ -112,7 +95,7 @@

<dependency>
<groupId>com.fasterxml.jackson.module</groupId>
<artifactId>jackson-module-scala_2.11</artifactId>
<artifactId>jackson-module-scala_${scala.binary.version}</artifactId>
<version>${jackson.version}</version>
<scope>${scope}</scope>
<exclusions>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -37,13 +37,14 @@
import io.datavines.engine.executor.core.executor.ShellCommandProcess;
import io.datavines.engine.spark.executor.parameter.SparkArgsUtils;
import io.datavines.engine.spark.executor.parameter.SparkParameters;
import io.datavines.engine.spark.executor.parameter.SparkVersion;

public class SparkEngineExecutor extends AbstractYarnEngineExecutor {

/**
* spark2 command
* spark command
*/
private static final String SPARK2_COMMAND = "${SPARK_HOME2}/bin/spark-submit";
private static final String SPARK_COMMAND = "${SPARK_HOME}/bin/spark-submit";

private Configurations configurations;

Expand All @@ -64,7 +65,7 @@ public void init(JobExecutionRequest jobExecutionRequest, Logger logger, Configu
public void execute() throws Exception {
try {
this.processResult = shellCommandProcess.run(buildCommand());
logger.info("process result: "+ JSONUtils.toJsonString(this.processResult));
logger.info("process result: {}", JSONUtils.toJsonString(this.processResult));
} catch (Exception e) {
logger.error("yarn process failure", e);
throw e;
Expand Down Expand Up @@ -139,7 +140,7 @@ protected String buildCommand() {

List<String> args = new ArrayList<>();

args.add(SPARK2_COMMAND);
args.add(SPARK_COMMAND);

args.addAll(SparkArgsUtils.buildArgs(sparkParameters));

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -19,11 +19,9 @@
public enum SparkVersion {

/**
* 0 SPARK1
* 1 SPARK2
* 0 SPARK3
*/
SPARK2(0, "SPARK2"),
SPARK3(1, "SPARK3");
SPARK3(0, "SPARK3");

SparkVersion(int code, String description) {
this.code = code;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -42,9 +42,9 @@
</modules>

<properties>
<scala.binary.version>2.11</scala.binary.version>
<spark.version>2.4.0</spark.version>
<jackson.version>2.9.0</jackson.version>
<scala.binary.version>2.12</scala.binary.version>
<spark.version>3.4.4</spark.version>
<jackson.version>2.14.2</jackson.version>
<spark.scope>provided</spark.scope>
</properties>

Expand All @@ -64,7 +64,7 @@
<scope>${spark.scope}</scope>
<exclusions>
<exclusion>
<artifactId>jackson-module-scala_2.11</artifactId>
<artifactId>jackson-module-scala_${scala.binary.version}</artifactId>
<groupId>com.fasterxml.jackson.module</groupId>
</exclusion>
</exclusions>
Expand All @@ -86,32 +86,14 @@
<dependency>
<groupId>org.mongodb.spark</groupId>
<artifactId>mongo-spark-connector_${scala.binary.version}</artifactId>
<version>${spark.version}</version>
<version>10.2.0</version>
<scope>${spark.scope}</scope>
</dependency>

<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-hive_${scala.binary.version}</artifactId>
<version>${spark.version}</version>
<exclusions>
<exclusion>
<groupId>commons-httpclient</groupId>
<artifactId>commons-httpclient</artifactId>
</exclusion>
<exclusion>
<groupId>org.apache.httpcomponents</groupId>
<artifactId>httpclient</artifactId>
</exclusion>
<exclusion>
<artifactId>jackson-core-asl</artifactId>
<groupId>org.codehaus.jackson</groupId>
</exclusion>
<exclusion>
<artifactId>jackson-mapper-asl</artifactId>
<groupId>org.codehaus.jackson</groupId>
</exclusion>
</exclusions>
<scope>${spark.scope}</scope>
</dependency>
</dependencies>
Expand Down
7 changes: 4 additions & 3 deletions pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -49,13 +49,13 @@
<commons.version>1.0.0</commons.version>
<slf4j.api.version>1.7.25</slf4j.api.version>
<slf4j.log4j12.version>1.7.25</slf4j.log4j12.version>
<jackson.version>2.12.5</jackson.version>
<jackson.version>2.15.2</jackson.version>
<oshi.core.version>5.8.7</oshi.core.version>
<commons.collections4.version>4.1</commons.collections4.version>
<guava.version>20.0</guava.version>
<java.version>1.8</java.version>
<scala.version>2.11.8</scala.version>
<scala.binary.version>2.11</scala.binary.version>
<scala.version>2.12.15</scala.version>
<scala.binary.version>2.12</scala.binary.version>
<hadoop.version>2.7.5</hadoop.version>
<mysql.version>8.4.0</mysql.version>
<skip.pmd.check>false</skip.pmd.check>
Expand Down Expand Up @@ -113,6 +113,7 @@
<javax.mail.version>1.6.2</javax.mail.version>
<oracle.version>19.3.0.0</oracle.version>
<orai18n.version>12.1.0.2.0</orai18n.version>
<mongo-spark-connector.version>10.4.1</mongo-spark-connector.version>
<jmx_prometheus_javaagent.version>0.20.0</jmx_prometheus_javaagent.version>
<kerb.version>2.0.1</kerb.version>

Expand Down
Loading