modify stop group description

This commit is contained in:
judy0131 2018-12-14 17:03:45 +08:00
parent 04b8992a89
commit 98be6dca8d
108 changed files with 195 additions and 246 deletions

View File

@ -52,7 +52,7 @@ class EmailClean extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CleanGroup.toString)
List(StopGroup.CleanGroup.toString)
}
}

View File

@ -56,6 +56,6 @@ class IdentityNumberClean extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CleanGroup.toString)
List(StopGroup.CleanGroup.toString)
}
}

View File

@ -50,7 +50,7 @@ class PhoneNumberClean extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CleanGroup.toString)
List(StopGroup.CleanGroup.toString)
}
}

View File

@ -48,7 +48,7 @@ class TitleClean extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CleanGroup.toString)
List(StopGroup.CleanGroup.toString)
}
}

View File

@ -54,7 +54,7 @@ class ConvertSchema extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CommonGroup.toString)
List(StopGroup.CommonGroup.toString)
}
}

View File

@ -45,7 +45,7 @@ class DoFlatMapStop extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CommonGroup.toString)
List(StopGroup.CommonGroup.toString)
}

View File

@ -62,7 +62,7 @@ class DoMapStop extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CommonGroup.toString)
List(StopGroup.CommonGroup.toString)
}

View File

@ -86,7 +86,7 @@ class ExecuteSQLStop extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CommonGroup.toString)
List(StopGroup.CommonGroup.toString)
}

View File

@ -41,6 +41,6 @@ class Fork extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CommonGroup.toString)
List(StopGroup.CommonGroup.toString)
}
}

View File

@ -41,7 +41,7 @@ class Merge extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CommonGroup.toString)
List(StopGroup.CommonGroup.toString)
}
}

View File

@ -56,7 +56,7 @@ class SelectField extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CommonGroup.toString)
List(StopGroup.CommonGroup.toString)
}
}

View File

@ -70,7 +70,7 @@ class CsvParser extends ConfigurableStop{
var descriptor : List[PropertyDescriptor] = List()
val csvPath = new PropertyDescriptor().name("csvPath").displayName("csvPath").description("The path of csv file").defaultValue("").required(true)
descriptor = csvPath :: descriptor
val header = new PropertyDescriptor().name("header").displayName("header").description("Whether the csv file have header or not").defaultValue("").required(true)
val header = new PropertyDescriptor().name("header").displayName("header").description("Whether the csv file have header or not").defaultValue("false").allowableValues(Set("true","false")).required(true)
descriptor = header :: descriptor
val delimiter = new PropertyDescriptor().name("delimiter").displayName("delimiter").description("The delimiter of csv file").defaultValue("").required(true)
descriptor = delimiter :: descriptor
@ -85,7 +85,7 @@ class CsvParser extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CsvGroup.toString)
List(StopGroup.CsvGroup.toString)
}
}

View File

@ -45,7 +45,7 @@ class CsvSave extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CsvGroup.toString)
List(StopGroup.CsvGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.csv
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.SparkContext
@ -75,7 +75,7 @@ class CsvStringParser extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CsvGroup.toString)
List(StopGroup.CsvGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.csv
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.SparkContext
@ -77,7 +77,7 @@ class FolderCsvParser extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.CsvGroup.toString)
List(StopGroup.CsvGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -2,7 +2,7 @@ package cn.piflow.bundle.es
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.sql.SparkSession
@ -67,7 +67,7 @@ class FetchEs extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.ESGroup.toString)
List(StopGroup.ESGroup.toString)
}

View File

@ -2,7 +2,7 @@ package cn.piflow.bundle.es
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.sql.SparkSession
import org.elasticsearch.spark.sql.EsSparkSQL
@ -78,7 +78,7 @@ class PutEs extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.ESGroup.toString)
List(StopGroup.ESGroup.toString)
}
}

View File

@ -2,7 +2,7 @@ package cn.piflow.bundle.es
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.sql.SparkSession
import org.elasticsearch.spark.rdd.EsSpark
@ -114,6 +114,6 @@ class QueryEs extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.ESGroup.toString)
List(StopGroup.ESGroup.toString)
}
}

View File

@ -73,7 +73,7 @@ class ExcelParser extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.ExcelGroup.toString)
List(StopGroup.ExcelGroup.toString)
}
}

View File

@ -58,7 +58,7 @@ class FetchFile extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.FileGroup.toString)
List(StopGroup.FileGroup.toString)
}

View File

@ -61,7 +61,7 @@ class PutFile extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.FileGroup.toString)
List(StopGroup.FileGroup.toString)
}

View File

@ -60,6 +60,6 @@ class RegexTextProcess extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.FileGroup.toString)
List(StopGroup.FileGroup.toString)
}
}

View File

@ -5,7 +5,7 @@ import java.net.{HttpURLConnection, URL}
import java.util
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, FtpGroup, HttpGroup, PortEnum, StopGroup, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import sun.net.ftp.{FtpClient, FtpDirEntry}
@ -160,7 +160,7 @@ class LoadFromFtp extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.FtpGroup.toString)
List(StopGroup.FtpGroup.toString)
}

View File

@ -8,7 +8,7 @@ import java.util.{ArrayList, Date}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.sql.SparkSession
import org.jsoup.Jsoup
@ -246,7 +246,7 @@ class LoadFromFtpUrl extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.FtpGroup.toString)
List(StopGroup.FtpGroup.toString)
}
def initialize(ctx: ProcessContext): Unit = {

View File

@ -6,7 +6,7 @@ import java.util.ArrayList
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import sun.net.ftp.{FtpClient, FtpDirEntry}
@ -182,7 +182,7 @@ class NewLoadFromFtp extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.FtpGroup.toString)
List(StopGroup.FtpGroup.toString)
}

View File

@ -4,7 +4,7 @@ package cn.piflow.bundle.ftp
import cn.piflow.bundle.util.UnGzUtil
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.sql.{DataFrame, Row, SparkSession}
@ -121,7 +121,7 @@ class UnGz extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.FtpGroup.toString)
List(StopGroup.FtpGroup.toString)
}
def initialize(ctx: ProcessContext): Unit = {

View File

@ -4,7 +4,7 @@ import java.io.{DataOutputStream, File, InputStream, OutputStream}
import java.util
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, FtpGroup, PortEnum, StopGroup, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import sun.net.TelnetOutputStream
@ -174,6 +174,6 @@ class UploadToFtp extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.FtpGroup.toString)
List(StopGroup.FtpGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.graphx
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.graphx._
@ -61,7 +61,7 @@ class LabelPropagation extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.GraphX.toString)
List(StopGroup.GraphX.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.graphx
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.sql.SparkSession
@ -55,7 +55,7 @@ class LoadGraph extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.GraphX.toString)
List(StopGroup.GraphX.toString)
}
}

View File

@ -2,7 +2,7 @@ package cn.piflow.bundle.hbase
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.ImageUtil
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.hbase.HBaseConfiguration
@ -84,7 +84,7 @@ class FetchHbase extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HbaseGroup.toString)
List(StopGroup.HbaseGroup.toString)
}
}

View File

@ -4,7 +4,7 @@ package cn.piflow.bundle.Hbase
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.hbase.{HBaseConfiguration, HTableDescriptor}
@ -105,7 +105,7 @@ class GetHbase extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HbaseGroup.toString)
List(StopGroup.HbaseGroup.toString)
}
}

View File

@ -2,7 +2,7 @@ package cn.piflow.bundle.Hbase
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.ImageUtil
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.hbase.HBaseConfiguration
@ -93,7 +93,7 @@ class PutHbase extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HbaseGroup.toString)
List(StopGroup.HbaseGroup.toString)
}
}

View File

@ -4,7 +4,7 @@ package cn.piflow.bundle.hdfs
import cn.piflow._
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import org.apache.spark.sql.SparkSession
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.fs.FileSystem
@ -69,7 +69,7 @@ class DeleteHdfs extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HdfsGroup.toString)
List(StopGroup.HdfsGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -3,7 +3,7 @@ package cn.piflow.bundle.hdfs
import cn.piflow._
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import org.apache.spark.sql.SparkSession
@ -100,7 +100,7 @@ class GetHdfs extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HdfsGroup.toString)
List(StopGroup.HdfsGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -3,7 +3,7 @@ package cn.piflow.bundle.hdfs
import cn.piflow._
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.fs.{FileStatus, FileSystem, Path}
import org.apache.spark.sql. SparkSession
@ -82,7 +82,7 @@ class ListHdfs extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HdfsGroup.toString)
List(StopGroup.HdfsGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -4,7 +4,7 @@ package cn.piflow.bundle.hdfs
import cn.piflow._
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.fs.FileSystem
import org.apache.spark.sql.SparkSession
@ -69,7 +69,7 @@ class PutHdfs extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HdfsGroup.toString)
List(StopGroup.HdfsGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -62,7 +62,7 @@ class PutHiveQL extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HiveGroup.toString)
List(StopGroup.HiveGroup.toString)
}
}

View File

@ -53,7 +53,7 @@ class PutHiveStreaming extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HiveGroup.toString)
List(StopGroup.HiveGroup.toString)
}

View File

@ -49,7 +49,7 @@ class SelectHiveQL extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HiveGroup.toString)
List(StopGroup.HiveGroup.toString)
}

View File

@ -86,7 +86,7 @@ class FileDownHDFS extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HttpGroup.toString)
List(StopGroup.HttpGroup.toString)
}

View File

@ -4,7 +4,7 @@ import java.util
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.http.client.methods.{CloseableHttpResponse, HttpGet}
import org.apache.http.impl.client.HttpClients
@ -158,7 +158,7 @@ class GetUrl extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HttpGroup.toString)
List(StopGroup.HttpGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -7,7 +7,7 @@ import java.util
import cn.piflow._
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.fs.{FSDataInputStream, FileSystem, Path}
import org.apache.http.client.methods._
@ -273,7 +273,7 @@ class InvokeUrl extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HttpGroup.toString)
List(StopGroup.HttpGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -86,7 +86,7 @@ class LoadZipFromUrl extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HttpGroup.toString)
List(StopGroup.HttpGroup.toString)
}

View File

@ -5,7 +5,7 @@ import java.net.URI
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.commons.httpclient.HttpClient
import org.apache.hadoop.conf.Configuration
@ -79,7 +79,7 @@ class PostUrl extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HttpGroup.toString)
List(StopGroup.HttpGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -213,7 +213,7 @@ class UnGZip extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HttpGroup.toString)
List(StopGroup.HttpGroup.toString)
}
}

View File

@ -198,7 +198,7 @@ class UnzipFilesOnHDFS extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.HttpGroup.toString)
List(StopGroup.HttpGroup.toString)
}
}

View File

@ -3,7 +3,7 @@ package cn.piflow.bundle.impala
import java.sql.{Connection, DriverManager, ResultSet, Statement}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.rdd.RDD
@ -95,7 +95,7 @@ class SelectImpala extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.Mongodb.toString)
List(StopGroup.Mongodb.toString)
}

View File

@ -7,7 +7,7 @@ import java.util.Date
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.rdd.RDD
import org.apache.spark.sql.types.{StringType, StructField, StructType}
@ -190,7 +190,7 @@ class spider extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.Spider.toString)
List(StopGroup.Spider.toString)
}
override def initialize(ctx: ProcessContext): Unit = {}

View File

@ -78,7 +78,7 @@ class JdbcRead extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.JdbcGroup.toString)
List(StopGroup.JdbcGroup.toString)
}
}

View File

@ -177,7 +177,7 @@ class JdbcReadFromOracle extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.JdbcGroup.toString)
List(StopGroup.JdbcGroup.toString)
}

View File

@ -71,7 +71,7 @@ class JdbcWrite extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.JdbcGroup.toString)
List(StopGroup.JdbcGroup.toString)
}

View File

@ -87,7 +87,7 @@ class JdbcWriteToOracle extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.JdbcGroup.toString)
List(StopGroup.JdbcGroup.toString)
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.json
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.sql
@ -58,7 +58,7 @@ class EvaluateJsonPath extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.JsonGroup.toString)
List(StopGroup.JsonGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -5,7 +5,7 @@ import java.net.URI
import cn.piflow.bundle.util.JsonUtil
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.hadoop.conf.Configuration
@ -127,7 +127,7 @@ class FolderJsonParser extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.JsonGroup.toString)
List(StopGroup.JsonGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -58,7 +58,7 @@ class JsonParser extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.JsonGroup.toString)
List(StopGroup.JsonGroup.toString)
}
}

View File

@ -45,7 +45,7 @@ class JsonSave extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.JsonGroup.toString)
List(StopGroup.JsonGroup.toString)
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.json
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.sql.SparkSession
@ -44,7 +44,7 @@ class JsonStringParser extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.JsonGroup.toString)
List(StopGroup.JsonGroup.toString)
}

View File

@ -2,7 +2,7 @@ package cn.piflow.bundle.json
import cn.piflow.bundle.util.JsonUtil
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.sql
@ -76,7 +76,7 @@ class MultiFolderJsonParser extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.JsonGroup.toString)
List(StopGroup.JsonGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -96,7 +96,7 @@ class ReadFromKafka extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.KafkaGroup.toString)
List(StopGroup.KafkaGroup.toString)
}
override val authorEmail: String = "06whuxx@163.com"

View File

@ -79,7 +79,7 @@ class WriteToKafka extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.KafkaGroup.toString)
List(StopGroup.KafkaGroup.toString)
}
override val authorEmail: String = "06whuxx@163.com"

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.memcache
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import com.danga.MemCached.{MemCachedClient, SockIOPool}
@ -166,7 +166,7 @@ class ComplementByMemcache extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.Memcache.toString)
List(StopGroup.Memcache.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.memcache
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import com.danga.MemCached.{MemCachedClient, SockIOPool}
@ -164,7 +164,7 @@ class GetMemcache extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.Memcache.toString)
List(StopGroup.Memcache.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.memcache
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import com.danga.MemCached.{MemCachedClient, SockIOPool}
@ -120,7 +120,7 @@ class PutMemcache extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.Memcache.toString)
List(StopGroup.Memcache.toString)
}
override def initialize(ctx: ProcessContext): Unit = { }

View File

@ -7,7 +7,7 @@ import java.util.regex.Pattern
import cn.piflow.bundle.microorganism.util.BioProject
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.sql.{Row, SparkSession}
import org.biojava.bio.BioException
@ -227,7 +227,7 @@ class BioProjetDataParse extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MicroorganismGroup.toString)
List(StopGroup.MicroorganismGroup.toString)
}
def initialize(ctx: ProcessContext): Unit = {

View File

@ -7,7 +7,7 @@ import java.util.ArrayList
import cn.piflow.bundle.microorganism.util.{CustomIOTools, Process}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.sql.{DataFrame, Row, SparkSession}
import org.biojava.bio.BioException
@ -146,7 +146,7 @@ class GenBankParse extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MicroorganismGroup.toString)
List(StopGroup.MicroorganismGroup.toString)
}
def initialize(ctx: ProcessContext): Unit = {

View File

@ -3,7 +3,7 @@ package cn.piflow.bundle.microorganism
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.sql.{Row, SparkSession}
import org.elasticsearch.spark.sql.EsSparkSQL
@ -77,7 +77,7 @@ class GoldDataParse extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MicroorganismGroup.toString)
List(StopGroup.MicroorganismGroup.toString)
}
def initialize(ctx: ProcessContext): Unit = {

View File

@ -2,7 +2,7 @@ package cn.piflow.bundle.ml_classification
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.ml.classification.DecisionTreeClassificationModel
import org.apache.spark.sql.SparkSession
@ -55,6 +55,6 @@ class DecisionTreePrediction extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -2,7 +2,7 @@ package cn.piflow.bundle.ml_classification
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.ml.classification.DecisionTreeClassifier
import org.apache.spark.sql.SparkSession
@ -115,7 +115,7 @@ class DecisionTreeTraining extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_classification
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.classification.GBTClassificationModel
@ -55,7 +55,7 @@ class GBTPrediction extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_classification
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.classification.GBTClassifier
@ -144,7 +144,7 @@ class GBTTraining extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -2,7 +2,7 @@ package cn.piflow.bundle.ml_classification
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.ml.classification.LogisticRegressionModel
import org.apache.spark.sql.SparkSession
@ -55,7 +55,7 @@ class LogisticRegressionPrediction extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -2,7 +2,7 @@ package cn.piflow.bundle.ml_classification
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.sql.SparkSession
import org.apache.spark.ml.classification.LogisticRegression
@ -126,7 +126,7 @@ class LogisticRegressionTraining extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_classification
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
@ -55,7 +55,7 @@ class MultilayerPerceptronPrediction extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_classification
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.classification.MultilayerPerceptronClassifier
@ -120,7 +120,7 @@ class MultilayerPerceptronTraining extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -2,7 +2,7 @@ package cn.piflow.bundle.ml_classification
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.ml.classification.NaiveBayesModel
import org.apache.spark.sql.SparkSession
@ -55,7 +55,7 @@ class NaiveBayesPrediction extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}

View File

@ -2,7 +2,7 @@ package cn.piflow.bundle.ml_classification
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.ml.classification.NaiveBayes
import org.apache.spark.sql.SparkSession
@ -69,7 +69,7 @@ class NaiveBayesTraining extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_classification
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.classification.RandomForestClassificationModel
@ -55,7 +55,7 @@ class RandomForestPrediction extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_classification
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.classification.RandomForestClassifier
@ -144,7 +144,7 @@ class RandomForestTraining extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_clustering
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.clustering.BisectingKMeansModel
@ -55,7 +55,7 @@ class BisectingKMeansPrediction extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_clustering
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.clustering.BisectingKMeans
@ -78,7 +78,7 @@ class BisectingKMeansTraining extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_clustering
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.clustering.GaussianMixtureModel
@ -55,7 +55,7 @@ class GaussianMixturePrediction extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_clustering
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.clustering.GaussianMixture
@ -93,7 +93,7 @@ class GaussianMixtureTraining extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_clustering
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.clustering.KMeansModel
@ -55,7 +55,7 @@ class KmeansPrediction extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_clustering
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.clustering.KMeans
@ -89,7 +89,7 @@ class KmeansTraining extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_clustering
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.clustering.{DistributedLDAModel, LDAModel, LocalLDAModel}
@ -55,7 +55,7 @@ class LDAPrediction extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_clustering
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.clustering.LDA
@ -125,7 +125,7 @@ class LDATraining extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.ml_feature
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.spark.ml.feature.Word2Vec
@ -128,6 +128,6 @@ class WordToVec extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.MLGroup.toString)
List(StopGroup.MLGroup.toString)
}
}

View File

@ -4,7 +4,7 @@ import java.util
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import com.mongodb.client.{FindIterable, MongoCollection, MongoCursor, MongoDatabase}
import com.mongodb.{MongoClient, MongoCredential, ServerAddress}
@ -123,7 +123,7 @@ class GetMongo extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.Mongodb.toString)
List(StopGroup.Mongodb.toString)
}
override def initialize(ctx: ProcessContext): Unit = { }

View File

@ -3,7 +3,7 @@ package cn.piflow.bundle.mongodb
import java.util
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import com.mongodb.client.{MongoCollection, MongoDatabase}
@ -99,7 +99,7 @@ class PutMongo extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.Mongodb.toString)
List(StopGroup.Mongodb.toString)
}
override def initialize(ctx: ProcessContext): Unit = { }

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.rdf
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
@ -384,7 +384,7 @@ class CsvToNeo4J extends ConfigurableStop{
override def getIcon(): Array[Byte] = ImageUtil.getImage("neo4j.jpg")
override def getGroup(): List[String] = List(StopGroupEnum.RDFGroup.toString)
override def getGroup(): List[String] = List(StopGroup.RDFGroup.toString)
override def initialize(ctx: ProcessContext): Unit = {}

View File

@ -5,7 +5,7 @@ import java.util.regex.{Matcher, Pattern}
import cn.piflow.bundle.util.Entity
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import org.apache.spark.rdd.RDD
import org.apache.spark.sql.types.{DataTypes, StringType, StructField, StructType}
@ -141,7 +141,7 @@ class RdfToDF extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.RDFGroup.toString)
List(StopGroup.RDFGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -94,7 +94,7 @@ class ReadFromRedis extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.RedisGroup.toString)
List(StopGroup.RedisGroup.toString)
}
override val authorEmail: String = "06whuxx@163.com"

View File

@ -66,7 +66,7 @@ class WriteToRedis extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.RedisGroup.toString)
List(StopGroup.RedisGroup.toString)
}
override val authorEmail: String = "06whuxx@163.com"

View File

@ -38,7 +38,7 @@ class DataFrameRowParser extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.ScriptGroup.toString)
List(StopGroup.ScriptGroup.toString)
}

View File

@ -42,7 +42,7 @@ class ShellExecutor extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.ScriptGroup.toString)
List(StopGroup.ScriptGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -3,7 +3,7 @@ import java.util
import scala.collection.JavaConversions._
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.solr.client.solrj.SolrQuery
@ -141,7 +141,7 @@ class GetFromSolr extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.SolrGroup.toString)
List(StopGroup.SolrGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -1,7 +1,7 @@
package cn.piflow.bundle.solr
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.solr.client.solrj.impl.HttpSolrClient
@ -62,7 +62,7 @@ class PutIntoSolr extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.SolrGroup.toString)
List(StopGroup.SolrGroup.toString)
}
override def getPropertyDescriptor(): List[PropertyDescriptor] = {

View File

@ -5,7 +5,7 @@ import java.util
import cn.piflow.bundle.util.JsonUtil
import cn.piflow.{JobContext, JobInputStream, JobOutputStream, ProcessContext}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroupEnum}
import cn.piflow.conf.{ConfigurableStop, PortEnum, StopGroup}
import cn.piflow.conf.bean.PropertyDescriptor
import cn.piflow.conf.util.{ImageUtil, MapUtil}
import org.apache.hadoop.conf.Configuration
@ -171,7 +171,7 @@ class FlattenXmlParser extends ConfigurableStop{
ImageUtil.getImage("xml.png")
}
override def getGroup(): List[String] = {
List(StopGroupEnum.XmlGroup.toString)
List(StopGroup.XmlGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit = {

View File

@ -51,7 +51,7 @@ class FolderXmlParser extends ConfigurableStop{
}
override def getGroup(): List[String] = {
List(StopGroupEnum.XmlGroup.toString)
List(StopGroup.XmlGroup.toString)
}
override def initialize(ctx: ProcessContext): Unit ={}

View File

@ -61,7 +61,7 @@ class XmlParser extends ConfigurableStop {
}
override def getGroup(): List[String] = {
List(StopGroupEnum.XmlGroup.toString)
List(StopGroup.XmlGroup.toString)
}
}

Some files were not shown because too many files have changed in this diff Show More