|
@@ -3,6 +3,7 @@ package com.winhc.bigdata.spark.jobs
|
|
import com.winhc.bigdata.spark.config.{EsConfig, HBaseConfig}
|
|
import com.winhc.bigdata.spark.config.{EsConfig, HBaseConfig}
|
|
import com.winhc.bigdata.spark.const.BaseConst
|
|
import com.winhc.bigdata.spark.const.BaseConst
|
|
import com.winhc.bigdata.spark.udf.BaseFunc
|
|
import com.winhc.bigdata.spark.udf.BaseFunc
|
|
|
|
+import com.winhc.bigdata.spark.utils.BaseUtil.{atDaysAfter, nowDate}
|
|
import com.winhc.bigdata.spark.utils.{BaseUtil, LoggingUtils, SparkUtils}
|
|
import com.winhc.bigdata.spark.utils.{BaseUtil, LoggingUtils, SparkUtils}
|
|
import org.apache.hadoop.hbase.client.Put
|
|
import org.apache.hadoop.hbase.client.Put
|
|
import org.apache.hadoop.hbase.io.ImmutableBytesWritable
|
|
import org.apache.hadoop.hbase.io.ImmutableBytesWritable
|
|
@@ -70,16 +71,16 @@ object CompanyIncCompany2Es {
|
|
)
|
|
)
|
|
val outFields_Human = Seq(
|
|
val outFields_Human = Seq(
|
|
"NEW_CID"
|
|
"NEW_CID"
|
|
- ,"CID"
|
|
|
|
- ,"ID"
|
|
|
|
- ,"COMPANY_NAME"
|
|
|
|
- ,"HUMAN_NAME"
|
|
|
|
- ,"HID"
|
|
|
|
- ,"HUMAN_PID"
|
|
|
|
- ,"STATUS"
|
|
|
|
- ,"CREATE_TIME"
|
|
|
|
- ,"UPDATE_TIME"
|
|
|
|
- ,"DELETED"
|
|
|
|
|
|
+ , "CID"
|
|
|
|
+ , "ID"
|
|
|
|
+ , "COMPANY_NAME"
|
|
|
|
+ , "HUMAN_NAME"
|
|
|
|
+ , "HID"
|
|
|
|
+ , "HUMAN_PID"
|
|
|
|
+ , "STATUS"
|
|
|
|
+ , "CREATE_TIME"
|
|
|
|
+ , "UPDATE_TIME"
|
|
|
|
+ , "DELETED"
|
|
)
|
|
)
|
|
|
|
|
|
case class Company2Es(s: SparkSession, project: String, bizDate: String) extends LoggingUtils with BaseFunc {
|
|
case class Company2Es(s: SparkSession, project: String, bizDate: String) extends LoggingUtils with BaseFunc {
|
|
@@ -89,7 +90,7 @@ object CompanyIncCompany2Es {
|
|
val code = code2Name()
|
|
val code = code2Name()
|
|
val partition = bizDate.replaceAll("\\-", "")
|
|
val partition = bizDate.replaceAll("\\-", "")
|
|
if (partition.length != 8) {
|
|
if (partition.length != 8) {
|
|
- println("biz date is error!")
|
|
|
|
|
|
+ println("biz date is error! "+partition)
|
|
sys.exit(-99)
|
|
sys.exit(-99)
|
|
}
|
|
}
|
|
val inc_ods_partitions = BaseUtil.getPartitions(s"${project}.inc_ods_company", spark)
|
|
val inc_ods_partitions = BaseUtil.getPartitions(s"${project}.inc_ods_company", spark)
|
|
@@ -157,13 +158,14 @@ object CompanyIncCompany2Es {
|
|
|
|
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
+
|
|
case class Company_Human_Relation2HBase(s: SparkSession, project: String, bizDate: String) extends LoggingUtils {
|
|
case class Company_Human_Relation2HBase(s: SparkSession, project: String, bizDate: String) extends LoggingUtils {
|
|
@(transient@getter) val spark: SparkSession = s
|
|
@(transient@getter) val spark: SparkSession = s
|
|
|
|
|
|
def calc() {
|
|
def calc() {
|
|
val partition = bizDate.replaceAll("\\-", "")
|
|
val partition = bizDate.replaceAll("\\-", "")
|
|
if (partition.length != 8) {
|
|
if (partition.length != 8) {
|
|
- println("biz date is error!")
|
|
|
|
|
|
+ println("biz date is error! "+partition)
|
|
sys.exit(-99)
|
|
sys.exit(-99)
|
|
}
|
|
}
|
|
val inc_ods_partitions = BaseUtil.getPartitions(s"${project}.inc_ods_company_human_relation", spark)
|
|
val inc_ods_partitions = BaseUtil.getPartitions(s"${project}.inc_ods_company_human_relation", spark)
|
|
@@ -229,17 +231,18 @@ object CompanyIncCompany2Es {
|
|
}
|
|
}
|
|
|
|
|
|
def main(args: Array[String]): Unit = {
|
|
def main(args: Array[String]): Unit = {
|
|
- if (args.length != 2) {
|
|
|
|
- println("please enter project and bizDate!")
|
|
|
|
|
|
+ if (args.length != 1) {
|
|
|
|
+ println("please enter project!")
|
|
sys.exit(-99)
|
|
sys.exit(-99)
|
|
}
|
|
}
|
|
|
|
|
|
- val Array(project, bizDate) = args
|
|
|
|
|
|
+ val Array(project) = args
|
|
|
|
|
|
val config = EsConfig.getEsConfigMap ++ mutable.Map(
|
|
val config = EsConfig.getEsConfigMap ++ mutable.Map(
|
|
"spark.hadoop.odps.project.name" -> project,
|
|
"spark.hadoop.odps.project.name" -> project,
|
|
"spark.hadoop.odps.spark.local.partition.amt" -> "10"
|
|
"spark.hadoop.odps.spark.local.partition.amt" -> "10"
|
|
)
|
|
)
|
|
|
|
+ val bizDate = atDaysAfter(-1, nowDate("yyyyMMdd"))
|
|
|
|
|
|
val spark = SparkUtils.InitEnv("company2Es", config)
|
|
val spark = SparkUtils.InitEnv("company2Es", config)
|
|
|
|
|