xufei пре 4 година
родитељ
комит
ca1bab9f57

+ 5 - 5
src/main/scala/com/winhc/bigdata/spark/jobs/CompanyCourtAnnouncement.scala

@@ -54,7 +54,7 @@ case class CompanyCourtAnnouncement(s: SparkSession, project: String, //表所
   def calc(runOld: Boolean = false) = {
     import spark.implicits._
 
-    val inc_ads_company_tb_list = s"${project}.inc_ads_${tableName}_list" //增量ads_list表
+    val inc_ads_company_tb_list = s"${project}.inc_ads_${tableName}_list_v1" //增量ads_list表
 
     var adsListDs = getPartion(inc_ads_company_tb_list, spark)
     //跑存量取第一个分区
@@ -326,11 +326,11 @@ case class CompanyCourtAnnouncement(s: SparkSession, project: String, //表所
 
   def preCalc(): Boolean = {
 
-    val inc_ods_company_tb = s"${project}.inc_ods_$tableName" //增量ods表
-    val inc_ads_company_tb = s"${project}.inc_ads_$tableName" //增量ads表
-    val inc_ads_company_tb_list = s"${project}.inc_ads_${tableName}_list" //增量ads_list表
+    val inc_ods_company_tb = s"${project}.inc_ods_${tableName}" //增量ods表
+    val inc_ads_company_tb = s"${project}.inc_ads_${tableName}_v1" //增量ads表
+    val inc_ads_company_tb_list = s"${project}.inc_ads_${tableName}_list_v1" //增量ads_list表
 
-    val ads_company_tb = s"${project}.ads_$tableName" //存量ads表
+    val ads_company_tb = s"${project}.ads_${tableName}_v1" //存量ads表
 
     //table字段
     val columns: Seq[String] = spark.table(inc_ads_company_tb).schema.map(_.name).filter(s => {