|
@@ -232,7 +232,7 @@ case class deadbeat_info(s: SparkSession,
|
|
|
mapTables("company_zxr_restrict") = ("rowkey", "name", "id_card_trim(identity_num)", "case_create_time", "deleted")
|
|
|
mapTables.foreach(m => {
|
|
|
spark.sparkContext.setJobDescription(s"查老赖数据预处理:${m._1}")
|
|
|
- dishonest_info(spark, project = project, table = m._1, rowkey = m._2._1, cid = null, name = m._2._2, card_num = m._2._3, publish_date = m._2._4, deleted = m._2._5).calc()
|
|
|
+ deadbeat_info_pre(spark, project = project, table = m._1, rowkey = m._2._1, cid = null, name = m._2._2, card_num = m._2._3, publish_date = m._2._4, deleted = m._2._5).calc()
|
|
|
})
|
|
|
}
|
|
|
|
|
@@ -314,7 +314,7 @@ case class deadbeat_info(s: SparkSession,
|
|
|
mapTables("company_zxr_restrict") = ("rowkey", "cid", "name", "identity_num", "case_create_time", "deleted")
|
|
|
mapTables.foreach(m => {
|
|
|
spark.sparkContext.setJobDescription(s"查老赖数据预处理:${m._1}")
|
|
|
- dishonest_info(spark, project = project, table = m._1, rowkey = m._2._1, cid = m._2._2, name = m._2._3, card_num = m._2._4, publish_date = m._2._5, deleted = m._2._6).calc()
|
|
|
+ deadbeat_info_pre(spark, project = project, table = m._1, rowkey = m._2._1, cid = m._2._2, name = m._2._3, card_num = m._2._4, publish_date = m._2._5, deleted = m._2._6).calc()
|
|
|
})
|
|
|
}
|
|
|
|