Jelajahi Sumber

财产线索加金额

lyb 3 tahun lalu
induk
melakukan
ed0b4fedc0
22 mengubah file dengan 363 tambahan dan 27 penghapusan
  1. 5 0
      src/main/scala/com/winhc/bigdata/spark/jobs/chance/ChangeExtract.scala
  2. 23 0
      src/main/scala/com/winhc/bigdata/spark/jobs/chance/table/zxr_evaluate.scala
  3. 23 0
      src/main/scala/com/winhc/bigdata/spark/jobs/chance/table/zxr_evaluate_results.scala
  4. 61 21
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/CompanyMonitor.scala
  5. 47 3
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/CompanyMonitorHandle.scala
  6. 25 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/alldata/change/tables/company_dishonest_info.scala
  7. 19 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/alldata/change/tables/company_equity_info_list.scala
  8. 12 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/auction_tracking_list.scala
  9. 2 1
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company.scala
  10. 7 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_equity_info_list.scala
  11. 11 2
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_holder_v2.scala
  12. 7 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_land_announcement.scala
  13. 7 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_land_mortgage.scala
  14. 7 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_land_mortgage_v2.scala
  15. 7 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_mortgage_info.scala
  16. 11 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_zxr_list.scala
  17. 2 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_zxr_restrict.scala
  18. 6 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/increase_registered_capital_info.scala
  19. 10 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/wenshu_detail_combine.scala
  20. 12 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/wenshu_detail_combine_v2.scala
  21. 24 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/zxr_evaluate.scala
  22. 35 0
      src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/zxr_evaluate_results.scala

+ 5 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/chance/ChangeExtract.scala

@@ -373,6 +373,11 @@ object ChangeExtract {
     , Args(tableName = "company_annual_report_out_investment", primaryFields = "main_id")
     , Args(tableName = "increase_registered_capital_info", primaryFields = "change_time")
     , Args(tableName = "auction_tracking_list", primaryFields = "auction_items_id")
+
+
+    , Args(tableName = "zxr_evaluate", primaryFields = "name,case_no,asset_name")
+    , Args(tableName = "zxr_evaluate_results", primaryFields = "name,case_no,asset_name")
+
   )
 
 

+ 23 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/chance/table/zxr_evaluate.scala

@@ -0,0 +1,23 @@
+
+package com.winhc.bigdata.spark.jobs.chance.table
+
+import com.winhc.bigdata.spark.jobs.chance.CompanyChangeHandle
+import com.winhc.bigdata.spark.utils.{ChangeExtractUtils, DateUtils}
+
+/**
+ * @Author: lyb
+ * @Date: 2021-01-07
+ * @Description:询价
+ */
+
+case class zxr_evaluate(equCols: Seq[String]) extends CompanyChangeHandle with Serializable  {
+  override def getUpdateTitle(newMap: Map[String, String]): String = getValueOrNull(newMap("name"), s"询价评估发生变更")
+
+  override def getInsertTitle(newMap: Map[String, String]): String = getValueOrNull(newMap("name"), s"新增询价评估")
+
+  override def getLabel(oldMap: Map[String, String], newMap: Map[String, String]): String = ChangeExtractUtils.getTags(newMap, "询价评估", Array("case_no","asset_type", "asset_name","insert_time" ))
+
+  override def getBizTime(newMap: Map[String, String]): String = {
+    DateUtils.getBizDate(newMap("insert_time"), newMap("update_time"))
+  }
+}

+ 23 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/chance/table/zxr_evaluate_results.scala

@@ -0,0 +1,23 @@
+
+package com.winhc.bigdata.spark.jobs.chance.table
+
+import com.winhc.bigdata.spark.jobs.chance.CompanyChangeHandle
+import com.winhc.bigdata.spark.utils.{ChangeExtractUtils, DateUtils}
+
+/**
+ * @Author: lyb
+ * @Date: 2021-01-07
+ * @Description:询价
+ */
+
+case class zxr_evaluate_results(equCols: Seq[String]) extends CompanyChangeHandle with Serializable  {
+  override def getUpdateTitle(newMap: Map[String, String]): String = getValueOrNull(newMap("name"), s"询价评估结果发生变更")
+
+  override def getInsertTitle(newMap: Map[String, String]): String = getValueOrNull(newMap("name"), s"新增询价评估结果")
+
+  override def getLabel(oldMap: Map[String, String], newMap: Map[String, String]): String = ChangeExtractUtils.getTags(newMap, "询价评估结果", Array("case_no","asset_type", "asset_name","publish_time", "money" ))
+
+  override def getBizTime(newMap: Map[String, String]): String = {
+    DateUtils.getBizDate(newMap("publish_time"), newMap("update_time"))
+  }
+}

+ 61 - 21
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/CompanyMonitor.scala

@@ -48,7 +48,9 @@ object CompanyMonitor {
            |    change_time     STRING COMMENT '变更时间',
            |    biz_id          STRING COMMENT '业务id,数据行id',
            |    info_risk_level STRING COMMENT '变更风险等级',
-           |    create_time     STRING COMMENT '创建时间'
+           |    create_time     STRING COMMENT '创建时间',
+           |    amt             STRING COMMENT '金额'
+           |
            |)
            |COMMENT '企业财务监控输出表'
            |PARTITIONED BY
@@ -134,7 +136,7 @@ object CompanyMonitor {
           result.map(res => Row(CompanyDynamicHandleUtils.getDynamicId(res._1, res._5, res._7, res._6),
             res._1, res._2, res._3, res._4,
             res._5.replaceAll("null", ""), res._6, res._7, res._8, res._9,
-            DateFormatUtils.format(new Date(), "yyyy-MM-dd HH:mm:ss")))
+            DateFormatUtils.format(new Date(), "yyyy-MM-dd HH:mm:ss"), res._10))
         }
       })
 
@@ -150,30 +152,66 @@ object CompanyMonitor {
         , "info_risk_level" -> StringType
         , "type" -> StringType
         , "create_time" -> StringType
+        , "amt" -> StringType
       ))
       spark.createDataFrame(rdd, schema)
         .createOrReplaceTempView(s"company_monitor_tmp_$tableName")
 
       unescapeHtml4()
-      sql(
-        s"""
-           |INSERT ${if (isWindows) "INTO" else "OVERWRITE"} TABLE ${getEnvProjectName(env, project)}.$targetTab PARTITION(ds='$ds',tn='$tableName')
-           |SELECT  id
-           |        , cid
-           |        , cname
-           |        , table_type
-           |        , flow_type
-           |        , unescapeHtml4(rta_desc) rta_desc
-           |        , change_time
-           |        , biz_id
-           |        , info_risk_level
-           |        , type
-           |        , create_time
-           |FROM
-           |    company_monitor_tmp_$tableName
-           |WHERE id IS NOT NULL
-           |AND   to_timestamp(change_time) <= now()
-           |""".stripMargin)
+
+      if(tableName.equals("company_zxr_restrict") || tableName.equals("company_dishonest_info")){
+        var dataFlag = 3
+        if(tableName.equals("company_zxr_restrict")){
+          dataFlag = 5
+        }
+        sql(
+          s"""
+             |INSERT ${if (isWindows) "INTO" else "OVERWRITE"} TABLE ${getEnvProjectName(env, project)}.$targetTab PARTITION(ds='$ds',tn='$tableName')
+             |SELECT  id
+             |        , cid
+             |        , cname
+             |        , table_type
+             |        , flow_type
+             |        , unescapeHtml4(rta_desc) rta_desc
+             |        , change_time
+             |        , biz_id
+             |        , info_risk_level
+             |        , type
+             |        , create_time
+             |        , if(case_amt is null, 0, case_amt/10000)
+             |FROM
+             |    (
+             |     select c.*, d.case_amt,  ROW_NUMBER() OVER (PARTITION BY c.id  ORDER BY d.case_amt DESC)  num from
+             |          ( SELECT a.*, b.judicase_id FROM  company_monitor_tmp_$tableName a
+             |            LEFT JOIN winhc_eci_dev.ads_judicial_case_relation_replace_cids b  ON a.biz_id = b.detail_id where b.ds = '${getLastPartitionsOrElse("winhc_eci_dev.ads_judicial_case_relation_replace_cids", "0")}' and b.flag = '$dataFlag'
+             |          ) c LEFT JOIN winhc_eci.ads_judicial_case_relation_r1 d ON c.judicase_id = d.judicase_id
+             |    )
+             |WHERE id IS NOT NULL
+             |AND   to_timestamp(change_time) <= now() and num = 1
+             |""".stripMargin)
+      }else{
+        sql(
+          s"""
+             |INSERT ${if (isWindows) "INTO" else "OVERWRITE"} TABLE ${getEnvProjectName(env, project)}.$targetTab PARTITION(ds='$ds',tn='$tableName')
+             |SELECT  id
+             |        , cid
+             |        , cname
+             |        , table_type
+             |        , flow_type
+             |        , unescapeHtml4(rta_desc) rta_desc
+             |        , change_time
+             |        , biz_id
+             |        , info_risk_level
+             |        , type
+             |        , create_time
+             |        , amt
+             |FROM
+             |    company_monitor_tmp_$tableName
+             |WHERE id IS NOT NULL
+             |AND   to_timestamp(change_time) <= now()
+             |""".stripMargin)
+      }
+
     }
   }
 
@@ -200,6 +238,8 @@ object CompanyMonitor {
     , Args(tableName = "company_land_mortgage_v2", bName = 1)
     , Args(tableName = "auction_tracking_list", bName = 1)
     , Args(tableName = "increase_registered_capital_info", bName = 1)
+    , Args(tableName = "zxr_evaluate", bName = 1)
+    , Args(tableName = "zxr_evaluate_results", bName = 1)
 
   )
 

+ 47 - 3
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/CompanyMonitorHandle.scala

@@ -1,5 +1,12 @@
 package com.winhc.bigdata.spark.jobs.monitor
 
+import java.util
+import java.util.Collections
+
+import com.winhc.bigdata.spark.utils.BaseUtil
+import org.apache.http.entity.ContentType
+import org.apache.http.nio.entity.NStringEntity
+import org.apache.http.util.EntityUtils
 import org.apache.spark.internal.Logging
 
 /**
@@ -33,9 +40,12 @@ trait CompanyMonitorHandle extends Logging {
     , "company_holder_deleted" -> "1" //股东移除
     , "increase_registered_capital_info" -> "2" //增资记录
     , "auction_tracking_list" -> "3" //司法拍卖
+    , "zxr_evaluate" -> "3" //有资产选定询价评估机构
+    , "zxr_evaluate_results" -> "3" //有资产完成询价评估
+
   )
 
-  //财产流向 0 -> 流入 1 -> 流出
+  //财产流向 0 -> 流入 1 -> 流出  2->流向未知
   private val flow_type_map = Map(
     "wenshu_detail_combine_v2" -> "0" //新增胜诉案件
     , "company_dishonest_info" -> "1" //失信移除
@@ -55,10 +65,12 @@ trait CompanyMonitorHandle extends Logging {
     , "company_patent_list" -> "0" //专利
     , "company_copyright_reg_list" -> "0" //软件著作权
     , "company_copyright_works_list" -> "0" //作品著作权
-    , "company_holder_add" -> "0" //股东新增
+    , "company_holder_add" -> "2" //股东新增
     , "company_holder_deleted" -> "1" //股东移除
     , "increase_registered_capital_info" -> "0" //增资记录
     , "auction_tracking_list" -> "0" //司法拍卖
+    , "zxr_evaluate" -> "0" //有资产选定询价评估机构
+    , "zxr_evaluate_results" -> "0" //有资产完成询价评估
   )
 
 
@@ -85,6 +97,8 @@ trait CompanyMonitorHandle extends Logging {
     , "wenshu_detail_combine" -> "18" //WENSHU_DETAIL//ES
     , "wenshu_detail_combine_v2" -> "18" //WENSHU_DETAIL//ES
     , "auction_tracking_list" -> "19" //AUCTION_TRACKING_LIST//HBase
+    , "zxr_evaluate" -> "20" ////zxr_evaluate//HBase
+    , "zxr_evaluate_results" -> "21" ////zxr_evaluate_results//HBase
   )
 
   //类别
@@ -111,6 +125,8 @@ trait CompanyMonitorHandle extends Logging {
     , "wenshu_detail_combine_v2" -> "20" //新增胜诉案件
     , "company_land_mortgage_v2" -> "21" //接受了他人土地抵押
     , "auction_tracking_list" -> "22" //有资产即将被司法拍卖
+    , "zxr_evaluate" -> "23" //有资产选定询价评估机构
+    , "zxr_evaluate_results" -> "24" //有资产完成询价评估
   )
 
   /**
@@ -131,7 +147,7 @@ trait CompanyMonitorHandle extends Logging {
    * info_risk_level
    * type
    */
-  def handle(rowkey: String, bizDate: String, cid: String, change_fields: Seq[String], old_map: Map[String, String], new_map: Map[String, String], cname: String = null): Seq[(String, String, String, String, String, String, String, String, String)] = {
+  def handle(rowkey: String, bizDate: String, cid: String, change_fields: Seq[String], old_map: Map[String, String], new_map: Map[String, String], cname: String = null): Seq[(String, String, String, String, String, String, String, String, String, String)] = {
     val rta_desc = get_rta_desc(old_map, new_map)
     if (rta_desc == null) {
       return Seq.empty
@@ -146,6 +162,7 @@ trait CompanyMonitorHandle extends Logging {
         , get_biz_id(rowkey, new_map)
         , get_info_risk_level(old_map, new_map)
         , get_type()
+        , getMoney(new_map)
       ))
     } catch {
       case e: Exception => {
@@ -235,4 +252,31 @@ trait CompanyMonitorHandle extends Logging {
   def get_conditional_filter(): String = ""
 
 
+  protected def getMoney(new_map: Map[String, String]) : String = "0"
+
+  protected def processMoney(amt : String) :String ={
+    var amtstr = amt
+    if (amtstr == null ){
+      "0"
+    }else{
+      amtstr = amtstr.replaceAll("[^\\d.]", "")
+
+      if(amtstr.equals("") || amtstr.split("\\.").size > 2  || amtstr.startsWith(".") || amtstr.endsWith(".") ){
+        amtstr = "0.0"
+      }
+      try{
+        var x = amtstr.toDouble
+      }catch {
+        case ex : Exception =>{
+          amtstr = "0"
+        }
+
+      }
+
+
+      amtstr
+    }
+  }
+
+
 }

+ 25 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/alldata/change/tables/company_dishonest_info.scala

@@ -0,0 +1,25 @@
+package com.winhc.bigdata.spark.jobs.monitor.alldata.change.tables
+
+import com.winhc.bigdata.spark.jobs.chance.CompanyChangeHandle
+import com.winhc.bigdata.spark.jobs.monitor.alldata.change.CompanyChangeHandle1
+import com.winhc.bigdata.spark.utils.{ChangeExtractUtils, DateUtils}
+
+/**
+ * @Author: XuJiakai
+ * @Date: 2020/8/12 18:50
+ * @Description: 企业失信被执
+ */
+case class company_dishonest_info(equCols: Seq[String]) extends CompanyChangeHandle1 {
+  override def getUpdateTitle(newMap: Map[String, String]): String = "企业失信被执发生变化"
+
+  override def getInsertTitle(newMap: Map[String, String]): String = s"新增1条企业失信信息:${newMap.getOrElse("court","")}"
+
+  override def getLabel(oldMap: Map[String, String], newMap: Map[String, String]): String = ChangeExtractUtils.getTags(newMap, "企业失信被执", Array("name", "case_no", "court", "gist_dd", "gist_unit"))
+
+  override def getBizTime(newMap: Map[String, String]): String = DateUtils.getNotNullStr(newMap("pub_date"),newMap("reg_time"),newMap("appro_time"),newMap("update_time"))
+
+  override def getCondition(): String = {
+
+    " AND status =1"
+  }
+}

+ 19 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/alldata/change/tables/company_equity_info_list.scala

@@ -0,0 +1,19 @@
+package com.winhc.bigdata.spark.jobs.monitor.alldata.change.tables
+
+import com.winhc.bigdata.spark.jobs.chance.CompanyChangeHandle
+import com.winhc.bigdata.spark.jobs.monitor.alldata.change.CompanyChangeHandle1
+import com.winhc.bigdata.spark.utils.ChangeExtractUtils
+
+//股权出质打平
+case class company_equity_info_list(equCols: Seq[String]) extends CompanyChangeHandle1 {
+
+  override def getCid(rowkey: String, newMap: Map[String, String]): String = newMap("cid")
+
+  override def getUpdateTitle(newMap: Map[String, String]): String = s"股权出质信息发生变更"
+
+  override def getInsertTitle(newMap: Map[String, String]): String = s"股权出质信息发生变更"
+
+  override def getLabel(oldMap: Map[String, String], newMap: Map[String, String]): String  = ChangeExtractUtils.getTags(newMap, "股权出质", Array("pledgor", "pledgee", "cid", "pledgor_id", "pledgee_type", "pledgee_id", "pledgor_type"))
+
+  override def getBizTime(newMap: Map[String, String]): String = newMap("reg_date")
+}

+ 12 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/auction_tracking_list.scala

@@ -22,4 +22,16 @@ case class auction_tracking_list() extends CompanyMonitorHandle {
 
   //override def org_type(): Seq[String] = Seq("insert","update")
 
+
+  override protected def getMoney(new_map: Map[String, String]): String = {
+
+    var amtstr = new_map("query_price")
+    if (amtstr == null){
+      "0"
+    }else {
+      amtstr = processMoney(amtstr)
+      ((amtstr.toDouble)/ 10000).formatted("%.2f")
+    }
+  }
+
 }

+ 2 - 1
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company.scala

@@ -21,7 +21,7 @@ case class company() extends CompanyMonitorHandle {
        |变更时间:${new_map("update_time")}""".stripMargin
 
   //注册资本上升
-  override def handle(rowkey: String, bizDate: String, cid: String, change_fields: Seq[String], old_map: Map[String, String], new_map: Map[String, String], cname: String = null): Seq[(String, String, String, String, String, String, String, String, String)] = {
+  override def handle(rowkey: String, bizDate: String, cid: String, change_fields: Seq[String], old_map: Map[String, String], new_map: Map[String, String], cname: String = null): Seq[(String, String, String, String, String, String, String, String, String, String)] = {
     if (!compareAmount(old_map, new_map)) {
       return Seq.empty
     }
@@ -35,6 +35,7 @@ case class company() extends CompanyMonitorHandle {
         , get_biz_id(rowkey, new_map)
         , get_info_risk_level(old_map, new_map)
         , get_type()
+        , getMoney(new_map)
       ))
     } catch {
       case e: Exception => {

+ 7 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_equity_info_list.scala

@@ -24,4 +24,11 @@ case class company_equity_info_list() extends CompanyMonitorHandle {
     "AND data['type'] = '1'"
   }
 
+
+  override protected def getMoney(new_map: Map[String, String]): String = {
+
+    var amtstr = new_map("equity_amount")
+    amtstr = processMoney(amtstr)
+    amtstr
+  }
 }

+ 11 - 2
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_holder_v2.scala

@@ -34,7 +34,7 @@ case class company_holder_v2() extends CompanyMonitorHandle {
     s"""|认缴金额:${new_map("amount")}
         |认缴出资时间:${new_map("update_time")}""".stripMargin
 
-  override def handle(rowkey: String, bizDate: String, cid: String, change_fields: Seq[String], old_map: Map[String, String], new_map: Map[String, String], cname: String = null): Seq[(String, String, String, String, String, String, String, String, String)] = {
+  override def handle(rowkey: String, bizDate: String, cid: String, change_fields: Seq[String], old_map: Map[String, String], new_map: Map[String, String], cname: String = null): Seq[(String, String, String, String, String, String, String, String, String, String)] = {
     val rta_desc = get_rta_desc(old_map, new_map)
     if (rta_desc == null) {
       return Seq.empty
@@ -49,12 +49,13 @@ case class company_holder_v2() extends CompanyMonitorHandle {
         , cname
         //, if (flag) "company_holder_add" else "company_holder_deleted"
         , if (flag) "10" else "10"
-        , if (flag) "0" else "1"
+        , if (flag) "2" else "1"
         , rta_desc
         , get_change_time(bizDate, new_map)
         , get_biz_id(rowkey, new_map)
         , if (flag) "2" else "1"
         , if (flag) "10" else "11"
+        , getMoney(new_map)
       ))
     } catch {
       case e: Exception => {
@@ -64,4 +65,12 @@ case class company_holder_v2() extends CompanyMonitorHandle {
     }
   }
 
+
+  override protected def getMoney(new_map: Map[String, String]): String = {
+
+    var amtstr = new_map("amount")
+    amtstr = processMoney(amtstr)
+    amtstr
+  }
+
 }

+ 7 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_land_announcement.scala

@@ -21,4 +21,11 @@ case class company_land_announcement() extends CompanyMonitorHandle {
        |合同签订日期:${new_map("contract_date")}""".stripMargin
 
 
+  override protected def getMoney(new_map: Map[String, String]): String = {
+
+    var amtstr = new_map("tran_price")
+    amtstr = processMoney(amtstr)
+    amtstr
+  }
+
 }

+ 7 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_land_mortgage.scala

@@ -25,4 +25,11 @@ case class company_land_mortgage() extends CompanyMonitorHandle {
   override def get_conditional_filter(): String = {
     "AND  data['type'] in ('bothone','mortgagor') "
   }
+
+  override protected def getMoney(new_map: Map[String, String]): String = {
+
+    var amtstr = new_map("mortgage_amount")
+    amtstr = processMoney(amtstr)
+    amtstr
+  }
 }

+ 7 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_land_mortgage_v2.scala

@@ -25,4 +25,11 @@ case class company_land_mortgage_v2() extends CompanyMonitorHandle {
   override def get_conditional_filter(): String = {
     "AND  data['type'] in ('bothtwo','mortgagee') "
   }
+
+  override protected def getMoney(new_map: Map[String, String]): String = {
+
+    var amtstr = new_map("mortgage_amount")
+    amtstr = processMoney(amtstr)
+    amtstr
+  }
 }

+ 7 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_mortgage_info.scala

@@ -21,4 +21,11 @@ case class company_mortgage_info() extends CompanyMonitorHandle {
        |状态:${new_map("status")}""".stripMargin
 
 
+
+  override protected def getMoney(new_map: Map[String, String]): String = {
+
+    var amtstr = new_map("amount")
+    amtstr = processMoney(amtstr)
+    amtstr
+  }
 }

+ 11 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_zxr_list.scala

@@ -24,4 +24,15 @@ case class company_zxr_list() extends CompanyMonitorHandle {
   override def org_type() = {
     Seq("update")
   }
+
+
+  override protected def getMoney(new_map: Map[String, String]): String = {
+
+    var amtstr = new_map("exec_money")
+    amtstr = processMoney(amtstr)
+    amtstr = (amtstr.toDouble/10000 ).formatted("%.2f")
+    amtstr
+
+
+  }
 }

+ 2 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/company_zxr_restrict.scala

@@ -24,4 +24,6 @@ case class company_zxr_restrict() extends CompanyMonitorHandle {
   override def org_type() = {
     Seq("update")
   }
+
+
 }

+ 6 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/increase_registered_capital_info.scala

@@ -19,5 +19,11 @@ case class increase_registered_capital_info() extends CompanyMonitorHandle {
        |变更前资本:${new_map("content_before")}
        |变更时间:${new_map("change_time")}""".stripMargin
 
+  override protected def getMoney(new_map: Map[String, String]): String = {
 
+    var amtstr = new_map("content_after")
+
+    amtstr = processMoney(amtstr)
+    amtstr
+  }
 }

+ 10 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/wenshu_detail_combine.scala

@@ -34,4 +34,14 @@ case class wenshu_detail_combine() extends CompanyMonitorHandle {
   override def get_biz_id(rowkey: String, new_map: Map[String, String]): String = {
     new_map("case_id")
   }
+
+
+  override protected def getMoney(new_map: Map[String, String]): String = {
+
+    var amtstr = new_map("case_amt")
+
+    amtstr = processMoney(amtstr)
+    amtstr
+
+  }
 }

+ 12 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/wenshu_detail_combine_v2.scala

@@ -36,4 +36,16 @@ case class wenshu_detail_combine_v2() extends CompanyMonitorHandle {
     new_map("case_id")
   }
 
+
+  override protected def getMoney(new_map: Map[String, String]): String = {
+
+    var amtstr = new_map("case_amt")
+
+    amtstr = processMoney(amtstr)
+    amtstr
+
+
+
+  }
+
 }

+ 24 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/zxr_evaluate.scala

@@ -0,0 +1,24 @@
+package com.winhc.bigdata.spark.jobs.monitor.tables
+
+import com.winhc.bigdata.spark.jobs.monitor.CompanyMonitorHandle
+
+/**
+ * @Date: 2021/1/07 18:50
+ * @Description: 询价评估
+ */
+case class zxr_evaluate() extends CompanyMonitorHandle {
+  /**
+   * 信息描述
+   *
+   * @param old_map
+   * @param new_map
+   * @return
+   */
+  override protected def get_rta_desc(old_map: Map[String, String], new_map: Map[String, String]): String =
+    s"""案号:${new_map("case_no")}
+       |财产类型:${new_map("asset_type")}
+       |财产名称:${new_map("asset_name")}
+       |日期:${new_map("insert_time")}""".stripMargin
+
+
+}

+ 35 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/monitor/tables/zxr_evaluate_results.scala

@@ -0,0 +1,35 @@
+package com.winhc.bigdata.spark.jobs.monitor.tables
+
+import com.winhc.bigdata.spark.jobs.monitor.CompanyMonitorHandle
+
+/**
+ * @Date: 2021/1/07 18:50
+ * @Description: 询价评估结果
+ */
+case class zxr_evaluate_results() extends CompanyMonitorHandle {
+  /**
+   * 信息描述
+   *
+   * @param old_map
+   * @param new_map
+   * @return
+   */
+  override protected def get_rta_desc(old_map: Map[String, String], new_map: Map[String, String]): String =
+    s"""案号:${new_map("case_no")}
+       |财产类型:${new_map("asset_type")}
+       |财产名称:${new_map("asset_name")}
+       |日期:${new_map("publish_time")}""".stripMargin
+
+
+  override protected def getMoney(new_map: Map[String, String]): String = {
+
+    var amtstr = new_map("money")
+    if (amtstr == null){
+      "0"
+    }else {
+      var a1 = amtstr.replaceAll("\"|\\[|\\]", "").split(",").toArray.map(_.toDouble)
+      var d = a1.sum / a1.length / 10000
+      d.formatted("%.2f")
+    }
+  }
+}