Bläddra i källkod

local模式下读写hbase

许家凯 4 år sedan
förälder
incheckning
1c365c5664

+ 5 - 8
pom.xml

@@ -111,6 +111,11 @@
             <version>1.18.4</version>
             <scope>provided</scope>
         </dependency>
+        <dependency>
+            <groupId>com.alibaba</groupId>
+            <artifactId>dns-cache-manipulator</artifactId>
+            <version>1.5.1</version>
+        </dependency>
     </dependencies>
 
     <build>
@@ -129,19 +134,11 @@
                             <minimizeJar>false</minimizeJar>
                             <shadedArtifactAttached>true</shadedArtifactAttached>
                             <artifactSet>
-                                <!--                                <excludes>-->
-                                <!--                                    <exclude>jmock:*</exclude>-->
-                                <!--                                    <exclude>*:xml-apis</exclude>-->
-                                <!--                                    <exclude>org.apache.maven:lib:tests</exclude>-->
-                                <!--                                    <exclude>log4j:log4j:jar:</exclude>-->
-                                <!--                                </excludes>-->
                                 <includes>
-                                    <!--                                    <include>*:*</include>-->
                                     <include>cn.hutool:*</include>
                                     <include>com.aliyun.odps:*</include>
                                     <include>org.mongodb.*:*</include>
                                     <include>org.apache.hbase:*</include>
-                                    <!--                                    <include>com.aliyun.odps:odps-spark-datasource_2.11:*</include>-->
                                 </includes>
                             </artifactSet>
                             <filters>

+ 2 - 0
src/main/scala/com/winhc/bigdata/spark/jobs/CompanyNameMapping.scala

@@ -72,6 +72,8 @@ object CompanyNameMapping extends Logging {
     import spark._
     val df = sql(s"select cid,name,current_cid from $inputTable")
 
+    import com.alibaba.dcm.DnsCacheManipulator
+    DnsCacheManipulator.setDnsCache("hb-uf6as8i6h85k02092-001.hbase.rds.aliyuncs.com", "47.101.251.157")
     val jobConf = SparkUtils.HBaseOutputJobConf(hbaseKVTable)
     df.rdd.map(row => {
       val id = row(0).asInstanceOf[Long].toString

+ 2 - 1
src/main/scala/com/winhc/bigdata/spark/utils/SparkUtils.scala

@@ -11,7 +11,8 @@ object SparkUtils {
 
   def HBaseOutputJobConf(outputTable: String): JobConf = {
     val config = HBaseConfiguration.create()
-    val zkAddress = "hb-uf6as8i6h85k02092-001.hbase.rds.aliyuncs.com"
+//    val zkAddress = "hb-uf6as8i6h85k02092-001.hbase.rds.aliyuncs.com"
+    val zkAddress = "hb-proxy-pub-uf6as8i6h85k02092-001.hbase.rds.aliyuncs.com"
     config.set(HConstants.ZOOKEEPER_QUORUM, zkAddress);
 
     val jobConf = new JobConf(config)