[cdh-user] Failed shortcircuit with CDH5.3.1: shortcircuit.DomainSocketFactory: error creating DomainSocket
I am running hbase performanceEvaluation with MR1 installing of CDH5.3.1 but get below error: 2015-01-05 04:00:33,783 WARN  [main] mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.2015-01-05 04:00:37,178 INFO  [main] input.FileInputFormat: Total input paths to process : 12015-01-05 04:00:37,198 WARN  [main] shortcircuit.DomainSocketFactory: error creating DomainSocketjava.net.ConnectException: connect(2) error: No such file or directory when trying to connect to '/var/run/hadoop-hdfs/dn.50010'        at org.apache.hadoop.net.unix.DomainSocket.connect0(Native Method)        at org.apache.hadoop.net.unix.DomainSocket.connect(DomainSocket.java:250)        at org.apache.hadoop.hdfs.shortcircuit.DomainSocketFactory.createSocket(DomainSocketFactory.java:163)        at org.apache.hadoop.hdfs.BlockReaderFactory.nextDomainPeer(BlockReaderFactory.java:719)        at org.apache.hadoop.hdfs.BlockReaderFactory.createShortCircuitReplicaInfo(BlockReaderFactory.java:441)        at org.apache.hadoop.hdfs.shortcircuit.ShortCircuitCache.create(ShortCircuitCache.java:783)        at org.apache.hadoop.hdfs.shortcircuit.ShortCircuitCache.fetchOrCreate(ShortCircuitCache.java:717)        at org.apache.hadoop.hdfs.BlockReaderFactory.getBlockReaderLocal(BlockReaderFactory.java:394)        at org.apache.hadoop.hdfs.BlockReaderFactory.build(BlockReaderFactory.java:305)        at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:574)        at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy(DFSInputStream.java:797)        at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:844)        at java.io.DataInputStream.read(DataInputStream.java:100)        at org.apache.hadoop.util.LineReader.fillBuffer(LineReader.java:180)        at org.apache.hadoop.util.LineReader.readDefaultLine(LineReader.java:216)        at org.apache.hadoop.util.LineReader.readLine(LineReader.java:174)        at org.apache.hadoop.util.LineReader.readLine(LineReader.java:370)        at org.apache.hadoop.mapreduce.lib.input.NLineInputFormat.getSplitsForFile(NLineInputFormat.java:102)        at org.apache.hadoop.mapreduce.lib.input.NLineInputFormat.getSplits(NLineInputFormat.java:79)        at org.apache.hadoop.mapred.JobClient.writeNewSplits(JobClient.java:1107)        at org.apache.hadoop.mapred.JobClient.writeSplits(JobClient.java:1124)        at org.apache.hadoop.mapred.JobClient.access$600(JobClient.java:178)        at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:1023)        at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:976)        at java.security.AccessController.doPrivileged(Native Method)        at javax.security.auth.Subject.doAs(Subject.java:415)        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1642)        at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:976)        at org.apache.hadoop.mapreduce.Job.submit(Job.java:582)        at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:612)        at org.apache.hadoop.hbase.PerformanceEvaluation.doMapReduce(PerformanceEvaluation.java:409)        at org.apache.hadoop.hbase.PerformanceEvaluation.runTest(PerformanceEvaluation.java:1080)        at org.apache.hadoop.hbase.PerformanceEvaluation.run(PerformanceEvaluation.java:1282)        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)        at org.apache.hadoop.hbase.PerformanceEvaluation.main(PerformanceEvaluation.java:1303)2015-01-05 04:00:37,203 WARN  [main] shortcircuit.ShortCircuitCache: ShortCircuitCache(0x76b06322): failed to load 1073741837_BP-1525690555-10.154.8.10-1420448369642Configuration in my hdfs-site.xml:      <property>        <name>dfs.block.local-path-access.user</name>        <value>root</value>      </property>      <property>        <name>dfs.client.read.shortcircuit.streams.cache.size</name>        <value>1000</value>      </property>      <property>        <name>dfs.client.read.shortcircuit.streams.cache.size.expiry.ms</name>        <value>1000</value>      </property>      <property>        <name>dfs.client.read.shortcircuit.streams.cache.size</name>        <value>1000000</value>      </property>      <property>        <name>dfs.client.read.shortcircuit.streams.cache.size.expiry.ms</name>        <value>1600000</value>      </property>      <property>        <name>dfs.client.domain.socket.data.traffic</name>        <value>false</value>      </property>      <property>        <name>dfs.domain.socket.path</name>        <value>/var/run/hadoop-hdfs/dn._PORT</value>      </property>Configuration in my hbase-site.xml:      <property>        <name>dfs.client.read.shortcircuit</name>        <value>true</value>      </property>      <property>        <name>hbase.regionserver.checksum.verify</name>        <value>true</value>      </property>What wrong with it? I install hadoop MR1 and Hbase from RPM and modify the configs manually. Anything I forgot to enable the socket path?
 
 
 
          
      
 
  
 
 
 
 
 
 
 
 
 
 
댓글 없음:
댓글 쓰기