凤凰表选择连接返回失败,尝试次数为11次。

huangapple go评论71阅读模式
英文:

phoenix table select join return Failed after attempts=11

问题

我在执行以下语句时遇到了错误:

SELECT r.*, 
       l.productid, 
       l.prodname, 
       l.projnum, 
       l.projname, 
       l.capitalnumb, 
       l.fundname, 
       l.channelno, 
       l.channelname, 
       l.merchantno, 
       l.merchantname, 
       l.staff_name, 
       l.busipers 
FROM   lake.lake_retu_record r 
       LEFT JOIN lake.lake_loan l 
              ON r.loanacno = l.loanacno 
WHERE  r.acdate = '20200912000000' 
LIMIT  1000 

表信息:

lake.lake_retu_record 大小:40MB 数据:44000 条数据
lake.lake_loan 大小:1.5GB 数据: 1700000 条数据

版本:

HBase 版本: 2.1.0
Phoenix 核心: 5.0.0-cdh-6.2.0

我在网上搜索了很长时间,但没有解决这个问题。请帮忙或者尝试提供一些解决方法。
我尝试修改了这些参数 phoenix.query.queueSize phoenix.query.threadPoolSize phoenix.query.targetConcurrency phoenix.query.maxConcurrency

异常:

20/10/18 10:35:50 WARN client.SyncCoprocessorRpcChannel: IOException 上的调用失败
    org.apache.hadoop.hbase.client.RetriesExhaustedException: 尝试后失败=11次,异常:
    Sun Oct 18 10:35:08 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, java.io.IOException: 调用到 CQ-HM-ALIFC-test/10.10.10.10:16020 失败,本地异常: java.io.IOException: 连接被对等方重置
    Sun Oct 18 10:35:08 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: 调用到 CQ-HM-ALIFC-test/10.10.10.10:16020 失败,本地异常: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: 连接已关闭
    Sun Oct 18 10:35:09 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: 调用到 CQ-HM-ALIFC-test/10.10.10.10:16020 失败,本地异常: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: 连接已关闭
    ... 更多异常行...

有人可以帮帮我吗..!

英文:

I get an error when I execute the following statement:

SELECT r.*, 
       l.productid, 
       l.prodname, 
       l.projnum, 
       l.projname, 
       l.capitalnumb, 
       l.fundname, 
       l.channelno, 
       l.channelname, 
       l.merchantno, 
       l.merchantname, 
       l.staff_name, 
       l.busipers 
FROM   lake.lake_retu_record r 
       LEFT JOIN lake.lake_loan l 
              ON r.loanacno = l.loanacno 
WHERE  r.acdate = '20200912000000' 
LIMIT  1000 

table information:<br/>

lake.lake_retu_record size:40MB data:44000 pieces of data<br/>
lake.lake_loan size:1.5GB data: 1700000 pieces of data

version:<br/>

HBase-Version : 2.1.0<br/>
Phoenix-core: 5.0.0-cdh-6.2.0

I am searching for a long time on net. But it doesn't solve the problem. Please help or try to give some ideas how to solve this.<br/>
I tried to modify these parameters phoenix.query.queueSize phoenix.query.threadPoolSize phoenix.query.targetConcurrency phoenix.query.maxConcurrency

Exception:<br/>

    20/10/18 10:35:50 WARN client.SyncCoprocessorRpcChannel: Call failed on IOException
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=11, exceptions:
Sun Oct 18 10:35:08 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, java.io.IOException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: java.io.IOException: Connection reset by peer
Sun Oct 18 10:35:08 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
Sun Oct 18 10:35:09 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
Sun Oct 18 10:35:10 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
Sun Oct 18 10:35:11 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
Sun Oct 18 10:35:12 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
Sun Oct 18 10:35:15 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
Sun Oct 18 10:35:19 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
Sun Oct 18 10:35:29 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
Sun Oct 18 10:35:40 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
Sun Oct 18 10:35:50 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
at org.apache.hadoop.hbase.client.RpcRetryingCallerImpl.callWithRetries(RpcRetryingCallerImpl.java:145)
at org.apache.hadoop.hbase.client.RegionCoprocessorRpcChannel.callExecService(RegionCoprocessorRpcChannel.java:91)
at org.apache.hadoop.hbase.client.SyncCoprocessorRpcChannel.callMethod(SyncCoprocessorRpcChannel.java:52)
at org.apache.phoenix.coprocessor.generated.ServerCachingProtos$ServerCachingService$Stub.addServerCache(ServerCachingProtos.java:8484)
at org.apache.phoenix.cache.ServerCacheClient$3.call(ServerCacheClient.java:494)
at org.apache.phoenix.cache.ServerCacheClient$3.call(ServerCacheClient.java:462)
at org.apache.hadoop.hbase.client.HTable$12.call(HTable.java:997)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
at org.apache.hadoop.hbase.ipc.IPCUtil.wrapException(IPCUtil.java:202)
at org.apache.hadoop.hbase.ipc.AbstractRpcClient.onCallFinished(AbstractRpcClient.java:390)
at org.apache.hadoop.hbase.ipc.AbstractRpcClient.access$100(AbstractRpcClient.java:95)
at org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:410)
at org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:406)
at org.apache.hadoop.hbase.ipc.Call.callComplete(Call.java:103)
at org.apache.hadoop.hbase.ipc.Call.setException(Call.java:118)
at org.apache.hadoop.hbase.ipc.NettyRpcDuplexHandler.cleanupCalls(NettyRpcDuplexHandler.java:203)
at org.apache.hadoop.hbase.ipc.NettyRpcDuplexHandler.channelInactive(NettyRpcDuplexHandler.java:211)
at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:245)
at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:231)
at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:224)
at org.apache.hbase.thirdparty.io.netty.handler.codec.ByteToMessageDecoder.channelInputClosed(ByteToMessageDecoder.java:377)
at org.apache.hbase.thirdparty.io.netty.handler.codec.ByteToMessageDecoder.channelInactive(ByteToMessageDecoder.java:342)
at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:245)
at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:231)
at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:224)
at org.apache.hbase.thirdparty.io.netty.channel.ChannelInboundHandlerAdapter.channelInactive(ChannelInboundHandlerAdapter.java:75)
at org.apache.hbase.thirdparty.io.netty.handler.timeout.IdleStateHandler.channelInactive(IdleStateHandler.java:277)
at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:245)
at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:231)
at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:224)
at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPipeline$HeadContext.channelInactive(DefaultChannelPipeline.java:1354)
at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:245)
at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:231)
at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPipeline.fireChannelInactive(DefaultChannelPipeline.java:917)
at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannel$AbstractUnsafe$8.run(AbstractChannel.java:822)
at org.apache.hbase.thirdparty.io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163)
at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:403)
at org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:463)
at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:858)
at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:138)
... 1 more
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
... 25 more

Can some one help me out..!

答案1

得分: 0

谢谢大家,我解决了这个问题!<br/>

解决方法如下:<br/>
我找到了异常的主要原因<br/>

org.apache.hadoop.hbase.client.RetriesExhaustedException: 尝试11次后失败,异常:
2020年10月18日,中国标准时间上午10:35:08,RpcRetryingCaller{globalStartTime=1602988506012,暂停=100,最多尝试11次},java.io.IOException:对CQ-HM-ALIFC-test/10.10.10.10:16020的调用失败,本地异常:java.io.IOException:连接被对等方重置

初步判断是RPC出了问题,所以在执行SQL语句时,我打开了报告错误的服务器的RegionServer日志。我发现RegionServer生成了以下日志:<br/>

警告 org.apache.hadoop.hbase.ipc.NettyRpcServer:从10.10.10.10接收的RPC数据长度为353247227,超过了允许的最大值268435456。在服务器上设置“hbase.ipc.max.request.size”以覆盖此限制(不建议这样做)

我增加了这个参数,问题得以解决!

英文:

thx everyone,I solved the problem!<br/>

The solutions are as follows:<br/>
I found the main reason for the abnormality<br/>

org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=11, exceptions:
Sun Oct 18 10:35:08 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, java.io.IOException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: java.io.IOException: Connection reset by peer

The preliminary judgment is that the RPC has problems,So when I execute the SQL statement, I open the RegionServer log of the server where the error is reported.I found that RegionServer generates the following logs:<br/>

WARN org.apache.hadoop.hbase.ipc.NettyRpcServer: RPC data length of 353247227 received from 10.10.10.10 is greater than max allowed 268435456. Set &quot;hbase.ipc.max.request.size&quot; on server to override this limit (not recommended)

I increased the parameter and solved the problem!

huangapple
  • 本文由 发表于 2020年10月18日 13:07:31
  • 转载请务必保留本文链接:https://go.coder-hub.com/64409921.html
匿名

发表评论

匿名网友

:?: :razz: :sad: :evil: :!: :smile: :oops: :grin: :eek: :shock: :???: :cool: :lol: :mad: :twisted: :roll: :wink: :idea: :arrow: :neutral: :cry: :mrgreen:

确定