Open wangxianbin1987 opened 10 years ago
I had the same issue here. You get this error because the default property phoenix.query.timeoutMs is set to 10 minutes (https://github.com/forcedotcom/phoenix/wiki/Tuning) What you can do is to edit (or create) your hbase-site.xml in /usr/lib/phoenix/bin/ and add the phoenix.query.timeoutMs parameter as follow: (1 hour in my config)
<property>
<name>phoenix.query.timeoutMs</name>
<value>3600000</value>
</property>
ok,thanks,i will try and get back to you 2014-6-24 下午9:04于 "Charles Bernard" notifications@github.com写道:
I had the same issue here. You get this error because the default property phoenix.query.timeoutMs is set to 10 minutes (https://github.com/forcedotcom/phoenix/wiki/Tuning) What you can do is to edit (or create) your hbase-site.xml in /usr/lib/phoenix/bin/ and add the phoenix.query.timeoutMs parameter as follow: (1 hour in my config)
phoenix.query.timeoutMs 3600000 — Reply to this email directly or view it on GitHub.
Thanks @charlesb your solution worked for me.
same error i am gacing, i have set the phoenix.query.timeoutMs but cout not resolve it below is error help Error: org.apache.phoenix.exception.PhoenixIOException: Failed after attempts=36, exceptions: Fri Jan 09 09:21:07 CST 2015, null, java.net.SocketTimeoutException: callTimeout=60000, callDuration=62318: row '' on table 'PJM_DATASET' at region=PJM_DATASET,,1420633295836.4394a3aa2721f87f3e6216d20ebeec44., hostname=hadoopm1,60020,1420815633410, seqNum=34326 (state=08000,code=101)
Can you paste your hbase-site.xml? What version of hadoop (hadoop stack maybe? CDH, HDP,...) are you using?
Hi i am getting same error Please help.
ok below is my hbase configuration .
<configuration>
<property>
<name>dfs.domain.socket.path</name>
<value>/var/lib/hadoop-hdfs/dn_socket</value>
</property>
<property>
<name>hbase.client.keyvalue.maxsize</name>
<value>10485760</value>
</property>
<property>
<name>hbase.client.scanner.caching</name>
<value>100</value>
</property>
<property>
<name>hbase.client.scanner.timeout.period</name>
<value>60000000</value>
</property>
<property>
<name>hbase.cluster.distributed</name>
<value>true</value>
</property>
<property>
<name>hbase.coprocessor.master.classes</name>
<value>org.apache.phoenix.hbase.index.master.IndexMasterObserver</value>
</property>
<property>
<name>hbase.coprocessor.region.classes</name>
<value>org.apache.hadoop.hbase.coprocessor.AggregateImplementation</value>
</property>
<property>
<name>hbase.defaults.for.version.skip</name>
<value>true</value>
</property>
<property>
<name>hbase.hregion.majorcompaction</name>
<value>604800000</value>
</property>
<property>
<name>hbase.hregion.max.filesize</name>
<value>10737418240</value>
</property>
<property>
<name>hbase.hregion.memstore.block.multiplier</name>
<value>4</value>
</property>
<property>
<name>hbase.hregion.memstore.flush.size</name>
<value>536870912</value>
</property>
<property>
<name>hbase.hregion.memstore.mslab.enabled</name>
<value>true</value>
</property>
<property>
<name>hbase.hstore.blockingStoreFiles</name>
<value>10</value>
</property>
<property>
<name>hbase.hstore.compactionThreshold</name>
<value>3</value>
</property>
<property>
<name>hbase.local.dir</name>
<value>${hbase.tmp.dir}/local</value>
</property>
<property>
<name>hbase.master.info.bindAddress</name>
<value>0.0.0.0</value>
</property>
<property>
<name>hbase.master.info.port</name>
<value>60010</value>
</property>
<property>
<name>hbase.master.port</name>
<value>60000</value>
</property>
<property>
<name>hbase.regionserver.global.memstore.lowerLimit</name>
<value>0.38</value>
</property>
<property>
<name>hbase.regionserver.global.memstore.upperLimit</name>
<value>0.4</value>
</property>
<property>
<name>hbase.regionserver.handler.count</name>
<value>60</value>
</property>
<property>
<name>hbase.regionserver.info.port</name>
<value>60030</value>
</property>
<property>
<name>hbase.regionserver.wal.codec</name>
<value>org.apache.hadoop.hbase.regionserver.wal.IndexedWALEditCodec</value>
</property>
<property>
<name>hbase.rootdir</name>
<value>hdfs://hadoopctrl.dev.oati.local:8020/apps/hbase/data</value>
</property>
<property>
<name>hbase.rpc.timeout</name>
<value>1500000</value>
</property>
<property>
<name>hbase.security.authentication</name>
<value>simple</value>
</property>
<property>
<name>hbase.security.authorization</name>
<value>false</value>
</property>
<property>
<name>hbase.superuser</name>
<value>hbase</value>
</property>
<property>
<name>hbase.tmp.dir</name>
<value>/hadoop/hbase</value>
</property>
<property>
<name>hbase.zookeeper.property.clientPort</name>
<value>2181</value>
</property>
<property>
<name>hbase.zookeeper.quorum</name>
<value>hadoopm1.dev.oati.local,hadoopm2.dev.oati.local,hadoopm3.dev.oati.local</value>
</property>
<property>
<name>hbase.zookeeper.useMulti</name>
<value>true</value>
</property>
<property>
<name>hfile.block.cache.size</name>
<value>0.40</value>
</property>
<property>
<name>phoenix.client.maxMetaDataCacheSize</name>
<value>10240000</value>
</property>
<property>
<name>phoenix.clock.skew.interval</name>
<value>2000</value>
</property>
<property>
<name>phoenix.connection.autoCommit</name>
<value>false</value>
</property>
<property>
<name>phoenix.coprocessor.maxMetaDataCacheSize</name>
<value>20480000</value>
</property>
<property>
<name>phoenix.coprocessor.maxMetaDataCacheTimeToLiveM</name>
<value>180000</value>
</property>
<property>
<name>phoenix.coprocessor.maxServerCacheTimeToLiveMs</name>
<value>30000</value>
</property>
<property>
<name>phoenix.distinct.value.compress.threshold</name>
<value>1024000</value>
</property>
<property>
<name>phoenix.groupby.estimatedDistinctValues</name>
<value>1000</value>
</property>
<property>
<name>phoenix.groupby.maxCacheSize</name>
<value>102400000</value>
</property>
<property>
<name>phoenix.groupby.spillable</name>
<value>true</value>
</property>
<property>
<name>phoenix.groupby.spillFiles</name>
<value>2</value>
</property>
<property>
<name>phoenix.index.failure.handling.rebuild</name>
<value>true</value>
</property>
<property>
<name>phoenix.index.failure.handling.rebuild.interval</name>
<value>10000</value>
</property>
<property>
<name>phoenix.index.failure.handling.rebuild.overlap.time</name>
<value>300000</value>
</property>
<property>
<name>phoenix.index.maxDataFileSizePerc</name>
<value>50</value>
</property>
<property>
<name>phoenix.index.mutableBatchSizeThreshold</name>
<value>5</value>
</property>
<property>
<name>phoenix.mutate.batchSize</name>
<value>1000</value>
</property>
<property>
<name>phoenix.mutate.maxSize</name>
<value>500000</value>
</property>
<property>
<name>phoenix.query.dateFormat</name>
<value>yyyy-MM-dd HH:mm:ss</value>
</property>
<property>
<name>phoenix.query.maxGlobalMemoryPercentage</name>
<value>15</value>
</property>
<property>
<name>phoenix.query.maxGlobalMemorySize</name>
<value>2147483648</value>
</property>
<property>
<name>phoenix.query.maxGlobalMemoryWaitMs</name>
<value>10000</value>
</property>
<property>
<name>phoenix.query.maxServerCacheBytes</name>
<value>104857600</value>
</property>
<property>
<name>phoenix.query.maxSpoolToDiskBytes</name>
<value>1024000000</value>
</property>
<property>
<name>phoenix.query.maxTenantMemoryPercentage</name>
<value>100</value>
</property>
<property>
<name>phoenix.query.numberFormat</name>
<value>#,##0.###</value>
</property>
<property>
<name>phoenix.query.rowKeyOrderSaltedTable</name>
<value>true</value>
</property>
<property>
<name>phoenix.query.spoolThresholdBytes</name>
<value>20971520</value>
</property>
<property>
<name>phoenix.query.useIndexes</name>
<value>true</value>
</property>
<property>
<name>phoenix.schema.dropMetaData</name>
<value>true</value>
</property>
<property>
<name>phoenix.sequence.cacheSize</name>
<value>100</value>
</property>
<property>
<name>phoenix.stats.guidepost.per.region</name>
<value>None</value>
</property>
<property>
<name>phoenix.stats.minUpdateFrequency</name>
<value>450000</value>
</property>
<property>
<name>phoenix.stats.updateFrequency</name>
<value>900000</value>
</property>
<property>
<name>phoenix.stats.useCurrentTime</name>
<value>true</value>
</property>
<property>
<name>zookeeper.session.timeout</name>
<value>120000000</value>
</property>
<property>
<name>zookeeper.znode.parent</name>
<value>/hbase</value>
</property>
and i am using hdp 2.2 i have 4000000000 rows in my phoenix table . please help asap.
i have setup the phoenix.query.timeoutMs property from ambari web UI but it is not reflecting in hbase-site.xml file.
<configuration>
<property>
<name>dfs.domain.socket.path</name>
<value>/var/lib/hadoop-hdfs/dn_socket</value>
</property>
<property>
<name>hbase.client.keyvalue.maxsize</name>
<value>10485760</value>
</property>
<property>
<name>hbase.client.scanner.caching</name>
<value>100</value>
</property>
<property>
<name>hbase.client.scanner.timeout.period</name>
<value>60000000</value>
</property>
<property>
<name>hbase.cluster.distributed</name>
<value>true</value>
</property>
<property>
<name>hbase.coprocessor.master.classes</name>
<value>org.apache.phoenix.hbase.index.master.IndexMasterObserver</value>
</property>
<property>
<name>hbase.coprocessor.region.classes</name>
<value>org.apache.hadoop.hbase.coprocessor.AggregateImplementation</value>
</property>
<property>
<name>hbase.defaults.for.version.skip</name>
<value>true</value>
</property>
<property>
<name>hbase.hregion.majorcompaction</name>
<value>604800000</value>
</property>
<property>
<name>hbase.hregion.majorcompaction.jitter</name>
<value>0.50</value>
</property>
<property>
<name>hbase.hregion.max.filesize</name>
<value>10737418240</value>
</property>
<property>
<name>hbase.hregion.memstore.block.multiplier</name>
<value>4</value>
</property>
<property>
<name>hbase.hregion.memstore.flush.size</name>
<value>536870912</value>
</property>
<property>
<name>hbase.hregion.memstore.mslab.enabled</name>
<value>true</value>
</property>
<property>
<name>hbase.hstore.blockingStoreFiles</name>
<value>10</value>
</property>
<property>
<name>hbase.hstore.compactionThreshold</name>
<value>3</value>
</property>
<property>
<name>hbase.local.dir</name>
<value>${hbase.tmp.dir}/local</value>
</property>
<property>
<name>hbase.master.info.bindAddress</name>
<value>0.0.0.0</value>
</property>
<property>
<name>hbase.master.info.port</name>
<value>60010</value>
</property>
<property>
<name>hbase.master.loadbalancer.class</name>
<value>org.apache.phoenix.hbase.index.balancer.IndexLoadBalancer</value>
</property>
<property>
<name>hbase.master.port</name>
<value>60000</value>
</property>
<property>
<name>hbase.region.server.rpc.scheduler.factory.class</name>
<value>org.apache.phoenix.hbase.index.ipc.PhoenixIndexRpcSchedulerFactory</value>
</property>
<property>
<name>hbase.regionserver.global.memstore.lowerLimit</name>
<value>0.38</value>
</property>
<property>
<name>hbase.regionserver.global.memstore.upperLimit</name>
<value>0.4</value>
</property>
<property>
<name>hbase.regionserver.handler.count</name>
<value>60</value>
</property>
<property>
<name>hbase.regionserver.info.port</name>
<value>60030</value>
</property>
<property>
<name>hbase.regionserver.wal.codec</name>
<value>org.apache.hadoop.hbase.regionserver.wal.IndexedWALEditCodec</value>
</property>
<property>
<name>hbase.rootdir</name>
<value>hdfs://hadoopctrl.dev.oati.local:8020/apps/hbase/data</value>
</property>
<property>
<name>hbase.rpc.timeout</name>
<value>1500000</value>
</property>
<property>
<name>hbase.security.authentication</name>
<value>simple</value>
</property>
<property>
<name>hbase.security.authorization</name>
<value>false</value>
</property>
<property>
<name>hbase.superuser</name>
<value>hbase</value>
</property>
<property>
<name>hbase.tmp.dir</name>
<value>/hadoop/hbase</value>
</property>
<property>
<name>hbase.zookeeper.property.clientPort</name>
<value>2181</value>
</property>
<property>
<name>hbase.zookeeper.quorum</name>
<value>hadoopm1.dev.oati.local,hadoopm2.dev.oati.local,hadoopm3.dev.oati.local</value>
</property>
<property>
<name>hbase.zookeeper.useMulti</name>
<value>true</value>
</property>
<property>
<name>hfile.block.cache.size</name>
<value>0.40</value>
</property>
<property>
<name>phoenix.client.maxMetaDataCacheSize</name>
<value>10240000</value>
</property>
<property>
<name>phoenix.clock.skew.interval</name>
<value>2000</value>
</property>
<property>
<name>phoenix.connection.autoCommit</name>
<value>false</value>
</property>
<property>
<name>phoenix.coprocessor.maxMetaDataCacheSize</name>
<value>20480000</value>
</property>
<property>
<name>phoenix.coprocessor.maxMetaDataCacheTimeToLiveM</name>
<value>180000</value>
</property>
<property>
<name>phoenix.coprocessor.maxServerCacheTimeToLiveMs</name>
<value>30000</value>
</property>
<property>
<name>phoenix.distinct.value.compress.threshold</name>
<value>1024000</value>
</property>
<property>
<name>phoenix.groupby.estimatedDistinctValues</name>
<value>1000</value>
</property>
<property>
<name>phoenix.groupby.maxCacheSize</name>
<value>102400000</value>
</property>
<property>
<name>phoenix.groupby.spillable</name>
<value>true</value>
</property>
<property>
<name>phoenix.groupby.spillFiles</name>
<value>2</value>
</property>
<property>
<name>phoenix.index.failure.handling.rebuild</name>
<value>true</value>
</property>
<property>
<name>phoenix.index.failure.handling.rebuild.interval</name>
<value>10000</value>
</property>
<property>
<name>phoenix.index.failure.handling.rebuild.overlap.time</name>
<value>300000</value>
</property>
<property>
<name>phoenix.index.maxDataFileSizePerc</name>
<value>50</value>
</property>
<property>
<name>phoenix.index.mutableBatchSizeThreshold</name>
<value>5</value>
</property>
<property>
<name>phoenix.mutate.batchSize</name>
<value>1000</value>
</property>
<property>
<name>phoenix.mutate.maxSize</name>
<value>500000</value>
</property>
<property>
<name>phoenix.query.dateFormat</name>
<value>yyyy-MM-dd HH:mm:ss</value>
</property>
<property>
<name>phoenix.query.maxGlobalMemoryPercentage</name>
<value>15</value>
</property>
<property>
<name>phoenix.query.maxGlobalMemorySize</name>
<value>2147483648</value>
</property>
<property>
<name>phoenix.query.maxGlobalMemoryWaitMs</name>
<value>10000</value>
</property>
<property>
<name>phoenix.query.maxServerCacheBytes</name>
<value>104857600</value>
</property>
<property>
<name>phoenix.query.maxSpoolToDiskBytes</name>
<value>1024000000</value>
</property>
<property>
<name>phoenix.query.maxTenantMemoryPercentage</name>
<value>100</value>
</property>
<property>
<name>phoenix.query.numberFormat</name>
<value>#,##0.###</value>
</property>
<property>
<name>phoenix.query.rowKeyOrderSaltedTable</name>
<value>true</value>
</property>
<property>
<name>phoenix.query.spoolThresholdBytes</name>
<value>20971520</value>
</property>
<property>
<name>phoenix.query.timeoutMs</name>
<value>60000000</value>
</property>
<property>
<name>phoenix.query.useIndexes</name>
<value>true</value>
</property>
<property>
<name>phoenix.schema.dropMetaData</name>
<value>true</value>
</property>
<property>
<name>phoenix.sequence.cacheSize</name>
<value>100</value>
</property>
<property>
<name>phoenix.stats.guidepost.per.region</name>
<value>None</value>
</property>
<property>
<name>phoenix.stats.minUpdateFrequency</name>
<value>450000</value>
</property>
<property>
<name>phoenix.stats.updateFrequency</name>
<value>900000</value>
</property>
<property>
<name>phoenix.stats.useCurrentTime</name>
<value>true</value>
</property>
<property>
<name>zookeeper.session.timeout</name>
<value>120000000</value>
</property>
<property>
<name>zookeeper.znode.parent</name>
<value>/hbase</value>
</property>
ok here is final hbase-site.configuration
jdbc:phoenix:hadoopm1> Select count(*) from PJM_DATASET; here is query and below is the execption. java.lang.RuntimeException: org.apache.phoenix.exception.PhoenixIOException: org.apache.phoenix.exception.PhoenixIOException: Failed after attempts=36, exceptions: Sat Jan 10 05:06:02 CST 2015, null, java.net.SocketTimeoutException: callTimeout=60000, callDuration=62320: row '' on table 'PJM_DATASET' at region=PJM_DATASET,,1420633295836.4394a3aa2721f87f3e6216d20ebeec44., hostname=hadoopm1,60020,1420887782278, seqNum=34350
at sqlline.SqlLine$IncrementalRows.hasNext(SqlLine.java:2440)
at sqlline.SqlLine$TableOutputFormat.print(SqlLine.java:2074)
at sqlline.SqlLine.print(SqlLine.java:1735)
at sqlline.SqlLine$Commands.execute(SqlLine.java:3683)
at sqlline.SqlLine$Commands.sql(SqlLine.java:3584)
at sqlline.SqlLine.dispatch(SqlLine.java:821)
at sqlline.SqlLine.begin(SqlLine.java:699)
at sqlline.SqlLine.mainWithInputRedirection(SqlLine.java:441)
at sqlline.SqlLine.main(SqlLine.java:424)
0: jdbc:phoenix:hadoopm1>
please help what wrong i am doing.
Seems like something else times out. Have you tried to scan this table from your hbase client (hbase shell)? Check this: http://hbase.apache.org/book/ch15s15.html (pragraph titled Connection Timeouts)
i have one master and three slaves .. i uninstall and reinstall the hbase and phoenix at master and install the hbase at other slave machin but now i am even not able to start the hbase master from ambari.. web UI
RegionServer
at org.apache.hadoop.hbase.regionserver.HRegionServer.constructRegionServer(HRegionServer.java:2486)
at org.apache.hadoop.hbase.regionserver.HRegionServerCommandLine.start(HRegionServerCommandLine.java:61)
at org.apache.hadoop.hbase.regionserver.HRegionServerCommandLine.run(HRegionServerCommandLine.java:85)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at org.apache.hadoop.hbase.util.ServerCommandLine.doMain(ServerCommandLine.java:126)
at org.apache.hadoop.hbase.regionserver.HRegionServer.main(HRegionServer.java:2501)
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
at org.apache.hadoop.hbase.regionserver.HRegionServer.constructRegionServer(HRegionServer.java:2484)
... 5 more
Caused by: java.lang.RuntimeException: java.lang.ClassNotFoundException: Class org.apache.phoenix.hbase.index.ipc.PhoenixIndexRpcSchedulerFactory not found
at org.apache.hadoop.conf.Configuration.getClass(Configuration.java:2076)
at org.apache.hadoop.hbase.regionserver.HRegionServer.
above is the current error i got from hbase regionserver log.
ok one question do we need to installed phoenix at the same machine where we have region server. ?
Phoenix moved to Apache over a year ago, so this site is no longer active nor maintained. Please post your question on our Apache mailing list and you'll likely get more help: http://phoenix.apache.org/mailing_list.html
ok thanks i will
2015-01-14 02:27:57,512 WARN [DataStreamer for file /apps/hbase/data/WALs/hadoopm2.dev.oati.local,60020,1421221188209/hadoopm2.dev.oati.local%2C60020%2C1421221188209.1421223957430 block BP-337983189-10.100.227.107-1397418605845:blk_1073948934_216462] hdfs.DFSClient: DataStreamer Exception java.io.IOException: Failed to replace a bad datanode on the existing pipeline due to no more good datanodes being available to try. (Nodes: current=[10.100.227.107:50010, 10.100.227.104:50010], original=[10.100.227.107:50010, 10.100.227.104:50010]). The current failed datanode replacement policy is DEFAULT, and a client may configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' in its configuration. at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.findNewDatanode(DFSOutputStream.java:1041) at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.addDatanode2ExistingPipeline(DFSOutputStream.java:1107) at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1254) at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.processDatanodeError(DFSOutputStream.java:1005) at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:549) Wed Jan 14 04:26:19 CST 2015 Terminating regionserver 2015-01-14 04:26:19,509 INFO [Thread-11] regionserver.ShutdownHook: Shutdown hook starting; hbase.shutdown.hook=true; fsShutdownHook=org.apache.hadoop.fs.FileSystem$Cache$ClientFinalizer@48a9bc2b Wed Jan 14 04:30:46 CST 2015 Terminating regionserver Wed Jan 14 04:34:13 CST 2015 Terminating regionserver 2015-01-14 04:34:32,959 DEBUG [regionserver60020-EventThread] regionserver.SplitLogWorker: tasks arrived or departed Wed Jan 14 04:58:39 CST 2015 Terminating regionserver Wed Jan 14 05:34:36 CST 2015 Terminating regionserver
now i am getting this error on region server .
we use performance.py script to generate a bigtable with 50,000,000 rows, and then we create index on it using SQL: CREATE index iPERFORMANCE_50000000 on PERFORMANCE_50000000(core) include(db, ACTIVE_VISITOR); after about 10 mins, it end with Error: (state=08000,code=101), any suggest, thanks!