hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Ted Yu <yuzhih...@gmail.com>
Subject Re: Tool to to execute an benchmark for HBase.
Date Thu, 29 Jan 2015 15:26:05 GMT
What's the value for hfile.block.cache.size ?

By default it is 40%. You may want to increase its value if you're using
default.

Andrew published some ycsb results :
http://people.apache.org/~apurtell/results-ycsb-0.98.8/ycsb
-0.98.0-vs-0.98.8.pdf

However, I couldn't access the above now.

Cheers

On Thu, Jan 29, 2015 at 7:14 AM, Guillermo Ortiz <konstt2000@gmail.com>
wrote:

> Is there any result with that benchmark to compare??
> I'm executing the different workloads and for example for 100% Reads
> in a table with 10Millions of records I only get an performance of
> 2000operations/sec. I hoped much better performance but I could be
> wrong. I'd like to know if it's a normal performance or I could have
> something bad configured.
>
>
> I have splitted the tabled and all the records are balanced and used
> snappy.
> The cluster has a master and 4 regions servers with 256Gb,Cores 2 (32
> w/ Hyperthreading), 0.98.6-cdh5.3.0,
>
> RegionServer is executed with these parameters:
>  /usr/java/jdk1.7.0_67-cloudera/bin/java -Dproc_regionserver
> -XX:OnOutOfMemoryError=kill -9 %p -Xmx1000m
> -Djava.net.preferIPv4Stack=true -Xms640679936 -Xmx640679936
> -XX:+UseParNewGC -XX:+UseConcMarkSweepGC -XX:-CMSConcurrentMTEnabled
> -XX:CMSInitiatingOccupancyFraction=70 -XX:+CMSParallelRemarkEnabled
> -XX:OnOutOfMemoryError=/usr/lib64/cmf/service/common/killparent.sh
> -Dhbase.log.dir=/var/log/hbase
>
> -Dhbase.log.file=hbase-cmf-hbase-REGIONSERVER-cnsalbsrvcl23.lvtc.gsnet.corp.log.out
> -Dhbase.home.dir=/opt/cloudera/parcels/CDH-5.3.0-1.cdh5.3.0.p0.30/lib/hbase
> -Dhbase.id.str= -Dhbase.root.logger=INFO,RFA
>
> -Djava.library.path=/opt/cloudera/parcels/CDH-5.3.0-1.cdh5.3.0.p0.30/lib/hadoop/lib/native
> -Dhbase.security.logger=INFO,RFAS
> org.apache.hadoop.hbase.regionserver.HRegionServer start
>
>
> The results for 100% reads are
> [OVERALL], RunTime(ms), 42734.0
> [OVERALL], Throughput(ops/sec), 2340.0570973931763
> [UPDATE], Operations, 1.0
> [UPDATE], AverageLatency(us), 103170.0
> [UPDATE], MinLatency(us), 103168.0
> [UPDATE], MaxLatency(us), 103171.0
> [UPDATE], 95thPercentileLatency(ms), 103.0
> [UPDATE], 99thPercentileLatency(ms), 103.0
> [READ], Operations, 100000.0
> [READ], AverageLatency(us), 412.5534
> [READ], AverageLatency(us,corrected), 581.6249026771276
> [READ], MinLatency(us), 218.0
> [READ], MaxLatency(us), 268383.0
> [READ], MaxLatency(us,corrected), 268383.0
> [READ], 95thPercentileLatency(ms), 0.0
> [READ], 95thPercentileLatency(ms,corrected), 0.0
> [READ], 99thPercentileLatency(ms), 0.0
> [READ], 99thPercentileLatency(ms,corrected), 0.0
> [READ], Return=0, 100000
> [CLEANUP], Operations, 1.0
> [CLEANUP], AverageLatency(us), 103598.0
> [CLEANUP], MinLatency(us), 103596.0
> [CLEANUP], MaxLatency(us), 103599.0
> [CLEANUP], 95thPercentileLatency(ms), 103.0
> [CLEANUP], 99thPercentileLatency(ms), 103.0
>
> hbase(main):030:0> describe 'username'
> DESCRIPTION
>                                     ENABLED
>  'username', {NAME => 'cf', DATA_BLOCK_ENCODING => 'NONE', BLOOMFILTER
> => 'ROW', REPLICATION_SCOPE => '0', true
>   VERSIONS => '1', COMPRESSION => 'SNAPPY', MIN_VERSIONS => '0', TTL
> => 'FOREVER', KEEP_DELETED_CELLS => '
>  false', BLOCKSIZE => '65536', IN_MEMORY => 'false', BLOCKCACHE => 'true'}
> 1 row(s) in 0.0170 seconds
>
> 2015-01-29 5:27 GMT+01:00 Ted Yu <yuzhihong@gmail.com>:
> > Maybe ask on Cassandra mailing list for the benchmark tool they use ?
> >
> > Cheers
> >
> > On Wed, Jan 28, 2015 at 1:23 PM, Guillermo Ortiz <konstt2000@gmail.com>
> > wrote:
> >
> >> I was checking that web, do you know if there's another possibility
> >> since last updated for Cassandra was two years ago and I'd like to
> >> compare bothof them with kind of same tool/code.
> >>
> >> 2015-01-28 22:10 GMT+01:00 Ted Yu <yuzhihong@gmail.com>:
> >> > Guillermo:
> >> > If you use hbase 0.98.x, please consider Andrew's ycsb repo:
> >> >
> >> > https://github.com/apurtell/ycsb/tree/new_hbase_client
> >> >
> >> > Cheers
> >> >
> >> > On Wed, Jan 28, 2015 at 12:41 PM, Nishanth S <nishanth.2884@gmail.com
> >
> >> > wrote:
> >> >
> >> >> You can use ycsb for this purpose.See here
> >> >>
> >> >> https://github.com/brianfrankcooper/YCSB/wiki/Getting-Started
> >> >> -Nishanth
> >> >>
> >> >> On Wed, Jan 28, 2015 at 1:37 PM, Guillermo Ortiz <
> konstt2000@gmail.com>
> >> >> wrote:
> >> >>
> >> >> > Hi,
> >> >> >
> >> >> > I'd like to do some benchmarks fo HBase but I don't know what
tool
> >> >> > could use. I started to make some code but I guess that there're
> some
> >> >> > easier.
> >> >> >
> >> >> > I've taken a look to JMeter, but I guess that I'd attack directly
> from
> >> >> > Java, JMeter looks great but I don't know if it fits well in this
> >> >> > scenario. What tool could I use to take some measures as time
to
> >> >> > response some read and write request, etc. I'd like that to be
> able to
> >> >> > make the same benchmarks to Cassandra.
> >> >> >
> >> >>
> >>
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message