hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From rgaimari <rgaim...@gmail.com>
Subject Re: Endpoint and Observer work together?
Date Wed, 02 Oct 2013 14:21:21 GMT

I've created some demo code to show the problem.  Here's the Observer:

public class DemoObserver extends BaseRegionObserver {
  byte[] personFamily = Bytes.toBytes("Person");
  public boolean postScannerNext(
	ObserverContext<RegionCoprocessorEnvironment> e, InternalScanner s,
	List<Result> results, int limit, boolean hasMore)
	throws IOException {
    List<Result> newResults = new ArrayList<Result>();
    for (Result result : results) {
	List<KeyValue> newKVList = new ArrayList<KeyValue>();
	for (KeyValue kv : result.list()) {
	  String newVal = Bytes.toString(kv.getValue()).toUpperCase();
	  newKVList.add(new KeyValue(kv.getRow(), kv.getFamily(),
kv.getQualifier(), kv.getTimestamp(), 
	newResults.add(new Result(newKVList));
    return super.postScannerNext(e, s, results, limit, hasMore);

And here's the Endpoint:

public class DemoEndpoint extends BaseEndpointCoprocessor implements
	DemoProtocol {

  public List<KeyValue> scanRows(Filter filter) throws IOException {
    Scan scan = new Scan();
    RegionCoprocessorEnvironment env =
    InternalScanner scanner = env.getRegion().getScanner(scan);
    List<KeyValue> retValues = new ArrayList<KeyValue>();
    boolean more = false;
    List<KeyValue> res = new ArrayList<KeyValue>();
    do {
	more = scanner.next(res);
	if (res != null)
    } while (more);
    return retValues;

They are loaded in separate jar files, and they are both attached to the

1.9.3p448 :009 >   describe 'Demo'
DESCRIPTION                                                        ENABLED               
 {NAME => 'Demo', coprocessor$2 => 'hdfs:///user/hduser/DemoEndpoi true            
 nt.jar|demo.DemoEndpoint|1001', coprocessor$1 => 'hdfs:///user/hd                    
 user/DemoObserver.jar|demo.DemoObserver|1', FAMILIES => [{NAME =>                 
  'Person', DATA_BLOCK_ENCODING => 'NONE', BLOOMFILTER => 'NONE',                  
 REPLICATION_SCOPE => '0', VERSIONS => '3', COMPRESSION => 'NONE',              
  MIN_VERSIONS => '0', TTL => '2147483647', KEEP_DELETED_CELLS =>               
 'false', BLOCKSIZE => '65536', IN_MEMORY => 'false', ENCODE_ON_DI                 
 SK => 'true', BLOCKCACHE => 'true'}]}                                             
1 row(s) in 0.0880 seconds

If I run a test where I do a scan directly on the client (with no filter), I
get the following results:

<row01>, <Person>, <Address>, <123 MAIN STREET>
<row01>, <Person>, <Name>, <JOHN SMITH>
<row02>, <Person>, <Address>, <234 ELM STREET>
<row02>, <Person>, <Name>, <LARRY DAVID>
<row03>, <Person>, <Address>, <345 SCOTT STREET>
<row03>, <Person>, <Name>, <JOHN DOE>

The values are all capitalized, as the Observer was supposed to do. 
However, if I then run a scan through the Endpoint coprocessor (with a
filter just looking for the name "john"), I get the following results:

<row01>, <Person>, <Address>, <123 main street>
<row01>, <Person>, <Name>, <john smith>
<row03>, <Person>, <Address>, <345 scott street>
<row03>, <Person>, <Name>, <john doe>

It's filtered properly, but the values don't go through the Observer and
aren't capitalized.

If there is any other info you need to help diagnose this, please let me
know. Thanks.

- Bob Gaimari

View this message in context: http://apache-hbase.679495.n3.nabble.com/Endpoint-and-Observer-work-together-tp4051383p4051395.html
Sent from the HBase User mailing list archive at Nabble.com.

View raw message