org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Java Examples
The following examples show how to use
org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint.
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example #1
Source File: TestAvoidCellReferencesIntoShippedBlocks.java From hbase with Apache License 2.0 | 6 votes |
/** * @throws java.lang.Exception */ @BeforeClass public static void setUpBeforeClass() throws Exception { ROWS[0] = ROW; ROWS[1] = ROW1; Configuration conf = TEST_UTIL.getConfiguration(); conf.setStrings(CoprocessorHost.REGION_COPROCESSOR_CONF_KEY, MultiRowMutationEndpoint.class.getName()); conf.setInt("hbase.regionserver.handler.count", 20); conf.setInt("hbase.bucketcache.size", 400); conf.setStrings(HConstants.BUCKET_CACHE_IOENGINE_KEY, "offheap"); conf.setInt("hbase.hstore.compactionThreshold", 7); conf.setFloat("hfile.block.cache.size", 0.2f); conf.setFloat("hbase.regionserver.global.memstore.size", 0.1f); conf.setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 0);// do not retry conf.setInt(HConstants.HBASE_CLIENT_SCANNER_TIMEOUT_PERIOD, 500000); FAMILIES_1[0] = FAMILY; TEST_UTIL.startMiniCluster(SLAVES); compactReadLatch = new CountDownLatch(1); }
Example #2
Source File: TestConnection.java From hbase with Apache License 2.0 | 6 votes |
@Test(expected = DoNotRetryIOException.class) public void testClosedConnection() throws ServiceException, Throwable { byte[] family = Bytes.toBytes("cf"); TableName tableName = TableName.valueOf(name.getMethodName()); TableDescriptorBuilder builder = TableDescriptorBuilder.newBuilder(tableName) .setCoprocessor(MultiRowMutationEndpoint.class.getName()) .setColumnFamily(ColumnFamilyDescriptorBuilder.of(family)); TEST_UTIL.getAdmin().createTable(builder.build()); Connection conn = ConnectionFactory.createConnection(TEST_UTIL.getConfiguration()); // cache the location try (Table table = conn.getTable(tableName)) { table.get(new Get(Bytes.toBytes(0))); } finally { conn.close(); } Batch.Call<MultiRowMutationService, MutateRowsResponse> callable = service -> { throw new RuntimeException("Should not arrive here"); }; conn.getTable(tableName).coprocessorService(MultiRowMutationService.class, HConstants.EMPTY_START_ROW, HConstants.EMPTY_END_ROW, callable); }
Example #3
Source File: TestBlockEvictionFromClient.java From hbase with Apache License 2.0 | 6 votes |
/** * @throws java.lang.Exception */ @BeforeClass public static void setUpBeforeClass() throws Exception { ROWS[0] = ROW; ROWS[1] = ROW1; Configuration conf = TEST_UTIL.getConfiguration(); conf.setStrings(CoprocessorHost.REGION_COPROCESSOR_CONF_KEY, MultiRowMutationEndpoint.class.getName()); conf.setInt("hbase.regionserver.handler.count", 20); conf.setInt("hbase.bucketcache.size", 400); conf.setStrings(HConstants.BUCKET_CACHE_IOENGINE_KEY, "offheap"); conf.setFloat("hfile.block.cache.size", 0.2f); conf.setFloat("hbase.regionserver.global.memstore.size", 0.1f); conf.setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 0);// do not retry conf.setInt(HConstants.HBASE_CLIENT_SCANNER_TIMEOUT_PERIOD, 5000); FAMILIES_1[0] = FAMILY; TEST_UTIL.startMiniCluster(SLAVES); }
Example #4
Source File: TestFromClientSideScanExcpetionWithCoprocessor.java From hbase with Apache License 2.0 | 5 votes |
@BeforeClass public static void setUpBeforeClass() throws Exception { Configuration conf = TEST_UTIL.getConfiguration(); conf.setStrings(CoprocessorHost.REGION_COPROCESSOR_CONF_KEY, MultiRowMutationEndpoint.class.getName(), NoOpScanPolicyObserver.class.getName()); TestFromClientSideScanExcpetion.setUpBeforeClass(); }
Example #5
Source File: TestIncrementsFromClientSide.java From hbase with Apache License 2.0 | 5 votes |
@BeforeClass public static void beforeClass() throws Exception { Configuration conf = TEST_UTIL.getConfiguration(); conf.setStrings(CoprocessorHost.REGION_COPROCESSOR_CONF_KEY, MultiRowMutationEndpoint.class.getName()); // We need more than one region server in this test TEST_UTIL.startMiniCluster(SLAVES); }
Example #6
Source File: ConnectionQueryServicesImpl.java From phoenix with Apache License 2.0 | 4 votes |
private void addCoprocessors(byte[] tableName, HTableDescriptor descriptor, PTableType tableType) throws SQLException { // The phoenix jar must be available on HBase classpath int priority = props.getInt(QueryServices.COPROCESSOR_PRIORITY_ATTRIB, QueryServicesOptions.DEFAULT_COPROCESSOR_PRIORITY); try { if (!descriptor.hasCoprocessor(ScanRegionObserver.class.getName())) { descriptor.addCoprocessor(ScanRegionObserver.class.getName(), null, priority, null); } if (!descriptor.hasCoprocessor(UngroupedAggregateRegionObserver.class.getName())) { descriptor.addCoprocessor(UngroupedAggregateRegionObserver.class.getName(), null, priority, null); } if (!descriptor.hasCoprocessor(GroupedAggregateRegionObserver.class.getName())) { descriptor.addCoprocessor(GroupedAggregateRegionObserver.class.getName(), null, priority, null); } if (!descriptor.hasCoprocessor(ServerCachingEndpointImpl.class.getName())) { descriptor.addCoprocessor(ServerCachingEndpointImpl.class.getName(), null, priority, null); } // TODO: better encapsulation for this // Since indexes can't have indexes, don't install our indexing coprocessor for indexes. // Also don't install on the SYSTEM.CATALOG and SYSTEM.STATS table because we use // all-or-none mutate class which break when this coprocessor is installed (PHOENIX-1318). if ((tableType != PTableType.INDEX && tableType != PTableType.VIEW) && !SchemaUtil.isMetaTable(tableName) && !SchemaUtil.isStatsTable(tableName) && !descriptor.hasCoprocessor(Indexer.class.getName())) { Map<String, String> opts = Maps.newHashMapWithExpectedSize(1); opts.put(CoveredColumnsIndexBuilder.CODEC_CLASS_NAME_KEY, PhoenixIndexCodec.class.getName()); Indexer.enableIndexing(descriptor, PhoenixIndexBuilder.class, opts, priority); } if (SchemaUtil.isStatsTable(tableName) && !descriptor.hasCoprocessor(MultiRowMutationEndpoint.class.getName())) { descriptor.addCoprocessor(MultiRowMutationEndpoint.class.getName(), null, priority, null); } if (descriptor.getValue(MetaDataUtil.IS_LOCAL_INDEX_TABLE_PROP_BYTES) != null && Boolean.TRUE.equals(PBoolean.INSTANCE.toObject(descriptor .getValue(MetaDataUtil.IS_LOCAL_INDEX_TABLE_PROP_BYTES)))) { if (!descriptor.hasCoprocessor(IndexHalfStoreFileReaderGenerator.class.getName())) { descriptor.addCoprocessor(IndexHalfStoreFileReaderGenerator.class.getName(), null, priority, null); } } else { if (!descriptor.hasCoprocessor(LocalIndexSplitter.class.getName()) && !SchemaUtil.isMetaTable(tableName) && !SchemaUtil.isSequenceTable(tableName)) { descriptor.addCoprocessor(LocalIndexSplitter.class.getName(), null, priority, null); } } // Setup split policy on Phoenix metadata table to ensure that the key values of a Phoenix table // stay on the same region. if (SchemaUtil.isMetaTable(tableName)) { if (!descriptor.hasCoprocessor(MetaDataEndpointImpl.class.getName())) { descriptor.addCoprocessor(MetaDataEndpointImpl.class.getName(), null, priority, null); } if (!descriptor.hasCoprocessor(MetaDataRegionObserver.class.getName())) { descriptor.addCoprocessor(MetaDataRegionObserver.class.getName(), null, priority + 1, null); } } else if (SchemaUtil.isSequenceTable(tableName)) { if (!descriptor.hasCoprocessor(SequenceRegionObserver.class.getName())) { descriptor.addCoprocessor(SequenceRegionObserver.class.getName(), null, priority, null); } } } catch (IOException e) { throw ServerUtil.parseServerException(e); } }
Example #7
Source File: FSTableDescriptors.java From hbase with Apache License 2.0 | 4 votes |
private static TableDescriptorBuilder createMetaTableDescriptorBuilder(final Configuration conf) throws IOException { // TODO We used to set CacheDataInL1 for META table. When we have BucketCache in file mode, now // the META table data goes to File mode BC only. Test how that affect the system. If too much, // we have to rethink about adding back the setCacheDataInL1 for META table CFs. return TableDescriptorBuilder.newBuilder(TableName.META_TABLE_NAME) .setColumnFamily(ColumnFamilyDescriptorBuilder.newBuilder(HConstants.CATALOG_FAMILY) .setMaxVersions(conf.getInt(HConstants.HBASE_META_VERSIONS, HConstants.DEFAULT_HBASE_META_VERSIONS)) .setInMemory(true) .setBlocksize(conf.getInt(HConstants.HBASE_META_BLOCK_SIZE, HConstants.DEFAULT_HBASE_META_BLOCK_SIZE)) .setScope(HConstants.REPLICATION_SCOPE_LOCAL) .setBloomFilterType(BloomType.ROWCOL) .setDataBlockEncoding(org.apache.hadoop.hbase.io.encoding.DataBlockEncoding.ROW_INDEX_V1) .build()) .setColumnFamily(ColumnFamilyDescriptorBuilder.newBuilder(HConstants.TABLE_FAMILY) .setMaxVersions(conf.getInt(HConstants.HBASE_META_VERSIONS, HConstants.DEFAULT_HBASE_META_VERSIONS)) .setInMemory(true) .setBlocksize(8 * 1024) .setScope(HConstants.REPLICATION_SCOPE_LOCAL) .setDataBlockEncoding(org.apache.hadoop.hbase.io.encoding.DataBlockEncoding.ROW_INDEX_V1) .setBloomFilterType(BloomType.ROWCOL) .build()) .setColumnFamily(ColumnFamilyDescriptorBuilder .newBuilder(HConstants.REPLICATION_BARRIER_FAMILY) .setMaxVersions(HConstants.ALL_VERSIONS) .setInMemory(true) .setScope(HConstants.REPLICATION_SCOPE_LOCAL) .setDataBlockEncoding(org.apache.hadoop.hbase.io.encoding.DataBlockEncoding.ROW_INDEX_V1) .setBloomFilterType(BloomType.ROWCOL) .build()) .setColumnFamily(ColumnFamilyDescriptorBuilder .newBuilder(HConstants.NAMESPACE_FAMILY) .setMaxVersions(conf.getInt(HConstants.HBASE_META_VERSIONS, HConstants.DEFAULT_HBASE_META_VERSIONS)) .setInMemory(true) .setBlocksize(conf.getInt(HConstants.HBASE_META_BLOCK_SIZE, HConstants.DEFAULT_HBASE_META_BLOCK_SIZE)) .setScope(HConstants.REPLICATION_SCOPE_LOCAL) .setDataBlockEncoding(org.apache.hadoop.hbase.io.encoding.DataBlockEncoding.ROW_INDEX_V1) .setBloomFilterType(BloomType.ROWCOL) .build()) .setCoprocessor(CoprocessorDescriptorBuilder.newBuilder( MultiRowMutationEndpoint.class.getName()) .setPriority(Coprocessor.PRIORITY_SYSTEM).build()); }
Example #8
Source File: TestIncrementFromClientSideWithCoprocessor.java From hbase with Apache License 2.0 | 4 votes |
@Before public void before() throws Exception { Configuration conf = TEST_UTIL.getConfiguration(); conf.setStrings(CoprocessorHost.REGION_COPROCESSOR_CONF_KEY, MultiRowMutationEndpoint.class.getName(), NoOpScanPolicyObserver.class.getName()); }
Example #9
Source File: TestFromClientSide5.java From hbase with Apache License 2.0 | 4 votes |
public TestFromClientSide5(Class registry, int numHedgedReqs) throws Exception { initialize(registry, numHedgedReqs, MultiRowMutationEndpoint.class); }
Example #10
Source File: TestFromClientSideWithCoprocessor4.java From hbase with Apache License 2.0 | 4 votes |
public TestFromClientSideWithCoprocessor4(Class registry, int numHedgedReqs) throws Exception { initialize(registry, numHedgedReqs, NoOpScanPolicyObserver.class, MultiRowMutationEndpoint.class); }
Example #11
Source File: TestFromClientSide.java From hbase with Apache License 2.0 | 4 votes |
public TestFromClientSide(Class registry, int numHedgedReqs) throws Exception { initialize(registry, numHedgedReqs, MultiRowMutationEndpoint.class); }
Example #12
Source File: TestFromClientSideWithCoprocessor.java From hbase with Apache License 2.0 | 4 votes |
public TestFromClientSideWithCoprocessor(Class registry, int numHedgedReqs) throws Exception { initialize(registry, numHedgedReqs, NoOpScanPolicyObserver.class, MultiRowMutationEndpoint.class); }
Example #13
Source File: TestFromClientSideWithCoprocessor5.java From hbase with Apache License 2.0 | 4 votes |
public TestFromClientSideWithCoprocessor5(Class registry, int numHedgedReqs) throws Exception { initialize(registry, numHedgedReqs, NoOpScanPolicyObserver.class, MultiRowMutationEndpoint.class); }
Example #14
Source File: TestFromClientSide4.java From hbase with Apache License 2.0 | 4 votes |
public TestFromClientSide4(Class registry, int numHedgedReqs) throws Exception { initialize(registry, numHedgedReqs, MultiRowMutationEndpoint.class); }