001/*
002 * Licensed to the Apache Software Foundation (ASF) under one
003 * or more contributor license agreements.  See the NOTICE file
004 * distributed with this work for additional information
005 * regarding copyright ownership.  The ASF licenses this file
006 * to you under the Apache License, Version 2.0 (the
007 * "License"); you may not use this file except in compliance
008 * with the License.  You may obtain a copy of the License at
009 *
010 *     http://www.apache.org/licenses/LICENSE-2.0
011 *
012 * Unless required by applicable law or agreed to in writing, software
013 * distributed under the License is distributed on an "AS IS" BASIS,
014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
015 * See the License for the specific language governing permissions and
016 * limitations under the License.
017 */
018package org.apache.hadoop.hbase.regionserver;
019
020import static org.apache.hadoop.hbase.regionserver.HRegion.SPLIT_IGNORE_BLOCKING_ENABLED_KEY;
021import static org.apache.hadoop.hbase.regionserver.Store.PRIORITY_USER;
022import static org.junit.Assert.assertEquals;
023import static org.junit.Assert.assertFalse;
024import static org.junit.Assert.assertNotNull;
025import static org.junit.Assert.assertTrue;
026
027import java.util.List;
028import org.apache.hadoop.hbase.HBaseClassTestRule;
029import org.apache.hadoop.hbase.HBaseTestingUtil;
030import org.apache.hadoop.hbase.HConstants;
031import org.apache.hadoop.hbase.TableName;
032import org.apache.hadoop.hbase.client.Admin;
033import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
034import org.apache.hadoop.hbase.client.Put;
035import org.apache.hadoop.hbase.client.ResultScanner;
036import org.apache.hadoop.hbase.client.Scan;
037import org.apache.hadoop.hbase.client.Table;
038import org.apache.hadoop.hbase.client.TableDescriptor;
039import org.apache.hadoop.hbase.client.TableDescriptorBuilder;
040import org.apache.hadoop.hbase.master.assignment.SplitTableRegionProcedure;
041import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
042import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
043import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility;
044import org.apache.hadoop.hbase.testclassification.MediumTests;
045import org.apache.hadoop.hbase.util.Bytes;
046import org.junit.AfterClass;
047import org.junit.Assert;
048import org.junit.BeforeClass;
049import org.junit.ClassRule;
050import org.junit.Test;
051import org.junit.experimental.categories.Category;
052import org.slf4j.Logger;
053import org.slf4j.LoggerFactory;
054
055import org.apache.hbase.thirdparty.com.google.common.io.Closeables;
056
057@Category({ MediumTests.class })
058public class TestSplitWithBlockingFiles {
059
060  @ClassRule
061  public static final HBaseClassTestRule CLASS_RULE =
062    HBaseClassTestRule.forClass(TestSplitWithBlockingFiles.class);
063
064  private static final Logger LOG = LoggerFactory.getLogger(TestSplitWithBlockingFiles.class);
065
066  protected static final HBaseTestingUtil UTIL = new HBaseTestingUtil();
067  private static TableName TABLE_NAME = TableName.valueOf("test");
068  private static Admin ADMIN;
069  private static byte[] CF = Bytes.toBytes("cf");
070  private static Table TABLE;
071
072  @BeforeClass
073  public static void setupCluster() throws Exception {
074    UTIL.getConfiguration().setLong(HConstants.HREGION_MAX_FILESIZE, 8 * 2 * 10240L);
075    UTIL.getConfiguration().setInt(HStore.BLOCKING_STOREFILES_KEY, 1);
076    UTIL.getConfiguration().set(HConstants.HBASE_REGION_SPLIT_POLICY_KEY,
077      ConstantSizeRegionSplitPolicy.class.getName());
078    UTIL.getConfiguration().setBoolean(SPLIT_IGNORE_BLOCKING_ENABLED_KEY, true);
079    UTIL.startMiniCluster(1);
080    ADMIN = UTIL.getAdmin();
081    TableDescriptor td = TableDescriptorBuilder.newBuilder(TABLE_NAME)
082      .setColumnFamily(ColumnFamilyDescriptorBuilder.newBuilder(CF).setBlocksize(1000).build())
083      .build();
084    TABLE = UTIL.createTable(td, null);
085    UTIL.waitTableAvailable(TABLE_NAME);
086  }
087
088  @AfterClass
089  public static void cleanupTest() throws Exception {
090    Closeables.close(TABLE, true);
091    UTIL.shutdownMiniCluster();
092  }
093
094  @Test
095  public void testSplitIgnoreBlockingFiles() throws Exception {
096    ADMIN.splitSwitch(false, true);
097    byte[] value = new byte[1024];
098    for (int m = 0; m < 10; m++) {
099      String rowPrefix = "row" + m;
100      for (int i = 0; i < 10; i++) {
101        Put p = new Put(Bytes.toBytes(rowPrefix + i));
102        p.addColumn(CF, Bytes.toBytes("qualifier"), value);
103        p.addColumn(CF, Bytes.toBytes("qualifier2"), value);
104        TABLE.put(p);
105      }
106      ADMIN.flush(TABLE_NAME);
107    }
108    Scan scan = new Scan();
109    ResultScanner results = TABLE.getScanner(scan);
110    int count = 0;
111    while (results.next() != null) {
112      count++;
113    }
114    Assert.assertEquals("There should be 100 rows!", 100, count);
115    List<HRegion> regions = UTIL.getMiniHBaseCluster().getRegionServer(0).getRegions();
116    regions.removeIf(r -> !r.getRegionInfo().getTable().equals(TABLE_NAME));
117    assertEquals(1, regions.size());
118    assertNotNull(regions.get(0).getSplitPolicy().getSplitPoint());
119    assertTrue(regions.get(0).getCompactPriority() >= PRIORITY_USER);
120    assertTrue(UTIL.getMiniHBaseCluster().getRegionServer(0).getCompactSplitThread()
121      .requestSplit(regions.get(0)));
122
123    // split region
124    ADMIN.splitSwitch(true, true);
125    MasterProcedureEnv env =
126      UTIL.getMiniHBaseCluster().getMaster().getMasterProcedureExecutor().getEnvironment();
127    final ProcedureExecutor<MasterProcedureEnv> executor =
128      UTIL.getMiniHBaseCluster().getMaster().getMasterProcedureExecutor();
129    SplitTableRegionProcedure splitProcedure =
130      new SplitTableRegionProcedure(env, regions.get(0).getRegionInfo(), Bytes.toBytes("row5"));
131    executor.submitProcedure(splitProcedure);
132    ProcedureTestingUtility.waitProcedure(executor, splitProcedure.getProcId());
133
134    regions = UTIL.getMiniHBaseCluster().getRegionServer(0).getRegions();
135    regions.removeIf(r -> !r.getRegionInfo().getTable().equals(TABLE_NAME));
136    assertEquals(2, regions.size());
137    scan = new Scan();
138    results = TABLE.getScanner(scan);
139    count = 0;
140    while (results.next() != null) {
141      count++;
142    }
143    Assert.assertEquals("There should be 100 rows!", 100, count);
144    for (HRegion region : regions) {
145      assertTrue(region.getCompactPriority() < PRIORITY_USER);
146      assertFalse(
147        UTIL.getMiniHBaseCluster().getRegionServer(0).getCompactSplitThread().requestSplit(region));
148    }
149  }
150}