001/*
002 * Licensed to the Apache Software Foundation (ASF) under one
003 * or more contributor license agreements.  See the NOTICE file
004 * distributed with this work for additional information
005 * regarding copyright ownership.  The ASF licenses this file
006 * to you under the Apache License, Version 2.0 (the
007 * "License"); you may not use this file except in compliance
008 * with the License.  You may obtain a copy of the License at
009 *
010 *     http://www.apache.org/licenses/LICENSE-2.0
011 *
012 * Unless required by applicable law or agreed to in writing, software
013 * distributed under the License is distributed on an "AS IS" BASIS,
014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
015 * See the License for the specific language governing permissions and
016 * limitations under the License.
017 */
018package org.apache.hadoop.hbase.regionserver;
019
020import static org.junit.Assert.assertEquals;
021import static org.junit.Assert.assertTrue;
022import static org.junit.Assert.fail;
023import static org.mockito.Mockito.mock;
024import static org.mockito.Mockito.when;
025
026import java.io.IOException;
027import java.util.ArrayList;
028import java.util.Arrays;
029import java.util.List;
030import org.apache.hadoop.conf.Configuration;
031import org.apache.hadoop.fs.FileSystem;
032import org.apache.hadoop.fs.Path;
033import org.apache.hadoop.hbase.Cell;
034import org.apache.hadoop.hbase.CellUtil;
035import org.apache.hadoop.hbase.HBaseClassTestRule;
036import org.apache.hadoop.hbase.HBaseTestingUtil;
037import org.apache.hadoop.hbase.HConstants;
038import org.apache.hadoop.hbase.Stoppable;
039import org.apache.hadoop.hbase.TableName;
040import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
041import org.apache.hadoop.hbase.client.Durability;
042import org.apache.hadoop.hbase.client.Get;
043import org.apache.hadoop.hbase.client.Put;
044import org.apache.hadoop.hbase.client.RegionInfo;
045import org.apache.hadoop.hbase.client.RegionInfoBuilder;
046import org.apache.hadoop.hbase.client.Result;
047import org.apache.hadoop.hbase.client.TableDescriptor;
048import org.apache.hadoop.hbase.client.TableDescriptorBuilder;
049import org.apache.hadoop.hbase.testclassification.MediumTests;
050import org.apache.hadoop.hbase.testclassification.RegionServerTests;
051import org.apache.hadoop.hbase.util.Bytes;
052import org.apache.hadoop.hbase.util.CommonFSUtils;
053import org.apache.hadoop.hbase.util.StoppableImplementation;
054import org.apache.hadoop.hbase.wal.WALFactory;
055import org.junit.Before;
056import org.junit.ClassRule;
057import org.junit.Rule;
058import org.junit.Test;
059import org.junit.experimental.categories.Category;
060import org.junit.rules.TestName;
061
062@Category({ RegionServerTests.class, MediumTests.class })
063public class TestStoreFileRefresherChore {
064
065  @ClassRule
066  public static final HBaseClassTestRule CLASS_RULE =
067    HBaseClassTestRule.forClass(TestStoreFileRefresherChore.class);
068
069  private HBaseTestingUtil TEST_UTIL;
070  private Path testDir;
071
072  @Rule
073  public TestName name = new TestName();
074
075  @Before
076  public void setUp() throws IOException {
077    TEST_UTIL = new HBaseTestingUtil();
078    testDir = TEST_UTIL.getDataTestDir("TestStoreFileRefresherChore");
079    CommonFSUtils.setRootDir(TEST_UTIL.getConfiguration(), testDir);
080  }
081
082  private TableDescriptor getTableDesc(TableName tableName, int regionReplication,
083    byte[]... families) {
084    return getTableDesc(tableName, regionReplication, false, families);
085  }
086
087  private TableDescriptor getTableDesc(TableName tableName, int regionReplication, boolean readOnly,
088    byte[]... families) {
089    TableDescriptorBuilder builder = TableDescriptorBuilder.newBuilder(tableName)
090      .setRegionReplication(regionReplication).setReadOnly(readOnly);
091    Arrays.stream(families).map(family -> ColumnFamilyDescriptorBuilder.newBuilder(family)
092      .setMaxVersions(Integer.MAX_VALUE).build()).forEachOrdered(builder::setColumnFamily);
093    return builder.build();
094  }
095
096  static class FailingHRegionFileSystem extends HRegionFileSystem {
097    boolean fail = false;
098
099    FailingHRegionFileSystem(Configuration conf, FileSystem fs, Path tableDir,
100      RegionInfo regionInfo) {
101      super(conf, fs, tableDir, regionInfo);
102    }
103
104    @Override
105    public List<StoreFileInfo> getStoreFiles(String familyName) throws IOException {
106      if (fail) {
107        throw new IOException("simulating FS failure");
108      }
109      return super.getStoreFiles(familyName);
110    }
111  }
112
113  private HRegion initHRegion(TableDescriptor htd, byte[] startKey, byte[] stopKey, int replicaId)
114    throws IOException {
115    Configuration conf = TEST_UTIL.getConfiguration();
116    Path tableDir = CommonFSUtils.getTableDir(testDir, htd.getTableName());
117
118    RegionInfo info = RegionInfoBuilder.newBuilder(htd.getTableName()).setStartKey(startKey)
119      .setEndKey(stopKey).setRegionId(0L).setReplicaId(replicaId).build();
120    HRegionFileSystem fs =
121      new FailingHRegionFileSystem(conf, tableDir.getFileSystem(conf), tableDir, info);
122    final Configuration walConf = new Configuration(conf);
123    CommonFSUtils.setRootDir(walConf, tableDir);
124    final WALFactory wals = new WALFactory(walConf, "log_" + replicaId);
125    ChunkCreator.initialize(MemStoreLAB.CHUNK_SIZE_DEFAULT, false, 0, 0, 0, null,
126      MemStoreLAB.INDEX_CHUNK_SIZE_PERCENTAGE_DEFAULT);
127    HRegion region = new HRegion(fs, wals.getWAL(info), conf, htd, null);
128
129    region.initialize();
130
131    return region;
132  }
133
134  private void putData(Region region, int startRow, int numRows, byte[] qf, byte[]... families)
135    throws IOException {
136    for (int i = startRow; i < startRow + numRows; i++) {
137      Put put = new Put(Bytes.toBytes("" + i));
138      put.setDurability(Durability.SKIP_WAL);
139      for (byte[] family : families) {
140        put.addColumn(family, qf, null);
141      }
142      region.put(put);
143    }
144  }
145
146  private void verifyDataExpectFail(Region newReg, int startRow, int numRows, byte[] qf,
147    byte[]... families) throws IOException {
148    boolean threw = false;
149    try {
150      verifyData(newReg, startRow, numRows, qf, families);
151    } catch (AssertionError e) {
152      threw = true;
153    }
154    if (!threw) {
155      fail("Expected data verification to fail");
156    }
157  }
158
159  private void verifyData(Region newReg, int startRow, int numRows, byte[] qf, byte[]... families)
160    throws IOException {
161    for (int i = startRow; i < startRow + numRows; i++) {
162      byte[] row = Bytes.toBytes("" + i);
163      Get get = new Get(row);
164      for (byte[] family : families) {
165        get.addColumn(family, qf);
166      }
167      Result result = newReg.get(get);
168      Cell[] raw = result.rawCells();
169      assertEquals(families.length, result.size());
170      for (int j = 0; j < families.length; j++) {
171        assertTrue(CellUtil.matchingRows(raw[j], row));
172        assertTrue(CellUtil.matchingFamily(raw[j], families[j]));
173        assertTrue(CellUtil.matchingQualifier(raw[j], qf));
174      }
175    }
176  }
177
178  static class StaleStorefileRefresherChore extends StorefileRefresherChore {
179    boolean isStale = false;
180
181    public StaleStorefileRefresherChore(int period, HRegionServer regionServer,
182      Stoppable stoppable) {
183      super(period, false, regionServer, stoppable);
184    }
185
186    @Override
187    protected boolean isRegionStale(String encodedName, long time) {
188      return isStale;
189    }
190  }
191
192  @Test
193  public void testIsStale() throws IOException {
194    int period = 0;
195    byte[][] families = new byte[][] { Bytes.toBytes("cf") };
196    byte[] qf = Bytes.toBytes("cq");
197
198    HRegionServer regionServer = mock(HRegionServer.class);
199    List<HRegion> regions = new ArrayList<>();
200    when(regionServer.getOnlineRegionsLocalContext()).thenReturn(regions);
201    when(regionServer.getConfiguration()).thenReturn(TEST_UTIL.getConfiguration());
202
203    TableDescriptor htd = getTableDesc(TableName.valueOf(name.getMethodName()), 2, families);
204    HRegion primary = initHRegion(htd, HConstants.EMPTY_START_ROW, HConstants.EMPTY_END_ROW, 0);
205    HRegion replica1 = initHRegion(htd, HConstants.EMPTY_START_ROW, HConstants.EMPTY_END_ROW, 1);
206    regions.add(primary);
207    regions.add(replica1);
208
209    StaleStorefileRefresherChore chore =
210      new StaleStorefileRefresherChore(period, regionServer, new StoppableImplementation());
211
212    // write some data to primary and flush
213    putData(primary, 0, 100, qf, families);
214    primary.flush(true);
215    verifyData(primary, 0, 100, qf, families);
216
217    verifyDataExpectFail(replica1, 0, 100, qf, families);
218    chore.chore();
219    verifyData(replica1, 0, 100, qf, families);
220
221    // simulate an fs failure where we cannot refresh the store files for the replica
222    ((FailingHRegionFileSystem) replica1.getRegionFileSystem()).fail = true;
223
224    // write some more data to primary and flush
225    putData(primary, 100, 100, qf, families);
226    primary.flush(true);
227    verifyData(primary, 0, 200, qf, families);
228
229    chore.chore(); // should not throw ex, but we cannot refresh the store files
230
231    verifyData(replica1, 0, 100, qf, families);
232    verifyDataExpectFail(replica1, 100, 100, qf, families);
233
234    chore.isStale = true;
235    chore.chore(); // now after this, we cannot read back any value
236    try {
237      verifyData(replica1, 0, 100, qf, families);
238      fail("should have failed with IOException");
239    } catch (IOException ex) {
240      // expected
241    }
242  }
243
244  @Test
245  public void testRefreshReadOnlyTable() throws IOException {
246    int period = 0;
247    byte[][] families = new byte[][] { Bytes.toBytes("cf") };
248    byte[] qf = Bytes.toBytes("cq");
249
250    HRegionServer regionServer = mock(HRegionServer.class);
251    List<HRegion> regions = new ArrayList<>();
252    when(regionServer.getOnlineRegionsLocalContext()).thenReturn(regions);
253    when(regionServer.getConfiguration()).thenReturn(TEST_UTIL.getConfiguration());
254
255    TableDescriptor htd = getTableDesc(TableName.valueOf(name.getMethodName()), 2, families);
256    HRegion primary = initHRegion(htd, HConstants.EMPTY_START_ROW, HConstants.EMPTY_END_ROW, 0);
257    HRegion replica1 = initHRegion(htd, HConstants.EMPTY_START_ROW, HConstants.EMPTY_END_ROW, 1);
258    regions.add(primary);
259    regions.add(replica1);
260
261    StorefileRefresherChore chore =
262      new StorefileRefresherChore(period, false, regionServer, new StoppableImplementation());
263
264    // write some data to primary and flush
265    putData(primary, 0, 100, qf, families);
266    primary.flush(true);
267    verifyData(primary, 0, 100, qf, families);
268
269    verifyDataExpectFail(replica1, 0, 100, qf, families);
270    chore.chore();
271    verifyData(replica1, 0, 100, qf, families);
272
273    // write some data to primary and flush before refresh the store files for the replica
274    putData(primary, 100, 100, qf, families);
275    primary.flush(true);
276    verifyData(primary, 0, 200, qf, families);
277
278    // then the table is set to readonly
279    htd = getTableDesc(TableName.valueOf(name.getMethodName()), 2, true, families);
280    primary.setTableDescriptor(htd);
281    replica1.setTableDescriptor(htd);
282
283    chore.chore(); // we cannot refresh the store files
284    verifyDataExpectFail(replica1, 100, 100, qf, families);
285  }
286}