001/**
002 * Licensed to the Apache Software Foundation (ASF) under one
003 * or more contributor license agreements.  See the NOTICE file
004 * distributed with this work for additional information
005 * regarding copyright ownership.  The ASF licenses this file
006 * to you under the Apache License, Version 2.0 (the
007 * "License"); you may not use this file except in compliance
008 * with the License.  You may obtain a copy of the License at
009 *
010 *     http://www.apache.org/licenses/LICENSE-2.0
011 *
012 * Unless required by applicable law or agreed to in writing, software
013 * distributed under the License is distributed on an "AS IS" BASIS,
014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
015 * See the License for the specific language governing permissions and
016 * limitations under the License.
017 */
018package org.apache.hadoop.hbase.regionserver.compactions;
019
020import java.io.IOException;
021import java.util.List;
022
023import org.apache.hadoop.conf.Configuration;
024import org.apache.hadoop.fs.Path;
025import org.apache.hadoop.hbase.regionserver.HStore;
026import org.apache.hadoop.hbase.regionserver.InternalScanner;
027import org.apache.hadoop.hbase.regionserver.ScanInfo;
028import org.apache.hadoop.hbase.regionserver.ScanType;
029import org.apache.hadoop.hbase.regionserver.StoreFileScanner;
030import org.apache.hadoop.hbase.regionserver.StripeMultiFileWriter;
031import org.apache.hadoop.hbase.regionserver.throttle.ThroughputController;
032import org.apache.hadoop.hbase.security.User;
033import org.apache.hadoop.hbase.util.Bytes;
034import org.apache.yetus.audience.InterfaceAudience;
035import org.slf4j.Logger;
036import org.slf4j.LoggerFactory;
037
038/**
039 * This is the placeholder for stripe compactor. The implementation, as well as the proper javadoc,
040 * will be added in HBASE-7967.
041 */
042@InterfaceAudience.Private
043public class StripeCompactor extends AbstractMultiOutputCompactor<StripeMultiFileWriter> {
044  private static final Logger LOG = LoggerFactory.getLogger(StripeCompactor.class);
045
046  public StripeCompactor(Configuration conf, HStore store) {
047    super(conf, store);
048  }
049
050  private final class StripeInternalScannerFactory implements InternalScannerFactory {
051
052    private final byte[] majorRangeFromRow;
053
054    private final byte[] majorRangeToRow;
055
056    public StripeInternalScannerFactory(byte[] majorRangeFromRow, byte[] majorRangeToRow) {
057      this.majorRangeFromRow = majorRangeFromRow;
058      this.majorRangeToRow = majorRangeToRow;
059    }
060
061    @Override
062    public ScanType getScanType(CompactionRequestImpl request) {
063      // If majorRangeFromRow and majorRangeToRow are not null, then we will not use the return
064      // value to create InternalScanner. See the createScanner method below. The return value is
065      // also used when calling coprocessor hooks.
066      return ScanType.COMPACT_RETAIN_DELETES;
067    }
068
069    @Override
070    public InternalScanner createScanner(ScanInfo scanInfo, List<StoreFileScanner> scanners,
071        ScanType scanType, FileDetails fd, long smallestReadPoint) throws IOException {
072      return (majorRangeFromRow == null)
073          ? StripeCompactor.this.createScanner(store, scanInfo, scanners, scanType,
074            smallestReadPoint, fd.earliestPutTs)
075          : StripeCompactor.this.createScanner(store, scanInfo, scanners, smallestReadPoint,
076            fd.earliestPutTs, majorRangeFromRow, majorRangeToRow);
077    }
078  }
079
080  public List<Path> compact(CompactionRequestImpl request, final List<byte[]> targetBoundaries,
081      final byte[] majorRangeFromRow, final byte[] majorRangeToRow,
082      ThroughputController throughputController, User user) throws IOException {
083    if (LOG.isDebugEnabled()) {
084      StringBuilder sb = new StringBuilder();
085      sb.append("Executing compaction with " + targetBoundaries.size() + " boundaries:");
086      for (byte[] tb : targetBoundaries) {
087        sb.append(" [").append(Bytes.toString(tb)).append("]");
088      }
089      LOG.debug(sb.toString());
090    }
091    return compact(request, new StripeInternalScannerFactory(majorRangeFromRow, majorRangeToRow),
092      new CellSinkFactory<StripeMultiFileWriter>() {
093
094        @Override
095        public StripeMultiFileWriter createWriter(InternalScanner scanner, FileDetails fd,
096            boolean shouldDropBehind) throws IOException {
097          StripeMultiFileWriter writer = new StripeMultiFileWriter.BoundaryMultiWriter(
098              store.getComparator(), targetBoundaries, majorRangeFromRow, majorRangeToRow);
099          initMultiWriter(writer, scanner, fd, shouldDropBehind);
100          return writer;
101        }
102      }, throughputController, user);
103  }
104
105  public List<Path> compact(CompactionRequestImpl request, final int targetCount, final long targetSize,
106      final byte[] left, final byte[] right, byte[] majorRangeFromRow, byte[] majorRangeToRow,
107      ThroughputController throughputController, User user) throws IOException {
108    if (LOG.isDebugEnabled()) {
109      LOG.debug(
110        "Executing compaction with " + targetSize + " target file size, no more than " + targetCount
111            + " files, in [" + Bytes.toString(left) + "] [" + Bytes.toString(right) + "] range");
112    }
113    return compact(request, new StripeInternalScannerFactory(majorRangeFromRow, majorRangeToRow),
114      new CellSinkFactory<StripeMultiFileWriter>() {
115
116        @Override
117        public StripeMultiFileWriter createWriter(InternalScanner scanner, FileDetails fd,
118            boolean shouldDropBehind) throws IOException {
119          StripeMultiFileWriter writer = new StripeMultiFileWriter.SizeMultiWriter(
120              store.getComparator(), targetCount, targetSize, left, right);
121          initMultiWriter(writer, scanner, fd, shouldDropBehind);
122          return writer;
123        }
124      }, throughputController, user);
125  }
126
127  @Override
128  protected List<Path> commitWriter(StripeMultiFileWriter writer, FileDetails fd,
129      CompactionRequestImpl request) throws IOException {
130    List<Path> newFiles = writer.commitWriters(fd.maxSeqId, request.isMajor(), request.getFiles());
131    assert !newFiles.isEmpty() : "Should have produced an empty file to preserve metadata.";
132    return newFiles;
133  }
134}