001/* 002 * Licensed to the Apache Software Foundation (ASF) under one 003 * or more contributor license agreements. See the NOTICE file 004 * distributed with this work for additional information 005 * regarding copyright ownership. The ASF licenses this file 006 * to you under the Apache License, Version 2.0 (the 007 * "License"); you may not use this file except in compliance 008 * with the License. You may obtain a copy of the License at 009 * 010 * http://www.apache.org/licenses/LICENSE-2.0 011 * 012 * Unless required by applicable law or agreed to in writing, software 013 * distributed under the License is distributed on an "AS IS" BASIS, 014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ 018package org.apache.hadoop.hbase.regionserver; 019 020import static org.junit.Assert.assertEquals; 021 022import java.util.ArrayList; 023import java.util.List; 024import org.apache.hadoop.hbase.HBaseClassTestRule; 025import org.apache.hadoop.hbase.HBaseTestingUtility; 026import org.apache.hadoop.hbase.HConstants; 027import org.apache.hadoop.hbase.TableName; 028import org.apache.hadoop.hbase.client.Admin; 029import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder; 030import org.apache.hadoop.hbase.client.Put; 031import org.apache.hadoop.hbase.client.Scan; 032import org.apache.hadoop.hbase.client.Table; 033import org.apache.hadoop.hbase.client.TableDescriptorBuilder; 034import org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner; 035import org.apache.hadoop.hbase.regionserver.compactions.CompactionConfiguration; 036import org.apache.hadoop.hbase.testclassification.LargeTests; 037import org.apache.hadoop.hbase.testclassification.RegionServerTests; 038import org.apache.hadoop.hbase.util.Bytes; 039import org.apache.hadoop.hbase.util.JVMClusterUtil; 040import org.junit.After; 041import org.junit.AfterClass; 042import org.junit.Before; 043import org.junit.BeforeClass; 044import org.junit.ClassRule; 045import org.junit.Test; 046import org.junit.experimental.categories.Category; 047import org.slf4j.Logger; 048import org.slf4j.LoggerFactory; 049 050@Category({ LargeTests.class, RegionServerTests.class }) 051public class TestNotCleanupCompactedFileWhenRegionWarmup { 052 private static final Logger LOG = 053 LoggerFactory.getLogger(TestNotCleanupCompactedFileWhenRegionWarmup.class); 054 055 @ClassRule 056 public static final HBaseClassTestRule CLASS_RULE = 057 HBaseClassTestRule.forClass(TestNotCleanupCompactedFileWhenRegionWarmup.class); 058 059 private static HBaseTestingUtility TEST_UTIL; 060 private static Admin admin; 061 private static Table table; 062 063 private static TableName TABLE_NAME = TableName.valueOf("TestCleanupCompactedFileAfterFailover"); 064 private static byte[] ROW = Bytes.toBytes("row"); 065 private static byte[] FAMILY = Bytes.toBytes("cf"); 066 private static byte[] QUALIFIER = Bytes.toBytes("cq"); 067 private static byte[] VALUE = Bytes.toBytes("value"); 068 069 @BeforeClass 070 public static void beforeClass() throws Exception { 071 TEST_UTIL = new HBaseTestingUtility(); 072 // Set the scanner lease to 20min, so the scanner can't be closed by RegionServer 073 TEST_UTIL.getConfiguration().setInt(HConstants.HBASE_CLIENT_SCANNER_TIMEOUT_PERIOD, 1200000); 074 TEST_UTIL.getConfiguration().setInt(CompactionConfiguration.HBASE_HSTORE_COMPACTION_MIN_KEY, 075 100); 076 TEST_UTIL.getConfiguration().set("dfs.blocksize", "64000"); 077 TEST_UTIL.getConfiguration().set("dfs.namenode.fs-limits.min-block-size", "1024"); 078 TEST_UTIL.getConfiguration().set(TimeToLiveHFileCleaner.TTL_CONF_KEY, "0"); 079 TEST_UTIL.startMiniCluster(1); 080 admin = TEST_UTIL.getAdmin(); 081 } 082 083 @AfterClass 084 public static void afterClass() throws Exception { 085 TEST_UTIL.shutdownMiniCluster(); 086 } 087 088 @Before 089 public void before() throws Exception { 090 TableDescriptorBuilder builder = TableDescriptorBuilder.newBuilder(TABLE_NAME); 091 builder.setColumnFamily(ColumnFamilyDescriptorBuilder.of(FAMILY)); 092 admin.createTable(builder.build()); 093 TEST_UTIL.waitTableAvailable(TABLE_NAME); 094 table = TEST_UTIL.getConnection().getTable(TABLE_NAME); 095 } 096 097 @After 098 public void after() throws Exception { 099 admin.disableTable(TABLE_NAME); 100 admin.deleteTable(TABLE_NAME); 101 } 102 103 @Test 104 public void testRegionWarmup() throws Exception { 105 List<HRegion> regions = new ArrayList<>(); 106 for (JVMClusterUtil.RegionServerThread rsThread : TEST_UTIL.getHBaseCluster() 107 .getLiveRegionServerThreads()) { 108 HRegionServer rs = rsThread.getRegionServer(); 109 if (rs.getOnlineTables().contains(TABLE_NAME)) { 110 regions.addAll(rs.getRegions(TABLE_NAME)); 111 } 112 } 113 assertEquals("Table should only have one region", 1, regions.size()); 114 HRegion region = regions.get(0); 115 HStore store = region.getStore(FAMILY); 116 117 writeDataAndFlush(3, region); 118 assertEquals(3, store.getStorefilesCount()); 119 120 // Open a scanner and not close, then the storefile will be referenced 121 store.getScanner(new Scan(), null, 0); 122 region.compact(true); 123 assertEquals(1, store.getStorefilesCount()); 124 // The compacted file should not be archived as there are references by user scanner 125 assertEquals(3, store.getStoreEngine().getStoreFileManager().getCompactedfiles().size()); 126 127 HStore newStore = region.instantiateHStore(ColumnFamilyDescriptorBuilder.of(FAMILY), true); 128 // Should not archive the compacted storefiles when region warmup 129 assertEquals(4, newStore.getStorefilesCount()); 130 131 newStore = region.instantiateHStore(ColumnFamilyDescriptorBuilder.of(FAMILY), false); 132 // Archived the compacted storefiles when region real open 133 assertEquals(1, newStore.getStorefilesCount()); 134 } 135 136 private void writeDataAndFlush(int fileNum, HRegion region) throws Exception { 137 for (int i = 0; i < fileNum; i++) { 138 for (int j = 0; j < 100; j++) { 139 table.put(new Put(concat(ROW, j)).addColumn(FAMILY, QUALIFIER, concat(VALUE, j))); 140 } 141 region.flush(true); 142 } 143 } 144 145 private byte[] concat(byte[] base, int index) { 146 return Bytes.toBytes(Bytes.toString(base) + "-" + index); 147 } 148}