Uses of Class
org.apache.hadoop.hbase.util.Pair
Package
Description
Provides HBase Client
Restrict the domain of a data attribute, often times to fulfill business rules/requirements.
Table of Contents
Provides row-level filters applied to HRegion scan results during calls to
ResultScanner.next()
.Provides implementations of
HFile
and HFile
BlockCache
.Provides
BucketCache
, an implementation of
BlockCache
.Tools to help define network clients and servers.
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
Multi Cluster Replication
HBase REST
-
Uses of Pair in org.apache.hadoop.hbase
Modifier and TypeMethodDescriptionstatic Pair<RegionInfo,
ServerName> MetaTableAccessor.getRegion
(Connection connection, byte[] regionName) Deprecated.TagUtil.readVIntValuePart
(Tag tag, int offset) Reads an int value stored as a VInt at tag's given offset.Modifier and TypeMethodDescriptionMetaTableAccessor.getTableEncodedRegionNameAndLastBarrier
(Connection conn, TableName tableName) private static CompletableFuture<List<Pair<RegionInfo,
ServerName>>> AsyncMetaTableAccessor.getTableRegionsAndLocations
(AsyncTable<AdvancedScanResultConsumer> metaTable, TableName tableName, boolean excludeOfflinedSplitParents) Used to get table regions' info and server.static List<Pair<RegionInfo,
ServerName>> MetaTableAccessor.getTableRegionsAndLocations
(Connection connection, TableName tableName) Do not use this method to get meta table regions, use methods in MetaTableLocator instead.static List<Pair<RegionInfo,
ServerName>> MetaTableAccessor.getTableRegionsAndLocations
(Connection connection, TableName tableName, boolean excludeOfflinedSplitParents) Do not use this method to get meta table regions, use methods in MetaTableLocator instead.Modifier and TypeMethodDescriptionprivate static List<RegionInfo>
MetaTableAccessor.getListOfRegionInfos
(List<Pair<RegionInfo, ServerName>> pairs) -
Uses of Pair in org.apache.hadoop.hbase.backup
Modifier and TypeMethodDescriptionvoid
BackupObserver.postBulkLoadHFile
(ObserverContext<RegionCoprocessorEnvironment> ctx, List<Pair<byte[], String>> stagingFamilyPaths, Map<byte[], List<org.apache.hadoop.fs.Path>> finalPaths) void
BackupObserver.preCommitStoreFile
(ObserverContext<RegionCoprocessorEnvironment> ctx, byte[] family, List<Pair<org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path>> pairs) -
Uses of Pair in org.apache.hadoop.hbase.backup.impl
Modifier and TypeMethodDescriptionBackupManager.readBulkloadRows
(List<TableName> tableList) BackupSystemTable.readBulkloadRows
(List<TableName> tableList) Modifier and TypeMethodDescriptionBackupManager.readBulkloadRows
(List<TableName> tableList) BackupSystemTable.readBulkloadRows
(List<TableName> tableList) Modifier and TypeMethodDescriptionBackupSystemTable.createPutForPreparedBulkload
(TableName table, byte[] region, byte[] family, List<Pair<org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path>> pairs) void
BackupSystemTable.writeFilesForBulkLoadPreCommit
(TableName tabName, byte[] region, byte[] family, List<Pair<org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path>> pairs) -
Uses of Pair in org.apache.hadoop.hbase.backup.mapreduce
Modifier and TypeMethodDescriptionprotected List<org.apache.hadoop.fs.Path>
MapReduceBackupMergeJob.toPathList
(List<Pair<TableName, org.apache.hadoop.fs.Path>> processedTableList) MapReduceBackupMergeJob.toTableNameList
(List<Pair<TableName, org.apache.hadoop.fs.Path>> processedTableList) -
Uses of Pair in org.apache.hadoop.hbase.chaos
-
Uses of Pair in org.apache.hadoop.hbase.client
Modifier and TypeMethodDescriptionScannerCallableWithReplicas.RetryingRPC.call
(int callTimeout) Admin.getAlterStatus
(byte[] tableName) Deprecated.Since 2.0.0.Admin.getAlterStatus
(TableName tableName) Deprecated.Since 2.0.0.HBaseAdmin.getAlterStatus
(byte[] tableName) HBaseAdmin.getAlterStatus
(TableName tableName) private Pair<List<byte[]>,
List<HRegionLocation>> HTable.getKeysAndRegionsInRange
(byte[] startKey, byte[] endKey, boolean includeEndKey) Get the corresponding start keys and regions for an arbitrary range of keys.private Pair<List<byte[]>,
List<HRegionLocation>> HTable.getKeysAndRegionsInRange
(byte[] startKey, byte[] endKey, boolean includeEndKey, boolean reload) Get the corresponding start keys and regions for an arbitrary range of keys.(package private) Pair<RegionInfo,
ServerName> HBaseAdmin.getRegion
(byte[] regionName) default Pair<byte[][],
byte[][]> RegionLocator.getStartEndKeys()
Gets the starting and ending row keys for every region in the currently open table.private Pair<RegionState.State,
ServerName> ZKConnectionRegistry.getStateAndServerName
(org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer proto) Deprecated.private Pair<HRegionLocation,
byte[]> ReversedScannerCallable.locateLastRegionInRange
(byte[] startKey, byte[] endKey) Get the last region before the endkey, which will be used to execute the reverse scanModifier and TypeMethodDescriptiondefault CompletableFuture<List<Pair<byte[],
byte[]>>> AsyncTableRegionLocator.getStartEndKeys()
Gets the starting and ending row keys for every region in the currently open table.Modifier and TypeMethodDescriptionprivate void
ScannerCallableWithReplicas.addCallsForCurrentReplica
(ResultBoundedCompletionService<Pair<Result[], ScannerCallable>> cs, int rpcTimeout) private void
ScannerCallableWithReplicas.addCallsForOtherReplicas
(ResultBoundedCompletionService<Pair<Result[], ScannerCallable>> cs, int min, int max, int rpcTimeout) boolean
SecureBulkLoadClient.secureBulkLoadHFiles
(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService.BlockingInterface client, List<Pair<byte[], String>> familyPaths, byte[] regionName, boolean assignSeqNum, org.apache.hadoop.security.token.Token<?> userToken, String bulkToken) Securely bulk load a list of HFiles using client protocol.boolean
SecureBulkLoadClient.secureBulkLoadHFiles
(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService.BlockingInterface client, List<Pair<byte[], String>> familyPaths, byte[] regionName, boolean assignSeqNum, org.apache.hadoop.security.token.Token<?> userToken, String bulkToken, boolean copyFiles) Securely bulk load a list of HFiles using client protocol.boolean
SecureBulkLoadClient.secureBulkLoadHFiles
(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService.BlockingInterface client, List<Pair<byte[], String>> familyPaths, byte[] regionName, boolean assignSeqNum, org.apache.hadoop.security.token.Token<?> userToken, String bulkToken, boolean copyFiles, List<String> clusterIds, boolean replicate) -
Uses of Pair in org.apache.hadoop.hbase.constraint
Modifier and TypeMethodDescriptionConstraints.getKeyValueForClass
(HTableDescriptor desc, Class<? extends Constraint> clazz) Get the kvMap.Entry
in the descriptor for the specified classModifier and TypeMethodDescriptionstatic void
Constraints.add
(HTableDescriptor desc, Pair<Class<? extends Constraint>, org.apache.hadoop.conf.Configuration>... constraints) Add constraints and their associated configurations to the table. -
Uses of Pair in org.apache.hadoop.hbase.coprocessor
Modifier and TypeMethodDescriptionRegionObserver.postAppendBeforeWAL
(ObserverContext<RegionCoprocessorEnvironment> ctx, Mutation mutation, List<Pair<Cell, Cell>> cellPairs) Called after a list of new cells has been created during an append operation, but before they are committed to the WAL or memstore.RegionObserver.postIncrementBeforeWAL
(ObserverContext<RegionCoprocessorEnvironment> ctx, Mutation mutation, List<Pair<Cell, Cell>> cellPairs) Called after a list of new cells has been created during an increment operation, but before they are committed to the WAL or memstore.Modifier and TypeMethodDescriptionRegionObserver.postAppendBeforeWAL
(ObserverContext<RegionCoprocessorEnvironment> ctx, Mutation mutation, List<Pair<Cell, Cell>> cellPairs) Called after a list of new cells has been created during an append operation, but before they are committed to the WAL or memstore.default void
RegionObserver.postBulkLoadHFile
(ObserverContext<RegionCoprocessorEnvironment> ctx, List<Pair<byte[], String>> stagingFamilyPaths, Map<byte[], List<org.apache.hadoop.fs.Path>> finalPaths) Called after bulkLoadHFile.RegionObserver.postIncrementBeforeWAL
(ObserverContext<RegionCoprocessorEnvironment> ctx, Mutation mutation, List<Pair<Cell, Cell>> cellPairs) Called after a list of new cells has been created during an increment operation, but before they are committed to the WAL or memstore.default void
RegionObserver.preBulkLoadHFile
(ObserverContext<RegionCoprocessorEnvironment> ctx, List<Pair<byte[], String>> familyPaths) Called before bulkLoadHFile.default void
RegionObserver.preCommitStoreFile
(ObserverContext<RegionCoprocessorEnvironment> ctx, byte[] family, List<Pair<org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path>> pairs) Called before moving bulk loaded hfile to region directory. -
Uses of Pair in org.apache.hadoop.hbase.favored
Modifier and TypeMethodDescriptionprivate Pair<Map<ServerName,
List<RegionInfo>>, List<RegionInfo>> FavoredNodeLoadBalancer.segregateRegionsAndAssignRegionsWithFavoredNodes
(List<RegionInfo> regions, List<ServerName> availableServers) -
Uses of Pair in org.apache.hadoop.hbase.filter
Modifier and TypeFieldDescriptionFuzzyRowFilter.fuzzyKeysData
private final PriorityQueue<Pair<byte[],
Pair<byte[], byte[]>>> FuzzyRowFilter.RowTracker.nextRows
private final PriorityQueue<Pair<byte[],
Pair<byte[], byte[]>>> FuzzyRowFilter.RowTracker.nextRows
Modifier and TypeMethodDescriptionprivate void
FuzzyRowFilter.preprocessSearchKey
(Pair<byte[], byte[]> p) (package private) void
FuzzyRowFilter.RowTracker.updateWith
(Cell currentCell, Pair<byte[], byte[]> fuzzyData) ModifierConstructorDescriptionFuzzyRowFilter
(List<Pair<byte[], byte[]>> fuzzyKeysData) (package private)
FuzzyRowFilter
(List<Pair<byte[], byte[]>> fuzzyKeysData, byte processedWildcardMask) -
Uses of Pair in org.apache.hadoop.hbase.io
-
Uses of Pair in org.apache.hadoop.hbase.io.crypto
Modifier and TypeFieldDescription(package private) static final Map<Pair<String,
String>, KeyProvider> Encryption.keyProviderCache
-
Uses of Pair in org.apache.hadoop.hbase.io.hfile
Modifier and TypeMethodDescriptionBlockCache.getFullyCachedFiles()
Returns an Optional containing the map of files that have been fully cached (all its blocks are present in the cache.CombinedBlockCache.getFullyCachedFiles()
Returns the list of fully cached files -
Uses of Pair in org.apache.hadoop.hbase.io.hfile.bucket
Modifier and TypeFieldDescriptionBucketCache.fullyCachedFiles
Map of hFile -> Region -> File size.Modifier and TypeMethodDescription(package private) static Pair<ConcurrentHashMap<BlockCacheKey,
BucketEntry>, NavigableSet<BlockCacheKey>> BucketProtoUtils.fromPB
(Map<Integer, String> deserializers, org.apache.hadoop.hbase.shaded.protobuf.generated.BucketCacheProtos.BackingMap backingMap, Function<BucketEntry, ByteBuffAllocator.Recycler> createRecycler) Modifier and TypeMethodDescriptionBucketProtoUtils.fromPB
(Map<String, org.apache.hadoop.hbase.shaded.protobuf.generated.BucketCacheProtos.RegionFileSizeMap> prefetchHFileNames) BucketCache.getFullyCachedFiles()
-
Uses of Pair in org.apache.hadoop.hbase.io.util
Modifier and TypeMethodDescriptionstatic Pair<Long,
MemoryType> MemorySizeUtil.getGlobalMemStoreSize
(org.apache.hadoop.conf.Configuration conf) Returns Pair of global memstore size and memory type(ie.StreamUtils.readRawVarint32
(byte[] input, int offset) Reads a varInt value stored in an array.StreamUtils.readRawVarint32
(ByteBuffer input, int offset) -
Uses of Pair in org.apache.hadoop.hbase.ipc
Modifier and TypeFieldDescriptionprivate static Pair<org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroup,
Class<? extends org.apache.hbase.thirdparty.io.netty.channel.Channel>> NettyRpcClientConfigHelper.DEFAULT_EVENT_LOOP
Modifier and TypeFieldDescriptionprivate static final Map<String,
Pair<org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroup, Class<? extends org.apache.hbase.thirdparty.io.netty.channel.Channel>>> NettyRpcClientConfigHelper.EVENT_LOOP_CONFIG_MAP
Modifier and TypeMethodDescriptionPair<org.apache.hbase.thirdparty.com.google.protobuf.Message,
CellScanner> NettyRpcServer.call
(org.apache.hbase.thirdparty.com.google.protobuf.BlockingService service, org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor md, org.apache.hbase.thirdparty.com.google.protobuf.Message param, CellScanner cellScanner, long receiveTime, MonitoredRPCHandler status) Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,
CellScanner> NettyRpcServer.call
(org.apache.hbase.thirdparty.com.google.protobuf.BlockingService service, org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor md, org.apache.hbase.thirdparty.com.google.protobuf.Message param, CellScanner cellScanner, long receiveTime, MonitoredRPCHandler status, long startTime, int timeout) Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,
CellScanner> RpcServer.call
(RpcCall call, MonitoredRPCHandler status) This is a server side method, which is invoked over RPC.Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,
CellScanner> RpcServerInterface.call
(RpcCall call, MonitoredRPCHandler status) Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,
CellScanner> RpcServerInterface.call
(org.apache.hbase.thirdparty.com.google.protobuf.BlockingService service, org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor md, org.apache.hbase.thirdparty.com.google.protobuf.Message param, CellScanner cellScanner, long receiveTime, MonitoredRPCHandler status) Deprecated.As of release 1.3, this will be removed in HBase 3.0Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,
CellScanner> RpcServerInterface.call
(org.apache.hbase.thirdparty.com.google.protobuf.BlockingService service, org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor md, org.apache.hbase.thirdparty.com.google.protobuf.Message param, CellScanner cellScanner, long receiveTime, MonitoredRPCHandler status, long startTime, int timeout) Deprecated.As of release 2.0, this will be removed in HBase 3.0Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,
CellScanner> SimpleRpcServer.call
(org.apache.hbase.thirdparty.com.google.protobuf.BlockingService service, org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor md, org.apache.hbase.thirdparty.com.google.protobuf.Message param, CellScanner cellScanner, long receiveTime, MonitoredRPCHandler status) Deprecated.Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,
CellScanner> SimpleRpcServer.call
(org.apache.hbase.thirdparty.com.google.protobuf.BlockingService service, org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor md, org.apache.hbase.thirdparty.com.google.protobuf.Message param, CellScanner cellScanner, long receiveTime, MonitoredRPCHandler status, long startTime, int timeout) Deprecated.private static Pair<org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroup,
Class<? extends org.apache.hbase.thirdparty.io.netty.channel.Channel>> NettyRpcClientConfigHelper.getDefaultEventLoopConfig
(org.apache.hadoop.conf.Configuration conf) (package private) static Pair<org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroup,
Class<? extends org.apache.hbase.thirdparty.io.netty.channel.Channel>> NettyRpcClientConfigHelper.getEventLoopConfig
(org.apache.hadoop.conf.Configuration conf) MetricsHBaseServerWrapper.getTotalAndMaxNettyOutboundBytes()
These two metrics are calculated together, so we want to return them in one callMetricsHBaseServerWrapperImpl.getTotalAndMaxNettyOutboundBytes()
NettyRpcServer.getTotalAndMaxNettyOutboundBytes()
private Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeaderResponse,
CryptoAES> ServerRpcConnection.setupCryptoCipher()
Set up cipher for rpc encryption with Apache Commons Crypto. -
Uses of Pair in org.apache.hadoop.hbase.mapreduce
Modifier and TypeMethodDescriptionprotected Pair<byte[][],
byte[][]> TableInputFormat.getStartEndKeys()
protected Pair<byte[][],
byte[][]> TableInputFormatBase.getStartEndKeys()
ImportTsv.TsvParser.parseRowKey
(byte[] lineBytes, int length) Return starting position and length of row key from the specified line bytes.Modifier and TypeMethodDescription(package private) void
HashTable.TableHash.selectPartitions
(Pair<byte[][], byte[][]> regionStartEndKeys) Choose partitions between row ranges to hash to a single output file Selects region boundaries that fall within the scan range, and groups them into the desired number of partitions. -
Uses of Pair in org.apache.hadoop.hbase.mapreduce.replication
Modifier and TypeMethodDescriptionprivate static Pair<ReplicationPeerConfig,
org.apache.hadoop.conf.Configuration> VerifyReplication.getPeerQuorumConfig
(org.apache.hadoop.conf.Configuration conf, String peerId) -
Uses of Pair in org.apache.hadoop.hbase.master
Modifier and TypeMethodDescription(package private) List<Pair<ServerName,
Long>> DeadServer.copyDeadServersSince
(long ts) Extract all the servers dead since a given time, and sort them.protected List<Pair<ServerName,
Long>> ClusterStatusPublisher.getDeadServers
(long since) Get the servers which died since a given timestamp.HMaster.getReplicationLoad
(ServerName[] serverNames) -
Uses of Pair in org.apache.hadoop.hbase.master.assignment
Modifier and TypeMethodDescriptionPair<org.apache.hadoop.fs.Path,
org.apache.hadoop.fs.Path> SplitTableRegionProcedure.StoreFileSplitter.call()
AssignmentManager.getReopenStatus
(TableName tableName) Used by the client (via master) to identify if all regions have the schema updatesprivate Pair<org.apache.hadoop.fs.Path,
org.apache.hadoop.fs.Path> SplitTableRegionProcedure.splitStoreFile
(HRegionFileSystem regionFs, byte[] family, HStoreFile sf) SplitTableRegionProcedure.splitStoreFiles
(MasterProcedureEnv env, HRegionFileSystem regionFs) Create Split directoryModifier and TypeMethodDescriptionAssignmentManager.getTableRegionsAndLocations
(TableName tableName, boolean excludeOfflinedSplitParents) -
Uses of Pair in org.apache.hadoop.hbase.master.balancer
Modifier and TypeFieldDescription(package private) Map<String,
Pair<ServerName, Float>> BalancerClusterState.regionCacheRatioOnOldServerMap
(package private) Map<String,
Pair<ServerName, Float>> StochasticLoadBalancer.regionCacheRatioOnOldServerMap
BalancerClusterState.regionIndexServerIndexRegionCachedRatio
Modifier and TypeMethodDescriptionprivate Pair<Map<ServerName,
List<RegionInfo>>, List<RegionInfo>> FavoredStochasticBalancer.segregateRegionsAndAssignRegionsWithFavoredNodes
(Collection<RegionInfo> regions, List<ServerName> onlineServers) -
Uses of Pair in org.apache.hadoop.hbase.master.hbck
Modifier and TypeFieldDescriptionprivate final Map<String,
Pair<ServerName, List<ServerName>>> HbckReport.inconsistentRegions
-
Uses of Pair in org.apache.hadoop.hbase.master.janitor
Modifier and TypeFieldDescription(package private) final List<Pair<RegionInfo,
RegionInfo>> CatalogJanitorReport.holes
(package private) final List<Pair<RegionInfo,
RegionInfo>> CatalogJanitorReport.overlaps
(package private) final List<Pair<RegionInfo,
ServerName>> CatalogJanitorReport.unknownServers
TODO: If CatalogJanitor finds an 'Unknown Server', it should 'fix' it by queuing aHBCKServerCrashProcedure
for found server for it to clean up meta.Modifier and TypeMethodDescriptionCatalogJanitor.checkRegionReferences
(MasterServices services, TableName tableName, RegionInfo region) Checks if a region still holds references to parent.Modifier and TypeMethodDescriptionCatalogJanitorReport.getHoles()
CatalogJanitorReport.getOverlaps()
CatalogJanitorReport.getUnknownServers()
Modifier and TypeMethodDescriptionprivate static Optional<RegionInfo>
MetaFixer.getHoleCover
(Pair<RegionInfo, RegionInfo> hole) private static boolean
CatalogJanitor.hasNoReferences
(Pair<Boolean, Boolean> p) (package private) static boolean
MetaFixer.isOverlap
(RegionInfo ri, Pair<RegionInfo, RegionInfo> pair) Modifier and TypeMethodDescription(package private) static List<SortedSet<RegionInfo>>
MetaFixer.calculateMerges
(int maxMergeCount, List<Pair<RegionInfo, RegionInfo>> overlaps) Run throughoverlaps
and return a list of merges to run.private static void
MetaFixer.calculateTableMerges
(int maxMergeCount, List<SortedSet<RegionInfo>> merges, Collection<Pair<RegionInfo, RegionInfo>> overlaps) private static List<RegionInfo>
MetaFixer.createRegionInfosForHoles
(List<Pair<RegionInfo, RegionInfo>> holes) Create a newRegionInfo
corresponding to each provided "hole" pair. -
Uses of Pair in org.apache.hadoop.hbase.master.procedure
-
Uses of Pair in org.apache.hadoop.hbase.master.snapshot
Modifier and TypeMethodDescriptionvoid
DisabledTableSnapshotHandler.snapshotRegions
(List<Pair<RegionInfo, ServerName>> regionsAndLocations) protected void
EnabledTableSnapshotHandler.snapshotRegions
(List<Pair<RegionInfo, ServerName>> regions) This method kicks off a snapshot procedure.protected abstract void
TakeSnapshotHandler.snapshotRegions
(List<Pair<RegionInfo, ServerName>> regions) Snapshot the specified regions -
Uses of Pair in org.apache.hadoop.hbase.regionserver
Modifier and TypeFieldDescriptionprivate ConcurrentMap<String,
Pair<Long, Long>> RegionServerAccounting.retainedRegionRWRequestsCnt
private final org.apache.hbase.thirdparty.com.google.common.cache.LoadingCache<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription,
Pair<org.apache.hadoop.fs.FileSystem, Map<String, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest>>> RSSnapshotVerifier.SNAPSHOT_MANIFEST_CACHE
Modifier and TypeMethodDescription(package private) Pair<org.apache.hadoop.fs.Path,
org.apache.hadoop.fs.Path> HRegionFileSystem.bulkLoadStoreFile
(String familyName, org.apache.hadoop.fs.Path srcPath, long seqNum) Bulk load: Add a specified store file to the specified family.StoreFileInfo.getReferredToRegionAndFile
(String referenceFile) Pair<org.apache.hadoop.fs.FileSystem,
Map<String, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest>> RSSnapshotVerifier.SnapshotManifestCacheLoader.load
(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot) private Pair<String,
RSRpcServices.RegionScannerHolder> RSRpcServices.newRegionScanner
(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ScanRequest request, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ScanResponse.Builder builder) Pair<org.apache.hadoop.fs.Path,
org.apache.hadoop.fs.Path> HStore.preBulkLoadHFile
(String srcPathStr, long seqNum) This method should only be called from Region.Modifier and TypeMethodDescriptionHRegion.BatchOperation.buildWALEdits
(MiniBatchOperationInProgress<Mutation> miniBatchOp) Builds separate WALEdit per nonce by applying input mutations.HRegion.MutationBatchOperation.buildWALEdits
(MiniBatchOperationInProgress<Mutation> miniBatchOp) protected ConcurrentMap<String,
Pair<Long, Long>> RegionServerAccounting.getRetainedRegionRWRequestsCnt()
Returns the retained metrics of region's read and write requests countModifier and TypeMethodDescriptionHRegion.bulkLoadHFiles
(Collection<Pair<byte[], String>> familyPaths, boolean assignSeqId, HRegion.BulkLoadListener bulkLoadListener) Attempts to atomically load a group of hfiles.HRegion.bulkLoadHFiles
(Collection<Pair<byte[], String>> familyPaths, boolean assignSeqId, HRegion.BulkLoadListener bulkLoadListener, boolean copyFile, List<String> clusterIds, boolean replicate) Attempts to atomically load a group of hfiles.private static boolean
HRegion.hasMultipleColumnFamilies
(Collection<Pair<byte[], String>> familyPaths) Determines whether multiple column families are present Precondition: familyPaths is not nullvoid
RegionCoprocessorHost.postBulkLoadHFile
(List<Pair<byte[], String>> familyPaths, Map<byte[], List<org.apache.hadoop.fs.Path>> map) void
RegionCoprocessorHost.preBulkLoadHFile
(List<Pair<byte[], String>> familyPaths) boolean
RegionCoprocessorHost.preCommitStoreFile
(byte[] family, List<Pair<org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path>> pairs) -
Uses of Pair in org.apache.hadoop.hbase.regionserver.compactions
Modifier and TypeMethodDescriptionStripeCompactionPolicy.estimateTargetKvs
(Collection<HStoreFile> files, double splitCount) -
Uses of Pair in org.apache.hadoop.hbase.regionserver.querymatcher
Modifier and TypeMethodDescriptionprotected static Pair<DeleteTracker,
ColumnTracker> ScanQueryMatcher.getTrackers
(RegionCoprocessorHost host, NavigableSet<byte[]> columns, ScanInfo scanInfo, long oldestUnexpiredTS, Scan userScan) -
Uses of Pair in org.apache.hadoop.hbase.regionserver.wal
Modifier and TypeMethodDescriptionprotected final Pair<org.apache.hadoop.fs.FSDataInputStream,
org.apache.hadoop.fs.FileStatus> AbstractProtobufWALReader.open()
private Pair<org.apache.hadoop.fs.FSDataInputStream,
org.apache.hadoop.fs.FileStatus> AbstractProtobufWALReader.openArchivedWAL()
-
Uses of Pair in org.apache.hadoop.hbase.replication
Modifier and TypeMethodDescriptionReplicationQueueStorage.claimQueue
(ServerName sourceServerName, String queueId, ServerName destServerName) Change ownership for the queue identified by queueId and belongs to a dead region server.ZKReplicationQueueStorage.claimQueue
(ServerName sourceServerName, String queueId, ServerName destServerName) This implement must update the cversion of rootZKReplicationQueueStorage.queuesZNode
.ZKReplicationQueueStorage.getLastSequenceIdWithVersion
(String encodedRegionName, String peerId) Return the {lastPushedSequenceId, ZNodeDataVersion} pair.Modifier and TypeMethodDescriptionvoid
ReplicationQueueStorage.addHFileRefs
(String peerId, List<Pair<org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path>> pairs) Add new hfile references to the queue.void
ZKReplicationQueueStorage.addHFileRefs
(String peerId, List<Pair<org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path>> pairs) -
Uses of Pair in org.apache.hadoop.hbase.replication.regionserver
Modifier and TypeFieldDescriptionHFileReplicator.bulkLoadHFileMap
WALEntryBatch.walEntriesWithSize
Modifier and TypeMethodDescriptionNoopReplicationQueueStorage.claimQueue
(ServerName sourceServerName, String queueId, ServerName destServerName) ReplicationSourceWALReader.countDistinctRowKeysAndHFiles
(WALEdit edit) Count the number of different row keys in the given edit because of mini-batching.private Pair<WALTailingReader.State,
Boolean> WALEntryStream.readNextEntryAndRecordReaderPosition()
Returns whether the file is opened for writing.Modifier and TypeMethodDescriptionprivate void
ReplicationSink.addFamilyAndItsHFilePathToTableInMap
(byte[] family, String pathToHfileFromNS, List<Pair<byte[], List<String>>> familyHFilePathsList) void
NoopReplicationQueueStorage.addHFileRefs
(String peerId, List<Pair<org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path>> pairs) void
ReplicationSource.addHFileRefs
(TableName tableName, byte[] family, List<Pair<org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path>> pairs) void
ReplicationSourceInterface.addHFileRefs
(TableName tableName, byte[] family, List<Pair<org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path>> pairs) Add hfile names to the queue to be replicated.void
ReplicationSourceManager.addHFileRefs
(TableName tableName, byte[] family, List<Pair<org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path>> pairs) (package private) void
Replication.addHFileRefsToQueue
(TableName tableName, byte[] family, List<Pair<org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path>> pairs) private void
ReplicationSink.addNewTableEntryInMap
(Map<String, List<Pair<byte[], List<String>>>> bulkLoadHFileMap, byte[] family, String pathToHfileFromNS, String tableName) private void
ReplicationSink.buildBulkLoadHFileMap
(Map<String, List<Pair<byte[], List<String>>>> bulkLoadHFileMap, TableName table, org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.BulkLoadDescriptor bld) void
ReplicationObserver.preCommitStoreFile
(ObserverContext<RegionCoprocessorEnvironment> ctx, byte[] family, List<Pair<org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path>> pairs) void
MetricsSource.updateTableLevelMetrics
(List<Pair<WAL.Entry, Long>> walEntries) Update the table level replication metrics per tableModifierConstructorDescriptionHFileReplicator
(org.apache.hadoop.conf.Configuration sourceClusterConf, String sourceBaseNamespaceDirPath, String sourceHFileArchiveDirPath, Map<String, List<Pair<byte[], List<String>>>> tableQueueMap, org.apache.hadoop.conf.Configuration conf, Connection connection, List<String> sourceClusterIds) -
Uses of Pair in org.apache.hadoop.hbase.rest
Modifier and TypeMethodDescriptionprivate static Pair<org.apache.hbase.thirdparty.org.eclipse.jetty.servlet.FilterHolder,
Class<? extends org.apache.hbase.thirdparty.org.glassfish.jersey.servlet.ServletContainer>> RESTServer.loginServerPrincipal
(UserProvider userProvider, org.apache.hadoop.conf.Configuration conf) -
Uses of Pair in org.apache.hadoop.hbase.rsgroup
Modifier and TypeMethodDescriptionprivate Pair<Map<TableName,
Map<ServerName, List<RegionInfo>>>, List<RegionPlan>> RSGroupBasedLoadBalancer.correctAssignments
(Map<TableName, Map<ServerName, List<RegionInfo>>> existingAssignments) Modifier and TypeMethodDescriptionprivate List<Pair<List<RegionInfo>,
List<ServerName>>> RSGroupBasedLoadBalancer.generateGroupAssignments
(List<RegionInfo> regions, List<ServerName> servers) Modifier and TypeMethodDescriptionprivate boolean
RSGroupAdminServer.waitForRegionMovement
(List<Pair<RegionInfo, Future<byte[]>>> regionMoveFutures, String groupName, int retryCount) Wait for all the region move to complete. -
Uses of Pair in org.apache.hadoop.hbase.security.access
Modifier and TypeMethodDescriptionSnapshotScannerHDFSAclController.SnapshotScannerHDFSAclStorage.getUserNamespaceAndTable
(Table aclTable, String userName) private static Pair<String,
Permission> PermissionStorage.parsePermissionRecord
(byte[] entryName, Cell kv, byte[] cf, byte[] cq, boolean filterPerms, String filterUser) Modifier and TypeMethodDescriptionAccessController.postAppendBeforeWAL
(ObserverContext<RegionCoprocessorEnvironment> ctx, Mutation mutation, List<Pair<Cell, Cell>> cellPairs) AccessController.postIncrementBeforeWAL
(ObserverContext<RegionCoprocessorEnvironment> ctx, Mutation mutation, List<Pair<Cell, Cell>> cellPairs) Modifier and TypeMethodDescriptionAccessController.postAppendBeforeWAL
(ObserverContext<RegionCoprocessorEnvironment> ctx, Mutation mutation, List<Pair<Cell, Cell>> cellPairs) AccessController.postIncrementBeforeWAL
(ObserverContext<RegionCoprocessorEnvironment> ctx, Mutation mutation, List<Pair<Cell, Cell>> cellPairs) void
AccessController.preBulkLoadHFile
(ObserverContext<RegionCoprocessorEnvironment> ctx, List<Pair<byte[], String>> familyPaths) Verifies user has CREATE or ADMIN privileges on the Column Families involved in the bulkLoadHFile request. -
Uses of Pair in org.apache.hadoop.hbase.security.provider
Modifier and TypeMethodDescriptionPair<SaslClientAuthenticationProvider,
org.apache.hadoop.security.token.Token<? extends org.apache.hadoop.security.token.TokenIdentifier>> SaslClientAuthenticationProviders.getSimpleProvider()
Returns the provider and token pair for SIMPLE authentication.Pair<SaslClientAuthenticationProvider,
org.apache.hadoop.security.token.Token<? extends org.apache.hadoop.security.token.TokenIdentifier>> AuthenticationProviderSelector.selectProvider
(String clusterId, User user) Chooses the authentication provider which should be used given the provided client context from the authentication providers passed in viaAuthenticationProviderSelector.configure(Configuration, Collection)
.Pair<SaslClientAuthenticationProvider,
org.apache.hadoop.security.token.Token<? extends org.apache.hadoop.security.token.TokenIdentifier>> BuiltInProviderSelector.selectProvider
(String clusterId, User user) Pair<SaslClientAuthenticationProvider,
org.apache.hadoop.security.token.Token<? extends org.apache.hadoop.security.token.TokenIdentifier>> SaslClientAuthenticationProviders.selectProvider
(String clusterId, User clientUser) Chooses the best authentication provider and corresponding token given the HBase cluster identifier and the user. -
Uses of Pair in org.apache.hadoop.hbase.security.provider.example
Modifier and TypeMethodDescriptionPair<SaslClientAuthenticationProvider,
org.apache.hadoop.security.token.Token<? extends org.apache.hadoop.security.token.TokenIdentifier>> ShadeProviderSelector.selectProvider
(String clusterId, User user) -
Uses of Pair in org.apache.hadoop.hbase.security.visibility
Modifier and TypeFieldDescriptionVisibilityScanDeleteTracker.visibilityTagsDeleteColumns
VisibilityScanDeleteTracker.visiblityTagsDeleteColumnVersion
Modifier and TypeMethodDescriptionVisibilityController.checkForReservedVisibilityTagPresence
(Cell cell, Pair<Boolean, Tag> pair) Checks whether cell contains any tag with type as VISIBILITY_TAG_TYPE.DefaultVisibilityLabelServiceImpl.extractLabelsAndAuths
(List<List<Cell>> labelDetails) Modifier and TypeMethodDescriptionVisibilityController.postAppendBeforeWAL
(ObserverContext<RegionCoprocessorEnvironment> ctx, Mutation mutation, List<Pair<Cell, Cell>> cellPairs) VisibilityController.postIncrementBeforeWAL
(ObserverContext<RegionCoprocessorEnvironment> ctx, Mutation mutation, List<Pair<Cell, Cell>> cellPairs) Modifier and TypeMethodDescriptionVisibilityController.checkForReservedVisibilityTagPresence
(Cell cell, Pair<Boolean, Tag> pair) Checks whether cell contains any tag with type as VISIBILITY_TAG_TYPE.Modifier and TypeMethodDescriptionVisibilityController.postAppendBeforeWAL
(ObserverContext<RegionCoprocessorEnvironment> ctx, Mutation mutation, List<Pair<Cell, Cell>> cellPairs) VisibilityController.postIncrementBeforeWAL
(ObserverContext<RegionCoprocessorEnvironment> ctx, Mutation mutation, List<Pair<Cell, Cell>> cellPairs) -
Uses of Pair in org.apache.hadoop.hbase.snapshot
Modifier and TypeFieldDescriptionExportSnapshot.ExportSnapshotInputFormat.ExportSnapshotInputSplit.files
ExportSnapshot.ExportSnapshotInputFormat.ExportSnapshotRecordReader.files
RestoreSnapshotHelper.parentsMap
RestoreSnapshotHelper.RestoreMetaChanges.parentsMap
Modifier and TypeMethodDescriptionprivate static Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo,
Long> ExportSnapshot.getSnapshotFileAndSize
(org.apache.hadoop.fs.FileSystem fs, org.apache.hadoop.conf.Configuration conf, TableName table, String region, String family, String hfile, long size) Modifier and TypeMethodDescription(package private) static List<List<Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo,
Long>>> ExportSnapshot.getBalancedSplits
(List<Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo, Long>> files, int ngroups) Given a list of file paths and sizes, create around ngroups in as balanced a way as possible.RestoreSnapshotHelper.RestoreMetaChanges.getParentToChildrenPairMap()
Returns the map of parent-children_pair.private static List<Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo,
Long>> ExportSnapshot.getSnapshotFiles
(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.fs.FileSystem fs, org.apache.hadoop.fs.Path snapshotDir) Extract the list of files (HFiles/WALs) to copy using Map-Reduce.ExportSnapshot.ExportSnapshotInputFormat.ExportSnapshotInputSplit.getSplitKeys()
Modifier and TypeMethodDescription(package private) static List<List<Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo,
Long>>> ExportSnapshot.getBalancedSplits
(List<Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo, Long>> files, int ngroups) Given a list of file paths and sizes, create around ngroups in as balanced a way as possible.ModifierConstructorDescriptionExportSnapshotInputSplit
(List<Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo, Long>> snapshotFiles) (package private)
ExportSnapshotRecordReader
(List<Pair<org.apache.hadoop.io.BytesWritable, Long>> files) RestoreMetaChanges
(TableDescriptor htd, Map<String, Pair<String, String>> parentsMap) -
Uses of Pair in org.apache.hadoop.hbase.thrift2.client
Modifier and TypeMethodDescriptionThriftAdmin.getAlterStatus
(byte[] tableName) ThriftAdmin.getAlterStatus
(TableName tableName) abstract Pair<org.apache.hadoop.hbase.thrift2.generated.THBaseService.Client,
org.apache.thrift.transport.TTransport> ThriftClientBuilder.getClient()
Pair<org.apache.hadoop.hbase.thrift2.generated.THBaseService.Client,
org.apache.thrift.transport.TTransport> ThriftConnection.DefaultThriftClientBuilder.getClient()
Pair<org.apache.hadoop.hbase.thrift2.generated.THBaseService.Client,
org.apache.thrift.transport.TTransport> ThriftConnection.HTTPThriftClientBuilder.getClient()
-
Uses of Pair in org.apache.hadoop.hbase.tool
Modifier and TypeMethodDescriptionprotected Pair<List<LoadIncrementalHFiles.LoadQueueItem>,
String> LoadIncrementalHFiles.groupOrSplit
(org.apache.hbase.thirdparty.com.google.common.collect.Multimap<ByteBuffer, LoadIncrementalHFiles.LoadQueueItem> regionGroups, LoadIncrementalHFiles.LoadQueueItem item, Table table, Pair<byte[][], byte[][]> startEndKeys) Deprecated.Attempt to assign the given load queue item into its target region group.private Pair<org.apache.hbase.thirdparty.com.google.common.collect.Multimap<ByteBuffer,
LoadIncrementalHFiles.LoadQueueItem>, Set<String>> LoadIncrementalHFiles.groupOrSplitPhase
(Table table, ExecutorService pool, Deque<LoadIncrementalHFiles.LoadQueueItem> queue, Pair<byte[][], byte[][]> startEndKeys) Deprecated.Modifier and TypeMethodDescriptionprivate void
LoadIncrementalHFiles.checkRegionIndexValid
(int idx, Pair<byte[][], byte[][]> startEndKeys, TableName tableName) Deprecated.we can consider there is a region hole in following conditions.private int
LoadIncrementalHFiles.getRegionIndex
(Pair<byte[][], byte[][]> startEndKeys, byte[] key) Deprecated.protected Pair<List<LoadIncrementalHFiles.LoadQueueItem>,
String> LoadIncrementalHFiles.groupOrSplit
(org.apache.hbase.thirdparty.com.google.common.collect.Multimap<ByteBuffer, LoadIncrementalHFiles.LoadQueueItem> regionGroups, LoadIncrementalHFiles.LoadQueueItem item, Table table, Pair<byte[][], byte[][]> startEndKeys) Deprecated.Attempt to assign the given load queue item into its target region group.private Pair<org.apache.hbase.thirdparty.com.google.common.collect.Multimap<ByteBuffer,
LoadIncrementalHFiles.LoadQueueItem>, Set<String>> LoadIncrementalHFiles.groupOrSplitPhase
(Table table, ExecutorService pool, Deque<LoadIncrementalHFiles.LoadQueueItem> queue, Pair<byte[][], byte[][]> startEndKeys) Deprecated.void
LoadIncrementalHFiles.loadHFileQueue
(Table table, Connection conn, Deque<LoadIncrementalHFiles.LoadQueueItem> queue, Pair<byte[][], byte[][]> startEndKeys) Deprecated.Used by the replication sink to load the hfiles from the source cluster.void
LoadIncrementalHFiles.loadHFileQueue
(Table table, Connection conn, Deque<LoadIncrementalHFiles.LoadQueueItem> queue, Pair<byte[][], byte[][]> startEndKeys, boolean copyFile) Deprecated.Used by the replication sink to load the hfiles from the source cluster. -
Uses of Pair in org.apache.hadoop.hbase.util
Modifier and TypeMethodDescriptionstatic Pair<org.apache.hadoop.fs.Path,
org.apache.hadoop.fs.FSDataOutputStream> HBaseFsck.checkAndMarkRunningHbck
(org.apache.hadoop.conf.Configuration conf, RetryCounter retryCounter) Deprecated.This method maintains a lock using a file.MunkresAssignment.findUncoveredZero()
Find a zero cost assignment which is not covered.private static Pair<org.apache.hadoop.fs.Path,
org.apache.hadoop.fs.Path> RegionSplitter.getTableDirAndSplitFile
(org.apache.hadoop.conf.Configuration conf, TableName tableName) static <T1,
T2> Pair<T1, T2> Pair.newPair
(T1 a, T2 b) Constructs a new pair, inferring the type via the passed argumentsMunkresAssignment.primeInRow
(int r) Find a primed zero in the specified row.MunkresAssignment.starInCol
(int c) Find a starred zero in the specified column.MunkresAssignment.starInRow
(int r) Find a starred zero in a specified row.Modifier and TypeMethodDescription(package private) static LinkedList<Pair<byte[],
byte[]>> RegionSplitter.getSplits
(Connection connection, TableName tableName, RegionSplitter.SplitAlgorithm splitAlgo) private static Optional<Pair<org.apache.hadoop.fs.FileStatus,
TableDescriptor>> FSTableDescriptors.getTableDescriptorFromFs
(org.apache.hadoop.fs.FileSystem fs, org.apache.hadoop.fs.Path tableDir, boolean readonly) (package private) static LinkedList<Pair<byte[],
byte[]>> RegionSplitter.splitScan
(LinkedList<Pair<byte[], byte[]>> regionList, Connection connection, TableName tableName, RegionSplitter.SplitAlgorithm splitAlgo) Modifier and TypeMethodDescription(package private) static LinkedList<Pair<byte[],
byte[]>> RegionSplitter.splitScan
(LinkedList<Pair<byte[], byte[]>> regionList, Connection connection, TableName tableName, RegionSplitter.SplitAlgorithm splitAlgo) -
Uses of Pair in org.apache.hadoop.hbase.wal
Modifier and TypeFieldDescriptionprivate static final Map<String,
Pair<org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroup, Class<? extends org.apache.hbase.thirdparty.io.netty.channel.Channel>>> NettyAsyncFSWALConfigHelper.EVENT_LOOP_CONFIG_MAP
Modifier and TypeMethodDescription(package private) static Pair<org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroup,
Class<? extends org.apache.hbase.thirdparty.io.netty.channel.Channel>> NettyAsyncFSWALConfigHelper.getEventLoopConfig
(org.apache.hadoop.conf.Configuration conf) Modifier and TypeMethodDescriptionstatic List<WALSplitUtil.MutationReplay>
WALSplitUtil.getMutationsFromWALEntry
(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.WALEntry entry, CellScanner cells, Pair<WALKey, WALEdit> logEntry, Durability durability) This function is used to construct mutations from a WALEntry.
MetaTableAccessor.getRegionLocation(Connection, byte[])
instead