Uses of Class
org.apache.hadoop.hbase.client.Result
Package
Description
Provides HBase Client
Table of Contents
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
HBase REST
Provides an HBase Thrift
service.
Provides an HBase Thrift
service.
-
Uses of Result in org.apache.hadoop.hbase
Modifier and TypeMethodDescriptionprivate static Result
static Result
MetaTableAccessor.getCatalogFamilyRow
(Connection connection, RegionInfo ri) Returns Return theHConstants.CATALOG_FAMILY
row from hbase:meta.static Result
MetaTableAccessor.getRegionResult
(Connection connection, RegionInfo regionInfo) Gets the result in hbase:meta for the specified region.static Result
MetaTableAccessor.scanByRegionEncodedName
(Connection connection, String regionEncodedName) Scans META table for a row whose key contains the specified regionEncodedName, returning a single relatedResult
instance if any row is found, null otherwise.Modifier and TypeMethodDescriptionMetaTableAccessor.fullScan
(Connection connection, MetaTableAccessor.QueryType type) Performs a full scan ofhbase:meta
.MetaTableAccessor.fullScanRegions
(Connection connection) Performs a full scan ofhbase:meta
for regions.static NavigableMap<RegionInfo,
Result> MetaTableAccessor.getServerUserRegions
(Connection connection, ServerName serverName) Get the user regions a given server is hosting.Modifier and TypeMethodDescription(package private) void
(package private) abstract void
static PairOfSameType<RegionInfo>
MetaTableAccessor.getDaughterRegions
(Result data) Returns the daughter regions by reading the corresponding columns of the catalog table Result.private static Optional<RegionInfo>
AsyncMetaTableAccessor.getHRegionInfo
(Result r, byte[] qualifier) Returns the RegionInfo object from the columnHConstants.CATALOG_FAMILY
andqualifier
of the catalog table result.static RegionInfo
MetaTableAccessor.getRegionInfo
(Result data) Returns RegionInfo object from the column HConstants.CATALOG_FAMILY:HConstants.REGIONINFO_QUALIFIER of the catalog table Result.static RegionInfo
MetaTableAccessor.getRegionInfo
(Result r, byte[] qualifier) Returns the RegionInfo object from the columnHConstants.CATALOG_FAMILY
andqualifier
of the catalog table result.private static HRegionLocation
AsyncMetaTableAccessor.getRegionLocation
(Result r, RegionInfo regionInfo, int replicaId) Returns the HRegionLocation parsed from the given meta row Result for the given regionInfo and replicaId.private static HRegionLocation
MetaTableAccessor.getRegionLocation
(Result r, RegionInfo regionInfo, int replicaId) Returns the HRegionLocation parsed from the given meta row Result for the given regionInfo and replicaId.private static Optional<RegionLocations>
AsyncMetaTableAccessor.getRegionLocations
(Result r) Returns an HRegionLocationList extracted from the result.static RegionLocations
MetaTableAccessor.getRegionLocations
(Result r) Returns an HRegionLocationList extracted from the result.private static MetaTableAccessor.ReplicationBarrierResult
MetaTableAccessor.getReplicationBarrierResult
(Result result) static long[]
MetaTableAccessor.getReplicationBarriers
(Result result) private static RSGroupInfo
RSGroupTableAccessor.getRSGroupInfo
(Result result) private static long
AsyncMetaTableAccessor.getSeqNumDuringOpen
(Result r, int replicaId) The latest seqnum that the server writing to meta observed when opening the region.private static long
MetaTableAccessor.getSeqNumDuringOpen
(Result r, int replicaId) The latest seqnum that the server writing to meta observed when opening the region.private static Optional<ServerName>
AsyncMetaTableAccessor.getServerName
(Result r, int replicaId) Returns aServerName
from catalog tableResult
.static ServerName
MetaTableAccessor.getServerName
(Result r, int replicaId) Returns aServerName
from catalog tableResult
.private static Optional<TableState>
AsyncMetaTableAccessor.getTableState
(Result r) static TableState
MetaTableAccessor.getTableState
(Result r) Decode table state from META Result.static ServerName
MetaTableAccessor.getTargetServerName
(Result r, int replicaId) Returns theServerName
from catalog tableResult
where the region is transitioning on.void
AsyncMetaTableAccessor.MetaTableScanResultConsumer.onNext
(Result[] results, AdvancedScanResultConsumer.ScanController controller) boolean
boolean
final boolean
boolean
Visit the catalog table row.abstract boolean
MetaTableAccessor.DefaultVisitorBase.visitInternal
(Result rowResult) -
Uses of Result in org.apache.hadoop.hbase.backup.impl
Modifier and TypeMethodDescriptionprivate BackupInfo
BackupSystemTable.resultToBackupInfo
(Result res) Converts Result to BackupInfo -
Uses of Result in org.apache.hadoop.hbase.client
Modifier and TypeFieldDescriptionstatic final Result
Result.EMPTY_RESULT
static final Result[]
ScanResultCache.EMPTY_RESULT_ARRAY
protected Result
ClientScanner.lastResult
private Result
ScannerCallableWithReplicas.lastResult
private final Result
CheckAndMutateResult.result
private Result
SingleResponse.Entry.result
Modifier and TypeFieldDescriptionClientScanner.cache
BatchScanResultCache.partialResults
CompleteScanResultCache.partialResults
AsyncTableResultScanner.queue
Modifier and TypeMethodDescriptionResult[]
Result[]
Result[]
Result[]
Add the given results to cache and get valid results back.default Result
Appends values to one or more columns within a single row.private Result[]
ClientScanner.call
(ScannerCallableWithReplicas callable, RpcRetryingCaller<Result[]> caller, int scannerTimeout, boolean updateCurrentRegion) RpcRetryingCallerWithReadReplicas.call
(int operationTimeout) Algo: - we put the query into the execution pool.Result[]
ScannerCallableWithReplicas.call
(int timeout) private Result
CompleteScanResultCache.combine()
static Result
Instantiate a Result with the specified List of KeyValues.static Result
static Result
static Result
static Result
Instantiate a Result with the specified array of KeyValues.static Result
static Result
private Result
BatchScanResultCache.createCompletedResult()
static Result
Result.createCompleteResult
(Iterable<Result> partialResults) Forms a single result from the partial results in the partialResults list.static Result
Result.createCursorResult
(Cursor cursor) (package private) static Result
ConnectionUtils.filterCells
(Result result, Cell keepCellsAfter) Result[]
private Result
default Result[]
Extracts specified cells from the given rows, as a batch.default Result
Extracts certain cells from a given row.CheckAndMutateResult.getResult()
Returns It is used only for CheckAndMutate operations with Increment/Append.SingleResponse.Entry.getResult()
default Result
Increments one or more columns within a single row.HTable.mutateRow
(RowMutations rm) default Result
Table.mutateRow
(RowMutations rm) Performs multiple mutations atomically on a single row.AsyncTableResultScanner.next()
ClientAsyncPrefetchScanner.next()
ClientScanner.next()
ClientSideRegionScanner.next()
ResultScanner.next()
Grab the next row's worth of values.default Result[]
ResultScanner.next
(int nbRows) Get nbRows rows.TableSnapshotScanner.next()
protected Result
ClientScanner.nextWithSyncCache()
private Result
ClientAsyncPrefetchScanner.pollCache()
private Result[]
CompleteScanResultCache.prependCombined
(Result[] results, int length) private Result
BatchScanResultCache.regroupResults
(Result result) protected Result
RpcRetryingCallerWithReadReplicas.ReplicaRegionServerCallable.rpcCall()
protected Result[]
ScannerCallable.rpcCall()
private static Result
RawAsyncTableImpl.toResult
(HBaseRpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.MutateResponse resp) private Result[]
CompleteScanResultCache.updateNumberOfCompleteResultsAndReturn
(Result... results) Modifier and TypeMethodDescriptionAppends values to one or more columns within a single row.Extracts certain cells from the given rows, in batch.Extracts certain cells from a given row.private CompletableFuture<Result>
default CompletableFuture<List<Result>>
A simple version for batch get.Increments one or more columns within a single row.ResultScanner.iterator()
AsyncTable.mutateRow
(RowMutations mutation) Performs multiple mutations atomically on a single row.AsyncTableImpl.mutateRow
(RowMutations mutation) RawAsyncTableImpl.mutateRow
(RowMutations mutations) Return all the results that match the given scan object.Modifier and TypeMethodDescriptionResult[]
Result[]
Result[]
Result[]
Add the given results to cache and get valid results back.private void
AsyncTableResultScanner.addToCache
(Result result) (package private) static long
ConnectionUtils.calcEstimatedSize
(Result rs) static void
Result.compareResults
(Result res1, Result res2) Does a deep comparison of two Results, down to the byte arrays.static void
Result.compareResults
(Result res1, Result res2, boolean verbose) Does a deep comparison of two Results, down to the byte arrays.void
Copy another Result into this one.(package private) static Result
ConnectionUtils.filterCells
(Result result, Cell keepCellsAfter) static long
Result.getTotalSizeOfCells
(Result result) Get total size of raw cellsvoid
AdvancedScanResultConsumer.onNext
(Result[] results, AdvancedScanResultConsumer.ScanController controller) Indicate that we have receive some data.void
AsyncTableResultScanner.onNext
(Result[] results, AdvancedScanResultConsumer.ScanController controller) boolean
Returnfalse
if you want to terminate the scan process.private boolean
AsyncNonMetaRegionLocator.onScanNext
(TableName tableName, AsyncNonMetaRegionLocator.LocateRequest req, Result result) private Result[]
CompleteScanResultCache.prependCombined
(Result[] results, int length) private void
AllowPartialScanResultCache.recordLastResult
(Result result) private void
BatchScanResultCache.recordLastResult
(Result result) private boolean
ClientScanner.regionExhausted
(Result[] values) private Result
BatchScanResultCache.regroupResults
(Result result) void
private void
ScannerCallableWithReplicas.updateCurrentlyServingReplica
(ScannerCallable scanner, Result[] result, AtomicBoolean done, ExecutorService pool) private void
AsyncScanSingleRegionRpcRetryingCaller.updateNextStartRowWhenError
(Result result) private Result[]
CompleteScanResultCache.updateNumberOfCompleteResultsAndReturn
(Result... results) (package private) static void
ConnectionUtils.updateResultsMetrics
(ScanMetrics scanMetrics, Result[] rrs, boolean isRegionServerRemote) Modifier and TypeMethodDescriptionprivate void
RpcRetryingCallerWithReadReplicas.addCallsForReplica
(ResultBoundedCompletionService<Result> cs, RegionLocations rl, int min, int max) Creates the calls and submit themstatic Result
Result.createCompleteResult
(Iterable<Result> partialResults) Forms a single result from the partial results in the partialResults list. -
Uses of Result in org.apache.hadoop.hbase.coprocessor
Modifier and TypeMethodDescriptiondefault Result
RegionObserver.postAppend
(ObserverContext<RegionCoprocessorEnvironment> c, Append append, Result result) Deprecated.since 2.5.0 and will be removed in 4.0.0.default Result
RegionObserver.postAppend
(ObserverContext<RegionCoprocessorEnvironment> c, Append append, Result result, WALEdit edit) Called after Appenddefault Result
RegionObserver.postIncrement
(ObserverContext<RegionCoprocessorEnvironment> c, Increment increment, Result result) Deprecated.since 2.5.0 and will be removed in 4.0.0.default Result
RegionObserver.postIncrement
(ObserverContext<RegionCoprocessorEnvironment> c, Increment increment, Result result, WALEdit edit) Called after incrementdefault Result
RegionObserver.preAppend
(ObserverContext<RegionCoprocessorEnvironment> c, Append append) Deprecated.since 2.5.0 and will be removed in 4.0.0.default Result
RegionObserver.preAppend
(ObserverContext<RegionCoprocessorEnvironment> c, Append append, WALEdit edit) Called before Append.default Result
RegionObserver.preAppendAfterRowLock
(ObserverContext<RegionCoprocessorEnvironment> c, Append append) Deprecated.since 2.5.0 and will be removed in 4.0.0.default Result
RegionObserver.preIncrement
(ObserverContext<RegionCoprocessorEnvironment> c, Increment increment) Deprecated.since 2.5.0 and will be removed in 4.0.0.default Result
RegionObserver.preIncrement
(ObserverContext<RegionCoprocessorEnvironment> c, Increment increment, WALEdit edit) Called before Increment.default Result
RegionObserver.preIncrementAfterRowLock
(ObserverContext<RegionCoprocessorEnvironment> c, Increment increment) Deprecated.since 2.5.0 and will be removed in 4.0.0.Modifier and TypeMethodDescriptiondefault Result
RegionObserver.postAppend
(ObserverContext<RegionCoprocessorEnvironment> c, Append append, Result result) Deprecated.since 2.5.0 and will be removed in 4.0.0.default Result
RegionObserver.postAppend
(ObserverContext<RegionCoprocessorEnvironment> c, Append append, Result result, WALEdit edit) Called after Appenddefault Result
RegionObserver.postIncrement
(ObserverContext<RegionCoprocessorEnvironment> c, Increment increment, Result result) Deprecated.since 2.5.0 and will be removed in 4.0.0.default Result
RegionObserver.postIncrement
(ObserverContext<RegionCoprocessorEnvironment> c, Increment increment, Result result, WALEdit edit) Called after incrementModifier and TypeMethodDescriptiondefault boolean
RegionObserver.postScannerNext
(ObserverContext<RegionCoprocessorEnvironment> c, InternalScanner s, List<Result> result, int limit, boolean hasNext) Called after the client asks for the next row on a scanner.default boolean
RegionObserver.preScannerNext
(ObserverContext<RegionCoprocessorEnvironment> c, InternalScanner s, List<Result> result, int limit, boolean hasNext) Called before the client asks for the next row on a scanner. -
Uses of Result in org.apache.hadoop.hbase.coprocessor.example
Modifier and TypeMethodDescriptionWriteHeavyIncrementObserver.preIncrement
(ObserverContext<RegionCoprocessorEnvironment> c, Increment increment) -
Uses of Result in org.apache.hadoop.hbase.mapred
Modifier and TypeMethodDescriptionTableRecordReader.createValue()
TableRecordReaderImpl.createValue()
TableSnapshotInputFormat.TableSnapshotRecordReader.createValue()
Modifier and TypeMethodDescriptionorg.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,
Result> MultiTableSnapshotInputFormat.getRecordReader
(org.apache.hadoop.mapred.InputSplit split, org.apache.hadoop.mapred.JobConf job, org.apache.hadoop.mapred.Reporter reporter) org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,
Result> TableInputFormatBase.getRecordReader
(org.apache.hadoop.mapred.InputSplit split, org.apache.hadoop.mapred.JobConf job, org.apache.hadoop.mapred.Reporter reporter) Builds a TableRecordReader.org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,
Result> TableSnapshotInputFormat.getRecordReader
(org.apache.hadoop.mapred.InputSplit split, org.apache.hadoop.mapred.JobConf job, org.apache.hadoop.mapred.Reporter reporter) Modifier and TypeMethodDescriptionprotected byte[][]
GroupingTableMap.extractKeyValues
(Result r) Extract columns values from the current record.void
GroupingTableMap.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable, Result> output, org.apache.hadoop.mapred.Reporter reporter) Extract the grouping columns from value to construct a new key.void
IdentityTableMap.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable, Result> output, org.apache.hadoop.mapred.Reporter reporter) Pass the key, value to reducevoid
RowCounter.RowCounterMapper.map
(ImmutableBytesWritable row, Result values, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable, Result> output, org.apache.hadoop.mapred.Reporter reporter) boolean
TableRecordReader.next
(ImmutableBytesWritable key, Result value) boolean
TableRecordReaderImpl.next
(ImmutableBytesWritable key, Result value) boolean
TableSnapshotInputFormat.TableSnapshotRecordReader.next
(ImmutableBytesWritable key, Result value) Modifier and TypeMethodDescriptionvoid
GroupingTableMap.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable, Result> output, org.apache.hadoop.mapred.Reporter reporter) Extract the grouping columns from value to construct a new key.void
IdentityTableMap.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable, Result> output, org.apache.hadoop.mapred.Reporter reporter) Pass the key, value to reducevoid
RowCounter.RowCounterMapper.map
(ImmutableBytesWritable row, Result values, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable, Result> output, org.apache.hadoop.mapred.Reporter reporter) -
Uses of Result in org.apache.hadoop.hbase.mapreduce
Modifier and TypeFieldDescriptionprivate Result
SyncTable.SyncMapper.CellScanner.currentRowResult
private Result
SyncTable.SyncMapper.CellScanner.nextRowResult
private Result
TableSnapshotInputFormatImpl.RecordReader.result
private Result
MultithreadedTableMapper.SubMapRecordReader.value
private Result
TableRecordReaderImpl.value
Modifier and TypeFieldDescriptionprivate Class<? extends org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable,
Result, K2, V2>> MultithreadedTableMapper.mapClass
private org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable,
Result, K2, V2> MultithreadedTableMapper.MapRunner.mapper
SyncTable.SyncMapper.CellScanner.results
Modifier and TypeMethodDescriptionResultSerialization.Result94Deserializer.deserialize
(Result mutation) ResultSerialization.ResultDeserializer.deserialize
(Result mutation) MultithreadedTableMapper.SubMapRecordReader.getCurrentValue()
TableRecordReader.getCurrentValue()
Returns the current value.TableRecordReaderImpl.getCurrentValue()
Returns the current value.TableSnapshotInputFormat.TableSnapshotRegionRecordReader.getCurrentValue()
TableSnapshotInputFormatImpl.RecordReader.getCurrentValue()
Modifier and TypeMethodDescriptionorg.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,
Result> MultiTableInputFormatBase.createRecordReader
(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context) Builds a TableRecordReader.org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,
Result> TableInputFormatBase.createRecordReader
(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context) Builds aTableRecordReader
.org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,
Result> TableSnapshotInputFormat.createRecordReader
(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context) org.apache.hadoop.io.serializer.Deserializer<Result>
ResultSerialization.getDeserializer
(Class<Result> c) static <K2,
V2> Class<org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, K2, V2>> MultithreadedTableMapper.getMapperClass
(org.apache.hadoop.mapreduce.JobContext job) Get the application's mapper class.org.apache.hadoop.io.serializer.Serializer<Result>
ResultSerialization.getSerializer
(Class<Result> c) Modifier and TypeMethodDescriptionResultSerialization.Result94Deserializer.deserialize
(Result mutation) ResultSerialization.ResultDeserializer.deserialize
(Result mutation) protected byte[][]
GroupingTableMapper.extractKeyValues
(Result r) Extract columns values from the current record.void
HashTable.ResultHasher.hashResult
(Result result) void
CellCounter.CellCounterMapper.map
(ImmutableBytesWritable row, Result values, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, org.apache.hadoop.io.Text, org.apache.hadoop.io.LongWritable>.org.apache.hadoop.mapreduce.Mapper.Context context) Maps the data.void
GroupingTableMapper.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Result>.org.apache.hadoop.mapreduce.Mapper.Context context) Extract the grouping columns from value to construct a new key.protected void
HashTable.HashMapper.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, ImmutableBytesWritable>.org.apache.hadoop.mapreduce.Mapper.Context context) void
IdentityTableMapper.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Result>.org.apache.hadoop.mapreduce.Mapper.Context context) Pass the key, value to reduce.void
Import.CellImporter.map
(ImmutableBytesWritable row, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Cell>.org.apache.hadoop.mapreduce.Mapper.Context context) void
Import.CellSortImporter.map
(ImmutableBytesWritable row, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, Import.CellWritableComparable, Cell>.org.apache.hadoop.mapreduce.Mapper.Context context) void
Import.Importer.map
(ImmutableBytesWritable row, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Mutation>.org.apache.hadoop.mapreduce.Mapper.Context context) void
Import.KeyValueImporter.map
(ImmutableBytesWritable row, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, KeyValue>.org.apache.hadoop.mapreduce.Mapper.Context context) Deprecated.void
Import.KeyValueSortImporter.map
(ImmutableBytesWritable row, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, Import.KeyValueWritableComparable, KeyValue>.org.apache.hadoop.mapreduce.Mapper.Context context) Deprecated.protected void
IndexBuilder.Map.map
(ImmutableBytesWritable rowKey, Result result, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Put>.org.apache.hadoop.mapreduce.Mapper.Context context) void
RowCounter.RowCounterMapper.map
(ImmutableBytesWritable row, Result values, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Result>.org.apache.hadoop.mapreduce.Mapper.Context context) Maps the data.protected void
SyncTable.SyncMapper.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Mutation>.org.apache.hadoop.mapreduce.Mapper.Context context) protected void
Import.Importer.processKV
(ImmutableBytesWritable key, Result result, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Mutation>.org.apache.hadoop.mapreduce.Mapper.Context context, Put put, Delete delete) void
private void
Import.Importer.writeResult
(ImmutableBytesWritable key, Result result, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Mutation>.org.apache.hadoop.mapreduce.Mapper.Context context) Modifier and TypeMethodDescriptionorg.apache.hadoop.io.serializer.Deserializer<Result>
ResultSerialization.getDeserializer
(Class<Result> c) org.apache.hadoop.io.serializer.Serializer<Result>
ResultSerialization.getSerializer
(Class<Result> c) static <K2,
V2> void MultithreadedTableMapper.setMapperClass
(org.apache.hadoop.mapreduce.Job job, Class<? extends org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, K2, V2>> cls) Set the application's mapper class. -
Uses of Result in org.apache.hadoop.hbase.mapreduce.replication
Modifier and TypeFieldDescriptionprivate Result
VerifyReplication.Verifier.currentCompareRowInPeerTable
private Result
VerifyReplicationRecompareRunnable.replicatedResult
private Result
VerifyReplicationRecompareRunnable.sourceResult
Modifier and TypeMethodDescriptionprotected static byte[]
private void
VerifyReplication.Verifier.logFailRowAndIncreaseCounter
(org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Put>.org.apache.hadoop.mapreduce.Mapper.Context context, VerifyReplication.Verifier.Counters counter, Result row, Result replicatedRow) void
VerifyReplication.Verifier.map
(ImmutableBytesWritable row, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Put>.org.apache.hadoop.mapreduce.Mapper.Context context) Map method that compares every scanned row with the equivalent from a distant cluster.private boolean
VerifyReplicationRecompareRunnable.matches
(Result original, Result updated, VerifyReplication.Verifier.Counters failCounter) ModifierConstructorDescriptionVerifyReplicationRecompareRunnable
(org.apache.hadoop.mapreduce.Mapper.Context context, Result sourceResult, Result replicatedResult, VerifyReplication.Verifier.Counters originalCounter, String delimiter, Scan tableScan, Table sourceTable, Table replicatedTable, int reCompareTries, int sleepMsBeforeReCompare, int reCompareBackoffExponent, boolean verbose) -
Uses of Result in org.apache.hadoop.hbase.master
Modifier and TypeMethodDescriptionprivate void
SnapshotOfRegionAssignmentFromMeta.processMetaRecord
(Result result) -
Uses of Result in org.apache.hadoop.hbase.master.assignment
Modifier and TypeMethodDescriptionstatic RegionState.State
RegionStateStore.getRegionState
(Result r, RegionInfo regionInfo) Pull the region state from a catalog tableResult
.static void
RegionStateStore.visitMetaEntry
(RegionStateStore.RegionStateVisitor visitor, Result result) void
AssignmentManager.RegionMetaLoadingVisitor.visitRegionState
(Result result, RegionInfo regionInfo, RegionState.State state, ServerName regionLocation, ServerName lastHost, long openSeqNum) void
RegionStateStore.RegionStateVisitor.visitRegionState
(Result result, RegionInfo regionInfo, RegionState.State state, ServerName regionLocation, ServerName lastHost, long openSeqNum) -
Uses of Result in org.apache.hadoop.hbase.master.http
-
Uses of Result in org.apache.hadoop.hbase.master.janitor
Modifier and TypeFieldDescription(package private) final Map<RegionInfo,
Result> CatalogJanitorReport.mergedRegions
(package private) final Map<RegionInfo,
Result> CatalogJanitorReport.splitParents
Modifier and TypeMethodDescriptionprivate boolean
CatalogJanitor.cleanParent
(RegionInfo parent, Result rowContent) If daughters no longer hold reference to the parents, delete the parent.(package private) static boolean
CatalogJanitor.cleanParent
(MasterServices services, RegionInfo parent, Result rowContent) private RegionInfo
ReportMakingVisitor.metaTableConsistencyCheck
(Result metaTableRow) Check row.boolean
-
Uses of Result in org.apache.hadoop.hbase.master.procedure
-
Uses of Result in org.apache.hadoop.hbase.master.region
-
Uses of Result in org.apache.hadoop.hbase.mob.mapreduce
Modifier and TypeMethodDescriptionvoid
MobRefReporter.MobRefMapper.map
(ImmutableBytesWritable r, Result columns, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, org.apache.hadoop.io.Text, ImmutableBytesWritable>.org.apache.hadoop.mapreduce.Mapper.Context context) -
Uses of Result in org.apache.hadoop.hbase.quotas
Modifier and TypeMethodDescriptionprotected static Result[]
QuotaTableUtil.doGet
(Connection connection, List<Get> gets) protected static Result
QuotaTableUtil.doGet
(Connection connection, Get get) Modifier and TypeMethodDescriptionvoid
DefaultOperationQuota.addGetResult
(Result result) void
NoopOperationQuota.addGetResult
(Result result) void
OperationQuota.addGetResult
(Result result) Add a get result.static long
QuotaUtil.calculateResultSize
(Result result) static void
QuotaTableUtil.extractQuotaSnapshot
(Result result, Map<TableName, SpaceQuotaSnapshot> snapshots) Extracts theSpaceViolationPolicy
andTableName
from the providedResult
and adds them to the givenMap
.(package private) void
SpaceQuotaRefresherChore.extractQuotaSnapshot
(Result result, Map<TableName, SpaceQuotaSnapshot> snapshots) Wrapper aroundQuotaTableUtil.extractQuotaSnapshot(Result, Map)
for testing.(package private) long
FileArchiverNotifierImpl.getSnapshotSizeFromResult
(Result r) Extracts the size component from a serializedSpaceQuotaSnapshot
protobuf.protected static void
QuotaTableUtil.parseNamespaceResult
(String namespace, Result result, QuotaTableUtil.NamespaceQuotasVisitor visitor) static void
QuotaTableUtil.parseNamespaceResult
(Result result, QuotaTableUtil.NamespaceQuotasVisitor visitor) private static void
QuotaTableUtil.parseRegionServerResult
(String regionServer, Result result, QuotaTableUtil.RegionServerQuotasVisitor visitor) private static void
QuotaTableUtil.parseRegionServerResult
(Result result, QuotaTableUtil.RegionServerQuotasVisitor visitor) static void
QuotaTableUtil.parseResult
(Result result, QuotaTableUtil.QuotasVisitor visitor) static void
QuotaTableUtil.parseResultToCollection
(Result result, Collection<QuotaSettings> quotaSettings) static void
QuotaTableUtil.parseTableResult
(Result result, QuotaTableUtil.TableQuotasVisitor visitor) protected static void
QuotaTableUtil.parseTableResult
(TableName table, Result result, QuotaTableUtil.TableQuotasVisitor visitor) protected static void
QuotaTableUtil.parseUserResult
(String userName, Result result, QuotaTableUtil.UserQuotasVisitor visitor) static void
QuotaTableUtil.parseUserResult
(Result result, QuotaTableUtil.UserQuotasVisitor visitor) Modifier and TypeMethodDescriptionvoid
DefaultOperationQuota.addScanResult
(List<Result> results) void
NoopOperationQuota.addScanResult
(List<Result> results) void
OperationQuota.addScanResult
(List<Result> results) Add a scan result.static long
QuotaUtil.calculateResultSize
(List<Result> results) -
Uses of Result in org.apache.hadoop.hbase.regionserver
Modifier and TypeFieldDescriptionprivate final Result
OperationStatus.result
protected final Result[]
HRegion.BatchOperation.results
Modifier and TypeMethodDescriptionPerform one or more append operations on a row.private Result
RSRpcServices.append
(HRegion region, OperationQuota quota, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.MutationProto mutation, CellScanner cellScanner, long nonceGroup, ActivePolicyEnforcement spaceQuota, RpcCallContext context) Execute an append mutation.Do a get based on the get parameter.private Result
RSRpcServices.get
(Get get, HRegion region, RSRpcServices.RegionScannersCloseCallBack closeCallBack, RpcCallContext context) OperationStatus.getResult()
Perform one or more increment operations on a row.private Result
RSRpcServices.increment
(HRegion region, OperationQuota quota, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.MutationProto mutation, CellScanner cells, long nonceGroup, ActivePolicyEnforcement spaceQuota, RpcCallContext context) Execute an increment mutation.HRegion.mutateRow
(RowMutations rm) HRegion.mutateRow
(RowMutations rm, long nonceGroup, long nonce) Region.mutateRow
(RowMutations mutations) Performs multiple mutations atomically on a single row.RegionCoprocessorHost.postAppend
(Append append, Result result, WALEdit edit) RegionCoprocessorHost.postIncrement
(Increment increment, Result result, WALEdit edit) Supports Coprocessor 'bypass'.RegionCoprocessorHost.preAppendAfterRowLock
(Append append) Supports Coprocessor 'bypass'.RegionCoprocessorHost.preIncrement
(Increment increment, WALEdit edit) Supports Coprocessor 'bypass'.RegionCoprocessorHost.preIncrementAfterRowLock
(Increment increment) Supports Coprocessor 'bypass'.Modifier and TypeMethodDescriptionprivate void
RSRpcServices.addResult
(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.MutateResponse.Builder builder, Result result, HBaseRpcController rpcc, boolean clientCellBlockSupported) (package private) void
RSRpcServices.addSize
(RpcCallContext context, Result r) Method to account for the size of retained cells.RegionCoprocessorHost.postAppend
(Append append, Result result, WALEdit edit) RegionCoprocessorHost.postIncrement
(Increment increment, Result result, WALEdit edit) Modifier and TypeMethodDescriptionprivate void
RSRpcServices.addResults
(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ScanResponse.Builder builder, List<Result> results, HBaseRpcController controller, boolean isDefaultRegion, boolean clientCellBlockSupported) boolean
RegionCoprocessorHost.postScannerNext
(InternalScanner s, List<Result> results, int limit, boolean hasMore) RegionCoprocessorHost.preScannerNext
(InternalScanner s, List<Result> results, int limit) private void
RSRpcServices.scan
(HBaseRpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ScanRequest request, RSRpcServices.RegionScannerHolder rsh, long maxQuotaResultSize, int maxResults, int limitOfRows, List<Result> results, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ScanResponse.Builder builder, RpcCall rpcCall) ModifierConstructorDescriptionOperationStatus
(HConstants.OperationStatusCode code, Result result) private
OperationStatus
(HConstants.OperationStatusCode code, Result result, String exceptionMsg) -
Uses of Result in org.apache.hadoop.hbase.rest
Modifier and TypeMethodDescriptionprivate CellSetModel
ProtobufStreamingOutput.createModelFromResults
(Result[] results) -
Uses of Result in org.apache.hadoop.hbase.security.access
Modifier and TypeMethodDescriptionAccessController.preAppend
(ObserverContext<RegionCoprocessorEnvironment> c, Append append) AccessController.preIncrement
(ObserverContext<RegionCoprocessorEnvironment> c, Increment increment) Modifier and TypeMethodDescriptionprivate static org.apache.hbase.thirdparty.com.google.common.collect.ListMultimap<String,
UserPermission> PermissionStorage.parsePermissions
(byte[] entryName, Result result, byte[] cf, byte[] cq, String user, boolean hasFilterUser) Parse and filter permission based on the specified column family, column qualifier and user name.Modifier and TypeMethodDescriptionboolean
AccessController.preScannerNext
(ObserverContext<RegionCoprocessorEnvironment> c, InternalScanner s, List<Result> result, int limit, boolean hasNext) -
Uses of Result in org.apache.hadoop.hbase.security.visibility
Modifier and TypeMethodDescriptionboolean
VisibilityController.preScannerNext
(ObserverContext<RegionCoprocessorEnvironment> c, InternalScanner s, List<Result> result, int limit, boolean hasNext) -
Uses of Result in org.apache.hadoop.hbase.thrift
Modifier and TypeMethodDescriptionprivate Result
ThriftHBaseServiceHandler.getReverseScanResult
(byte[] tableName, byte[] row, byte[] family) Modifier and TypeMethodDescriptionstatic List<org.apache.hadoop.hbase.thrift.generated.TRowResult>
ThriftUtilities.rowResultFromHBase
(Result in) static List<org.apache.hadoop.hbase.thrift.generated.TRowResult>
ThriftUtilities.rowResultFromHBase
(Result[] in) This utility method creates a list of Thrift TRowResult "struct" based on an array of Hbase RowResult objects.static List<org.apache.hadoop.hbase.thrift.generated.TRowResult>
ThriftUtilities.rowResultFromHBase
(Result[] in, boolean sortColumns) This utility method creates a list of Thrift TRowResult "struct" based on an Hbase RowResult object. -
Uses of Result in org.apache.hadoop.hbase.thrift2
Modifier and TypeFieldDescriptionprivate static final Result
ThriftUtilities.EMPTY_RESULT
private static final Result
ThriftUtilities.EMPTY_RESULT_STALE
Modifier and TypeMethodDescriptionstatic Result
ThriftUtilities.resultFromThrift
(org.apache.hadoop.hbase.thrift2.generated.TResult in) static Result[]
ThriftUtilities.resultsFromThrift
(List<org.apache.hadoop.hbase.thrift2.generated.TResult> in) Modifier and TypeMethodDescriptionstatic org.apache.hadoop.hbase.thrift2.generated.TResult
ThriftUtilities.resultFromHBase
(Result in) Creates aTResult
(Thrift) from aResult
(HBase).static List<org.apache.hadoop.hbase.thrift2.generated.TResult>
ThriftUtilities.resultsFromHBase
(Result[] in) Converts multipleResult
s (HBase) into a list ofTResult
s (Thrift). -
Uses of Result in org.apache.hadoop.hbase.thrift2.client
-
Uses of Result in org.apache.hadoop.hbase.util