tidb admin 源码
tidb admin 代码
文件路径:/executor/admin.go
// Copyright 2018 PingCAP, Inc.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package executor
import (
"context"
"math"
"github.com/pingcap/errors"
"github.com/pingcap/kvproto/pkg/kvrpcpb"
"github.com/pingcap/tidb/distsql"
"github.com/pingcap/tidb/infoschema"
"github.com/pingcap/tidb/kv"
"github.com/pingcap/tidb/parser/ast"
"github.com/pingcap/tidb/parser/model"
"github.com/pingcap/tidb/parser/terror"
plannercore "github.com/pingcap/tidb/planner/core"
"github.com/pingcap/tidb/sessionctx/stmtctx"
"github.com/pingcap/tidb/statistics"
"github.com/pingcap/tidb/table"
"github.com/pingcap/tidb/table/tables"
"github.com/pingcap/tidb/tablecodec"
"github.com/pingcap/tidb/types"
"github.com/pingcap/tidb/util"
"github.com/pingcap/tidb/util/chunk"
"github.com/pingcap/tidb/util/logutil"
"github.com/pingcap/tidb/util/ranger"
"github.com/pingcap/tidb/util/timeutil"
"github.com/pingcap/tipb/go-tipb"
"go.uber.org/zap"
)
var (
_ Executor = &CheckIndexRangeExec{}
_ Executor = &RecoverIndexExec{}
_ Executor = &CleanupIndexExec{}
)
// CheckIndexRangeExec outputs the index values which has handle between begin and end.
type CheckIndexRangeExec struct {
baseExecutor
table *model.TableInfo
index *model.IndexInfo
is infoschema.InfoSchema
startKey []types.Datum
handleRanges []ast.HandleRange
srcChunk *chunk.Chunk
result distsql.SelectResult
cols []*model.ColumnInfo
}
// Next implements the Executor Next interface.
func (e *CheckIndexRangeExec) Next(ctx context.Context, req *chunk.Chunk) error {
req.Reset()
handleIdx := e.schema.Len() - 1
for {
err := e.result.Next(ctx, e.srcChunk)
if err != nil {
return err
}
if e.srcChunk.NumRows() == 0 {
return nil
}
iter := chunk.NewIterator4Chunk(e.srcChunk)
appendRows := make([]chunk.Row, 0, e.srcChunk.NumRows())
for row := iter.Begin(); row != iter.End(); row = iter.Next() {
handle := row.GetInt64(handleIdx)
for _, hr := range e.handleRanges {
if handle >= hr.Begin && handle < hr.End {
appendRows = append(appendRows, row)
break
}
}
}
if len(appendRows) > 0 {
req.AppendRows(appendRows)
}
if req.NumRows() > 0 {
return nil
}
}
}
// Open implements the Executor Open interface.
func (e *CheckIndexRangeExec) Open(ctx context.Context) error {
tCols := e.table.Cols()
for _, ic := range e.index.Columns {
col := tCols[ic.Offset]
e.cols = append(e.cols, col)
}
colTypeForHandle := e.schema.Columns[len(e.cols)].RetType
e.cols = append(e.cols, &model.ColumnInfo{
ID: model.ExtraHandleID,
Name: model.ExtraHandleName,
FieldType: *colTypeForHandle,
})
e.srcChunk = newFirstChunk(e)
dagPB, err := e.buildDAGPB()
if err != nil {
return err
}
sc := e.ctx.GetSessionVars().StmtCtx
txn, err := e.ctx.Txn(true)
if err != nil {
return nil
}
var builder distsql.RequestBuilder
kvReq, err := builder.SetIndexRanges(sc, e.table.ID, e.index.ID, ranger.FullRange()).
SetDAGRequest(dagPB).
SetStartTS(txn.StartTS()).
SetKeepOrder(true).
SetFromSessionVars(e.ctx.GetSessionVars()).
SetFromInfoSchema(e.ctx.GetInfoSchema()).
Build()
if err != nil {
return err
}
e.result, err = distsql.Select(ctx, e.ctx, kvReq, e.retFieldTypes, statistics.NewQueryFeedback(0, nil, 0, false))
if err != nil {
return err
}
return nil
}
func (e *CheckIndexRangeExec) buildDAGPB() (*tipb.DAGRequest, error) {
dagReq := &tipb.DAGRequest{}
dagReq.TimeZoneName, dagReq.TimeZoneOffset = timeutil.Zone(e.ctx.GetSessionVars().Location())
sc := e.ctx.GetSessionVars().StmtCtx
dagReq.Flags = sc.PushDownFlags()
for i := range e.schema.Columns {
dagReq.OutputOffsets = append(dagReq.OutputOffsets, uint32(i))
}
execPB := e.constructIndexScanPB()
dagReq.Executors = append(dagReq.Executors, execPB)
err := plannercore.SetPBColumnsDefaultValue(e.ctx, dagReq.Executors[0].IdxScan.Columns, e.cols)
if err != nil {
return nil, err
}
distsql.SetEncodeType(e.ctx, dagReq)
return dagReq, nil
}
func (e *CheckIndexRangeExec) constructIndexScanPB() *tipb.Executor {
idxExec := &tipb.IndexScan{
TableId: e.table.ID,
IndexId: e.index.ID,
Columns: util.ColumnsToProto(e.cols, e.table.PKIsHandle),
}
return &tipb.Executor{Tp: tipb.ExecType_TypeIndexScan, IdxScan: idxExec}
}
// Close implements the Executor Close interface.
func (e *CheckIndexRangeExec) Close() error {
return nil
}
// RecoverIndexExec represents a recover index executor.
// It is built from "admin recover index" statement, is used to backfill
// corrupted index.
type RecoverIndexExec struct {
baseExecutor
done bool
index table.Index
table table.Table
physicalID int64
batchSize int
columns []*model.ColumnInfo
colFieldTypes []*types.FieldType
srcChunk *chunk.Chunk
handleCols plannercore.HandleCols
// below buf is used to reduce allocations.
recoverRows []recoverRows
idxValsBufs [][]types.Datum
idxKeyBufs [][]byte
batchKeys []kv.Key
}
func (e *RecoverIndexExec) columnsTypes() []*types.FieldType {
if e.colFieldTypes != nil {
return e.colFieldTypes
}
e.colFieldTypes = make([]*types.FieldType, 0, len(e.columns))
for _, col := range e.columns {
e.colFieldTypes = append(e.colFieldTypes, &col.FieldType)
}
return e.colFieldTypes
}
// Open implements the Executor Open interface.
func (e *RecoverIndexExec) Open(ctx context.Context) error {
if err := e.baseExecutor.Open(ctx); err != nil {
return err
}
e.srcChunk = chunk.New(e.columnsTypes(), e.initCap, e.maxChunkSize)
e.batchSize = 2048
e.recoverRows = make([]recoverRows, 0, e.batchSize)
e.idxValsBufs = make([][]types.Datum, e.batchSize)
e.idxKeyBufs = make([][]byte, e.batchSize)
return nil
}
func (e *RecoverIndexExec) constructTableScanPB(tblInfo *model.TableInfo, colInfos []*model.ColumnInfo) (*tipb.Executor, error) {
tblScan := tables.BuildTableScanFromInfos(tblInfo, colInfos)
tblScan.TableId = e.physicalID
err := plannercore.SetPBColumnsDefaultValue(e.ctx, tblScan.Columns, colInfos)
return &tipb.Executor{Tp: tipb.ExecType_TypeTableScan, TblScan: tblScan}, err
}
func (e *RecoverIndexExec) constructLimitPB(count uint64) *tipb.Executor {
limitExec := &tipb.Limit{
Limit: count,
}
return &tipb.Executor{Tp: tipb.ExecType_TypeLimit, Limit: limitExec}
}
func (e *RecoverIndexExec) buildDAGPB(txn kv.Transaction, limitCnt uint64) (*tipb.DAGRequest, error) {
dagReq := &tipb.DAGRequest{}
dagReq.TimeZoneName, dagReq.TimeZoneOffset = timeutil.Zone(e.ctx.GetSessionVars().Location())
sc := e.ctx.GetSessionVars().StmtCtx
dagReq.Flags = sc.PushDownFlags()
for i := range e.columns {
dagReq.OutputOffsets = append(dagReq.OutputOffsets, uint32(i))
}
tblScanExec, err := e.constructTableScanPB(e.table.Meta(), e.columns)
if err != nil {
return nil, err
}
dagReq.Executors = append(dagReq.Executors, tblScanExec)
limitExec := e.constructLimitPB(limitCnt)
dagReq.Executors = append(dagReq.Executors, limitExec)
distsql.SetEncodeType(e.ctx, dagReq)
return dagReq, nil
}
func (e *RecoverIndexExec) buildTableScan(ctx context.Context, txn kv.Transaction, startHandle kv.Handle, limitCnt uint64) (distsql.SelectResult, error) {
dagPB, err := e.buildDAGPB(txn, limitCnt)
if err != nil {
return nil, err
}
var builder distsql.RequestBuilder
builder.KeyRanges, err = buildRecoverIndexKeyRanges(e.ctx.GetSessionVars().StmtCtx, e.physicalID, startHandle)
if err != nil {
return nil, err
}
kvReq, err := builder.
SetDAGRequest(dagPB).
SetStartTS(txn.StartTS()).
SetKeepOrder(true).
SetFromSessionVars(e.ctx.GetSessionVars()).
SetFromInfoSchema(e.ctx.GetInfoSchema()).
Build()
if err != nil {
return nil, err
}
// Actually, with limitCnt, the match datas maybe only in one region, so let the concurrency to be 1,
// avoid unnecessary region scan.
kvReq.Concurrency = 1
result, err := distsql.Select(ctx, e.ctx, kvReq, e.columnsTypes(), statistics.NewQueryFeedback(0, nil, 0, false))
if err != nil {
return nil, err
}
return result, nil
}
// buildRecoverIndexKeyRanges build a KeyRange: (startHandle, unlimited).
func buildRecoverIndexKeyRanges(sctx *stmtctx.StatementContext, tid int64, startHandle kv.Handle) ([]kv.KeyRange, error) {
var startKey []byte
if startHandle == nil {
startKey = tablecodec.GenTableRecordPrefix(tid).Next()
} else {
startKey = tablecodec.EncodeRowKey(tid, startHandle.Encoded()).PrefixNext()
}
endKey := tablecodec.GenTableRecordPrefix(tid).PrefixNext()
return []kv.KeyRange{{StartKey: startKey, EndKey: endKey}}, nil
}
type backfillResult struct {
currentHandle kv.Handle
addedCount int64
scanRowCount int64
}
func (e *RecoverIndexExec) backfillIndex(ctx context.Context) (int64, int64, error) {
var (
currentHandle kv.Handle = nil
totalAddedCnt = int64(0)
totalScanCnt = int64(0)
lastLogCnt = int64(0)
result backfillResult
)
for {
ctx = kv.WithInternalSourceType(ctx, kv.InternalTxnAdmin)
errInTxn := kv.RunInNewTxn(ctx, e.ctx.GetStore(), true, func(ctx context.Context, txn kv.Transaction) error {
setOptionForTopSQL(e.ctx.GetSessionVars().StmtCtx, txn)
var err error
result, err = e.backfillIndexInTxn(ctx, txn, currentHandle)
return err
})
if errInTxn != nil {
return totalAddedCnt, totalScanCnt, errInTxn
}
totalAddedCnt += result.addedCount
totalScanCnt += result.scanRowCount
if totalScanCnt-lastLogCnt >= 50000 {
lastLogCnt = totalScanCnt
logutil.Logger(ctx).Info("recover index", zap.String("table", e.table.Meta().Name.O),
zap.String("index", e.index.Meta().Name.O), zap.Int64("totalAddedCnt", totalAddedCnt),
zap.Int64("totalScanCnt", totalScanCnt), zap.Stringer("currentHandle", result.currentHandle))
}
// no more rows
if result.scanRowCount == 0 {
break
}
currentHandle = result.currentHandle
if currentHandle.Next().Compare(result.currentHandle) <= 0 {
break // There is no more handles in the table.
}
}
return totalAddedCnt, totalScanCnt, nil
}
type recoverRows struct {
handle kv.Handle
idxVals []types.Datum
rsData []types.Datum
skip bool
}
func (e *RecoverIndexExec) fetchRecoverRows(ctx context.Context, srcResult distsql.SelectResult, result *backfillResult) ([]recoverRows, error) {
e.recoverRows = e.recoverRows[:0]
idxValLen := len(e.index.Meta().Columns)
result.scanRowCount = 0
for {
err := srcResult.Next(ctx, e.srcChunk)
if err != nil {
return nil, err
}
if e.srcChunk.NumRows() == 0 {
break
}
iter := chunk.NewIterator4Chunk(e.srcChunk)
for row := iter.Begin(); row != iter.End(); row = iter.Next() {
if result.scanRowCount >= int64(e.batchSize) {
return e.recoverRows, nil
}
handle, err := e.handleCols.BuildHandle(row)
if err != nil {
return nil, err
}
idxVals := extractIdxVals(row, e.idxValsBufs[result.scanRowCount], e.colFieldTypes, idxValLen)
e.idxValsBufs[result.scanRowCount] = idxVals
rsData := tables.TryGetHandleRestoredDataWrapper(e.table, plannercore.GetCommonHandleDatum(e.handleCols, row), nil, e.index.Meta())
e.recoverRows = append(e.recoverRows, recoverRows{handle: handle, idxVals: idxVals, rsData: rsData, skip: false})
result.scanRowCount++
result.currentHandle = handle
}
}
return e.recoverRows, nil
}
func (e *RecoverIndexExec) batchMarkDup(txn kv.Transaction, rows []recoverRows) error {
if len(rows) == 0 {
return nil
}
e.batchKeys = e.batchKeys[:0]
sc := e.ctx.GetSessionVars().StmtCtx
distinctFlags := make([]bool, len(rows))
for i, row := range rows {
idxKey, distinct, err := e.index.GenIndexKey(sc, row.idxVals, row.handle, e.idxKeyBufs[i])
if err != nil {
return err
}
e.idxKeyBufs[i] = idxKey
e.batchKeys = append(e.batchKeys, idxKey)
distinctFlags[i] = distinct
}
values, err := txn.BatchGet(context.Background(), e.batchKeys)
if err != nil {
return err
}
// 1. unique-key is duplicate and the handle is equal, skip it.
// 2. unique-key is duplicate and the handle is not equal, data is not consistent, log it and skip it.
// 3. non-unique-key is duplicate, skip it.
isCommonHandle := e.table.Meta().IsCommonHandle
for i, key := range e.batchKeys {
if val, found := values[string(key)]; found {
if distinctFlags[i] {
handle, err1 := tablecodec.DecodeHandleInUniqueIndexValue(val, isCommonHandle)
if err1 != nil {
return err1
}
if handle.Compare(rows[i].handle) != 0 {
logutil.BgLogger().Warn("recover index: the constraint of unique index is broken, handle in index is not equal to handle in table",
zap.String("index", e.index.Meta().Name.O), zap.ByteString("indexKey", key),
zap.Stringer("handleInTable", rows[i].handle), zap.Stringer("handleInIndex", handle))
}
}
rows[i].skip = true
}
}
return nil
}
func (e *RecoverIndexExec) backfillIndexInTxn(ctx context.Context, txn kv.Transaction, currentHandle kv.Handle) (result backfillResult, err error) {
srcResult, err := e.buildTableScan(ctx, txn, currentHandle, uint64(e.batchSize))
if err != nil {
return result, err
}
defer terror.Call(srcResult.Close)
rows, err := e.fetchRecoverRows(ctx, srcResult, &result)
if err != nil {
return result, err
}
err = e.batchMarkDup(txn, rows)
if err != nil {
return result, err
}
// Constrains is already checked.
e.ctx.GetSessionVars().StmtCtx.BatchCheck = true
for _, row := range rows {
if row.skip {
continue
}
recordKey := tablecodec.EncodeRecordKey(e.table.RecordPrefix(), row.handle)
err := txn.LockKeys(ctx, new(kv.LockCtx), recordKey)
if err != nil {
return result, err
}
_, err = e.index.Create(e.ctx, txn, row.idxVals, row.handle, row.rsData, table.WithIgnoreAssertion)
if err != nil {
return result, err
}
result.addedCount++
}
return result, nil
}
// Next implements the Executor Next interface.
func (e *RecoverIndexExec) Next(ctx context.Context, req *chunk.Chunk) error {
req.Reset()
if e.done {
return nil
}
recoveringClusteredIndex := e.index.Meta().Primary && e.table.Meta().IsCommonHandle
if recoveringClusteredIndex {
req.AppendInt64(0, 0)
req.AppendInt64(1, 0)
e.done = true
return nil
}
var totalAddedCnt, totalScanCnt int64
var err error
if tbl, ok := e.table.(table.PartitionedTable); ok {
pi := e.table.Meta().GetPartitionInfo()
for _, p := range pi.Definitions {
e.table = tbl.GetPartition(p.ID)
e.index = tables.GetWritableIndexByName(e.index.Meta().Name.L, e.table)
e.physicalID = p.ID
addedCnt, scanCnt, err := e.backfillIndex(ctx)
totalAddedCnt += addedCnt
totalScanCnt += scanCnt
if err != nil {
return err
}
}
} else {
totalAddedCnt, totalScanCnt, err = e.backfillIndex(ctx)
if err != nil {
return err
}
}
req.AppendInt64(0, totalAddedCnt)
req.AppendInt64(1, totalScanCnt)
e.done = true
return nil
}
// CleanupIndexExec represents a cleanup index executor.
// It is built from "admin cleanup index" statement, is used to delete
// dangling index data.
type CleanupIndexExec struct {
baseExecutor
done bool
removeCnt uint64
index table.Index
table table.Table
physicalID int64
columns []*model.ColumnInfo
idxColFieldTypes []*types.FieldType
idxChunk *chunk.Chunk
handleCols plannercore.HandleCols
idxValues *kv.HandleMap // kv.Handle -> [][]types.Datum
batchSize uint64
batchKeys []kv.Key
idxValsBufs [][]types.Datum
lastIdxKey []byte
scanRowCnt uint64
}
func (e *CleanupIndexExec) getIdxColTypes() []*types.FieldType {
if e.idxColFieldTypes != nil {
return e.idxColFieldTypes
}
e.idxColFieldTypes = make([]*types.FieldType, 0, len(e.columns))
for _, col := range e.columns {
e.idxColFieldTypes = append(e.idxColFieldTypes, &col.FieldType)
}
return e.idxColFieldTypes
}
func (e *CleanupIndexExec) batchGetRecord(txn kv.Transaction) (map[string][]byte, error) {
e.idxValues.Range(func(h kv.Handle, _ interface{}) bool {
e.batchKeys = append(e.batchKeys, tablecodec.EncodeRecordKey(e.table.RecordPrefix(), h))
return true
})
values, err := txn.BatchGet(context.Background(), e.batchKeys)
if err != nil {
return nil, err
}
return values, nil
}
func (e *CleanupIndexExec) deleteDanglingIdx(txn kv.Transaction, values map[string][]byte) error {
for _, k := range e.batchKeys {
if _, found := values[string(k)]; !found {
_, handle, err := tablecodec.DecodeRecordKey(k)
if err != nil {
return err
}
handleIdxValsGroup, ok := e.idxValues.Get(handle)
if !ok {
return errors.Trace(errors.Errorf("batch keys are inconsistent with handles"))
}
for _, handleIdxVals := range handleIdxValsGroup.([][]types.Datum) {
if err := e.index.Delete(e.ctx.GetSessionVars().StmtCtx, txn, handleIdxVals, handle); err != nil {
return err
}
e.removeCnt++
if e.removeCnt%e.batchSize == 0 {
logutil.BgLogger().Info("clean up dangling index", zap.String("table", e.table.Meta().Name.String()),
zap.String("index", e.index.Meta().Name.String()), zap.Uint64("count", e.removeCnt))
}
}
}
}
return nil
}
func extractIdxVals(row chunk.Row, idxVals []types.Datum,
fieldTypes []*types.FieldType, idxValLen int) []types.Datum {
if cap(idxVals) < idxValLen {
idxVals = make([]types.Datum, idxValLen)
} else {
idxVals = idxVals[:idxValLen]
}
for i := 0; i < idxValLen; i++ {
colVal := row.GetDatum(i, fieldTypes[i])
colVal.Copy(&idxVals[i])
}
return idxVals
}
func (e *CleanupIndexExec) fetchIndex(ctx context.Context, txn kv.Transaction) error {
result, err := e.buildIndexScan(ctx, txn)
if err != nil {
return err
}
defer terror.Call(result.Close)
sc := e.ctx.GetSessionVars().StmtCtx
idxColLen := len(e.index.Meta().Columns)
for {
err := result.Next(ctx, e.idxChunk)
if err != nil {
return err
}
if e.idxChunk.NumRows() == 0 {
return nil
}
iter := chunk.NewIterator4Chunk(e.idxChunk)
for row := iter.Begin(); row != iter.End(); row = iter.Next() {
handle, err := e.handleCols.BuildHandle(row)
if err != nil {
return err
}
idxVals := extractIdxVals(row, e.idxValsBufs[e.scanRowCnt], e.idxColFieldTypes, idxColLen)
e.idxValsBufs[e.scanRowCnt] = idxVals
existingIdxVals, ok := e.idxValues.Get(handle)
if ok {
updatedIdxVals := append(existingIdxVals.([][]types.Datum), idxVals)
e.idxValues.Set(handle, updatedIdxVals)
} else {
e.idxValues.Set(handle, [][]types.Datum{idxVals})
}
idxKey, _, err := e.index.GenIndexKey(sc, idxVals, handle, nil)
if err != nil {
return err
}
e.scanRowCnt++
e.lastIdxKey = idxKey
if e.scanRowCnt >= e.batchSize {
return nil
}
}
}
}
// Next implements the Executor Next interface.
func (e *CleanupIndexExec) Next(ctx context.Context, req *chunk.Chunk) error {
req.Reset()
if e.done {
return nil
}
cleaningClusteredPrimaryKey := e.table.Meta().IsCommonHandle && e.index.Meta().Primary
if cleaningClusteredPrimaryKey {
e.done = true
req.AppendUint64(0, 0)
return nil
}
var err error
if tbl, ok := e.table.(table.PartitionedTable); ok {
pi := e.table.Meta().GetPartitionInfo()
for _, p := range pi.Definitions {
e.table = tbl.GetPartition(p.ID)
e.index = tables.GetWritableIndexByName(e.index.Meta().Name.L, e.table)
e.physicalID = p.ID
err = e.init()
if err != nil {
return err
}
err = e.cleanTableIndex(ctx)
if err != nil {
return err
}
}
} else {
err = e.cleanTableIndex(ctx)
if err != nil {
return err
}
}
e.done = true
req.AppendUint64(0, e.removeCnt)
return nil
}
func (e *CleanupIndexExec) cleanTableIndex(ctx context.Context) error {
for {
ctx = kv.WithInternalSourceType(ctx, kv.InternalTxnAdmin)
errInTxn := kv.RunInNewTxn(ctx, e.ctx.GetStore(), true, func(ctx context.Context, txn kv.Transaction) error {
txn.SetDiskFullOpt(kvrpcpb.DiskFullOpt_AllowedOnAlmostFull)
setOptionForTopSQL(e.ctx.GetSessionVars().StmtCtx, txn)
err := e.fetchIndex(ctx, txn)
if err != nil {
return err
}
values, err := e.batchGetRecord(txn)
if err != nil {
return err
}
err = e.deleteDanglingIdx(txn, values)
if err != nil {
return err
}
return nil
})
if errInTxn != nil {
return errInTxn
}
if e.scanRowCnt == 0 {
break
}
e.scanRowCnt = 0
e.batchKeys = e.batchKeys[:0]
e.idxValues.Range(func(h kv.Handle, val interface{}) bool {
e.idxValues.Delete(h)
return true
})
}
return nil
}
func (e *CleanupIndexExec) buildIndexScan(ctx context.Context, txn kv.Transaction) (distsql.SelectResult, error) {
dagPB, err := e.buildIdxDAGPB(txn)
if err != nil {
return nil, err
}
sc := e.ctx.GetSessionVars().StmtCtx
var builder distsql.RequestBuilder
ranges := ranger.FullRange()
kvReq, err := builder.SetIndexRanges(sc, e.physicalID, e.index.Meta().ID, ranges).
SetDAGRequest(dagPB).
SetStartTS(txn.StartTS()).
SetKeepOrder(true).
SetFromSessionVars(e.ctx.GetSessionVars()).
SetFromInfoSchema(e.ctx.GetInfoSchema()).
Build()
if err != nil {
return nil, err
}
kvReq.KeyRanges[0].StartKey = kv.Key(e.lastIdxKey).PrefixNext()
kvReq.Concurrency = 1
result, err := distsql.Select(ctx, e.ctx, kvReq, e.getIdxColTypes(), statistics.NewQueryFeedback(0, nil, 0, false))
if err != nil {
return nil, err
}
return result, nil
}
// Open implements the Executor Open interface.
func (e *CleanupIndexExec) Open(ctx context.Context) error {
if err := e.baseExecutor.Open(ctx); err != nil {
return err
}
return e.init()
}
func (e *CleanupIndexExec) init() error {
e.idxChunk = chunk.New(e.getIdxColTypes(), e.initCap, e.maxChunkSize)
e.idxValues = kv.NewHandleMap()
e.batchKeys = make([]kv.Key, 0, e.batchSize)
e.idxValsBufs = make([][]types.Datum, e.batchSize)
sc := e.ctx.GetSessionVars().StmtCtx
idxKey, _, err := e.index.GenIndexKey(sc, []types.Datum{{}}, kv.IntHandle(math.MinInt64), nil)
if err != nil {
return err
}
e.lastIdxKey = idxKey
return nil
}
func (e *CleanupIndexExec) buildIdxDAGPB(txn kv.Transaction) (*tipb.DAGRequest, error) {
dagReq := &tipb.DAGRequest{}
dagReq.TimeZoneName, dagReq.TimeZoneOffset = timeutil.Zone(e.ctx.GetSessionVars().Location())
sc := e.ctx.GetSessionVars().StmtCtx
dagReq.Flags = sc.PushDownFlags()
for i := range e.columns {
dagReq.OutputOffsets = append(dagReq.OutputOffsets, uint32(i))
}
execPB := e.constructIndexScanPB()
dagReq.Executors = append(dagReq.Executors, execPB)
err := plannercore.SetPBColumnsDefaultValue(e.ctx, dagReq.Executors[0].IdxScan.Columns, e.columns)
if err != nil {
return nil, err
}
limitExec := e.constructLimitPB()
dagReq.Executors = append(dagReq.Executors, limitExec)
distsql.SetEncodeType(e.ctx, dagReq)
return dagReq, nil
}
func (e *CleanupIndexExec) constructIndexScanPB() *tipb.Executor {
idxExec := &tipb.IndexScan{
TableId: e.physicalID,
IndexId: e.index.Meta().ID,
Columns: util.ColumnsToProto(e.columns, e.table.Meta().PKIsHandle),
PrimaryColumnIds: tables.TryGetCommonPkColumnIds(e.table.Meta()),
}
return &tipb.Executor{Tp: tipb.ExecType_TypeIndexScan, IdxScan: idxExec}
}
func (e *CleanupIndexExec) constructLimitPB() *tipb.Executor {
limitExec := &tipb.Limit{
Limit: e.batchSize,
}
return &tipb.Executor{Tp: tipb.ExecType_TypeLimit, Limit: limitExec}
}
// Close implements the Executor Close interface.
func (e *CleanupIndexExec) Close() error {
return nil
}
相关信息
相关文章
0
赞
热门推荐
-
2、 - 优质文章
-
3、 gate.io
-
8、 golang
-
9、 openharmony
-
10、 Vue中input框自动聚焦