Skip to content

Commit

Permalink
*: fix some typos in code comments (pingcap#53371)
Browse files Browse the repository at this point in the history
  • Loading branch information
lilin90 authored and terry1purcell committed May 20, 2024
1 parent 1fc2825 commit 5293d2a
Show file tree
Hide file tree
Showing 43 changed files with 68 additions and 68 deletions.
2 changes: 1 addition & 1 deletion pkg/disttask/framework/storage/task_table.go
Original file line number Diff line number Diff line change
Expand Up @@ -802,7 +802,7 @@ func (mgr *TaskManager) GetAllSubtasks(ctx context.Context) ([]*proto.SubtaskBas
// a stuck issue if the new version TiDB has less than 16 CPU count.
// We don't adjust the concurrency in subtask table because this field does not exist in v7.5.0.
// For details, see https://github.com/pingcap/tidb/issues/50894.
// For the following versions, there is a check when submiting a new task. This function should be a no-op.
// For the following versions, there is a check when submitting a new task. This function should be a no-op.
func (mgr *TaskManager) AdjustTaskOverflowConcurrency(ctx context.Context, se sessionctx.Context) error {
cpuCount, err := mgr.getCPUCountOfNode(ctx, se)
if err != nil {
Expand Down
12 changes: 6 additions & 6 deletions pkg/domain/domain.go
Original file line number Diff line number Diff line change
Expand Up @@ -277,7 +277,7 @@ func (do *Domain) loadInfoSchema(startTS uint64) (infoschema.InfoSchema, bool, i
// 1. Not first time bootstrap loading, which needs a full load.
// 2. It is newer than the current one, so it will be "the current one" after this function call.
// 3. There are less 100 diffs.
// 4. No regenrated schema diff.
// 4. No regenerated schema diff.
startTime := time.Now()
if currentSchemaVersion != 0 && neededSchemaVersion > currentSchemaVersion && neededSchemaVersion-currentSchemaVersion < LoadSchemaDiffVersionGapThreshold {
is, relatedChanges, diffTypes, err := do.tryLoadSchemaDiffs(m, currentSchemaVersion, neededSchemaVersion, startTS)
Expand Down Expand Up @@ -511,7 +511,7 @@ func (do *Domain) InfoSchema() infoschema.InfoSchema {

// GetSnapshotInfoSchema gets a snapshot information schema.
func (do *Domain) GetSnapshotInfoSchema(snapshotTS uint64) (infoschema.InfoSchema, error) {
// if the snapshotTS is new enough, we can get infoschema directly through sanpshotTS.
// if the snapshotTS is new enough, we can get infoschema directly through snapshotTS.
if is := do.infoCache.GetBySnapshotTS(snapshotTS); is != nil {
return is, nil
}
Expand Down Expand Up @@ -613,7 +613,7 @@ func (do *Domain) Reload() error {
is, hitCache, oldSchemaVersion, changes, err := do.loadInfoSchema(version)
if err != nil {
if version = getFlashbackStartTSFromErrorMsg(err); version != 0 {
// use the lastest available version to create domain
// use the latest available version to create domain
version--
is, hitCache, oldSchemaVersion, changes, err = do.loadInfoSchema(version)
}
Expand Down Expand Up @@ -949,7 +949,7 @@ func (do *Domain) loadSchemaInLoop(ctx context.Context, lease time.Duration) {
}
// The schema maybe changed, must reload schema then the schema validator can restart.
exitLoop := do.mustReload()
// domain is cosed.
// domain is closed.
if exitLoop {
logutil.BgLogger().Error("domain is closed, exit loadSchemaInLoop")
return
Expand All @@ -968,7 +968,7 @@ func (do *Domain) loadSchemaInLoop(ctx context.Context, lease time.Duration) {
}

// mustRestartSyncer tries to restart the SchemaSyncer.
// It returns until it's successful or the domain is stoped.
// It returns until it's successful or the domain is stopped.
func (do *Domain) mustRestartSyncer(ctx context.Context) error {
syncer := do.ddl.SchemaSyncer()

Expand Down Expand Up @@ -2098,7 +2098,7 @@ func (do *Domain) GetHistoricalStatsWorker() *HistoricalStatsWorker {
return do.historicalStatsWorker
}

// EnableDumpHistoricalStats used to control whether enbale dump stats for unit test
// EnableDumpHistoricalStats used to control whether enable dump stats for unit test
var enableDumpHistoricalStats atomic.Bool

// StartHistoricalStatsWorker start historical workers running
Expand Down
2 changes: 1 addition & 1 deletion pkg/domain/plan_replayer.go
Original file line number Diff line number Diff line change
Expand Up @@ -501,7 +501,7 @@ func (h *planReplayerTaskDumpHandle) GetWorker() *planReplayerTaskDumpWorker {
return h.workers[0]
}

// Close make finished flag ture
// Close make finished flag true
func (h *planReplayerTaskDumpHandle) Close() {
close(h.taskCH)
}
Expand Down
2 changes: 1 addition & 1 deletion pkg/domain/resourcegroup/runaway.go
Original file line number Diff line number Diff line change
Expand Up @@ -317,7 +317,7 @@ func (rm *RunawayManager) addWatchList(record *QuarantineRecord, ttl time.Durati
rm.queryLock.Lock()
defer rm.queryLock.Unlock()
if item != nil {
// check the ID because of the eariler scan.
// check the ID because of the earlier scan.
if item.ID == record.ID {
return
}
Expand Down
2 changes: 1 addition & 1 deletion pkg/domain/ru_stats.go
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ import (
const (
maxRetryCount int = 10
ruStatsInterval time.Duration = 24 * time.Hour
// only keep stats rows for last 3 monthes(92 days at most).
// only keep stats rows for last 3 months(92 days at most).
ruStatsGCDuration time.Duration = 92 * ruStatsInterval
gcBatchSize int64 = 1000
)
Expand Down
2 changes: 1 addition & 1 deletion pkg/domain/runaway.go
Original file line number Diff line number Diff line change
Expand Up @@ -240,7 +240,7 @@ func (do *Domain) RemoveRunawayWatch(recordID int64) error {
func (do *Domain) runawayRecordFlushLoop() {
defer util.Recover(metrics.LabelDomain, "runawayRecordFlushLoop", nil, false)

// this times is used to batch flushing rocords, with 1s duration,
// this times is used to batch flushing records, with 1s duration,
// we can guarantee a watch record can be seen by the user within 1s.
runawayRecordFluashTimer := time.NewTimer(runawayRecordFlushInterval)
runawayRecordGCTicker := time.NewTicker(runawayRecordGCInterval)
Expand Down
2 changes: 1 addition & 1 deletion pkg/domain/schema_validator.go
Original file line number Diff line number Diff line change
Expand Up @@ -112,7 +112,7 @@ func (s *schemaValidator) Restart() {
defer s.mux.Unlock()
s.isStarted = true
if s.do != nil {
// When this instance reconnects PD, we should record the latest schema verion after mustReload(),
// When this instance reconnects PD, we should record the latest schema version after mustReload(),
// to prevent write txns using a stale schema version by aborting them before commit.
// However, the problem still exists for read-only txns.
s.restartSchemaVer = s.do.InfoSchema().SchemaMetaVersion()
Expand Down
2 changes: 1 addition & 1 deletion pkg/executor/admin.go
Original file line number Diff line number Diff line change
Expand Up @@ -285,7 +285,7 @@ func (e *RecoverIndexExec) buildTableScan(ctx context.Context, txn kv.Transactio
return nil, err
}

// Actually, with limitCnt, the match datas maybe only in one region, so let the concurrency to be 1,
// Actually, with limitCnt, the match data maybe only in one region, so let the concurrency to be 1,
// avoid unnecessary region scan.
kvReq.Concurrency = 1
result, err := distsql.Select(ctx, e.Ctx().GetDistSQLCtx(), kvReq, e.columnsTypes())
Expand Down
4 changes: 2 additions & 2 deletions pkg/executor/builder.go
Original file line number Diff line number Diff line change
Expand Up @@ -3682,7 +3682,7 @@ func buildTableReq(b *executorBuilder, schemaLen int, plans []base.PhysicalPlan)

// buildIndexReq is designed to create a DAG for index request.
// If len(ByItems) != 0 means index request should return related columns
// to sort result rows in TiDB side for parition tables.
// to sort result rows in TiDB side for partition tables.
func buildIndexReq(ctx sessionctx.Context, columns []*model.IndexColumn, handleLen int, plans []base.PhysicalPlan) (dagReq *tipb.DAGRequest, err error) {
indexReq, err := builder.ConstructDAGReq(ctx, plans, kv.TiKV)
if err != nil {
Expand Down Expand Up @@ -4574,7 +4574,7 @@ func buildRangesForIndexJoin(ctx sessionctx.Context, lookUpContents []*join.Inde
}
}
if cwc == nil {
// A deep copy is need here because the old []*range.Range is overwriten
// A deep copy is need here because the old []*range.Range is overwritten
for _, ran := range ranges {
retRanges = append(retRanges, ran.Clone())
}
Expand Down
4 changes: 2 additions & 2 deletions pkg/executor/distsql.go
Original file line number Diff line number Diff line change
Expand Up @@ -650,7 +650,7 @@ func (e *IndexLookUpExecutor) startIndexWorker(ctx context.Context, workCh chan<
kvRanges = e.partitionKVRanges
}
// When len(kvrange) = 1, no sorting is required,
// so remove byItems and non-necessary output colums
// so remove byItems and non-necessary output columns
if len(kvRanges) == 1 {
e.dagPB.OutputOffsets = e.dagPB.OutputOffsets[len(e.byItems):]
e.byItems = nil
Expand Down Expand Up @@ -738,7 +738,7 @@ func (e *IndexLookUpExecutor) startIndexWorker(ctx context.Context, workCh chan<
return nil
}

// startTableWorker launchs some background goroutines which pick tasks from workCh and execute the task.
// startTableWorker launches some background goroutines which pick tasks from workCh and execute the task.
func (e *IndexLookUpExecutor) startTableWorker(ctx context.Context, workCh <-chan *lookupTableTask) {
lookupConcurrencyLimit := e.Ctx().GetSessionVars().IndexLookupConcurrency()
e.tblWorkerWg.Add(lookupConcurrencyLimit)
Expand Down
2 changes: 1 addition & 1 deletion pkg/executor/infoschema_reader.go
Original file line number Diff line number Diff line change
Expand Up @@ -535,7 +535,7 @@ func (e *memtableRetriever) setDataFromReferConst(sctx sessionctx.Context, schem

func (e *memtableRetriever) updateStatsCacheIfNeed() bool {
for _, col := range e.columns {
// only the following columns need stats cahce.
// only the following columns need stats cache.
if col.Name.O == "AVG_ROW_LENGTH" || col.Name.O == "DATA_LENGTH" || col.Name.O == "INDEX_LENGTH" || col.Name.O == "TABLE_ROWS" {
return true
}
Expand Down
2 changes: 1 addition & 1 deletion pkg/executor/load_data.go
Original file line number Diff line number Diff line change
Expand Up @@ -356,7 +356,7 @@ type commitTask struct {
rows [][]types.Datum
}

// processStream always trys to build a parser from channel and process it. When
// processStream always tries to build a parser from channel and process it. When
// it returns nil, it means all data is read.
func (w *encodeWorker) processStream(
ctx context.Context,
Expand Down
2 changes: 1 addition & 1 deletion pkg/executor/slow_query.go
Original file line number Diff line number Diff line change
Expand Up @@ -928,7 +928,7 @@ func (e *slowQueryRetriever) getAllFiles(ctx context.Context, sctx sessionctx.Co
}

// If we want to get the end time from a compressed file,
// we need uncompress the whole file which is very slow and consume a lot of memeory.
// we need uncompress the whole file which is very slow and consume a lot of memory.
if !compressed {
// Get the file end time.
fileEndTime, err := e.getFileEndTime(ctx, file)
Expand Down
10 changes: 5 additions & 5 deletions pkg/expression/builtin_compare.go
Original file line number Diff line number Diff line change
Expand Up @@ -368,7 +368,7 @@ func ResolveType4Between(args [3]Expression) types.EvalType {
return cmpTp
}

// GLCmpStringMode represents Greatest/Least interal string comparison mode
// GLCmpStringMode represents Greatest/Least integral string comparison mode
type GLCmpStringMode uint8

const (
Expand Down Expand Up @@ -448,7 +448,7 @@ func (c *greatestFunctionClass) getFunction(ctx BuildContext, args []Expression)
resTp := resFieldType.EvalType()
argTp := resTp
if cmpStringMode != GLCmpStringDirectly {
// Args are temporal and string mixed, we cast all args as string and parse it to temporal mannualy to compare.
// Args are temporal and string mixed, we cast all args as string and parse it to temporal manually to compare.
argTp = types.ETString
} else if resTp == types.ETJson {
unsupportedJSONComparison(ctx, args)
Expand Down Expand Up @@ -761,7 +761,7 @@ func (c *leastFunctionClass) getFunction(ctx BuildContext, args []Expression) (s
resTp := resFieldType.EvalType()
argTp := resTp
if cmpStringMode != GLCmpStringDirectly {
// Args are temporal and string mixed, we cast all args as string and parse it to temporal mannualy to compare.
// Args are temporal and string mixed, we cast all args as string and parse it to temporal manually to compare.
argTp = types.ETString
} else if resTp == types.ETJson {
unsupportedJSONComparison(ctx, args)
Expand Down Expand Up @@ -1373,7 +1373,7 @@ func isTemporalColumn(expr Expression) bool {

// tryToConvertConstantInt tries to convert a constant with other type to a int constant.
// isExceptional indicates whether the 'int column [cmp] const' might be true/false.
// If isExceptional is true, ExecptionalVal is returned. Or, CorrectVal is returned.
// If isExceptional is true, ExceptionalVal is returned. Or, CorrectVal is returned.
// CorrectVal: The computed result. If the constant can be converted to int without exception, return the val. Else return 'con'(the input).
// ExceptionalVal : It is used to get more information to check whether 'int column [cmp] const' is true/false
//
Expand Down Expand Up @@ -1412,7 +1412,7 @@ func tryToConvertConstantInt(ctx BuildContext, targetFieldType *types.FieldType,

// RefineComparedConstant changes a non-integer constant argument to its ceiling or floor result by the given op.
// isExceptional indicates whether the 'int column [cmp] const' might be true/false.
// If isExceptional is true, ExecptionalVal is returned. Or, CorrectVal is returned.
// If isExceptional is true, ExceptionalVal is returned. Or, CorrectVal is returned.
// CorrectVal: The computed result. If the constant can be converted to int without exception, return the val. Else return 'con'(the input).
// ExceptionalVal : It is used to get more information to check whether 'int column [cmp] const' is true/false
//
Expand Down
2 changes: 1 addition & 1 deletion pkg/expression/builtin_encryption_vec.go
Original file line number Diff line number Diff line change
Expand Up @@ -694,7 +694,7 @@ func (b *builtinAesEncryptSig) vecEvalString(ctx EvalContext, input *chunk.Chunk
}

// NOTE: we can't use GetBytes, because in AESEncryptWithECB padding is automatically
// added to str and this will damange the data layout in chunk.Column
// added to str and this will damage the data layout in chunk.Column
str := []byte(strBuf.GetString(i))
cipherText, err := encrypt.AESEncryptWithECB(str, key)
if err != nil {
Expand Down
2 changes: 1 addition & 1 deletion pkg/expression/builtin_miscellaneous.go
Original file line number Diff line number Diff line change
Expand Up @@ -624,7 +624,7 @@ func (b *builtinInetNtoaSig) evalString(ctx EvalContext, row chunk.Row) (string,
binary.BigEndian.PutUint32(ip, uint32(val))
ipv4 := ip.To4()
if ipv4 == nil {
// Not a vaild ipv4 address.
// Not a valid ipv4 address.
return "", true, nil
}

Expand Down
2 changes: 1 addition & 1 deletion pkg/expression/builtin_miscellaneous_vec.go
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ func (b *builtinInetNtoaSig) vecEvalString(ctx EvalContext, input *chunk.Chunk,
binary.BigEndian.PutUint32(ip, uint32(val))
ipv4 := ip.To4()
if ipv4 == nil {
// Not a vaild ipv4 address.
// Not a valid ipv4 address.
result.AppendNull()
continue
}
Expand Down
2 changes: 1 addition & 1 deletion pkg/expression/builtin_string.go
Original file line number Diff line number Diff line change
Expand Up @@ -1751,7 +1751,7 @@ type trimFunctionClass struct {

// getFunction sets trim built-in function signature.
// The syntax of trim in mysql is 'TRIM([{BOTH | LEADING | TRAILING} [remstr] FROM] str), TRIM([remstr FROM] str)',
// but we wil convert it into trim(str), trim(str, remstr) and trim(str, remstr, direction) in AST.
// but we will convert it into trim(str), trim(str, remstr) and trim(str, remstr, direction) in AST.
func (c *trimFunctionClass) getFunction(ctx BuildContext, args []Expression) (builtinFunc, error) {
if err := c.verifyArgs(args); err != nil {
return nil, err
Expand Down
4 changes: 2 additions & 2 deletions pkg/expression/builtin_time.go
Original file line number Diff line number Diff line change
Expand Up @@ -63,10 +63,10 @@ const ( // GET_FORMAT location.
)

var (
// durationPattern checks whether a string matchs the format of duration.
// durationPattern checks whether a string matches the format of duration.
durationPattern = regexp.MustCompile(`^\s*[-]?(((\d{1,2}\s+)?0*\d{0,3}(:0*\d{1,2}){0,2})|(\d{1,7}))?(\.\d*)?\s*$`)

// timestampPattern checks whether a string matchs the format of timestamp.
// timestampPattern checks whether a string matches the format of timestamp.
timestampPattern = regexp.MustCompile(`^\s*0*\d{1,4}([^\d]0*\d{1,2}){2}\s+(0*\d{0,2}([^\d]0*\d{1,2}){2})?(\.\d*)?\s*$`)

// datePattern determine whether to match the format of date.
Expand Down
2 changes: 1 addition & 1 deletion pkg/expression/builtin_time_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -845,7 +845,7 @@ func TestNowAndUTCTimestamp(t *testing.T) {
ts := x.now()
require.NoError(t, err)
mt := v.GetMysqlTime()
// we canot use a constant value to check timestamp funcs, so here
// we cannot use a constant value to check timestamp funcs, so here
// just to check the fractional seconds part and the time delta.
require.False(t, strings.Contains(mt.String(), "."))
require.LessOrEqual(t, ts.Sub(gotime(mt, ts.Location())), 5*time.Second)
Expand Down
2 changes: 1 addition & 1 deletion pkg/expression/column.go
Original file line number Diff line number Diff line change
Expand Up @@ -276,7 +276,7 @@ func (col *Column) EqualColumn(expr Expression) bool {
return false
}

// EqualByExprAndID extends Equal by comparing virual expression
// EqualByExprAndID extends Equal by comparing virtual expression
func (col *Column) EqualByExprAndID(ctx EvalContext, expr Expression) bool {
if newCol, ok := expr.(*Column); ok {
expr, isOk := col.VirtualExpr.(*ScalarFunction)
Expand Down
2 changes: 1 addition & 1 deletion pkg/expression/expression.go
Original file line number Diff line number Diff line change
Expand Up @@ -219,7 +219,7 @@ type Expression interface {
// resolveIndices is called inside the `ResolveIndices` It will perform on the expression itself.
resolveIndices(schema *Schema) error

// ResolveIndicesByVirtualExpr resolves indices by the given schema in terms of virual expression. It will copy the original expression and return the copied one.
// ResolveIndicesByVirtualExpr resolves indices by the given schema in terms of virtual expression. It will copy the original expression and return the copied one.
ResolveIndicesByVirtualExpr(ctx EvalContext, schema *Schema) (Expression, bool)

// resolveIndicesByVirtualExpr is called inside the `ResolveIndicesByVirtualExpr` It will perform on the expression itself.
Expand Down
6 changes: 3 additions & 3 deletions pkg/expression/util.go
Original file line number Diff line number Diff line change
Expand Up @@ -997,7 +997,7 @@ func Contains(exprs []Expression, e Expression) bool {

// ExtractFiltersFromDNFs checks whether the cond is DNF. If so, it will get the extracted part and the remained part.
// The original DNF will be replaced by the remained part or just be deleted if remained part is nil.
// And the extracted part will be appended to the end of the orignal slice.
// And the extracted part will be appended to the end of the original slice.
func ExtractFiltersFromDNFs(ctx BuildContext, conditions []Expression) []Expression {
var allExtracted []Expression
for i := len(conditions) - 1; i >= 0; i-- {
Expand Down Expand Up @@ -1249,7 +1249,7 @@ func GetStringFromConstant(ctx EvalContext, value Expression) (string, bool, err
return str, false, nil
}

// GetIntFromConstant gets an interger value from the Constant expression.
// GetIntFromConstant gets an integer value from the Constant expression.
func GetIntFromConstant(ctx EvalContext, value Expression) (int, bool, error) {
str, isNull, err := GetStringFromConstant(ctx, value)
if err != nil || isNull {
Expand Down Expand Up @@ -1494,7 +1494,7 @@ func RemoveMutableConst(ctx BuildContext, exprs []Expression) (err error) {
case *Constant:
v.ParamMarker = nil
if v.DeferredExpr != nil { // evaluate and update v.Value to convert v to a complete immutable constant.
// TODO: remove or hide DefferedExpr since it's too dangerous (hard to be consistent with v.Value all the time).
// TODO: remove or hide DeferredExpr since it's too dangerous (hard to be consistent with v.Value all the time).
v.Value, err = v.DeferredExpr.Eval(ctx.GetEvalCtx(), chunk.Row{})
if err != nil {
return err
Expand Down
2 changes: 1 addition & 1 deletion pkg/session/bootstrap_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -781,7 +781,7 @@ func TestIndexMergeUpgradeFrom300To540(t *testing.T) {
require.NoError(t, err)
require.Equal(t, int64(ver300), ver)

// We are now in 3.0.0, check tidb_enable_index_merge shoudle not exist.
// We are now in 3.0.0, check tidb_enable_index_merge should not exist.
res := MustExecToRecodeSet(t, seV3, fmt.Sprintf("select * from mysql.GLOBAL_VARIABLES where variable_name='%s'", variable.TiDBEnableIndexMerge))
chk := res.NewChunk(nil)
err = res.Next(ctx, chk)
Expand Down
2 changes: 1 addition & 1 deletion pkg/session/bootstraptest/bootstrap_upgrade_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -474,7 +474,7 @@ func checkDDLJobExecSucc(t *testing.T, se sessiontypes.Session, jobID int64) {
// TestUpgradeVersionForSystemPausedJob tests mock the first upgrade failed, and it has a mock system DDL in queue.
// Then we do re-upgrade(This operation will pause all DDL jobs by the system).
func TestUpgradeVersionForSystemPausedJob(t *testing.T) {
// Mock a general and a reorg job in boostrap.
// Mock a general and a reorg job in bootstrap.
mock := true
session.WithMockUpgrade = &mock
session.MockUpgradeToVerLatestKind = session.MockSimpleUpgradeToVerLatest
Expand Down
4 changes: 2 additions & 2 deletions pkg/session/session.go
Original file line number Diff line number Diff line change
Expand Up @@ -747,7 +747,7 @@ func (s *session) commitTxnWithTemporaryData(ctx context.Context, txn kv.Transac
return nil
}

// errIsNoisy is used to filter DUPLCATE KEY errors.
// errIsNoisy is used to filter DUPLICATE KEY errors.
// These can observed by users in INFORMATION_SCHEMA.CLIENT_ERRORS_SUMMARY_GLOBAL instead.
//
// The rationale for filtering these errors is because they are "client generated errors". i.e.
Expand Down Expand Up @@ -2340,7 +2340,7 @@ const ExecStmtVarKey ExecStmtVarKeyType = 0
// execStmtResult is the return value of ExecuteStmt and it implements the sqlexec.RecordSet interface.
// Why we need a struct to wrap a RecordSet and provide another RecordSet?
// This is because there are so many session state related things that definitely not belongs to the original
// RecordSet, so this struct exists and RecordSet.Close() is overrided handle that.
// RecordSet, so this struct exists and RecordSet.Close() is overridden to handle that.
type execStmtResult struct {
sqlexec.RecordSet
se *session
Expand Down
2 changes: 1 addition & 1 deletion pkg/session/test/session_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -287,7 +287,7 @@ func TestParseWithParams(t *testing.T) {
se := tk.Session()
exec := se.GetRestrictedSQLExecutor()

// test compatibility with ExcuteInternal
// test compatibility with ExecuteInternal
_, err := exec.ParseWithParams(context.TODO(), "SELECT 4")
require.NoError(t, err)

Expand Down
Loading

0 comments on commit 5293d2a

Please sign in to comment.