Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

*: fix some typos in code comments #53371

Merged
merged 1 commit into from
May 20, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion pkg/disttask/framework/storage/task_table.go
Original file line number Diff line number Diff line change
Expand Up @@ -802,7 +802,7 @@ func (mgr *TaskManager) GetAllSubtasks(ctx context.Context) ([]*proto.SubtaskBas
// a stuck issue if the new version TiDB has less than 16 CPU count.
// We don't adjust the concurrency in subtask table because this field does not exist in v7.5.0.
// For details, see https://github.com/pingcap/tidb/issues/50894.
// For the following versions, there is a check when submiting a new task. This function should be a no-op.
// For the following versions, there is a check when submitting a new task. This function should be a no-op.
func (mgr *TaskManager) AdjustTaskOverflowConcurrency(ctx context.Context, se sessionctx.Context) error {
cpuCount, err := mgr.getCPUCountOfNode(ctx, se)
if err != nil {
Expand Down
12 changes: 6 additions & 6 deletions pkg/domain/domain.go
Original file line number Diff line number Diff line change
Expand Up @@ -277,7 +277,7 @@ func (do *Domain) loadInfoSchema(startTS uint64) (infoschema.InfoSchema, bool, i
// 1. Not first time bootstrap loading, which needs a full load.
// 2. It is newer than the current one, so it will be "the current one" after this function call.
// 3. There are less 100 diffs.
// 4. No regenrated schema diff.
// 4. No regenerated schema diff.
startTime := time.Now()
if currentSchemaVersion != 0 && neededSchemaVersion > currentSchemaVersion && neededSchemaVersion-currentSchemaVersion < LoadSchemaDiffVersionGapThreshold {
is, relatedChanges, diffTypes, err := do.tryLoadSchemaDiffs(m, currentSchemaVersion, neededSchemaVersion, startTS)
Expand Down Expand Up @@ -511,7 +511,7 @@ func (do *Domain) InfoSchema() infoschema.InfoSchema {

// GetSnapshotInfoSchema gets a snapshot information schema.
func (do *Domain) GetSnapshotInfoSchema(snapshotTS uint64) (infoschema.InfoSchema, error) {
// if the snapshotTS is new enough, we can get infoschema directly through sanpshotTS.
// if the snapshotTS is new enough, we can get infoschema directly through snapshotTS.
if is := do.infoCache.GetBySnapshotTS(snapshotTS); is != nil {
return is, nil
}
Expand Down Expand Up @@ -613,7 +613,7 @@ func (do *Domain) Reload() error {
is, hitCache, oldSchemaVersion, changes, err := do.loadInfoSchema(version)
if err != nil {
if version = getFlashbackStartTSFromErrorMsg(err); version != 0 {
// use the lastest available version to create domain
// use the latest available version to create domain
version--
is, hitCache, oldSchemaVersion, changes, err = do.loadInfoSchema(version)
}
Expand Down Expand Up @@ -949,7 +949,7 @@ func (do *Domain) loadSchemaInLoop(ctx context.Context, lease time.Duration) {
}
// The schema maybe changed, must reload schema then the schema validator can restart.
exitLoop := do.mustReload()
// domain is cosed.
// domain is closed.
if exitLoop {
logutil.BgLogger().Error("domain is closed, exit loadSchemaInLoop")
return
Expand All @@ -968,7 +968,7 @@ func (do *Domain) loadSchemaInLoop(ctx context.Context, lease time.Duration) {
}

// mustRestartSyncer tries to restart the SchemaSyncer.
// It returns until it's successful or the domain is stoped.
// It returns until it's successful or the domain is stopped.
func (do *Domain) mustRestartSyncer(ctx context.Context) error {
syncer := do.ddl.SchemaSyncer()

Expand Down Expand Up @@ -2098,7 +2098,7 @@ func (do *Domain) GetHistoricalStatsWorker() *HistoricalStatsWorker {
return do.historicalStatsWorker
}

// EnableDumpHistoricalStats used to control whether enbale dump stats for unit test
// EnableDumpHistoricalStats used to control whether enable dump stats for unit test
var enableDumpHistoricalStats atomic.Bool

// StartHistoricalStatsWorker start historical workers running
Expand Down
2 changes: 1 addition & 1 deletion pkg/domain/plan_replayer.go
Original file line number Diff line number Diff line change
Expand Up @@ -501,7 +501,7 @@ func (h *planReplayerTaskDumpHandle) GetWorker() *planReplayerTaskDumpWorker {
return h.workers[0]
}

// Close make finished flag ture
// Close make finished flag true
func (h *planReplayerTaskDumpHandle) Close() {
close(h.taskCH)
}
Expand Down
2 changes: 1 addition & 1 deletion pkg/domain/resourcegroup/runaway.go
Original file line number Diff line number Diff line change
Expand Up @@ -317,7 +317,7 @@ func (rm *RunawayManager) addWatchList(record *QuarantineRecord, ttl time.Durati
rm.queryLock.Lock()
defer rm.queryLock.Unlock()
if item != nil {
// check the ID because of the eariler scan.
// check the ID because of the earlier scan.
if item.ID == record.ID {
return
}
Expand Down
2 changes: 1 addition & 1 deletion pkg/domain/ru_stats.go
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ import (
const (
maxRetryCount int = 10
ruStatsInterval time.Duration = 24 * time.Hour
// only keep stats rows for last 3 monthes(92 days at most).
// only keep stats rows for last 3 months(92 days at most).
ruStatsGCDuration time.Duration = 92 * ruStatsInterval
gcBatchSize int64 = 1000
)
Expand Down
2 changes: 1 addition & 1 deletion pkg/domain/runaway.go
Original file line number Diff line number Diff line change
Expand Up @@ -240,7 +240,7 @@ func (do *Domain) RemoveRunawayWatch(recordID int64) error {
func (do *Domain) runawayRecordFlushLoop() {
defer util.Recover(metrics.LabelDomain, "runawayRecordFlushLoop", nil, false)

// this times is used to batch flushing rocords, with 1s duration,
// this times is used to batch flushing records, with 1s duration,
// we can guarantee a watch record can be seen by the user within 1s.
runawayRecordFluashTimer := time.NewTimer(runawayRecordFlushInterval)
runawayRecordGCTicker := time.NewTicker(runawayRecordGCInterval)
Expand Down
2 changes: 1 addition & 1 deletion pkg/domain/schema_validator.go
Original file line number Diff line number Diff line change
Expand Up @@ -112,7 +112,7 @@ func (s *schemaValidator) Restart() {
defer s.mux.Unlock()
s.isStarted = true
if s.do != nil {
// When this instance reconnects PD, we should record the latest schema verion after mustReload(),
// When this instance reconnects PD, we should record the latest schema version after mustReload(),
// to prevent write txns using a stale schema version by aborting them before commit.
// However, the problem still exists for read-only txns.
s.restartSchemaVer = s.do.InfoSchema().SchemaMetaVersion()
Expand Down
2 changes: 1 addition & 1 deletion pkg/executor/admin.go
Original file line number Diff line number Diff line change
Expand Up @@ -285,7 +285,7 @@ func (e *RecoverIndexExec) buildTableScan(ctx context.Context, txn kv.Transactio
return nil, err
}

// Actually, with limitCnt, the match datas maybe only in one region, so let the concurrency to be 1,
// Actually, with limitCnt, the match data maybe only in one region, so let the concurrency to be 1,
// avoid unnecessary region scan.
kvReq.Concurrency = 1
result, err := distsql.Select(ctx, e.Ctx().GetDistSQLCtx(), kvReq, e.columnsTypes())
Expand Down
4 changes: 2 additions & 2 deletions pkg/executor/builder.go
Original file line number Diff line number Diff line change
Expand Up @@ -3682,7 +3682,7 @@ func buildTableReq(b *executorBuilder, schemaLen int, plans []base.PhysicalPlan)

// buildIndexReq is designed to create a DAG for index request.
// If len(ByItems) != 0 means index request should return related columns
// to sort result rows in TiDB side for parition tables.
// to sort result rows in TiDB side for partition tables.
func buildIndexReq(ctx sessionctx.Context, columns []*model.IndexColumn, handleLen int, plans []base.PhysicalPlan) (dagReq *tipb.DAGRequest, err error) {
indexReq, err := builder.ConstructDAGReq(ctx, plans, kv.TiKV)
if err != nil {
Expand Down Expand Up @@ -4574,7 +4574,7 @@ func buildRangesForIndexJoin(ctx sessionctx.Context, lookUpContents []*join.Inde
}
}
if cwc == nil {
// A deep copy is need here because the old []*range.Range is overwriten
// A deep copy is need here because the old []*range.Range is overwritten
for _, ran := range ranges {
retRanges = append(retRanges, ran.Clone())
}
Expand Down
4 changes: 2 additions & 2 deletions pkg/executor/distsql.go
Original file line number Diff line number Diff line change
Expand Up @@ -650,7 +650,7 @@ func (e *IndexLookUpExecutor) startIndexWorker(ctx context.Context, workCh chan<
kvRanges = e.partitionKVRanges
}
// When len(kvrange) = 1, no sorting is required,
// so remove byItems and non-necessary output colums
// so remove byItems and non-necessary output columns
if len(kvRanges) == 1 {
e.dagPB.OutputOffsets = e.dagPB.OutputOffsets[len(e.byItems):]
e.byItems = nil
Expand Down Expand Up @@ -738,7 +738,7 @@ func (e *IndexLookUpExecutor) startIndexWorker(ctx context.Context, workCh chan<
return nil
}

// startTableWorker launchs some background goroutines which pick tasks from workCh and execute the task.
// startTableWorker launches some background goroutines which pick tasks from workCh and execute the task.
func (e *IndexLookUpExecutor) startTableWorker(ctx context.Context, workCh <-chan *lookupTableTask) {
lookupConcurrencyLimit := e.Ctx().GetSessionVars().IndexLookupConcurrency()
e.tblWorkerWg.Add(lookupConcurrencyLimit)
Expand Down
2 changes: 1 addition & 1 deletion pkg/executor/infoschema_reader.go
Original file line number Diff line number Diff line change
Expand Up @@ -535,7 +535,7 @@ func (e *memtableRetriever) setDataFromReferConst(sctx sessionctx.Context, schem

func (e *memtableRetriever) updateStatsCacheIfNeed() bool {
for _, col := range e.columns {
// only the following columns need stats cahce.
// only the following columns need stats cache.
if col.Name.O == "AVG_ROW_LENGTH" || col.Name.O == "DATA_LENGTH" || col.Name.O == "INDEX_LENGTH" || col.Name.O == "TABLE_ROWS" {
return true
}
Expand Down
2 changes: 1 addition & 1 deletion pkg/executor/load_data.go
Original file line number Diff line number Diff line change
Expand Up @@ -356,7 +356,7 @@ type commitTask struct {
rows [][]types.Datum
}

// processStream always trys to build a parser from channel and process it. When
// processStream always tries to build a parser from channel and process it. When
// it returns nil, it means all data is read.
func (w *encodeWorker) processStream(
ctx context.Context,
Expand Down
2 changes: 1 addition & 1 deletion pkg/executor/slow_query.go
Original file line number Diff line number Diff line change
Expand Up @@ -928,7 +928,7 @@ func (e *slowQueryRetriever) getAllFiles(ctx context.Context, sctx sessionctx.Co
}

// If we want to get the end time from a compressed file,
// we need uncompress the whole file which is very slow and consume a lot of memeory.
// we need uncompress the whole file which is very slow and consume a lot of memory.
if !compressed {
// Get the file end time.
fileEndTime, err := e.getFileEndTime(ctx, file)
Expand Down
10 changes: 5 additions & 5 deletions pkg/expression/builtin_compare.go
Original file line number Diff line number Diff line change
Expand Up @@ -368,7 +368,7 @@ func ResolveType4Between(args [3]Expression) types.EvalType {
return cmpTp
}

// GLCmpStringMode represents Greatest/Least interal string comparison mode
// GLCmpStringMode represents Greatest/Least integral string comparison mode
type GLCmpStringMode uint8

const (
Expand Down Expand Up @@ -448,7 +448,7 @@ func (c *greatestFunctionClass) getFunction(ctx BuildContext, args []Expression)
resTp := resFieldType.EvalType()
argTp := resTp
if cmpStringMode != GLCmpStringDirectly {
// Args are temporal and string mixed, we cast all args as string and parse it to temporal mannualy to compare.
// Args are temporal and string mixed, we cast all args as string and parse it to temporal manually to compare.
argTp = types.ETString
} else if resTp == types.ETJson {
unsupportedJSONComparison(ctx, args)
Expand Down Expand Up @@ -761,7 +761,7 @@ func (c *leastFunctionClass) getFunction(ctx BuildContext, args []Expression) (s
resTp := resFieldType.EvalType()
argTp := resTp
if cmpStringMode != GLCmpStringDirectly {
// Args are temporal and string mixed, we cast all args as string and parse it to temporal mannualy to compare.
// Args are temporal and string mixed, we cast all args as string and parse it to temporal manually to compare.
argTp = types.ETString
} else if resTp == types.ETJson {
unsupportedJSONComparison(ctx, args)
Expand Down Expand Up @@ -1373,7 +1373,7 @@ func isTemporalColumn(expr Expression) bool {

// tryToConvertConstantInt tries to convert a constant with other type to a int constant.
// isExceptional indicates whether the 'int column [cmp] const' might be true/false.
// If isExceptional is true, ExecptionalVal is returned. Or, CorrectVal is returned.
// If isExceptional is true, ExceptionalVal is returned. Or, CorrectVal is returned.
// CorrectVal: The computed result. If the constant can be converted to int without exception, return the val. Else return 'con'(the input).
// ExceptionalVal : It is used to get more information to check whether 'int column [cmp] const' is true/false
//
Expand Down Expand Up @@ -1412,7 +1412,7 @@ func tryToConvertConstantInt(ctx BuildContext, targetFieldType *types.FieldType,

// RefineComparedConstant changes a non-integer constant argument to its ceiling or floor result by the given op.
// isExceptional indicates whether the 'int column [cmp] const' might be true/false.
// If isExceptional is true, ExecptionalVal is returned. Or, CorrectVal is returned.
// If isExceptional is true, ExceptionalVal is returned. Or, CorrectVal is returned.
// CorrectVal: The computed result. If the constant can be converted to int without exception, return the val. Else return 'con'(the input).
// ExceptionalVal : It is used to get more information to check whether 'int column [cmp] const' is true/false
//
Expand Down
2 changes: 1 addition & 1 deletion pkg/expression/builtin_encryption_vec.go
Original file line number Diff line number Diff line change
Expand Up @@ -694,7 +694,7 @@ func (b *builtinAesEncryptSig) vecEvalString(ctx EvalContext, input *chunk.Chunk
}

// NOTE: we can't use GetBytes, because in AESEncryptWithECB padding is automatically
// added to str and this will damange the data layout in chunk.Column
// added to str and this will damage the data layout in chunk.Column
str := []byte(strBuf.GetString(i))
cipherText, err := encrypt.AESEncryptWithECB(str, key)
if err != nil {
Expand Down
2 changes: 1 addition & 1 deletion pkg/expression/builtin_miscellaneous.go
Original file line number Diff line number Diff line change
Expand Up @@ -624,7 +624,7 @@ func (b *builtinInetNtoaSig) evalString(ctx EvalContext, row chunk.Row) (string,
binary.BigEndian.PutUint32(ip, uint32(val))
ipv4 := ip.To4()
if ipv4 == nil {
// Not a vaild ipv4 address.
// Not a valid ipv4 address.
return "", true, nil
}

Expand Down
2 changes: 1 addition & 1 deletion pkg/expression/builtin_miscellaneous_vec.go
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ func (b *builtinInetNtoaSig) vecEvalString(ctx EvalContext, input *chunk.Chunk,
binary.BigEndian.PutUint32(ip, uint32(val))
ipv4 := ip.To4()
if ipv4 == nil {
// Not a vaild ipv4 address.
// Not a valid ipv4 address.
result.AppendNull()
continue
}
Expand Down
2 changes: 1 addition & 1 deletion pkg/expression/builtin_string.go
Original file line number Diff line number Diff line change
Expand Up @@ -1751,7 +1751,7 @@ type trimFunctionClass struct {

// getFunction sets trim built-in function signature.
// The syntax of trim in mysql is 'TRIM([{BOTH | LEADING | TRAILING} [remstr] FROM] str), TRIM([remstr FROM] str)',
// but we wil convert it into trim(str), trim(str, remstr) and trim(str, remstr, direction) in AST.
// but we will convert it into trim(str), trim(str, remstr) and trim(str, remstr, direction) in AST.
func (c *trimFunctionClass) getFunction(ctx BuildContext, args []Expression) (builtinFunc, error) {
if err := c.verifyArgs(args); err != nil {
return nil, err
Expand Down
4 changes: 2 additions & 2 deletions pkg/expression/builtin_time.go
Original file line number Diff line number Diff line change
Expand Up @@ -63,10 +63,10 @@ const ( // GET_FORMAT location.
)

var (
// durationPattern checks whether a string matchs the format of duration.
// durationPattern checks whether a string matches the format of duration.
durationPattern = regexp.MustCompile(`^\s*[-]?(((\d{1,2}\s+)?0*\d{0,3}(:0*\d{1,2}){0,2})|(\d{1,7}))?(\.\d*)?\s*$`)

// timestampPattern checks whether a string matchs the format of timestamp.
// timestampPattern checks whether a string matches the format of timestamp.
timestampPattern = regexp.MustCompile(`^\s*0*\d{1,4}([^\d]0*\d{1,2}){2}\s+(0*\d{0,2}([^\d]0*\d{1,2}){2})?(\.\d*)?\s*$`)

// datePattern determine whether to match the format of date.
Expand Down
2 changes: 1 addition & 1 deletion pkg/expression/builtin_time_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -845,7 +845,7 @@ func TestNowAndUTCTimestamp(t *testing.T) {
ts := x.now()
require.NoError(t, err)
mt := v.GetMysqlTime()
// we canot use a constant value to check timestamp funcs, so here
// we cannot use a constant value to check timestamp funcs, so here
// just to check the fractional seconds part and the time delta.
require.False(t, strings.Contains(mt.String(), "."))
require.LessOrEqual(t, ts.Sub(gotime(mt, ts.Location())), 5*time.Second)
Expand Down
2 changes: 1 addition & 1 deletion pkg/expression/column.go
Original file line number Diff line number Diff line change
Expand Up @@ -276,7 +276,7 @@ func (col *Column) EqualColumn(expr Expression) bool {
return false
}

// EqualByExprAndID extends Equal by comparing virual expression
// EqualByExprAndID extends Equal by comparing virtual expression
func (col *Column) EqualByExprAndID(ctx EvalContext, expr Expression) bool {
if newCol, ok := expr.(*Column); ok {
expr, isOk := col.VirtualExpr.(*ScalarFunction)
Expand Down
2 changes: 1 addition & 1 deletion pkg/expression/expression.go
Original file line number Diff line number Diff line change
Expand Up @@ -219,7 +219,7 @@ type Expression interface {
// resolveIndices is called inside the `ResolveIndices` It will perform on the expression itself.
resolveIndices(schema *Schema) error

// ResolveIndicesByVirtualExpr resolves indices by the given schema in terms of virual expression. It will copy the original expression and return the copied one.
// ResolveIndicesByVirtualExpr resolves indices by the given schema in terms of virtual expression. It will copy the original expression and return the copied one.
ResolveIndicesByVirtualExpr(ctx EvalContext, schema *Schema) (Expression, bool)

// resolveIndicesByVirtualExpr is called inside the `ResolveIndicesByVirtualExpr` It will perform on the expression itself.
Expand Down
6 changes: 3 additions & 3 deletions pkg/expression/util.go
Original file line number Diff line number Diff line change
Expand Up @@ -997,7 +997,7 @@ func Contains(exprs []Expression, e Expression) bool {

// ExtractFiltersFromDNFs checks whether the cond is DNF. If so, it will get the extracted part and the remained part.
// The original DNF will be replaced by the remained part or just be deleted if remained part is nil.
// And the extracted part will be appended to the end of the orignal slice.
// And the extracted part will be appended to the end of the original slice.
func ExtractFiltersFromDNFs(ctx BuildContext, conditions []Expression) []Expression {
var allExtracted []Expression
for i := len(conditions) - 1; i >= 0; i-- {
Expand Down Expand Up @@ -1249,7 +1249,7 @@ func GetStringFromConstant(ctx EvalContext, value Expression) (string, bool, err
return str, false, nil
}

// GetIntFromConstant gets an interger value from the Constant expression.
// GetIntFromConstant gets an integer value from the Constant expression.
func GetIntFromConstant(ctx EvalContext, value Expression) (int, bool, error) {
str, isNull, err := GetStringFromConstant(ctx, value)
if err != nil || isNull {
Expand Down Expand Up @@ -1494,7 +1494,7 @@ func RemoveMutableConst(ctx BuildContext, exprs []Expression) (err error) {
case *Constant:
v.ParamMarker = nil
if v.DeferredExpr != nil { // evaluate and update v.Value to convert v to a complete immutable constant.
// TODO: remove or hide DefferedExpr since it's too dangerous (hard to be consistent with v.Value all the time).
// TODO: remove or hide DeferredExpr since it's too dangerous (hard to be consistent with v.Value all the time).
v.Value, err = v.DeferredExpr.Eval(ctx.GetEvalCtx(), chunk.Row{})
if err != nil {
return err
Expand Down
2 changes: 1 addition & 1 deletion pkg/session/bootstrap_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -781,7 +781,7 @@ func TestIndexMergeUpgradeFrom300To540(t *testing.T) {
require.NoError(t, err)
require.Equal(t, int64(ver300), ver)

// We are now in 3.0.0, check tidb_enable_index_merge shoudle not exist.
// We are now in 3.0.0, check tidb_enable_index_merge should not exist.
res := MustExecToRecodeSet(t, seV3, fmt.Sprintf("select * from mysql.GLOBAL_VARIABLES where variable_name='%s'", variable.TiDBEnableIndexMerge))
chk := res.NewChunk(nil)
err = res.Next(ctx, chk)
Expand Down
2 changes: 1 addition & 1 deletion pkg/session/bootstraptest/bootstrap_upgrade_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -474,7 +474,7 @@ func checkDDLJobExecSucc(t *testing.T, se sessiontypes.Session, jobID int64) {
// TestUpgradeVersionForSystemPausedJob tests mock the first upgrade failed, and it has a mock system DDL in queue.
// Then we do re-upgrade(This operation will pause all DDL jobs by the system).
func TestUpgradeVersionForSystemPausedJob(t *testing.T) {
// Mock a general and a reorg job in boostrap.
// Mock a general and a reorg job in bootstrap.
mock := true
session.WithMockUpgrade = &mock
session.MockUpgradeToVerLatestKind = session.MockSimpleUpgradeToVerLatest
Expand Down
4 changes: 2 additions & 2 deletions pkg/session/session.go
Original file line number Diff line number Diff line change
Expand Up @@ -747,7 +747,7 @@ func (s *session) commitTxnWithTemporaryData(ctx context.Context, txn kv.Transac
return nil
}

// errIsNoisy is used to filter DUPLCATE KEY errors.
// errIsNoisy is used to filter DUPLICATE KEY errors.
// These can observed by users in INFORMATION_SCHEMA.CLIENT_ERRORS_SUMMARY_GLOBAL instead.
//
// The rationale for filtering these errors is because they are "client generated errors". i.e.
Expand Down Expand Up @@ -2340,7 +2340,7 @@ const ExecStmtVarKey ExecStmtVarKeyType = 0
// execStmtResult is the return value of ExecuteStmt and it implements the sqlexec.RecordSet interface.
// Why we need a struct to wrap a RecordSet and provide another RecordSet?
// This is because there are so many session state related things that definitely not belongs to the original
// RecordSet, so this struct exists and RecordSet.Close() is overrided handle that.
// RecordSet, so this struct exists and RecordSet.Close() is overridden to handle that.
type execStmtResult struct {
sqlexec.RecordSet
se *session
Expand Down
2 changes: 1 addition & 1 deletion pkg/session/test/session_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -287,7 +287,7 @@ func TestParseWithParams(t *testing.T) {
se := tk.Session()
exec := se.GetRestrictedSQLExecutor()

// test compatibility with ExcuteInternal
// test compatibility with ExecuteInternal
_, err := exec.ParseWithParams(context.TODO(), "SELECT 4")
require.NoError(t, err)

Expand Down
Loading