fix: add forgejo doctor cleanup-commit-status command (#10686)

```
NAME:
   forgejo doctor cleanup-commit-status - Cleanup extra records in commit_status table

USAGE:
   forgejo doctor cleanup-commit-status

DESCRIPTION:
   Forgejo suffered from a bug which caused the creation of more entries in the
   "commit_status" table than necessary. This operation removes the redundant
   data caused by the bug. Removing this data is almost always safe.
   These reundant records can be accessed by users through the API, making it
   possible, but unlikely, that removing it could have an impact to
   integrating services (API: /repos/{owner}/{repo}/commits/{ref}/statuses).

   It is safe to run while Forgejo is online.

   On very large Forgejo instances, the performance of operation will improve
   if the buffer-size option is used with large values. Approximately 130 MB of
   memory is required for every 100,000 records in the buffer.

   Bug reference: https://codeberg.org/forgejo/forgejo/issues/10671

OPTIONS:
   --help, -h                       show help
   --custom-path string, -C string  Set custom path (defaults to '{WorkPath}/custom')
   --config string, -c string       Set custom config file (defaults to '{WorkPath}/custom/conf/app.ini')
   --work-path string, -w string    Set Forgejo's working path (defaults to the directory of the Forgejo binary)
   --verbose, -V                    Show process details
   --dry-run                        Report statistics from the operation but do not modify the database
   --buffer-size int                Record count per query while iterating records; larger values are typically faster but use more memory (default: 100000)
   --delete-chunk-size int          Number of records to delete per DELETE query (default: 1000)
```

The cleanup effectively performs `SELECT * FROM commit_status ORDER BY repo_id, sha, context, index, id`, and iterates through the records.  Whenever `index, id` changes without the other fields changing, then it's a useless record that can be deleted.  The major complication is doing that at scale without bringing the entire database table into memory, which is performed through a new iteration method `IterateByKeyset`.

Manually tested against a 455,303 record table in PostgreSQL, MySQL, and SQLite, which was reduced to 10,781 records, dropping 97.5% of the records.

Reviewed-on: https://codeberg.org/forgejo/forgejo/pulls/10686
Reviewed-by: Gusted <gusted@noreply.codeberg.org>
Co-authored-by: Mathieu Fenniak <mathieu@fenniak.net>
Co-committed-by: Mathieu Fenniak <mathieu@fenniak.net>
This commit is contained in:
Mathieu Fenniak 2026-01-10 23:15:21 +01:00 committed by Gusted
parent ba2a6fbc41
commit 270317a3ad
6 changed files with 460 additions and 1 deletions

View file

@ -15,6 +15,7 @@ import (
"text/tabwriter"
"forgejo.org/models/db"
git_model "forgejo.org/models/git"
"forgejo.org/models/gitea_migrations"
migrate_base "forgejo.org/models/gitea_migrations/base"
repo_model "forgejo.org/models/repo"
@ -41,6 +42,7 @@ func cmdDoctor() *cli.Command {
cmdRecreateTable(),
cmdDoctorConvert(),
cmdAvatarStripExif(),
cmdCleanupCommitStatuses(),
},
}
}
@ -115,6 +117,54 @@ func cmdAvatarStripExif() *cli.Command {
}
}
func cmdCleanupCommitStatuses() *cli.Command {
return &cli.Command{
Name: "cleanup-commit-status",
Usage: "Cleanup extra records in commit_status table",
Description: `Forgejo suffered from a bug which caused the creation of more entries in the
"commit_status" table than necessary. This operation removes the redundant
data caused by the bug. Removing this data is almost always safe.
These reundant records can be accessed by users through the API, making it
possible, but unlikely, that removing it could have an impact to
integrating services (API: /repos/{owner}/{repo}/commits/{ref}/statuses).
It is safe to run while Forgejo is online.
On very large Forgejo instances, the performance of operation will improve
if the buffer-size option is used with large values. Approximately 130 MB of
memory is required for every 100,000 records in the buffer.
Bug reference: https://codeberg.org/forgejo/forgejo/issues/10671
`,
Before: multipleBefore(noDanglingArgs, PrepareConsoleLoggerLevel(log.INFO)),
Action: runCleanupCommitStatus,
Flags: []cli.Flag{
&cli.BoolFlag{
Name: "verbose",
Aliases: []string{"V"},
Usage: "Show process details",
},
&cli.BoolFlag{
Name: "dry-run",
Usage: "Report statistics from the operation but do not modify the database",
},
&cli.IntFlag{
Name: "buffer-size",
Usage: "Record count per query while iterating records; larger values are typically faster but use more memory",
// See IterateByKeyset's documentation for performance notes which led to the choice of the default
// buffer size for this operation.
Value: 100000,
},
&cli.IntFlag{
Name: "delete-chunk-size",
Usage: "Number of records to delete per DELETE query",
Value: 1000,
},
},
}
}
func runRecreateTable(stdCtx context.Context, ctx *cli.Command) error {
stdCtx, cancel := installSignals(stdCtx)
defer cancel()
@ -322,3 +372,19 @@ func runAvatarStripExif(ctx context.Context, c *cli.Command) error {
return nil
}
func runCleanupCommitStatus(ctx context.Context, cli *cli.Command) error {
ctx, cancel := installSignals(ctx)
defer cancel()
if err := initDB(ctx); err != nil {
return err
}
bufferSize := cli.Int("buffer-size")
deleteChunkSize := cli.Int("delete-chunk-size")
dryRun := cli.Bool("dry-run")
log.Debug("bufferSize = %d, deleteChunkSize = %d, dryRun = %v", bufferSize, deleteChunkSize, dryRun)
return git_model.CleanupCommitStatus(ctx, bufferSize, deleteChunkSize, dryRun)
}

View file

@ -5,8 +5,10 @@ package db
import (
"context"
"errors"
"fmt"
"reflect"
"strings"
"forgejo.org/modules/setting"
@ -84,3 +86,123 @@ func extractFieldValue(bean any, fieldName string) any {
field := v.FieldByName(fieldName)
return field.Interface()
}
// IterateByKeyset iterates all the records on a database (matching the provided condition) in the order of specified
// order fields, and invokes the provided handler function for each record. It is safe to UPDATE or DELETE the record in
// the handler function, as long as the order fields are not mutated on the record (which could cause records to be
// missed or iterated multiple times).
//
// Assuming order fields a, b, and c, then database queries will be performed as "SELECT * FROM table WHERE (a, b, c) >
// (last_a, last_b, last_c) ORDER BY a, b, c LIMIT buffer_size" repeatedly until the query returns no records (except
// the first query will have no WHERE clause).
//
// Critical requirements for proper usage:
//
// - the order fields encompass at least one UNIQUE or PRIMARY KEY constraint of the table to ensure that records are
// not duplicated -- for example, if the table has a unique index covering `(repo_id, index)`, then it would be safe to
// use this function as long as both fields (in either order) are provided as order fields.
//
// - none of the order fields may have NULL values in them, as the `=` and `>` comparisons being performed by the
// iterative queries will not operate on these records consistently as they do with other values.
//
// This implementation could be a much simpler streaming scan of the query results, except that doesn't permit making
// any additional database queries or data modifications in the target function -- SQLite cannot write while holding a
// read lock. Buffering pages of data in-memory avoids that issue.
//
// Performance:
//
// - High performance will result from an alignment of an index on the table with the order fields, in the same field
// order, even if additional ordering fields could be provided after the index fields. In the absence of this index
// alignment, it is reasonable to expect that every extra page of data accessed will require a query that will perform
// an index scan (if available) or sequential scan of the target table. In testing on the `commit_status` table with
// 455k records, a fully index-supported ordering allowed each query page to execute in 0.18ms, as opposed to 80ms
// per-query without matching supporting index.
//
// - In the absence of a matching index, slower per-query performance can be compensated with a larger `batchSize`
// parameter, which controls how many records to fetch at once and therefore reduces the number of queries required.
// This requires more memory. Similar `commit_status` table testing showed these stats for iteration time and memory
// usage for different buffer sizes; specifics will vary depending on the target table:
// - buffer size = 1,000,000 - iterates in 2.8 seconds, consumes 363 MB of RAM
// - buffer size = 100,000 - iterates in 3.5 seconds, consume 130 MB of RAM
// - buffer size = 10,000 - iterates in 7.1 seconds, consumes 59 MB of RAM
// - buffer size = 1,000 - iterates in 33.9 seconds, consumes 42 MB of RAM
func IterateByKeyset[Bean any](ctx context.Context, cond builder.Cond, orderFields []string, batchSize int, f func(ctx context.Context, bean *Bean) error) error {
var dummy Bean
if len(orderFields) == 0 {
return errors.New("orderFields must be provided")
}
table, err := TableInfo(&dummy)
if err != nil {
return fmt.Errorf("unable to fetch table info for bean %v: %w", dummy, err)
}
goFieldNames := make([]string, len(orderFields))
for i, f := range orderFields {
goFieldNames[i] = table.GetColumn(f).FieldName
}
sqlFieldNames := make([]string, len(orderFields))
for i, f := range orderFields {
// Support field names like "index" which need quoting in builder.Cond & OrderBy
sqlFieldNames[i] = x.Dialect().Quoter().Quote(f)
}
var lastKey []any
// For the order fields, generate clauses (a, b, c) and (?, ?, ?) which will be used in the WHERE clause when
// reading additional pages of data.
rowValue := strings.Builder{}
rowParameterValue := strings.Builder{}
rowValue.WriteString("(")
rowParameterValue.WriteString("(")
for i, f := range sqlFieldNames {
rowValue.WriteString(f)
rowParameterValue.WriteString("?")
if i != len(sqlFieldNames)-1 {
rowValue.WriteString(", ")
rowParameterValue.WriteString(", ")
}
}
rowValue.WriteString(")")
rowParameterValue.WriteString(")")
for {
select {
case <-ctx.Done():
return ctx.Err()
default:
beans := make([]*Bean, 0, batchSize)
sess := GetEngine(ctx)
for _, f := range sqlFieldNames {
sess = sess.OrderBy(f)
}
if cond != nil {
sess = sess.Where(cond)
}
if lastKey != nil {
sess = sess.Where(
builder.Expr(fmt.Sprintf("%s > %s", rowValue.String(), rowParameterValue.String()), lastKey...))
}
if err := sess.Limit(batchSize).Find(&beans); err != nil {
return err
}
if len(beans) == 0 {
return nil
}
for _, bean := range beans {
if err := f(ctx, bean); err != nil {
return err
}
}
lastBean := beans[len(beans)-1]
lastKey = make([]any, len(goFieldNames))
for i := range goFieldNames {
lastKey[i] = extractFieldValue(lastBean, goFieldNames[i])
}
}
}
}

View file

@ -10,6 +10,7 @@ import (
"testing"
"forgejo.org/models/db"
git_model "forgejo.org/models/git"
repo_model "forgejo.org/models/repo"
"forgejo.org/models/unittest"
"forgejo.org/modules/setting"
@ -21,7 +22,6 @@ import (
)
func TestIterate(t *testing.T) {
db.SetLogSQL(t.Context(), true)
defer test.MockVariableValue(&setting.Database.IterateBufferSize, 50)()
t.Run("No Modifications", func(t *testing.T) {
@ -115,3 +115,31 @@ func TestIterate(t *testing.T) {
assert.Empty(t, remainingRepoIDs)
})
}
func TestIterateMultipleFields(t *testing.T) {
for _, bufferSize := range []int{1, 2, 3, 10} { // 8 records in fixture
t.Run(fmt.Sprintf("No Modifications bufferSize=%d", bufferSize), func(t *testing.T) {
require.NoError(t, unittest.PrepareTestDatabase())
// Fetch all the commit status IDs...
var remainingIDs []int64
err := db.GetEngine(t.Context()).Table(&git_model.CommitStatus{}).Cols("id").Find(&remainingIDs)
require.NoError(t, err)
require.NotEmpty(t, remainingIDs)
// Ensure that every repo unit ID is found when doing iterate:
err = db.IterateByKeyset(t.Context(),
nil,
[]string{"repo_id", "sha", "context", "index", "id"},
bufferSize,
func(ctx context.Context, commit_status *git_model.CommitStatus) error {
remainingIDs = slices.DeleteFunc(remainingIDs, func(n int64) bool {
return commit_status.ID == n
})
return nil
})
require.NoError(t, err)
assert.Empty(t, remainingIDs)
})
}
}

View file

@ -0,0 +1,135 @@
# Fields that should, if changed, prevent deletion: repo_id, sha, context, state, description. The first test sets will
# be varying each of these fields independently to confirm they're kept.
# Vary description:
-
id: 10
index: 1
repo_id: 62
state: "pending"
sha: "01"
description: "Waiting for wake up"
context: deploy/awesomeness
-
id: 11
index: 2
repo_id: 62
state: "pending"
sha: "01"
description: "Almost woke up..."
context: deploy/awesomeness
# Vary state:
-
id: 12
index: 1
repo_id: 62
state: "pending"
sha: "02"
description: "Waiting for wake up"
context: deploy/awesomeness
-
id: 13
index: 2
repo_id: 62
state: "success"
sha: "02"
description: "Waiting for wake up"
context: deploy/awesomeness
# Vary context:
-
id: 14
index: 1
repo_id: 62
state: "pending"
sha: "03"
description: "Waiting for wake up"
context: deploy/awesomeness-v1
-
id: 15
index: 2
repo_id: 62
state: "pending"
sha: "03"
description: "Waiting for wake up"
context: deploy/awesomeness-v2
# Vary sha:
-
id: 16
index: 1
repo_id: 62
state: "pending"
sha: "04"
description: "Waiting for wake up"
context: deploy/awesomeness-v1
-
id: 17
index: 2
repo_id: 62
state: "pending"
sha: "05"
description: "Waiting for wake up"
context: deploy/awesomeness-v1
# Vary Repo ID:
-
id: 18
index: 1
repo_id: 62
state: "pending"
sha: "06"
description: "Waiting for wake up"
context: deploy/awesomeness-v1
-
id: 19
index: 2
repo_id: 63
state: "pending"
sha: "06"
description: "Waiting for wake up"
context: deploy/awesomeness-v1
# That's all the varying cases, now here's the data that should be affected by the delete:
-
id: 20
index: 1
repo_id: 62
state: "pending"
sha: "07"
description: "Waiting for wake up"
context: deploy/awesomeness-v1
- # Dupe 1
id: 21
index: 2
repo_id: 62
state: "pending"
sha: "07"
description: "Waiting for wake up"
context: deploy/awesomeness-v1
- # Dupe 2
id: 22
index: 3
repo_id: 62
state: "pending"
sha: "07"
description: "Waiting for wake up"
context: deploy/awesomeness-v1
- # Switched to "success", keep
id: 23
index: 4
repo_id: 62
state: "success"
sha: "07"
description: "Successful!"
context: deploy/awesomeness-v1
- # Dupe reporting success again
id: 24
index: 5
repo_id: 62
state: "success"
sha: "07"
description: "Successful!"
context: deploy/awesomeness-v1

View file

@ -467,3 +467,68 @@ func ParseCommitsWithStatus(ctx context.Context, commits []*git.Commit, repo *re
func hashCommitStatusContext(context string) string {
return fmt.Sprintf("%x", sha1.Sum([]byte(context)))
}
func CleanupCommitStatus(ctx context.Context, bufferSize, deleteChunkSize int, dryRun bool) error {
startTime := time.Now()
var lastCommitStatus CommitStatus
deleteTargets := make([]int64, 0, deleteChunkSize)
recordCount := 0
deleteCount := 0
err := db.IterateByKeyset(ctx,
nil,
[]string{"repo_id", "sha", "context", "index", "id"},
bufferSize,
func(ctx context.Context, commitStatus *CommitStatus) error {
if commitStatus.RepoID != lastCommitStatus.RepoID ||
commitStatus.SHA != lastCommitStatus.SHA ||
commitStatus.Context != lastCommitStatus.Context ||
commitStatus.State != lastCommitStatus.State ||
commitStatus.Description != lastCommitStatus.Description {
// New context, or changed state/description; keep it, start looking for duplicates of it.
lastCommitStatus = *commitStatus
} else {
// Same context as previous record, and same state -- this record shouldn't have been stored.
deleteTargets = append(deleteTargets, commitStatus.ID)
if len(deleteTargets) == deleteChunkSize {
// Flush delete chunk
log.Debug("deleting chunk of %d records (dryRun=%v)", len(deleteTargets), dryRun)
if !dryRun {
if err := db.DeleteByIDs[CommitStatus](ctx, deleteTargets...); err != nil {
return err
}
}
deleteCount += len(deleteTargets)
deleteTargets = make([]int64, 0, deleteChunkSize)
}
}
recordCount++
return nil
})
if err != nil {
return err
}
if len(deleteTargets) > 0 {
log.Debug("deleting final chunk of %d records (dryRun=%v)", len(deleteTargets), dryRun)
if !dryRun {
if err := db.DeleteByIDs[CommitStatus](ctx, deleteTargets...); err != nil {
return err
}
}
deleteCount += len(deleteTargets)
}
duration := time.Since(startTime)
if dryRun {
log.Info("Reviewed %d records in commit_status, and would delete %d", recordCount, deleteCount)
} else {
log.Info("Reviewed %d records in commit_status, and deleted %d", recordCount, deleteCount)
}
log.Info("Cleanup commit status took %d milliseconds", duration.Milliseconds())
return nil
}

View file

@ -244,3 +244,46 @@ func TestFindRepoRecentCommitStatusContexts(t *testing.T) {
assert.Equal(t, "compliance/lint-backend", contexts[0])
}
}
func TestCleanupCommitStatus(t *testing.T) {
defer unittest.OverrideFixtures("models/git/TestCleanupCommitStatus")()
require.NoError(t, unittest.PrepareTestDatabase())
// No changes after a dry run:
originalCount := unittest.GetCount(t, &git_model.CommitStatus{})
err := git_model.CleanupCommitStatus(t.Context(), 100, 100, true)
require.NoError(t, err)
countAfterDryRun := unittest.GetCount(t, &git_model.CommitStatus{})
assert.Equal(t, originalCount, countAfterDryRun)
// Perform actual cleanup
err = git_model.CleanupCommitStatus(t.Context(), 100, 100, false)
require.NoError(t, err)
// Varying descriptions
unittest.AssertExistsAndLoadBean(t, &git_model.CommitStatus{ID: 10})
unittest.AssertExistsAndLoadBean(t, &git_model.CommitStatus{ID: 11})
// Varying state
unittest.AssertExistsAndLoadBean(t, &git_model.CommitStatus{ID: 12})
unittest.AssertExistsAndLoadBean(t, &git_model.CommitStatus{ID: 13})
// Varying context
unittest.AssertExistsAndLoadBean(t, &git_model.CommitStatus{ID: 14})
unittest.AssertExistsAndLoadBean(t, &git_model.CommitStatus{ID: 15})
// Varying sha
unittest.AssertExistsAndLoadBean(t, &git_model.CommitStatus{ID: 16})
unittest.AssertExistsAndLoadBean(t, &git_model.CommitStatus{ID: 17})
// Varying repo ID
unittest.AssertExistsAndLoadBean(t, &git_model.CommitStatus{ID: 18})
unittest.AssertExistsAndLoadBean(t, &git_model.CommitStatus{ID: 19})
// Expected to remain or be removed from cleanup of fixture data:
unittest.AssertExistsAndLoadBean(t, &git_model.CommitStatus{ID: 20})
unittest.AssertNotExistsBean(t, &git_model.CommitStatus{ID: 21})
unittest.AssertNotExistsBean(t, &git_model.CommitStatus{ID: 22})
unittest.AssertExistsAndLoadBean(t, &git_model.CommitStatus{ID: 23})
unittest.AssertNotExistsBean(t, &git_model.CommitStatus{ID: 24})
}