summaryrefslogtreecommitdiffstats
path: root/tests/test_utils.go
diff options
context:
space:
mode:
authorDaniel Baumann <daniel@debian.org>2024-10-18 20:33:49 +0200
committerDaniel Baumann <daniel@debian.org>2024-10-18 20:33:49 +0200
commitdd136858f1ea40ad3c94191d647487fa4f31926c (patch)
tree58fec94a7b2a12510c9664b21793f1ed560c6518 /tests/test_utils.go
parentInitial commit. (diff)
downloadforgejo-upstream/9.0.0.tar.xz
forgejo-upstream/9.0.0.zip
Adding upstream version 9.0.0.upstream/9.0.0upstreamdebian
Signed-off-by: Daniel Baumann <daniel@debian.org>
Diffstat (limited to 'tests/test_utils.go')
-rw-r--r--tests/test_utils.go452
1 files changed, 452 insertions, 0 deletions
diff --git a/tests/test_utils.go b/tests/test_utils.go
new file mode 100644
index 0000000..ee10689
--- /dev/null
+++ b/tests/test_utils.go
@@ -0,0 +1,452 @@
+// Copyright 2017 The Gitea Authors. All rights reserved.
+// SPDX-License-Identifier: MIT
+
+//nolint:forbidigo
+package tests
+
+import (
+ "context"
+ "database/sql"
+ "fmt"
+ "os"
+ "path"
+ "path/filepath"
+ "runtime"
+ "strings"
+ "testing"
+ "time"
+
+ "code.gitea.io/gitea/models/db"
+ packages_model "code.gitea.io/gitea/models/packages"
+ repo_model "code.gitea.io/gitea/models/repo"
+ unit_model "code.gitea.io/gitea/models/unit"
+ "code.gitea.io/gitea/models/unittest"
+ user_model "code.gitea.io/gitea/models/user"
+ "code.gitea.io/gitea/modules/base"
+ "code.gitea.io/gitea/modules/git"
+ "code.gitea.io/gitea/modules/graceful"
+ "code.gitea.io/gitea/modules/log"
+ "code.gitea.io/gitea/modules/optional"
+ "code.gitea.io/gitea/modules/process"
+ repo_module "code.gitea.io/gitea/modules/repository"
+ "code.gitea.io/gitea/modules/setting"
+ "code.gitea.io/gitea/modules/storage"
+ "code.gitea.io/gitea/modules/testlogger"
+ "code.gitea.io/gitea/modules/util"
+ "code.gitea.io/gitea/routers"
+ repo_service "code.gitea.io/gitea/services/repository"
+ files_service "code.gitea.io/gitea/services/repository/files"
+ wiki_service "code.gitea.io/gitea/services/wiki"
+
+ "github.com/google/uuid"
+ "github.com/stretchr/testify/assert"
+ "github.com/stretchr/testify/require"
+)
+
+func exitf(format string, args ...any) {
+ fmt.Printf(format+"\n", args...)
+ os.Exit(1)
+}
+
+func InitTest(requireGitea bool) {
+ log.RegisterEventWriter("test", testlogger.NewTestLoggerWriter)
+
+ giteaRoot := base.SetupGiteaRoot()
+ if giteaRoot == "" {
+ exitf("Environment variable $GITEA_ROOT not set")
+ }
+
+ // TODO: Speedup tests that rely on the event source ticker, confirm whether there is any bug or failure.
+ // setting.UI.Notification.EventSourceUpdateTime = time.Second
+
+ setting.IsInTesting = true
+ setting.AppWorkPath = giteaRoot
+ setting.CustomPath = filepath.Join(setting.AppWorkPath, "custom")
+ if requireGitea {
+ giteaBinary := "gitea"
+ if setting.IsWindows {
+ giteaBinary += ".exe"
+ }
+ setting.AppPath = path.Join(giteaRoot, giteaBinary)
+ if _, err := os.Stat(setting.AppPath); err != nil {
+ exitf("Could not find gitea binary at %s", setting.AppPath)
+ }
+ }
+ giteaConf := os.Getenv("GITEA_CONF")
+ if giteaConf == "" {
+ // By default, use sqlite.ini for testing, then IDE like GoLand can start the test process with debugger.
+ // It's easier for developers to debug bugs step by step with a debugger.
+ // Notice: when doing "ssh push", Gitea executes sub processes, debugger won't work for the sub processes.
+ giteaConf = "tests/sqlite.ini"
+ _ = os.Setenv("GITEA_CONF", giteaConf)
+ fmt.Printf("Environment variable $GITEA_CONF not set, use default: %s\n", giteaConf)
+ if !setting.EnableSQLite3 {
+ exitf(`sqlite3 requires: import _ "github.com/mattn/go-sqlite3" or -tags sqlite,sqlite_unlock_notify`)
+ }
+ }
+ if !path.IsAbs(giteaConf) {
+ setting.CustomConf = filepath.Join(giteaRoot, giteaConf)
+ } else {
+ setting.CustomConf = giteaConf
+ }
+
+ unittest.InitSettings()
+ setting.Repository.DefaultBranch = "master" // many test code still assume that default branch is called "master"
+ _ = util.RemoveAll(repo_module.LocalCopyPath())
+
+ if err := git.InitFull(context.Background()); err != nil {
+ log.Fatal("git.InitOnceWithSync: %v", err)
+ }
+
+ setting.LoadDBSetting()
+ if err := storage.Init(); err != nil {
+ exitf("Init storage failed: %v", err)
+ }
+
+ switch {
+ case setting.Database.Type.IsMySQL():
+ connType := "tcp"
+ if len(setting.Database.Host) > 0 && setting.Database.Host[0] == '/' { // looks like a unix socket
+ connType = "unix"
+ }
+
+ db, err := sql.Open("mysql", fmt.Sprintf("%s:%s@%s(%s)/",
+ setting.Database.User, setting.Database.Passwd, connType, setting.Database.Host))
+ defer db.Close()
+ if err != nil {
+ log.Fatal("sql.Open: %v", err)
+ }
+ if _, err = db.Exec(fmt.Sprintf("CREATE DATABASE IF NOT EXISTS %s", strings.SplitN(setting.Database.Name, "?", 2)[0])); err != nil {
+ log.Fatal("db.Exec: %v", err)
+ }
+ case setting.Database.Type.IsPostgreSQL():
+ var db *sql.DB
+ var err error
+ if setting.Database.Host[0] == '/' {
+ db, err = sql.Open("postgres", fmt.Sprintf("postgres://%s:%s@/%s?sslmode=%s&host=%s",
+ setting.Database.User, setting.Database.Passwd, setting.Database.Name, setting.Database.SSLMode, setting.Database.Host))
+ } else {
+ db, err = sql.Open("postgres", fmt.Sprintf("postgres://%s:%s@%s/%s?sslmode=%s",
+ setting.Database.User, setting.Database.Passwd, setting.Database.Host, setting.Database.Name, setting.Database.SSLMode))
+ }
+
+ defer db.Close()
+ if err != nil {
+ log.Fatal("sql.Open: %v", err)
+ }
+ dbrows, err := db.Query(fmt.Sprintf("SELECT 1 FROM pg_database WHERE datname = '%s'", setting.Database.Name))
+ if err != nil {
+ log.Fatal("db.Query: %v", err)
+ }
+ defer dbrows.Close()
+
+ if !dbrows.Next() {
+ if _, err = db.Exec(fmt.Sprintf("CREATE DATABASE %s", setting.Database.Name)); err != nil {
+ log.Fatal("db.Exec: CREATE DATABASE: %v", err)
+ }
+ }
+ // Check if we need to setup a specific schema
+ if len(setting.Database.Schema) == 0 {
+ break
+ }
+ db.Close()
+
+ if setting.Database.Host[0] == '/' {
+ db, err = sql.Open("postgres", fmt.Sprintf("postgres://%s:%s@/%s?sslmode=%s&host=%s",
+ setting.Database.User, setting.Database.Passwd, setting.Database.Name, setting.Database.SSLMode, setting.Database.Host))
+ } else {
+ db, err = sql.Open("postgres", fmt.Sprintf("postgres://%s:%s@%s/%s?sslmode=%s",
+ setting.Database.User, setting.Database.Passwd, setting.Database.Host, setting.Database.Name, setting.Database.SSLMode))
+ }
+ // This is a different db object; requires a different Close()
+ defer db.Close()
+ if err != nil {
+ log.Fatal("sql.Open: %v", err)
+ }
+ schrows, err := db.Query(fmt.Sprintf("SELECT 1 FROM information_schema.schemata WHERE schema_name = '%s'", setting.Database.Schema))
+ if err != nil {
+ log.Fatal("db.Query: %v", err)
+ }
+ defer schrows.Close()
+
+ if !schrows.Next() {
+ // Create and setup a DB schema
+ if _, err = db.Exec(fmt.Sprintf("CREATE SCHEMA %s", setting.Database.Schema)); err != nil {
+ log.Fatal("db.Exec: CREATE SCHEMA: %v", err)
+ }
+ }
+ }
+
+ routers.InitWebInstalled(graceful.GetManager().HammerContext())
+}
+
+func PrepareAttachmentsStorage(t testing.TB) {
+ // prepare attachments directory and files
+ require.NoError(t, storage.Clean(storage.Attachments))
+
+ s, err := storage.NewStorage(setting.LocalStorageType, &setting.Storage{
+ Path: filepath.Join(filepath.Dir(setting.AppPath), "tests", "testdata", "data", "attachments"),
+ })
+ require.NoError(t, err)
+ require.NoError(t, s.IterateObjects("", func(p string, obj storage.Object) error {
+ _, err = storage.Copy(storage.Attachments, p, s, p)
+ return err
+ }))
+}
+
+// cancelProcesses cancels all processes of the [process.Manager].
+// Returns immediately if delay is 0, otherwise wait until all processes are done
+// and fails the test if it takes longer that the given delay.
+func cancelProcesses(t testing.TB, delay time.Duration) {
+ processManager := process.GetManager()
+ processes, _ := processManager.Processes(true, true)
+ for _, p := range processes {
+ processManager.Cancel(p.PID)
+ t.Logf("PrepareTestEnv:Process %q cancelled", p.Description)
+ }
+ if delay == 0 || len(processes) == 0 {
+ return
+ }
+
+ start := time.Now()
+ processes, _ = processManager.Processes(true, true)
+ for len(processes) > 0 {
+ if time.Since(start) > delay {
+ t.Errorf("ERROR PrepareTestEnv: could not cancel all processes within %s", delay)
+ for _, p := range processes {
+ t.Logf("PrepareTestEnv:Remaining Process: %q", p.Description)
+ }
+ return
+ }
+ runtime.Gosched() // let the context cancellation propagate
+ processes, _ = processManager.Processes(true, true)
+ }
+ t.Logf("PrepareTestEnv: all processes cancelled within %s", time.Since(start))
+}
+
+func PrepareTestEnv(t testing.TB, skip ...int) func() {
+ t.Helper()
+ ourSkip := 1
+ if len(skip) > 0 {
+ ourSkip += skip[0]
+ }
+ deferFn := PrintCurrentTest(t, ourSkip)
+
+ // kill all background processes to prevent them from interfering with the fixture loading
+ // see https://codeberg.org/forgejo/forgejo/issues/2962
+ cancelProcesses(t, 30*time.Second)
+ t.Cleanup(func() { cancelProcesses(t, 0) }) // cancel remaining processes in a non-blocking way
+
+ // load database fixtures
+ require.NoError(t, unittest.LoadFixtures())
+
+ // load git repo fixtures
+ require.NoError(t, util.RemoveAll(setting.RepoRootPath))
+ require.NoError(t, unittest.CopyDir(path.Join(filepath.Dir(setting.AppPath), "tests/gitea-repositories-meta"), setting.RepoRootPath))
+ ownerDirs, err := os.ReadDir(setting.RepoRootPath)
+ if err != nil {
+ require.NoError(t, err, "unable to read the new repo root: %v\n", err)
+ }
+ for _, ownerDir := range ownerDirs {
+ if !ownerDir.Type().IsDir() {
+ continue
+ }
+ repoDirs, err := os.ReadDir(filepath.Join(setting.RepoRootPath, ownerDir.Name()))
+ if err != nil {
+ require.NoError(t, err, "unable to read the new repo root: %v\n", err)
+ }
+ for _, repoDir := range repoDirs {
+ _ = os.MkdirAll(filepath.Join(setting.RepoRootPath, ownerDir.Name(), repoDir.Name(), "objects", "pack"), 0o755)
+ _ = os.MkdirAll(filepath.Join(setting.RepoRootPath, ownerDir.Name(), repoDir.Name(), "objects", "info"), 0o755)
+ _ = os.MkdirAll(filepath.Join(setting.RepoRootPath, ownerDir.Name(), repoDir.Name(), "refs", "heads"), 0o755)
+ _ = os.MkdirAll(filepath.Join(setting.RepoRootPath, ownerDir.Name(), repoDir.Name(), "refs", "tag"), 0o755)
+ }
+ }
+
+ // load LFS object fixtures
+ // (LFS storage can be on any of several backends, including remote servers, so we init it with the storage API)
+ lfsFixtures, err := storage.NewStorage(setting.LocalStorageType, &setting.Storage{
+ Path: filepath.Join(filepath.Dir(setting.AppPath), "tests/gitea-lfs-meta"),
+ })
+ require.NoError(t, err)
+ require.NoError(t, storage.Clean(storage.LFS))
+ require.NoError(t, lfsFixtures.IterateObjects("", func(path string, _ storage.Object) error {
+ _, err := storage.Copy(storage.LFS, path, lfsFixtures, path)
+ return err
+ }))
+
+ // clear all package data
+ require.NoError(t, db.TruncateBeans(db.DefaultContext,
+ &packages_model.Package{},
+ &packages_model.PackageVersion{},
+ &packages_model.PackageFile{},
+ &packages_model.PackageBlob{},
+ &packages_model.PackageProperty{},
+ &packages_model.PackageBlobUpload{},
+ &packages_model.PackageCleanupRule{},
+ ))
+ require.NoError(t, storage.Clean(storage.Packages))
+
+ return deferFn
+}
+
+func PrintCurrentTest(t testing.TB, skip ...int) func() {
+ t.Helper()
+ actualSkip := 1
+ if len(skip) > 0 {
+ actualSkip = skip[0] + 1
+ }
+ return testlogger.PrintCurrentTest(t, actualSkip)
+}
+
+// Printf takes a format and args and prints the string to os.Stdout
+func Printf(format string, args ...any) {
+ testlogger.Printf(format, args...)
+}
+
+func AddFixtures(dirs ...string) func() {
+ return unittest.OverrideFixtures(
+ unittest.FixturesOptions{
+ Dir: filepath.Join(setting.AppWorkPath, "models/fixtures/"),
+ Base: setting.AppWorkPath,
+ Dirs: dirs,
+ },
+ )
+}
+
+type DeclarativeRepoOptions struct {
+ Name optional.Option[string]
+ EnabledUnits optional.Option[[]unit_model.Type]
+ DisabledUnits optional.Option[[]unit_model.Type]
+ Files optional.Option[[]*files_service.ChangeRepoFile]
+ WikiBranch optional.Option[string]
+ AutoInit optional.Option[bool]
+ IsTemplate optional.Option[bool]
+}
+
+func CreateDeclarativeRepoWithOptions(t *testing.T, owner *user_model.User, opts DeclarativeRepoOptions) (*repo_model.Repository, string, func()) {
+ t.Helper()
+
+ // Not using opts.Name.ValueOrDefault() here to avoid unnecessarily
+ // generating an UUID when a name is specified.
+ var repoName string
+ if opts.Name.Has() {
+ repoName = opts.Name.Value()
+ } else {
+ repoName = uuid.NewString()
+ }
+
+ var autoInit bool
+ if opts.AutoInit.Has() {
+ autoInit = opts.AutoInit.Value()
+ } else {
+ autoInit = true
+ }
+
+ // Create the repository
+ repo, err := repo_service.CreateRepository(db.DefaultContext, owner, owner, repo_service.CreateRepoOptions{
+ Name: repoName,
+ Description: "Temporary Repo",
+ AutoInit: autoInit,
+ Gitignores: "",
+ License: "WTFPL",
+ Readme: "Default",
+ DefaultBranch: "main",
+ IsTemplate: opts.IsTemplate.Value(),
+ })
+ require.NoError(t, err)
+ assert.NotEmpty(t, repo)
+
+ // Populate `enabledUnits` if we have any enabled.
+ var enabledUnits []repo_model.RepoUnit
+ if opts.EnabledUnits.Has() {
+ units := opts.EnabledUnits.Value()
+ enabledUnits = make([]repo_model.RepoUnit, len(units))
+
+ for i, unitType := range units {
+ enabledUnits[i] = repo_model.RepoUnit{
+ RepoID: repo.ID,
+ Type: unitType,
+ }
+ }
+ }
+
+ // Adjust the repo units according to our parameters.
+ if opts.EnabledUnits.Has() || opts.DisabledUnits.Has() {
+ err := repo_service.UpdateRepositoryUnits(db.DefaultContext, repo, enabledUnits, opts.DisabledUnits.ValueOrDefault(nil))
+ require.NoError(t, err)
+ }
+
+ // Add files, if any.
+ var sha string
+ if opts.Files.Has() {
+ assert.True(t, autoInit, "Files cannot be specified if AutoInit is disabled")
+ files := opts.Files.Value()
+
+ resp, err := files_service.ChangeRepoFiles(git.DefaultContext, repo, owner, &files_service.ChangeRepoFilesOptions{
+ Files: files,
+ Message: "add files",
+ OldBranch: "main",
+ NewBranch: "main",
+ Author: &files_service.IdentityOptions{
+ Name: owner.Name,
+ Email: owner.Email,
+ },
+ Committer: &files_service.IdentityOptions{
+ Name: owner.Name,
+ Email: owner.Email,
+ },
+ Dates: &files_service.CommitDateOptions{
+ Author: time.Now(),
+ Committer: time.Now(),
+ },
+ })
+ require.NoError(t, err)
+ assert.NotEmpty(t, resp)
+
+ sha = resp.Commit.SHA
+ }
+
+ // If there's a Wiki branch specified, create a wiki, and a default wiki page.
+ if opts.WikiBranch.Has() {
+ // Set the wiki branch in the database first
+ repo.WikiBranch = opts.WikiBranch.Value()
+ err := repo_model.UpdateRepositoryCols(db.DefaultContext, repo, "wiki_branch")
+ require.NoError(t, err)
+
+ // Initialize the wiki
+ err = wiki_service.InitWiki(db.DefaultContext, repo)
+ require.NoError(t, err)
+
+ // Add a new wiki page
+ err = wiki_service.AddWikiPage(db.DefaultContext, owner, repo, "Home", "Welcome to the wiki!", "Add a Home page")
+ require.NoError(t, err)
+ }
+
+ // Return the repo, the top commit, and a defer-able function to delete the
+ // repo.
+ return repo, sha, func() {
+ _ = repo_service.DeleteRepository(db.DefaultContext, owner, repo, false)
+ }
+}
+
+func CreateDeclarativeRepo(t *testing.T, owner *user_model.User, name string, enabledUnits, disabledUnits []unit_model.Type, files []*files_service.ChangeRepoFile) (*repo_model.Repository, string, func()) {
+ t.Helper()
+
+ var opts DeclarativeRepoOptions
+
+ if name != "" {
+ opts.Name = optional.Some(name)
+ }
+ if enabledUnits != nil {
+ opts.EnabledUnits = optional.Some(enabledUnits)
+ }
+ if disabledUnits != nil {
+ opts.DisabledUnits = optional.Some(disabledUnits)
+ }
+ if files != nil {
+ opts.Files = optional.Some(files)
+ }
+
+ return CreateDeclarativeRepoWithOptions(t, owner, opts)
+}