mirror of
https://github.com/hrfee/jfa-go.git
synced 2024-11-09 20:00:12 +00:00
Harvey Tindall
69569e556a
mautrix-go now include a cryptohelper package, which solves all my issues and just works. the setting is now on by default, however packages are not yet built with it.
176 lines
4.4 KiB
Go
176 lines
4.4 KiB
Go
package main
|
|
|
|
import (
|
|
"fmt"
|
|
"os"
|
|
"path/filepath"
|
|
"sort"
|
|
"strings"
|
|
"time"
|
|
|
|
lm "github.com/hrfee/jfa-go/logmessages"
|
|
)
|
|
|
|
const (
|
|
BACKUP_PREFIX = "jfa-go-db-"
|
|
BACKUP_UPLOAD_PREFIX = "upload-"
|
|
BACKUP_DATEFMT = "2006-01-02T15-04-05"
|
|
BACKUP_SUFFIX = ".bak"
|
|
)
|
|
|
|
type BackupList struct {
|
|
files []os.DirEntry
|
|
dates []time.Time
|
|
count int
|
|
}
|
|
|
|
func (bl BackupList) Len() int { return len(bl.files) }
|
|
func (bl BackupList) Swap(i, j int) {
|
|
bl.files[i], bl.files[j] = bl.files[j], bl.files[i]
|
|
bl.dates[i], bl.dates[j] = bl.dates[j], bl.dates[i]
|
|
}
|
|
|
|
func (bl BackupList) Less(i, j int) bool {
|
|
// Push non-backup files to the end of the array,
|
|
// Since they didn't have a date parsed.
|
|
if bl.dates[i].IsZero() {
|
|
return false
|
|
}
|
|
if bl.dates[j].IsZero() {
|
|
return true
|
|
}
|
|
// Sort by oldest first
|
|
return bl.dates[j].After(bl.dates[i])
|
|
}
|
|
|
|
// Get human-readable file size from f.Size() result.
|
|
// https://programming.guide/go/formatting-byte-size-to-human-readable-format.html
|
|
func fileSize(l int64) string {
|
|
const unit = 1000
|
|
if l < unit {
|
|
return fmt.Sprintf("%dB", l)
|
|
}
|
|
div, exp := int64(unit), 0
|
|
for n := l / unit; n >= unit; n /= unit {
|
|
div *= unit
|
|
exp++
|
|
}
|
|
return fmt.Sprintf("%.1f%c", float64(l)/float64(div), "KMGTPE"[exp])
|
|
}
|
|
|
|
func (app *appContext) getBackups() *BackupList {
|
|
path := app.config.Section("backups").Key("path").String()
|
|
err := os.MkdirAll(path, 0755)
|
|
if err != nil {
|
|
app.err.Printf(lm.FailedCreateDir, path, err)
|
|
return nil
|
|
}
|
|
items, err := os.ReadDir(path)
|
|
if err != nil {
|
|
app.err.Printf(lm.FailedReading, path, err)
|
|
return nil
|
|
}
|
|
backups := &BackupList{}
|
|
backups.files = items
|
|
backups.dates = make([]time.Time, len(items))
|
|
backups.count = 0
|
|
for i, item := range items {
|
|
if item.IsDir() || !(strings.HasSuffix(item.Name(), BACKUP_SUFFIX)) {
|
|
continue
|
|
}
|
|
t, err := time.Parse(BACKUP_DATEFMT, strings.TrimSuffix(strings.TrimPrefix(strings.TrimPrefix(item.Name(), BACKUP_UPLOAD_PREFIX), BACKUP_PREFIX), BACKUP_SUFFIX))
|
|
if err != nil {
|
|
app.debug.Printf(lm.FailedParseTime, err)
|
|
continue
|
|
}
|
|
backups.dates[i] = t
|
|
backups.count++
|
|
}
|
|
return backups
|
|
}
|
|
|
|
func (app *appContext) makeBackup() (fileDetails CreateBackupDTO) {
|
|
toKeep := app.config.Section("backups").Key("keep_n_backups").MustInt(20)
|
|
fname := BACKUP_PREFIX + time.Now().Local().Format(BACKUP_DATEFMT) + BACKUP_SUFFIX
|
|
path := app.config.Section("backups").Key("path").String()
|
|
backups := app.getBackups()
|
|
if backups == nil {
|
|
return
|
|
}
|
|
toDelete := backups.count + 1 - toKeep
|
|
// fmt.Printf("toDelete: %d, backCount: %d, keep: %d, length: %d\n", toDelete, backups.count, toKeep, len(backups.files))
|
|
if toDelete > 0 && toDelete <= backups.count {
|
|
sort.Sort(backups)
|
|
for _, item := range backups.files[:toDelete] {
|
|
fullpath := filepath.Join(path, item.Name())
|
|
err := os.Remove(fullpath)
|
|
if err != nil {
|
|
app.err.Printf(lm.FailedDeleteOldBackup, fullpath, err)
|
|
return
|
|
}
|
|
app.debug.Printf(lm.DeleteOldBackup, fullpath)
|
|
}
|
|
}
|
|
fullpath := filepath.Join(path, fname)
|
|
f, err := os.Create(fullpath)
|
|
if err != nil {
|
|
app.err.Printf(lm.FailedOpen, fullpath, err)
|
|
return
|
|
}
|
|
defer f.Close()
|
|
_, err = app.storage.db.Badger().Backup(f, 0)
|
|
if err != nil {
|
|
app.err.Printf(lm.FailedCreateBackup, err)
|
|
return
|
|
}
|
|
|
|
fstat, err := f.Stat()
|
|
if err != nil {
|
|
app.err.Printf(lm.FailedStat, fullpath, err)
|
|
return
|
|
}
|
|
fileDetails.Size = fileSize(fstat.Size())
|
|
fileDetails.Name = fname
|
|
fileDetails.Path = fullpath
|
|
app.debug.Printf(lm.CreateBackup, fileDetails)
|
|
return
|
|
}
|
|
|
|
func (app *appContext) loadPendingBackup() {
|
|
if LOADBAK == "" {
|
|
return
|
|
}
|
|
oldPath := filepath.Join(app.dataPath, "db-"+string(time.Now().Unix())+"-pre-"+filepath.Base(LOADBAK))
|
|
err := os.Rename(app.storage.db_path, oldPath)
|
|
if err != nil {
|
|
app.err.Fatalf(lm.FailedMoveOldDB, oldPath, err)
|
|
}
|
|
app.info.Printf(lm.MoveOldDB, oldPath)
|
|
|
|
app.ConnectDB()
|
|
defer app.storage.db.Close()
|
|
|
|
f, err := os.Open(LOADBAK)
|
|
if err != nil {
|
|
app.err.Fatalf(lm.FailedOpen, LOADBAK, err)
|
|
}
|
|
err = app.storage.db.Badger().Load(f, 256)
|
|
f.Close()
|
|
if err != nil {
|
|
app.err.Fatalf(lm.FailedRestoreDB, LOADBAK, err)
|
|
}
|
|
app.info.Printf(lm.RestoreDB, LOADBAK)
|
|
LOADBAK = ""
|
|
}
|
|
|
|
func newBackupDaemon(app *appContext) *GenericDaemon {
|
|
interval := time.Duration(app.config.Section("backups").Key("every_n_minutes").MustInt(1440)) * time.Minute
|
|
d := NewGenericDaemon(interval, app,
|
|
func(app *appContext) {
|
|
app.makeBackup()
|
|
},
|
|
)
|
|
d.Name("Backup")
|
|
return d
|
|
}
|