package main import ( "errors" "log" "time" "strconv" "sync/atomic" "database/sql" c "github.com/Azareal/Gosora/common" "github.com/Azareal/Gosora/query_gen" ) // TODO: Name the tasks so we can figure out which one it was when something goes wrong? Or maybe toss it up WithStack down there? func runTasks(tasks []func() error) { for _, task := range tasks { if err := task(); err != nil { c.LogError(err) } } } func startTick() (abort bool) { isDBDown := atomic.LoadInt32(&c.IsDBDown) if err := db.Ping(); err != nil { // TODO: There's a bit of a race here, but it doesn't matter if this error appears multiple times in the logs as it's capped at three times, we just want to cut it down 99% of the time if isDBDown == 0 { db.SetConnMaxLifetime(time.Second) // Drop all the connections and start over c.LogWarning(err) c.LogWarning(errors.New("The database is down")) } atomic.StoreInt32(&c.IsDBDown, 1) return true } if isDBDown == 1 { log.Print("The database is back") } //db.SetConnMaxLifetime(time.Second * 60 * 5) // Make this infinite as the temporary lifetime change will purge the stale connections? db.SetConnMaxLifetime(-1) atomic.StoreInt32(&c.IsDBDown, 0) return false } func runHook(name string) { if err := c.RunTaskHook(name); err != nil { c.LogError(err, "Failed at task '"+name+"'") } } func tickLoop(thumbChan chan bool) { lastDailyStr, err := c.Meta.Get("lastDaily") // TODO: Report this error back correctly... if err != nil && err != sql.ErrNoRows { c.LogError(err) } lastDaily, _ := strconv.ParseInt(lastDailyStr, 10, 64) low := time.Now().Unix() - (60 * 60 * 24) if lastDaily < low { dailies() } // TODO: Write tests for these // Run this goroutine once every half second halfSecondTicker := time.NewTicker(time.Second / 2) secondTicker := time.NewTicker(time.Second) fifteenMinuteTicker := time.NewTicker(15 * time.Minute) hourTicker := time.NewTicker(time.Hour) dailyTicker := time.NewTicker(time.Hour * 24) for { select { case <-halfSecondTicker.C: if startTick() { continue } runHook("before_half_second_tick") runTasks(c.ScheduledHalfSecondTasks) runHook("after_half_second_tick") case <-secondTicker.C: if startTick() { continue } runHook("before_second_tick") go func() { thumbChan <- true }() runTasks(c.ScheduledSecondTasks) // TODO: Stop hard-coding this if err := c.HandleExpiredScheduledGroups(); err != nil { c.LogError(err) } // TODO: Handle delayed moderation tasks // Sync with the database, if there are any changes if err = c.HandleServerSync(); err != nil { c.LogError(err) } // TODO: Manage the TopicStore, UserStore, and ForumStore // TODO: Alert the admin, if CPU usage, RAM usage, or the number of posts in the past second are too high // TODO: Clean-up alerts with no unread matches which are over two weeks old. Move this to a 24 hour task? // TODO: Rescan the static files for changes runHook("after_second_tick") case <-fifteenMinuteTicker.C: if startTick() { continue } runHook("before_fifteen_minute_tick") runTasks(c.ScheduledFifteenMinuteTasks) // TODO: Automatically lock topics, if they're really old, and the associated setting is enabled. // TODO: Publish scheduled posts. runHook("after_fifteen_minute_tick") case <-hourTicker.C: if startTick() { continue } runHook("before_hour_tick") jsToken, err := c.GenerateSafeString(80) if err != nil { c.LogError(err) } c.JSTokenBox.Store(jsToken) c.OldSessionSigningKeyBox.Store(c.SessionSigningKeyBox.Load().(string)) // TODO: We probably don't need this type conversion sessionSigningKey, err := c.GenerateSafeString(80) if err != nil { c.LogError(err) } c.SessionSigningKeyBox.Store(sessionSigningKey) runTasks(c.ScheduledHourTasks) runHook("after_hour_tick") // TODO: Handle the instance going down a lot better case <-dailyTicker.C: dailies() } // TODO: Handle the daily clean-up. } } func asmMatches() { // TODO: Find a more efficient way of doing this acc := qgen.NewAcc() countStmt := acc.Count("activity_stream_matches").Where("asid=?").Prepare() if err := acc.FirstError(); err != nil { c.LogError(err) return } err := acc.Select("activity_stream").Cols("asid").EachInt(func(asid int) error { var count int err := countStmt.QueryRow(asid).Scan(&count) if err != sql.ErrNoRows { return err } if count > 0 { return nil } _, err = qgen.NewAcc().Delete("activity_stream").Where("asid=?").Run(asid) return err }) if err != nil && err != sql.ErrNoRows { c.LogError(err) } } func dailies() { asmMatches() if c.Config.LogPruneCutoff > -1 { f := func(tbl string) { _, err := qgen.NewAcc().Delete(tbl).DateOlderThan("doneAt",c.Config.LogPruneCutoff,"day").Run() if err != nil { c.LogError(err) } } f("login_logs") f("registration_logs") } if c.Config.PostIPCutoff > -1 { // TODO: Use unixtime to remove this MySQLesque logic? f := func(tbl string) { _, err := qgen.NewAcc().Update(tbl).Set("ipaddress='0'").DateOlderThan("createdAt",c.Config.PostIPCutoff,"day").Where("ipaddress!='0'").Exec() if err != nil { c.LogError(err) } } f("topics") f("replies") f("users_replies") // TODO: Find some way of purging the ip data in polls_votes without breaking any anti-cheat measures which might be running... maybe hash it instead? } // TODO: lastActiveAt isn't currently set, so we can't rely on this to purge last_ips of users who haven't been on in a while /*if c.Config.LastIPCutoff == -1 { _, err := qgen.NewAcc().Update("users").Set("last_ip=0").Where("last_ip!=0").Exec() if err != nil { c.LogError(err) } } else */if c.Config.LastIPCutoff > 0 { /*_, err = qgen.NewAcc().Update("users").Set("last_ip='0'").DateOlderThan("lastActiveAt",c.Config.PostIPCutoff,"day").Where("last_ip!='0'").Exec() if err != nil { c.LogError(err) }*/ mon := time.Now().Month() _, err := qgen.NewAcc().Update("users").Set("last_ip=0").Where("last_ip!=0 AND last_ip NOT LIKE '"+strconv.Itoa(int(mon))+"-%'").Exec() if err != nil { c.LogError(err) } } { err := c.Meta.Set("lastDaily", strconv.FormatInt(time.Now().Unix(), 10)) if err != nil { c.LogError(err) } } }