job.go 7.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329
  1. package worker
  2. import (
  3. "errors"
  4. "fmt"
  5. "sync"
  6. "sync/atomic"
  7. "time"
  8. tunasync "github.com/tuna/tunasync/internal"
  9. )
  10. // this file contains the workflow of a mirror jb
  11. type ctrlAction uint8
  12. const (
  13. jobStart ctrlAction = iota
  14. jobStop // stop syncing keep the job
  15. jobDisable // disable the job (stops goroutine)
  16. jobRestart // restart syncing
  17. jobPing // ensure the goroutine is alive
  18. jobHalt // worker halts
  19. jobForceStart // ignore concurrent limit
  20. )
  21. type jobMessage struct {
  22. status tunasync.SyncStatus
  23. name string
  24. msg string
  25. schedule bool
  26. }
  27. const (
  28. // empty state
  29. stateNone uint32 = iota
  30. // ready to run, able to schedule
  31. stateReady
  32. // paused by jobStop
  33. statePaused
  34. // disabled by jobDisable
  35. stateDisabled
  36. // worker is halting
  37. stateHalting
  38. )
  39. // use to ensure all jobs are finished before
  40. // worker exit
  41. var jobsDone sync.WaitGroup
  42. type mirrorJob struct {
  43. provider mirrorProvider
  44. ctrlChan chan ctrlAction
  45. disabled chan empty
  46. state uint32
  47. size string
  48. }
  49. func newMirrorJob(provider mirrorProvider) *mirrorJob {
  50. return &mirrorJob{
  51. provider: provider,
  52. ctrlChan: make(chan ctrlAction, 1),
  53. state: stateNone,
  54. }
  55. }
  56. func (m *mirrorJob) Name() string {
  57. return m.provider.Name()
  58. }
  59. func (m *mirrorJob) State() uint32 {
  60. return atomic.LoadUint32(&(m.state))
  61. }
  62. func (m *mirrorJob) SetState(state uint32) {
  63. atomic.StoreUint32(&(m.state), state)
  64. }
  65. func (m *mirrorJob) SetProvider(provider mirrorProvider) error {
  66. s := m.State()
  67. if (s != stateNone) && (s != stateDisabled) {
  68. return fmt.Errorf("Provider cannot be switched when job state is %d", s)
  69. }
  70. m.provider = provider
  71. return nil
  72. }
  73. // runMirrorJob is the goroutine where syncing job runs in
  74. // arguments:
  75. // provider: mirror provider object
  76. // ctrlChan: receives messages from the manager
  77. // managerChan: push messages to the manager, this channel should have a larger buffer
  78. // sempaphore: make sure the concurrent running syncing job won't explode
  79. // TODO: message struct for managerChan
  80. func (m *mirrorJob) Run(managerChan chan<- jobMessage, semaphore chan empty) error {
  81. jobsDone.Add(1)
  82. m.disabled = make(chan empty)
  83. defer func() {
  84. close(m.disabled)
  85. jobsDone.Done()
  86. }()
  87. provider := m.provider
  88. // to make code shorter
  89. runHooks := func(Hooks []jobHook, action func(h jobHook) error, hookname string) error {
  90. for _, hook := range Hooks {
  91. if err := action(hook); err != nil {
  92. logger.Errorf(
  93. "failed at %s hooks for %s: %s",
  94. hookname, m.Name(), err.Error(),
  95. )
  96. managerChan <- jobMessage{
  97. tunasync.Failed, m.Name(),
  98. fmt.Sprintf("error exec hook %s: %s", hookname, err.Error()),
  99. true,
  100. }
  101. return err
  102. }
  103. }
  104. return nil
  105. }
  106. runJobWrapper := func(kill <-chan empty, jobDone chan<- empty) error {
  107. defer close(jobDone)
  108. managerChan <- jobMessage{tunasync.PreSyncing, m.Name(), "", false}
  109. logger.Noticef("start syncing: %s", m.Name())
  110. Hooks := provider.Hooks()
  111. rHooks := []jobHook{}
  112. for i := len(Hooks); i > 0; i-- {
  113. rHooks = append(rHooks, Hooks[i-1])
  114. }
  115. logger.Debug("hooks: pre-job")
  116. err := runHooks(Hooks, func(h jobHook) error { return h.preJob() }, "pre-job")
  117. if err != nil {
  118. return err
  119. }
  120. for retry := 0; retry < provider.Retry(); retry++ {
  121. stopASAP := false // stop job as soon as possible
  122. if retry > 0 {
  123. logger.Noticef("retry syncing: %s, retry: %d", m.Name(), retry)
  124. }
  125. err := runHooks(Hooks, func(h jobHook) error { return h.preExec() }, "pre-exec")
  126. if err != nil {
  127. return err
  128. }
  129. // start syncing
  130. managerChan <- jobMessage{tunasync.Syncing, m.Name(), "", false}
  131. var syncErr error
  132. syncDone := make(chan error, 1)
  133. started := make(chan empty, 10) // we may receive "started" more than one time (e.g. two_stage_rsync)
  134. go func() {
  135. err := provider.Run(started)
  136. syncDone <- err
  137. }()
  138. select { // Wait until provider started or error happened
  139. case err := <-syncDone:
  140. logger.Errorf("failed to start provider %s: %s", m.Name(), err.Error())
  141. syncDone <- err // it will be read again later
  142. case <-started:
  143. logger.Debug("provider started")
  144. }
  145. // Now terminating the provider is feasible
  146. var termErr error
  147. timeout := provider.Timeout()
  148. if timeout <= 0 {
  149. timeout = 100000 * time.Hour // never time out
  150. }
  151. select {
  152. case syncErr = <-syncDone:
  153. logger.Debug("syncing done")
  154. case <-time.After(timeout):
  155. logger.Notice("provider timeout")
  156. stopASAP = true
  157. termErr = provider.Terminate()
  158. syncErr = fmt.Errorf("%s timeout after %v", m.Name(), timeout)
  159. case <-kill:
  160. logger.Debug("received kill")
  161. stopASAP = true
  162. termErr = provider.Terminate()
  163. syncErr = errors.New("killed by manager")
  164. }
  165. if termErr != nil {
  166. logger.Errorf("failed to terminate provider %s: %s", m.Name(), err.Error())
  167. return termErr
  168. }
  169. // post-exec hooks
  170. herr := runHooks(rHooks, func(h jobHook) error { return h.postExec() }, "post-exec")
  171. if herr != nil {
  172. return herr
  173. }
  174. if syncErr == nil {
  175. // syncing success
  176. logger.Noticef("succeeded syncing %s", m.Name())
  177. // post-success hooks
  178. logger.Debug("post-success hooks")
  179. err := runHooks(rHooks, func(h jobHook) error { return h.postSuccess() }, "post-success")
  180. if err != nil {
  181. return err
  182. }
  183. } else {
  184. // syncing failed
  185. logger.Warningf("failed syncing %s: %s", m.Name(), syncErr.Error())
  186. // post-fail hooks
  187. logger.Debug("post-fail hooks")
  188. err := runHooks(rHooks, func(h jobHook) error { return h.postFail() }, "post-fail")
  189. if err != nil {
  190. return err
  191. }
  192. }
  193. if syncErr == nil {
  194. // syncing success
  195. m.size = provider.DataSize()
  196. managerChan <- jobMessage{tunasync.Success, m.Name(), "", (m.State() == stateReady)}
  197. return nil
  198. }
  199. // syncing failed
  200. managerChan <- jobMessage{tunasync.Failed, m.Name(), syncErr.Error(), (retry == provider.Retry()-1) && (m.State() == stateReady)}
  201. // gracefully exit
  202. if stopASAP {
  203. logger.Debug("No retry, exit directly")
  204. return nil
  205. }
  206. // continue to next retry
  207. } // for retry
  208. return nil
  209. }
  210. runJob := func(kill <-chan empty, jobDone chan<- empty, bypassSemaphore <-chan empty) {
  211. select {
  212. case semaphore <- empty{}:
  213. defer func() { <-semaphore }()
  214. runJobWrapper(kill, jobDone)
  215. case <-bypassSemaphore:
  216. logger.Noticef("Concurrent limit ignored by %s", m.Name())
  217. runJobWrapper(kill, jobDone)
  218. case <-kill:
  219. jobDone <- empty{}
  220. return
  221. }
  222. }
  223. bypassSemaphore := make(chan empty, 1)
  224. for {
  225. if m.State() == stateReady {
  226. kill := make(chan empty)
  227. jobDone := make(chan empty)
  228. go runJob(kill, jobDone, bypassSemaphore)
  229. _wait_for_job:
  230. select {
  231. case <-jobDone:
  232. logger.Debug("job done")
  233. case ctrl := <-m.ctrlChan:
  234. switch ctrl {
  235. case jobStop:
  236. m.SetState(statePaused)
  237. close(kill)
  238. <-jobDone
  239. case jobDisable:
  240. m.SetState(stateDisabled)
  241. close(kill)
  242. <-jobDone
  243. return nil
  244. case jobRestart:
  245. m.SetState(stateReady)
  246. close(kill)
  247. <-jobDone
  248. time.Sleep(time.Second) // Restart may fail if the process was not exited yet
  249. continue
  250. case jobForceStart:
  251. select { //non-blocking
  252. default:
  253. case bypassSemaphore <- empty{}:
  254. }
  255. fallthrough
  256. case jobStart:
  257. m.SetState(stateReady)
  258. goto _wait_for_job
  259. case jobHalt:
  260. m.SetState(stateHalting)
  261. close(kill)
  262. <-jobDone
  263. return nil
  264. default:
  265. // TODO: implement this
  266. close(kill)
  267. return nil
  268. }
  269. }
  270. }
  271. ctrl := <-m.ctrlChan
  272. switch ctrl {
  273. case jobStop:
  274. m.SetState(statePaused)
  275. case jobDisable:
  276. m.SetState(stateDisabled)
  277. return nil
  278. case jobForceStart:
  279. select { //non-blocking
  280. default:
  281. case bypassSemaphore <- empty{}:
  282. }
  283. fallthrough
  284. case jobRestart:
  285. fallthrough
  286. case jobStart:
  287. m.SetState(stateReady)
  288. default:
  289. // TODO
  290. return nil
  291. }
  292. }
  293. }