0
0
Fork 0
mirror of https://github.com/go-gitea/gitea synced 2024-11-17 15:31:06 +01:00

Simplify Boost/Pause logic (#18673)

* Simplify Boost/Pause logic

#18658 has added a check to see if we need to boost because there is still work to do
however the check is slightly complex and not ideal. There's no point boosting if
the queue is paused or can't scale. Therefore merge the two selects into one and add
a check to p.paused.

Signed-off-by: Andrew Thornton <art27@cantab.net>

* And on resume add a zeroboost if necessary

Signed-off-by: Andrew Thornton <art27@cantab.net>

* simplify

Signed-off-by: Andrew Thornton <art27@cantab.net>

Co-authored-by: Lauris BH <lauris@nix.lv>
This commit is contained in:
zeripath 2022-02-08 18:53:34 +00:00 committed by GitHub
parent c14ca34d57
commit f8b21ac04a
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23

View file

@ -308,22 +308,18 @@ func (p *WorkerPool) addWorkers(ctx context.Context, cancel context.CancelFunc,
p.cond.Broadcast() p.cond.Broadcast()
cancel() cancel()
} }
if p.hasNoWorkerScaling() {
select { select {
case <-p.baseCtx.Done(): case <-p.baseCtx.Done():
// Don't warn if the baseCtx is shutdown // Don't warn or check for ongoing work if the baseCtx is shutdown
case <-p.paused:
// Don't warn or check for ongoing work if the pool is paused
default: default:
if p.hasNoWorkerScaling() {
log.Warn( log.Warn(
"Queue: %d is configured to be non-scaling and has no workers - this configuration is likely incorrect.\n"+ "Queue: %d is configured to be non-scaling and has no workers - this configuration is likely incorrect.\n"+
"The queue will be paused to prevent data-loss with the assumption that you will add workers and unpause as required.", p.qid) "The queue will be paused to prevent data-loss with the assumption that you will add workers and unpause as required.", p.qid)
}
p.pause() p.pause()
} } else if p.numberOfWorkers == 0 && atomic.LoadInt64(&p.numInQueue) > 0 {
select {
case <-p.baseCtx.Done():
// this worker queue is shut-down don't reboost
default:
if p.numberOfWorkers == 0 && atomic.LoadInt64(&p.numInQueue) > 0 {
// OK there are no workers but... there's still work to be done -> Reboost // OK there are no workers but... there's still work to be done -> Reboost
p.zeroBoost() p.zeroBoost()
// p.lock will be unlocked by zeroBoost // p.lock will be unlocked by zeroBoost
@ -385,14 +381,37 @@ func (p *WorkerPool) pause() {
// Resume resumes the WorkerPool // Resume resumes the WorkerPool
func (p *WorkerPool) Resume() { func (p *WorkerPool) Resume() {
p.lock.Lock() p.lock.Lock() // can't defer unlock because of the zeroBoost at the end
defer p.lock.Unlock()
select { select {
case <-p.resumed: case <-p.resumed:
// already resumed - there's nothing to do
p.lock.Unlock()
return
default: default:
}
p.paused = make(chan struct{}) p.paused = make(chan struct{})
close(p.resumed) close(p.resumed)
// OK now we need to check if we need to add some workers...
if p.numberOfWorkers > 0 || p.hasNoWorkerScaling() || atomic.LoadInt64(&p.numInQueue) == 0 {
// We either have workers, can't scale or there's no work to be done -> so just resume
p.lock.Unlock()
return
} }
// OK we got some work but no workers we need to think about boosting
select {
case <-p.baseCtx.Done():
// don't bother boosting if the baseCtx is done
p.lock.Unlock()
return
default:
}
// OK we'd better add some boost workers!
p.zeroBoost()
// p.zeroBoost will unlock the lock
} }
// CleanUp will drain the remaining contents of the channel // CleanUp will drain the remaining contents of the channel