| // SPDX-License-Identifier: GPL-2.0 |
| #include "bcachefs.h" |
| #include "clock.h" |
| |
| #include <linux/freezer.h> |
| #include <linux/kthread.h> |
| #include <linux/preempt.h> |
| |
| static inline long io_timer_cmp(io_timer_heap *h, |
| struct io_timer *l, |
| struct io_timer *r) |
| { |
| return l->expire - r->expire; |
| } |
| |
| void bch2_io_timer_add(struct io_clock *clock, struct io_timer *timer) |
| { |
| size_t i; |
| |
| spin_lock(&clock->timer_lock); |
| |
| if (time_after_eq((unsigned long) atomic64_read(&clock->now), |
| timer->expire)) { |
| spin_unlock(&clock->timer_lock); |
| timer->fn(timer); |
| return; |
| } |
| |
| for (i = 0; i < clock->timers.used; i++) |
| if (clock->timers.data[i] == timer) |
| goto out; |
| |
| BUG_ON(!heap_add(&clock->timers, timer, io_timer_cmp, NULL)); |
| out: |
| spin_unlock(&clock->timer_lock); |
| } |
| |
| void bch2_io_timer_del(struct io_clock *clock, struct io_timer *timer) |
| { |
| size_t i; |
| |
| spin_lock(&clock->timer_lock); |
| |
| for (i = 0; i < clock->timers.used; i++) |
| if (clock->timers.data[i] == timer) { |
| heap_del(&clock->timers, i, io_timer_cmp, NULL); |
| break; |
| } |
| |
| spin_unlock(&clock->timer_lock); |
| } |
| |
| struct io_clock_wait { |
| struct io_timer io_timer; |
| struct timer_list cpu_timer; |
| struct task_struct *task; |
| int expired; |
| }; |
| |
| static void io_clock_wait_fn(struct io_timer *timer) |
| { |
| struct io_clock_wait *wait = container_of(timer, |
| struct io_clock_wait, io_timer); |
| |
| wait->expired = 1; |
| wake_up_process(wait->task); |
| } |
| |
| static void io_clock_cpu_timeout(struct timer_list *timer) |
| { |
| struct io_clock_wait *wait = container_of(timer, |
| struct io_clock_wait, cpu_timer); |
| |
| wait->expired = 1; |
| wake_up_process(wait->task); |
| } |
| |
| void bch2_io_clock_schedule_timeout(struct io_clock *clock, unsigned long until) |
| { |
| struct io_clock_wait wait; |
| |
| /* XXX: calculate sleep time rigorously */ |
| wait.io_timer.expire = until; |
| wait.io_timer.fn = io_clock_wait_fn; |
| wait.task = current; |
| wait.expired = 0; |
| bch2_io_timer_add(clock, &wait.io_timer); |
| |
| schedule(); |
| |
| bch2_io_timer_del(clock, &wait.io_timer); |
| } |
| |
| void bch2_kthread_io_clock_wait(struct io_clock *clock, |
| unsigned long io_until, |
| unsigned long cpu_timeout) |
| { |
| bool kthread = (current->flags & PF_KTHREAD) != 0; |
| struct io_clock_wait wait; |
| |
| wait.io_timer.expire = io_until; |
| wait.io_timer.fn = io_clock_wait_fn; |
| wait.task = current; |
| wait.expired = 0; |
| bch2_io_timer_add(clock, &wait.io_timer); |
| |
| timer_setup_on_stack(&wait.cpu_timer, io_clock_cpu_timeout, 0); |
| |
| if (cpu_timeout != MAX_SCHEDULE_TIMEOUT) |
| mod_timer(&wait.cpu_timer, cpu_timeout + jiffies); |
| |
| while (1) { |
| set_current_state(TASK_INTERRUPTIBLE); |
| if (kthread && kthread_should_stop()) |
| break; |
| |
| if (wait.expired) |
| break; |
| |
| schedule(); |
| try_to_freeze(); |
| } |
| |
| __set_current_state(TASK_RUNNING); |
| del_timer_sync(&wait.cpu_timer); |
| destroy_timer_on_stack(&wait.cpu_timer); |
| bch2_io_timer_del(clock, &wait.io_timer); |
| } |
| |
| static struct io_timer *get_expired_timer(struct io_clock *clock, |
| unsigned long now) |
| { |
| struct io_timer *ret = NULL; |
| |
| spin_lock(&clock->timer_lock); |
| |
| if (clock->timers.used && |
| time_after_eq(now, clock->timers.data[0]->expire)) |
| heap_pop(&clock->timers, ret, io_timer_cmp, NULL); |
| |
| spin_unlock(&clock->timer_lock); |
| |
| return ret; |
| } |
| |
| void __bch2_increment_clock(struct io_clock *clock, unsigned sectors) |
| { |
| struct io_timer *timer; |
| unsigned long now = atomic64_add_return(sectors, &clock->now); |
| |
| while ((timer = get_expired_timer(clock, now))) |
| timer->fn(timer); |
| } |
| |
| void bch2_io_timers_to_text(struct printbuf *out, struct io_clock *clock) |
| { |
| unsigned long now; |
| unsigned i; |
| |
| out->atomic++; |
| spin_lock(&clock->timer_lock); |
| now = atomic64_read(&clock->now); |
| |
| for (i = 0; i < clock->timers.used; i++) |
| prt_printf(out, "%ps:\t%li\n", |
| clock->timers.data[i]->fn, |
| clock->timers.data[i]->expire - now); |
| spin_unlock(&clock->timer_lock); |
| --out->atomic; |
| } |
| |
| void bch2_io_clock_exit(struct io_clock *clock) |
| { |
| free_heap(&clock->timers); |
| free_percpu(clock->pcpu_buf); |
| } |
| |
| int bch2_io_clock_init(struct io_clock *clock) |
| { |
| atomic64_set(&clock->now, 0); |
| spin_lock_init(&clock->timer_lock); |
| |
| clock->max_slop = IO_CLOCK_PCPU_SECTORS * num_possible_cpus(); |
| |
| clock->pcpu_buf = alloc_percpu(*clock->pcpu_buf); |
| if (!clock->pcpu_buf) |
| return -ENOMEM; |
| |
| if (!init_heap(&clock->timers, NR_IO_TIMERS, GFP_KERNEL)) |
| return -ENOMEM; |
| |
| return 0; |
| } |