| /* SPDX-License-Identifier: GPL-2.0 */ |
| #ifndef _BCACHEFS_CLOCK_H |
| #define _BCACHEFS_CLOCK_H |
| |
| void bch2_io_timer_add(struct io_clock *, struct io_timer *); |
| void bch2_io_timer_del(struct io_clock *, struct io_timer *); |
| void bch2_kthread_io_clock_wait(struct io_clock *, u64, unsigned long); |
| |
| void __bch2_increment_clock(struct io_clock *, u64); |
| |
| static inline void bch2_increment_clock(struct bch_fs *c, u64 sectors, |
| int rw) |
| { |
| struct io_clock *clock = &c->io_clock[rw]; |
| |
| if (unlikely(this_cpu_add_return(*clock->pcpu_buf, sectors) >= |
| IO_CLOCK_PCPU_SECTORS)) |
| __bch2_increment_clock(clock, this_cpu_xchg(*clock->pcpu_buf, 0)); |
| } |
| |
| void bch2_io_clock_schedule_timeout(struct io_clock *, u64); |
| |
| #define bch2_kthread_wait_event_ioclock_timeout(condition, clock, timeout)\ |
| ({ \ |
| long __ret = timeout; \ |
| might_sleep(); \ |
| if (!___wait_cond_timeout(condition)) \ |
| __ret = __wait_event_timeout(wq, condition, timeout); \ |
| __ret; \ |
| }) |
| |
| void bch2_io_timers_to_text(struct printbuf *, struct io_clock *); |
| |
| void bch2_io_clock_exit(struct io_clock *); |
| int bch2_io_clock_init(struct io_clock *); |
| |
| #endif /* _BCACHEFS_CLOCK_H */ |