| /* SPDX-License-Identifier: MIT */ |
| /* |
| * Copyright © 2018 Intel Corporation |
| */ |
| |
| #ifndef _I915_FIXED_H_ |
| #define _I915_FIXED_H_ |
| |
| #include <linux/bug.h> |
| #include <linux/kernel.h> |
| #include <linux/math64.h> |
| #include <linux/types.h> |
| |
| typedef struct { |
| u32 val; |
| } uint_fixed_16_16_t; |
| |
| #define FP_16_16_MAX ((uint_fixed_16_16_t){ .val = UINT_MAX }) |
| |
| static inline bool is_fixed16_zero(uint_fixed_16_16_t val) |
| { |
| return val.val == 0; |
| } |
| |
| static inline uint_fixed_16_16_t u32_to_fixed16(u32 val) |
| { |
| uint_fixed_16_16_t fp = { .val = val << 16 }; |
| |
| WARN_ON(val > U16_MAX); |
| |
| return fp; |
| } |
| |
| static inline u32 fixed16_to_u32_round_up(uint_fixed_16_16_t fp) |
| { |
| return DIV_ROUND_UP(fp.val, 1 << 16); |
| } |
| |
| static inline u32 fixed16_to_u32(uint_fixed_16_16_t fp) |
| { |
| return fp.val >> 16; |
| } |
| |
| static inline uint_fixed_16_16_t min_fixed16(uint_fixed_16_16_t min1, |
| uint_fixed_16_16_t min2) |
| { |
| uint_fixed_16_16_t min = { .val = min(min1.val, min2.val) }; |
| |
| return min; |
| } |
| |
| static inline uint_fixed_16_16_t max_fixed16(uint_fixed_16_16_t max1, |
| uint_fixed_16_16_t max2) |
| { |
| uint_fixed_16_16_t max = { .val = max(max1.val, max2.val) }; |
| |
| return max; |
| } |
| |
| static inline uint_fixed_16_16_t clamp_u64_to_fixed16(u64 val) |
| { |
| uint_fixed_16_16_t fp = { .val = (u32)val }; |
| |
| WARN_ON(val > U32_MAX); |
| |
| return fp; |
| } |
| |
| static inline u32 div_round_up_fixed16(uint_fixed_16_16_t val, |
| uint_fixed_16_16_t d) |
| { |
| return DIV_ROUND_UP(val.val, d.val); |
| } |
| |
| static inline u32 mul_round_up_u32_fixed16(u32 val, uint_fixed_16_16_t mul) |
| { |
| u64 tmp; |
| |
| tmp = mul_u32_u32(val, mul.val); |
| tmp = DIV_ROUND_UP_ULL(tmp, 1 << 16); |
| WARN_ON(tmp > U32_MAX); |
| |
| return (u32)tmp; |
| } |
| |
| static inline uint_fixed_16_16_t mul_fixed16(uint_fixed_16_16_t val, |
| uint_fixed_16_16_t mul) |
| { |
| u64 tmp; |
| |
| tmp = mul_u32_u32(val.val, mul.val); |
| tmp = tmp >> 16; |
| |
| return clamp_u64_to_fixed16(tmp); |
| } |
| |
| static inline uint_fixed_16_16_t div_fixed16(u32 val, u32 d) |
| { |
| u64 tmp; |
| |
| tmp = (u64)val << 16; |
| tmp = DIV_ROUND_UP_ULL(tmp, d); |
| |
| return clamp_u64_to_fixed16(tmp); |
| } |
| |
| static inline u32 div_round_up_u32_fixed16(u32 val, uint_fixed_16_16_t d) |
| { |
| u64 tmp; |
| |
| tmp = (u64)val << 16; |
| tmp = DIV_ROUND_UP_ULL(tmp, d.val); |
| WARN_ON(tmp > U32_MAX); |
| |
| return (u32)tmp; |
| } |
| |
| static inline uint_fixed_16_16_t mul_u32_fixed16(u32 val, uint_fixed_16_16_t mul) |
| { |
| u64 tmp; |
| |
| tmp = mul_u32_u32(val, mul.val); |
| |
| return clamp_u64_to_fixed16(tmp); |
| } |
| |
| static inline uint_fixed_16_16_t add_fixed16(uint_fixed_16_16_t add1, |
| uint_fixed_16_16_t add2) |
| { |
| u64 tmp; |
| |
| tmp = (u64)add1.val + add2.val; |
| |
| return clamp_u64_to_fixed16(tmp); |
| } |
| |
| static inline uint_fixed_16_16_t add_fixed16_u32(uint_fixed_16_16_t add1, |
| u32 add2) |
| { |
| uint_fixed_16_16_t tmp_add2 = u32_to_fixed16(add2); |
| u64 tmp; |
| |
| tmp = (u64)add1.val + tmp_add2.val; |
| |
| return clamp_u64_to_fixed16(tmp); |
| } |
| |
| #endif /* _I915_FIXED_H_ */ |