mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-28 01:57:06 +07:00
c75299aea2
Add the minimal includes/declarations to make the header self-contained, and ensure it stays that way. Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk> Signed-off-by: Jani Nikula <jani.nikula@intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/20190626144020.2155-4-jani.nikula@intel.com
149 lines
2.8 KiB
C
149 lines
2.8 KiB
C
/* SPDX-License-Identifier: MIT */
|
|
/*
|
|
* Copyright © 2018 Intel Corporation
|
|
*/
|
|
|
|
#ifndef _I915_FIXED_H_
|
|
#define _I915_FIXED_H_
|
|
|
|
#include <linux/bug.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/math64.h>
|
|
#include <linux/types.h>
|
|
|
|
typedef struct {
|
|
u32 val;
|
|
} uint_fixed_16_16_t;
|
|
|
|
#define FP_16_16_MAX ((uint_fixed_16_16_t){ .val = UINT_MAX })
|
|
|
|
static inline bool is_fixed16_zero(uint_fixed_16_16_t val)
|
|
{
|
|
return val.val == 0;
|
|
}
|
|
|
|
static inline uint_fixed_16_16_t u32_to_fixed16(u32 val)
|
|
{
|
|
uint_fixed_16_16_t fp = { .val = val << 16 };
|
|
|
|
WARN_ON(val > U16_MAX);
|
|
|
|
return fp;
|
|
}
|
|
|
|
static inline u32 fixed16_to_u32_round_up(uint_fixed_16_16_t fp)
|
|
{
|
|
return DIV_ROUND_UP(fp.val, 1 << 16);
|
|
}
|
|
|
|
static inline u32 fixed16_to_u32(uint_fixed_16_16_t fp)
|
|
{
|
|
return fp.val >> 16;
|
|
}
|
|
|
|
static inline uint_fixed_16_16_t min_fixed16(uint_fixed_16_16_t min1,
|
|
uint_fixed_16_16_t min2)
|
|
{
|
|
uint_fixed_16_16_t min = { .val = min(min1.val, min2.val) };
|
|
|
|
return min;
|
|
}
|
|
|
|
static inline uint_fixed_16_16_t max_fixed16(uint_fixed_16_16_t max1,
|
|
uint_fixed_16_16_t max2)
|
|
{
|
|
uint_fixed_16_16_t max = { .val = max(max1.val, max2.val) };
|
|
|
|
return max;
|
|
}
|
|
|
|
static inline uint_fixed_16_16_t clamp_u64_to_fixed16(u64 val)
|
|
{
|
|
uint_fixed_16_16_t fp = { .val = (u32)val };
|
|
|
|
WARN_ON(val > U32_MAX);
|
|
|
|
return fp;
|
|
}
|
|
|
|
static inline u32 div_round_up_fixed16(uint_fixed_16_16_t val,
|
|
uint_fixed_16_16_t d)
|
|
{
|
|
return DIV_ROUND_UP(val.val, d.val);
|
|
}
|
|
|
|
static inline u32 mul_round_up_u32_fixed16(u32 val, uint_fixed_16_16_t mul)
|
|
{
|
|
u64 tmp;
|
|
|
|
tmp = mul_u32_u32(val, mul.val);
|
|
tmp = DIV_ROUND_UP_ULL(tmp, 1 << 16);
|
|
WARN_ON(tmp > U32_MAX);
|
|
|
|
return (u32)tmp;
|
|
}
|
|
|
|
static inline uint_fixed_16_16_t mul_fixed16(uint_fixed_16_16_t val,
|
|
uint_fixed_16_16_t mul)
|
|
{
|
|
u64 tmp;
|
|
|
|
tmp = mul_u32_u32(val.val, mul.val);
|
|
tmp = tmp >> 16;
|
|
|
|
return clamp_u64_to_fixed16(tmp);
|
|
}
|
|
|
|
static inline uint_fixed_16_16_t div_fixed16(u32 val, u32 d)
|
|
{
|
|
u64 tmp;
|
|
|
|
tmp = (u64)val << 16;
|
|
tmp = DIV_ROUND_UP_ULL(tmp, d);
|
|
|
|
return clamp_u64_to_fixed16(tmp);
|
|
}
|
|
|
|
static inline u32 div_round_up_u32_fixed16(u32 val, uint_fixed_16_16_t d)
|
|
{
|
|
u64 tmp;
|
|
|
|
tmp = (u64)val << 16;
|
|
tmp = DIV_ROUND_UP_ULL(tmp, d.val);
|
|
WARN_ON(tmp > U32_MAX);
|
|
|
|
return (u32)tmp;
|
|
}
|
|
|
|
static inline uint_fixed_16_16_t mul_u32_fixed16(u32 val, uint_fixed_16_16_t mul)
|
|
{
|
|
u64 tmp;
|
|
|
|
tmp = mul_u32_u32(val, mul.val);
|
|
|
|
return clamp_u64_to_fixed16(tmp);
|
|
}
|
|
|
|
static inline uint_fixed_16_16_t add_fixed16(uint_fixed_16_16_t add1,
|
|
uint_fixed_16_16_t add2)
|
|
{
|
|
u64 tmp;
|
|
|
|
tmp = (u64)add1.val + add2.val;
|
|
|
|
return clamp_u64_to_fixed16(tmp);
|
|
}
|
|
|
|
static inline uint_fixed_16_16_t add_fixed16_u32(uint_fixed_16_16_t add1,
|
|
u32 add2)
|
|
{
|
|
uint_fixed_16_16_t tmp_add2 = u32_to_fixed16(add2);
|
|
u64 tmp;
|
|
|
|
tmp = (u64)add1.val + tmp_add2.val;
|
|
|
|
return clamp_u64_to_fixed16(tmp);
|
|
}
|
|
|
|
#endif /* _I915_FIXED_H_ */
|