/* SPDX-License-Identifier: GPL-2.0 */ #ifndef _ASM_POWERPC_VDSO_GETTIMEOFDAY_H #define _ASM_POWERPC_VDSO_GETTIMEOFDAY_H #ifndef __ASSEMBLY__ #include #include #include #include #define VDSO_HAS_CLOCK_GETRES 1 #define VDSO_HAS_TIME 1 /* * powerpc specific delta calculation. * * This variant removes the masking of the subtraction because the * clocksource mask of all VDSO capable clocksources on powerpc is U64_MAX * which would result in a pointless operation. The compiler cannot * optimize it away as the mask comes from the vdso data and is not compile * time constant. */ #define VDSO_DELTA_NOMASK 1 static __always_inline int do_syscall_2(const unsigned long _r0, const unsigned long _r3, const unsigned long _r4) { register long r0 asm("r0") = _r0; register unsigned long r3 asm("r3") = _r3; register unsigned long r4 asm("r4") = _r4; register int ret asm ("r3"); asm volatile( " sc\n" " bns+ 1f\n" " neg %0, %0\n" "1:\n" : "=r" (ret), "+r" (r4), "+r" (r0) : "r" (r3) : "memory", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12", "cr0", "ctr"); return ret; } static __always_inline int gettimeofday_fallback(struct __kernel_old_timeval *_tv, struct timezone *_tz) { return do_syscall_2(__NR_gettimeofday, (unsigned long)_tv, (unsigned long)_tz); } #ifdef __powerpc64__ static __always_inline int clock_gettime_fallback(clockid_t _clkid, struct __kernel_timespec *_ts) { return do_syscall_2(__NR_clock_gettime, _clkid, (unsigned long)_ts); } static __always_inline int clock_getres_fallback(clockid_t _clkid, struct __kernel_timespec *_ts) { return do_syscall_2(__NR_clock_getres, _clkid, (unsigned long)_ts); } #else #define BUILD_VDSO32 1 static __always_inline int clock_gettime_fallback(clockid_t _clkid, struct __kernel_timespec *_ts) { return do_syscall_2(__NR_clock_gettime64, _clkid, (unsigned long)_ts); } static __always_inline int clock_getres_fallback(clockid_t _clkid, struct __kernel_timespec *_ts) { return do_syscall_2(__NR_clock_getres_time64, _clkid, (unsigned long)_ts); } static __always_inline int clock_gettime32_fallback(clockid_t _clkid, struct old_timespec32 *_ts) { return do_syscall_2(__NR_clock_gettime, _clkid, (unsigned long)_ts); } static __always_inline int clock_getres32_fallback(clockid_t _clkid, struct old_timespec32 *_ts) { return do_syscall_2(__NR_clock_getres, _clkid, (unsigned long)_ts); } #endif static __always_inline u64 __arch_get_hw_counter(s32 clock_mode, const struct vdso_data *vd) { return get_tb(); } const struct vdso_data *__arch_get_vdso_data(void); #ifdef CONFIG_TIME_NS static __always_inline const struct vdso_data *__arch_get_timens_vdso_data(const struct vdso_data *vd) { return (void *)vd + (1U << CONFIG_PAGE_SHIFT); } #endif static inline bool vdso_clocksource_ok(const struct vdso_data *vd) { return true; } #define vdso_clocksource_ok vdso_clocksource_ok #ifndef __powerpc64__ static __always_inline u64 vdso_shift_ns(u64 ns, unsigned long shift) { u32 hi = ns >> 32; u32 lo = ns; lo >>= shift; lo |= hi << (32 - shift); hi >>= shift; if (likely(hi == 0)) return lo; return ((u64)hi << 32) | lo; } #define vdso_shift_ns vdso_shift_ns #endif #ifdef __powerpc64__ int __c_kernel_clock_gettime(clockid_t clock, struct __kernel_timespec *ts, const struct vdso_data *vd); int __c_kernel_clock_getres(clockid_t clock_id, struct __kernel_timespec *res, const struct vdso_data *vd); #else int __c_kernel_clock_gettime(clockid_t clock, struct old_timespec32 *ts, const struct vdso_data *vd); int __c_kernel_clock_gettime64(clockid_t clock, struct __kernel_timespec *ts, const struct vdso_data *vd); int __c_kernel_clock_getres(clockid_t clock_id, struct old_timespec32 *res, const struct vdso_data *vd); #endif int __c_kernel_gettimeofday(struct __kernel_old_timeval *tv, struct timezone *tz, const struct vdso_data *vd); __kernel_old_time_t __c_kernel_time(__kernel_old_time_t *time, const struct vdso_data *vd); #endif /* __ASSEMBLY__ */ #endif /* _ASM_POWERPC_VDSO_GETTIMEOFDAY_H */