2 * Userland implementation of gettimeofday() for 32 bits processes in a
3 * ppc64 kernel for use in the vDSO
5 * Copyright (C) 2004 Benjamin Herrenschmuidt (benh@kernel.crashing.org,
8 * This program is free software; you can redistribute it and/or
9 * modify it under the terms of the GNU General Public License
10 * as published by the Free Software Foundation; either version
11 * 2 of the License, or (at your option) any later version.
13 #include <linux/config.h>
14 #include <asm/processor.h>
15 #include <asm/ppc_asm.h>
17 #include <asm/asm-offsets.h>
18 #include <asm/unistd.h>
22 * Exact prototype of gettimeofday
24 * int __kernel_gettimeofday(struct timeval *tv, struct timezone *tz);
27 V_FUNCTION_BEGIN(__kernel_gettimeofday)
32 mr r10,r3 /* r10 saves tv */
33 mr r11,r4 /* r11 saves tz */
34 bl __get_datapage@local /* get data page */
35 mr r9, r3 /* datapage ptr in r9 */
36 bl __do_get_xsec@local /* get xsec from tb & kernel */
37 bne- 2f /* out of line -> do syscall */
39 /* seconds are xsec >> 20 */
42 stw r5,TVAL32_TV_SEC(r10)
44 /* get remaining xsec and convert to usec. we scale
45 * up remaining xsec by 12 bits and get the top 32 bits
46 * of the multiplication
52 stw r5,TVAL32_TV_USEC(r10)
54 cmpli cr0,r11,0 /* check if tz is NULL */
56 lwz r4,CFG_TZ_MINUTEWEST(r9)/* fill tz */
57 lwz r5,CFG_TZ_DSTTIME(r9)
58 stw r4,TZONE_TZ_MINWEST(r11)
59 stw r5,TZONE_TZ_DSTTIME(r11)
70 li r0,__NR_gettimeofday
74 V_FUNCTION_END(__kernel_gettimeofday)
77 * Exact prototype of clock_gettime()
79 * int __kernel_clock_gettime(clockid_t clock_id, struct timespec *tp);
82 V_FUNCTION_BEGIN(__kernel_clock_gettime)
84 /* Check for supported clock IDs */
85 cmpli cr0,r3,CLOCK_REALTIME
86 cmpli cr1,r3,CLOCK_MONOTONIC
87 cror cr0*4+eq,cr0*4+eq,cr1*4+eq
90 mflr r12 /* r12 saves lr */
92 mr r10,r3 /* r10 saves id */
93 mr r11,r4 /* r11 saves tp */
94 bl __get_datapage@local /* get data page */
95 mr r9,r3 /* datapage ptr in r9 */
96 beq cr1,50f /* if monotonic -> jump there */
102 bl __do_get_xsec@local /* get xsec from tb & kernel */
103 bne- 98f /* out of line -> do syscall */
105 /* seconds are xsec >> 20 */
106 rlwinm r5,r4,12,20,31
108 stw r5,TSPC32_TV_SEC(r11)
110 /* get remaining xsec and convert to nsec. we scale
111 * up remaining xsec by 12 bits and get the top 32 bits
112 * of the multiplication, then we multiply by 1000
119 stw r5,TSPC32_TV_NSEC(r11)
129 50: bl __do_get_xsec@local /* get xsec from tb & kernel */
130 bne- 98f /* out of line -> do syscall */
132 /* seconds are xsec >> 20 */
133 rlwinm r6,r4,12,20,31
136 /* get remaining xsec and convert to nsec. we scale
137 * up remaining xsec by 12 bits and get the top 32 bits
138 * of the multiplication, then we multiply by 1000
146 /* now we must fixup using wall to monotonic. We need to snapshot
147 * that value and do the counter trick again. Fortunately, we still
148 * have the counter value in r8 that was returned by __do_get_xsec.
149 * At this point, r6,r7 contain our sec/nsec values, r3,r4 and r5
153 lwz r3,WTOM_CLOCK_SEC(r9)
154 lwz r4,WTOM_CLOCK_NSEC(r9)
156 /* We now have our result in r3,r4. We create a fake dependency
157 * on that result and re-check the counter
163 lwz r0,(CFG_TB_UPDATE_COUNT+4)(r9)
165 lwz r0,(CFG_TB_UPDATE_COUNT)(r9)
167 cmpl cr0,r8,r0 /* check if updated */
170 /* Calculate and store result. Note that this mimmics the C code,
171 * which may cause funny results if nsec goes negative... is that
176 lis r5,NSEC_PER_SEC@h
177 ori r5,r5,NSEC_PER_SEC@l
186 1: stw r3,TSPC32_TV_SEC(r11)
187 stw r4,TSPC32_TV_NSEC(r11)
202 li r0,__NR_clock_gettime
206 V_FUNCTION_END(__kernel_clock_gettime)
210 * Exact prototype of clock_getres()
212 * int __kernel_clock_getres(clockid_t clock_id, struct timespec *res);
215 V_FUNCTION_BEGIN(__kernel_clock_getres)
217 /* Check for supported clock IDs */
218 cmpwi cr0,r3,CLOCK_REALTIME
219 cmpwi cr1,r3,CLOCK_MONOTONIC
220 cror cr0*4+eq,cr0*4+eq,cr1*4+eq
227 lis r5,CLOCK_REALTIME_RES@h
228 ori r5,r5,CLOCK_REALTIME_RES@l
229 stw r3,TSPC32_TV_SEC(r4)
230 stw r5,TSPC32_TV_NSEC(r4)
237 li r0,__NR_clock_getres
241 V_FUNCTION_END(__kernel_clock_getres)
245 * This is the core of gettimeofday() & friends, it returns the xsec
246 * value in r3 & r4 and expects the datapage ptr (non clobbered)
247 * in r9. clobbers r0,r4,r5,r6,r7,r8.
248 * When returning, r8 contains the counter value that can be reused
249 * by the monotonic clock implementation
253 /* Check for update count & load values. We use the low
254 * order 32 bits of the update count
257 1: lwz r8,(CFG_TB_UPDATE_COUNT+4)(r9)
259 1: lwz r8,(CFG_TB_UPDATE_COUNT)(r9)
261 andi. r0,r8,1 /* pending update ? loop */
263 xor r0,r8,r8 /* create dependency */
266 /* Load orig stamp (offset to TB) */
267 lwz r5,CFG_TB_ORIG_STAMP(r9)
268 lwz r6,(CFG_TB_ORIG_STAMP+4)(r9)
270 /* Get a stable TB value */
277 /* Substract tb orig stamp. If the high part is non-zero, we jump to
278 * the slow path which call the syscall.
279 * If it's ok, then we have our 32 bits tb_ticks value in r7
285 /* Load scale factor & do multiplication */
286 lwz r5,CFG_TB_TO_XS(r9) /* load values */
287 lwz r6,(CFG_TB_TO_XS+4)(r9)
293 /* At this point, we have the scaled xsec value in r4 + XER:CA
294 * we load & add the stamp since epoch
296 lwz r5,CFG_STAMP_XSEC(r9)
297 lwz r6,(CFG_STAMP_XSEC+4)(r9)
301 /* We now have our result in r3,r4. We create a fake dependency
302 * on that result and re-check the counter
308 lwz r0,(CFG_TB_UPDATE_COUNT+4)(r9)
310 lwz r0,(CFG_TB_UPDATE_COUNT)(r9)
312 cmpl cr0,r8,r0 /* check if updated */
315 /* Warning ! The caller expects CR:EQ to be set to indicate a
316 * successful calculation (so it won't fallback to the syscall
317 * method). We have overriden that CR bit in the counter check,
318 * but fortunately, the loop exit condition _is_ CR:EQ set, so
319 * we can exit safely here. If you change this code, be careful
320 * of that side effect.