1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
|
/* fpumacro.h: FPU related macros.
*
* Copyright (C) 1997 Jakub Jelinek (jj@sunsite.mff.cuni.cz)
*/
#ifndef _SPARC64_FPUMACRO_H
#define _SPARC64_FPUMACRO_H
extern __inline__ void fpsave32(unsigned long *fpregs, unsigned long *fsr)
{
__asm__ __volatile__ ("
wr %%g0, %2, %%asi
stx %%fsr, [%1]
stda %%f0, [%0] %%asi
stda %%f16, [%0 + 64] %%asi
" : : "r" (fpregs), "r" (fsr), "i" (ASI_BLK_P));
}
extern __inline__ void fpload32(unsigned long *fpregs, unsigned long *fsr)
{
__asm__ __volatile__ ("
wr %%g0, %2, %%asi
ldda [%0] %%asi, %%f0
ldda [%0 + 64] %%asi, %%f16
ldx [%1], %%fsr
" : : "r" (fpregs), "r" (fsr), "i" (ASI_BLK_P));
}
extern __inline__ void fpsave(unsigned long *fpregs, unsigned long *fsr)
{
__asm__ __volatile__ ("
wr %%g0, %2, %%asi
stx %%fsr, [%1]
stda %%f0, [%0] %%asi
stda %%f16, [%0 + 64] %%asi
stda %%f32, [%0 + 128] %%asi
stda %%f48, [%0 + 192] %%asi
" : : "r" (fpregs), "r" (fsr), "i" (ASI_BLK_P));
}
extern __inline__ void fpload(unsigned long *fpregs, unsigned long *fsr)
{
__asm__ __volatile__ ("
wr %%g0, %2, %%asi
ldda [%0] %%asi, %%f0
ldda [%0 + 64] %%asi, %%f16
ldda [%0 + 128] %%asi, %%f32
ldda [%0 + 192] %%asi, %%f48
ldx [%1], %%fsr
" : : "r" (fpregs), "r" (fsr), "i" (ASI_BLK_P));
}
#endif /* !(_SPARC64_FPUMACRO_H) */
|