Commit | Line | Data |
---|---|---|
1da177e4 LT |
1 | #ifndef _ASM_GENERIC_PERCPU_H_ |
2 | #define _ASM_GENERIC_PERCPU_H_ | |
3 | #include <linux/compiler.h> | |
ae1ee11b | 4 | #include <linux/threads.h> |
1da177e4 | 5 | |
acdac872 | 6 | /* |
7 | * Determine the real variable name from the name visible in the | |
8 | * kernel sources. | |
9 | */ | |
10 | #define per_cpu_var(var) per_cpu__##var | |
11 | ||
1da177e4 LT |
12 | #ifdef CONFIG_SMP |
13 | ||
acdac872 | 14 | /* |
15 | * per_cpu_offset() is the offset that has to be added to a | |
16 | * percpu variable to get to the instance for a certain processor. | |
17 | * | |
18 | * Most arches use the __per_cpu_offset array for those offsets but | |
19 | * some arches have their own ways of determining the offset (x86_64, s390). | |
20 | */ | |
21 | #ifndef __per_cpu_offset | |
1da177e4 LT |
22 | extern unsigned long __per_cpu_offset[NR_CPUS]; |
23 | ||
a875a69f | 24 | #define per_cpu_offset(x) (__per_cpu_offset[x]) |
acdac872 | 25 | #endif |
26 | ||
27 | /* | |
28 | * Determine the offset for the currently active processor. | |
29 | * An arch may define __my_cpu_offset to provide a more effective | |
30 | * means of obtaining the offset to the per cpu variables of the | |
31 | * current processor. | |
32 | */ | |
33 | #ifndef __my_cpu_offset | |
34 | #define __my_cpu_offset per_cpu_offset(raw_smp_processor_id()) | |
35 | #define my_cpu_offset per_cpu_offset(smp_processor_id()) | |
36 | #else | |
37 | #define my_cpu_offset __my_cpu_offset | |
38 | #endif | |
39 | ||
40 | /* | |
41 | * Add a offset to a pointer but keep the pointer as is. | |
42 | * | |
43 | * Only S390 provides its own means of moving the pointer. | |
44 | */ | |
45 | #ifndef SHIFT_PERCPU_PTR | |
46 | #define SHIFT_PERCPU_PTR(__p, __offset) RELOC_HIDE((__p), (__offset)) | |
47 | #endif | |
a875a69f | 48 | |
acdac872 | 49 | /* |
50 | * A percpu variable may point to a discarded reghions. The following are | |
51 | * established ways to produce a usable pointer from the percpu variable | |
52 | * offset. | |
53 | */ | |
54 | #define per_cpu(var, cpu) \ | |
55 | (*SHIFT_PERCPU_PTR(&per_cpu_var(var), per_cpu_offset(cpu))) | |
56 | #define __get_cpu_var(var) \ | |
57 | (*SHIFT_PERCPU_PTR(&per_cpu_var(var), my_cpu_offset)) | |
58 | #define __raw_get_cpu_var(var) \ | |
59 | (*SHIFT_PERCPU_PTR(&per_cpu_var(var), __my_cpu_offset)) | |
60 | ||
61 | ||
62 | #ifdef CONFIG_ARCH_SETS_UP_PER_CPU_AREA | |
63 | extern void setup_per_cpu_areas(void); | |
64 | #endif | |
1da177e4 LT |
65 | |
66 | /* A macro to avoid #include hell... */ | |
67 | #define percpu_modcopy(pcpudst, src, size) \ | |
68 | do { \ | |
69 | unsigned int __i; \ | |
0a945022 | 70 | for_each_possible_cpu(__i) \ |
acdac872 | 71 | memcpy((pcpudst)+per_cpu_offset(__i), \ |
394e3902 | 72 | (src), (size)); \ |
1da177e4 LT |
73 | } while (0) |
74 | #else /* ! SMP */ | |
75 | ||
acdac872 | 76 | #define per_cpu(var, cpu) (*((void)(cpu), &per_cpu_var(var))) |
77 | #define __get_cpu_var(var) per_cpu_var(var) | |
78 | #define __raw_get_cpu_var(var) per_cpu_var(var) | |
1da177e4 LT |
79 | |
80 | #endif /* SMP */ | |
81 | ||
acdac872 | 82 | #ifndef PER_CPU_ATTRIBUTES |
83 | #define PER_CPU_ATTRIBUTES | |
84 | #endif | |
85 | ||
86 | #define DECLARE_PER_CPU(type, name) extern PER_CPU_ATTRIBUTES \ | |
87 | __typeof__(type) per_cpu_var(name) | |
1da177e4 | 88 | |
1da177e4 | 89 | #endif /* _ASM_GENERIC_PERCPU_H_ */ |