2 * Hibernation support for x86-64
4 * Distribute under GPLv2
6 * Copyright (c) 2007 Rafael J. Wysocki <rjw@sisk.pl>
7 * Copyright (c) 2002 Pavel Machek <pavel@ucw.cz>
8 * Copyright (c) 2001 Patrick Mochel <mochel@osdl.org>
11 #include <linux/gfp.h>
12 #include <linux/smp.h>
13 #include <linux/suspend.h>
16 #include <asm/proto.h>
18 #include <asm/pgtable.h>
20 #include <asm/suspend.h>
22 /* References to section boundaries */
23 extern __visible
const void __nosave_begin
, __nosave_end
;
25 /* Defined in hibernate_asm_64.S */
26 extern asmlinkage
int restore_image(void);
29 * Address to jump to in the last phase of restore in order to get to the image
30 * kernel's text (this value is passed in the image header).
32 unsigned long restore_jump_address __visible
;
35 * Value of the cr3 register from before the hibernation (this value is passed
36 * in the image header).
38 unsigned long restore_cr3 __visible
;
40 pgd_t
*temp_level4_pgt __visible
;
42 void *relocated_restore_code __visible
;
44 static void *alloc_pgt_page(void *context
)
46 return (void *)get_safe_page(GFP_ATOMIC
);
49 static int set_up_temporary_mappings(void)
51 struct x86_mapping_info info
= {
52 .alloc_pgt_page
= alloc_pgt_page
,
53 .pmd_flag
= __PAGE_KERNEL_LARGE_EXEC
,
54 .kernel_mapping
= true,
56 unsigned long mstart
, mend
;
60 temp_level4_pgt
= (pgd_t
*)get_safe_page(GFP_ATOMIC
);
64 /* It is safe to reuse the original kernel mapping */
65 set_pgd(temp_level4_pgt
+ pgd_index(__START_KERNEL_map
),
66 init_level4_pgt
[pgd_index(__START_KERNEL_map
)]);
68 /* Set up the direct mapping from scratch */
69 for (i
= 0; i
< nr_pfn_mapped
; i
++) {
70 mstart
= pfn_mapped
[i
].start
<< PAGE_SHIFT
;
71 mend
= pfn_mapped
[i
].end
<< PAGE_SHIFT
;
73 result
= kernel_ident_mapping_init(&info
, temp_level4_pgt
,
83 int swsusp_arch_resume(void)
87 /* We have got enough memory and from now on we cannot recover */
88 if ((error
= set_up_temporary_mappings()))
91 relocated_restore_code
= (void *)get_safe_page(GFP_ATOMIC
);
92 if (!relocated_restore_code
)
94 memcpy(relocated_restore_code
, &core_restore_code
,
95 &restore_registers
- &core_restore_code
);
102 * pfn_is_nosave - check if given pfn is in the 'nosave' section
105 int pfn_is_nosave(unsigned long pfn
)
107 unsigned long nosave_begin_pfn
= __pa_symbol(&__nosave_begin
) >> PAGE_SHIFT
;
108 unsigned long nosave_end_pfn
= PAGE_ALIGN(__pa_symbol(&__nosave_end
)) >> PAGE_SHIFT
;
109 return (pfn
>= nosave_begin_pfn
) && (pfn
< nosave_end_pfn
);
112 struct restore_data_record
{
113 unsigned long jump_address
;
118 #define RESTORE_MAGIC 0x0123456789ABCDEFUL
121 * arch_hibernation_header_save - populate the architecture specific part
122 * of a hibernation image header
123 * @addr: address to save the data at
125 int arch_hibernation_header_save(void *addr
, unsigned int max_size
)
127 struct restore_data_record
*rdr
= addr
;
129 if (max_size
< sizeof(struct restore_data_record
))
131 rdr
->jump_address
= restore_jump_address
;
132 rdr
->cr3
= restore_cr3
;
133 rdr
->magic
= RESTORE_MAGIC
;
138 * arch_hibernation_header_restore - read the architecture specific data
139 * from the hibernation image header
140 * @addr: address to read the data from
142 int arch_hibernation_header_restore(void *addr
)
144 struct restore_data_record
*rdr
= addr
;
146 restore_jump_address
= rdr
->jump_address
;
147 restore_cr3
= rdr
->cr3
;
148 return (rdr
->magic
== RESTORE_MAGIC
) ? 0 : -EINVAL
;