Blame view

arch/powerpc/kernel/crash_dump.c 4.23 KB
0cc4746ca   Michael Ellerman   [PATCH] powerpc: ...
1
2
3
4
5
6
7
8
9
10
11
12
  /*
   * Routines for doing kexec-based kdump.
   *
   * Copyright (C) 2005, IBM Corp.
   *
   * Created by: Michael Ellerman
   *
   * This source code is licensed under the GNU General Public License,
   * Version 2.  See the file COPYING for more details.
   */
  
  #undef DEBUG
cc5329152   Michael Ellerman   [PATCH] powerpc: ...
13
14
  #include <linux/crash_dump.h>
  #include <linux/bootmem.h>
95f72d1ed   Yinghai Lu   lmb: rename to me...
15
  #include <linux/memblock.h>
aaddd3eac   Michael Ellerman   powerpc: Move cod...
16
  #include <asm/code-patching.h>
0cc4746ca   Michael Ellerman   [PATCH] powerpc: ...
17
  #include <asm/kdump.h>
d9b2b2a27   David S. Miller   [LIB]: Make Power...
18
  #include <asm/prom.h>
0cc4746ca   Michael Ellerman   [PATCH] powerpc: ...
19
  #include <asm/firmware.h>
54c32021e   Michael Ellerman   [PATCH] powerpc: ...
20
  #include <asm/uaccess.h>
d72e063bb   Anton Blanchard   powerpc/kdump: Ov...
21
  #include <asm/rtas.h>
0cc4746ca   Michael Ellerman   [PATCH] powerpc: ...
22
23
24
25
26
27
28
  
  #ifdef DEBUG
  #include <asm/udbg.h>
  #define DBG(fmt...) udbg_printf(fmt)
  #else
  #define DBG(fmt...)
  #endif
0f890c8d2   Suzuki Poulose   powerpc: Rename m...
29
  #ifndef CONFIG_NONSTATIC_KERNEL
d56c3aaa9   Stephen Rothwell   [POWERPC] Fix sec...
30
  void __init reserve_kdump_trampoline(void)
473104134   Michael Ellerman   [PATCH] powerpc: ...
31
  {
95f72d1ed   Yinghai Lu   lmb: rename to me...
32
  	memblock_reserve(0, KDUMP_RESERVE_LIMIT);
473104134   Michael Ellerman   [PATCH] powerpc: ...
33
  }
0cc4746ca   Michael Ellerman   [PATCH] powerpc: ...
34
35
  static void __init create_trampoline(unsigned long addr)
  {
e7a57273c   Michael Ellerman   powerpc: Allow cr...
36
  	unsigned int *p = (unsigned int *)addr;
0cc4746ca   Michael Ellerman   [PATCH] powerpc: ...
37
38
39
40
41
42
43
44
  	/* The maximum range of a single instruction branch, is the current
  	 * instruction's address + (32 MB - 4) bytes. For the trampoline we
  	 * need to branch to current address + 32 MB. So we insert a nop at
  	 * the trampoline address, then the next instruction (+ 4 bytes)
  	 * does a branch to (32 MB - 4). The net effect is that when we
  	 * branch to "addr" we jump to ("addr" + 32 MB). Although it requires
  	 * two instructions it doesn't require any registers.
  	 */
16c57b362   Kumar Gala   powerpc: Unify op...
45
  	patch_instruction(p, PPC_INST_NOP);
e7a57273c   Michael Ellerman   powerpc: Allow cr...
46
  	patch_branch(++p, addr + PHYSICAL_START, 0);
0cc4746ca   Michael Ellerman   [PATCH] powerpc: ...
47
  }
473104134   Michael Ellerman   [PATCH] powerpc: ...
48
  void __init setup_kdump_trampoline(void)
0cc4746ca   Michael Ellerman   [PATCH] powerpc: ...
49
50
  {
  	unsigned long i;
473104134   Michael Ellerman   [PATCH] powerpc: ...
51
52
  	DBG(" -> setup_kdump_trampoline()
  ");
0cc4746ca   Michael Ellerman   [PATCH] powerpc: ...
53
54
55
56
  
  	for (i = KDUMP_TRAMPOLINE_START; i < KDUMP_TRAMPOLINE_END; i += 8) {
  		create_trampoline(i);
  	}
9e4859ef5   Stephen Rothwell   [POWERPC] FWNMI i...
57
  #ifdef CONFIG_PPC_PSERIES
0cc4746ca   Michael Ellerman   [PATCH] powerpc: ...
58
59
  	create_trampoline(__pa(system_reset_fwnmi) - PHYSICAL_START);
  	create_trampoline(__pa(machine_check_fwnmi) - PHYSICAL_START);
9e4859ef5   Stephen Rothwell   [POWERPC] FWNMI i...
60
  #endif /* CONFIG_PPC_PSERIES */
0cc4746ca   Michael Ellerman   [PATCH] powerpc: ...
61

473104134   Michael Ellerman   [PATCH] powerpc: ...
62
63
  	DBG(" <- setup_kdump_trampoline()
  ");
0cc4746ca   Michael Ellerman   [PATCH] powerpc: ...
64
  }
0f890c8d2   Suzuki Poulose   powerpc: Rename m...
65
  #endif /* CONFIG_NONSTATIC_KERNEL */
cc5329152   Michael Ellerman   [PATCH] powerpc: ...
66

cc5329152   Michael Ellerman   [PATCH] powerpc: ...
67
68
69
70
  static int __init parse_savemaxmem(char *p)
  {
  	if (p)
  		saved_max_pfn = (memparse(p, &p) >> PAGE_SHIFT) - 1;
9b41046cd   OGAWA Hirofumi   [PATCH] Don't pas...
71
  	return 1;
cc5329152   Michael Ellerman   [PATCH] powerpc: ...
72
73
  }
  __setup("savemaxmem=", parse_savemaxmem);
54c32021e   Michael Ellerman   [PATCH] powerpc: ...
74

7230ced49   Michael Ellerman   powerpc: Fix /dev...
75
76
77
78
79
80
81
82
83
84
85
86
  
  static size_t copy_oldmem_vaddr(void *vaddr, char *buf, size_t csize,
                                 unsigned long offset, int userbuf)
  {
  	if (userbuf) {
  		if (copy_to_user((char __user *)buf, (vaddr + offset), csize))
  			return -EFAULT;
  	} else
  		memcpy(buf, (vaddr + offset), csize);
  
  	return csize;
  }
40681b95a   Michael Ellerman   [POWERPC] Make do...
87
  /**
54c32021e   Michael Ellerman   [PATCH] powerpc: ...
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
   * copy_oldmem_page - copy one page from "oldmem"
   * @pfn: page frame number to be copied
   * @buf: target memory address for the copy; this can be in kernel address
   *      space or user address space (see @userbuf)
   * @csize: number of bytes to copy
   * @offset: offset in bytes into the page (based on pfn) to begin the copy
   * @userbuf: if set, @buf is in user address space, use copy_to_user(),
   *      otherwise @buf is in kernel address space, use memcpy().
   *
   * Copy a page from "oldmem". For this page, there is no pte mapped
   * in the current kernel. We stitch up a pte, similar to kmap_atomic.
   */
  ssize_t copy_oldmem_page(unsigned long pfn, char *buf,
  			size_t csize, unsigned long offset, int userbuf)
  {
  	void  *vaddr;
  
  	if (!csize)
  		return 0;
bbc8e30f1   Matthew McClintock   powerpc/crashdump...
107
  	csize = min_t(size_t, csize, PAGE_SIZE);
54c32021e   Michael Ellerman   [PATCH] powerpc: ...
108

bbc8e30f1   Matthew McClintock   powerpc/crashdump...
109
  	if ((min_low_pfn < pfn) && (pfn < max_pfn)) {
7230ced49   Michael Ellerman   powerpc: Fix /dev...
110
111
112
113
114
115
116
  		vaddr = __va(pfn << PAGE_SHIFT);
  		csize = copy_oldmem_vaddr(vaddr, buf, csize, offset, userbuf);
  	} else {
  		vaddr = __ioremap(pfn << PAGE_SHIFT, PAGE_SIZE, 0);
  		csize = copy_oldmem_vaddr(vaddr, buf, csize, offset, userbuf);
  		iounmap(vaddr);
  	}
54c32021e   Michael Ellerman   [PATCH] powerpc: ...
117

54c32021e   Michael Ellerman   [PATCH] powerpc: ...
118
119
  	return csize;
  }
d72e063bb   Anton Blanchard   powerpc/kdump: Ov...
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
  
  #ifdef CONFIG_PPC_RTAS
  /*
   * The crashkernel region will almost always overlap the RTAS region, so
   * we have to be careful when shrinking the crashkernel region.
   */
  void crash_free_reserved_phys_range(unsigned long begin, unsigned long end)
  {
  	unsigned long addr;
  	const u32 *basep, *sizep;
  	unsigned int rtas_start = 0, rtas_end = 0;
  
  	basep = of_get_property(rtas.dev, "linux,rtas-base", NULL);
  	sizep = of_get_property(rtas.dev, "rtas-size", NULL);
  
  	if (basep && sizep) {
  		rtas_start = *basep;
  		rtas_end = *basep + *sizep;
  	}
  
  	for (addr = begin; addr < end; addr += PAGE_SIZE) {
  		/* Does this page overlap with the RTAS region? */
  		if (addr <= rtas_end && ((addr + PAGE_SIZE) > rtas_start))
  			continue;
  
  		ClearPageReserved(pfn_to_page(addr >> PAGE_SHIFT));
  		init_page_count(pfn_to_page(addr >> PAGE_SHIFT));
  		free_page((unsigned long)__va(addr));
  		totalram_pages++;
  	}
  }
  #endif