1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3 * Kernel traps/events for Hexagon processor
4 *
5 * Copyright (c) 2010-2014, The Linux Foundation. All rights reserved.
6 */
7
8 #include <linux/init.h>
9 #include <linux/sched/signal.h>
10 #include <linux/sched/debug.h>
11 #include <linux/sched/task_stack.h>
12 #include <linux/module.h>
13 #include <linux/kallsyms.h>
14 #include <linux/kdebug.h>
15 #include <linux/syscalls.h>
16 #include <linux/signal.h>
17 #include <linux/ptrace.h>
18 #include <asm/traps.h>
19 #include <asm/vm_fault.h>
20 #include <asm/syscall.h>
21 #include <asm/registers.h>
22 #include <asm/unistd.h>
23 #include <asm/sections.h>
24 #ifdef CONFIG_KGDB
25 # include <linux/kgdb.h>
26 #endif
27
28 #define TRAP_SYSCALL 1
29 #define TRAP_DEBUG 0xdb
30
31 #ifdef CONFIG_GENERIC_BUG
32 /* Maybe should resemble arch/sh/kernel/traps.c ?? */
is_valid_bugaddr(unsigned long addr)33 int is_valid_bugaddr(unsigned long addr)
34 {
35 return 1;
36 }
37 #endif /* CONFIG_GENERIC_BUG */
38
ex_name(int ex)39 static const char *ex_name(int ex)
40 {
41 switch (ex) {
42 case HVM_GE_C_XPROT:
43 case HVM_GE_C_XUSER:
44 return "Execute protection fault";
45 case HVM_GE_C_RPROT:
46 case HVM_GE_C_RUSER:
47 return "Read protection fault";
48 case HVM_GE_C_WPROT:
49 case HVM_GE_C_WUSER:
50 return "Write protection fault";
51 case HVM_GE_C_XMAL:
52 return "Misaligned instruction";
53 case HVM_GE_C_WREG:
54 return "Multiple writes to same register in packet";
55 case HVM_GE_C_PCAL:
56 return "Program counter values that are not properly aligned";
57 case HVM_GE_C_RMAL:
58 return "Misaligned data load";
59 case HVM_GE_C_WMAL:
60 return "Misaligned data store";
61 case HVM_GE_C_INVI:
62 case HVM_GE_C_PRIVI:
63 return "Illegal instruction";
64 case HVM_GE_C_BUS:
65 return "Precise bus error";
66 case HVM_GE_C_CACHE:
67 return "Cache error";
68
69 case 0xdb:
70 return "Debugger trap";
71
72 default:
73 return "Unrecognized exception";
74 }
75 }
76
do_show_stack(struct task_struct * task,unsigned long * fp,unsigned long ip,const char * loglvl)77 static void do_show_stack(struct task_struct *task, unsigned long *fp,
78 unsigned long ip, const char *loglvl)
79 {
80 int kstack_depth_to_print = 24;
81 unsigned long offset, size;
82 const char *name = NULL;
83 unsigned long *newfp;
84 unsigned long low, high;
85 char tmpstr[128];
86 char *modname;
87 int i;
88
89 if (task == NULL)
90 task = current;
91
92 printk("%sCPU#%d, %s/%d, Call Trace:\n", loglvl, raw_smp_processor_id(),
93 task->comm, task_pid_nr(task));
94
95 if (fp == NULL) {
96 if (task == current) {
97 asm("%0 = r30" : "=r" (fp));
98 } else {
99 fp = (unsigned long *)
100 ((struct hexagon_switch_stack *)
101 task->thread.switch_sp)->fp;
102 }
103 }
104
105 if ((((unsigned long) fp) & 0x3) || ((unsigned long) fp < 0x1000)) {
106 printk("%s-- Corrupt frame pointer %p\n", loglvl, fp);
107 return;
108 }
109
110 /* Saved link reg is one word above FP */
111 if (!ip)
112 ip = *(fp+1);
113
114 /* Expect kernel stack to be in-bounds */
115 low = (unsigned long)task_stack_page(task);
116 high = low + THREAD_SIZE - 8;
117 low += sizeof(struct thread_info);
118
119 for (i = 0; i < kstack_depth_to_print; i++) {
120
121 name = kallsyms_lookup(ip, &size, &offset, &modname, tmpstr);
122
123 printk("%s[%p] 0x%lx: %s + 0x%lx", loglvl, fp, ip, name, offset);
124 if (((unsigned long) fp < low) || (high < (unsigned long) fp))
125 printk(KERN_CONT " (FP out of bounds!)");
126 if (modname)
127 printk(KERN_CONT " [%s] ", modname);
128 printk(KERN_CONT "\n");
129
130 newfp = (unsigned long *) *fp;
131
132 if (((unsigned long) newfp) & 0x3) {
133 printk("%s-- Corrupt frame pointer %p\n", loglvl, newfp);
134 break;
135 }
136
137 /* Attempt to continue past exception. */
138 if (0 == newfp) {
139 struct pt_regs *regs = (struct pt_regs *) (((void *)fp)
140 + 8);
141
142 if (regs->syscall_nr != -1) {
143 printk("%s-- trap0 -- syscall_nr: %ld", loglvl,
144 regs->syscall_nr);
145 printk(KERN_CONT " psp: %lx elr: %lx\n",
146 pt_psp(regs), pt_elr(regs));
147 break;
148 } else {
149 /* really want to see more ... */
150 kstack_depth_to_print += 6;
151 printk("%s-- %s (0x%lx) badva: %lx\n", loglvl,
152 ex_name(pt_cause(regs)), pt_cause(regs),
153 pt_badva(regs));
154 }
155
156 newfp = (unsigned long *) regs->r30;
157 ip = pt_elr(regs);
158 } else {
159 ip = *(newfp + 1);
160 }
161
162 /* If link reg is null, we are done. */
163 if (ip == 0x0)
164 break;
165
166 /* If newfp isn't larger, we're tracing garbage. */
167 if (newfp > fp)
168 fp = newfp;
169 else
170 break;
171 }
172 }
173
show_stack(struct task_struct * task,unsigned long * fp,const char * loglvl)174 void show_stack(struct task_struct *task, unsigned long *fp, const char *loglvl)
175 {
176 /* Saved link reg is one word above FP */
177 do_show_stack(task, fp, 0, loglvl);
178 }
179
die(const char * str,struct pt_regs * regs,long err)180 int die(const char *str, struct pt_regs *regs, long err)
181 {
182 static struct {
183 spinlock_t lock;
184 int counter;
185 } die = {
186 .lock = __SPIN_LOCK_UNLOCKED(die.lock),
187 .counter = 0
188 };
189
190 console_verbose();
191 oops_enter();
192
193 spin_lock_irq(&die.lock);
194 bust_spinlocks(1);
195 printk(KERN_EMERG "Oops: %s[#%d]:\n", str, ++die.counter);
196
197 if (notify_die(DIE_OOPS, str, regs, err, pt_cause(regs), SIGSEGV) ==
198 NOTIFY_STOP)
199 return 1;
200
201 print_modules();
202 show_regs(regs);
203 do_show_stack(current, ®s->r30, pt_elr(regs), KERN_EMERG);
204
205 bust_spinlocks(0);
206 add_taint(TAINT_DIE, LOCKDEP_NOW_UNRELIABLE);
207
208 spin_unlock_irq(&die.lock);
209
210 if (in_interrupt())
211 panic("Fatal exception in interrupt");
212
213 if (panic_on_oops)
214 panic("Fatal exception");
215
216 oops_exit();
217 make_task_dead(err);
218 return 0;
219 }
220
die_if_kernel(char * str,struct pt_regs * regs,long err)221 int die_if_kernel(char *str, struct pt_regs *regs, long err)
222 {
223 if (!user_mode(regs))
224 return die(str, regs, err);
225 else
226 return 0;
227 }
228
229 /*
230 * It's not clear that misaligned fetches are ever recoverable.
231 */
misaligned_instruction(struct pt_regs * regs)232 static void misaligned_instruction(struct pt_regs *regs)
233 {
234 die_if_kernel("Misaligned Instruction", regs, 0);
235 force_sig(SIGBUS);
236 }
237
238 /*
239 * Misaligned loads and stores, on the other hand, can be
240 * emulated, and probably should be, some day. But for now
241 * they will be considered fatal.
242 */
misaligned_data_load(struct pt_regs * regs)243 static void misaligned_data_load(struct pt_regs *regs)
244 {
245 die_if_kernel("Misaligned Data Load", regs, 0);
246 force_sig(SIGBUS);
247 }
248
misaligned_data_store(struct pt_regs * regs)249 static void misaligned_data_store(struct pt_regs *regs)
250 {
251 die_if_kernel("Misaligned Data Store", regs, 0);
252 force_sig(SIGBUS);
253 }
254
illegal_instruction(struct pt_regs * regs)255 static void illegal_instruction(struct pt_regs *regs)
256 {
257 die_if_kernel("Illegal Instruction", regs, 0);
258 force_sig(SIGILL);
259 }
260
261 /*
262 * Precise bus errors may be recoverable with a a retry,
263 * but for now, treat them as irrecoverable.
264 */
precise_bus_error(struct pt_regs * regs)265 static void precise_bus_error(struct pt_regs *regs)
266 {
267 die_if_kernel("Precise Bus Error", regs, 0);
268 force_sig(SIGBUS);
269 }
270
271 /*
272 * If anything is to be done here other than panic,
273 * it will probably be complex and migrate to another
274 * source module. For now, just die.
275 */
cache_error(struct pt_regs * regs)276 static void cache_error(struct pt_regs *regs)
277 {
278 die("Cache Error", regs, 0);
279 }
280
281 /*
282 * General exception handler
283 */
do_genex(struct pt_regs * regs)284 void do_genex(struct pt_regs *regs)
285 {
286 /*
287 * Decode Cause and Dispatch
288 */
289 switch (pt_cause(regs)) {
290 case HVM_GE_C_XPROT:
291 case HVM_GE_C_XUSER:
292 execute_protection_fault(regs);
293 break;
294 case HVM_GE_C_RPROT:
295 case HVM_GE_C_RUSER:
296 read_protection_fault(regs);
297 break;
298 case HVM_GE_C_WPROT:
299 case HVM_GE_C_WUSER:
300 write_protection_fault(regs);
301 break;
302 case HVM_GE_C_XMAL:
303 misaligned_instruction(regs);
304 break;
305 case HVM_GE_C_WREG:
306 illegal_instruction(regs);
307 break;
308 case HVM_GE_C_PCAL:
309 misaligned_instruction(regs);
310 break;
311 case HVM_GE_C_RMAL:
312 misaligned_data_load(regs);
313 break;
314 case HVM_GE_C_WMAL:
315 misaligned_data_store(regs);
316 break;
317 case HVM_GE_C_INVI:
318 case HVM_GE_C_PRIVI:
319 illegal_instruction(regs);
320 break;
321 case HVM_GE_C_BUS:
322 precise_bus_error(regs);
323 break;
324 case HVM_GE_C_CACHE:
325 cache_error(regs);
326 break;
327 default:
328 /* Halt and catch fire */
329 panic("Unrecognized exception 0x%lx\n", pt_cause(regs));
330 break;
331 }
332 }
333
334 /* Indirect system call dispatch */
sys_syscall(void)335 long sys_syscall(void)
336 {
337 printk(KERN_ERR "sys_syscall invoked!\n");
338 return -ENOSYS;
339 }
340
do_trap0(struct pt_regs * regs)341 void do_trap0(struct pt_regs *regs)
342 {
343 syscall_fn syscall;
344
345 switch (pt_cause(regs)) {
346 case TRAP_SYSCALL:
347 /* System call is trap0 #1 */
348
349 /* allow strace to catch syscall args */
350 if (unlikely(test_thread_flag(TIF_SYSCALL_TRACE) &&
351 ptrace_report_syscall_entry(regs)))
352 return; /* return -ENOSYS somewhere? */
353
354 /* Interrupts should be re-enabled for syscall processing */
355 __vmsetie(VM_INT_ENABLE);
356
357 /*
358 * System call number is in r6, arguments in r0..r5.
359 * Fortunately, no Linux syscall has more than 6 arguments,
360 * and Hexagon ABI passes first 6 arguments in registers.
361 * 64-bit arguments are passed in odd/even register pairs.
362 * Fortunately, we have no system calls that take more
363 * than three arguments with more than one 64-bit value.
364 * Should that change, we'd need to redesign to copy
365 * between user and kernel stacks.
366 */
367 regs->syscall_nr = regs->r06;
368
369 /*
370 * GPR R0 carries the first parameter, and is also used
371 * to report the return value. We need a backup of
372 * the user's value in case we need to do a late restart
373 * of the system call.
374 */
375 regs->restart_r0 = regs->r00;
376
377 if ((unsigned long) regs->syscall_nr >= __NR_syscalls) {
378 regs->r00 = -1;
379 } else {
380 syscall = (syscall_fn)
381 (sys_call_table[regs->syscall_nr]);
382 regs->r00 = syscall(regs->r00, regs->r01,
383 regs->r02, regs->r03,
384 regs->r04, regs->r05);
385 }
386
387 /* allow strace to get the syscall return state */
388 if (unlikely(test_thread_flag(TIF_SYSCALL_TRACE)))
389 ptrace_report_syscall_exit(regs, 0);
390
391 break;
392 case TRAP_DEBUG:
393 /* Trap0 0xdb is debug breakpoint */
394 if (user_mode(regs)) {
395 /*
396 * Some architecures add some per-thread state
397 * to distinguish between breakpoint traps and
398 * trace traps. We may want to do that, and
399 * set the si_code value appropriately, or we
400 * may want to use a different trap0 flavor.
401 */
402 force_sig_fault(SIGTRAP, TRAP_BRKPT,
403 (void __user *) pt_elr(regs));
404 } else {
405 #ifdef CONFIG_KGDB
406 kgdb_handle_exception(pt_cause(regs), SIGTRAP,
407 TRAP_BRKPT, regs);
408 #endif
409 }
410 break;
411 }
412 /* Ignore other trap0 codes for now, especially 0 (Angel calls) */
413 }
414
415 /*
416 * Machine check exception handler
417 */
do_machcheck(struct pt_regs * regs)418 void do_machcheck(struct pt_regs *regs)
419 {
420 /* Halt and catch fire */
421 __vmstop();
422 }
423
424 /*
425 * Treat this like the old 0xdb trap.
426 */
427
do_debug_exception(struct pt_regs * regs)428 void do_debug_exception(struct pt_regs *regs)
429 {
430 regs->hvmer.vmest &= ~HVM_VMEST_CAUSE_MSK;
431 regs->hvmer.vmest |= (TRAP_DEBUG << HVM_VMEST_CAUSE_SFT);
432 do_trap0(regs);
433 }
434