forked from luck/tmp_suning_uos_patched
d7c9661115
This chip is no longer being actively developed for (it was superceded by the TILEPro64 in 2008), and in any case the existing compiler and toolchain in the community do not support it. It's unlikely that the kernel works with TILE64 at this point as the configuration has not been tested in years. The support is also awkward as it requires maintaining a significant number of ifdefs. So, just remove it altogether. Signed-off-by: Chris Metcalf <cmetcalf@tilera.com>
270 lines
5.9 KiB
ArmAsm
270 lines
5.9 KiB
ArmAsm
/*
|
|
* Copyright 2010 Tilera Corporation. All Rights Reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation, version 2.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
|
|
* NON INFRINGEMENT. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* copy new kernel into place and then call hv_reexec
|
|
*
|
|
*/
|
|
|
|
#include <linux/linkage.h>
|
|
#include <arch/chip.h>
|
|
#include <asm/page.h>
|
|
#include <hv/hypervisor.h>
|
|
|
|
#undef RELOCATE_NEW_KERNEL_VERBOSE
|
|
|
|
STD_ENTRY(relocate_new_kernel)
|
|
|
|
move r30, r0 /* page list */
|
|
move r31, r1 /* address of page we are on */
|
|
move r32, r2 /* start address of new kernel */
|
|
|
|
shri r1, r1, PAGE_SHIFT
|
|
addi r1, r1, 1
|
|
shli sp, r1, PAGE_SHIFT
|
|
addi sp, sp, -8
|
|
/* we now have a stack (whether we need one or not) */
|
|
|
|
moveli r40, lo16(hv_console_putc)
|
|
auli r40, r40, ha16(hv_console_putc)
|
|
|
|
#ifdef RELOCATE_NEW_KERNEL_VERBOSE
|
|
moveli r0, 'r'
|
|
jalr r40
|
|
|
|
moveli r0, '_'
|
|
jalr r40
|
|
|
|
moveli r0, 'n'
|
|
jalr r40
|
|
|
|
moveli r0, '_'
|
|
jalr r40
|
|
|
|
moveli r0, 'k'
|
|
jalr r40
|
|
|
|
moveli r0, '\n'
|
|
jalr r40
|
|
#endif
|
|
|
|
/*
|
|
* Throughout this code r30 is pointer to the element of page
|
|
* list we are working on.
|
|
*
|
|
* Normally we get to the next element of the page list by
|
|
* incrementing r30 by four. The exception is if the element
|
|
* on the page list is an IND_INDIRECTION in which case we use
|
|
* the element with the low bits masked off as the new value
|
|
* of r30.
|
|
*
|
|
* To get this started, we need the value passed to us (which
|
|
* will always be an IND_INDIRECTION) in memory somewhere with
|
|
* r30 pointing at it. To do that, we push the value passed
|
|
* to us on the stack and make r30 point to it.
|
|
*/
|
|
|
|
sw sp, r30
|
|
move r30, sp
|
|
addi sp, sp, -8
|
|
|
|
/*
|
|
* On TILEPro, we need to flush all tiles' caches, since we may
|
|
* have been doing hash-for-home caching there. Note that we
|
|
* must do this _after_ we're completely done modifying any memory
|
|
* other than our output buffer (which we know is locally cached).
|
|
* We want the caches to be fully clean when we do the reexec,
|
|
* because the hypervisor is going to do this flush again at that
|
|
* point, and we don't want that second flush to overwrite any memory.
|
|
*/
|
|
{
|
|
move r0, zero /* cache_pa */
|
|
move r1, zero
|
|
}
|
|
{
|
|
auli r2, zero, ha16(HV_FLUSH_EVICT_L2) /* cache_control */
|
|
movei r3, -1 /* cache_cpumask; -1 means all client tiles */
|
|
}
|
|
{
|
|
move r4, zero /* tlb_va */
|
|
move r5, zero /* tlb_length */
|
|
}
|
|
{
|
|
move r6, zero /* tlb_pgsize */
|
|
move r7, zero /* tlb_cpumask */
|
|
}
|
|
{
|
|
move r8, zero /* asids */
|
|
moveli r20, lo16(hv_flush_remote)
|
|
}
|
|
{
|
|
move r9, zero /* asidcount */
|
|
auli r20, r20, ha16(hv_flush_remote)
|
|
}
|
|
|
|
jalr r20
|
|
|
|
/* r33 is destination pointer, default to zero */
|
|
|
|
moveli r33, 0
|
|
|
|
.Lloop: lw r10, r30
|
|
|
|
andi r9, r10, 0xf /* low 4 bits tell us what type it is */
|
|
xor r10, r10, r9 /* r10 is now value with low 4 bits stripped */
|
|
|
|
seqi r0, r9, 0x1 /* IND_DESTINATION */
|
|
bzt r0, .Ltry2
|
|
|
|
move r33, r10
|
|
|
|
#ifdef RELOCATE_NEW_KERNEL_VERBOSE
|
|
moveli r0, 'd'
|
|
jalr r40
|
|
#endif
|
|
|
|
addi r30, r30, 4
|
|
j .Lloop
|
|
|
|
.Ltry2:
|
|
seqi r0, r9, 0x2 /* IND_INDIRECTION */
|
|
bzt r0, .Ltry4
|
|
|
|
move r30, r10
|
|
|
|
#ifdef RELOCATE_NEW_KERNEL_VERBOSE
|
|
moveli r0, 'i'
|
|
jalr r40
|
|
#endif
|
|
|
|
j .Lloop
|
|
|
|
.Ltry4:
|
|
seqi r0, r9, 0x4 /* IND_DONE */
|
|
bzt r0, .Ltry8
|
|
|
|
mf
|
|
|
|
#ifdef RELOCATE_NEW_KERNEL_VERBOSE
|
|
moveli r0, 'D'
|
|
jalr r40
|
|
moveli r0, '\n'
|
|
jalr r40
|
|
#endif
|
|
|
|
move r0, r32
|
|
moveli r1, 0 /* arg to hv_reexec is 64 bits */
|
|
|
|
moveli r41, lo16(hv_reexec)
|
|
auli r41, r41, ha16(hv_reexec)
|
|
|
|
jalr r41
|
|
|
|
/* we should not get here */
|
|
|
|
moveli r0, '?'
|
|
jalr r40
|
|
moveli r0, '\n'
|
|
jalr r40
|
|
|
|
j .Lhalt
|
|
|
|
.Ltry8: seqi r0, r9, 0x8 /* IND_SOURCE */
|
|
bz r0, .Lerr /* unknown type */
|
|
|
|
/* copy page at r10 to page at r33 */
|
|
|
|
move r11, r33
|
|
|
|
moveli r0, lo16(PAGE_SIZE)
|
|
auli r0, r0, ha16(PAGE_SIZE)
|
|
add r33, r33, r0
|
|
|
|
/* copy word at r10 to word at r11 until r11 equals r33 */
|
|
|
|
/* We know page size must be multiple of 16, so we can unroll
|
|
* 16 times safely without any edge case checking.
|
|
*
|
|
* Issue a flush of the destination every 16 words to avoid
|
|
* incoherence when starting the new kernel. (Now this is
|
|
* just good paranoia because the hv_reexec call will also
|
|
* take care of this.)
|
|
*/
|
|
|
|
1:
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0; addi r11, r11, 4 }
|
|
{ lw r0, r10; addi r10, r10, 4 }
|
|
{ sw r11, r0 }
|
|
{ flush r11 ; addi r11, r11, 4 }
|
|
|
|
seq r0, r33, r11
|
|
bzt r0, 1b
|
|
|
|
#ifdef RELOCATE_NEW_KERNEL_VERBOSE
|
|
moveli r0, 's'
|
|
jalr r40
|
|
#endif
|
|
|
|
addi r30, r30, 4
|
|
j .Lloop
|
|
|
|
|
|
.Lerr: moveli r0, 'e'
|
|
jalr r40
|
|
moveli r0, 'r'
|
|
jalr r40
|
|
moveli r0, 'r'
|
|
jalr r40
|
|
moveli r0, '\n'
|
|
jalr r40
|
|
.Lhalt:
|
|
moveli r41, lo16(hv_halt)
|
|
auli r41, r41, ha16(hv_halt)
|
|
|
|
jalr r41
|
|
STD_ENDPROC(relocate_new_kernel)
|
|
|
|
.section .rodata,"a"
|
|
|
|
.globl relocate_new_kernel_size
|
|
relocate_new_kernel_size:
|
|
.long .Lend_relocate_new_kernel - relocate_new_kernel
|