forked from luck/tmp_suning_uos_patched
cde9f2f420
There is no need to create a new section for these. Consolidate with 32-bit and just use .text. Signed-off-by: Nicholas Piggin <npiggin@gmail.com> Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
546 lines
11 KiB
ArmAsm
546 lines
11 KiB
ArmAsm
/*
|
|
* Special support for eabi and SVR4
|
|
*
|
|
* Copyright (C) 1995, 1996, 1998, 2000, 2001 Free Software Foundation, Inc.
|
|
* Copyright 2008 Freescale Semiconductor, Inc.
|
|
* Written By Michael Meissner
|
|
*
|
|
* Based on gcc/config/rs6000/crtsavres.asm from gcc
|
|
* 64 bit additions from reading the PPC elf64abi document.
|
|
*
|
|
* This file is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License as published by the
|
|
* Free Software Foundation; either version 2, or (at your option) any
|
|
* later version.
|
|
*
|
|
* In addition to the permissions in the GNU General Public License, the
|
|
* Free Software Foundation gives you unlimited permission to link the
|
|
* compiled version of this file with other programs, and to distribute
|
|
* those programs without any restriction coming from the use of this
|
|
* file. (The General Public License restrictions do apply in other
|
|
* respects; for example, they cover modification of the file, and
|
|
* distribution when not linked into another program.)
|
|
*
|
|
* This file is distributed in the hope that it will be useful, but
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; see the file COPYING. If not, write to
|
|
* the Free Software Foundation, 51 Franklin Street, Fifth Floor,
|
|
* Boston, MA 02110-1301, USA.
|
|
*
|
|
* As a special exception, if you link this library with files
|
|
* compiled with GCC to produce an executable, this does not cause
|
|
* the resulting executable to be covered by the GNU General Public License.
|
|
* This exception does not however invalidate any other reasons why
|
|
* the executable file might be covered by the GNU General Public License.
|
|
*/
|
|
|
|
#include <asm/ppc_asm.h>
|
|
|
|
.file "crtsavres.S"
|
|
|
|
#ifdef CONFIG_CC_OPTIMIZE_FOR_SIZE
|
|
|
|
.section ".text"
|
|
|
|
#ifndef CONFIG_PPC64
|
|
|
|
/* Routines for saving integer registers, called by the compiler. */
|
|
/* Called with r11 pointing to the stack header word of the caller of the */
|
|
/* function, just beyond the end of the integer save area. */
|
|
|
|
_GLOBAL(_savegpr_14)
|
|
_GLOBAL(_save32gpr_14)
|
|
stw 14,-72(11) /* save gp registers */
|
|
_GLOBAL(_savegpr_15)
|
|
_GLOBAL(_save32gpr_15)
|
|
stw 15,-68(11)
|
|
_GLOBAL(_savegpr_16)
|
|
_GLOBAL(_save32gpr_16)
|
|
stw 16,-64(11)
|
|
_GLOBAL(_savegpr_17)
|
|
_GLOBAL(_save32gpr_17)
|
|
stw 17,-60(11)
|
|
_GLOBAL(_savegpr_18)
|
|
_GLOBAL(_save32gpr_18)
|
|
stw 18,-56(11)
|
|
_GLOBAL(_savegpr_19)
|
|
_GLOBAL(_save32gpr_19)
|
|
stw 19,-52(11)
|
|
_GLOBAL(_savegpr_20)
|
|
_GLOBAL(_save32gpr_20)
|
|
stw 20,-48(11)
|
|
_GLOBAL(_savegpr_21)
|
|
_GLOBAL(_save32gpr_21)
|
|
stw 21,-44(11)
|
|
_GLOBAL(_savegpr_22)
|
|
_GLOBAL(_save32gpr_22)
|
|
stw 22,-40(11)
|
|
_GLOBAL(_savegpr_23)
|
|
_GLOBAL(_save32gpr_23)
|
|
stw 23,-36(11)
|
|
_GLOBAL(_savegpr_24)
|
|
_GLOBAL(_save32gpr_24)
|
|
stw 24,-32(11)
|
|
_GLOBAL(_savegpr_25)
|
|
_GLOBAL(_save32gpr_25)
|
|
stw 25,-28(11)
|
|
_GLOBAL(_savegpr_26)
|
|
_GLOBAL(_save32gpr_26)
|
|
stw 26,-24(11)
|
|
_GLOBAL(_savegpr_27)
|
|
_GLOBAL(_save32gpr_27)
|
|
stw 27,-20(11)
|
|
_GLOBAL(_savegpr_28)
|
|
_GLOBAL(_save32gpr_28)
|
|
stw 28,-16(11)
|
|
_GLOBAL(_savegpr_29)
|
|
_GLOBAL(_save32gpr_29)
|
|
stw 29,-12(11)
|
|
_GLOBAL(_savegpr_30)
|
|
_GLOBAL(_save32gpr_30)
|
|
stw 30,-8(11)
|
|
_GLOBAL(_savegpr_31)
|
|
_GLOBAL(_save32gpr_31)
|
|
stw 31,-4(11)
|
|
blr
|
|
|
|
/* Routines for restoring integer registers, called by the compiler. */
|
|
/* Called with r11 pointing to the stack header word of the caller of the */
|
|
/* function, just beyond the end of the integer restore area. */
|
|
|
|
_GLOBAL(_restgpr_14)
|
|
_GLOBAL(_rest32gpr_14)
|
|
lwz 14,-72(11) /* restore gp registers */
|
|
_GLOBAL(_restgpr_15)
|
|
_GLOBAL(_rest32gpr_15)
|
|
lwz 15,-68(11)
|
|
_GLOBAL(_restgpr_16)
|
|
_GLOBAL(_rest32gpr_16)
|
|
lwz 16,-64(11)
|
|
_GLOBAL(_restgpr_17)
|
|
_GLOBAL(_rest32gpr_17)
|
|
lwz 17,-60(11)
|
|
_GLOBAL(_restgpr_18)
|
|
_GLOBAL(_rest32gpr_18)
|
|
lwz 18,-56(11)
|
|
_GLOBAL(_restgpr_19)
|
|
_GLOBAL(_rest32gpr_19)
|
|
lwz 19,-52(11)
|
|
_GLOBAL(_restgpr_20)
|
|
_GLOBAL(_rest32gpr_20)
|
|
lwz 20,-48(11)
|
|
_GLOBAL(_restgpr_21)
|
|
_GLOBAL(_rest32gpr_21)
|
|
lwz 21,-44(11)
|
|
_GLOBAL(_restgpr_22)
|
|
_GLOBAL(_rest32gpr_22)
|
|
lwz 22,-40(11)
|
|
_GLOBAL(_restgpr_23)
|
|
_GLOBAL(_rest32gpr_23)
|
|
lwz 23,-36(11)
|
|
_GLOBAL(_restgpr_24)
|
|
_GLOBAL(_rest32gpr_24)
|
|
lwz 24,-32(11)
|
|
_GLOBAL(_restgpr_25)
|
|
_GLOBAL(_rest32gpr_25)
|
|
lwz 25,-28(11)
|
|
_GLOBAL(_restgpr_26)
|
|
_GLOBAL(_rest32gpr_26)
|
|
lwz 26,-24(11)
|
|
_GLOBAL(_restgpr_27)
|
|
_GLOBAL(_rest32gpr_27)
|
|
lwz 27,-20(11)
|
|
_GLOBAL(_restgpr_28)
|
|
_GLOBAL(_rest32gpr_28)
|
|
lwz 28,-16(11)
|
|
_GLOBAL(_restgpr_29)
|
|
_GLOBAL(_rest32gpr_29)
|
|
lwz 29,-12(11)
|
|
_GLOBAL(_restgpr_30)
|
|
_GLOBAL(_rest32gpr_30)
|
|
lwz 30,-8(11)
|
|
_GLOBAL(_restgpr_31)
|
|
_GLOBAL(_rest32gpr_31)
|
|
lwz 31,-4(11)
|
|
blr
|
|
|
|
/* Routines for restoring integer registers, called by the compiler. */
|
|
/* Called with r11 pointing to the stack header word of the caller of the */
|
|
/* function, just beyond the end of the integer restore area. */
|
|
|
|
_GLOBAL(_restgpr_14_x)
|
|
_GLOBAL(_rest32gpr_14_x)
|
|
lwz 14,-72(11) /* restore gp registers */
|
|
_GLOBAL(_restgpr_15_x)
|
|
_GLOBAL(_rest32gpr_15_x)
|
|
lwz 15,-68(11)
|
|
_GLOBAL(_restgpr_16_x)
|
|
_GLOBAL(_rest32gpr_16_x)
|
|
lwz 16,-64(11)
|
|
_GLOBAL(_restgpr_17_x)
|
|
_GLOBAL(_rest32gpr_17_x)
|
|
lwz 17,-60(11)
|
|
_GLOBAL(_restgpr_18_x)
|
|
_GLOBAL(_rest32gpr_18_x)
|
|
lwz 18,-56(11)
|
|
_GLOBAL(_restgpr_19_x)
|
|
_GLOBAL(_rest32gpr_19_x)
|
|
lwz 19,-52(11)
|
|
_GLOBAL(_restgpr_20_x)
|
|
_GLOBAL(_rest32gpr_20_x)
|
|
lwz 20,-48(11)
|
|
_GLOBAL(_restgpr_21_x)
|
|
_GLOBAL(_rest32gpr_21_x)
|
|
lwz 21,-44(11)
|
|
_GLOBAL(_restgpr_22_x)
|
|
_GLOBAL(_rest32gpr_22_x)
|
|
lwz 22,-40(11)
|
|
_GLOBAL(_restgpr_23_x)
|
|
_GLOBAL(_rest32gpr_23_x)
|
|
lwz 23,-36(11)
|
|
_GLOBAL(_restgpr_24_x)
|
|
_GLOBAL(_rest32gpr_24_x)
|
|
lwz 24,-32(11)
|
|
_GLOBAL(_restgpr_25_x)
|
|
_GLOBAL(_rest32gpr_25_x)
|
|
lwz 25,-28(11)
|
|
_GLOBAL(_restgpr_26_x)
|
|
_GLOBAL(_rest32gpr_26_x)
|
|
lwz 26,-24(11)
|
|
_GLOBAL(_restgpr_27_x)
|
|
_GLOBAL(_rest32gpr_27_x)
|
|
lwz 27,-20(11)
|
|
_GLOBAL(_restgpr_28_x)
|
|
_GLOBAL(_rest32gpr_28_x)
|
|
lwz 28,-16(11)
|
|
_GLOBAL(_restgpr_29_x)
|
|
_GLOBAL(_rest32gpr_29_x)
|
|
lwz 29,-12(11)
|
|
_GLOBAL(_restgpr_30_x)
|
|
_GLOBAL(_rest32gpr_30_x)
|
|
lwz 30,-8(11)
|
|
_GLOBAL(_restgpr_31_x)
|
|
_GLOBAL(_rest32gpr_31_x)
|
|
lwz 0,4(11)
|
|
lwz 31,-4(11)
|
|
mtlr 0
|
|
mr 1,11
|
|
blr
|
|
|
|
#ifdef CONFIG_ALTIVEC
|
|
/* Called with r0 pointing just beyond the end of the vector save area. */
|
|
|
|
_GLOBAL(_savevr_20)
|
|
li r11,-192
|
|
stvx v20,r11,r0
|
|
_GLOBAL(_savevr_21)
|
|
li r11,-176
|
|
stvx v21,r11,r0
|
|
_GLOBAL(_savevr_22)
|
|
li r11,-160
|
|
stvx v22,r11,r0
|
|
_GLOBAL(_savevr_23)
|
|
li r11,-144
|
|
stvx v23,r11,r0
|
|
_GLOBAL(_savevr_24)
|
|
li r11,-128
|
|
stvx v24,r11,r0
|
|
_GLOBAL(_savevr_25)
|
|
li r11,-112
|
|
stvx v25,r11,r0
|
|
_GLOBAL(_savevr_26)
|
|
li r11,-96
|
|
stvx v26,r11,r0
|
|
_GLOBAL(_savevr_27)
|
|
li r11,-80
|
|
stvx v27,r11,r0
|
|
_GLOBAL(_savevr_28)
|
|
li r11,-64
|
|
stvx v28,r11,r0
|
|
_GLOBAL(_savevr_29)
|
|
li r11,-48
|
|
stvx v29,r11,r0
|
|
_GLOBAL(_savevr_30)
|
|
li r11,-32
|
|
stvx v30,r11,r0
|
|
_GLOBAL(_savevr_31)
|
|
li r11,-16
|
|
stvx v31,r11,r0
|
|
blr
|
|
|
|
_GLOBAL(_restvr_20)
|
|
li r11,-192
|
|
lvx v20,r11,r0
|
|
_GLOBAL(_restvr_21)
|
|
li r11,-176
|
|
lvx v21,r11,r0
|
|
_GLOBAL(_restvr_22)
|
|
li r11,-160
|
|
lvx v22,r11,r0
|
|
_GLOBAL(_restvr_23)
|
|
li r11,-144
|
|
lvx v23,r11,r0
|
|
_GLOBAL(_restvr_24)
|
|
li r11,-128
|
|
lvx v24,r11,r0
|
|
_GLOBAL(_restvr_25)
|
|
li r11,-112
|
|
lvx v25,r11,r0
|
|
_GLOBAL(_restvr_26)
|
|
li r11,-96
|
|
lvx v26,r11,r0
|
|
_GLOBAL(_restvr_27)
|
|
li r11,-80
|
|
lvx v27,r11,r0
|
|
_GLOBAL(_restvr_28)
|
|
li r11,-64
|
|
lvx v28,r11,r0
|
|
_GLOBAL(_restvr_29)
|
|
li r11,-48
|
|
lvx v29,r11,r0
|
|
_GLOBAL(_restvr_30)
|
|
li r11,-32
|
|
lvx v30,r11,r0
|
|
_GLOBAL(_restvr_31)
|
|
li r11,-16
|
|
lvx v31,r11,r0
|
|
blr
|
|
|
|
#endif /* CONFIG_ALTIVEC */
|
|
|
|
#else /* CONFIG_PPC64 */
|
|
|
|
.globl _savegpr0_14
|
|
_savegpr0_14:
|
|
std r14,-144(r1)
|
|
.globl _savegpr0_15
|
|
_savegpr0_15:
|
|
std r15,-136(r1)
|
|
.globl _savegpr0_16
|
|
_savegpr0_16:
|
|
std r16,-128(r1)
|
|
.globl _savegpr0_17
|
|
_savegpr0_17:
|
|
std r17,-120(r1)
|
|
.globl _savegpr0_18
|
|
_savegpr0_18:
|
|
std r18,-112(r1)
|
|
.globl _savegpr0_19
|
|
_savegpr0_19:
|
|
std r19,-104(r1)
|
|
.globl _savegpr0_20
|
|
_savegpr0_20:
|
|
std r20,-96(r1)
|
|
.globl _savegpr0_21
|
|
_savegpr0_21:
|
|
std r21,-88(r1)
|
|
.globl _savegpr0_22
|
|
_savegpr0_22:
|
|
std r22,-80(r1)
|
|
.globl _savegpr0_23
|
|
_savegpr0_23:
|
|
std r23,-72(r1)
|
|
.globl _savegpr0_24
|
|
_savegpr0_24:
|
|
std r24,-64(r1)
|
|
.globl _savegpr0_25
|
|
_savegpr0_25:
|
|
std r25,-56(r1)
|
|
.globl _savegpr0_26
|
|
_savegpr0_26:
|
|
std r26,-48(r1)
|
|
.globl _savegpr0_27
|
|
_savegpr0_27:
|
|
std r27,-40(r1)
|
|
.globl _savegpr0_28
|
|
_savegpr0_28:
|
|
std r28,-32(r1)
|
|
.globl _savegpr0_29
|
|
_savegpr0_29:
|
|
std r29,-24(r1)
|
|
.globl _savegpr0_30
|
|
_savegpr0_30:
|
|
std r30,-16(r1)
|
|
.globl _savegpr0_31
|
|
_savegpr0_31:
|
|
std r31,-8(r1)
|
|
std r0,16(r1)
|
|
blr
|
|
|
|
.globl _restgpr0_14
|
|
_restgpr0_14:
|
|
ld r14,-144(r1)
|
|
.globl _restgpr0_15
|
|
_restgpr0_15:
|
|
ld r15,-136(r1)
|
|
.globl _restgpr0_16
|
|
_restgpr0_16:
|
|
ld r16,-128(r1)
|
|
.globl _restgpr0_17
|
|
_restgpr0_17:
|
|
ld r17,-120(r1)
|
|
.globl _restgpr0_18
|
|
_restgpr0_18:
|
|
ld r18,-112(r1)
|
|
.globl _restgpr0_19
|
|
_restgpr0_19:
|
|
ld r19,-104(r1)
|
|
.globl _restgpr0_20
|
|
_restgpr0_20:
|
|
ld r20,-96(r1)
|
|
.globl _restgpr0_21
|
|
_restgpr0_21:
|
|
ld r21,-88(r1)
|
|
.globl _restgpr0_22
|
|
_restgpr0_22:
|
|
ld r22,-80(r1)
|
|
.globl _restgpr0_23
|
|
_restgpr0_23:
|
|
ld r23,-72(r1)
|
|
.globl _restgpr0_24
|
|
_restgpr0_24:
|
|
ld r24,-64(r1)
|
|
.globl _restgpr0_25
|
|
_restgpr0_25:
|
|
ld r25,-56(r1)
|
|
.globl _restgpr0_26
|
|
_restgpr0_26:
|
|
ld r26,-48(r1)
|
|
.globl _restgpr0_27
|
|
_restgpr0_27:
|
|
ld r27,-40(r1)
|
|
.globl _restgpr0_28
|
|
_restgpr0_28:
|
|
ld r28,-32(r1)
|
|
.globl _restgpr0_29
|
|
_restgpr0_29:
|
|
ld r0,16(r1)
|
|
ld r29,-24(r1)
|
|
mtlr r0
|
|
ld r30,-16(r1)
|
|
ld r31,-8(r1)
|
|
blr
|
|
|
|
.globl _restgpr0_30
|
|
_restgpr0_30:
|
|
ld r30,-16(r1)
|
|
.globl _restgpr0_31
|
|
_restgpr0_31:
|
|
ld r0,16(r1)
|
|
ld r31,-8(r1)
|
|
mtlr r0
|
|
blr
|
|
|
|
#ifdef CONFIG_ALTIVEC
|
|
/* Called with r0 pointing just beyond the end of the vector save area. */
|
|
|
|
.globl _savevr_20
|
|
_savevr_20:
|
|
li r12,-192
|
|
stvx v20,r12,r0
|
|
.globl _savevr_21
|
|
_savevr_21:
|
|
li r12,-176
|
|
stvx v21,r12,r0
|
|
.globl _savevr_22
|
|
_savevr_22:
|
|
li r12,-160
|
|
stvx v22,r12,r0
|
|
.globl _savevr_23
|
|
_savevr_23:
|
|
li r12,-144
|
|
stvx v23,r12,r0
|
|
.globl _savevr_24
|
|
_savevr_24:
|
|
li r12,-128
|
|
stvx v24,r12,r0
|
|
.globl _savevr_25
|
|
_savevr_25:
|
|
li r12,-112
|
|
stvx v25,r12,r0
|
|
.globl _savevr_26
|
|
_savevr_26:
|
|
li r12,-96
|
|
stvx v26,r12,r0
|
|
.globl _savevr_27
|
|
_savevr_27:
|
|
li r12,-80
|
|
stvx v27,r12,r0
|
|
.globl _savevr_28
|
|
_savevr_28:
|
|
li r12,-64
|
|
stvx v28,r12,r0
|
|
.globl _savevr_29
|
|
_savevr_29:
|
|
li r12,-48
|
|
stvx v29,r12,r0
|
|
.globl _savevr_30
|
|
_savevr_30:
|
|
li r12,-32
|
|
stvx v30,r12,r0
|
|
.globl _savevr_31
|
|
_savevr_31:
|
|
li r12,-16
|
|
stvx v31,r12,r0
|
|
blr
|
|
|
|
.globl _restvr_20
|
|
_restvr_20:
|
|
li r12,-192
|
|
lvx v20,r12,r0
|
|
.globl _restvr_21
|
|
_restvr_21:
|
|
li r12,-176
|
|
lvx v21,r12,r0
|
|
.globl _restvr_22
|
|
_restvr_22:
|
|
li r12,-160
|
|
lvx v22,r12,r0
|
|
.globl _restvr_23
|
|
_restvr_23:
|
|
li r12,-144
|
|
lvx v23,r12,r0
|
|
.globl _restvr_24
|
|
_restvr_24:
|
|
li r12,-128
|
|
lvx v24,r12,r0
|
|
.globl _restvr_25
|
|
_restvr_25:
|
|
li r12,-112
|
|
lvx v25,r12,r0
|
|
.globl _restvr_26
|
|
_restvr_26:
|
|
li r12,-96
|
|
lvx v26,r12,r0
|
|
.globl _restvr_27
|
|
_restvr_27:
|
|
li r12,-80
|
|
lvx v27,r12,r0
|
|
.globl _restvr_28
|
|
_restvr_28:
|
|
li r12,-64
|
|
lvx v28,r12,r0
|
|
.globl _restvr_29
|
|
_restvr_29:
|
|
li r12,-48
|
|
lvx v29,r12,r0
|
|
.globl _restvr_30
|
|
_restvr_30:
|
|
li r12,-32
|
|
lvx v30,r12,r0
|
|
.globl _restvr_31
|
|
_restvr_31:
|
|
li r12,-16
|
|
lvx v31,r12,r0
|
|
blr
|
|
|
|
#endif /* CONFIG_ALTIVEC */
|
|
|
|
#endif /* CONFIG_PPC64 */
|
|
|
|
#endif
|