mail archive of the barebox mailing list
 help / color / mirror / Atom feed
From: "Raphaël Poggi" <poggi.raph@gmail.com>
To: Sascha Hauer <s.hauer@pengutronix.de>
Cc: barebox@lists.infradead.org
Subject: Re: [PATCH 0/3] Add basic support for arm64
Date: Mon, 23 May 2016 10:33:29 +0200	[thread overview]
Message-ID: <CACqcpZBSAGG5KF9R+dSNHEvpDS62UNvksZQd6mb9hm0YhrLzxw@mail.gmail.com> (raw)
In-Reply-To: <20160523073228.GM11148@pengutronix.de>

Hi Sascha,

2016-05-23 9:32 GMT+02:00 Sascha Hauer <s.hauer@pengutronix.de>:
> Hi Raphael,
>
> On Fri, May 20, 2016 at 03:11:06AM +0200, Raphael Poggi wrote:
>> Hello,
>>
>> This patch series introduces a basic support for arm64.
>>
>> [PATCH 1/3] arch: add minimal aarch64 support :
>>         Introduce new architecture by creating a new root directory,
>>         I choose this approach for now, because it is simpler for now, maybe later we will merge this in arch/arm.
>
> I just applied this series to a temporary branch, did a cp -r
> arch/arm64/* arch/arm and committed everything, see the result below.
> This of course breaks arm32 support, but it nicely reveals all places
> that need fixup for arm64. How about we proceed like this: We continue
> to work on the patch below. We continuously find proper solutions for the
> different places that need fixup. For every item that is fixed properly
> we apply the resulting patch mainline and rebase the big
> work-in-progress patch ontop of it. This way the patch should
> continuously get smaller until we finally have working arm32 and arm64
> support in a single architecture.
> One of the first things we'll need is ARM32/ARM64 Kconfig options which we
> can use for the different lib/ functions to depend on (maybe create a
> lib32/ and a lib64/ directory?) This should already make the
> work-in-progress patch much smaller. What do you think?

Perfect, I am fine with this approach.


Thanks,
Raphaël


>
> Sascha
>
> ----------------------------8<------------------------------
>
> From 79e852820d19e3620bfe63b87161317e616546d5 Mon Sep 17 00:00:00 2001
> From: Sascha Hauer <s.hauer@pengutronix.de>
> Date: Mon, 23 May 2016 08:47:36 +0200
> Subject: [PATCH] wip
>
> Signed-off-by: Sascha Hauer <s.hauer@pengutronix.de>
> ---
>  arch/arm/Kconfig                           | 276 ++------------
>  arch/arm/Makefile                          | 190 +---------
>  arch/arm/boards/Makefile                   |   1 +
>  arch/arm/boards/virt/Kconfig               |   8 +
>  arch/arm/boards/virt/Makefile              |   1 +
>  arch/arm/boards/virt/env/bin/_update       |  36 ++
>  arch/arm/boards/virt/env/bin/boot          |  38 ++
>  arch/arm/boards/virt/env/bin/init          |  20 +
>  arch/arm/boards/virt/env/bin/update_kernel |   8 +
>  arch/arm/boards/virt/env/bin/update_root   |   8 +
>  arch/arm/boards/virt/env/config            |  38 ++
>  arch/arm/boards/virt/env/init/mtdparts-nor |  11 +
>  arch/arm/boards/virt/init.c                |  67 ++++
>  arch/arm/configs/virt_defconfig            |  53 +++
>  arch/arm/cpu/Kconfig                       | 102 +----
>  arch/arm/cpu/Makefile                      |  29 +-
>  arch/arm/cpu/cache-armv8.S                 | 168 +++++++++
>  arch/arm/cpu/cache.c                       |  69 +---
>  arch/arm/cpu/cpu.c                         |  44 +--
>  arch/arm/cpu/cpuinfo.c                     |  86 ++---
>  arch/arm/cpu/entry.c                       |   5 +-
>  arch/arm/cpu/exceptions.S                  | 313 ++++++----------
>  arch/arm/cpu/interrupts.c                  |  91 ++---
>  arch/arm/cpu/lowlevel.S                    |  89 ++---
>  arch/arm/cpu/mmu.c                         | 578 ++++++++---------------------
>  arch/arm/cpu/mmu.h                         | 155 ++++++++
>  arch/arm/cpu/start.c                       |  12 +-
>  arch/arm/cpu/uncompress.c                  |   2 +-
>  arch/arm/include/asm/barebox-arm.h         |   2 +-
>  arch/arm/include/asm/bitops.h              | 192 ++--------
>  arch/arm/include/asm/boarddata.h           |   5 +
>  arch/arm/include/asm/cache-l2x0.h          |   8 -
>  arch/arm/include/asm/cache.h               |   4 +-
>  arch/arm/include/asm/errata.h              |   9 -
>  arch/arm/include/asm/gic.h                 | 128 +++++++
>  arch/arm/include/asm/mmu.h                 |   6 +-
>  arch/arm/include/asm/pgtable.h             |   5 +-
>  arch/arm/include/asm/ptrace.h              | 111 +-----
>  arch/arm/include/asm/system.h              | 173 +++++----
>  arch/arm/include/asm/system_info.h         |  73 ++--
>  arch/arm/lib/Makefile                      |  13 -
>  arch/arm/lib/armlinux.c                    |   6 -
>  arch/arm/lib/barebox.lds.S                 |   5 +-
>  arch/arm/lib/bootm.c                       | 109 +++---
>  arch/arm/lib/copy_template.S               | 438 +++++++++-------------
>  arch/arm/lib/memcpy.S                      |  84 +++--
>  arch/arm/lib/memset.S                      | 305 +++++++++------
>  arch/arm/lib/runtime-offset.S              |  18 +-
>  arch/arm/mach-virt/Kconfig                 |  15 +
>  arch/arm/mach-virt/Makefile                |   3 +
>  arch/arm/mach-virt/devices.c               |  30 ++
>  arch/arm/mach-virt/include/mach/debug_ll.h |  24 ++
>  arch/arm/mach-virt/include/mach/devices.h  |  13 +
>  arch/arm/mach-virt/lowlevel.c              |  19 +
>  arch/arm/mach-virt/reset.c                 |  24 ++
>  55 files changed, 2008 insertions(+), 2312 deletions(-)
>  create mode 100644 arch/arm/boards/virt/Kconfig
>  create mode 100644 arch/arm/boards/virt/Makefile
>  create mode 100644 arch/arm/boards/virt/env/bin/_update
>  create mode 100644 arch/arm/boards/virt/env/bin/boot
>  create mode 100644 arch/arm/boards/virt/env/bin/init
>  create mode 100644 arch/arm/boards/virt/env/bin/update_kernel
>  create mode 100644 arch/arm/boards/virt/env/bin/update_root
>  create mode 100644 arch/arm/boards/virt/env/config
>  create mode 100644 arch/arm/boards/virt/env/init/mtdparts-nor
>  create mode 100644 arch/arm/boards/virt/init.c
>  create mode 100644 arch/arm/configs/virt_defconfig
>  create mode 100644 arch/arm/cpu/cache-armv8.S
>  create mode 100644 arch/arm/include/asm/boarddata.h
>  create mode 100644 arch/arm/include/asm/gic.h
>  create mode 100644 arch/arm/mach-virt/Kconfig
>  create mode 100644 arch/arm/mach-virt/Makefile
>  create mode 100644 arch/arm/mach-virt/devices.c
>  create mode 100644 arch/arm/mach-virt/include/mach/debug_ll.h
>  create mode 100644 arch/arm/mach-virt/include/mach/devices.h
>  create mode 100644 arch/arm/mach-virt/lowlevel.c
>  create mode 100644 arch/arm/mach-virt/reset.c
>
> diff --git a/arch/arm/Kconfig b/arch/arm/Kconfig
> index 1fc887b..34085f6 100644
> --- a/arch/arm/Kconfig
> +++ b/arch/arm/Kconfig
> @@ -22,15 +22,6 @@ config ARM_USE_COMPRESSED_DTB
>         select UNCOMPRESS
>         select LZO_DECOMPRESS
>
> -config ARCH_BCM283X
> -       bool
> -       select GPIOLIB
> -       select CLKDEV_LOOKUP
> -       select COMMON_CLK
> -       select CLOCKSOURCE_BCM283X
> -       select ARM_AMBA
> -       select HAS_DEBUG_LL
> -
>  menu "System Type"
>
>  config BUILTIN_DTB
> @@ -50,236 +41,18 @@ config BUILTIN_DTB_NAME
>  choice
>         prompt "ARM system type"
>
> -config ARCH_AT91
> -       bool "Atmel AT91"
> -       select GPIOLIB
> -       select CLKDEV_LOOKUP
> -       select HAS_DEBUG_LL
> -       select HAVE_MACH_ARM_HEAD
> -       select HAVE_CLK
> -       select PINCTRL_AT91
> -
> -config ARCH_BCM2835
> -       bool "Broadcom BCM2835 boards"
> -       select ARCH_BCM283X
> -       select CPU_ARM1176
> -
> -config ARCH_BCM2836
> -       bool "Broadcom BCM2836 boards"
> -       select ARCH_BCM283X
> -       select CPU_V7
> -
> -config ARCH_CLPS711X
> -       bool "Cirrus Logic EP711x/EP721x/EP731x"
> -       select CLKDEV_LOOKUP
> -       select CLOCKSOURCE_CLPS711X
> -       select COMMON_CLK
> -       select CPU_32v4T
> -       select GPIOLIB
> -       select MFD_SYSCON
> -
> -config ARCH_DAVINCI
> -       bool "TI Davinci"
> -       select CPU_ARM926T
> -       select HAS_DEBUG_LL
> -       select GPIOLIB
> -
> -config ARCH_DIGIC
> -       bool "Canon DIGIC-based cameras"
> -       select CPU_ARM946E
> -       select HAS_DEBUG_LL
> -       select CLOCKSOURCE_DIGIC
> -       select GPIOLIB
> -       help
> -         Support for Canon's digital cameras that use the DIGIC4 chip.
> -
> -config ARCH_EP93XX
> -       bool "Cirrus Logic EP93xx"
> -       select CPU_ARM920T
> -       select GENERIC_GPIO
> -
> -config ARCH_HIGHBANK
> -       bool "Calxeda Highbank"
> +config ARCH_VIRT
> +       bool "ARM QEMU virt boards"
>         select HAS_DEBUG_LL
> -       select HAS_POWEROFF
> -       select ARCH_HAS_L2X0
> -       select CPU_V7
> +       select CPU_V8
> +       select SYS_SUPPORTS_64BIT_KERNEL
>         select ARM_AMBA
> -       select AMBA_SP804
> -       select CLKDEV_LOOKUP
> -       select COMMON_CLK
> -       select GPIOLIB
> -
> -config ARCH_IMX
> -       bool "Freescale iMX-based"
> -       select GPIOLIB
> -       select COMMON_CLK
> -       select CLKDEV_LOOKUP
> -       select WATCHDOG_IMX_RESET_SOURCE
> -       select HAS_DEBUG_LL
> -
> -config ARCH_MVEBU
> -       bool "Marvell EBU platforms"
> -       select COMMON_CLK
> -       select COMMON_CLK_OF_PROVIDER
> -       select CLKDEV_LOOKUP
> -       select GPIOLIB
> -       select HAS_DEBUG_LL
> -       select HAVE_PBL_MULTI_IMAGES
> -       select HW_HAS_PCI
> -       select MVEBU_MBUS
> -       select OFTREE
> -       select OF_ADDRESS_PCI
> -       select PINCTRL
> -
> -config ARCH_MXS
> -       bool "Freescale i.MX23/28 (mxs) based"
> -       select GPIOLIB
> -       select GENERIC_GPIO
> -       select COMMON_CLK
> -       select CLKDEV_LOOKUP
> -       select HAS_DEBUG_LL
> -
> -config ARCH_NETX
> -       bool "Hilscher NetX based"
> -       select CPU_ARM926T
> -
> -config ARCH_NOMADIK
> -       bool "STMicroelectronics Nomadik"
> -       select CPU_ARM926T
> -       select CLOCKSOURCE_NOMADIK
> -       select HAVE_CLK
> -       help
> -         Support for the Nomadik platform by ST-Ericsson
> -
> -config ARCH_OMAP
> -       bool "TI OMAP"
> -       select HAS_DEBUG_LL
> -       select GPIOLIB
> -
> -config ARCH_PXA
> -       bool "Intel/Marvell PXA based"
> -       select GENERIC_GPIO
> -       select HAS_POWEROFF
> -
> -config ARCH_ROCKCHIP
> -       bool "Rockchip RX3xxx"
> -       select CPU_V7
> -       select ARM_SMP_TWD
> -       select COMMON_CLK
> -       select CLKDEV_LOOKUP
> -       select COMMON_CLK_OF_PROVIDER
> -       select GPIOLIB
> -       select PINCTRL
> -       select PINCTRL_ROCKCHIP
> -       select OFTREE
> -       select HAVE_PBL_MULTI_IMAGES
> -       select HAS_DEBUG_LL
> -       select ARCH_HAS_L2X0
> -
> -config ARCH_SOCFPGA
> -       bool "Altera SOCFPGA cyclone5"
> -       select HAS_DEBUG_LL
> -       select ARM_SMP_TWD
> -       select CPU_V7
> -       select COMMON_CLK
> -       select CLKDEV_LOOKUP
> -       select GPIOLIB
> -       select HAVE_PBL_MULTI_IMAGES
> -       select OFDEVICE if !ARCH_SOCFPGA_XLOAD
> -       select OFTREE if !ARCH_SOCFPGA_XLOAD
> -
> -config ARCH_S3C24xx
> -       bool "Samsung S3C2410, S3C2440"
> -       select ARCH_SAMSUNG
> -       select CPU_ARM920T
> -       select GENERIC_GPIO
> -
> -config ARCH_S5PCxx
> -       bool "Samsung S5PC110, S5PV210"
> -       select ARCH_SAMSUNG
> -       select CPU_V7
> -       select GENERIC_GPIO
> -
> -config ARCH_S3C64xx
> -       bool "Samsung S3C64xx"
> -       select ARCH_SAMSUNG
> -       select CPU_V6
> -       select GENERIC_GPIO
> -
> -config ARCH_VERSATILE
> -       bool "ARM Versatile boards (ARM926EJ-S)"
> -       select GPIOLIB
> -       select HAVE_CLK
> -       select HAS_DEBUG_LL
> -
> -config ARCH_VEXPRESS
> -       bool "ARM Vexpres boards"
> -       select HAS_DEBUG_LL
> -       select CPU_V7
> -       select ARM_AMBA
> -       select AMBA_SP804
> -       select CLKDEV_LOOKUP
> -       select COMMON_CLK
> -
> -config ARCH_TEGRA
> -       bool "NVIDIA Tegra"
> -       select CPU_V7
> -       select HAS_DEBUG_LL
> -       select HW_HAS_PCI
> -       select COMMON_CLK
> -       select COMMON_CLK_OF_PROVIDER
> -       select CLKDEV_LOOKUP
> -       select GPIOLIB
> -       select GPIO_TEGRA
> -       select HAVE_DEFAULT_ENVIRONMENT_NEW
> -       select HAVE_PBL_MULTI_IMAGES
> -       select OFDEVICE
> -       select OFTREE
> -       select RELOCATABLE
> -       select RESET_CONTROLLER
> -       select PINCTRL
> -
> -config ARCH_UEMD
> -       bool "RC Module UEMD Platform"
> -       select CPU_ARM1176
> -       select COMMON_CLK
> -       select COMMON_CLK_OF_PROVIDER
> -       select CLKDEV_LOOKUP
> -       select OFDEVICE
> -       select OFTREE
> -       select CLOCKSOURCE_UEMD
> -       select HAS_DEBUG_LL
> -
> -config ARCH_ZYNQ
> -       bool "Xilinx Zynq-based boards"
> -       select HAS_DEBUG_LL
> +       select HAVE_CONFIGURABLE_MEMORY_LAYOUT
>
>  endchoice
>
>  source arch/arm/cpu/Kconfig
> -source arch/arm/mach-at91/Kconfig
> -source arch/arm/mach-bcm283x/Kconfig
> -source arch/arm/mach-clps711x/Kconfig
> -source arch/arm/mach-davinci/Kconfig
> -source arch/arm/mach-digic/Kconfig
> -source arch/arm/mach-ep93xx/Kconfig
> -source arch/arm/mach-highbank/Kconfig
> -source arch/arm/mach-imx/Kconfig
> -source arch/arm/mach-mxs/Kconfig
> -source arch/arm/mach-mvebu/Kconfig
> -source arch/arm/mach-netx/Kconfig
> -source arch/arm/mach-nomadik/Kconfig
> -source arch/arm/mach-omap/Kconfig
> -source arch/arm/mach-pxa/Kconfig
> -source arch/arm/mach-rockchip/Kconfig
> -source arch/arm/mach-samsung/Kconfig
> -source arch/arm/mach-socfpga/Kconfig
> -source arch/arm/mach-versatile/Kconfig
> -source arch/arm/mach-vexpress/Kconfig
> -source arch/arm/mach-tegra/Kconfig
> -source arch/arm/mach-uemd/Kconfig
> -source arch/arm/mach-zynq/Kconfig
> +source arch/arm/mach-virt/Kconfig
>
>  config ARM_ASM_UNIFIED
>         bool
> @@ -292,20 +65,6 @@ config AEABI
>
>           To use this you need GCC version 4.0.0 or later.
>
> -config THUMB2_BAREBOX
> -       select ARM_ASM_UNIFIED
> -       select AEABI
> -       depends on !ARCH_TEGRA && !ARCH_AT91
> -       depends on CPU_V7 && !CPU_32v4T && !CPU_32v5 && !CPU_32v6
> -       bool "Compile barebox in thumb-2 mode (read help)"
> -       help
> -         This enables compilation of barebox in thumb-2 mode which generates
> -         ~25% smaller binaries. ARM assembly code needs some fixups to be able
> -         to work correctly in thumb-2 mode. the barebox core should have these
> -         fixups since most assembly code is derived from the Kernel. However,
> -         your board lowlevel init code may break in thumb-2 mode. You have been
> -         warned.
> -
>  config ARM_BOARD_APPEND_ATAG
>         bool "Let board specific code to add ATAGs to be passed to the kernel"
>         depends on ARM_LINUX
> @@ -315,6 +74,29 @@ config ARM_BOARD_APPEND_ATAG
>
>  endmenu
>
> +choice
> +       prompt "Barebox code model"
> +       help
> +         You should only select this option if you have a workload that
> +         actually benefits from 64-bit processing or if your machine has
> +         large memory. You will only be presented a single option in this
> +         menu if your system does not support both 32-bit and 64-bit modes.
> +
> +config 32BIT
> +       bool "32-bit barebox"
> +       depends on CPU_SUPPORTS_32BIT_KERNEL && SYS_SUPPORTS_32BIT_KERNEL
> +       help
> +         Select this option if you want to build a 32-bit barebox.
> +
> +config 64BIT
> +       bool "64-bit barebox"
> +       depends on CPU_SUPPORTS_64BIT_KERNEL && SYS_SUPPORTS_64BIT_KERNEL
> +       select ARCH_DMA_ADDR_T_64BIT
> +       help
> +         Select this option if you want to build a 64-bit barebox.
> +
> +endchoice
> +
>  menu "ARM specific settings"
>
>  config ARM_OPTIMZED_STRING_FUNCTIONS
> diff --git a/arch/arm/Makefile b/arch/arm/Makefile
> index 5ccdb83..ad250c4 100644
> --- a/arch/arm/Makefile
> +++ b/arch/arm/Makefile
> @@ -1,7 +1,6 @@
>
>  CPPFLAGS       += -D__ARM__ -fno-strict-aliasing
> -# Explicitly specifiy 32-bit ARM ISA since toolchain default can be -mthumb:
> -CPPFLAGS       +=$(call cc-option,-marm,)
> +CPPFLAGS       +=$(call cc-option,-maarch64,)
>
>  ifeq ($(CONFIG_CPU_BIG_ENDIAN),y)
>  CPPFLAGS       += -mbig-endian
> @@ -13,91 +12,27 @@ AS          += -EL
>  LD             += -EL
>  endif
>
> -# Unaligned access is not supported when MMU is disabled, so given how
> -# at least some of the code would be executed with MMU off, lets be
> -# conservative and instruct the compiler not to generate any unaligned
> -# accesses
> -CFLAGS += -mno-unaligned-access
> -
> -
>  # This selects which instruction set is used.
>  # Note that GCC does not numerically define an architecture version
>  # macro, but instead defines a whole series of macros which makes
>  # testing for a specific architecture or later rather impossible.
> -arch-$(CONFIG_CPU_32v7)                :=-D__LINUX_ARM_ARCH__=7 $(call cc-option,-march=armv7-a,-march=armv5t -Wa$(comma)-march=armv7-a)
> -arch-$(CONFIG_CPU_32v6)            :=-D__LINUX_ARM_ARCH__=6 $(call cc-option,-march=armv6,-march=armv5t -Wa$(comma)-march=armv6)
> -arch-$(CONFIG_CPU_32v5)                :=-D__LINUX_ARM_ARCH__=5 $(call cc-option,-march=armv5te,-march=armv4t)
> -arch-$(CONFIG_CPU_32v4T)       :=-D__LINUX_ARM_ARCH__=4 -march=armv4t
> -
> -# This selects how we optimise for the processor.
> -tune-$(CONFIG_CPU_ARM920T)     :=-mtune=arm9tdmi
> -tune-$(CONFIG_CPU_ARM926T)     :=-mtune=arm9tdmi
> -tune-$(CONFIG_CPU_XSCALE)      :=$(call cc-option,-mtune=xscale,-mtune=strongarm110) -Wa,-mcpu=xscale
> +arch-$(CONFIG_CPU_64v8)                := -D__LINUX_ARM_ARCH__=8 $(call cc-option,-march=armv8-a)
>
> -ifeq ($(CONFIG_AEABI),y)
> -CFLAGS_ABI     :=-mabi=aapcs-linux -mno-thumb-interwork
> -else
> -CFLAGS_ABI     :=$(call cc-option,-mapcs-32,-mabi=apcs-gnu) $(call cc-option,-mno-thumb-interwork,)
> -endif
> +CFLAGS_ABI     :=-mabi=lp64
>
>  ifeq ($(CONFIG_ARM_UNWIND),y)
>  CFLAGS_ABI     +=-funwind-tables
>  endif
>
> -ifeq ($(CONFIG_THUMB2_BAREBOX),y)
> -AFLAGS_AUTOIT  :=$(call as-option,-Wa$(comma)-mimplicit-it=always,-Wa$(comma)-mauto-it)
> -AFLAGS_NOWARN  :=$(call as-option,-Wa$(comma)-mno-warn-deprecated,-Wa$(comma)-W)
> -CFLAGS_THUMB2  :=-mthumb $(AFLAGS_AUTOIT) $(AFLAGS_NOWARN)
> -AFLAGS_THUMB2  :=$(CFLAGS_THUMB2) -Wa$(comma)-mthumb
> -endif
> +CPPFLAGS += $(CFLAGS_ABI) $(arch-y) $(tune-y)
>
> -CPPFLAGS += $(CFLAGS_ABI) $(arch-y) $(tune-y) -msoft-float $(CFLAGS_THUMB2)
> -AFLAGS   += -include asm/unified.h -msoft-float $(AFLAGS_THUMB2)
> +AFLAGS   += -include asm/unified.h
>
>  # Machine directory name.  This list is sorted alphanumerically
>  # by CONFIG_* macro name.
> -machine-$(CONFIG_ARCH_AT91)            := at91
> -machine-$(CONFIG_ARCH_BCM283X)         := bcm283x
> -machine-$(CONFIG_ARCH_CLPS711X)                := clps711x
> -machine-$(CONFIG_ARCH_DAVINCI)         := davinci
> -machine-$(CONFIG_ARCH_DIGIC)           := digic
> -machine-$(CONFIG_ARCH_EP93XX)          := ep93xx
> -machine-$(CONFIG_ARCH_HIGHBANK)                := highbank
> -machine-$(CONFIG_ARCH_IMX)             := imx
> -machine-$(CONFIG_ARCH_MXS)             := mxs
> -machine-$(CONFIG_ARCH_MVEBU)           := mvebu
> -machine-$(CONFIG_ARCH_NOMADIK)         := nomadik
> -machine-$(CONFIG_ARCH_NETX)            := netx
> -machine-$(CONFIG_ARCH_OMAP)            := omap
> -machine-$(CONFIG_ARCH_PXA)             := pxa
> -machine-$(CONFIG_ARCH_ROCKCHIP)                := rockchip
> -machine-$(CONFIG_ARCH_SAMSUNG)         := samsung
> -machine-$(CONFIG_ARCH_SOCFPGA)         := socfpga
> -machine-$(CONFIG_ARCH_VERSATILE)       := versatile
> -machine-$(CONFIG_ARCH_VEXPRESS)                := vexpress
> -machine-$(CONFIG_ARCH_TEGRA)           := tegra
> -machine-$(CONFIG_ARCH_UEMD)            := uemd
> -machine-$(CONFIG_ARCH_ZYNQ)            := zynq
> +machine-$(CONFIG_ARCH_VIRT)            := virt
>
>
> -# Board directory name.  This list is sorted alphanumerically
> -# by CONFIG_* macro name.
> -#
> -# DO NOT ADD NEW ENTRIES TO THIS LIST!
> -# Add to arch/arm/boards/Makefile instead.
> -#
> -# These are here only because they have a board specific config.h.
> -# TODO: Get rid of board specific config.h and move these to
> -# arch/arm/boards/Makefile aswell.
> -board-$(CONFIG_MACH_A9M2410)                   += a9m2410
> -board-$(CONFIG_MACH_A9M2440)                   += a9m2440
> -board-$(CONFIG_MACH_AT91RM9200EK)              += at91rm9200ek
> -board-$(CONFIG_MACH_MINI2440)                  += friendlyarm-mini2440
> -board-$(CONFIG_MACH_MINI6410)                  += friendlyarm-mini6410
> -board-$(CONFIG_MACH_PCM027)                    += phytec-phycore-pxa270
> -board-$(CONFIG_MACH_TINY210)                   += friendlyarm-tiny210
> -board-$(CONFIG_MACH_TINY6410)                  += friendlyarm-tiny6410
> -
>  machdirs := $(patsubst %,arch/arm/mach-%/,$(machine-y))
>
>  ifeq ($(KBUILD_SRC),)
> @@ -141,106 +76,7 @@ endif
>  barebox.s5p: $(KBUILD_BINARY)
>         $(Q)scripts/s5p_cksum $< barebox.s5p
>
> -ifeq ($(CONFIG_ARCH_S5PCxx),y)
> -KBUILD_IMAGE := barebox.s5p
> -endif
> -
> -quiet_cmd_mlo ?= IFT     $@
> -       cmd_mlo ?= scripts/omap_signGP -o MLO -l $(TEXT_BASE) -c $<
> -
> -MLO: $(KBUILD_BINARY)
> -       $(call if_changed,mlo)
> -
> -ifeq ($(CONFIG_OMAP_BUILD_IFT),y)
> -KBUILD_IMAGE := MLO
> -endif
> -
> -quiet_cmd_davinci_ubl_image = UBL-IMG $@
> -      cmd_davinci_ubl_image = set -e; \
> -        scripts/mkublheader $< > $@; \
> -        cat $< >> $@
> -
> -barebox.ubl: $(KBUILD_BINARY) FORCE
> -       $(call if_changed,davinci_ubl_image)
> -
> -ifeq ($(CONFIG_ARCH_DAVINCI),y)
> -KBUILD_IMAGE := barebox.ubl
> -endif
> -
> -quiet_cmd_am35xx_spi_image = SPI-IMG $@
> -      cmd_am35xx_spi_image = scripts/mk-omap-image -s -a $(TEXT_BASE) $< > $@
> -
> -barebox.spi: $(KBUILD_BINARY) FORCE
> -       $(call if_changed,am35xx_spi_image)
> -
> -MLO.spi: MLO FORCE
> -       $(call if_changed,am35xx_spi_image)
> -
> -ifeq ($(CONFIG_OMAP_BUILD_SPI),y)
> -KBUILD_IMAGE := MLO.spi
> -endif
> -
> -quiet_cmd_zynq_image = ZYNQ-IMG $@
> -      cmd_zynq_image = scripts/zynq_mkimage $< $@
> -
> -barebox.zynq: $(KBUILD_BINARY) FORCE
> -       $(call if_changed,zynq_image)
> -
> -ifeq ($(machine-y),zynq)
> -KBUILD_IMAGE := barebox.zynq
> -endif
> -
> -quiet_cmd_canon_a1100_image = DD      $@
> -      cmd_canon_a1100_image = scripts/canon-a1100-image $< $@ || \
> -       echo "WARNING: Couldn't create Canon A1100 image due to previous errors."
> -barebox.canon-a1100.bin: $(KBUILD_BINARY) FORCE
> -       $(call if_changed,canon_a1100_image)
> -
> -ifeq ($(CONFIG_MACH_CANON_A1100),y)
> -KBUILD_IMAGE := barebox.canon-a1100.bin
> -endif
> -
> -KWBIMAGE_OPTS = \
> -       -c -i $(srctree)/$(BOARD)/kwbimage.cfg -d $(TEXT_BASE) -e $(TEXT_BASE)
> -
> -quiet_cmd_kwbimage = KWB     $@
> -      cmd_kwbimage = scripts/kwbimage -p $< $(KWBIMAGE_OPTS) -o $@ || \
> -       echo "WARNING: Couldn't create KWB image due to previous errors."
> -
> -quiet_cmd_kwbimage_uart = KWBUART $@
> -      cmd_kwbimage_uart = scripts/kwbimage -m uart -p $< $(KWBIMAGE_OPTS) -o $@ || \
> -       echo "WARNING Couldn't create KWB image due to previous errors."
> -
> -barebox.kwb: $(KBUILD_BINARY) FORCE
> -       $(call if_changed,kwbimage)
> -
> -barebox.kwbuart: $(KBUILD_BINARY) FORCE
> -       $(call if_changed,kwbimage_uart)
> -
> -ifeq ($(CONFIG_ARCH_MVEBU),y)
> -KBUILD_IMAGE  := barebox.kwb barebox.kwbuart
> -endif
> -
> -barebox.imximg: $(KBUILD_BINARY) FORCE
> -       $(call if_changed,imx_image,$(CFG_$(@F)),)
> -
>  boarddir = $(srctree)/arch/arm/boards
> -imxcfg-$(CONFIG_MACH_FREESCALE_MX53_SMD) += $(boarddir)/freescale-mx53-smd/flash-header.imxcfg
> -imxcfg-$(CONFIG_MACH_CCMX51) += $(boarddir)/ccxmx51/flash-header.imxcfg
> -imxcfg-$(CONFIG_MACH_TX51) += $(boarddir)/karo-tx51/flash-header-karo-tx51.imxcfg
> -imxcfg-$(CONFIG_MACH_GUF_VINCELL) += $(boarddir)/guf-vincell/flash-header.imxcfg
> -imxcfg-$(CONFIG_MACH_EUKREA_CPUIMX51SD) += $(boarddir)/eukrea_cpuimx51/flash-header.imxcfg
> -imxcfg-$(CONFIG_MACH_FREESCALE_MX25_3STACK) += $(boarddir)/freescale-mx25-3ds/flash-header.imxcfg
> -imxcfg-$(CONFIG_MACH_FREESCALE_MX35_3STACK) += $(boarddir)/freescale-mx35-3ds/flash-header.imxcfg
> -imxcfg-$(CONFIG_MACH_TQMA53) += $(boarddir)/tqma53/flash-header.imxcfg
> -imxcfg-$(CONFIG_MACH_EUKREA_CPUIMX25) += $(boarddir)/eukrea_cpuimx25/flash-header.imxcfg
> -imxcfg-$(CONFIG_MACH_EUKREA_CPUIMX35) += $(boarddir)/eukrea_cpuimx35/flash-header.imxcfg
> -imxcfg-$(CONFIG_TX53_REV_1011) += $(boarddir)/karo-tx53/flash-header-tx53-rev1011.imxcfg
> -imxcfg-$(CONFIG_TX53_REV_XX30) += $(boarddir)/karo-tx53/flash-header-tx53-revxx30.imxcfg
> -ifneq ($(imxcfg-y),)
> -CFG_barebox.imximg := $(imxcfg-y)
> -KBUILD_IMAGE  := barebox.imximg
> -endif
>
>  pbl := arch/arm/pbl
>  $(pbl)/zbarebox.S $(pbl)/zbarebox.bin $(pbl)/zbarebox: barebox.bin FORCE
> @@ -249,19 +85,8 @@ $(pbl)/zbarebox.S $(pbl)/zbarebox.bin $(pbl)/zbarebox: barebox.bin FORCE
>  archclean:
>         $(MAKE) $(clean)=$(pbl)
>
> -dts := arch/arm/dts
> -
> -%.dtb: scripts
> -       $(Q)$(MAKE) $(build)=$(dts) $(dts)/$@
> -
>  KBUILD_IMAGE ?= $(KBUILD_BINARY)
>
> -archprepare: maketools
> -maketools:
> -       $(Q)$(MAKE) $(build)=arch/arm/tools include/generated/mach-types.h
> -
> -PHONY += maketools
> -
>  ifneq ($(board-y),)
>  BOARD := arch/arm/boards/$(board-y)/
>  else
> @@ -276,9 +101,6 @@ endif
>
>  common-y += $(BOARD) arch/arm/boards/ $(MACH)
>  common-y += arch/arm/lib/ arch/arm/cpu/
> -common-y += arch/arm/crypto/
> -
> -common-$(CONFIG_OFTREE) += arch/arm/dts/
>
>  lds-y  := arch/arm/lib/barebox.lds
>
> diff --git a/arch/arm/boards/Makefile b/arch/arm/boards/Makefile
> index 9241b66..f9cb059 100644
> --- a/arch/arm/boards/Makefile
> +++ b/arch/arm/boards/Makefile
> @@ -135,3 +135,4 @@ obj-$(CONFIG_MACH_VIRT2REAL)                        += virt2real/
>  obj-$(CONFIG_MACH_ZEDBOARD)                    += avnet-zedboard/
>  obj-$(CONFIG_MACH_ZYLONITE)                    += zylonite/
>  obj-$(CONFIG_MACH_VARISCITE_MX6)               += variscite-mx6/
> +obj-$(CONFIG_MACH_VIRT)                                += virt/
> diff --git a/arch/arm/boards/virt/Kconfig b/arch/arm/boards/virt/Kconfig
> new file mode 100644
> index 0000000..b239127
> --- /dev/null
> +++ b/arch/arm/boards/virt/Kconfig
> @@ -0,0 +1,8 @@
> +
> +if MACH_VIRT
> +
> +config ARCH_TEXT_BASE
> +       hex
> +       default 0x40000000
> +
> +endif
> diff --git a/arch/arm/boards/virt/Makefile b/arch/arm/boards/virt/Makefile
> new file mode 100644
> index 0000000..eb072c0
> --- /dev/null
> +++ b/arch/arm/boards/virt/Makefile
> @@ -0,0 +1 @@
> +obj-y += init.o
> diff --git a/arch/arm/boards/virt/env/bin/_update b/arch/arm/boards/virt/env/bin/_update
> new file mode 100644
> index 0000000..014bce3
> --- /dev/null
> +++ b/arch/arm/boards/virt/env/bin/_update
> @@ -0,0 +1,36 @@
> +#!/bin/sh
> +
> +if [ -z "$part" -o -z "$image" ]; then
> +       echo "define \$part and \$image"
> +       exit 1
> +fi
> +
> +if [ ! -e "$part" ]; then
> +       echo "Partition $part does not exist"
> +       exit 1
> +fi
> +
> +if [ $# = 1 ]; then
> +       image=$1
> +fi
> +
> +if [ x$ip = xdhcp ]; then
> +       dhcp
> +fi
> +
> +ping $eth0.serverip
> +if [ $? -ne 0 ] ; then
> +       echo "update aborted"
> +       exit 1
> +fi
> +
> +unprotect $part
> +
> +echo
> +echo "erasing partition $part"
> +erase $part
> +
> +echo
> +echo "flashing $image to $part"
> +echo
> +tftp $image $part
> diff --git a/arch/arm/boards/virt/env/bin/boot b/arch/arm/boards/virt/env/bin/boot
> new file mode 100644
> index 0000000..3859dc1
> --- /dev/null
> +++ b/arch/arm/boards/virt/env/bin/boot
> @@ -0,0 +1,38 @@
> +#!/bin/sh
> +
> +. /env/config
> +
> +if [ x$1 = xflash ]; then
> +       root=flash
> +       kernel=flash
> +fi
> +
> +if [ x$1 = xnet ]; then
> +       root=net
> +       kernel=net
> +fi
> +
> +if [ x$ip = xdhcp ]; then
> +       bootargs="$bootargs ip=dhcp"
> +else
> +       bootargs="$bootargs ip=$eth0.ipaddr:$eth0.serverip:$eth0.gateway:$eth0.netmask:::"
> +fi
> +
> +if [ x$root = xflash ]; then
> +       bootargs="$bootargs root=$rootpart rootfstype=jffs2"
> +else
> +       bootargs="$bootargs root=/dev/nfs nfsroot=$eth0.serverip:$nfsroot,v3,tcp"
> +fi
> +
> +bootargs="$bootargs mtdparts=physmap-flash.0:$mtdparts"
> +
> +if [ $kernel = net ]; then
> +       if [ x$ip = xdhcp ]; then
> +               dhcp
> +       fi
> +       tftp $uimage uImage || exit 1
> +       bootm uImage
> +else
> +       bootm /dev/nor0.kernel
> +fi
> +
> diff --git a/arch/arm/boards/virt/env/bin/init b/arch/arm/boards/virt/env/bin/init
> new file mode 100644
> index 0000000..48e2139
> --- /dev/null
> +++ b/arch/arm/boards/virt/env/bin/init
> @@ -0,0 +1,20 @@
> +#!/bin/sh
> +
> +PATH=/env/bin
> +export PATH
> +
> +. /env/config
> +addpart /dev/nor0 $mtdparts
> +
> +echo
> +echo -n "Hit any key to stop autoboot: "
> +timeout -a $autoboot_timeout
> +if [ $? != 0 ]; then
> +       echo
> +       echo "type update_kernel [<imagename>] to update kernel into flash"
> +       echo "type udate_root [<imagename>] to update rootfs into flash"
> +       echo
> +       exit
> +fi
> +
> +boot
> \ No newline at end of file
> diff --git a/arch/arm/boards/virt/env/bin/update_kernel b/arch/arm/boards/virt/env/bin/update_kernel
> new file mode 100644
> index 0000000..1ad95fc
> --- /dev/null
> +++ b/arch/arm/boards/virt/env/bin/update_kernel
> @@ -0,0 +1,8 @@
> +#!/bin/sh
> +
> +. /env/config
> +
> +image=$uimage
> +part=/dev/nor0.kernel
> +
> +. /env/bin/_update $1
> diff --git a/arch/arm/boards/virt/env/bin/update_root b/arch/arm/boards/virt/env/bin/update_root
> new file mode 100644
> index 0000000..b757a5b
> --- /dev/null
> +++ b/arch/arm/boards/virt/env/bin/update_root
> @@ -0,0 +1,8 @@
> +#!/bin/sh
> +
> +. /env/config
> +
> +image=$jffs2
> +part=/dev/nor0.root
> +
> +. /env/bin/_update $1
> diff --git a/arch/arm/boards/virt/env/config b/arch/arm/boards/virt/env/config
> new file mode 100644
> index 0000000..6c0abda
> --- /dev/null
> +++ b/arch/arm/boards/virt/env/config
> @@ -0,0 +1,38 @@
> +#!/bin/sh
> +
> +# use 'dhcp' to do dhcp in barebox and in kernel
> +# use 'none' if you want to skip kernel ip autoconfiguration
> +ip=dhcp
> +global.dhcp.vendor_id=barebox-${global.hostname}
> +
> +# or set your networking parameters here
> +#eth0.ipaddr=a.b.c.d
> +#eth0.netmask=a.b.c.d
> +#eth0.gateway=a.b.c.d
> +#eth0.serverip=a.b.c.d
> +
> +# can be either 'nfs', 'tftp' or 'nor'
> +kernel_loc=tftp
> +# can be either 'net', 'nor' or 'initrd'
> +rootfs_loc=initrd
> +
> +# can be either 'jffs2' or 'ubifs'
> +rootfs_type=ubifs
> +rootfsimage=root.$rootfs_type
> +
> +kernelimage=zImage
> +#kernelimage=uImage
> +#kernelimage=Image
> +#kernelimage=Image.lzo
> +
> +nfsroot="$eth0.serverip:/opt/work/busybox/arm9/rootfs_arm"
> +
> +nor_parts="256k(barebox)ro,64k(bareboxenv),1536k(kernel),-(root)"
> +rootfs_mtdblock_nor=3
> +
> +autoboot_timeout=3
> +
> +bootargs="console=ttyAMA0,115200n8 CONSOLE=/dev/ttyAMA0"
> +
> +# set a fancy prompt (if support is compiled in)
> +PS1="\e[1;31m[barebox@\h]:\w\e[0m\n# "
> diff --git a/arch/arm/boards/virt/env/init/mtdparts-nor b/arch/arm/boards/virt/env/init/mtdparts-nor
> new file mode 100644
> index 0000000..3307596
> --- /dev/null
> +++ b/arch/arm/boards/virt/env/init/mtdparts-nor
> @@ -0,0 +1,11 @@
> +#!/bin/sh
> +
> +if [ "$1" = menu ]; then
> +        init-menu-add-entry "$0" "NOR partitions"
> +        exit
> +fi
> +
> +mtdparts="2048k@0(nor0.barebox)ro,256k(nor0.barebox-env),256k(nor0.barebox-logo),256k(nor0.barebox-logo2),5120k(nor0.kernel),-(nor0.root)"
> +kernelname="application-flash"
> +
> +mtdparts-add -d nor0 -k ${kernelname} -p ${mtdparts}
> diff --git a/arch/arm/boards/virt/init.c b/arch/arm/boards/virt/init.c
> new file mode 100644
> index 0000000..9626067
> --- /dev/null
> +++ b/arch/arm/boards/virt/init.c
> @@ -0,0 +1,67 @@
> +/*
> + * Copyright (C) 2016 Raphaël Poggi <poggi.raph@gmail.com>
> + *
> + * GPLv2 only
> + */
> +
> +#include <common.h>
> +#include <init.h>
> +#include <asm/armlinux.h>
> +#include <asm/system_info.h>
> +#include <mach/devices.h>
> +#include <environment.h>
> +#include <linux/sizes.h>
> +#include <io.h>
> +#include <globalvar.h>
> +#include <asm/mmu.h>
> +
> +static int virt_mem_init(void)
> +{
> +       virt_add_ddram(SZ_512M);
> +
> +       add_cfi_flash_device(0, 0x00000000, SZ_4M, 0);
> +
> +       devfs_add_partition("nor0", 0x00000, 0x40000, DEVFS_PARTITION_FIXED, "self0");
> +       devfs_add_partition("nor0", 0x40000, 0x20000, DEVFS_PARTITION_FIXED, "env0");
> +
> +       return 0;
> +}
> +mem_initcall(virt_mem_init);
> +
> +static int virt_console_init(void)
> +{
> +       virt_register_uart(0);
> +
> +       return 0;
> +}
> +console_initcall(virt_console_init);
> +
> +static int virt_core_init(void)
> +{
> +       char *hostname = "virt";
> +
> +       if (cpu_is_cortex_a53())
> +               hostname = "virt-a53";
> +       else if (cpu_is_cortex_a57())
> +               hostname = "virt-a57";
> +
> +       barebox_set_model("ARM QEMU virt");
> +       barebox_set_hostname(hostname);
> +
> +       return 0;
> +}
> +postcore_initcall(virt_core_init);
> +
> +static int virt_mmu_enable(void)
> +{
> +       /* Mapping all periph range */
> +       arch_remap_range(0x09000000, 0x01000000, PMD_SECT_DEF_CACHED);
> +
> +       /* Mapping all flash range */
> +       arch_remap_range(0x00000000, 0x08000000, PMD_SECT_DEF_CACHED);
> +
> +       mmu_enable();
> +
> +       return 0;
> +}
> +postmmu_initcall(virt_mmu_enable);
> diff --git a/arch/arm/configs/virt_defconfig b/arch/arm/configs/virt_defconfig
> new file mode 100644
> index 0000000..ae928a2
> --- /dev/null
> +++ b/arch/arm/configs/virt_defconfig
> @@ -0,0 +1,53 @@
> +CONFIG_AEABI=y
> +CONFIG_ARM_OPTIMZED_STRING_FUNCTIONS=y
> +CONFIG_BAREBOX_MAX_IMAGE_SIZE=0x05000000
> +CONFIG_BAREBOX_MAX_BARE_INIT_SIZE=0x01000000
> +CONFIG_MEMORY_LAYOUT_FIXED=y
> +CONFIG_STACK_BASE=0x60000000
> +CONFIG_MALLOC_BASE=0x50000000
> +CONFIG_PROMPT="virt: "
> +CONFIG_HUSH_FANCY_PROMPT=y
> +CONFIG_CMDLINE_EDITING=y
> +CONFIG_AUTO_COMPLETE=y
> +CONFIG_MENU=y
> +CONFIG_PASSWORD=y
> +CONFIG_PARTITION=y
> +CONFIG_DEFAULT_ENVIRONMENT_GENERIC_NEW=y
> +CONFIG_DEFAULT_ENVIRONMENT_PATH="arch/arm/boards/virt/env"
> +CONFIG_DEBUG_INFO=y
> +# CONFIG_CMD_ARM_CPUINFO is not set
> +CONFIG_LONGHELP=y
> +# CONFIG_CMD_BOOTM is not set
> +# CONFIG_CMD_BOOTU is not set
> +# CONFIG_CMD_MOUNT is not set
> +# CONFIG_CMD_UMOUNT is not set
> +# CONFIG_CMD_CAT is not set
> +# CONFIG_CMD_CD is not set
> +# CONFIG_CMD_CP is not set
> +# CONFIG_CMD_LS is not set
> +# CONFIG_CMD_MKDIR is not set
> +# CONFIG_CMD_PWD is not set
> +# CONFIG_CMD_RM is not set
> +# CONFIG_CMD_RMDIR is not set
> +# CONFIG_CMD_FALSE is not set
> +# CONFIG_CMD_TEST is not set
> +# CONFIG_CMD_TRUE is not set
> +# CONFIG_CMD_CLEAR is not set
> +# CONFIG_CMD_ECHO is not set
> +CONFIG_CMD_CRC=y
> +CONFIG_CMD_CRC_CMP=y
> +# CONFIG_CMD_MD is not set
> +# CONFIG_CMD_MEMCMP is not set
> +# CONFIG_CMD_MEMCPY is not set
> +# CONFIG_CMD_MEMSET is not set
> +# CONFIG_CMD_MW is not set
> +CONFIG_SERIAL_AMBA_PL011=y
> +# CONFIG_SPI is not set
> +CONFIG_MTD=y
> +CONFIG_DRIVER_CFI=y
> +CONFIG_DRIVER_CFI_BANK_WIDTH_8=y
> +CONFIG_CFI_BUFFER_WRITE=y
> +CONFIG_NAND=y
> +# CONFIG_FS_RAMFS is not set
> +CONFIG_DIGEST_SHA1_GENERIC=y
> +CONFIG_DIGEST_SHA256_GENERIC=y
> diff --git a/arch/arm/cpu/Kconfig b/arch/arm/cpu/Kconfig
> index 4f5d9b6..86d64a4 100644
> --- a/arch/arm/cpu/Kconfig
> +++ b/arch/arm/cpu/Kconfig
> @@ -1,6 +1,6 @@
>  comment "Processor Type"
>
> -config CPU_32
> +config CPU_64
>         bool
>         default y
>
> @@ -8,90 +8,13 @@ config CPU_32
>  # which CPUs we support in the kernel image, and the compiler instruction
>  # optimiser behaviour.
>
> -# ARM1176
> -config CPU_ARM1176
> +# ARMv8
> +config CPU_V8
>         bool
> -       select CPU_V6
> +       select CPU_64v8
> +       select CPU_SUPPORTS_64BIT_KERNEL
>
> -# ARM920T
> -config CPU_ARM920T
> -       bool
> -       select CPU_32v4T
> -       help
> -         The ARM920T is licensed to be produced by numerous vendors,
> -         and is used in the Maverick EP9312 and the Samsung S3C2410.
> -
> -         More information on the Maverick EP9312 at
> -         <http://www.cirrus.com/en/products/ep9312.html>.
> -
> -         Say Y if you want support for the ARM920T processor.
> -         Otherwise, say N.
> -
> -# ARM926T
> -config CPU_ARM926T
> -       bool
> -       select CPU_32v5
> -       help
> -         This is a variant of the ARM920. It has slightly different
> -         instruction sequences for cache and TLB operations. Curiously,
> -         there is no documentation on it at the ARM corporate website.
> -
> -         Say Y if you want support for the ARM926T processor.
> -         Otherwise, say N.
> -
> -# ARM946E-S
> -config CPU_ARM946E
> -       bool
> -       select CPU_32v4T
> -       help
> -         ARM946E-S is a member of the ARM9E-S family of high-
> -         performance, 32-bit system-on-chip processor solutions.
> -         The TCM and ARMv5TE 32-bit instruction set is supported.
> -
> -         Say Y if you want support for the ARM946E-S processor.
> -         Otherwise, say N.
> -
> -# Feroceon
> -config CPU_FEROCEON
> -       bool
> -       select CPU_32v5
> -       help
> -         This is a Marvell implementation of an ARMv5TE compatible
> -         ARM core, used in the Marvell Kirkwood SoC family.
> -
> -# ARMv6
> -config CPU_V6
> -       bool
> -       select CPU_32v6
> -
> -# ARMv7
> -config CPU_V7
> -       bool
> -       select CPU_32v7
> -
> -config CPU_XSC3
> -        bool
> -        select CPU_32v4T
> -        help
> -          Select code specific to PXA3xx variants
> -
> -# Xscale PXA25x, PXA27x
> -config CPU_XSCALE
> -       bool
> -       select CPU_32v4T
> -
> -# Figure out what processor architecture version we should be using.
> -# This defines the compiler instruction set which depends on the machine type.
> -config CPU_32v4T
> -       bool
> -
> -config CPU_32v5
> -       bool
> -
> -config CPU_32v6
> -       bool
> -
> -config CPU_32v7
> +config CPU_64v8
>         bool
>
>  comment "processor features"
> @@ -117,10 +40,15 @@ config BOOT_ENDIANNESS_SWITCH
>
>           Currently implemented only by "bootz" command.
>
> -config ARCH_HAS_L2X0
> +config SYS_SUPPORTS_32BIT_KERNEL
> +       bool
> +
> +config SYS_SUPPORTS_64BIT_KERNEL
> +       bool
> +
> +config CPU_SUPPORTS_32BIT_KERNEL
>         bool
>
> -config CACHE_L2X0
> -       bool "Enable L2x0 PrimeCell"
> -       depends on MMU && ARCH_HAS_L2X0
> +config CPU_SUPPORTS_64BIT_KERNEL
> +       bool
>
> diff --git a/arch/arm/cpu/Makefile b/arch/arm/cpu/Makefile
> index 854df60e..fe6e7af 100644
> --- a/arch/arm/cpu/Makefile
> +++ b/arch/arm/cpu/Makefile
> @@ -1,38 +1,25 @@
>  obj-y += cpu.o
>  obj-$(CONFIG_ARM_EXCEPTIONS) += exceptions.o
>  obj-$(CONFIG_ARM_EXCEPTIONS) += interrupts.o
> -obj-y += start.o setupc.o entry.o
> +obj-y += start.o entry.o
>
>  #
>  # Any variants can be called as start-armxyz.S
>  #
>  obj-$(CONFIG_CMD_ARM_CPUINFO) += cpuinfo.o
>  obj-$(CONFIG_CMD_ARM_MMUINFO) += mmuinfo.o
> -obj-$(CONFIG_OFDEVICE) += dtb.o
> -obj-$(CONFIG_MMU) += mmu.o cache.o mmu-early.o
> -pbl-$(CONFIG_MMU) += mmu-early.o
> +obj-$(CONFIG_MMU) += mmu.o cache.o
>
> -ifeq ($(CONFIG_MMU),)
> -obj-y += no-mmu.o
> -endif
> -
> -obj-$(CONFIG_CPU_32v4T) += cache-armv4.o
> -pbl-$(CONFIG_CPU_32v4T) += cache-armv4.o
> -obj-$(CONFIG_CPU_32v5) += cache-armv5.o
> -pbl-$(CONFIG_CPU_32v5) += cache-armv5.o
> -obj-$(CONFIG_CPU_32v6) += cache-armv6.o
> -pbl-$(CONFIG_CPU_32v6) += cache-armv6.o
> -AFLAGS_cache-armv7.o       :=-Wa,-march=armv7-a
> -obj-$(CONFIG_CPU_32v7) += cache-armv7.o
> -AFLAGS_pbl-cache-armv7.o       :=-Wa,-march=armv7-a
> -pbl-$(CONFIG_CPU_32v7) += cache-armv7.o
> -obj-$(CONFIG_CACHE_L2X0) += cache-l2x0.o
> +AFLAGS_cache-armv8.o       :=-Wa,-march=armv8-a
> +obj-$(CONFIG_CPU_64v8) += cache-armv8.o
> +AFLAGS_pbl-cache-armv8.o       :=-Wa,-march=armv8-a
> +pbl-$(CONFIG_CPU_64v8) += cache-armv8.o
>
>  pbl-y += setupc.o entry.o
>  pbl-$(CONFIG_PBL_SINGLE_IMAGE) += start-pbl.o
>  pbl-$(CONFIG_PBL_MULTI_IMAGES) += uncompress.o
>
> -obj-y += common.o cache.o
> -pbl-y += common.o cache.o
> +obj-y += cache.o
> +pbl-y += cache.o
>
>  lwl-y += lowlevel.o
> diff --git a/arch/arm/cpu/cache-armv8.S b/arch/arm/cpu/cache-armv8.S
> new file mode 100644
> index 0000000..82b2f81
> --- /dev/null
> +++ b/arch/arm/cpu/cache-armv8.S
> @@ -0,0 +1,168 @@
> +/*
> + * (C) Copyright 2013
> + * David Feng <fenghua@phytium.com.cn>
> + *
> + * This file is based on sample code from ARMv8 ARM.
> + *
> + * SPDX-License-Identifier:    GPL-2.0+
> + */
> +
> +#include <config.h>
> +#include <linux/linkage.h>
> +#include <init.h>
> +
> +/*
> + * void v8_flush_dcache_level(level)
> + *
> + * clean and invalidate one level cache.
> + *
> + * x0: cache level
> + * x1: 0 flush & invalidate, 1 invalidate only
> + * x2~x9: clobbered
> + */
> +.section .text.v8_flush_dcache_level
> +ENTRY(v8_flush_dcache_level)
> +       lsl     x12, x0, #1
> +       msr     csselr_el1, x12         /* select cache level */
> +       isb                             /* sync change of cssidr_el1 */
> +       mrs     x6, ccsidr_el1          /* read the new cssidr_el1 */
> +       and     x2, x6, #7              /* x2 <- log2(cache line size)-4 */
> +       add     x2, x2, #4              /* x2 <- log2(cache line size) */
> +       mov     x3, #0x3ff
> +       and     x3, x3, x6, lsr #3      /* x3 <- max number of #ways */
> +       clz     w5, w3                  /* bit position of #ways */
> +       mov     x4, #0x7fff
> +       and     x4, x4, x6, lsr #13     /* x4 <- max number of #sets */
> +       /* x12 <- cache level << 1 */
> +       /* x2 <- line length offset */
> +       /* x3 <- number of cache ways - 1 */
> +       /* x4 <- number of cache sets - 1 */
> +       /* x5 <- bit position of #ways */
> +
> +loop_set:
> +       mov     x6, x3                  /* x6 <- working copy of #ways */
> +loop_way:
> +       lsl     x7, x6, x5
> +       orr     x9, x12, x7             /* map way and level to cisw value */
> +       lsl     x7, x4, x2
> +       orr     x9, x9, x7              /* map set number to cisw value */
> +       tbz     w1, #0, 1f
> +       dc      isw, x9
> +       b       2f
> +1:     dc      cisw, x9                /* clean & invalidate by set/way */
> +2:     subs    x6, x6, #1              /* decrement the way */
> +       b.ge    loop_way
> +       subs    x4, x4, #1              /* decrement the set */
> +       b.ge    loop_set
> +
> +       ret
> +ENDPROC(v8_flush_dcache_level)
> +
> +/*
> + * void v8_flush_dcache_all(int invalidate_only)
> + *
> + * x0: 0 flush & invalidate, 1 invalidate only
> + *
> + * clean and invalidate all data cache by SET/WAY.
> + */
> +.section .text.v8_dcache_all
> +ENTRY(v8_dcache_all)
> +       mov     x1, x0
> +       dsb     sy
> +       mrs     x10, clidr_el1          /* read clidr_el1 */
> +       lsr     x11, x10, #24
> +       and     x11, x11, #0x7          /* x11 <- loc */
> +       cbz     x11, finished           /* if loc is 0, exit */
> +       mov     x15, x30
> +       mov     x0, #0                  /* start flush at cache level 0 */
> +       /* x0  <- cache level */
> +       /* x10 <- clidr_el1 */
> +       /* x11 <- loc */
> +       /* x15 <- return address */
> +
> +loop_level:
> +       lsl     x12, x0, #1
> +       add     x12, x12, x0            /* x0 <- tripled cache level */
> +       lsr     x12, x10, x12
> +       and     x12, x12, #7            /* x12 <- cache type */
> +       cmp     x12, #2
> +       b.lt    skip                    /* skip if no cache or icache */
> +       bl      v8_flush_dcache_level   /* x1 = 0 flush, 1 invalidate */
> +skip:
> +       add     x0, x0, #1              /* increment cache level */
> +       cmp     x11, x0
> +       b.gt    loop_level
> +
> +       mov     x0, #0
> +       msr     csselr_el1, x0          /* restore csselr_el1 */
> +       dsb     sy
> +       isb
> +       mov     x30, x15
> +
> +finished:
> +       ret
> +ENDPROC(v8_dcache_all)
> +
> +.section .text.v8_flush_dcache_all
> +ENTRY(v8_flush_dcache_all)
> +       mov     x16, x30
> +       mov     x0, #0
> +       bl      v8_dcache_all
> +       mov     x30, x16
> +       ret
> +ENDPROC(v8_flush_dcache_all)
> +
> +.section .text.v8_invalidate_dcache_all
> +ENTRY(v8_invalidate_dcache_all)
> +       mov     x16, x30
> +       mov     x0, #0x1
> +       bl      v8_dcache_all
> +       mov     x30, x16
> +       ret
> +ENDPROC(v8_invalidate_dcache_all)
> +
> +/*
> + * void v8_flush_dcache_range(start, end)
> + *
> + * clean & invalidate data cache in the range
> + *
> + * x0: start address
> + * x1: end address
> + */
> +.section .text.v8_flush_dcache_range
> +ENTRY(v8_flush_dcache_range)
> +       mrs     x3, ctr_el0
> +       lsr     x3, x3, #16
> +       and     x3, x3, #0xf
> +       mov     x2, #4
> +       lsl     x2, x2, x3              /* cache line size */
> +
> +       /* x2 <- minimal cache line size in cache system */
> +       sub     x3, x2, #1
> +       bic     x0, x0, x3
> +1:     dc      civac, x0       /* clean & invalidate data or unified cache */
> +       add     x0, x0, x2
> +       cmp     x0, x1
> +       b.lo    1b
> +       dsb     sy
> +       ret
> +ENDPROC(v8_flush_dcache_range)
> +
> +/*
> + * void v8_invalidate_icache_all(void)
> + *
> + * invalidate all tlb entries.
> + */
> +.section .text.v8_invalidate_icache_all
> +ENTRY(v8_invalidate_icache_all)
> +       ic      ialluis
> +       isb     sy
> +       ret
> +ENDPROC(v8_invalidate_icache_all)
> +
> +.section .text.v8_flush_l3_cache
> +ENTRY(v8_flush_l3_cache)
> +       mov     x0, #0                  /* return status as success */
> +       ret
> +ENDPROC(v8_flush_l3_cache)
> +       .weak   v8_flush_l3_cache
> diff --git a/arch/arm/cpu/cache.c b/arch/arm/cpu/cache.c
> index 27ead1c..8465cf9 100644
> --- a/arch/arm/cpu/cache.c
> +++ b/arch/arm/cpu/cache.c
> @@ -32,10 +32,7 @@ struct cache_fns *cache_fns;
>                 .mmu_cache_flush = arch##_mmu_cache_flush,                      \
>         };
>
> -DEFINE_CPU_FNS(v4)
> -DEFINE_CPU_FNS(v5)
> -DEFINE_CPU_FNS(v6)
> -DEFINE_CPU_FNS(v7)
> +DEFINE_CPU_FNS(v8)
>
>  void __dma_clean_range(unsigned long start, unsigned long end)
>  {
> @@ -78,29 +75,9 @@ void __mmu_cache_flush(void)
>  int arm_set_cache_functions(void)
>  {
>         switch (cpu_architecture()) {
> -#ifdef CONFIG_CPU_32v4T
> -       case CPU_ARCH_ARMv4T:
> -               cache_fns = &cache_fns_armv4;
> +       case CPU_ARCH_ARMv8:
> +               cache_fns = &cache_fns_armv8;
>                 break;
> -#endif
> -#ifdef CONFIG_CPU_32v5
> -       case CPU_ARCH_ARMv5:
> -       case CPU_ARCH_ARMv5T:
> -       case CPU_ARCH_ARMv5TE:
> -       case CPU_ARCH_ARMv5TEJ:
> -               cache_fns = &cache_fns_armv5;
> -               break;
> -#endif
> -#ifdef CONFIG_CPU_32v6
> -       case CPU_ARCH_ARMv6:
> -               cache_fns = &cache_fns_armv6;
> -               break;
> -#endif
> -#ifdef CONFIG_CPU_32v7
> -       case CPU_ARCH_ARMv7:
> -               cache_fns = &cache_fns_armv7;
> -               break;
> -#endif
>         default:
>                 while(1);
>         }
> @@ -115,49 +92,19 @@ int arm_set_cache_functions(void)
>  void arm_early_mmu_cache_flush(void)
>  {
>         switch (arm_early_get_cpu_architecture()) {
> -#ifdef CONFIG_CPU_32v4T
> -       case CPU_ARCH_ARMv4T:
> -               v4_mmu_cache_flush();
> -               return;
> -#endif
> -#ifdef CONFIG_CPU_32v5
> -       case CPU_ARCH_ARMv5:
> -       case CPU_ARCH_ARMv5T:
> -       case CPU_ARCH_ARMv5TE:
> -       case CPU_ARCH_ARMv5TEJ:
> -               v5_mmu_cache_flush();
> +       case CPU_ARCH_ARMv8:
> +//             v7_mmu_cache_flush();
>                 return;
> -#endif
> -#ifdef CONFIG_CPU_32v6
> -       case CPU_ARCH_ARMv6:
> -               v6_mmu_cache_flush();
> -               return;
> -#endif
> -#ifdef CONFIG_CPU_32v7
> -       case CPU_ARCH_ARMv7:
> -               v7_mmu_cache_flush();
> -               return;
> -#endif
>         }
>  }
>
> -void v7_mmu_cache_invalidate(void);
> +//void v7_mmu_cache_invalidate(void);
>
>  void arm_early_mmu_cache_invalidate(void)
>  {
>         switch (arm_early_get_cpu_architecture()) {
> -       case CPU_ARCH_ARMv4T:
> -       case CPU_ARCH_ARMv5:
> -       case CPU_ARCH_ARMv5T:
> -       case CPU_ARCH_ARMv5TE:
> -       case CPU_ARCH_ARMv5TEJ:
> -       case CPU_ARCH_ARMv6:
> -               asm volatile("mcr p15, 0, %0, c7, c6, 0\n" : : "r"(0));
> -               return;
> -#ifdef CONFIG_CPU_32v7
> -       case CPU_ARCH_ARMv7:
> -               v7_mmu_cache_invalidate();
> +       case CPU_ARCH_ARMv8:
> +//             v7_mmu_cache_invalidate();
>                 return;
> -#endif
>         }
>  }
> diff --git a/arch/arm/cpu/cpu.c b/arch/arm/cpu/cpu.c
> index eb12166..19cd944 100644
> --- a/arch/arm/cpu/cpu.c
> +++ b/arch/arm/cpu/cpu.c
> @@ -40,11 +40,8 @@
>   */
>  void icache_enable(void)
>  {
> -       u32 r;
> -
> -       r = get_cr();
> -       r |= CR_I;
> -       set_cr(r);
> +       v8_invalidate_icache_all();
> +       set_sctlr(get_sctlr() | CR_I);
>  }
>
>  /**
> @@ -52,11 +49,7 @@ void icache_enable(void)
>   */
>  void icache_disable(void)
>  {
> -       u32 r;
> -
> -       r = get_cr();
> -       r &= ~CR_I;
> -       set_cr(r);
> +       set_sctlr(get_sctlr() & ~CR_I);
>  }
>
>  /**
> @@ -65,26 +58,7 @@ void icache_disable(void)
>   */
>  int icache_status(void)
>  {
> -       return (get_cr () & CR_I) != 0;
> -}
> -
> -/*
> - * SoC like the ux500 have the l2x0 always enable
> - * with or without MMU enable
> - */
> -struct outer_cache_fns outer_cache;
> -
> -/*
> - * Clean and invalide caches, disable MMU
> - */
> -void mmu_disable(void)
> -{
> -       __mmu_cache_flush();
> -       if (outer_cache.disable) {
> -               outer_cache.flush_all();
> -               outer_cache.disable();
> -       }
> -       __mmu_cache_off();
> +       return (get_sctlr() & CR_I) != 0;
>  }
>
>  /**
> @@ -96,18 +70,8 @@ void mmu_disable(void)
>   */
>  static void arch_shutdown(void)
>  {
> -       uint32_t r;
> -
>         mmu_disable();
>         flush_icache();
> -       /*
> -        * barebox normally does not use interrupts, but some functionalities
> -        * (eg. OMAP4_USBBOOT) require them enabled. So be sure interrupts are
> -        * disabled before exiting.
> -        */
> -       __asm__ __volatile__("mrs %0, cpsr" : "=r"(r));
> -       r |= PSR_I_BIT;
> -       __asm__ __volatile__("msr cpsr, %0" : : "r"(r));
>  }
>  archshutdown_exitcall(arch_shutdown);
>
> diff --git a/arch/arm/cpu/cpuinfo.c b/arch/arm/cpu/cpuinfo.c
> index 8b22e9b..2306101 100644
> --- a/arch/arm/cpu/cpuinfo.c
> +++ b/arch/arm/cpu/cpuinfo.c
> @@ -21,21 +21,10 @@
>  #include <complete.h>
>
>  #define CPU_ARCH_UNKNOWN       0
> -#define CPU_ARCH_ARMv3         1
> -#define CPU_ARCH_ARMv4         2
> -#define CPU_ARCH_ARMv4T                3
> -#define CPU_ARCH_ARMv5         4
> -#define CPU_ARCH_ARMv5T                5
> -#define CPU_ARCH_ARMv5TE       6
> -#define CPU_ARCH_ARMv5TEJ      7
> -#define CPU_ARCH_ARMv6         8
> -#define CPU_ARCH_ARMv7         9
> -
> -#define ARM_CPU_PART_CORTEX_A5      0xC050
> -#define ARM_CPU_PART_CORTEX_A7      0xC070
> -#define ARM_CPU_PART_CORTEX_A8      0xC080
> -#define ARM_CPU_PART_CORTEX_A9      0xC090
> -#define ARM_CPU_PART_CORTEX_A15     0xC0F0
> +#define CPU_ARCH_ARMv8         10
> +
> +#define ARM_CPU_PART_CORTEX_A53            0xD034
> +#define ARM_CPU_PART_CORTEX_A57            0xD070
>
>  static void decode_cache(unsigned long size)
>  {
> @@ -61,22 +50,22 @@ static int do_cpuinfo(int argc, char *argv[])
>         int cpu_arch;
>
>         __asm__ __volatile__(
> -               "mrc    p15, 0, %0, c0, c0, 0   @ read control reg\n"
> +               "mrs    %0, midr_el1\n"
>                 : "=r" (mainid)
>                 :
>                 : "memory");
>
> -       __asm__ __volatile__(
> -               "mrc    p15, 0, %0, c0, c0, 1   @ read control reg\n"
> -               : "=r" (cache)
> -               :
> -               : "memory");
> -
> -       __asm__ __volatile__(
> -               "mrc    p15, 0, %0, c1, c0, 0   @ read control reg\n"
> -                       : "=r" (cr)
> -                       :
> -                       : "memory");
> +//     __asm__ __volatile__(
> +//             "mrc    p15, 0, %0, c0, c0, 1   @ read control reg\n"
> +//             : "=r" (cache)
> +//             :
> +//             : "memory");
> +//
> +//     __asm__ __volatile__(
> +//             "mrc    p15, 0, %0, c1, c0, 0   @ read control reg\n"
> +//                     : "=r" (cr)
> +//                     :
> +//                     : "memory");
>
>         switch (mainid >> 24) {
>         case 0x41:
> @@ -111,8 +100,8 @@ static int do_cpuinfo(int argc, char *argv[])
>
>                 /* Revised CPUID format. Read the Memory Model Feature
>                  * Register 0 and check for VMSAv7 or PMSAv7 */
> -               asm("mrc        p15, 0, %0, c0, c1, 4"
> -                   : "=r" (mmfr0));
> +//             asm("mrc        p15, 0, %0, c0, c1, 4"
> +//                 : "=r" (mmfr0));
>                 if ((mmfr0 & 0x0000000f) >= 0x00000003 ||
>                     (mmfr0 & 0x000000f0) >= 0x00000030)
>                         cpu_arch = CPU_ARCH_ARMv7;
> @@ -152,6 +141,9 @@ static int do_cpuinfo(int argc, char *argv[])
>         case CPU_ARCH_ARMv7:
>                 architecture = "v7";
>                 break;
> +       case CPU_ARCH_ARMv8:
> +               architecture = "v8";
> +               break;
>         case CPU_ARCH_UNKNOWN:
>         default:
>                 architecture = "Unknown";
> @@ -181,29 +173,31 @@ static int do_cpuinfo(int argc, char *argv[])
>                 case ARM_CPU_PART_CORTEX_A15:
>                         part = "Cortex-A15";
>                         break;
> +               case ARM_CPU_PART_CORTEX_A53:
> +                       part = "Cortex-A53";
>                 default:
>                         part = "unknown";
>                 }
>                 printf("core: %s r%up%u\n", part, major, minor);
>         }
>
> -       if (cache & (1 << 24)) {
> -               /* separate I/D cache */
> -               printf("I-cache: ");
> -               decode_cache(cache & 0xfff);
> -               printf("D-cache: ");
> -               decode_cache((cache >> 12) & 0xfff);
> -       } else {
> -               /* unified I/D cache */
> -               printf("cache: ");
> -               decode_cache(cache & 0xfff);
> -       }
> -
> -       printf("Control register: ");
> -       for (i = 0; i < ARRAY_SIZE(crbits); i++)
> -               if (cr & (1 << i))
> -                       printf("%s ", crbits[i]);
> -       printf("\n");
> +//     if (cache & (1 << 24)) {
> +//             /* separate I/D cache */
> +//             printf("I-cache: ");
> +//             decode_cache(cache & 0xfff);
> +//             printf("D-cache: ");
> +//             decode_cache((cache >> 12) & 0xfff);
> +//     } else {
> +//             /* unified I/D cache */
> +//             printf("cache: ");
> +//             decode_cache(cache & 0xfff);
> +//     }
> +
> +//     printf("Control register: ");
> +//     for (i = 0; i < ARRAY_SIZE(crbits); i++)
> +//             if (cr & (1 << i))
> +//                     printf("%s ", crbits[i]);
> +//     printf("\n");
>
>         return 0;
>  }
> diff --git a/arch/arm/cpu/entry.c b/arch/arm/cpu/entry.c
> index 0cdcfec..a029f09 100644
> --- a/arch/arm/cpu/entry.c
> +++ b/arch/arm/cpu/entry.c
> @@ -1,7 +1,6 @@
>  #include <types.h>
>
>  #include <asm/cache.h>
> -#include <asm/barebox-arm.h>
>
>  #include "entry.h"
>
> @@ -24,10 +23,10 @@
>   * be fine.
>   */
>
> -void __naked __noreturn barebox_arm_entry(unsigned long membase,
> +void __noreturn barebox_arm_entry(unsigned long membase,
>                                           unsigned long memsize, void *boarddata)
>  {
> -       arm_setup_stack(arm_mem_stack(membase, membase + memsize) + STACK_SIZE - 16);
> +       arm_setup_stack(membase + memsize - 16);
>         arm_early_mmu_cache_invalidate();
>
>         if (IS_ENABLED(CONFIG_PBL_MULTI_IMAGES))
> diff --git a/arch/arm/cpu/exceptions.S b/arch/arm/cpu/exceptions.S
> index eda0d6a..5812025 100644
> --- a/arch/arm/cpu/exceptions.S
> +++ b/arch/arm/cpu/exceptions.S
> @@ -1,220 +1,119 @@
> -#include <config.h>
> -#include <linux/linkage.h>
> -#include <asm-generic/memory_layout.h>
> -
>  /*
> - *************************************************************************
> - *
> - * Interrupt handling
> + * (C) Copyright 2013
> + * David Feng <fenghua@phytium.com.cn>
>   *
> - *************************************************************************
> + * SPDX-License-Identifier:    GPL-2.0+
>   */
>
> -@
> -@ IRQ stack frame.
> -@
> -#define S_FRAME_SIZE   72
> -
> -#define S_OLD_R0       68
> -#define S_PSR          64
> -#define S_PC           60
> -#define S_LR           56
> -#define S_SP           52
> -
> -#define S_IP           48
> -#define S_FP           44
> -#define S_R10          40
> -#define S_R9           36
> -#define S_R8           32
> -#define S_R7           28
> -#define S_R6           24
> -#define S_R5           20
> -#define S_R4           16
> -#define S_R3           12
> -#define S_R2           8
> -#define S_R1           4
> -#define S_R0           0
> -
> -#define MODE_SVC       0x13
> +#include <config.h>
> +#include <asm/ptrace.h>
> +#include <linux/linkage.h>
>
>  /*
> - * use bad_save_user_regs for abort/prefetch/undef/swi ...
> - * use irq_save_user_regs / irq_restore_user_regs for IRQ/FIQ handling
> + * Enter Exception.
> + * This will save the processor state that is ELR/X0~X30
> + * to the stack frame.
>   */
> -
> -       .macro  bad_save_user_regs
> -       sub     sp, sp, #S_FRAME_SIZE
> -       stmia   sp, {r0 - r12}                  @ Calling r0-r12
> -       ldr     r2, =abort_stack
> -       ldmia   r2, {r2 - r3}                   @ get pc, cpsr
> -       add     r0, sp, #S_FRAME_SIZE           @ restore sp_SVC
> -
> -       add     r5, sp, #S_SP
> -       mov     r1, lr
> -       stmia   r5, {r0 - r3}                   @ save sp_SVC, lr_SVC, pc, cpsr
> -       mov     r0, sp
> -       .endm
> -
> -       .macro  irq_save_user_regs
> -       sub     sp, sp, #S_FRAME_SIZE
> -       stmia   sp, {r0 - r12}                  @ Calling r0-r12
> -       add     r8, sp, #S_PC
> -       stmdb   r8, {sp, lr}^                   @ Calling SP, LR
> -       str     lr, [r8, #0]                    @ Save calling PC
> -       mrs     r6, spsr
> -       str     r6, [r8, #4]                    @ Save CPSR
> -       str     r0, [r8, #8]                    @ Save OLD_R0
> -       mov     r0, sp
> -       .endm
> -
> -       .macro  irq_restore_user_regs
> -       ldmia   sp, {r0 - lr}^                  @ Calling r0 - lr
> -       mov     r0, r0
> -       ldr     lr, [sp, #S_PC]                 @ Get PC
> -       add     sp, sp, #S_FRAME_SIZE
> -       subs    pc, lr, #4                      @ return & move spsr_svc into cpsr
> -       .endm
> -
> -       .macro get_bad_stack
> -       ldr     r13, =abort_stack
> -       str     lr, [r13]                       @ save caller lr / spsr
> -       mrs     lr, spsr
> -       str     lr, [r13, #4]
> -
> -       mov     r13, #MODE_SVC                  @ prepare SVC-Mode
> -       @ msr   spsr_c, r13
> -       msr     spsr, r13
> -       mov     lr, pc
> -       movs    pc, lr
> -       .endm
> -
> -       .macro try_data_abort
> -       ldr     r13, =arm_ignore_data_abort     @ check try mode
> -       ldr     r13, [r13]
> -       cmp     r13, #0
> -       beq     do_abort_\@
> -       ldr     r13, =arm_data_abort_occurred
> -       str     r13, [r13]
> -       mrs     r13, spsr                       @ read saved CPSR
> -       tst     r13, #1<<5                      @ check Thumb mode
> -       subeq   lr, #4                          @ next ARM instr
> -       subne   lr, #6                          @ next Thumb instr
> -       movs    pc, lr
> -do_abort_\@:
> -       .endm
> -
> -       .macro get_irq_stack                    @ setup IRQ stack
> -       ldr     sp, IRQ_STACK_START
> -       .endm
> -
> -       .macro get_fiq_stack                    @ setup FIQ stack
> -       ldr     sp, FIQ_STACK_START
> -       .endm
> +.macro exception_entry
> +       stp     x29, x30, [sp, #-16]!
> +       stp     x27, x28, [sp, #-16]!
> +       stp     x25, x26, [sp, #-16]!
> +       stp     x23, x24, [sp, #-16]!
> +       stp     x21, x22, [sp, #-16]!
> +       stp     x19, x20, [sp, #-16]!
> +       stp     x17, x18, [sp, #-16]!
> +       stp     x15, x16, [sp, #-16]!
> +       stp     x13, x14, [sp, #-16]!
> +       stp     x11, x12, [sp, #-16]!
> +       stp     x9, x10, [sp, #-16]!
> +       stp     x7, x8, [sp, #-16]!
> +       stp     x5, x6, [sp, #-16]!
> +       stp     x3, x4, [sp, #-16]!
> +       stp     x1, x2, [sp, #-16]!
> +
> +       /* Could be running at EL3/EL2/EL1 */
> +       mrs     x11, CurrentEL
> +       cmp     x11, #0xC               /* Check EL3 state */
> +       b.eq    1f
> +       cmp     x11, #0x8               /* Check EL2 state */
> +       b.eq    2f
> +       cmp     x11, #0x4               /* Check EL1 state */
> +       b.eq    3f
> +3:     mrs     x1, esr_el3
> +       mrs     x2, elr_el3
> +       b       0f
> +2:     mrs     x1, esr_el2
> +       mrs     x2, elr_el2
> +       b       0f
> +1:     mrs     x1, esr_el1
> +       mrs     x2, elr_el1
> +0:
> +       stp     x2, x0, [sp, #-16]!
> +       mov     x0, sp
> +.endm
>
>  /*
> - * exception handlers
> + * Exception vectors.
>   */
> -       .section ".text","ax"
> -       .arm
> -
> -       .align  5
> -undefined_instruction:
> -       get_bad_stack
> -       bad_save_user_regs
> -       bl      do_undefined_instruction
> -
> -       .align  5
> -software_interrupt:
> -       get_bad_stack
> -       bad_save_user_regs
> -       bl      do_software_interrupt
> -
> -       .align  5
> -prefetch_abort:
> -       get_bad_stack
> -       bad_save_user_regs
> -       bl      do_prefetch_abort
> -
> -       .align  5
> -data_abort:
> -       try_data_abort
> -       get_bad_stack
> -       bad_save_user_regs
> -       bl      do_data_abort
> -
> -       .align  5
> -irq:
> -       get_bad_stack
> -       bad_save_user_regs
> -       bl      do_irq
> -
> -       .align  5
> -fiq:
> -       get_bad_stack
> -       bad_save_user_regs
> -       bl      do_fiq
> -
> -#if defined(CONFIG_RELOCATABLE) && defined(CONFIG_ARM_EXCEPTIONS)
> -/*
> - * With relocatable binary support the runtime exception vectors do not match
> - * the addresses in the binary. We have to fix them up during runtime
> - */
> -ENTRY(arm_fixup_vectors)
> -       ldr     r0, =undefined_instruction
> -       ldr     r1, =_undefined_instruction
> -       str     r0, [r1]
> -       ldr     r0, =software_interrupt
> -       ldr     r1, =_software_interrupt
> -       str     r0, [r1]
> -       ldr     r0, =prefetch_abort
> -       ldr     r1, =_prefetch_abort
> -       str     r0, [r1]
> -       ldr     r0, =data_abort
> -       ldr     r1, =_data_abort
> -       str     r0, [r1]
> -       ldr     r0, =irq
> -       ldr     r1, =_irq
> -       str     r0, [r1]
> -       ldr     r0, =fiq
> -       ldr     r1, =_fiq
> -       str     r0, [r1]
> -       bx      lr
> -ENDPROC(arm_fixup_vectors)
> -#endif
> -
> -.section .text_exceptions
> -.globl extable
> -extable:
> -1:     b 1b                            /* barebox_arm_reset_vector */
> -#ifdef CONFIG_ARM_EXCEPTIONS
> -       ldr pc, _undefined_instruction  /* undefined instruction */
> -       ldr pc, _software_interrupt     /* software interrupt (SWI) */
> -       ldr pc, _prefetch_abort         /* prefetch abort */
> -       ldr pc, _data_abort             /* data abort */
> -1:     b 1b                            /* (reserved) */
> -       ldr pc, _irq                    /* irq (interrupt) */
> -       ldr pc, _fiq                    /* fiq (fast interrupt) */
> -.globl _undefined_instruction
> -_undefined_instruction: .word undefined_instruction
> -.globl _software_interrupt
> -_software_interrupt: .word software_interrupt
> -.globl _prefetch_abort
> -_prefetch_abort: .word prefetch_abort
> -.globl _data_abort
> -_data_abort: .word data_abort
> -.globl _irq
> -_irq: .word irq
> -.globl _fiq
> -_fiq: .word fiq
> -#else
> -1:     b 1b                            /* undefined instruction */
> -1:     b 1b                            /* software interrupt (SWI) */
> -1:     b 1b                            /* prefetch abort */
> -1:     b 1b                            /* data abort */
> -1:     b 1b                            /* (reserved) */
> -1:     b 1b                            /* irq (interrupt) */
> -1:     b 1b                            /* fiq (fast interrupt) */
> -#endif
> +       .align  11
> +       .globl  vectors
> +vectors:
> +       .align  7
> +       b       _do_bad_sync    /* Current EL Synchronous Thread */
> +
> +       .align  7
> +       b       _do_bad_irq     /* Current EL IRQ Thread */
> +
> +       .align  7
> +       b       _do_bad_fiq     /* Current EL FIQ Thread */
> +
> +       .align  7
> +       b       _do_bad_error   /* Current EL Error Thread */
> +
> +       .align  7
> +       b       _do_sync        /* Current EL Synchronous Handler */
> +
> +       .align  7
> +       b       _do_irq         /* Current EL IRQ Handler */
> +
> +       .align  7
> +       b       _do_fiq         /* Current EL FIQ Handler */
> +
> +       .align  7
> +       b       _do_error       /* Current EL Error Handler */
> +
> +
> +_do_bad_sync:
> +       exception_entry
> +       bl      do_bad_sync
> +
> +_do_bad_irq:
> +       exception_entry
> +       bl      do_bad_irq
> +
> +_do_bad_fiq:
> +       exception_entry
> +       bl      do_bad_fiq
> +
> +_do_bad_error:
> +       exception_entry
> +       bl      do_bad_error
> +
> +_do_sync:
> +       exception_entry
> +       bl      do_sync
> +
> +_do_irq:
> +       exception_entry
> +       bl      do_irq
> +
> +_do_fiq:
> +       exception_entry
> +       bl      do_fiq
> +
> +_do_error:
> +       exception_entry
> +       bl      do_error
>
>  .section .data
>  .align 4
> diff --git a/arch/arm/cpu/interrupts.c b/arch/arm/cpu/interrupts.c
> index fb4bb78..d42a5b1 100644
> --- a/arch/arm/cpu/interrupts.c
> +++ b/arch/arm/cpu/interrupts.c
> @@ -27,54 +27,8 @@
>  #include <asm/ptrace.h>
>  #include <asm/unwind.h>
>
> -/**
> - * Display current register set content
> - * @param[in] regs Guess what
> - */
> -void show_regs (struct pt_regs *regs)
> -{
> -       unsigned long flags;
> -       const char *processor_modes[] = {
> -       "USER_26",      "FIQ_26",       "IRQ_26",       "SVC_26",
> -       "UK4_26",       "UK5_26",       "UK6_26",       "UK7_26",
> -       "UK8_26",       "UK9_26",       "UK10_26",      "UK11_26",
> -       "UK12_26",      "UK13_26",      "UK14_26",      "UK15_26",
> -       "USER_32",      "FIQ_32",       "IRQ_32",       "SVC_32",
> -       "UK4_32",       "UK5_32",       "UK6_32",       "ABT_32",
> -       "UK8_32",       "UK9_32",       "UK10_32",      "UND_32",
> -       "UK12_32",      "UK13_32",      "UK14_32",      "SYS_32",
> -       };
> -
> -       flags = condition_codes (regs);
> -
> -       printf ("pc : [<%08lx>]    lr : [<%08lx>]\n"
> -               "sp : %08lx  ip : %08lx  fp : %08lx\n",
> -               instruction_pointer (regs),
> -               regs->ARM_lr, regs->ARM_sp, regs->ARM_ip, regs->ARM_fp);
> -       printf ("r10: %08lx  r9 : %08lx  r8 : %08lx\n",
> -               regs->ARM_r10, regs->ARM_r9, regs->ARM_r8);
> -       printf ("r7 : %08lx  r6 : %08lx  r5 : %08lx  r4 : %08lx\n",
> -               regs->ARM_r7, regs->ARM_r6, regs->ARM_r5, regs->ARM_r4);
> -       printf ("r3 : %08lx  r2 : %08lx  r1 : %08lx  r0 : %08lx\n",
> -               regs->ARM_r3, regs->ARM_r2, regs->ARM_r1, regs->ARM_r0);
> -       printf ("Flags: %c%c%c%c",
> -               flags & PSR_N_BIT ? 'N' : 'n',
> -               flags & PSR_Z_BIT ? 'Z' : 'z',
> -               flags & PSR_C_BIT ? 'C' : 'c', flags & PSR_V_BIT ? 'V' : 'v');
> -       printf ("  IRQs %s  FIQs %s  Mode %s%s\n",
> -               interrupts_enabled (regs) ? "on" : "off",
> -               fast_interrupts_enabled (regs) ? "on" : "off",
> -               processor_modes[processor_mode (regs)],
> -               thumb_mode (regs) ? " (T)" : "");
> -#ifdef CONFIG_ARM_UNWIND
> -       unwind_backtrace(regs);
> -#endif
> -}
> -
>  static void __noreturn do_exception(struct pt_regs *pt_regs)
>  {
> -       show_regs(pt_regs);
> -
>         panic("");
>  }
>
> @@ -121,14 +75,6 @@ void do_prefetch_abort (struct pt_regs *pt_regs)
>   */
>  void do_data_abort (struct pt_regs *pt_regs)
>  {
> -       u32 far;
> -
> -       asm volatile ("mrc     p15, 0, %0, c6, c0, 0" : "=r" (far) : : "cc");
> -
> -       printf("unable to handle %s at address 0x%08x\n",
> -                       far < PAGE_SIZE ? "NULL pointer dereference" :
> -                       "paging request", far);
> -
>         do_exception(pt_regs);
>  }
>
> @@ -156,6 +102,43 @@ void do_irq (struct pt_regs *pt_regs)
>         do_exception(pt_regs);
>  }
>
> +void do_bad_sync(struct pt_regs *pt_regs)
> +{
> +       printf("bad sync\n");
> +       do_exception(pt_regs);
> +}
> +
> +void do_bad_irq(struct pt_regs *pt_regs)
> +{
> +       printf("bad irq\n");
> +       do_exception(pt_regs);
> +}
> +
> +void do_bad_fiq(struct pt_regs *pt_regs)
> +{
> +       printf("bad fiq\n");
> +       do_exception(pt_regs);
> +}
> +
> +void do_bad_error(struct pt_regs *pt_regs)
> +{
> +       printf("bad error\n");
> +       do_exception(pt_regs);
> +}
> +
> +void do_sync(struct pt_regs *pt_regs)
> +{
> +       printf("sync exception\n");
> +       do_exception(pt_regs);
> +}
> +
> +
> +void do_error(struct pt_regs *pt_regs)
> +{
> +       printf("error exception\n");
> +       do_exception(pt_regs);
> +}
> +
>  extern volatile int arm_ignore_data_abort;
>  extern volatile int arm_data_abort_occurred;
>
> diff --git a/arch/arm/cpu/lowlevel.S b/arch/arm/cpu/lowlevel.S
> index e5baa12..0691b2b 100644
> --- a/arch/arm/cpu/lowlevel.S
> +++ b/arch/arm/cpu/lowlevel.S
> @@ -1,60 +1,43 @@
>  #include <linux/linkage.h>
>  #include <init.h>
>  #include <asm/system.h>
> +#include <asm/gic.h>
> +#include <asm-generic/memory_layout.h>
>
>  .section ".text_bare_init_","ax"
> +
>  ENTRY(arm_cpu_lowlevel_init)
> -       /* save lr, since it may be banked away with a processor mode change */
> -       mov     r2, lr
> -       /* set the cpu to SVC32 mode, mask irq and fiq */
> -       mrs     r12, cpsr
> -       bic     r12, r12, #0x1f
> -       orr     r12, r12, #0xd3
> -       msr     cpsr, r12
> -
> -#if __LINUX_ARM_ARCH__ >= 6
> -       /*
> -        * ICIALLU: Invalidate all instruction caches to PoU,
> -        * includes flushing of branch predictors.
> -        * Even if the i-cache is off it might contain stale entries
> -        * that are better discarded before enabling the cache.
> -        * Architectually this is even possible after a cold reset.
> -        */
> -       mcr     p15, 0, r12, c7, c5, 0
> -       /* DSB, ensure completion of the invalidation */
> -       mcr     p15, 0, r12, c7, c10, 4
> -       /*
> -        * ISB, ensure instruction fetch path is in sync.
> -        * Note that the ARM Architecture Reference Manual, ARMv7-A and ARMv7-R
> -        * edition (ARM DDI 0406C.c) doesn't define this instruction in the
> -        * ARMv6 part (D12.7.10). It only has: "Support of additional
> -        * operations is IMPLEMENTATION DEFINED".
> -        * But an earlier version of the ARMARM (ARM DDI 0100I) does define it
> -        * as "Flush prefetch buffer (PrefetchFlush)".
> -        */
> -       mcr     p15, 0, r12, c7, c5, 4
> -#endif
> -
> -       /* disable MMU stuff and data/unified caches */
> -       mrc     p15, 0, r12, c1, c0, 0          /* SCTLR */
> -       bic     r12, r12, #(CR_M | CR_C | CR_B)
> -       bic     r12, r12, #(CR_S | CR_R | CR_V)
> -
> -       /* enable instruction cache */
> -       orr     r12, r12, #CR_I
> -
> -#if __LINUX_ARM_ARCH__ >= 6
> -       orr     r12, r12, #CR_U
> -       bic     r12, r12, #CR_A
> -#else
> -       orr     r12, r12, #CR_A
> -#endif
> -
> -#ifdef __ARMEB__
> -       orr     r12, r12, #CR_B
> -#endif
> -
> -       mcr     p15, 0, r12, c1, c0, 0          /* SCTLR */
> -
> -       mov     pc, r2
> +       adr     x0, vectors
> +       mrs     x1, CurrentEL
> +       cmp     x1, #0xC                /* Check EL3 state */
> +       b.eq    1f
> +       cmp     x1, #0x8                /* Check EL2 state */
> +       b.eq    2f
> +       cmp     x1, #0x4                /* Check EL1 state */
> +       b.eq    3f
> +
> +1:
> +       msr     vbar_el3, x0
> +       mov     x0, #1                  /* Non-Secure EL0/1 */
> +       orr     x0, x0, #(1 << 10)      /* 64-bit EL2 */
> +       msr     scr_el3, x0
> +       msr     cptr_el3, xzr
> +       b       done
> +
> +2:
> +       msr     vbar_el2, x0
> +       mov     x0, #0x33ff             /* Enable FP/SIMD */
> +       msr     cptr_el2, x0
> +       b       done
> +
> +
> +3:
> +       msr     vbar_el1, x0
> +       mov     x0, #(3 << 20)          /* Enable FP/SIMD */
> +       msr     cpacr_el1, x0
> +       b       done
> +
> +done:
> +       ret
> +
>  ENDPROC(arm_cpu_lowlevel_init)
> diff --git a/arch/arm/cpu/mmu.c b/arch/arm/cpu/mmu.c
> index a31bce4..b171f80 100644
> --- a/arch/arm/cpu/mmu.c
> +++ b/arch/arm/cpu/mmu.c
> @@ -32,54 +32,8 @@
>
>  #include "mmu.h"
>
> -static unsigned long *ttb;
> -
> -static void create_sections(unsigned long virt, unsigned long phys, int size_m,
> -               unsigned int flags)
> -{
> -       int i;
> -
> -       phys >>= 20;
> -       virt >>= 20;
> -
> -       for (i = size_m; i > 0; i--, virt++, phys++)
> -               ttb[virt] = (phys << 20) | flags;
> -
> -       __mmu_cache_flush();
> -}
> -
> -/*
> - * Do it the simple way for now and invalidate the entire
> - * tlb
> - */
> -static inline void tlb_invalidate(void)
> -{
> -       asm volatile (
> -               "mov    r0, #0\n"
> -               "mcr    p15, 0, r0, c7, c10, 4; @ drain write buffer\n"
> -               "mcr    p15, 0, r0, c8, c6, 0;  @ invalidate D TLBs\n"
> -               "mcr    p15, 0, r0, c8, c5, 0;  @ invalidate I TLBs\n"
> -               :
> -               :
> -               : "r0"
> -       );
> -}
> -
> -#define PTE_FLAGS_CACHED_V7 (PTE_EXT_TEX(1) | PTE_BUFFERABLE | PTE_CACHEABLE)
> -#define PTE_FLAGS_WC_V7 PTE_EXT_TEX(1)
> -#define PTE_FLAGS_UNCACHED_V7 (0)
> -#define PTE_FLAGS_CACHED_V4 (PTE_SMALL_AP_UNO_SRW | PTE_BUFFERABLE | PTE_CACHEABLE)
> -#define PTE_FLAGS_UNCACHED_V4 PTE_SMALL_AP_UNO_SRW
> -
> -/*
> - * PTE flags to set cached and uncached areas.
> - * This will be determined at runtime.
> - */
> -static uint32_t pte_flags_cached;
> -static uint32_t pte_flags_wc;
> -static uint32_t pte_flags_uncached;
> -
> -#define PTE_MASK ((1 << 12) - 1)
> +static uint64_t *ttb;
> +static int free_idx;
>
>  static void arm_mmu_not_initialized_error(void)
>  {
> @@ -92,329 +46,175 @@ static void arm_mmu_not_initialized_error(void)
>         panic("MMU not initialized\n");
>  }
>
> +
>  /*
> - * Create a second level translation table for the given virtual address.
> - * We initially create a flat uncached mapping on it.
> - * Not yet exported, but may be later if someone finds use for it.
> + * Do it the simple way for now and invalidate the entire
> + * tlb
>   */
> -static u32 *arm_create_pte(unsigned long virt)
> +static inline void tlb_invalidate(void)
>  {
> -       u32 *table;
> -       int i;
> +       unsigned int el = current_el();
>
> -       table = memalign(0x400, 0x400);
> +       dsb();
>
> -       if (!ttb)
> -               arm_mmu_not_initialized_error();
> -
> -       ttb[virt >> 20] = (unsigned long)table | PMD_TYPE_TABLE;
> -
> -       for (i = 0; i < 256; i++) {
> -               table[i] = virt | PTE_TYPE_SMALL | pte_flags_uncached;
> -               virt += PAGE_SIZE;
> -       }
> +       if (el == 1)
> +               __asm__ __volatile__("tlbi alle1\n\t" : : : "memory");
> +       else if (el == 2)
> +               __asm__ __volatile__("tlbi alle2\n\t" : : : "memory");
> +       else if (el == 3)
> +               __asm__ __volatile__("tlbi alle3\n\t" : : : "memory");
>
> -       return table;
> +       dsb();
> +       isb();
>  }
>
> -static u32 *find_pte(unsigned long adr)
> +static int level2shift(int level)
>  {
> -       u32 *table;
> -
> -       if (!ttb)
> -               arm_mmu_not_initialized_error();
> -
> -       if ((ttb[adr >> 20] & PMD_TYPE_MASK) != PMD_TYPE_TABLE) {
> -               struct memory_bank *bank;
> -               int i = 0;
> +       /* Page is 12 bits wide, every level translates 9 bits */
> +       return (12 + 9 * (3 - level));
> +}
>
> -               /*
> -                * This should only be called for page mapped memory inside our
> -                * memory banks. It's a bug to call it with section mapped memory
> -                * locations.
> -                */
> -               pr_crit("%s: TTB for address 0x%08lx is not of type table\n",
> -                               __func__, adr);
> -               pr_crit("Memory banks:\n");
> -               for_each_memory_bank(bank)
> -                       pr_crit("#%d 0x%08lx - 0x%08lx\n", i, bank->start,
> -                                       bank->start + bank->size - 1);
> -               BUG();
> -       }
> +static uint64_t level2mask(int level)
> +{
> +       uint64_t mask = -EINVAL;
>
> -       /* find the coarse page table base address */
> -       table = (u32 *)(ttb[adr >> 20] & ~0x3ff);
> +       if (level == 1)
> +               mask = L1_ADDR_MASK;
> +       else if (level == 2)
> +               mask = L2_ADDR_MASK;
> +       else if (level == 3)
> +               mask = L3_ADDR_MASK;
>
> -       /* find second level descriptor */
> -       return &table[(adr >> PAGE_SHIFT) & 0xff];
> +       return mask;
>  }
>
> -static void dma_flush_range(unsigned long start, unsigned long end)
> +static int pte_type(uint64_t *pte)
>  {
> -       __dma_flush_range(start, end);
> -       if (outer_cache.flush_range)
> -               outer_cache.flush_range(start, end);
> +       return *pte & PMD_TYPE_MASK;
>  }
>
> -static void dma_inv_range(unsigned long start, unsigned long end)
> +static void set_table(uint64_t *pt, uint64_t *table_addr)
>  {
> -       if (outer_cache.inv_range)
> -               outer_cache.inv_range(start, end);
> -       __dma_inv_range(start, end);
> +       uint64_t val;
> +
> +       val = PMD_TYPE_TABLE | (uint64_t)table_addr;
> +       *pt = val;
>  }
>
> -static int __remap_range(void *_start, size_t size, u32 pte_flags)
> +static uint64_t *create_table(void)
>  {
> -       unsigned long start = (unsigned long)_start;
> -       u32 *p;
> -       int numentries, i;
> -
> -       numentries = size >> PAGE_SHIFT;
> -       p = find_pte(start);
> -
> -       for (i = 0; i < numentries; i++) {
> -               p[i] &= ~PTE_MASK;
> -               p[i] |= pte_flags | PTE_TYPE_SMALL;
> -       }
> +       uint64_t *new_table = ttb + free_idx * GRANULE_SIZE;
>
> -       dma_flush_range((unsigned long)p,
> -                       (unsigned long)p + numentries * sizeof(u32));
> +       /* Mark all entries as invalid */
> +       memset(new_table, 0, GRANULE_SIZE);
>
> -       tlb_invalidate();
> +       free_idx++;
>
> -       return 0;
> +       return new_table;
>  }
>
> -int arch_remap_range(void *start, size_t size, unsigned flags)
> +static uint64_t *get_level_table(uint64_t *pte)
>  {
> -       u32 pte_flags;
> -
> -       switch (flags) {
> -       case MAP_CACHED:
> -               pte_flags = pte_flags_cached;
> -               break;
> -       case MAP_UNCACHED:
> -               pte_flags = pte_flags_uncached;
> -               break;
> -       default:
> -               return -EINVAL;
> +       uint64_t *table = (uint64_t *)(*pte & XLAT_ADDR_MASK);
> +
> +       if (pte_type(pte) != PMD_TYPE_TABLE) {
> +               table = create_table();
> +               set_table(pte, table);
>         }
>
> -       return __remap_range(start, size, pte_flags);
> +       return table;
>  }
>
> -void *map_io_sections(unsigned long phys, void *_start, size_t size)
> +static uint64_t *find_pte(uint64_t addr)
>  {
> -       unsigned long start = (unsigned long)_start, sec;
> +       uint64_t *pte;
> +       uint64_t block_shift;
> +       uint64_t idx;
> +       int i;
>
> -       phys >>= 20;
> -       for (sec = start; sec < start + size; sec += (1 << 20))
> -               ttb[sec >> 20] = (phys++ << 20) | PMD_SECT_DEF_UNCACHED;
> +       pte = ttb;
>
> -       dma_flush_range((unsigned long)ttb, (unsigned long)ttb + 0x4000);
> -       tlb_invalidate();
> -       return _start;
> +       for (i = 1; i < 4; i++) {
> +               block_shift = level2shift(i);
> +               idx = (addr & level2mask(i)) >> block_shift;
> +               pte += idx;
> +
> +               if ((pte_type(pte) != PMD_TYPE_TABLE) || (block_shift <= GRANULE_SIZE_SHIFT))
> +                       break;
> +               else
> +                       pte = (uint64_t *)(*pte & XLAT_ADDR_MASK);
> +       }
> +
> +       return pte;
>  }
>
> -/*
> - * remap the memory bank described by mem cachable and
> - * bufferable
> - */
> -static int arm_mmu_remap_sdram(struct memory_bank *bank)
> +static void map_region(uint64_t virt, uint64_t phys, uint64_t size, uint64_t attr)
>  {
> -       unsigned long phys = (unsigned long)bank->start;
> -       unsigned long ttb_start = phys >> 20;
> -       unsigned long ttb_end = (phys >> 20) + (bank->size >> 20);
> -       unsigned long num_ptes = bank->size >> 12;
> -       int i, pte;
> -       u32 *ptes;
> +       uint64_t block_size;
> +       uint64_t block_shift;
> +       uint64_t *pte;
> +       uint64_t idx;
> +       uint64_t addr;
> +       uint64_t *table;
> +       int level;
>
> -       pr_debug("remapping SDRAM from 0x%08lx (size 0x%08lx)\n",
> -                       phys, bank->size);
> +       if (!ttb)
> +               arm_mmu_not_initialized_error();
>
> -       /*
> -        * We replace each 1MiB section in this range with second level page
> -        * tables, therefore we must have 1Mib aligment here.
> -        */
> -       if ((phys & (SZ_1M - 1)) || (bank->size & (SZ_1M - 1)))
> -               return -EINVAL;
> +       addr = virt;
>
> -       ptes = xmemalign(PAGE_SIZE, num_ptes * sizeof(u32));
> +       attr &= ~(PMD_TYPE_SECT);
>
> -       pr_debug("ptes: 0x%p ttb_start: 0x%08lx ttb_end: 0x%08lx\n",
> -                       ptes, ttb_start, ttb_end);
> +       while (size) {
> +               table = ttb;
> +               for (level = 1; level < 4; level++) {
> +                       block_shift = level2shift(level);
> +                       idx = (addr & level2mask(level)) >> block_shift;
> +                       block_size = (1 << block_shift);
>
> -       for (i = 0; i < num_ptes; i++) {
> -               ptes[i] = (phys + i * PAGE_SIZE) | PTE_TYPE_SMALL |
> -                       pte_flags_cached;
> -       }
> +                       pte = table + idx;
>
> -       pte = 0;
> +                       if (level == 3)
> +                               attr |= PMD_TYPE_PAGE;
> +                       else
> +                               attr |= PMD_TYPE_SECT;
>
> -       for (i = ttb_start; i < ttb_end; i++) {
> -               ttb[i] = (unsigned long)(&ptes[pte]) | PMD_TYPE_TABLE |
> -                       (0 << 4);
> -               pte += 256;
> -       }
> +                       if (size >= block_size && IS_ALIGNED(addr, block_size)) {
> +                               *pte = phys | attr;
> +                               addr += block_size;
> +                               phys += block_size;
> +                               size -= block_size;
> +                               break;
>
> -       dma_flush_range((unsigned long)ttb, (unsigned long)ttb + 0x4000);
> -       dma_flush_range((unsigned long)ptes,
> -                       (unsigned long)ptes + num_ptes * sizeof(u32));
> +                       }
>
> -       tlb_invalidate();
> +                       table = get_level_table(pte);
> +               }
>
> -       return 0;
> +       }
>  }
> -/*
> - * We have 8 exception vectors and the table consists of absolute
> - * jumps, so we need 8 * 4 bytes for the instructions and another
> - * 8 * 4 bytes for the addresses.
> - */
> -#define ARM_VECTORS_SIZE       (sizeof(u32) * 8 * 2)
> -
> -#define ARM_HIGH_VECTORS       0xffff0000
> -#define ARM_LOW_VECTORS                0x0
>
> -/**
> - * create_vector_table - create a vector table at given address
> - * @adr - The address where the vector table should be created
> - *
> - * After executing this function the vector table is found at the
> - * virtual address @adr.
> - */
> -static void create_vector_table(unsigned long adr)
> +static void create_sections(uint64_t virt, uint64_t phys, uint64_t size_m, uint64_t flags)
>  {
> -       struct resource *vectors_sdram;
> -       void *vectors;
> -       u32 *exc;
> -       int idx;
> -
> -       vectors_sdram = request_sdram_region("vector table", adr, SZ_4K);
> -       if (vectors_sdram) {
> -               /*
> -                * The vector table address is inside the SDRAM physical
> -                * address space. Use the existing identity mapping for
> -                * the vector table.
> -                */
> -               pr_debug("Creating vector table, virt = phys = 0x%08lx\n", adr);
> -               vectors = (void *)vectors_sdram->start;
> -       } else {
> -               /*
> -                * The vector table address is outside of SDRAM. Create
> -                * a secondary page table for the section and map
> -                * allocated memory to the vector address.
> -                */
> -               vectors = xmemalign(PAGE_SIZE, PAGE_SIZE);
> -               pr_debug("Creating vector table, virt = 0x%p, phys = 0x%08lx\n",
> -                        vectors, adr);
> -               exc = arm_create_pte(adr);
> -               idx = (adr & (SZ_1M - 1)) >> PAGE_SHIFT;
> -               exc[idx] = (u32)vectors | PTE_TYPE_SMALL | pte_flags_cached;
> -       }
>
> -       arm_fixup_vectors();
> -
> -       memset(vectors, 0, PAGE_SIZE);
> -       memcpy(vectors, __exceptions_start, __exceptions_stop - __exceptions_start);
> +       map_region(virt, phys, size_m, flags);
>  }
>
> -/**
> - * set_vector_table - let CPU use the vector table at given address
> - * @adr - The address of the vector table
> - *
> - * Depending on the CPU the possibilities differ. ARMv7 and later allow
> - * to map the vector table to arbitrary addresses. Other CPUs only allow
> - * vectors at 0xffff0000 or at 0x0.
> - */
> -static int set_vector_table(unsigned long adr)
> +void *map_io_sections(uint64_t phys, void *_start, size_t size)
>  {
> -       u32 cr;
> -
> -       if (cpu_architecture() >= CPU_ARCH_ARMv7) {
> -               pr_debug("Vectors are at 0x%08lx\n", adr);
> -               set_vbar(adr);
> -               return 0;
> -       }
>
> -       if (adr == ARM_HIGH_VECTORS) {
> -               cr = get_cr();
> -               cr |= CR_V;
> -               set_cr(cr);
> -               cr = get_cr();
> -               if (cr & CR_V) {
> -                       pr_debug("Vectors are at 0x%08lx\n", adr);
> -                       return 0;
> -               } else {
> -                       return -EINVAL;
> -               }
> -       }
> -
> -       if (adr == ARM_LOW_VECTORS) {
> -               cr = get_cr();
> -               cr &= ~CR_V;
> -               set_cr(cr);
> -               cr = get_cr();
> -               if (cr & CR_V) {
> -                       return -EINVAL;
> -               } else {
> -                       pr_debug("Vectors are at 0x%08lx\n", adr);
> -                       return 0;
> -               }
> -       }
> +       map_region((uint64_t)_start, phys, (uint64_t)size, PMD_SECT_DEF_UNCACHED);
>
> -       return -EINVAL;
> +       tlb_invalidate();
> +       return _start;
>  }
>
> -static void create_zero_page(void)
> -{
> -       struct resource *zero_sdram;
> -       u32 *zero;
> -
> -       zero_sdram = request_sdram_region("zero page", 0x0, SZ_4K);
> -       if (zero_sdram) {
> -               /*
> -                * Here we would need to set the second level page table
> -                * entry to faulting. This is not yet implemented.
> -                */
> -               pr_debug("zero page is in SDRAM area, currently not supported\n");
> -       } else {
> -               zero = arm_create_pte(0x0);
> -               zero[0] = 0;
> -               pr_debug("Created zero page\n");
> -       }
> -}
>
> -/*
> - * Map vectors and zero page
> - */
> -static void vectors_init(void)
> +int arch_remap_range(void *_start, size_t size, unsigned flags)
>  {
> -       /*
> -        * First try to use the vectors where they actually are, works
> -        * on ARMv7 and later.
> -        */
> -       if (!set_vector_table((unsigned long)__exceptions_start)) {
> -               arm_fixup_vectors();
> -               create_zero_page();
> -               return;
> -       }
> -
> -       /*
> -        * Next try high vectors at 0xffff0000.
> -        */
> -       if (!set_vector_table(ARM_HIGH_VECTORS)) {
> -               create_zero_page();
> -               create_vector_table(ARM_HIGH_VECTORS);
> -               return;
> -       }
> +       map_region((uint64_t)_start, (uint64_t)_start, (uint64_t)size, flags);
>
> -       /*
> -        * As a last resort use low vectors at 0x0. With this we can't
> -        * set the zero page to faulting and can't catch NULL pointer
> -        * exceptions.
> -        */
> -       set_vector_table(ARM_LOW_VECTORS);
> -       create_vector_table(ARM_LOW_VECTORS);
> +       return 0;
>  }
>
>  /*
> @@ -423,7 +223,6 @@ static void vectors_init(void)
>  static int mmu_init(void)
>  {
>         struct memory_bank *bank;
> -       int i;
>
>         if (list_empty(&memory_banks))
>                 /*
> @@ -434,56 +233,31 @@ static int mmu_init(void)
>                  */
>                 panic("MMU: No memory bank found! Cannot continue\n");
>
> -       arm_set_cache_functions();
> -
> -       if (cpu_architecture() >= CPU_ARCH_ARMv7) {
> -               pte_flags_cached = PTE_FLAGS_CACHED_V7;
> -               pte_flags_wc = PTE_FLAGS_WC_V7;
> -               pte_flags_uncached = PTE_FLAGS_UNCACHED_V7;
> -       } else {
> -               pte_flags_cached = PTE_FLAGS_CACHED_V4;
> -               pte_flags_wc = PTE_FLAGS_UNCACHED_V4;
> -               pte_flags_uncached = PTE_FLAGS_UNCACHED_V4;
> -       }
> -
> -       if (get_cr() & CR_M) {
> -               /*
> -                * Early MMU code has already enabled the MMU. We assume a
> -                * flat 1:1 section mapping in this case.
> -                */
> -               asm volatile ("mrc  p15,0,%0,c2,c0,0" : "=r"(ttb));
> -
> -               /* Clear unpredictable bits [13:0] */
> -               ttb = (unsigned long *)((unsigned long)ttb & ~0x3fff);
> -
> +       if (get_sctlr() & CR_M) {
> +               ttb = (uint64_t *)get_ttbr(1);
>                 if (!request_sdram_region("ttb", (unsigned long)ttb, SZ_16K))
>                         /*
> -                        * This can mean that:
> -                        * - the early MMU code has put the ttb into a place
> -                        *   which we don't have inside our available memory
> -                        * - Somebody else has occupied the ttb region which means
> -                        *   the ttb will get corrupted.
> -                        */
> +                       * This can mean that:
> +                       * - the early MMU code has put the ttb into a place
> +                       *   which we don't have inside our available memory
> +                       * - Somebody else has occupied the ttb region which means
> +                       *   the ttb will get corrupted.
> +                       */
>                         pr_crit("Critical Error: Can't request SDRAM region for ttb at %p\n",
> -                                       ttb);
> +                               ttb);
>         } else {
> -               ttb = memalign(0x10000, 0x4000);
> -       }
> +               ttb = memalign(0x1000, SZ_16K);
> +               free_idx = 1;
>
> -       pr_debug("ttb: 0x%p\n", ttb);
> +               memset(ttb, 0, GRANULE_SIZE);
>
> -       /* Set the ttb register */
> -       asm volatile ("mcr  p15,0,%0,c2,c0,0" : : "r"(ttb) /*:*/);
> +               set_ttbr_tcr_mair(current_el(), (uint64_t)ttb, TCR_FLAGS, MEMORY_ATTR);
> +       }
>
> -       /* Set the Domain Access Control Register */
> -       i = 0x3;
> -       asm volatile ("mcr  p15,0,%0,c3,c0,0" : : "r"(i) /*:*/);
> +       pr_debug("ttb: 0x%p\n", ttb);
>
>         /* create a flat mapping using 1MiB sections */
> -       create_sections(0, 0, PAGE_SIZE, PMD_SECT_AP_WRITE | PMD_SECT_AP_READ |
> -                       PMD_TYPE_SECT);
> -
> -       vectors_init();
> +       create_sections(0, 0, GRANULE_SIZE, PMD_SECT_AP_WRITE | PMD_SECT_AP_READ | PMD_TYPE_SECT);
>
>         /*
>          * First remap sdram cached using sections.
> @@ -491,92 +265,70 @@ static int mmu_init(void)
>          * below
>          */
>         for_each_memory_bank(bank)
> -               create_sections(bank->start, bank->start, bank->size >> 20,
> -                               PMD_SECT_DEF_CACHED);
> -
> -       __mmu_cache_on();
> -
> -       /*
> -        * Now that we have the MMU and caches on remap sdram again using
> -        * page tables
> -        */
> -       for_each_memory_bank(bank)
> -               arm_mmu_remap_sdram(bank);
> +               create_sections(bank->start, bank->start, bank->size, PMD_SECT_DEF_CACHED);
>
>         return 0;
>  }
>  mmu_initcall(mmu_init);
>
> -void *dma_alloc_coherent(size_t size, dma_addr_t *dma_handle)
> +void mmu_enable(void)
>  {
> -       void *ret;
> -
> -       size = PAGE_ALIGN(size);
> -       ret = xmemalign(PAGE_SIZE, size);
> -       if (dma_handle)
> -               *dma_handle = (dma_addr_t)ret;
> -
> -       dma_inv_range((unsigned long)ret, (unsigned long)ret + size);
> +       if (!ttb)
> +               arm_mmu_not_initialized_error();
>
> -       __remap_range(ret, size, pte_flags_uncached);
> +       if (!(get_sctlr() & CR_M)) {
>
> -       return ret;
> +               isb();
> +               set_sctlr(get_sctlr() | CR_M | CR_C | CR_I);
> +       }
>  }
>
> -void *dma_alloc_writecombine(size_t size, dma_addr_t *dma_handle)
> +void mmu_disable(void)
>  {
> -       void *ret;
> +       unsigned int sctlr;
> +
> +       if (!ttb)
> +               arm_mmu_not_initialized_error();
>
> -       size = PAGE_ALIGN(size);
> -       ret = xmemalign(PAGE_SIZE, size);
> -       if (dma_handle)
> -               *dma_handle = (dma_addr_t)ret;
> +       sctlr = get_sctlr();
> +       sctlr &= ~(CR_M | CR_C | CR_I);
>
> -       dma_inv_range((unsigned long)ret, (unsigned long)ret + size);
> +       tlb_invalidate();
>
> -       __remap_range(ret, size, pte_flags_wc);
> +       dsb();
> +       isb();
>
> -       return ret;
> -}
> +       set_sctlr(sctlr);
>
> -unsigned long virt_to_phys(volatile void *virt)
> -{
> -       return (unsigned long)virt;
> +       dsb();
> +       isb();
>  }
>
> -void *phys_to_virt(unsigned long phys)
> +void mmu_early_enable(uint64_t membase, uint64_t memsize, uint64_t _ttb)
>  {
> -       return (void *)phys;
> -}
> +       ttb = (uint64_t *)_ttb;
>
> -void dma_free_coherent(void *mem, dma_addr_t dma_handle, size_t size)
> -{
> -       size = PAGE_ALIGN(size);
> -       __remap_range(mem, size, pte_flags_cached);
> +       memset(ttb, 0, GRANULE_SIZE);
> +       free_idx = 1;
> +
> +       set_ttbr_tcr_mair(current_el(), (uint64_t)ttb, TCR_FLAGS, MEMORY_ATTR);
>
> -       free(mem);
> +       create_sections(0, 0, 4096, PMD_SECT_AP_WRITE |
> +                       PMD_SECT_AP_READ | PMD_TYPE_SECT);
> +
> +       create_sections(membase, membase, memsize, PMD_SECT_AP_WRITE |
> +               PMD_SECT_AP_READ | PMD_TYPE_SECT | PMD_SECT_WB);
> +
> +       isb();
> +       set_sctlr(get_sctlr() | CR_M);
>  }
>
> -void dma_sync_single_for_cpu(unsigned long address, size_t size,
> -                            enum dma_data_direction dir)
> +unsigned long virt_to_phys(volatile void *virt)
>  {
> -       if (dir != DMA_TO_DEVICE) {
> -               if (outer_cache.inv_range)
> -                       outer_cache.inv_range(address, address + size);
> -               __dma_inv_range(address, address + size);
> -       }
> +       return (unsigned long)virt;
>  }
>
> -void dma_sync_single_for_device(unsigned long address, size_t size,
> -                               enum dma_data_direction dir)
> +void *phys_to_virt(unsigned long phys)
>  {
> -       if (dir == DMA_FROM_DEVICE) {
> -               __dma_inv_range(address, address + size);
> -               if (outer_cache.inv_range)
> -                       outer_cache.inv_range(address, address + size);
> -       } else {
> -               __dma_clean_range(address, address + size);
> -               if (outer_cache.clean_range)
> -                       outer_cache.clean_range(address, address + size);
> -       }
> +       return (void *)phys;
>  }
> diff --git a/arch/arm/cpu/mmu.h b/arch/arm/cpu/mmu.h
> index 79ebc80..a20adec 100644
> --- a/arch/arm/cpu/mmu.h
> +++ b/arch/arm/cpu/mmu.h
> @@ -1,6 +1,159 @@
>  #ifndef __ARM_MMU_H
>  #define __ARM_MMU_H
>
> +#define UL(x)          _AC(x, UL)
> +
> +#define UNUSED_DESC                0x6EbAAD0BBADbA6E0
> +
> +#define VA_START                   0x0
> +#define BITS_PER_VA                33
> +
> +/* Granule size of 4KB is being used */
> +#define GRANULE_SIZE_SHIFT         12
> +#define GRANULE_SIZE               (1 << GRANULE_SIZE_SHIFT)
> +#define XLAT_ADDR_MASK             ((1UL << BITS_PER_VA) - GRANULE_SIZE)
> +#define GRANULE_SIZE_MASK          ((1 << GRANULE_SIZE_SHIFT) - 1)
> +
> +#define BITS_RESOLVED_PER_LVL   (GRANULE_SIZE_SHIFT - 3)
> +#define L1_ADDR_SHIFT           (GRANULE_SIZE_SHIFT + BITS_RESOLVED_PER_LVL * 2)
> +#define L2_ADDR_SHIFT           (GRANULE_SIZE_SHIFT + BITS_RESOLVED_PER_LVL * 1)
> +#define L3_ADDR_SHIFT           (GRANULE_SIZE_SHIFT + BITS_RESOLVED_PER_LVL * 0)
> +
> +
> +#define L1_ADDR_MASK     (((1UL << BITS_RESOLVED_PER_LVL) - 1) << L1_ADDR_SHIFT)
> +#define L2_ADDR_MASK     (((1UL << BITS_RESOLVED_PER_LVL) - 1) << L2_ADDR_SHIFT)
> +#define L3_ADDR_MASK     (((1UL << BITS_RESOLVED_PER_LVL) - 1) << L3_ADDR_SHIFT)
> +
> +/* These macros give the size of the region addressed by each entry of a xlat
> +   table at any given level */
> +#define L3_XLAT_SIZE               (1UL << L3_ADDR_SHIFT)
> +#define L2_XLAT_SIZE               (1UL << L2_ADDR_SHIFT)
> +#define L1_XLAT_SIZE               (1UL << L1_ADDR_SHIFT)
> +
> +#define GRANULE_MASK   GRANULE_SIZE
> +
> +/*
> + * Memory types
> + */
> +#define MT_DEVICE_NGNRNE       0
> +#define MT_DEVICE_NGNRE                1
> +#define MT_DEVICE_GRE          2
> +#define MT_NORMAL_NC           3
> +#define MT_NORMAL              4
> +
> +#define MEMORY_ATTRIBUTES      ((0x00 << (MT_DEVICE_NGNRNE*8)) |       \
> +               (0x04 << (MT_DEVICE_NGNRE*8)) |         \
> +               (0x0c << (MT_DEVICE_GRE*8)) |           \
> +               (0x44 << (MT_NORMAL_NC*8)) |            \
> +               (UL(0xff) << (MT_NORMAL*8)))
> +
> +/*
> + * Hardware page table definitions.
> + *
> + * Level 2 descriptor (PMD).
> + */
> +#define PMD_TYPE_MASK          (3 << 0)
> +#define PMD_TYPE_FAULT         (0 << 0)
> +#define PMD_TYPE_TABLE         (3 << 0)
> +#define PMD_TYPE_PAGE          (3 << 0)
> +#define PMD_TYPE_SECT          (1 << 0)
> +
> +/*
> + * Section
> + */
> +#define PMD_SECT_NON_SHARE     (0 << 8)
> +#define PMD_SECT_OUTER_SHARE   (2 << 8)
> +#define PMD_SECT_INNER_SHARE   (3 << 8)
> +#define PMD_SECT_AF            (1 << 10)
> +#define PMD_SECT_NG            (1 << 11)
> +#define PMD_SECT_PXN           (UL(1) << 53)
> +#define PMD_SECT_UXN           (UL(1) << 54)
> +
> +/*
> + * AttrIndx[2:0]
> + */
> +#define PMD_ATTRINDX(t)                ((t) << 2)
> +#define PMD_ATTRINDX_MASK      (7 << 2)
> +
> +/*
> + * TCR flags.
> + */
> +#define TCR_T0SZ(x)            ((64 - (x)) << 0)
> +#define TCR_IRGN_NC            (0 << 8)
> +#define TCR_IRGN_WBWA          (1 << 8)
> +#define TCR_IRGN_WT            (2 << 8)
> +#define TCR_IRGN_WBNWA         (3 << 8)
> +#define TCR_IRGN_MASK          (3 << 8)
> +#define TCR_ORGN_NC            (0 << 10)
> +#define TCR_ORGN_WBWA          (1 << 10)
> +#define TCR_ORGN_WT            (2 << 10)
> +#define TCR_ORGN_WBNWA         (3 << 10)
> +#define TCR_ORGN_MASK          (3 << 10)
> +#define TCR_SHARED_NON         (0 << 12)
> +#define TCR_SHARED_OUTER       (2 << 12)
> +#define TCR_SHARED_INNER       (3 << 12)
> +#define TCR_TG0_4K             (0 << 14)
> +#define TCR_TG0_64K            (1 << 14)
> +#define TCR_TG0_16K            (2 << 14)
> +#define TCR_EL1_IPS_BITS       (UL(3) << 32)   /* 42 bits physical address */
> +#define TCR_EL2_IPS_BITS       (3 << 16)       /* 42 bits physical address */
> +#define TCR_EL3_IPS_BITS       (3 << 16)       /* 42 bits physical address */
> +
> +#define TCR_EL1_RSVD           (1 << 31)
> +#define TCR_EL2_RSVD           (1 << 31 | 1 << 23)
> +#define TCR_EL3_RSVD           (1 << 31 | 1 << 23)
> +
> +#define TCR_FLAGS              (TCR_TG0_4K | \
> +               TCR_SHARED_OUTER | \
> +               TCR_SHARED_INNER | \
> +               TCR_IRGN_WBWA | \
> +               TCR_ORGN_WBWA | \
> +               TCR_T0SZ(BITS_PER_VA))
> +
> +#define MEMORY_ATTR     (PMD_SECT_AF | PMD_SECT_INNER_SHARE |    \
> +               PMD_ATTRINDX(MT_NORMAL) |       \
> +               PMD_TYPE_SECT)
> +
> +#ifndef __ASSEMBLY__
> +
> +static inline void set_ttbr_tcr_mair(int el, uint64_t table, uint64_t tcr, uint64_t attr)
> +{
> +       asm volatile("dsb sy");
> +       if (el == 1) {
> +               asm volatile("msr ttbr0_el1, %0" : : "r" (table) : "memory");
> +               asm volatile("msr tcr_el1, %0" : : "r" (tcr) : "memory");
> +               asm volatile("msr mair_el1, %0" : : "r" (attr) : "memory");
> +       } else if (el == 2) {
> +               asm volatile("msr ttbr0_el2, %0" : : "r" (table) : "memory");
> +               asm volatile("msr tcr_el2, %0" : : "r" (tcr) : "memory");
> +               asm volatile("msr mair_el2, %0" : : "r" (attr) : "memory");
> +       } else if (el == 3) {
> +               asm volatile("msr ttbr0_el3, %0" : : "r" (table) : "memory");
> +               asm volatile("msr tcr_el3, %0" : : "r" (tcr) : "memory");
> +               asm volatile("msr mair_el3, %0" : : "r" (attr) : "memory");
> +       } else {
> +               hang();
> +       }
> +       asm volatile("isb");
> +}
> +
> +static inline uint64_t get_ttbr(int el)
> +{
> +       uint64_t val;
> +       if (el == 1) {
> +               asm volatile("mrs %0, ttbr0_el1" : "=r" (val));
> +       } else if (el == 2) {
> +               asm volatile("mrs %0, ttbr0_el2" : "=r" (val));
> +       } else if (el == 3) {
> +               asm volatile("mrs %0, ttbr0_el3" : "=r" (val));
> +       } else {
> +               hang();
> +       }
> +
> +       return val;
> +}
> +#endif
> +
>  #ifdef CONFIG_MMU
>  void __mmu_cache_on(void);
>  void __mmu_cache_off(void);
> @@ -11,4 +164,6 @@ static inline void __mmu_cache_off(void) {}
>  static inline void __mmu_cache_flush(void) {}
>  #endif
>
> +void mmu_early_enable(uint64_t membase, uint64_t memsize, uint64_t _ttb);
> +
>  #endif /* __ARM_MMU_H */
> diff --git a/arch/arm/cpu/start.c b/arch/arm/cpu/start.c
> index e037d91..1d017bc 100644
> --- a/arch/arm/cpu/start.c
> +++ b/arch/arm/cpu/start.c
> @@ -31,7 +31,7 @@
>  #include <malloc.h>
>
>  #include <debug_ll.h>
> -#include "mmu-early.h"
> +#include "mmu.h"
>
>  unsigned long arm_stack_top;
>  static unsigned long arm_head_bottom;
> @@ -151,7 +151,7 @@ __noreturn void barebox_non_pbl_start(unsigned long membase,
>                 relocate_to_adr(barebox_base);
>         }
>
> -       setup_c();
> +//     setup_c();
>
>         barrier();
>
> @@ -170,7 +170,7 @@ __noreturn void barebox_non_pbl_start(unsigned long membase,
>                 } else {
>                         pr_debug("enabling MMU, ttb @ 0x%08lx\n", ttb);
>                         arm_early_mmu_cache_invalidate();
> -                       mmu_early_enable(membase, memsize, ttb);
> +                       mmu_early_enable((uint64_t)membase, (uint64_t)memsize, (uint64_t)ttb);
>                 }
>         }
>
> @@ -193,7 +193,7 @@ __noreturn void barebox_non_pbl_start(unsigned long membase,
>                 if (totalsize) {
>                         unsigned long mem = arm_mem_boarddata(membase, endmem,
>                                                               totalsize);
> -                       pr_debug("found %s in boarddata, copying to 0x%08lx\n",
> +                       pr_debug("found %s in boarddata, copying to 0x%lu\n",
>                                  name, mem);
>                         barebox_boarddata = memcpy((void *)mem, boarddata,
>                                                    totalsize);
> @@ -229,7 +229,7 @@ __noreturn void barebox_non_pbl_start(unsigned long membase,
>
>  #ifndef CONFIG_PBL_IMAGE
>
> -void __naked __section(.text_entry) start(void)
> +void __section(.text_entry) start(void)
>  {
>         barebox_arm_head();
>  }
> @@ -239,7 +239,7 @@ void __naked __section(.text_entry) start(void)
>   * First function in the uncompressed image. We get here from
>   * the pbl. The stack already has been set up by the pbl.
>   */
> -void __naked __section(.text_entry) start(unsigned long membase,
> +void __section(.text_entry) start(unsigned long membase,
>                 unsigned long memsize, void *boarddata)
>  {
>         barebox_non_pbl_start(membase, memsize, boarddata);
> diff --git a/arch/arm/cpu/uncompress.c b/arch/arm/cpu/uncompress.c
> index b8e2e9f..5bcce6b 100644
> --- a/arch/arm/cpu/uncompress.c
> +++ b/arch/arm/cpu/uncompress.c
> @@ -60,7 +60,7 @@ void __noreturn barebox_multi_pbl_start(unsigned long membase,
>                  * to the current address. Otherwise it may be a readonly location.
>                  * Copy and relocate to the start of the memory in this case.
>                  */
> -               if (pc > membase && pc - membase < memsize)
> +               if (pc > membase && pc < membase + memsize)
>                         relocate_to_current_adr();
>                 else
>                         relocate_to_adr(membase);
> diff --git a/arch/arm/include/asm/barebox-arm.h b/arch/arm/include/asm/barebox-arm.h
> index 8e7b45c..6713326 100644
> --- a/arch/arm/include/asm/barebox-arm.h
> +++ b/arch/arm/include/asm/barebox-arm.h
> @@ -97,7 +97,7 @@ void *barebox_arm_boot_dtb(void);
>  static inline unsigned long arm_mem_stack(unsigned long membase,
>                                           unsigned long endmem)
>  {
> -       return endmem - SZ_64K - STACK_SIZE;
> +       return endmem - STACK_SIZE;
>  }
>
>  static inline unsigned long arm_mem_ttb(unsigned long membase,
> diff --git a/arch/arm/include/asm/bitops.h b/arch/arm/include/asm/bitops.h
> index 138ebe2..ac85a0a 100644
> --- a/arch/arm/include/asm/bitops.h
> +++ b/arch/arm/include/asm/bitops.h
> @@ -1,184 +1,48 @@
>  /*
> - * Copyright 1995, Russell King.
> - * Various bits and pieces copyrights include:
> - *  Linus Torvalds (test_bit).
> + * Copyright (C) 2012 ARM Ltd.
>   *
> - * bit 0 is the LSB of addr; bit 32 is the LSB of (addr+1).
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License version 2 as
> + * published by the Free Software Foundation.
>   *
> - * Please note that the code in this file should never be included
> - * from user space.  Many of these are not implemented in assembler
> - * since they would be too costly.  Also, they require priviledged
> - * instructions (which are not available from user mode) to ensure
> - * that they are atomic.
> + * This program is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
> + * GNU General Public License for more details.
> + *
> + * You should have received a copy of the GNU General Public License
> + * along with this program.  If not, see <http://www.gnu.org/licenses/>.
>   */
> +#ifndef __ASM_BITOPS_H
> +#define __ASM_BITOPS_H
>
> -#ifndef __ASM_ARM_BITOPS_H
> -#define __ASM_ARM_BITOPS_H
> +#include <linux/compiler.h>
>
>  #ifndef _LINUX_BITOPS_H
>  #error only <linux/bitops.h> can be included directly
>  #endif
>
>  /*
> - * Functions equivalent of ops.h
> - */
> -static inline void __set_bit(int nr, volatile void *addr)
> -{
> -       ((unsigned char *) addr)[nr >> 3] |= (1U << (nr & 7));
> -}
> -
> -static inline void __clear_bit(int nr, volatile void *addr)
> -{
> -       ((unsigned char *) addr)[nr >> 3] &= ~(1U << (nr & 7));
> -}
> -
> -static inline void __change_bit(int nr, volatile void *addr)
> -{
> -       ((unsigned char *) addr)[nr >> 3] ^= (1U << (nr & 7));
> -}
> -
> -static inline int __test_and_set_bit(int nr, volatile void *addr)
> -{
> -       unsigned int mask = 1 << (nr & 7);
> -       unsigned int oldval;
> -
> -       oldval = ((unsigned char *) addr)[nr >> 3];
> -       ((unsigned char *) addr)[nr >> 3] = oldval | mask;
> -       return oldval & mask;
> -}
> -
> -static inline int __test_and_clear_bit(int nr, volatile void *addr)
> -{
> -       unsigned int mask = 1 << (nr & 7);
> -       unsigned int oldval;
> -
> -       oldval = ((unsigned char *) addr)[nr >> 3];
> -       ((unsigned char *) addr)[nr >> 3] = oldval & ~mask;
> -       return oldval & mask;
> -}
> -
> -static inline int __test_and_change_bit(int nr, volatile void *addr)
> -{
> -       unsigned int mask = 1 << (nr & 7);
> -       unsigned int oldval;
> -
> -       oldval = ((unsigned char *) addr)[nr >> 3];
> -       ((unsigned char *) addr)[nr >> 3] = oldval ^ mask;
> -       return oldval & mask;
> -}
> -
> -/*
> - * This routine doesn't need to be atomic.
> - */
> -static inline int test_bit(int nr, const void * addr)
> -{
> -    return ((unsigned char *) addr)[nr >> 3] & (1U << (nr & 7));
> -}
> -
> -#define set_bit(x, y)                  __set_bit(x, y)
> -#define clear_bit(x, y)                        __clear_bit(x, y)
> -#define change_bit(x, y)               __change_bit(x, y)
> -#define test_and_set_bit(x, y)         __test_and_set_bit(x, y)
> -#define test_and_clear_bit(x, y)       __test_and_clear_bit(x, y)
> -#define test_and_change_bit(x, y)      __test_and_change_bit(x, y)
> -
> -#ifndef __ARMEB__
> -/*
> - * These are the little endian definitions.
> + * Little endian assembly atomic bitops.
>   */
> -extern int _find_first_zero_bit_le(const void *p, unsigned size);
> -extern int _find_next_zero_bit_le(const void *p, int size, int offset);
> -extern int _find_first_bit_le(const unsigned long *p, unsigned size);
> -extern int _find_next_bit_le(const unsigned long *p, int size, int offset);
> -#define find_first_zero_bit(p, sz)     _find_first_zero_bit_le(p, sz)
> -#define find_next_zero_bit(p, sz, off) _find_next_zero_bit_le(p, sz, off)
> -#define find_first_bit(p, sz)          _find_first_bit_le(p, sz)
> -#define find_next_bit(p, sz, off)      _find_next_bit_le(p, sz, off)
> +extern void set_bit(int nr, volatile unsigned long *p);
> +extern void clear_bit(int nr, volatile unsigned long *p);
> +extern void change_bit(int nr, volatile unsigned long *p);
> +extern int test_and_set_bit(int nr, volatile unsigned long *p);
> +extern int test_and_clear_bit(int nr, volatile unsigned long *p);
> +extern int test_and_change_bit(int nr, volatile unsigned long *p);
>
> -#define WORD_BITOFF_TO_LE(x)           ((x))
> -
> -#else          /* ! __ARMEB__ */
> -
> -/*
> - * These are the big endian definitions.
> - */
> -extern int _find_first_zero_bit_be(const void *p, unsigned size);
> -extern int _find_next_zero_bit_be(const void *p, int size, int offset);
> -extern int _find_first_bit_be(const unsigned long *p, unsigned size);
> -extern int _find_next_bit_be(const unsigned long *p, int size, int offset);
> -#define find_first_zero_bit(p, sz)     _find_first_zero_bit_be(p, sz)
> -#define find_next_zero_bit(p, sz, off) _find_next_zero_bit_be(p, sz, off)
> -#define find_first_bit(p, sz)          _find_first_bit_be(p, sz)
> -#define find_next_bit(p, sz, off)      _find_next_bit_be(p, sz, off)
> -
> -#define WORD_BITOFF_TO_LE(x)           ((x) ^ 0x18)
> -
> -#endif         /* __ARMEB__ */
> -
> -#if defined(__LINUX_ARM_ARCH__) && (__LINUX_ARM_ARCH__ >= 5)
> -static inline int constant_fls(int x)
> -{
> -       int r = 32;
> -
> -       if (!x)
> -               return 0;
> -       if (!(x & 0xffff0000u)) {
> -               x <<= 16;
> -               r -= 16;
> -       }
> -       if (!(x & 0xff000000u)) {
> -               x <<= 8;
> -               r -= 8;
> -       }
> -       if (!(x & 0xf0000000u)) {
> -               x <<= 4;
> -               r -= 4;
> -       }
> -       if (!(x & 0xc0000000u)) {
> -               x <<= 2;
> -               r -= 2;
> -       }
> -       if (!(x & 0x80000000u)) {
> -               x <<= 1;
> -               r -= 1;
> -       }
> -       return r;
> -}
> -
> -/*
> - * On ARMv5 and above those functions can be implemented around
> - * the clz instruction for much better code efficiency.
> - */
> -#define fls(x) \
> -       (__builtin_constant_p(x) ? constant_fls(x) : \
> -       ({ int __r; asm("clz\t%0, %1" : "=r"(__r) : "r"(x) : "cc"); 32-__r; }))
> -#define ffs(x) ({ unsigned long __t = (x); fls(__t &-__t); })
> -#define __ffs(x) (ffs(x) - 1)
> -#define ffz(x) __ffs(~(x))
> -#else          /* ! __ARM__USE_GENERIC_FF */
> -/*
> - * ffz = Find First Zero in word. Undefined if no zero exists,
> - * so code should check against ~0UL first..
> - */
> -static inline unsigned long ffz(unsigned long word)
> -{
> -       int k;
> -
> -       word = ~word;
> -       k = 31;
> -       if (word & 0x0000ffff) { k -= 16; word <<= 16; }
> -       if (word & 0x00ff0000) { k -= 8;  word <<= 8;  }
> -       if (word & 0x0f000000) { k -= 4;  word <<= 4;  }
> -       if (word & 0x30000000) { k -= 2;  word <<= 2;  }
> -       if (word & 0x40000000) { k -= 1; }
> -       return k;
> -}
>  #include <asm-generic/bitops/__ffs.h>
>  #include <asm-generic/bitops/ffs.h>
> +#include <asm-generic/bitops/__fls.h>
>  #include <asm-generic/bitops/fls.h>
> -#endif         /* __ARM__USE_GENERIC_FF */
> +
> +#include <asm-generic/bitops/ffz.h>
>  #include <asm-generic/bitops/fls64.h>
> +#include <asm-generic/bitops/find.h>
>
>  #include <asm-generic/bitops/hweight.h>
>
> -#endif /* _ARM_BITOPS_H */
> +#include <asm-generic/bitops/ops.h>
> +
> +#endif /* __ASM_BITOPS_H */
> diff --git a/arch/arm/include/asm/boarddata.h b/arch/arm/include/asm/boarddata.h
> new file mode 100644
> index 0000000..8c3c5f0
> --- /dev/null
> +++ b/arch/arm/include/asm/boarddata.h
> @@ -0,0 +1,5 @@
> +#ifndef __ASM_BOARDDATA_H
> +#define __ASM_BOARDDATA_H
> +
> +
> +#endif /* __ASM_BOARDDATA_H */
> diff --git a/arch/arm/include/asm/cache-l2x0.h b/arch/arm/include/asm/cache-l2x0.h
> index 9bb245b..963dd99 100644
> --- a/arch/arm/include/asm/cache-l2x0.h
> +++ b/arch/arm/include/asm/cache-l2x0.h
> @@ -56,14 +56,6 @@
>  #define L2X0_LINE_TAG                  0xF30
>  #define L2X0_DEBUG_CTRL                        0xF40
>  #define L2X0_PREFETCH_CTRL             0xF60
> -#define   L2X0_DOUBLE_LINEFILL_EN                      (1 << 30)
> -#define   L2X0_INSTRUCTION_PREFETCH_EN                 (1 << 29)
> -#define   L2X0_DATA_PREFETCH_EN                                (1 << 28)
> -#define   L2X0_DOUBLE_LINEFILL_ON_WRAP_READ_DIS                (1 << 27)
> -#define   L2X0_PREFETCH_DROP_EN                                (1 << 24)
> -#define   L2X0_INCR_DOUBLE_LINEFILL_EN                 (1 << 23)
> -#define   L2X0_ESCLUSIVE_SEQUENCE_EN                   (1 << 21)
> -
>  #define L2X0_POWER_CTRL                        0xF80
>  #define   L2X0_DYNAMIC_CLK_GATING_EN   (1 << 1)
>  #define   L2X0_STNDBY_MODE_EN          (1 << 0)
> diff --git a/arch/arm/include/asm/cache.h b/arch/arm/include/asm/cache.h
> index 2f6eab0..5a524f3 100644
> --- a/arch/arm/include/asm/cache.h
> +++ b/arch/arm/include/asm/cache.h
> @@ -1,9 +1,11 @@
>  #ifndef __ASM_CACHE_H
>  #define __ASM_CACHE_H
>
> +extern void v8_invalidate_icache_all(void);
> +
>  static inline void flush_icache(void)
>  {
> -       asm volatile("mcr p15, 0, %0, c7, c5, 0" : : "r" (0));
> +       v8_invalidate_icache_all();
>  }
>
>  int arm_set_cache_functions(void);
> diff --git a/arch/arm/include/asm/errata.h b/arch/arm/include/asm/errata.h
> index 98137b5..9525823 100644
> --- a/arch/arm/include/asm/errata.h
> +++ b/arch/arm/include/asm/errata.h
> @@ -77,12 +77,3 @@ static inline void enable_arm_errata_794072_war(void)
>                 "mcr    p15, 0, r0, c15, c0, 1\n"
>         );
>  }
> -
> -static inline void enable_arm_errata_845369_war(void)
> -{
> -       __asm__ __volatile__ (
> -               "mrc    p15, 0, r0, c15, c0, 1\n"
> -               "orr    r0, r0, #1 << 22\n"
> -               "mcr    p15, 0, r0, c15, c0, 1\n"
> -       );
> -}
> diff --git a/arch/arm/include/asm/gic.h b/arch/arm/include/asm/gic.h
> new file mode 100644
> index 0000000..c7c17e3
> --- /dev/null
> +++ b/arch/arm/include/asm/gic.h
> @@ -0,0 +1,128 @@
> +#ifndef __GIC_H__
> +#define __GIC_H__
> +
> +/* Generic Interrupt Controller Definitions */
> +//#ifdef CONFIG_GICV3
> +//#define GICD_BASE                       (0x2f000000)
> +//#define GICR_BASE                       (0x2f100000)
> +//#else
> +
> +//#if defined(CONFIG_TARGET_VEXPRESS64_BASE_FVP) || \
> +        defined(CONFIG_TARGET_VEXPRESS64_BASE_FVP_DRAM)
> +//#define GICD_BASE                       (0x2f000000)
> +//#define GICC_BASE                       (0x2c000000)
> +//#elif CONFIG_TARGET_VEXPRESS64_JUNO
> +#define GIC_DIST_BASE                       (0x2C010000)
> +#define GIC_CPU_BASE                       (0x2C02f000)
> +//#else
> +//#error "Unknown board variant"
> +//#endif
> +//#endif /* !CONFIG_GICV3 */
> +
> +/* Register offsets for the ARM generic interrupt controller (GIC) */
> +
> +#define GIC_DIST_OFFSET                0x1000
> +#define GIC_CPU_OFFSET_A9      0x0100
> +#define GIC_CPU_OFFSET_A15     0x2000
> +
> +/* Distributor Registers */
> +#define GICD_CTLR              0x0000
> +#define GICD_TYPER             0x0004
> +#define GICD_IIDR              0x0008
> +#define GICD_STATUSR           0x0010
> +#define GICD_SETSPI_NSR                0x0040
> +#define GICD_CLRSPI_NSR                0x0048
> +#define GICD_SETSPI_SR         0x0050
> +#define GICD_CLRSPI_SR         0x0058
> +#define GICD_SEIR              0x0068
> +#define GICD_IGROUPRn          0x0080
> +#define GICD_ISENABLERn                0x0100
> +#define GICD_ICENABLERn                0x0180
> +#define GICD_ISPENDRn          0x0200
> +#define GICD_ICPENDRn          0x0280
> +#define GICD_ISACTIVERn                0x0300
> +#define GICD_ICACTIVERn                0x0380
> +#define GICD_IPRIORITYRn       0x0400
> +#define GICD_ITARGETSRn                0x0800
> +#define GICD_ICFGR             0x0c00
> +#define GICD_IGROUPMODRn       0x0d00
> +#define GICD_NSACRn            0x0e00
> +#define GICD_SGIR              0x0f00
> +#define GICD_CPENDSGIRn                0x0f10
> +#define GICD_SPENDSGIRn                0x0f20
> +#define GICD_IROUTERn          0x6000
> +
> +/* Cpu Interface Memory Mapped Registers */
> +#define GICC_CTLR              0x0000
> +#define GICC_PMR               0x0004
> +#define GICC_BPR               0x0008
> +#define GICC_IAR               0x000C
> +#define GICC_EOIR              0x0010
> +#define GICC_RPR               0x0014
> +#define GICC_HPPIR             0x0018
> +#define GICC_ABPR              0x001c
> +#define GICC_AIAR              0x0020
> +#define GICC_AEOIR             0x0024
> +#define GICC_AHPPIR            0x0028
> +#define GICC_APRn              0x00d0
> +#define GICC_NSAPRn            0x00e0
> +#define GICC_IIDR              0x00fc
> +#define GICC_DIR               0x1000
> +
> +/* ReDistributor Registers for Control and Physical LPIs */
> +#define GICR_CTLR              0x0000
> +#define GICR_IIDR              0x0004
> +#define GICR_TYPER             0x0008
> +#define GICR_STATUSR           0x0010
> +#define GICR_WAKER             0x0014
> +#define GICR_SETLPIR           0x0040
> +#define GICR_CLRLPIR           0x0048
> +#define GICR_SEIR              0x0068
> +#define GICR_PROPBASER         0x0070
> +#define GICR_PENDBASER         0x0078
> +#define GICR_INVLPIR           0x00a0
> +#define GICR_INVALLR           0x00b0
> +#define GICR_SYNCR             0x00c0
> +#define GICR_MOVLPIR           0x0100
> +#define GICR_MOVALLR           0x0110
> +
> +/* ReDistributor Registers for SGIs and PPIs */
> +#define GICR_IGROUPRn          0x0080
> +#define GICR_ISENABLERn                0x0100
> +#define GICR_ICENABLERn                0x0180
> +#define GICR_ISPENDRn          0x0200
> +#define GICR_ICPENDRn          0x0280
> +#define GICR_ISACTIVERn                0x0300
> +#define GICR_ICACTIVERn                0x0380
> +#define GICR_IPRIORITYRn       0x0400
> +#define GICR_ICFGR0            0x0c00
> +#define GICR_ICFGR1            0x0c04
> +#define GICR_IGROUPMODRn       0x0d00
> +#define GICR_NSACRn            0x0e00
> +
> +/* Cpu Interface System Registers */
> +#define ICC_IAR0_EL1           S3_0_C12_C8_0
> +#define ICC_IAR1_EL1           S3_0_C12_C12_0
> +#define ICC_EOIR0_EL1          S3_0_C12_C8_1
> +#define ICC_EOIR1_EL1          S3_0_C12_C12_1
> +#define ICC_HPPIR0_EL1         S3_0_C12_C8_2
> +#define ICC_HPPIR1_EL1         S3_0_C12_C12_2
> +#define ICC_BPR0_EL1           S3_0_C12_C8_3
> +#define ICC_BPR1_EL1           S3_0_C12_C12_3
> +#define ICC_DIR_EL1            S3_0_C12_C11_1
> +#define ICC_PMR_EL1            S3_0_C4_C6_0
> +#define ICC_RPR_EL1            S3_0_C12_C11_3
> +#define ICC_CTLR_EL1           S3_0_C12_C12_4
> +#define ICC_CTLR_EL3           S3_6_C12_C12_4
> +#define ICC_SRE_EL1            S3_0_C12_C12_5
> +#define ICC_SRE_EL2            S3_4_C12_C9_5
> +#define ICC_SRE_EL3            S3_6_C12_C12_5
> +#define ICC_IGRPEN0_EL1                S3_0_C12_C12_6
> +#define ICC_IGRPEN1_EL1                S3_0_C12_C12_7
> +#define ICC_IGRPEN1_EL3                S3_6_C12_C12_7
> +#define ICC_SEIEN_EL1          S3_0_C12_C13_0
> +#define ICC_SGI0R_EL1          S3_0_C12_C11_7
> +#define ICC_SGI1R_EL1          S3_0_C12_C11_5
> +#define ICC_ASGI1R_EL1         S3_0_C12_C11_6
> +
> +#endif /* __GIC_H__ */
> diff --git a/arch/arm/include/asm/mmu.h b/arch/arm/include/asm/mmu.h
> index 8de6544..8a1d80a 100644
> --- a/arch/arm/include/asm/mmu.h
> +++ b/arch/arm/include/asm/mmu.h
> @@ -13,9 +13,7 @@
>
>  struct arm_memory;
>
> -static inline void mmu_enable(void)
> -{
> -}
> +void mmu_enable(void);
>  void mmu_disable(void);
>  static inline void arm_create_section(unsigned long virt, unsigned long phys, int size_m,
>                 unsigned int flags)
> @@ -30,7 +28,7 @@ static inline void setup_dma_coherent(unsigned long offset)
>  #define ARCH_HAS_REMAP
>  #define MAP_ARCH_DEFAULT MAP_CACHED
>  int arch_remap_range(void *_start, size_t size, unsigned flags);
> -void *map_io_sections(unsigned long physaddr, void *start, size_t size);
> +void *map_io_sections(uint64_t phys, void *_start, size_t size);
>  #else
>  #define MAP_ARCH_DEFAULT MAP_UNCACHED
>  static inline void *map_io_sections(unsigned long phys, void *start, size_t size)
> diff --git a/arch/arm/include/asm/pgtable.h b/arch/arm/include/asm/pgtable.h
> index fd1521d..e4a3c53 100644
> --- a/arch/arm/include/asm/pgtable.h
> +++ b/arch/arm/include/asm/pgtable.h
> @@ -18,8 +18,9 @@
>   */
>  #define PMD_TYPE_MASK          (3 << 0)
>  #define PMD_TYPE_FAULT         (0 << 0)
> -#define PMD_TYPE_TABLE         (1 << 0)
> -#define PMD_TYPE_SECT          (2 << 0)
> +#define PMD_TYPE_TABLE         (3 << 0)
> +#define PMD_TYPE_PAGE          (3 << 0)
> +#define PMD_TYPE_SECT          (1 << 0)
>  #define PMD_BIT4               (1 << 4)
>  #define PMD_DOMAIN(x)          ((x) << 5)
>  #define PMD_PROTECTION         (1 << 9)        /* v5 */
> diff --git a/arch/arm/include/asm/ptrace.h b/arch/arm/include/asm/ptrace.h
> index 022d365..450b63a 100644
> --- a/arch/arm/include/asm/ptrace.h
> +++ b/arch/arm/include/asm/ptrace.h
> @@ -20,124 +20,15 @@
>  /* options set using PTRACE_SETOPTIONS */
>  #define PTRACE_O_TRACESYSGOOD  0x00000001
>
> -/*
> - * PSR bits
> - */
> -#define USR26_MODE     0x00000000
> -#define FIQ26_MODE     0x00000001
> -#define IRQ26_MODE     0x00000002
> -#define SVC26_MODE     0x00000003
> -#define USR_MODE       0x00000010
> -#define FIQ_MODE       0x00000011
> -#define IRQ_MODE       0x00000012
> -#define SVC_MODE       0x00000013
> -#define ABT_MODE       0x00000017
> -#define UND_MODE       0x0000001b
> -#define SYSTEM_MODE    0x0000001f
> -#define MODE32_BIT     0x00000010
> -#define MODE_MASK      0x0000001f
> -#define PSR_T_BIT      0x00000020
> -#define PSR_F_BIT      0x00000040
> -#define PSR_I_BIT      0x00000080
> -#define PSR_A_BIT      0x00000100
> -#define PSR_E_BIT      0x00000200
> -#define PSR_J_BIT      0x01000000
> -#define PSR_Q_BIT      0x08000000
> -#define PSR_V_BIT      0x10000000
> -#define PSR_C_BIT      0x20000000
> -#define PSR_Z_BIT      0x40000000
> -#define PSR_N_BIT      0x80000000
> -#define PCMASK         0
> -
>  #ifndef __ASSEMBLY__
>
>  /* this struct defines the way the registers are stored on the
>     stack during a system call. */
>
>  struct pt_regs {
> -       long uregs[18];
> +       long uregs[31];
>  };
>
> -#define ARM_cpsr       uregs[16]
> -#define ARM_pc         uregs[15]
> -#define ARM_lr         uregs[14]
> -#define ARM_sp         uregs[13]
> -#define ARM_ip         uregs[12]
> -#define ARM_fp         uregs[11]
> -#define ARM_r10                uregs[10]
> -#define ARM_r9         uregs[9]
> -#define ARM_r8         uregs[8]
> -#define ARM_r7         uregs[7]
> -#define ARM_r6         uregs[6]
> -#define ARM_r5         uregs[5]
> -#define ARM_r4         uregs[4]
> -#define ARM_r3         uregs[3]
> -#define ARM_r2         uregs[2]
> -#define ARM_r1         uregs[1]
> -#define ARM_r0         uregs[0]
> -#define ARM_ORIG_r0    uregs[17]
> -
> -#ifdef __KERNEL__
> -
> -#define user_mode(regs)        \
> -       (((regs)->ARM_cpsr & 0xf) == 0)
> -
> -#ifdef CONFIG_ARM_THUMB
> -#define thumb_mode(regs) \
> -       (((regs)->ARM_cpsr & PSR_T_BIT))
> -#else
> -#define thumb_mode(regs) (0)
> -#endif
> -
> -#define processor_mode(regs) \
> -       ((regs)->ARM_cpsr & MODE_MASK)
> -
> -#define interrupts_enabled(regs) \
> -       (!((regs)->ARM_cpsr & PSR_I_BIT))
> -
> -#define fast_interrupts_enabled(regs) \
> -       (!((regs)->ARM_cpsr & PSR_F_BIT))
> -
> -#define condition_codes(regs) \
> -       ((regs)->ARM_cpsr & (PSR_V_BIT | PSR_C_BIT | PSR_Z_BIT | PSR_N_BIT))
> -
> -/* Are the current registers suitable for user mode?
> - * (used to maintain security in signal handlers)
> - */
> -static inline int valid_user_regs(struct pt_regs *regs)
> -{
> -       if ((regs->ARM_cpsr & 0xf) == 0 &&
> -           (regs->ARM_cpsr & (PSR_F_BIT | PSR_I_BIT)) == 0)
> -               return 1;
> -
> -       /*
> -        * Force CPSR to something logical...
> -        */
> -       regs->ARM_cpsr &= (PSR_V_BIT | PSR_C_BIT | PSR_Z_BIT | PSR_N_BIT |
> -                               0x10);
> -
> -       return 0;
> -}
> -
> -#endif /* __KERNEL__ */
> -
>  #endif /* __ASSEMBLY__ */
>
> -#ifndef __ASSEMBLY__
> -#define pc_pointer(v) \
> -       ((v) & ~PCMASK)
> -
> -#define instruction_pointer(regs) \
> -       (pc_pointer((regs)->ARM_pc))
> -
> -#ifdef __KERNEL__
> -extern void show_regs(struct pt_regs *);
> -
> -#define predicate(x)   (x & 0xf0000000)
> -#define PREDICATE_ALWAYS       0xe0000000
> -
> -#endif
> -
> -#endif /* __ASSEMBLY__ */
> -
>  #endif
> diff --git a/arch/arm/include/asm/system.h b/arch/arm/include/asm/system.h
> index b118a42..04a79c4 100644
> --- a/arch/arm/include/asm/system.h
> +++ b/arch/arm/include/asm/system.h
> @@ -1,96 +1,125 @@
>  #ifndef __ASM_ARM_SYSTEM_H
>  #define __ASM_ARM_SYSTEM_H
>
> -#if __LINUX_ARM_ARCH__ >= 7
>  #define isb() __asm__ __volatile__ ("isb" : : : "memory")
> -#define dsb() __asm__ __volatile__ ("dsb" : : : "memory")
> +#define dsb() __asm__ __volatile__ ("dsb sy" : : : "memory")
>  #define dmb() __asm__ __volatile__ ("dmb" : : : "memory")
> -#elif defined(CONFIG_CPU_XSC3) || __LINUX_ARM_ARCH__ == 6
> -#define isb() __asm__ __volatile__ ("mcr p15, 0, %0, c7, c5, 4" \
> -                                    : : "r" (0) : "memory")
> -#define dsb() __asm__ __volatile__ ("mcr p15, 0, %0, c7, c10, 4" \
> -                                    : : "r" (0) : "memory")
> -#define dmb() __asm__ __volatile__ ("mcr p15, 0, %0, c7, c10, 5" \
> -                                    : : "r" (0) : "memory")
> -#elif defined(CONFIG_CPU_FA526)
> -#define isb() __asm__ __volatile__ ("mcr p15, 0, %0, c7, c5, 4" \
> -                                    : : "r" (0) : "memory")
> -#define dsb() __asm__ __volatile__ ("mcr p15, 0, %0, c7, c10, 4" \
> -                                    : : "r" (0) : "memory")
> -#define dmb() __asm__ __volatile__ ("" : : : "memory")
> -#else
> -#define isb() __asm__ __volatile__ ("" : : : "memory")
> -#define dsb() __asm__ __volatile__ ("mcr p15, 0, %0, c7, c10, 4" \
> -                                    : : "r" (0) : "memory")
> -#define dmb() __asm__ __volatile__ ("" : : : "memory")
> -#endif
>
>  /*
> - * CR1 bits (CP#15 CR1)
> + * SCTLR_EL1/SCTLR_EL2/SCTLR_EL3 bits definitions
>   */
> -#define CR_M    (1 << 0)       /* MMU enable                           */
> -#define CR_A    (1 << 1)       /* Alignment abort enable               */
> -#define CR_C    (1 << 2)       /* Dcache enable                        */
> -#define CR_W    (1 << 3)       /* Write buffer enable                  */
> -#define CR_P    (1 << 4)       /* 32-bit exception handler             */
> -#define CR_D    (1 << 5)       /* 32-bit data address range            */
> -#define CR_L    (1 << 6)       /* Implementation defined               */
> -#define CR_B    (1 << 7)       /* Big endian                           */
> -#define CR_S    (1 << 8)       /* System MMU protection                */
> -#define CR_R    (1 << 9)       /* ROM MMU protection                   */
> -#define CR_F    (1 << 10)      /* Implementation defined               */
> -#define CR_Z    (1 << 11)      /* Implementation defined               */
> -#define CR_I    (1 << 12)      /* Icache enable                        */
> -#define CR_V    (1 << 13)      /* Vectors relocated to 0xffff0000      */
> -#define CR_RR   (1 << 14)      /* Round Robin cache replacement        */
> -#define CR_L4   (1 << 15)      /* LDR pc can set T bit                 */
> -#define CR_DT   (1 << 16)
> -#define CR_IT   (1 << 18)
> -#define CR_ST   (1 << 19)
> -#define CR_FI   (1 << 21)      /* Fast interrupt (lower latency mode)  */
> -#define CR_U    (1 << 22)      /* Unaligned access operation           */
> -#define CR_XP   (1 << 23)      /* Extended page tables                 */
> -#define CR_VE   (1 << 24)      /* Vectored interrupts                  */
> -#define CR_EE   (1 << 25)      /* Exception (Big) Endian               */
> -#define CR_TRE  (1 << 28)      /* TEX remap enable                     */
> -#define CR_AFE  (1 << 29)      /* Access flag enable                   */
> -#define CR_TE   (1 << 30)      /* Thumb exception enable               */
> +#define CR_M           (1 << 0)        /* MMU enable                   */
> +#define CR_A           (1 << 1)        /* Alignment abort enable       */
> +#define CR_C           (1 << 2)        /* Dcache enable                */
> +#define CR_SA          (1 << 3)        /* Stack Alignment Check Enable */
> +#define CR_I           (1 << 12)       /* Icache enable                */
> +#define CR_WXN         (1 << 19)       /* Write Permision Imply XN     */
> +#define CR_EE          (1 << 25)       /* Exception (Big) Endian       */
> +
> +#ifndef CONFIG_SYS_FULL_VA
> +#define PGTABLE_SIZE   (0x10000)
> +#else
> +#define PGTABLE_SIZE   CONFIG_SYS_PGTABLE_SIZE
> +#endif
> +
> +/* 2MB granularity */
> +#define MMU_SECTION_SHIFT      21
> +#define MMU_SECTION_SIZE       (1 << MMU_SECTION_SHIFT)
>
>  #ifndef __ASSEMBLY__
> -static inline unsigned int get_cr(void)
> +
> +enum dcache_option {
> +       DCACHE_OFF = 0x3,
> +};
> +
> +#define wfi()                          \
> +       ({asm volatile(                 \
> +       "wfi" : : : "memory");          \
> +       })
> +
> +static inline unsigned int current_el(void)
>  {
> -       unsigned int val;
> -       asm volatile ("mrc p15, 0, %0, c1, c0, 0  @ get CR" : "=r" (val) : : "cc");
> -       return val;
> +       unsigned int el;
> +       asm volatile("mrs %0, CurrentEL" : "=r" (el) : : "cc");
> +       return el >> 2;
>  }
>
> -static inline void set_cr(unsigned int val)
> +static inline unsigned int get_sctlr(void)
>  {
> -       asm volatile("mcr p15, 0, %0, c1, c0, 0 @ set CR"
> -         : : "r" (val) : "cc");
> -       isb();
> +       unsigned int el, val;
> +
> +       el = current_el();
> +       if (el == 1)
> +               asm volatile("mrs %0, sctlr_el1" : "=r" (val) : : "cc");
> +       else if (el == 2)
> +               asm volatile("mrs %0, sctlr_el2" : "=r" (val) : : "cc");
> +       else
> +               asm volatile("mrs %0, sctlr_el3" : "=r" (val) : : "cc");
> +
> +       return val;
>  }
>
> -#ifdef CONFIG_CPU_32v7
> -static inline unsigned int get_vbar(void)
> +static inline void set_sctlr(unsigned int val)
>  {
> -       unsigned int vbar;
> -       asm volatile("mrc p15, 0, %0, c12, c0, 0 @ get VBAR"
> -                    : "=r" (vbar) : : "cc");
> -       return vbar;
> +       unsigned int el;
> +
> +       el = current_el();
> +       if (el == 1)
> +               asm volatile("msr sctlr_el1, %0" : : "r" (val) : "cc");
> +       else if (el == 2)
> +               asm volatile("msr sctlr_el2, %0" : : "r" (val) : "cc");
> +       else
> +               asm volatile("msr sctlr_el3, %0" : : "r" (val) : "cc");
> +
> +       asm volatile("isb");
>  }
>
> -static inline void set_vbar(unsigned int vbar)
> +static inline unsigned long read_mpidr(void)
>  {
> -       asm volatile("mcr p15, 0, %0, c12, c0, 0 @ set VBAR"
> -                    : : "r" (vbar) : "cc");
> -       isb();
> +       unsigned long val;
> +
> +       asm volatile("mrs %0, mpidr_el1" : "=r" (val));
> +
> +       return val;
>  }
> -#else
> -static inline unsigned int get_vbar(void) { return 0; }
> -static inline void set_vbar(unsigned int vbar) {}
> -#endif
>
> -#endif
> +#define BSP_COREID     0
> +
> +void __asm_flush_dcache_all(void);
> +void __asm_invalidate_dcache_all(void);
> +void __asm_flush_dcache_range(u64 start, u64 end);
> +void __asm_invalidate_tlb_all(void);
> +void __asm_invalidate_icache_all(void);
> +int __asm_flush_l3_cache(void);
> +
> +void armv8_switch_to_el2(void);
> +void armv8_switch_to_el1(void);
> +void gic_init(void);
> +void gic_send_sgi(unsigned long sgino);
> +void wait_for_wakeup(void);
> +void protect_secure_region(void);
> +void smp_kick_all_cpus(void);
> +
> +void flush_l3_cache(void);
> +
> +/*
> + *Issue a hypervisor call in accordance with ARM "SMC Calling convention",
> + * DEN0028A
> + *
> + * @args: input and output arguments
> + *
> + */
> +void hvc_call(struct pt_regs *args);
> +
> +/*
> + *Issue a secure monitor call in accordance with ARM "SMC Calling convention",
> + * DEN0028A
> + *
> + * @args: input and output arguments
> + *
> + */
> +void smc_call(struct pt_regs *args);
> +
> +#endif /* __ASSEMBLY__ */
>
>  #endif /* __ASM_ARM_SYSTEM_H */
> diff --git a/arch/arm/include/asm/system_info.h b/arch/arm/include/asm/system_info.h
> index 0761848..f595aae 100644
> --- a/arch/arm/include/asm/system_info.h
> +++ b/arch/arm/include/asm/system_info.h
> @@ -13,6 +13,7 @@
>  #define CPU_ARCH_ARMv5TEJ      7
>  #define CPU_ARCH_ARMv6         8
>  #define CPU_ARCH_ARMv7         9
> +#define CPU_ARCH_ARMv8         10
>
>  #define CPU_IS_ARM720          0x41007200
>  #define CPU_IS_ARM720_MASK     0xff00fff0
> @@ -41,6 +42,12 @@
>  #define CPU_IS_CORTEX_A15      0x410fc0f0
>  #define CPU_IS_CORTEX_A15_MASK 0xff0ffff0
>
> +#define CPU_IS_CORTEX_A53      0x410fd034
> +#define CPU_IS_CORTEX_A53_MASK 0xff0ffff0
> +
> +#define CPU_IS_CORTEX_A57      0x411fd070
> +#define CPU_IS_CORTEX_A57_MASK 0xff0ffff0
> +
>  #define CPU_IS_PXA250          0x69052100
>  #define CPU_IS_PXA250_MASK     0xfffff7f0
>
> @@ -112,6 +119,20 @@
>  #define cpu_is_cortex_a15() (0)
>  #endif
>
> +
> +#ifdef CONFIG_CPU_64v8
> +#ifdef ARM_ARCH
> +#define ARM_MULTIARCH
> +#else
> +#define ARM_ARCH CPU_ARCH_ARMv8
> +#endif
> +#define cpu_is_cortex_a53() cpu_is_arm(CORTEX_A53)
> +#define cpu_is_cortex_a57() cpu_is_arm(CORTEX_A57)
> +#else
> +#define cpu_is_cortex_a53() (0)
> +#define cpu_is_cortex_a57() (0)
> +#endif
> +
>  #ifndef __ASSEMBLY__
>
>  #ifdef ARM_MULTIARCH
> @@ -124,31 +145,33 @@ static inline int arm_early_get_cpu_architecture(void)
>  {
>         int cpu_arch;
>
> -       if ((read_cpuid_id() & 0x0008f000) == 0) {
> -               cpu_arch = CPU_ARCH_UNKNOWN;
> -       } else if ((read_cpuid_id() & 0x0008f000) == 0x00007000) {
> -               cpu_arch = (read_cpuid_id() & (1 << 23)) ? CPU_ARCH_ARMv4T : CPU_ARCH_ARMv3;
> -       } else if ((read_cpuid_id() & 0x00080000) == 0x00000000) {
> -               cpu_arch = (read_cpuid_id() >> 16) & 7;
> -               if (cpu_arch)
> -                       cpu_arch += CPU_ARCH_ARMv3;
> -       } else if ((read_cpuid_id() & 0x000f0000) == 0x000f0000) {
> -               unsigned int mmfr0;
> -
> -               /* Revised CPUID format. Read the Memory Model Feature
> -                * Register 0 and check for VMSAv7 or PMSAv7 */
> -               asm("mrc        p15, 0, %0, c0, c1, 4"
> -                   : "=r" (mmfr0));
> -               if ((mmfr0 & 0x0000000f) >= 0x00000003 ||
> -                   (mmfr0 & 0x000000f0) >= 0x00000030)
> -                       cpu_arch = CPU_ARCH_ARMv7;
> -               else if ((mmfr0 & 0x0000000f) == 0x00000002 ||
> -                        (mmfr0 & 0x000000f0) == 0x00000020)
> -                       cpu_arch = CPU_ARCH_ARMv6;
> -               else
> -                       cpu_arch = CPU_ARCH_UNKNOWN;
> -       } else
> -               cpu_arch = CPU_ARCH_UNKNOWN;
> +//     if ((read_cpuid_id() & 0x0008f000) == 0) {
> +//             cpu_arch = CPU_ARCH_UNKNOWN;
> +//     } else if ((read_cpuid_id() & 0x0008f000) == 0x00007000) {
> +//             cpu_arch = (read_cpuid_id() & (1 << 23)) ? CPU_ARCH_ARMv4T : CPU_ARCH_ARMv3;
> +//     } else if ((read_cpuid_id() & 0x00080000) == 0x00000000) {
> +//             cpu_arch = (read_cpuid_id() >> 16) & 7;
> +//             if (cpu_arch)
> +//                     cpu_arch += CPU_ARCH_ARMv3;
> +//     } else if ((read_cpuid_id() & 0x000f0000) == 0x000f0000) {
> +//             unsigned int mmfr0;
> +//
> +//             /* Revised CPUID format. Read the Memory Model Feature
> +//              * Register 0 and check for VMSAv7 or PMSAv7 */
> +//             asm("mrc        p15, 0, %0, c0, c1, 4"
> +//                 : "=r" (mmfr0));
> +//             if ((mmfr0 & 0x0000000f) >= 0x00000003 ||
> +//                 (mmfr0 & 0x000000f0) >= 0x00000030)
> +//                     cpu_arch = CPU_ARCH_ARMv7;
> +//             else if ((mmfr0 & 0x0000000f) == 0x00000002 ||
> +//                      (mmfr0 & 0x000000f0) == 0x00000020)
> +//                     cpu_arch = CPU_ARCH_ARMv6;
> +//             else
> +//                     cpu_arch = CPU_ARCH_UNKNOWN;
> +//     } else
> +//             cpu_arch = CPU_ARCH_UNKNOWN;
> +
> +       cpu_arch = CPU_ARCH_ARMv8;
>
>         return cpu_arch;
>  }
> diff --git a/arch/arm/lib/Makefile b/arch/arm/lib/Makefile
> index e1c6f5b..5b9d4a5 100644
> --- a/arch/arm/lib/Makefile
> +++ b/arch/arm/lib/Makefile
> @@ -3,24 +3,11 @@ obj-$(CONFIG_BOOTM)   += bootm.o
>  obj-$(CONFIG_CMD_BOOTZ)        += bootz.o
>  obj-$(CONFIG_CMD_BOOTU)        += bootu.o
>  obj-y  += div0.o
> -obj-y  += findbit.o
> -obj-y  += io.o
> -obj-y  += io-readsb.o
> -obj-y  += io-readsw-armv4.o
> -obj-y  += io-readsl.o
> -obj-y  += io-writesb.o
> -obj-y  += io-writesw-armv4.o
> -obj-y  += io-writesl.o
> -obj-y  += lib1funcs.o
> -obj-y  += ashrdi3.o
> -obj-y  += ashldi3.o
> -obj-y  += lshrdi3.o
>  obj-y  += runtime-offset.o
>  pbl-y  += runtime-offset.o
>  obj-$(CONFIG_ARM_OPTIMZED_STRING_FUNCTIONS)    += memcpy.o
>  obj-$(CONFIG_ARM_OPTIMZED_STRING_FUNCTIONS)    += memset.o
>  obj-$(CONFIG_ARM_UNWIND) += unwind.o
> -obj-$(CONFIG_ARM_SEMIHOSTING) += semihosting-trap.o semihosting.o
>  obj-$(CONFIG_MODULES) += module.o
>  extra-y += barebox.lds
>
> diff --git a/arch/arm/lib/armlinux.c b/arch/arm/lib/armlinux.c
> index 47b9bd3..21a2292 100644
> --- a/arch/arm/lib/armlinux.c
> +++ b/arch/arm/lib/armlinux.c
> @@ -270,12 +270,6 @@ void start_linux(void *adr, int swap, unsigned long initrd_address,
>         architecture = armlinux_get_architecture();
>
>         shutdown_barebox();
> -       if (swap) {
> -               u32 reg;
> -               __asm__ __volatile__("mrc p15, 0, %0, c1, c0" : "=r" (reg));
> -               reg ^= CR_B; /* swap big-endian flag */
> -               __asm__ __volatile__("mcr p15, 0, %0, c1, c0" :: "r" (reg));
> -       }
>
>         kernel(0, architecture, params);
>  }
> diff --git a/arch/arm/lib/barebox.lds.S b/arch/arm/lib/barebox.lds.S
> index 6dc8bd2..240699f 100644
> --- a/arch/arm/lib/barebox.lds.S
> +++ b/arch/arm/lib/barebox.lds.S
> @@ -20,8 +20,8 @@
>
>  #include <asm-generic/barebox.lds.h>
>
> -OUTPUT_FORMAT("elf32-littlearm", "elf32-littlearm", "elf32-littlearm")
> -OUTPUT_ARCH(arm)
> +OUTPUT_FORMAT("elf64-littleaarch64", "elf64-littleaarch64", "elf64-littleaarch64")
> +OUTPUT_ARCH(aarch64)
>  ENTRY(start)
>  SECTIONS
>  {
> @@ -43,7 +43,6 @@ SECTIONS
>                 __bare_init_start = .;
>                 *(.text_bare_init*)
>                 __bare_init_end = .;
> -               . = ALIGN(4);
>                 __exceptions_start = .;
>                 KEEP(*(.text_exceptions*))
>                 __exceptions_stop = .;
> diff --git a/arch/arm/lib/bootm.c b/arch/arm/lib/bootm.c
> index f6024c8..1913d5f 100644
> --- a/arch/arm/lib/bootm.c
> +++ b/arch/arm/lib/bootm.c
> @@ -67,55 +67,6 @@ static int sdram_start_and_size(unsigned long *start, unsigned long *size)
>         return 0;
>  }
>
> -static int get_kernel_addresses(size_t image_size,
> -                                int verbose, unsigned long *load_address,
> -                                unsigned long *mem_free)
> -{
> -       unsigned long mem_start, mem_size;
> -       int ret;
> -       size_t image_decomp_size;
> -       unsigned long spacing;
> -
> -       ret = sdram_start_and_size(&mem_start, &mem_size);
> -       if (ret)
> -               return ret;
> -
> -       /*
> -        * We don't know the exact decompressed size so just use a conservative
> -        * default of 4 times the size of the compressed image.
> -        */
> -       image_decomp_size = PAGE_ALIGN(image_size * 4);
> -
> -       /*
> -        * By default put oftree/initrd close behind compressed kernel image to
> -        * avoid placing it outside of the kernels lowmem region.
> -        */
> -       spacing = SZ_1M;
> -
> -       if (*load_address == UIMAGE_INVALID_ADDRESS) {
> -               /*
> -                * Place the kernel at an address where it does not need to
> -                * relocate itself before decompression.
> -                */
> -               *load_address = mem_start + image_decomp_size;
> -               if (verbose)
> -                       printf("no OS load address, defaulting to 0x%08lx\n",
> -                               *load_address);
> -       } else if (*load_address <= mem_start + image_decomp_size) {
> -               /*
> -                * If the user/image specified an address where the kernel needs
> -                * to relocate itself before decompression we need to extend the
> -                * spacing to allow this relocation to happen without
> -                * overwriting anything placed behind the kernel.
> -                */
> -               spacing += image_decomp_size;
> -       }
> -
> -       *mem_free = PAGE_ALIGN(*load_address + image_size + spacing);
> -
> -       return 0;
> -}
> -
>  static int __do_bootm_linux(struct image_data *data, unsigned long free_mem, int swap)
>  {
>         unsigned long kernel;
> @@ -173,20 +124,38 @@ static int __do_bootm_linux(struct image_data *data, unsigned long free_mem, int
>
>  static int do_bootm_linux(struct image_data *data)
>  {
> -       unsigned long load_address, mem_free;
> +       unsigned long load_address, mem_start, mem_size, mem_free;
>         int ret;
>
> -       load_address = data->os_address;
> -
> -       ret = get_kernel_addresses(bootm_get_os_size(data),
> -                            bootm_verbose(data), &load_address, &mem_free);
> +       ret = sdram_start_and_size(&mem_start, &mem_size);
>         if (ret)
>                 return ret;
>
> +       load_address = data->os_address;
> +
> +       if (load_address == UIMAGE_INVALID_ADDRESS) {
> +               /*
> +                * Just use a conservative default of 4 times the size of the
> +                * compressed image, to avoid the need for the kernel to
> +                * relocate itself before decompression.
> +                */
> +               load_address = mem_start + PAGE_ALIGN(
> +                              bootm_get_os_size(data) * 4);
> +               if (bootm_verbose(data))
> +                       printf("no OS load address, defaulting to 0x%08lx\n",
> +                               load_address);
> +       }
> +
>         ret = bootm_load_os(data, load_address);
>         if (ret)
>                 return ret;
>
> +       /*
> +        * put oftree/initrd close behind compressed kernel image to avoid
> +        * placing it outside of the kernels lowmem.
> +        */
> +       mem_free = PAGE_ALIGN(data->os_res->end + SZ_1M);
> +
>         return __do_bootm_linux(data, mem_free, 0);
>  }
>
> @@ -282,7 +251,11 @@ static int do_bootz_linux(struct image_data *data)
>         u32 end, start;
>         size_t image_size;
>         unsigned long load_address = data->os_address;
> -       unsigned long mem_free;
> +       unsigned long mem_start, mem_size, mem_free;
> +
> +       ret = sdram_start_and_size(&mem_start, &mem_size);
> +       if (ret)
> +               return ret;
>
>         fd = open(data->os_file, O_RDONLY);
>         if (fd < 0) {
> @@ -318,12 +291,20 @@ static int do_bootz_linux(struct image_data *data)
>         }
>
>         image_size = end - start;
> -       load_address = data->os_address;
>
> -       ret = get_kernel_addresses(image_size, bootm_verbose(data),
> -                            &load_address, &mem_free);
> -       if (ret)
> -               return ret;
> +       if (load_address == UIMAGE_INVALID_ADDRESS) {
> +               /*
> +                * Just use a conservative default of 4 times the size of the
> +                * compressed image, to avoid the need for the kernel to
> +                * relocate itself before decompression.
> +                */
> +               data->os_address = mem_start + PAGE_ALIGN(image_size * 4);
> +
> +               load_address = data->os_address;
> +               if (bootm_verbose(data))
> +                       printf("no OS load address, defaulting to 0x%08lx\n",
> +                               load_address);
> +       }
>
>         data->os_res = request_sdram_region("zimage", load_address, image_size);
>         if (!data->os_res) {
> @@ -359,6 +340,12 @@ static int do_bootz_linux(struct image_data *data)
>
>         close(fd);
>
> +       /*
> +        * put oftree/initrd close behind compressed kernel image to avoid
> +        * placing it outside of the kernels lowmem.
> +        */
> +       mem_free = PAGE_ALIGN(data->os_res->end + SZ_1M);
> +
>         return __do_bootm_linux(data, mem_free, swap);
>
>  err_out:
> @@ -575,7 +562,7 @@ static int armlinux_register_image_handler(void)
>                 register_image_handler(&aimage_handler);
>                 binfmt_register(&binfmt_aimage_hook);
>         }
> -       if (IS_BUILTIN(CONFIG_FITIMAGE))
> +       if (IS_BUILTIN(CONFIG_CMD_BOOTM_FITIMAGE))
>                 register_image_handler(&arm_fit_handler);
>         binfmt_register(&binfmt_arm_zimage_hook);
>         binfmt_register(&binfmt_barebox_hook);
> diff --git a/arch/arm/lib/copy_template.S b/arch/arm/lib/copy_template.S
> index d8eb063..cc9a842 100644
> --- a/arch/arm/lib/copy_template.S
> +++ b/arch/arm/lib/copy_template.S
> @@ -1,268 +1,192 @@
>  /*
> - *  linux/arch/arm/lib/copy_template.s
> + * Copyright (C) 2013 ARM Ltd.
> + * Copyright (C) 2013 Linaro.
>   *
> - *  Code template for optimized memory copy functions
> + * This code is based on glibc cortex strings work originally authored by Linaro
> + * and re-licensed under GPLv2 for the Linux kernel. The original code can
> + * be found @
>   *
> - *  Author:    Nicolas Pitre
> - *  Created:   Sep 28, 2005
> - *  Copyright: MontaVista Software, Inc.
> + * http://bazaar.launchpad.net/~linaro-toolchain-dev/cortex-strings/trunk/
> + * files/head:/src/aarch64/
>   *
> - *  This program is free software; you can redistribute it and/or modify
> - *  it under the terms of the GNU General Public License version 2 as
> - *  published by the Free Software Foundation.
> - */
> -
> -/*
> - * Theory of operation
> - * -------------------
> - *
> - * This file provides the core code for a forward memory copy used in
> - * the implementation of memcopy(), copy_to_user() and copy_from_user().
> - *
> - * The including file must define the following accessor macros
> - * according to the need of the given function:
> - *
> - * ldr1w ptr reg abort
> - *
> - *     This loads one word from 'ptr', stores it in 'reg' and increments
> - *     'ptr' to the next word. The 'abort' argument is used for fixup tables.
> - *
> - * ldr4w ptr reg1 reg2 reg3 reg4 abort
> - * ldr8w ptr, reg1 reg2 reg3 reg4 reg5 reg6 reg7 reg8 abort
> - *
> - *     This loads four or eight words starting from 'ptr', stores them
> - *     in provided registers and increments 'ptr' past those words.
> - *     The'abort' argument is used for fixup tables.
> - *
> - * ldr1b ptr reg cond abort
> - *
> - *     Similar to ldr1w, but it loads a byte and increments 'ptr' one byte.
> - *     It also must apply the condition code if provided, otherwise the
> - *     "al" condition is assumed by default.
> - *
> - * str1w ptr reg abort
> - * str8w ptr reg1 reg2 reg3 reg4 reg5 reg6 reg7 reg8 abort
> - * str1b ptr reg cond abort
> - *
> - *     Same as their ldr* counterparts, but data is stored to 'ptr' location
> - *     rather than being loaded.
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License version 2 as
> + * published by the Free Software Foundation.
>   *
> - * enter reg1 reg2
> + * This program is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
> + * GNU General Public License for more details.
>   *
> - *     Preserve the provided registers on the stack plus any additional
> - *     data as needed by the implementation including this code. Called
> - *     upon code entry.
> - *
> - * exit reg1 reg2
> - *
> - *     Restore registers with the values previously saved with the
> - *     'preserv' macro. Called upon code termination.
> - *
> - * LDR1W_SHIFT
> - * STR1W_SHIFT
> - *
> - *     Correction to be applied to the "ip" register when branching into
> - *     the ldr1w or str1w instructions (some of these macros may expand to
> - *     than one 32bit instruction in Thumb-2)
> + * You should have received a copy of the GNU General Public License
> + * along with this program.  If not, see <http://www.gnu.org/licenses/>.
>   */
>
>
> -               enter   r4, lr
> -
> -               subs    r2, r2, #4
> -               blt     8f
> -               ands    ip, r0, #3
> -       PLD(    pld     [r1, #0]                )
> -               bne     9f
> -               ands    ip, r1, #3
> -               bne     10f
> -
> -1:             subs    r2, r2, #(28)
> -               stmfd   sp!, {r5 - r8}
> -               blt     5f
> -
> -       CALGN(  ands    ip, r0, #31             )
> -       CALGN(  rsb     r3, ip, #32             )
> -       CALGN(  sbcnes  r4, r3, r2              )  @ C is always set here
> -       CALGN(  bcs     2f                      )
> -       CALGN(  adr     r4, 6f                  )
> -       CALGN(  subs    r2, r2, r3              )  @ C gets set
> -       CALGN(  add     pc, r4, ip              )
> -
> -       PLD(    pld     [r1, #0]                )
> -2:     PLD(    subs    r2, r2, #96             )
> -       PLD(    pld     [r1, #28]               )
> -       PLD(    blt     4f                      )
> -       PLD(    pld     [r1, #60]               )
> -       PLD(    pld     [r1, #92]               )
> -
> -3:     PLD(    pld     [r1, #124]              )
> -4:             ldr8w   r1, r3, r4, r5, r6, r7, r8, ip, lr, abort=20f
> -               subs    r2, r2, #32
> -               str8w   r0, r3, r4, r5, r6, r7, r8, ip, lr, abort=20f
> -               bge     3b
> -       PLD(    cmn     r2, #96                 )
> -       PLD(    bge     4b                      )
> -
> -5:             ands    ip, r2, #28
> -               rsb     ip, ip, #32
> -#if LDR1W_SHIFT > 0
> -               lsl     ip, ip, #LDR1W_SHIFT
> -#endif
> -               addne   pc, pc, ip              @ C is always clear here
> -               b       7f
> -6:
> -               .rept   (1 << LDR1W_SHIFT)
> -               W(nop)
> -               .endr
> -               ldr1w   r1, r3, abort=20f
> -               ldr1w   r1, r4, abort=20f
> -               ldr1w   r1, r5, abort=20f
> -               ldr1w   r1, r6, abort=20f
> -               ldr1w   r1, r7, abort=20f
> -               ldr1w   r1, r8, abort=20f
> -               ldr1w   r1, lr, abort=20f
> -
> -#if LDR1W_SHIFT < STR1W_SHIFT
> -               lsl     ip, ip, #STR1W_SHIFT - LDR1W_SHIFT
> -#elif LDR1W_SHIFT > STR1W_SHIFT
> -               lsr     ip, ip, #LDR1W_SHIFT - STR1W_SHIFT
> -#endif
> -               add     pc, pc, ip
> -               nop
> -               .rept   (1 << STR1W_SHIFT)
> -               W(nop)
> -               .endr
> -               str1w   r0, r3, abort=20f
> -               str1w   r0, r4, abort=20f
> -               str1w   r0, r5, abort=20f
> -               str1w   r0, r6, abort=20f
> -               str1w   r0, r7, abort=20f
> -               str1w   r0, r8, abort=20f
> -               str1w   r0, lr, abort=20f
> -
> -       CALGN(  bcs     2b                      )
> -
> -7:             ldmfd   sp!, {r5 - r8}
> -
> -8:             movs    r2, r2, lsl #31
> -               ldr1b   r1, r3, ne, abort=21f
> -               ldr1b   r1, r4, cs, abort=21f
> -               ldr1b   r1, ip, cs, abort=21f
> -               str1b   r0, r3, ne, abort=21f
> -               str1b   r0, r4, cs, abort=21f
> -               str1b   r0, ip, cs, abort=21f
> -
> -               exit    r4, pc
> -
> -9:             rsb     ip, ip, #4
> -               cmp     ip, #2
> -               ldr1b   r1, r3, gt, abort=21f
> -               ldr1b   r1, r4, ge, abort=21f
> -               ldr1b   r1, lr, abort=21f
> -               str1b   r0, r3, gt, abort=21f
> -               str1b   r0, r4, ge, abort=21f
> -               subs    r2, r2, ip
> -               str1b   r0, lr, abort=21f
> -               blt     8b
> -               ands    ip, r1, #3
> -               beq     1b
> -
> -10:            bic     r1, r1, #3
> -               cmp     ip, #2
> -               ldr1w   r1, lr, abort=21f
> -               beq     17f
> -               bgt     18f
> -
> -
> -               .macro  forward_copy_shift pull push
> -
> -               subs    r2, r2, #28
> -               blt     14f
> -
> -       CALGN(  ands    ip, r0, #31             )
> -       CALGN(  rsb     ip, ip, #32             )
> -       CALGN(  sbcnes  r4, ip, r2              )  @ C is always set here
> -       CALGN(  subcc   r2, r2, ip              )
> -       CALGN(  bcc     15f                     )
> -
> -11:            stmfd   sp!, {r5 - r9}
> -
> -       PLD(    pld     [r1, #0]                )
> -       PLD(    subs    r2, r2, #96             )
> -       PLD(    pld     [r1, #28]               )
> -       PLD(    blt     13f                     )
> -       PLD(    pld     [r1, #60]               )
> -       PLD(    pld     [r1, #92]               )
> -
> -12:    PLD(    pld     [r1, #124]              )
> -13:            ldr4w   r1, r4, r5, r6, r7, abort=19f
> -               mov     r3, lr, pull #\pull
> -               subs    r2, r2, #32
> -               ldr4w   r1, r8, r9, ip, lr, abort=19f
> -               orr     r3, r3, r4, push #\push
> -               mov     r4, r4, pull #\pull
> -               orr     r4, r4, r5, push #\push
> -               mov     r5, r5, pull #\pull
> -               orr     r5, r5, r6, push #\push
> -               mov     r6, r6, pull #\pull
> -               orr     r6, r6, r7, push #\push
> -               mov     r7, r7, pull #\pull
> -               orr     r7, r7, r8, push #\push
> -               mov     r8, r8, pull #\pull
> -               orr     r8, r8, r9, push #\push
> -               mov     r9, r9, pull #\pull
> -               orr     r9, r9, ip, push #\push
> -               mov     ip, ip, pull #\pull
> -               orr     ip, ip, lr, push #\push
> -               str8w   r0, r3, r4, r5, r6, r7, r8, r9, ip, , abort=19f
> -               bge     12b
> -       PLD(    cmn     r2, #96                 )
> -       PLD(    bge     13b                     )
> -
> -               ldmfd   sp!, {r5 - r9}
> -
> -14:            ands    ip, r2, #28
> -               beq     16f
> -
> -15:            mov     r3, lr, pull #\pull
> -               ldr1w   r1, lr, abort=21f
> -               subs    ip, ip, #4
> -               orr     r3, r3, lr, push #\push
> -               str1w   r0, r3, abort=21f
> -               bgt     15b
> -       CALGN(  cmp     r2, #0                  )
> -       CALGN(  bge     11b                     )
> -
> -16:            sub     r1, r1, #(\push / 8)
> -               b       8b
> -
> -               .endm
> -
> -
> -               forward_copy_shift      pull=8  push=24
> -
> -17:            forward_copy_shift      pull=16 push=16
> -
> -18:            forward_copy_shift      pull=24 push=8
> -
> -
>  /*
> - * Abort preamble and completion macros.
> - * If a fixup handler is required then those macros must surround it.
> - * It is assumed that the fixup code will handle the private part of
> - * the exit macro.
> + * Copy a buffer from src to dest (alignment handled by the hardware)
> + *
> + * Parameters:
> + *     x0 - dest
> + *     x1 - src
> + *     x2 - n
> + * Returns:
> + *     x0 - dest
>   */
> -
> -       .macro  copy_abort_preamble
> -19:    ldmfd   sp!, {r5 - r9}
> -       b       21f
> -20:    ldmfd   sp!, {r5 - r8}
> -21:
> -       .endm
> -
> -       .macro  copy_abort_end
> -       ldmfd   sp!, {r4, pc}
> -       .endm
> -
> -
> +dstin  .req    x0
> +src    .req    x1
> +count  .req    x2
> +tmp1   .req    x3
> +tmp1w  .req    w3
> +tmp2   .req    x4
> +tmp2w  .req    w4
> +dst    .req    x6
> +
> +A_l    .req    x7
> +A_h    .req    x8
> +B_l    .req    x9
> +B_h    .req    x10
> +C_l    .req    x11
> +C_h    .req    x12
> +D_l    .req    x13
> +D_h    .req    x14
> +
> +       mov     dst, dstin
> +       cmp     count, #16
> +       /*When memory length is less than 16, the accessed are not aligned.*/
> +       b.lo    .Ltiny15
> +
> +       neg     tmp2, src
> +       ands    tmp2, tmp2, #15/* Bytes to reach alignment. */
> +       b.eq    .LSrcAligned
> +       sub     count, count, tmp2
> +       /*
> +       * Copy the leading memory data from src to dst in an increasing
> +       * address order.By this way,the risk of overwritting the source
> +       * memory data is eliminated when the distance between src and
> +       * dst is less than 16. The memory accesses here are alignment.
> +       */
> +       tbz     tmp2, #0, 1f
> +       ldrb1   tmp1w, src, #1
> +       strb1   tmp1w, dst, #1
> +1:
> +       tbz     tmp2, #1, 2f
> +       ldrh1   tmp1w, src, #2
> +       strh1   tmp1w, dst, #2
> +2:
> +       tbz     tmp2, #2, 3f
> +       ldr1    tmp1w, src, #4
> +       str1    tmp1w, dst, #4
> +3:
> +       tbz     tmp2, #3, .LSrcAligned
> +       ldr1    tmp1, src, #8
> +       str1    tmp1, dst, #8
> +
> +.LSrcAligned:
> +       cmp     count, #64
> +       b.ge    .Lcpy_over64
> +       /*
> +       * Deal with small copies quickly by dropping straight into the
> +       * exit block.
> +       */
> +.Ltail63:
> +       /*
> +       * Copy up to 48 bytes of data. At this point we only need the
> +       * bottom 6 bits of count to be accurate.
> +       */
> +       ands    tmp1, count, #0x30
> +       b.eq    .Ltiny15
> +       cmp     tmp1w, #0x20
> +       b.eq    1f
> +       b.lt    2f
> +       ldp1    A_l, A_h, src, #16
> +       stp1    A_l, A_h, dst, #16
> +1:
> +       ldp1    A_l, A_h, src, #16
> +       stp1    A_l, A_h, dst, #16
> +2:
> +       ldp1    A_l, A_h, src, #16
> +       stp1    A_l, A_h, dst, #16
> +.Ltiny15:
> +       /*
> +       * Prefer to break one ldp/stp into several load/store to access
> +       * memory in an increasing address order,rather than to load/store 16
> +       * bytes from (src-16) to (dst-16) and to backward the src to aligned
> +       * address,which way is used in original cortex memcpy. If keeping
> +       * the original memcpy process here, memmove need to satisfy the
> +       * precondition that src address is at least 16 bytes bigger than dst
> +       * address,otherwise some source data will be overwritten when memove
> +       * call memcpy directly. To make memmove simpler and decouple the
> +       * memcpy's dependency on memmove, withdrew the original process.
> +       */
> +       tbz     count, #3, 1f
> +       ldr1    tmp1, src, #8
> +       str1    tmp1, dst, #8
> +1:
> +       tbz     count, #2, 2f
> +       ldr1    tmp1w, src, #4
> +       str1    tmp1w, dst, #4
> +2:
> +       tbz     count, #1, 3f
> +       ldrh1   tmp1w, src, #2
> +       strh1   tmp1w, dst, #2
> +3:
> +       tbz     count, #0, .Lexitfunc
> +       ldrb1   tmp1w, src, #1
> +       strb1   tmp1w, dst, #1
> +
> +       b       .Lexitfunc
> +
> +.Lcpy_over64:
> +       subs    count, count, #128
> +       b.ge    .Lcpy_body_large
> +       /*
> +       * Less than 128 bytes to copy, so handle 64 here and then jump
> +       * to the tail.
> +       */
> +       ldp1    A_l, A_h, src, #16
> +       stp1    A_l, A_h, dst, #16
> +       ldp1    B_l, B_h, src, #16
> +       ldp1    C_l, C_h, src, #16
> +       stp1    B_l, B_h, dst, #16
> +       stp1    C_l, C_h, dst, #16
> +       ldp1    D_l, D_h, src, #16
> +       stp1    D_l, D_h, dst, #16
> +
> +       tst     count, #0x3f
> +       b.ne    .Ltail63
> +       b       .Lexitfunc
> +
> +       /*
> +       * Critical loop.  Start at a new cache line boundary.  Assuming
> +       * 64 bytes per line this ensures the entire loop is in one line.
> +       */
> +.Lcpy_body_large:
> +       /* pre-get 64 bytes data. */
> +       ldp1    A_l, A_h, src, #16
> +       ldp1    B_l, B_h, src, #16
> +       ldp1    C_l, C_h, src, #16
> +       ldp1    D_l, D_h, src, #16
> +1:
> +       /*
> +       * interlace the load of next 64 bytes data block with store of the last
> +       * loaded 64 bytes data.
> +       */
> +       stp1    A_l, A_h, dst, #16
> +       ldp1    A_l, A_h, src, #16
> +       stp1    B_l, B_h, dst, #16
> +       ldp1    B_l, B_h, src, #16
> +       stp1    C_l, C_h, dst, #16
> +       ldp1    C_l, C_h, src, #16
> +       stp1    D_l, D_h, dst, #16
> +       ldp1    D_l, D_h, src, #16
> +       subs    count, count, #64
> +       b.ge    1b
> +       stp1    A_l, A_h, dst, #16
> +       stp1    B_l, B_h, dst, #16
> +       stp1    C_l, C_h, dst, #16
> +       stp1    D_l, D_h, dst, #16
> +
> +       tst     count, #0x3f
> +       b.ne    .Ltail63
> +.Lexitfunc:
> diff --git a/arch/arm/lib/memcpy.S b/arch/arm/lib/memcpy.S
> index 5123691..cfed319 100644
> --- a/arch/arm/lib/memcpy.S
> +++ b/arch/arm/lib/memcpy.S
> @@ -1,64 +1,74 @@
>  /*
> - *  linux/arch/arm/lib/memcpy.S
> + * Copyright (C) 2013 ARM Ltd.
> + * Copyright (C) 2013 Linaro.
>   *
> - *  Author:    Nicolas Pitre
> - *  Created:   Sep 28, 2005
> - *  Copyright: MontaVista Software, Inc.
> + * This code is based on glibc cortex strings work originally authored by Linaro
> + * and re-licensed under GPLv2 for the Linux kernel. The original code can
> + * be found @
>   *
> - *  This program is free software; you can redistribute it and/or modify
> - *  it under the terms of the GNU General Public License version 2 as
> - *  published by the Free Software Foundation.
> + * http://bazaar.launchpad.net/~linaro-toolchain-dev/cortex-strings/trunk/
> + * files/head:/src/aarch64/
> + *
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License version 2 as
> + * published by the Free Software Foundation.
> + *
> + * This program is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
> + * GNU General Public License for more details.
> + *
> + * You should have received a copy of the GNU General Public License
> + * along with this program.  If not, see <http://www.gnu.org/licenses/>.
>   */
>
>  #include <linux/linkage.h>
>  #include <asm/assembler.h>
>
> -#define LDR1W_SHIFT    0
> -#define STR1W_SHIFT    0
> -
> -       .macro ldr1w ptr reg abort
> -       W(ldr) \reg, [\ptr], #4
> -       .endm
> -
> -       .macro ldr4w ptr reg1 reg2 reg3 reg4 abort
> -       ldmia \ptr!, {\reg1, \reg2, \reg3, \reg4}
> +/*
> + * Copy a buffer from src to dest (alignment handled by the hardware)
> + *
> + * Parameters:
> + *     x0 - dest
> + *     x1 - src
> + *     x2 - n
> + * Returns:
> + *     x0 - dest
> + */
> +       .macro ldrb1 ptr, regB, val
> +       ldrb  \ptr, [\regB], \val
>         .endm
>
> -       .macro ldr8w ptr reg1 reg2 reg3 reg4 reg5 reg6 reg7 reg8 abort
> -       ldmia \ptr!, {\reg1, \reg2, \reg3, \reg4, \reg5, \reg6, \reg7, \reg8}
> +       .macro strb1 ptr, regB, val
> +       strb \ptr, [\regB], \val
>         .endm
>
> -       .macro ldr1b ptr reg cond=al abort
> -       ldr\cond\()b \reg, [\ptr], #1
> +       .macro ldrh1 ptr, regB, val
> +       ldrh  \ptr, [\regB], \val
>         .endm
>
> -       .macro str1w ptr reg abort
> -       W(str) \reg, [\ptr], #4
> +       .macro strh1 ptr, regB, val
> +       strh \ptr, [\regB], \val
>         .endm
>
> -       .macro str8w ptr reg1 reg2 reg3 reg4 reg5 reg6 reg7 reg8 abort
> -       stmia \ptr!, {\reg1, \reg2, \reg3, \reg4, \reg5, \reg6, \reg7, \reg8}
> +       .macro ldr1 ptr, regB, val
> +       ldr \ptr, [\regB], \val
>         .endm
>
> -       .macro str1b ptr reg cond=al abort
> -       str\cond\()b \reg, [\ptr], #1
> +       .macro str1 ptr, regB, val
> +       str \ptr, [\regB], \val
>         .endm
>
> -       .macro enter reg1 reg2
> -       stmdb sp!, {r0, \reg1, \reg2}
> +       .macro ldp1 ptr, regB, regC, val
> +       ldp \ptr, \regB, [\regC], \val
>         .endm
>
> -       .macro exit reg1 reg2
> -       ldmfd sp!, {r0, \reg1, \reg2}
> +       .macro stp1 ptr, regB, regC, val
> +       stp \ptr, \regB, [\regC], \val
>         .endm
>
> -       .text
> -
> -/* Prototype: void *memcpy(void *dest, const void *src, size_t n); */
> -
> +       .weak memcpy
>  ENTRY(memcpy)
> -
>  #include "copy_template.S"
> -
> +       ret
>  ENDPROC(memcpy)
> -
> diff --git a/arch/arm/lib/memset.S b/arch/arm/lib/memset.S
> index c4d2672..380a540 100644
> --- a/arch/arm/lib/memset.S
> +++ b/arch/arm/lib/memset.S
> @@ -1,124 +1,215 @@
>  /*
> - *  linux/arch/arm/lib/memset.S
> + * Copyright (C) 2013 ARM Ltd.
> + * Copyright (C) 2013 Linaro.
>   *
> - *  Copyright (C) 1995-2000 Russell King
> + * This code is based on glibc cortex strings work originally authored by Linaro
> + * and re-licensed under GPLv2 for the Linux kernel. The original code can
> + * be found @
> + *
> + * http://bazaar.launchpad.net/~linaro-toolchain-dev/cortex-strings/trunk/
> + * files/head:/src/aarch64/
>   *
>   * This program is free software; you can redistribute it and/or modify
>   * it under the terms of the GNU General Public License version 2 as
>   * published by the Free Software Foundation.
>   *
> - *  ASM optimised string functions
> + * This program is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
> + * GNU General Public License for more details.
> + *
> + * You should have received a copy of the GNU General Public License
> + * along with this program.  If not, see <http://www.gnu.org/licenses/>.
>   */
> +
>  #include <linux/linkage.h>
>  #include <asm/assembler.h>
>
> -       .text
> -       .align  5
> -
> -ENTRY(memset)
> -       ands    r3, r0, #3              @ 1 unaligned?
> -       mov     ip, r0                  @ preserve r0 as return value
> -       bne     6f                      @ 1
>  /*
> - * we know that the pointer in ip is aligned to a word boundary.
> - */
> -1:     orr     r1, r1, r1, lsl #8
> -       orr     r1, r1, r1, lsl #16
> -       mov     r3, r1
> -       cmp     r2, #16
> -       blt     4f
> -
> -#if ! CALGN(1)+0
> -
> -/*
> - * We need an 2 extra registers for this loop - use r8 and the LR
> - */
> -       stmfd   sp!, {r8, lr}
> -       mov     r8, r1
> -       mov     lr, r1
> -
> -2:     subs    r2, r2, #64
> -       stmgeia ip!, {r1, r3, r8, lr}   @ 64 bytes at a time.
> -       stmgeia ip!, {r1, r3, r8, lr}
> -       stmgeia ip!, {r1, r3, r8, lr}
> -       stmgeia ip!, {r1, r3, r8, lr}
> -       bgt     2b
> -       ldmeqfd sp!, {r8, pc}           @ Now <64 bytes to go.
> -/*
> - * No need to correct the count; we're only testing bits from now on
> + * Fill in the buffer with character c (alignment handled by the hardware)
> + *
> + * Parameters:
> + *     x0 - buf
> + *     x1 - c
> + *     x2 - n
> + * Returns:
> + *     x0 - buf
>   */
> -       tst     r2, #32
> -       stmneia ip!, {r1, r3, r8, lr}
> -       stmneia ip!, {r1, r3, r8, lr}
> -       tst     r2, #16
> -       stmneia ip!, {r1, r3, r8, lr}
> -       ldmfd   sp!, {r8, lr}
> -
> -#else
>
> +dstin          .req    x0
> +val            .req    w1
> +count          .req    x2
> +tmp1           .req    x3
> +tmp1w          .req    w3
> +tmp2           .req    x4
> +tmp2w          .req    w4
> +zva_len_x      .req    x5
> +zva_len                .req    w5
> +zva_bits_x     .req    x6
> +
> +A_l            .req    x7
> +A_lw           .req    w7
> +dst            .req    x8
> +tmp3w          .req    w9
> +tmp3           .req    x9
> +
> +       .weak memset
> +ENTRY(memset)
> +       mov     dst, dstin      /* Preserve return value.  */
> +       and     A_lw, val, #255
> +       orr     A_lw, A_lw, A_lw, lsl #8
> +       orr     A_lw, A_lw, A_lw, lsl #16
> +       orr     A_l, A_l, A_l, lsl #32
> +
> +       cmp     count, #15
> +       b.hi    .Lover16_proc
> +       /*All store maybe are non-aligned..*/
> +       tbz     count, #3, 1f
> +       str     A_l, [dst], #8
> +1:
> +       tbz     count, #2, 2f
> +       str     A_lw, [dst], #4
> +2:
> +       tbz     count, #1, 3f
> +       strh    A_lw, [dst], #2
> +3:
> +       tbz     count, #0, 4f
> +       strb    A_lw, [dst]
> +4:
> +       ret
> +
> +.Lover16_proc:
> +       /*Whether  the start address is aligned with 16.*/
> +       neg     tmp2, dst
> +       ands    tmp2, tmp2, #15
> +       b.eq    .Laligned
>  /*
> - * This version aligns the destination pointer in order to write
> - * whole cache lines at once.
> - */
> -
> -       stmfd   sp!, {r4-r8, lr}
> -       mov     r4, r1
> -       mov     r5, r1
> -       mov     r6, r1
> -       mov     r7, r1
> -       mov     r8, r1
> -       mov     lr, r1
> -
> -       cmp     r2, #96
> -       tstgt   ip, #31
> -       ble     3f
> -
> -       and     r8, ip, #31
> -       rsb     r8, r8, #32
> -       sub     r2, r2, r8
> -       movs    r8, r8, lsl #(32 - 4)
> -       stmcsia ip!, {r4, r5, r6, r7}
> -       stmmiia ip!, {r4, r5}
> -       tst     r8, #(1 << 30)
> -       mov     r8, r1
> -       strne   r1, [ip], #4
> -
> -3:     subs    r2, r2, #64
> -       stmgeia ip!, {r1, r3-r8, lr}
> -       stmgeia ip!, {r1, r3-r8, lr}
> -       bgt     3b
> -       ldmeqfd sp!, {r4-r8, pc}
> -
> -       tst     r2, #32
> -       stmneia ip!, {r1, r3-r8, lr}
> -       tst     r2, #16
> -       stmneia ip!, {r4-r7}
> -       ldmfd   sp!, {r4-r8, lr}
> -
> -#endif
> -
> -4:     tst     r2, #8
> -       stmneia ip!, {r1, r3}
> -       tst     r2, #4
> -       strne   r1, [ip], #4
> +* The count is not less than 16, we can use stp to store the start 16 bytes,
> +* then adjust the dst aligned with 16.This process will make the current
> +* memory address at alignment boundary.
> +*/
> +       stp     A_l, A_l, [dst] /*non-aligned store..*/
> +       /*make the dst aligned..*/
> +       sub     count, count, tmp2
> +       add     dst, dst, tmp2
> +
> +.Laligned:
> +       cbz     A_l, .Lzero_mem
> +
> +.Ltail_maybe_long:
> +       cmp     count, #64
> +       b.ge    .Lnot_short
> +.Ltail63:
> +       ands    tmp1, count, #0x30
> +       b.eq    3f
> +       cmp     tmp1w, #0x20
> +       b.eq    1f
> +       b.lt    2f
> +       stp     A_l, A_l, [dst], #16
> +1:
> +       stp     A_l, A_l, [dst], #16
> +2:
> +       stp     A_l, A_l, [dst], #16
>  /*
> - * When we get here, we've got less than 4 bytes to zero.  We
> - * may have an unaligned pointer as well.
> - */
> -5:     tst     r2, #2
> -       strneb  r1, [ip], #1
> -       strneb  r1, [ip], #1
> -       tst     r2, #1
> -       strneb  r1, [ip], #1
> -       mov     pc, lr
> -
> -6:     subs    r2, r2, #4              @ 1 do we have enough
> -       blt     5b                      @ 1 bytes to align with?
> -       cmp     r3, #2                  @ 1
> -       strltb  r1, [ip], #1            @ 1
> -       strleb  r1, [ip], #1            @ 1
> -       strb    r1, [ip], #1            @ 1
> -       add     r2, r2, r3              @ 1 (r2 = r2 - (4 - r3))
> -       b       1b
> +* The last store length is less than 16,use stp to write last 16 bytes.
> +* It will lead some bytes written twice and the access is non-aligned.
> +*/
> +3:
> +       ands    count, count, #15
> +       cbz     count, 4f
> +       add     dst, dst, count
> +       stp     A_l, A_l, [dst, #-16]   /* Repeat some/all of last store. */
> +4:
> +       ret
> +
> +       /*
> +       * Critical loop. Start at a new cache line boundary. Assuming
> +       * 64 bytes per line, this ensures the entire loop is in one line.
> +       */
> +.Lnot_short:
> +       sub     dst, dst, #16/* Pre-bias.  */
> +       sub     count, count, #64
> +1:
> +       stp     A_l, A_l, [dst, #16]
> +       stp     A_l, A_l, [dst, #32]
> +       stp     A_l, A_l, [dst, #48]
> +       stp     A_l, A_l, [dst, #64]!
> +       subs    count, count, #64
> +       b.ge    1b
> +       tst     count, #0x3f
> +       add     dst, dst, #16
> +       b.ne    .Ltail63
> +.Lexitfunc:
> +       ret
> +
> +       /*
> +       * For zeroing memory, check to see if we can use the ZVA feature to
> +       * zero entire 'cache' lines.
> +       */
> +.Lzero_mem:
> +       cmp     count, #63
> +       b.le    .Ltail63
> +       /*
> +       * For zeroing small amounts of memory, it's not worth setting up
> +       * the line-clear code.
> +       */
> +       cmp     count, #128
> +       b.lt    .Lnot_short /*count is at least  128 bytes*/
> +
> +       mrs     tmp1, dczid_el0
> +       tbnz    tmp1, #4, .Lnot_short
> +       mov     tmp3w, #4
> +       and     zva_len, tmp1w, #15     /* Safety: other bits reserved.  */
> +       lsl     zva_len, tmp3w, zva_len
> +
> +       ands    tmp3w, zva_len, #63
> +       /*
> +       * ensure the zva_len is not less than 64.
> +       * It is not meaningful to use ZVA if the block size is less than 64.
> +       */
> +       b.ne    .Lnot_short
> +.Lzero_by_line:
> +       /*
> +       * Compute how far we need to go to become suitably aligned. We're
> +       * already at quad-word alignment.
> +       */
> +       cmp     count, zva_len_x
> +       b.lt    .Lnot_short             /* Not enough to reach alignment.  */
> +       sub     zva_bits_x, zva_len_x, #1
> +       neg     tmp2, dst
> +       ands    tmp2, tmp2, zva_bits_x
> +       b.eq    2f                      /* Already aligned.  */
> +       /* Not aligned, check that there's enough to copy after alignment.*/
> +       sub     tmp1, count, tmp2
> +       /*
> +       * grantee the remain length to be ZVA is bigger than 64,
> +       * avoid to make the 2f's process over mem range.*/
> +       cmp     tmp1, #64
> +       ccmp    tmp1, zva_len_x, #8, ge /* NZCV=0b1000 */
> +       b.lt    .Lnot_short
> +       /*
> +       * We know that there's at least 64 bytes to zero and that it's safe
> +       * to overrun by 64 bytes.
> +       */
> +       mov     count, tmp1
> +1:
> +       stp     A_l, A_l, [dst]
> +       stp     A_l, A_l, [dst, #16]
> +       stp     A_l, A_l, [dst, #32]
> +       subs    tmp2, tmp2, #64
> +       stp     A_l, A_l, [dst, #48]
> +       add     dst, dst, #64
> +       b.ge    1b
> +       /* We've overrun a bit, so adjust dst downwards.*/
> +       add     dst, dst, tmp2
> +2:
> +       sub     count, count, zva_len_x
> +3:
> +       dc      zva, dst
> +       add     dst, dst, zva_len_x
> +       subs    count, count, zva_len_x
> +       b.ge    3b
> +       ands    count, count, zva_bits_x
> +       b.ne    .Ltail_maybe_long
> +       ret
>  ENDPROC(memset)
> -
> diff --git a/arch/arm/lib/runtime-offset.S b/arch/arm/lib/runtime-offset.S
> index f10c4c8..e368baa 100644
> --- a/arch/arm/lib/runtime-offset.S
> +++ b/arch/arm/lib/runtime-offset.S
> @@ -8,11 +8,11 @@
>   * we are currently running at.
>   */
>  ENTRY(get_runtime_offset)
> -1:     adr r0, 1b
> -       ldr r1, linkadr
> -       subs r0, r1, r0
> -THUMB( subs r0, r0, #1)
> -       mov pc, lr
> +1:     adr x0, 1b
> +       adr x1, get_runtime_offset
> +       subs x0, x1, x0
> +       subs x0, x0, #1
> +       ret
>
>  linkadr:
>  .word get_runtime_offset
> @@ -28,7 +28,7 @@ __ld_var_base:
>   */
>  .macro ld_var_entry name
>         ENTRY(__ld_var_\name)
> -               ldr r0, __\name
> +               ldr x0, __\name
>                 b 1f
>         __\name: .word \name - __ld_var_base
>         ENDPROC(__ld_var_\name)
> @@ -47,6 +47,6 @@ ld_var_entry __image_end
>  #endif
>
>  1:
> -       ldr r1, =__ld_var_base
> -       adds r0, r0, r1
> -       mov pc, lr
> +       ldr x1, =__ld_var_base
> +       adds x0, x0, x1
> +       ret
> diff --git a/arch/arm/mach-virt/Kconfig b/arch/arm/mach-virt/Kconfig
> new file mode 100644
> index 0000000..1f43606
> --- /dev/null
> +++ b/arch/arm/mach-virt/Kconfig
> @@ -0,0 +1,15 @@
> +if ARCH_VIRT
> +
> +config ARCH_TEXT_BASE
> +       hex
> +       default 0x40000000
> +
> +choice
> +       prompt "ARM Board type"
> +
> +config MACH_VIRT
> +       bool "ARM QEMU virt"
> +
> +endchoice
> +
> +endif
> diff --git a/arch/arm/mach-virt/Makefile b/arch/arm/mach-virt/Makefile
> new file mode 100644
> index 0000000..3924a10
> --- /dev/null
> +++ b/arch/arm/mach-virt/Makefile
> @@ -0,0 +1,3 @@
> +obj-y += devices.o reset.o
> +
> +lwl-y += lowlevel.o
> diff --git a/arch/arm/mach-virt/devices.c b/arch/arm/mach-virt/devices.c
> new file mode 100644
> index 0000000..999f463
> --- /dev/null
> +++ b/arch/arm/mach-virt/devices.c
> @@ -0,0 +1,30 @@
> +/*
> + * Copyright (C) 2016 Raphaël Poggi <poggi.raph@gmail.com>
> + *
> + * GPLv2 only
> + */
> +
> +#include <common.h>
> +#include <linux/amba/bus.h>
> +#include <asm/memory.h>
> +#include <mach/devices.h>
> +#include <linux/ioport.h>
> +
> +void virt_add_ddram(u32 size)
> +{
> +       arm_add_mem_device("ram0", 0x40000000, size);
> +}
> +
> +void virt_register_uart(unsigned id)
> +{
> +       resource_size_t start;
> +
> +       switch (id) {
> +       case 0:
> +               start = 0x09000000;
> +               break;
> +       default:
> +               return;
> +       }
> +       amba_apb_device_add(NULL, "uart-pl011", id, start, 4096, NULL, 0);
> +}
> diff --git a/arch/arm/mach-virt/include/mach/debug_ll.h b/arch/arm/mach-virt/include/mach/debug_ll.h
> new file mode 100644
> index 0000000..89b0692
> --- /dev/null
> +++ b/arch/arm/mach-virt/include/mach/debug_ll.h
> @@ -0,0 +1,24 @@
> +/*
> + * Copyright 2013 Jean-Christophe PLAGNIOL-VILLARD <plagniol@jcrosoft.com>
> + *
> + * GPLv2 only
> + */
> +
> +#ifndef __MACH_DEBUG_LL_H__
> +#define   __MACH_DEBUG_LL_H__
> +
> +#include <linux/amba/serial.h>
> +#include <io.h>
> +
> +#define DEBUG_LL_PHYS_BASE             0x10000000
> +#define DEBUG_LL_PHYS_BASE_RS1         0x1c000000
> +
> +#ifdef MP
> +#define DEBUG_LL_UART_ADDR DEBUG_LL_PHYS_BASE
> +#else
> +#define DEBUG_LL_UART_ADDR DEBUG_LL_PHYS_BASE_RS1
> +#endif
> +
> +#include <asm/debug_ll_pl011.h>
> +
> +#endif
> diff --git a/arch/arm/mach-virt/include/mach/devices.h b/arch/arm/mach-virt/include/mach/devices.h
> new file mode 100644
> index 0000000..9872c61
> --- /dev/null
> +++ b/arch/arm/mach-virt/include/mach/devices.h
> @@ -0,0 +1,13 @@
> +/*
> + * Copyright (C) 2016 Raphaël Poggi <poggi.raph@gmail.com>
> + *
> + * GPLv2 only
> + */
> +
> +#ifndef __ASM_ARCH_DEVICES_H__
> +#define __ASM_ARCH_DEVICES_H__
> +
> +void virt_add_ddram(u32 size);
> +void virt_register_uart(unsigned id);
> +
> +#endif /* __ASM_ARCH_DEVICES_H__ */
> diff --git a/arch/arm/mach-virt/lowlevel.c b/arch/arm/mach-virt/lowlevel.c
> new file mode 100644
> index 0000000..6f695a5
> --- /dev/null
> +++ b/arch/arm/mach-virt/lowlevel.c
> @@ -0,0 +1,19 @@
> +/*
> + * Copyright (C) 2013 Jean-Christophe PLAGNIOL-VILLARD <plagnio@jcrosoft.com>
> + *
> + * GPLv2 only
> + */
> +
> +#include <common.h>
> +#include <linux/sizes.h>
> +#include <asm/barebox-arm-head.h>
> +#include <asm/barebox-arm.h>
> +#include <asm/system_info.h>
> +
> +void barebox_arm_reset_vector(void)
> +{
> +       arm_cpu_lowlevel_init();
> +       arm_setup_stack(STACK_BASE);
> +
> +       barebox_arm_entry(0x40000000, SZ_512M, NULL);
> +}
> diff --git a/arch/arm/mach-virt/reset.c b/arch/arm/mach-virt/reset.c
> new file mode 100644
> index 0000000..fb895eb
> --- /dev/null
> +++ b/arch/arm/mach-virt/reset.c
> @@ -0,0 +1,24 @@
> +/*
> + * Copyright (C) 2016 Raphaël Poggi <poggi.raph@gmail.com>
> + *
> + * GPLv2 only
> + */
> +
> +#include <common.h>
> +#include <io.h>
> +#include <init.h>
> +#include <restart.h>
> +#include <mach/devices.h>
> +
> +static void virt_reset_soc(struct restart_handler *rst)
> +{
> +       hang();
> +}
> +
> +static int restart_register_feature(void)
> +{
> +       restart_handler_register_fn(virt_reset_soc);
> +
> +       return 0;
> +}
> +coredevice_initcall(restart_register_feature);
> --
> 2.8.0.rc3
>
>
> --
> Pengutronix e.K.                           |                             |
> Industrial Linux Solutions                 | http://www.pengutronix.de/  |
> Peiner Str. 6-8, 31137 Hildesheim, Germany | Phone: +49-5121-206917-0    |
> Amtsgericht Hildesheim, HRA 2686           | Fax:   +49-5121-206917-5555 |
>
> _______________________________________________
> barebox mailing list
> barebox@lists.infradead.org
> http://lists.infradead.org/mailman/listinfo/barebox

_______________________________________________
barebox mailing list
barebox@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/barebox

      reply	other threads:[~2016-05-23  8:34 UTC|newest]

Thread overview: 6+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2016-05-20  1:11 Raphael Poggi
2016-05-20  1:11 ` [PATCH 1/3] arch: add minimal aarch64 support Raphael Poggi
2016-05-20  1:11 ` [PATCH 2/3] arm64: add basic mmu support Raphael Poggi
2016-05-20  1:11 ` [PATCH 3/3] arm64: board: virt: set up mmu Raphael Poggi
2016-05-23  7:32 ` [PATCH 0/3] Add basic support for arm64 Sascha Hauer
2016-05-23  8:33   ` Raphaël Poggi [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=CACqcpZBSAGG5KF9R+dSNHEvpDS62UNvksZQd6mb9hm0YhrLzxw@mail.gmail.com \
    --to=poggi.raph@gmail.com \
    --cc=barebox@lists.infradead.org \
    --cc=s.hauer@pengutronix.de \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox