HelenOS sources

root/kernel/arch/amd64/src/asm.S

/* [<][>][^][v][top][bottom][index][help] */
/*
 * Copyright (c) 2005 Ondrej Palkovsky
 * All rights reserved.
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions
 * are met:
 *
 * - Redistributions of source code must retain the above copyright
 *   notice, this list of conditions and the following disclaimer.
 * - Redistributions in binary form must reproduce the above copyright
 *   notice, this list of conditions and the following disclaimer in the
 *   documentation and/or other materials provided with the distribution.
 * - The name of the author may not be used to endorse or promote products
 *   derived from this software without specific prior written permission.
 *
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 */

#include <abi/asmtool.h>
#include <arch/pm.h>
#include <arch/mm/page.h>
#include <arch/istate_struct.h>
#include <arch/kseg_struct.h>
#include <arch/cpu.h>
#include <arch/smp/apic.h>

.text

#define MEMCPY_DST   %rdi
#define MEMCPY_SRC   %rsi
#define MEMCPY_SIZE  %rdx

/**
 * Copy memory from/to userspace.
 *
 * This is almost conventional memcpy().
 * The difference is that there is a failover part
 * to where control is returned from a page fault if
 * the page fault occurs during copy_from_uspace()
 * or copy_to_uspace().
 *
 * @param MEMCPY_DST  Destination address.
 * @param MEMCPY_SRC  Source address.
 * @param MEMCPY_SIZE Number of bytes to copy.
 *
 * @retrun MEMCPY_DST on success, 0 on failure.
 *
 */
FUNCTION_BEGIN(memcpy_from_uspace)
FUNCTION_BEGIN(memcpy_to_uspace)
        movq MEMCPY_DST, %rax

        movq MEMCPY_SIZE, %rcx
        shrq $3, %rcx           /* size / 8 */

        rep movsq               /* copy as much as possible word by word */

        movq MEMCPY_SIZE, %rcx
        andq $7, %rcx           /* size % 8 */
        jz 0f

        rep movsb               /* copy the rest byte by byte */

        0:
                ret                 /* return MEMCPY_SRC, success */
FUNCTION_END(memcpy_from_uspace)
FUNCTION_END(memcpy_to_uspace)

SYMBOL(memcpy_from_uspace_failover_address)
SYMBOL(memcpy_to_uspace_failover_address)
        xorl %eax, %eax         /* return 0, failure */
        ret

/** Determine CPUID support
*
* @return 0 in EAX if CPUID is not support, 1 if supported.
*
*/
FUNCTION_BEGIN(has_cpuid)
        /* Load RFLAGS */
        pushfq
        popq %rax
        movq %rax, %rdx

        /* Flip the ID bit */
        xorl $RFLAGS_ID, %edx

        /* Store RFLAGS */
        pushq %rdx
        popfq
        pushfq

        /* Get the ID bit again */
        popq %rdx
        andl $RFLAGS_ID, %eax
        andl $RFLAGS_ID, %edx

        /* 0 if not supported, 1 if supported */
        xorl %edx, %eax
        ret
FUNCTION_END(has_cpuid)

FUNCTION_BEGIN(cpuid)
        /* Preserve %rbx across function calls */
        movq %rbx, %r10

        /* Load the command into %eax */
        movl %edi, %eax

        cpuid
        movl %eax, 0(%rsi)
        movl %ebx, 4(%rsi)
        movl %ecx, 8(%rsi)
        movl %edx, 12(%rsi)

        movq %r10, %rbx
        ret
FUNCTION_END(cpuid)

/** Enable local APIC
 *
 * Enable local APIC in MSR.
 *
 */
FUNCTION_BEGIN(enable_l_apic_in_msr)
        movl $AMD_MSR_APIC_BASE, %ecx
        rdmsr
        orl $(L_APIC_BASE | AMD_APIC_BASE_GE), %eax
        wrmsr
        ret
FUNCTION_END(enable_l_apic_in_msr)

/*
 * Size of the istate structure without the hardware-saved part and without the
 * error word.
 */
#define ISTATE_SOFT_SIZE        ISTATE_SIZE - (6 * 8)

/**
 * Mask for interrupts 0 - 31 (bits 0 - 31) where 0 means that int
 * has no error word  and 1 means interrupt with error word
 *
 */
#define ERROR_WORD_INTERRUPT_LIST  0x00027D00

.macro handler i
SYMBOL(int_\i)

        /*
         * Choose between version with error code and version without error
         * code.
         */

        .iflt \i-32
                .if (1 << \i) & ERROR_WORD_INTERRUPT_LIST
                        /*
                         * Version with error word.
                         */
                        subq $ISTATE_SOFT_SIZE, %rsp
                .else
                        /*
                         * Version without error word.
                         */
                        subq $(ISTATE_SOFT_SIZE + 8), %rsp
                .endif
        .else
                /*
                 * Version without error word.
                 */
                subq $(ISTATE_SOFT_SIZE + 8), %rsp
        .endif

        /*
         * Save the general purpose registers.
         */
        movq %rax, ISTATE_OFFSET_RAX(%rsp)
        movq %rbx, ISTATE_OFFSET_RBX(%rsp)
        movq %rcx, ISTATE_OFFSET_RCX(%rsp)
        movq %rdx, ISTATE_OFFSET_RDX(%rsp)
        movq %rsi, ISTATE_OFFSET_RSI(%rsp)
        movq %rdi, ISTATE_OFFSET_RDI(%rsp)
        movq %rbp, ISTATE_OFFSET_RBP(%rsp)
        movq %r8, ISTATE_OFFSET_R8(%rsp)
        movq %r9, ISTATE_OFFSET_R9(%rsp)
        movq %r10, ISTATE_OFFSET_R10(%rsp)
        movq %r11, ISTATE_OFFSET_R11(%rsp)
        movq %r12, ISTATE_OFFSET_R12(%rsp)
        movq %r13, ISTATE_OFFSET_R13(%rsp)
        movq %r14, ISTATE_OFFSET_R14(%rsp)
        movq %r15, ISTATE_OFFSET_R15(%rsp)

        /*
         * Is this trap from the kernel?
         */
        cmpq $(GDT_SELECTOR(KTEXT_DES)), ISTATE_OFFSET_CS(%rsp)
        jz 0f

        /*
         * Switch to kernel FS base.
         */
        swapgs
        movl $AMD_MSR_FS, %ecx
        movl %gs:KSEG_OFFSET_FSBASE, %eax
        movl %gs:KSEG_OFFSET_FSBASE+4, %edx
        wrmsr
        swapgs

        /*
         * Imitate a regular stack frame linkage.
         * Stop stack traces here if we came from userspace.
         */
0:      movl $0x0, %edx
        cmovnzq %rdx, %rbp

        movq %rbp, ISTATE_OFFSET_RBP_FRAME(%rsp)
        movq ISTATE_OFFSET_RIP(%rsp), %rax
        movq %rax, ISTATE_OFFSET_RIP_FRAME(%rsp)
        leaq ISTATE_OFFSET_RBP_FRAME(%rsp), %rbp

        movq $(\i), %rdi   /* pass intnum in the first argument */
        movq %rsp, %rsi    /* pass istate address in the second argument */

        cld

        /* Call exc_dispatch(i, istate) */
        call exc_dispatch

        /*
         * Restore all scratch registers and the preserved registers we have
         * clobbered in this handler (i.e. RBP).
         */
        movq ISTATE_OFFSET_RAX(%rsp), %rax
        movq ISTATE_OFFSET_RCX(%rsp), %rcx
        movq ISTATE_OFFSET_RDX(%rsp), %rdx
        movq ISTATE_OFFSET_RSI(%rsp), %rsi
        movq ISTATE_OFFSET_RDI(%rsp), %rdi
        movq ISTATE_OFFSET_RBP(%rsp), %rbp
        movq ISTATE_OFFSET_R8(%rsp), %r8
        movq ISTATE_OFFSET_R9(%rsp), %r9
        movq ISTATE_OFFSET_R10(%rsp), %r10
        movq ISTATE_OFFSET_R11(%rsp), %r11

        /* $8 = Skip error word */
        addq $(ISTATE_SOFT_SIZE + 8), %rsp
        iretq
.endm

#define LIST_0_63 \
        0, 1, 2, 3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,\
        28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,\
        53,54,55,56,57,58,59,60,61,62,63

SYMBOL(interrupt_handlers)
.irp cnt, LIST_0_63
        handler \cnt
.endr

/** Low-level syscall handler
 *
 * Registers on entry:
 *
 * @param %rcx Userspace return address.
 * @param %r11 Userspace RLFAGS.
 *
 * @param %rax Syscall number.
 * @param %rdi 1st syscall argument.
 * @param %rsi 2nd syscall argument.
 * @param %rdx 3rd syscall argument.
 * @param %r10 4th syscall argument. Used instead of RCX because
 *             the SYSCALL instruction clobbers it.
 * @param %r8  5th syscall argument.
 * @param %r9  6th syscall argument.
 *
 * @return Return value is in %rax.
 *
 */
SYMBOL(syscall_entry)
        /* Switch to hidden %gs */
        swapgs

        movq %rsp, %gs:KSEG_OFFSET_USTACK_RSP  /* save this thread's user RSP */
        movq %gs:KSEG_OFFSET_KSTACK_RSP, %rsp  /* set this thread's kernel RSP */

        /*
         * Note that the space needed for the imitated istate structure has been
         * preallocated for us in thread_create_arch() and set in
         * before_thread_runs_arch().
         */

        /*
         * Save the general purpose registers and push the 7th argument (syscall
         * number) onto the stack. Note that the istate structure has a layout
         * which supports this.
         */
        movq %rax, ISTATE_OFFSET_RAX(%rsp)  /* 7th argument, passed on stack */
        movq %rbx, ISTATE_OFFSET_RBX(%rsp)  /* observability */
        movq %rcx, ISTATE_OFFSET_RCX(%rsp)  /* userspace RIP */
        movq %rdx, ISTATE_OFFSET_RDX(%rsp)  /* 3rd argument, observability */
        movq %rsi, ISTATE_OFFSET_RSI(%rsp)  /* 2nd argument, observability */
        movq %rdi, ISTATE_OFFSET_RDI(%rsp)  /* 1st argument, observability */
        movq %rbp, ISTATE_OFFSET_RBP(%rsp)  /* need to preserve userspace RBP */
        movq %r8, ISTATE_OFFSET_R8(%rsp)    /* 5th argument, observability */
        movq %r9, ISTATE_OFFSET_R9(%rsp)    /* 6th argument, observability */
        movq %r10, ISTATE_OFFSET_R10(%rsp)  /* 4th argument, observability */
        movq %r11, ISTATE_OFFSET_R11(%rsp)  /* low 32 bits userspace RFLAGS */
        movq %r12, ISTATE_OFFSET_R12(%rsp)  /* observability */
        movq %r13, ISTATE_OFFSET_R13(%rsp)  /* observability */
        movq %r14, ISTATE_OFFSET_R14(%rsp)  /* observability */
        movq %r15, ISTATE_OFFSET_R15(%rsp)  /* observability */

        /*
         * Switch to kernel FS base.
         */
        movl $AMD_MSR_FS, %ecx
        movl %gs:KSEG_OFFSET_FSBASE, %eax
        movl %gs:KSEG_OFFSET_FSBASE+4, %edx
        wrmsr
        movq ISTATE_OFFSET_RDX(%rsp), %rdx      /* restore 3rd argument */

        /*
         * Save the return address and the userspace stack on locations that
         * would normally be taken by them.
         */
        movq %gs:KSEG_OFFSET_USTACK_RSP, %rax
        movq %rax, ISTATE_OFFSET_RSP(%rsp)
        movq %rcx, ISTATE_OFFSET_RIP(%rsp)

        /*
         * Imitate a regular stack frame linkage.
         */
        movq $0, ISTATE_OFFSET_RBP_FRAME(%rsp)
        movq %rcx, ISTATE_OFFSET_RIP_FRAME(%rsp)
        leaq ISTATE_OFFSET_RBP_FRAME(%rsp), %rbp

        /* Switch back to normal %gs */
        swapgs
        sti

        /* Copy the 4th argument where it is expected  */
        movq %r10, %rcx

        /*
         * Call syscall_handler() with the 7th argument passed on stack.
         */
        call syscall_handler

        /*
         * Test if the saved return address is canonical and not-kernel.
         * We do this by looking at the 16 most significant bits
         * of the saved return address (two bytes at offset 6).
         */
        testw $0xffff, ISTATE_OFFSET_RIP+6(%rsp)
        jnz bad_rip

        cli

        /*
         * Restore registers needed for return via the SYSRET instruction and
         * the clobbered preserved registers (i.e. RBP).
         */
        movq ISTATE_OFFSET_RBP(%rsp), %rbp
        movq ISTATE_OFFSET_RCX(%rsp), %rcx
        movq ISTATE_OFFSET_R11(%rsp), %r11
        movq ISTATE_OFFSET_RSP(%rsp), %rsp

        /*
         * Clear the rest of the scratch registers to prevent information leak.
         * The 32-bit XOR on the low GPRs actually clears the entire 64-bit
         * register and the instruction is shorter.
         */
        xorl %edx, %edx
        xorl %esi, %esi
        xorl %edi, %edi
        xorq %r8, %r8
        xorq %r9, %r9
        xorq %r10, %r10

        sysretq

bad_rip:
        movq %rsp, %rdi
        movabs $bad_rip_msg, %rsi
        xorb %al, %al
        callq fault_from_uspace
        /* not reached */

bad_rip_msg:
        .asciz "Invalid instruction pointer."

/** Print Unicode character to an early display device.
 *
 * Since the EGA can only display Extended ASCII (usually
 * ISO Latin 1) characters, some of the Unicode characters
 * can be displayed in a wrong way. Only newline and backspace
 * are interpreted, all other characters (even unprintable) are
 * printed verbatim.
 *
 * @param %rdi Unicode character to be printed.
 *
 */
FUNCTION_BEGIN(early_putuchar)

#if (defined(CONFIG_DEBUG_EARLY_PRINT))

#if (defined(CONFIG_L4RE_UVMM_EARLY_PRINT))
        xorl %eax, %eax  /* RAX==0: uvmm's print hypercall */
        mov %rdi, %rcx   /* RCX:    printed character */
        vmcall
#endif

#if ((defined(CONFIG_EGA)) && (!defined(CONFIG_FB)))

        /* Prologue, save preserved registers */
        pushq %rbp
        movq %rsp, %rbp
        pushq %rbx

        movq %rdi, %rsi
        movq $(PA2KA(0xb8000)), %rdi  /* base of EGA text mode memory */
        xorl %eax, %eax

        /* Read bits 8 - 15 of the cursor address */
        movw $0x3d4, %dx
        movb $0xe, %al
        outb %al, %dx

        movw $0x3d5, %dx
        inb %dx, %al
        shl $8, %ax

        /* Read bits 0 - 7 of the cursor address */
        movw $0x3d4, %dx
        movb $0xf, %al
        outb %al, %dx

        movw $0x3d5, %dx
        inb %dx, %al

        /* Sanity check for the cursor on screen */
        cmp $2000, %ax
        jb early_putuchar_cursor_ok

                movw $1998, %ax

        early_putuchar_cursor_ok:

        movw %ax, %bx
        shl $1, %rax
        addq %rax, %rdi

        movq %rsi, %rax

        cmp $0x0a, %al
        jne early_putuchar_backspace

                /* Interpret newline */

                movw %bx, %ax  /* %bx -> %dx:%ax */
                xorw %dx, %dx

                movw $80, %cx
                idivw %cx, %ax  /* %dx = %bx % 80 */

                /* %bx <- %bx + 80 - (%bx % 80) */
                addw %cx, %bx
                subw %dx, %bx

                jmp early_putuchar_skip

        early_putuchar_backspace:

                cmp $0x08, %al
                jne early_putuchar_print

                /* Interpret backspace */

                cmp $0x0000, %bx
                je early_putuchar_skip

                dec %bx
                jmp early_putuchar_skip

        early_putuchar_print:

                /* Print character */

                movb $0x0e, %ah  /* black background, yellow foreground */
                stosw
                inc %bx

        early_putuchar_skip:

        /* Sanity check for the cursor on the last line */
        cmp $2000, %bx
        jb early_putuchar_no_scroll

                /* Scroll the screen (24 rows) */
                movq $(PA2KA(0xb80a0)), %rsi
                movq $(PA2KA(0xb8000)), %rdi
                movl $480, %ecx
                rep movsq

                /* Clear the 24th row */
                xorl %eax, %eax
                movl $20, %ecx
                rep stosq

                /* Go to row 24 */
                movw $1920, %bx

        early_putuchar_no_scroll:

        /* Write bits 8 - 15 of the cursor address */
        movw $0x3d4, %dx
        movb $0xe, %al
        outb %al, %dx

        movw $0x3d5, %dx
        movb %bh, %al
        outb %al, %dx

        /* Write bits 0 - 7 of the cursor address */
        movw $0x3d4, %dx
        movb $0xf, %al
        outb %al, %dx

        movw $0x3d5, %dx
        movb %bl, %al
        outb %al, %dx

        /* Epilogue, restore preserved registers */
        popq %rbx
        leave

#endif

#endif

        ret
FUNCTION_END(early_putuchar)

/* [<][>][^][v][top][bottom][index][help] */
HelenOS homepage, sources at GitHub