config root man

Current Path : /sys/amd64/compile/hs32/modules/usr/src/sys/modules/drm2/drm2/@/boot/ia64/common/

FreeBSD hs32.drive.ne.jp 9.1-RELEASE FreeBSD 9.1-RELEASE #1: Wed Jan 14 12:18:08 JST 2015 root@hs32.drive.ne.jp:/sys/amd64/compile/hs32 amd64
Upload File :
Current File : //sys/amd64/compile/hs32/modules/usr/src/sys/modules/drm2/drm2/@/boot/ia64/common/copy.c

/*-
 * Copyright (c) 2006 Marcel Moolenaar
 * All rights reserved.
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions
 * are met:
 *
 * 1. Redistributions of source code must retain the above copyright
 *    notice, this list of conditions and the following disclaimer.
 * 2. Redistributions in binary form must reproduce the above copyright
 *    notice, this list of conditions and the following disclaimer in the
 *    documentation and/or other materials provided with the distribution.
 *
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 */

#include <sys/cdefs.h>
__FBSDID("$FreeBSD: release/9.1.0/sys/boot/ia64/common/copy.c 221269 2011-04-30 20:16:49Z marcel $");

#include <stand.h>
#include <machine/param.h>
#include <machine/pte.h>

#include "libia64.h"

u_int ia64_legacy_kernel;

uint64_t *ia64_pgtbl;
uint32_t ia64_pgtblsz;

static int
pgtbl_extend(u_int idx)
{
	vm_paddr_t pa;
	uint64_t *pgtbl;
	uint32_t pgtblsz;
	u_int pot;

	pgtblsz = (idx + 1) << 3;

	/* The minimum size is 4KB. */
	if (pgtblsz < 4096)
		pgtblsz = 4096;

	/* Find the next higher power of 2. */
	pgtblsz--;
	for (pot = 1; pot < 32; pot <<= 1)
		pgtblsz = pgtblsz | (pgtblsz >> pot);
	pgtblsz++;

	/* The maximum size is 1MB. */
	if (pgtblsz > 1048576)
		return (ENOMEM);

	/* Make sure the size is a valid (mappable) page size. */
	if (pgtblsz == 32*1024 || pgtblsz == 128*1024 || pgtblsz == 512*1024)
		pgtblsz <<= 1;

	/* Allocate naturally aligned memory. */
	pa = ia64_platform_alloc(0, pgtblsz);
	if (pa == ~0UL)
		return (ENOMEM);
	pgtbl = (void *)pa;

	/* Initialize new page table. */
	if (ia64_pgtbl != NULL && ia64_pgtbl != pgtbl)
		bcopy(ia64_pgtbl, pgtbl, ia64_pgtblsz);
	bzero(pgtbl + (ia64_pgtblsz >> 3), pgtblsz - ia64_pgtblsz);

	if (ia64_pgtbl != NULL && ia64_pgtbl != pgtbl)
		ia64_platform_free(0, (uintptr_t)ia64_pgtbl, ia64_pgtblsz);

	ia64_pgtbl = pgtbl;
	ia64_pgtblsz = pgtblsz;
	return (0);
}

void *
ia64_va2pa(vm_offset_t va, size_t *len)
{
	uint64_t pa, pte;
	u_int idx, ofs;
	int error;

	/* Backward compatibility. */
	if (va >= IA64_RR_BASE(7)) {
		ia64_legacy_kernel = 1;
		pa = IA64_RR_MASK(va);
		return ((void *)pa);
	}

	if (va < IA64_PBVM_BASE) {
		error = EINVAL;
		goto fail;
	}

	ia64_legacy_kernel = 0;

	idx = (va - IA64_PBVM_BASE) >> IA64_PBVM_PAGE_SHIFT;
	if (idx >= (ia64_pgtblsz >> 3)) {
		error = pgtbl_extend(idx);
		if (error)
			goto fail;
	}

	ofs = va & IA64_PBVM_PAGE_MASK;
	pte = ia64_pgtbl[idx];
	if ((pte & PTE_PRESENT) == 0) {
		pa = ia64_platform_alloc(va - ofs, IA64_PBVM_PAGE_SIZE);
		if (pa == ~0UL) {
			error = ENOMEM;
			goto fail;
		}
		pte = PTE_AR_RWX | PTE_DIRTY | PTE_ACCESSED | PTE_PRESENT;
		pte |= (pa & PTE_PPN_MASK);
		ia64_pgtbl[idx] = pte;
	}
	pa = (pte & PTE_PPN_MASK) + ofs;

	/* We can not cross page boundaries (in general). */
	if (*len + ofs > IA64_PBVM_PAGE_SIZE)
		*len = IA64_PBVM_PAGE_SIZE - ofs;

	return ((void *)pa);

 fail:
	*len = 0;
	return (NULL);
}

ssize_t
ia64_copyin(const void *src, vm_offset_t va, size_t len)
{
	void *pa;
	ssize_t res;
	size_t sz;

	res = 0;
	while (len > 0) {
		sz = len;
		pa = ia64_va2pa(va, &sz);
		if (sz == 0)
			break;
		bcopy(src, pa, sz);
		len -= sz;
		res += sz;
		va += sz;
	}
	return (res);
}

ssize_t
ia64_copyout(vm_offset_t va, void *dst, size_t len)
{
	void *pa;
	ssize_t res;
	size_t sz;

	res = 0;
	while (len > 0) {
		sz = len;
		pa = ia64_va2pa(va, &sz);
		if (sz == 0)
			break;
		bcopy(pa, dst, sz);
		len -= sz;
		res += sz;
		va += sz;
	}
	return (res);
}

uint64_t
ia64_loadaddr(u_int type, void *data, uint64_t addr)
{
	uint64_t align;

	/*
	 * Align ELF objects at PBVM page boundaries.  Align all other
	 * objects at cache line boundaries for good measure.
	 */
	align = (type == LOAD_ELF) ? IA64_PBVM_PAGE_SIZE : CACHE_LINE_SIZE;
	return ((addr + align - 1) & ~(align - 1));
}

ssize_t
ia64_readin(int fd, vm_offset_t va, size_t len)
{
	void *pa;
	ssize_t res, s;
	size_t sz;

	res = 0;
	while (len > 0) {
		sz = len;
		pa = ia64_va2pa(va, &sz);
		if (sz == 0)
			break;
		s = read(fd, pa, sz);
		if (s <= 0)
			break;
		len -= s;
		res += s;
		va += s;
	}
	return (res);
}

Man Man