[BACK]Return to htab.c CVS log [TXT][DIR] Up to [Development] / linux-2.4-xfs / arch / ppc64 / kernel

File: [Development] / linux-2.4-xfs / arch / ppc64 / kernel / htab.c (download)

Revision 1.1, Wed Dec 31 00:54:49 2003 UTC (13 years, 9 months ago) by cattelan
Branch: MAIN
CVS Tags: HEAD

Initial Import 2.4.24pre2

/*
 * PowerPC64 port by Mike Corrigan and Dave Engebretsen
 *   {mikejc|engebret}@us.ibm.com
 *
 *    Copyright (c) 2000 Mike Corrigan <mikejc@us.ibm.com>
 *
 * SMP scalability work:
 *    Copyright (C) 2001 Anton Blanchard <anton@au.ibm.com>, IBM
 * 
 *    Module name: htab.c
 *
 *    Description:
 *      PowerPC Hashed Page Table functions
 *
 * This program is free software; you can redistribute it and/or
 * modify it under the terms of the GNU General Public License
 * as published by the Free Software Foundation; either version
 * 2 of the License, or (at your option) any later version.
 */

#include <linux/config.h>
#include <linux/spinlock.h>
#include <linux/errno.h>
#include <linux/sched.h>
#include <linux/proc_fs.h>
#include <linux/stat.h>
#include <linux/sysctl.h>
#include <linux/ctype.h>
#include <linux/cache.h>

#include <asm/ppcdebug.h>
#include <asm/processor.h>
#include <asm/pgtable.h>
#include <asm/mmu.h>
#include <asm/mmu_context.h>
#include <asm/page.h>
#include <asm/types.h>
#include <asm/uaccess.h>
#include <asm/naca.h>
#include <asm/pmc.h>
#include <asm/machdep.h>
#include <asm/lmb.h>
#include <asm/abs_addr.h>
#include <asm/io.h>
#include <asm/eeh.h>
#include <asm/hvcall.h>
#include <asm/iSeries/LparData.h>
#include <asm/iSeries/HvCallHpt.h>
#include <asm/cputable.h>

/*
 * Note:  pte   --> Linux PTE
 *        HPTE  --> PowerPC Hashed Page Table Entry
 *
 * Execution context:
 *   htab_initialize is called with the MMU off (of course), but
 *   the kernel has been copied down to zero so it can directly
 *   reference global data.  At this point it is very difficult
 *   to print debug info.
 *
 */

HTAB htab_data = {NULL, 0, 0, 0, 0};

extern unsigned long _SDR1;
extern unsigned long klimit;

void make_pte(HPTE *htab, unsigned long va, unsigned long pa,
	      int mode, unsigned long hash_mask, int large);
long plpar_pte_enter(unsigned long flags,
		     unsigned long ptex,
		     unsigned long new_pteh, unsigned long new_ptel,
		     unsigned long *old_pteh_ret, unsigned long *old_ptel_ret);
static long hpte_remove(unsigned long hpte_group);
static long rpa_lpar_hpte_remove(unsigned long hpte_group);
static long iSeries_hpte_remove(unsigned long hpte_group);
inline unsigned long get_lock_slot(unsigned long vpn);

static spinlock_t pSeries_tlbie_lock = SPIN_LOCK_UNLOCKED;
static spinlock_t pSeries_lpar_tlbie_lock = SPIN_LOCK_UNLOCKED;

#define LOCK_SPLIT
#ifdef LOCK_SPLIT
hash_table_lock_t hash_table_lock[128] __cacheline_aligned_in_smp = { [0 ... 31] = {SPIN_LOCK_UNLOCKED}};
#else
hash_table_lock_t hash_table_lock[1] __cacheline_aligned_in_smp = { [0] = {SPIN_LOCK_UNLOCKED}};
#endif

#define KB (1024)
#define MB (1024*KB)

static inline void
loop_forever(void)
{
	volatile unsigned long x = 1;
	for(;x;x|=1)
		;
}

static inline void
create_pte_mapping(unsigned long start, unsigned long end,
		   unsigned long mode, unsigned long mask, int large)
{
	unsigned long addr;
	HPTE *htab = (HPTE *)__v2a(htab_data.htab);
	unsigned int step;

	if (large)
		step = 16*MB;
	else
		step = 4*KB;

	for (addr = start; addr < end; addr += step) {
		unsigned long vsid = get_kernel_vsid(addr);
		unsigned long va = (vsid << 28) | (addr & 0xfffffff);
		make_pte(htab, va, (unsigned long)__v2a(addr), 
			 mode, mask, large);
	}
}

void
htab_initialize(void)
{
	unsigned long table, htab_size_bytes;
	unsigned long pteg_count;
	unsigned long mode_rw, mask, lock_shift;

#if 0
	/* Can't really do the call below since it calls the normal RTAS
	 * entry point and we're still relocate off at the moment.
	 * Temporarily diabling until it can call through the relocate off
	 * RTAS entry point.  -Peter
	 */
	ppc64_boot_msg(0x05, "htab init");
#endif
	/*
	 * Calculate the required size of the htab.  We want the number of
	 * PTEGs to equal one half the number of real pages.
	 */ 
	htab_size_bytes = 1UL << naca->pftSize;
	pteg_count = htab_size_bytes >> 7;

	/* For debug, make the HTAB 1/8 as big as it normally would be. */
	ifppcdebug(PPCDBG_HTABSIZE) {
		pteg_count >>= 3;
		htab_size_bytes = pteg_count << 7;
	}

	htab_data.htab_num_ptegs = pteg_count;
	htab_data.htab_hash_mask = pteg_count - 1;

	/* 
	 * Calculate the number of bits to shift the pteg selector such that we
	 * use the high order 8 bits to select a page table lock.
	 */
	asm ("cntlzd %0,%1" : "=r" (lock_shift) : 
	     "r" (htab_data.htab_hash_mask));
	htab_data.htab_lock_shift = (64 - lock_shift) - 8;

	if(systemcfg->platform == PLATFORM_PSERIES) {
		/* Find storage for the HPT.  Must be contiguous in
		 * the absolute address space.
		 */
		table = lmb_alloc(htab_size_bytes, htab_size_bytes);
		if ( !table ) {
			ppc64_terminate_msg(0x20, "hpt space");
			loop_forever();
		}
		htab_data.htab = (HPTE *)__a2v(table);

		/* htab absolute addr + encoded htabsize */
		_SDR1 = table + __ilog2(pteg_count) - 11;

		/* Initialize the HPT with no entries */
		memset((void *)table, 0, htab_size_bytes);
	} else {
		/* Using a hypervisor which owns the htab */
		htab_data.htab = NULL;
		_SDR1 = 0; 
	}

	mode_rw = _PAGE_ACCESSED | _PAGE_COHERENT | PP_RWXX;
	mask = pteg_count-1;

	/* XXX we currently map kernel text rw, should fix this */
	if ((systemcfg->platform & PLATFORM_PSERIES) &&
	    (cur_cpu_spec->cpu_features & 
	     CPU_FTR_16M_PAGE) &&
	    (systemcfg->physicalMemorySize > 256*MB)) {
		create_pte_mapping((unsigned long)KERNELBASE, 
				   KERNELBASE + 256*MB, mode_rw, mask, 0);
		create_pte_mapping((unsigned long)KERNELBASE + 256*MB, 
				   KERNELBASE + (systemcfg->physicalMemorySize), 
				   mode_rw, mask, 1);
	} else {
		create_pte_mapping((unsigned long)KERNELBASE, 
				   KERNELBASE+(systemcfg->physicalMemorySize), 
				   mode_rw, mask, 0);
	}
#if 0
	/* Can't really do the call below since it calls the normal RTAS
	 * entry point and we're still relocate off at the moment.
	 * Temporarily diabling until it can call through the relocate off
	 * RTAS entry point.  -Peter
	 */
	ppc64_boot_msg(0x06, "htab done");
#endif
}
#undef KB
#undef MB

/*
 * Create a pte. Used during initialization only.
 * We assume the PTE will fit in the primary PTEG.
 */
void make_pte(HPTE *htab, unsigned long va, unsigned long pa,
	      int mode, unsigned long hash_mask, int large)
{
	HPTE *hptep, local_hpte, rhpte;
	unsigned long hash, vpn, flags, lpar_rc;
	unsigned long i, dummy1, dummy2;
	long slot;

	if (large)
		vpn = va >> LARGE_PAGE_SHIFT;
	else
		vpn = va >> PAGE_SHIFT;

	hash = hpt_hash(vpn, large);

	local_hpte.dw1.dword1 = pa | mode;
	local_hpte.dw0.dword0 = 0;
	local_hpte.dw0.dw0.avpn = va >> 23;
	local_hpte.dw0.dw0.bolted = 1;		/* bolted */
	if (large) {
		local_hpte.dw0.dw0.l = 1;	/* large page */
		local_hpte.dw0.dw0.avpn &= ~0x1UL;
	}
	local_hpte.dw0.dw0.v = 1;

	if (systemcfg->platform == PLATFORM_PSERIES) {
		hptep  = htab + ((hash & hash_mask)*HPTES_PER_GROUP);

		for (i = 0; i < 8; ++i, ++hptep) {
			if (hptep->dw0.dw0.v == 0) {		/* !valid */
				*hptep = local_hpte;
				return;
			}
		}
	} else if (systemcfg->platform == PLATFORM_PSERIES_LPAR) {
		slot = ((hash & hash_mask)*HPTES_PER_GROUP);
		
		/* Set CEC cookie to 0                   */
		/* Zero page = 0                         */
		/* I-cache Invalidate = 0                */
		/* I-cache synchronize = 0               */
		/* Exact = 0 - modify any entry in group */
		flags = 0;
		
		lpar_rc =  plpar_pte_enter(flags, slot, local_hpte.dw0.dword0,
					   local_hpte.dw1.dword1, 
					   &dummy1, &dummy2);
		return;
	} else if (systemcfg->platform == PLATFORM_ISERIES_LPAR) {
		slot = HvCallHpt_findValid(&rhpte, vpn);
		if (slot < 0) {
			/* Must find space in primary group */
			panic("hash_page: hpte already exists\n");
		}
		HvCallHpt_addValidate(slot, 0, (HPTE *)&local_hpte );
		return;
	}

	/* We should _never_ get here and too early to call xmon. */
	ppc64_terminate_msg(0x22, "hpte platform");
	loop_forever();
}

/*
 * find_linux_pte returns the address of a linux pte for a given 
 * effective address and directory.  If not found, it returns zero.
 */
pte_t *find_linux_pte(pgd_t *pgdir, unsigned long ea)
{
	pgd_t *pg;
	pmd_t *pm;
	pte_t *pt = NULL;
	pte_t pte;

	pg = pgdir + pgd_index(ea);
	if (!pgd_none(*pg)) {
		pm = pmd_offset(pg, ea);
		if (!pmd_none(*pm)) { 
			pt = pte_offset(pm, ea);
			pte = *pt;
			if (!pte_present(pte))
				pt = NULL;
		}
	}

	return pt;
}

static inline unsigned long computeHptePP(unsigned long pte)
{
	return (pte & _PAGE_USER) |
		(((pte & _PAGE_USER) >> 1) &
		 ((~((pte >> 2) &	/* _PAGE_RW */
		     (pte >> 7))) &	/* _PAGE_DIRTY */
		  1));
}

/*
 * Handle a fault by adding an HPTE. If the address can't be determined
 * to be valid via Linux page tables, return 1. If handled return 0
 */
int __hash_page(unsigned long ea, unsigned long access, 
		unsigned long vsid, pte_t *ptep)
{
	unsigned long va, vpn;
	unsigned long newpp, prpn;
	unsigned long hpteflags, lock_slot;
	long slot;
	pte_t old_pte, new_pte;

	/* Search the Linux page table for a match with va */
	va = (vsid << 28) | (ea & 0x0fffffff);
	vpn = va >> PAGE_SHIFT;
	lock_slot = get_lock_slot(vpn); 

	/* Acquire the hash table lock to guarantee that the linux
	 * pte we fetch will not change
	 */
	spin_lock(&hash_table_lock[lock_slot].lock);
	
	/* 
	 * Check the user's access rights to the page.  If access should be
	 * prevented then send the problem up to do_page_fault.
	 */
#ifdef CONFIG_SHARED_MEMORY_ADDRESSING
	access |= _PAGE_PRESENT;
	if (unlikely(access & ~(pte_val(*ptep)))) {
		if(!(((ea >> SMALLOC_EA_SHIFT) == 
		      (SMALLOC_START >> SMALLOC_EA_SHIFT)) &&
		     ((current->thread.flags) & PPC_FLAG_SHARED))) {
			spin_unlock(&hash_table_lock[lock_slot].lock);
			return 1;
		}
	}
#else
	access |= _PAGE_PRESENT;
	if (unlikely(access & ~(pte_val(*ptep)))) {
		spin_unlock(&hash_table_lock[lock_slot].lock);
		return 1;
	}
#endif

	/* 
	 * We have found a pte (which was present).
	 * The spinlocks prevent this status from changing
	 * The hash_table_lock prevents the _PAGE_HASHPTE status
	 * from changing (RPN, DIRTY and ACCESSED too)
	 * The page_table_lock prevents the pte from being 
	 * invalidated or modified
	 */

	/*
	 * At this point, we have a pte (old_pte) which can be used to build
	 * or update an HPTE. There are 2 cases:
	 *
	 * 1. There is a valid (present) pte with no associated HPTE (this is 
	 *	the most common case)
	 * 2. There is a valid (present) pte with an associated HPTE. The
	 *	current values of the pp bits in the HPTE prevent access
	 *	because we are doing software DIRTY bit management and the
	 *	page is currently not DIRTY. 
	 */

	old_pte = *ptep;
	new_pte = old_pte;

	/* If the attempted access was a store */
	if (access & _PAGE_RW)
		pte_val(new_pte) |= _PAGE_ACCESSED | _PAGE_DIRTY;
	else
		pte_val(new_pte) |= _PAGE_ACCESSED;

	newpp = computeHptePP(pte_val(new_pte));
	
	/* Check if pte already has an hpte (case 2) */
	if (unlikely(pte_val(old_pte) & _PAGE_HASHPTE)) {
		/* There MIGHT be an HPTE for this pte */
		unsigned long hash, slot, secondary;

		/* XXX fix large pte flag */
		hash = hpt_hash(vpn, 0);
		secondary = (pte_val(old_pte) & _PAGE_SECONDARY) >> 15;
		if (secondary)
			hash = ~hash;
		slot = (hash & htab_data.htab_hash_mask) * HPTES_PER_GROUP;
		slot += (pte_val(old_pte) & _PAGE_GROUP_IX) >> 12;

		/* XXX fix large pte flag */
		if (ppc_md.hpte_updatepp(slot, secondary, 
					 newpp, va, 0) == -1) {
			pte_val(old_pte) &= ~_PAGE_HPTEFLAGS;
		} else {
			if (!pte_same(old_pte, new_pte)) {
				*ptep = new_pte;
			}
		}
	}

	if (likely(!(pte_val(old_pte) & _PAGE_HASHPTE))) {
		/* Update the linux pte with the HPTE slot */
		pte_val(new_pte) &= ~_PAGE_HPTEFLAGS;
		pte_val(new_pte) |= _PAGE_HASHPTE;
		prpn = pte_val(old_pte) >> PTE_SHIFT;

		/* copy appropriate flags from linux pte */
		hpteflags = (pte_val(new_pte) & 0x1f8) | newpp;

		slot = ppc_md.hpte_insert(vpn, prpn, hpteflags, 0, 0);

		pte_val(new_pte) |= ((slot<<12) & 
				     (_PAGE_GROUP_IX | _PAGE_SECONDARY));

		*ptep = new_pte;
	}

	spin_unlock(&hash_table_lock[lock_slot].lock);

	return 0;
}

/*
 * Handle a fault by adding an HPTE. If the address can't be determined
 * to be valid via Linux page tables, return 1. If handled return 0
 */
int hash_page(unsigned long ea, unsigned long access)
{
	void *pgdir;
	unsigned long vsid;
	struct mm_struct *mm;
	pte_t *ptep;
	int ret;

	/* Check for invalid addresses. */
	if (!IS_VALID_EA(ea)) return 1;

 	switch (REGION_ID(ea)) {
	case USER_REGION_ID:
		mm = current->mm;
		if (mm == NULL) return 1;
		vsid = get_vsid(mm->context, ea);
		break;
	case IO_REGION_ID:
		mm = &ioremap_mm;
		vsid = get_kernel_vsid(ea);
		break;
	case VMALLOC_REGION_ID:
		mm = &init_mm;
		vsid = get_kernel_vsid(ea);
#ifdef CONFIG_SHARED_MEMORY_ADDRESSING
                /*
                 * Check if this is a user task with shared access to kernel
                 * data & we got a protection fault.  If it is, the kernel
                 * must have faulted in the segment and the protection flags
                 * on the segment are kernel access only.  Just flush the
                 * segment table & fault in the segment with the right flags.
                 */
                if(((current->thread.flags) & PPC_FLAG_SHARED) &&
                   (access & _PAGE_USER)) {
                        flush_stab();
                }
#endif
		break;
	case EEH_REGION_ID:
		/*
		 * Should only be hit if there is an access to MMIO space
		 * which is protected by EEH.
		 * Send the problem up to do_page_fault 
		 */
	case KERNEL_REGION_ID:
		/*
		 * Should never get here - entire 0xC0... region is bolted.
		 * Send the problem up to do_page_fault 
		 */
	default:
		/* Not a valid range
		 * Send the problem up to do_page_fault 
		 */
		return 1;
		break;
	}

	pgdir = mm->pgd;
	if (pgdir == NULL) return 1;

	/*
	 * Lock the Linux page table to prevent mmap and kswapd
	 * from modifying entries while we search and update
	 */
	spin_lock(&mm->page_table_lock);

	ptep = find_linux_pte(pgdir, ea);
	/*
	 * If no pte found or not present, send the problem up to
	 * do_page_fault
	 */
	if (ptep && pte_present(*ptep)) {
		ret = __hash_page(ea, access, vsid, ptep);
	} else {	
		/* If no pte, send the problem up to do_page_fault */
		ret = 1;
	}

	spin_unlock(&mm->page_table_lock);

	return ret;
}

void flush_hash_page(unsigned long context, unsigned long ea, pte_t *ptep)
{
	unsigned long vsid, vpn, va, hash, secondary, slot, flags, lock_slot;
	unsigned long large = 0, local = 0;
	pte_t pte;

	if ((ea >= USER_START) && (ea <= USER_END))
		vsid = get_vsid(context, ea);
	else
		vsid = get_kernel_vsid(ea);

	va = (vsid << 28) | (ea & 0x0fffffff);
	if (large)
		vpn = va >> LARGE_PAGE_SHIFT;
	else
		vpn = va >> PAGE_SHIFT;

	lock_slot = get_lock_slot(vpn); 
	hash = hpt_hash(vpn, large);

	spin_lock_irqsave(&hash_table_lock[lock_slot].lock, flags);

	pte = __pte(pte_update(ptep, _PAGE_HPTEFLAGS, 0));
	secondary = (pte_val(pte) & _PAGE_SECONDARY) >> 15;
	if (secondary) hash = ~hash;
	slot = (hash & htab_data.htab_hash_mask) * HPTES_PER_GROUP;
	slot += (pte_val(pte) & _PAGE_GROUP_IX) >> 12;
	
	if (pte_val(pte) & _PAGE_HASHPTE) {
		ppc_md.hpte_invalidate(slot, secondary, va, large, local);
	}

	spin_unlock_irqrestore(&hash_table_lock[lock_slot].lock, flags);
}

long plpar_pte_enter(unsigned long flags,
		     unsigned long ptex,
		     unsigned long new_pteh, unsigned long new_ptel,
		     unsigned long *old_pteh_ret, unsigned long *old_ptel_ret)
{
	unsigned long dummy, ret;
	ret = plpar_hcall(H_ENTER, flags, ptex, new_pteh, new_ptel,
			   old_pteh_ret, old_ptel_ret, &dummy);
	return(ret);
}

long plpar_pte_remove(unsigned long flags,
		      unsigned long ptex,
		      unsigned long avpn,
		      unsigned long *old_pteh_ret, unsigned long *old_ptel_ret)
{
	unsigned long dummy;
	return plpar_hcall(H_REMOVE, flags, ptex, avpn, 0,
			   old_pteh_ret, old_ptel_ret, &dummy);
}

long plpar_pte_read(unsigned long flags,
		    unsigned long ptex,
		    unsigned long *old_pteh_ret, unsigned long *old_ptel_ret)
{
	unsigned long dummy;
	return plpar_hcall(H_READ, flags, ptex, 0, 0,
			   old_pteh_ret, old_ptel_ret, &dummy);
}

long plpar_pte_protect(unsigned long flags,
		       unsigned long ptex,
		       unsigned long avpn)
{
	return plpar_hcall_norets(H_PROTECT, flags, ptex, avpn);
}

static __inline__ void set_pp_bit(unsigned long pp, HPTE *addr)
{
	unsigned long old;
	unsigned long *p = &addr->dw1.dword1;

	__asm__ __volatile__(
        "1:	ldarx	%0,0,%3\n\
                rldimi  %0,%2,0,62\n\
                stdcx.	%0,0,%3\n\
            	bne	1b"
        : "=&r" (old), "=m" (*p)
        : "r" (pp), "r" (p), "m" (*p)
        : "cc");
}

/*
 * Calculate which hash_table_lock to use, based on the pteg being used.
 *
 * Given a VPN, use the high order 8 bits to select one of 2^7 locks.  The
 * highest order bit is used to indicate primary vs. secondary group.  If the
 * secondary is selected, complement the lock select bits.  This results in
 * both the primary and secondary groups being covered under the same lock.
 */
inline unsigned long get_lock_slot(unsigned long vpn)
{
	unsigned long lock_slot;
#ifdef LOCK_SPLIT
	lock_slot = (hpt_hash(vpn,0) >> htab_data.htab_lock_shift) & 0xff;
	if(lock_slot & 0x80) lock_slot = (~lock_slot) & 0x7f;
#else
	lock_slot = 0;
#endif
	return(lock_slot);
}

/*
 * Functions used to retrieve word 0 of a given page table entry.
 *
 * Input : slot : PTE index within the page table of the entry to retrieve 
 * Output: Contents of word 0 of the specified entry
 */
static unsigned long rpa_lpar_hpte_getword0(unsigned long slot)
{
	unsigned long dword0;
	unsigned long lpar_rc;
	unsigned long dummy_word1;
	unsigned long flags;

	/* Read 1 pte at a time                        */
	/* Do not need RPN to logical page translation */
	/* No cross CEC PFT access                     */
	flags = 0;
	
	lpar_rc = plpar_pte_read(flags, slot, &dword0, &dummy_word1);

	if (lpar_rc != H_Success)
		panic("Error on pte read in get_hpte0 rc = %lx\n", lpar_rc);

	return dword0;
}

unsigned long iSeries_hpte_getword0(unsigned long slot)
{
	unsigned long dword0;

	HPTE hpte;
	HvCallHpt_get(&hpte, slot);
	dword0 = hpte.dw0.dword0;

	return dword0;
}

/*
 * Functions used to find the PTE for a particular virtual address. 
 * Only used during boot when bolting pages.
 *
 * Input : vpn      : virtual page number
 * Output: PTE index within the page table of the entry
 *         -1 on failure
 */
static long hpte_find(unsigned long vpn)
{
	HPTE *hptep;
	unsigned long hash;
	unsigned long i, j;
	long slot;
	Hpte_dword0 dw0;

	hash = hpt_hash(vpn, 0);

	for (j = 0; j < 2; j++) {
		slot = (hash & htab_data.htab_hash_mask) * HPTES_PER_GROUP;
		for (i = 0; i < HPTES_PER_GROUP; i++) {
			hptep = htab_data.htab + slot;
			dw0 = hptep->dw0.dw0;

			if ((dw0.avpn == (vpn >> 11)) && dw0.v &&
			    (dw0.h == j)) {
				/* HPTE matches */
				if (j)
					slot = -slot;
				return slot;
			}
			++slot;
		}
		hash = ~hash;
	}

	return -1;
}

static long rpa_lpar_hpte_find(unsigned long vpn)
{
	unsigned long hash;
	unsigned long i, j;
	long slot;
	union {
		unsigned long dword0;
		Hpte_dword0 dw0;
	} hpte_dw0;
	Hpte_dword0 dw0;

	hash = hpt_hash(vpn, 0);

	for (j = 0; j < 2; j++) {
		slot = (hash & htab_data.htab_hash_mask) * HPTES_PER_GROUP;
		for (i = 0; i < HPTES_PER_GROUP; i++) {
			hpte_dw0.dword0 = rpa_lpar_hpte_getword0(slot);
			dw0 = hpte_dw0.dw0;

			if ((dw0.avpn == (vpn >> 11)) && dw0.v &&
			    (dw0.h == j)) {
				/* HPTE matches */
				if (j)
					slot = -slot;
				return slot;
			}
			++slot;
		}
		hash = ~hash;
	}

	return -1;
} 

static long iSeries_hpte_find(unsigned long vpn)
{
	HPTE hpte;
	long slot;

	/*
	 * The HvCallHpt_findValid interface is as follows:
	 * 0xffffffffffffffff : No entry found.
	 * 0x00000000xxxxxxxx : Entry found in primary group, slot x
	 * 0x80000000xxxxxxxx : Entry found in secondary group, slot x
	 */
	slot = HvCallHpt_findValid(&hpte, vpn); 
	if (hpte.dw0.dw0.v) {
		if (slot < 0) {
			slot &= 0x7fffffffffffffff;
			slot = -slot;
		}
	} else {
		slot = -1;
	}

	return slot;
}

/*
 * Functions used to invalidate a page table entry from the page table
 * and tlb.
 *
 * Input : slot  : PTE index within the page table of the entry to invalidated
 *         va    : Virtual address of the entry being invalidated
 *         large : 1 = large page (16M)
 *         local : 1 = Use tlbiel to only invalidate the local tlb 
 */
static void hpte_invalidate(unsigned long slot, 
			    unsigned long secondary,
			    unsigned long va,
			    int large, int local)
{
	HPTE *hptep = htab_data.htab + slot;
	Hpte_dword0 dw0;
	unsigned long vpn, avpn;
	unsigned long flags;

	if (large)
		vpn = va >> LARGE_PAGE_SHIFT;
	else
		vpn = va >> PAGE_SHIFT;

	avpn = vpn >> 11;

	dw0 = hptep->dw0.dw0;

	/*
	 * Do not remove bolted entries.  Alternatively, we could check
	 * the AVPN, hash group, and valid bits.  By doing it this way,
	 * it is common with the pSeries LPAR optimal path.
	 */
	if (dw0.bolted) return;

	/* Invalidate the hpte. */
	hptep->dw0.dword0 = 0;

	/* Invalidate the tlb */
	spin_lock_irqsave(&pSeries_tlbie_lock, flags);
	_tlbie(va, large);
	spin_unlock_irqrestore(&pSeries_tlbie_lock, flags);
}

static void rpa_lpar_hpte_invalidate(unsigned long slot, 
				     unsigned long secondary,
				     unsigned long va,
				     int large, int local)
{
	unsigned long lpar_rc;
	unsigned long dummy1, dummy2;

	/* 
	 * Don't remove a bolted entry.  This case can occur when we bolt
	 * pages dynamically after initial boot.
	 */
	lpar_rc = plpar_pte_remove(H_ANDCOND, slot, (0x1UL << 4), 
				   &dummy1, &dummy2);

	if (lpar_rc != H_Success)
		panic("Bad return code from invalidate rc = %lx\n", lpar_rc);
}

static void iSeries_hpte_invalidate(unsigned long slot, 
				    unsigned long secondary,
				    unsigned long va,
				    int large, int local)
{
	HPTE lhpte;
	unsigned long vpn, avpn;

	if (large)
		vpn = va >> LARGE_PAGE_SHIFT;
	else
		vpn = va >> PAGE_SHIFT;

	avpn = vpn >> 11;

	lhpte.dw0.dword0 = iSeries_hpte_getword0(slot);
	
	if ((lhpte.dw0.dw0.avpn == avpn) && 
	    (lhpte.dw0.dw0.v) &&
	    (lhpte.dw0.dw0.h == secondary)) {
		HvCallHpt_invalidateSetSwBitsGet(slot, 0, 0);
	}
}

/*
 * Functions used to update page protection bits.
 *
 * Input : slot  : PTE index within the page table of the entry to update
 *         newpp : new page protection bits
 *         va    : Virtual address of the entry being updated
 *         large : 1 = large page (16M)
 * Output: 0 on success, -1 on failure
 */
static long hpte_updatepp(unsigned long slot, 
			  unsigned long secondary,
			  unsigned long newpp,
			  unsigned long va, int large)
{
	HPTE *hptep = htab_data.htab + slot;
	Hpte_dword0 dw0;
	Hpte_dword1 dw1;
	unsigned long vpn, avpn;
	unsigned long flags;

	if (large)
		vpn = va >> LARGE_PAGE_SHIFT;
	else
		vpn = va >> PAGE_SHIFT;

	avpn = vpn >> 11;

	dw0 = hptep->dw0.dw0;
	if ((dw0.avpn == avpn) && 
	    (dw0.v) && (dw0.h == secondary)) {
		/* Turn off valid bit in HPTE */
		dw0.v = 0;
		hptep->dw0.dw0 = dw0;
		
		/* Ensure it is out of the tlb too */
		spin_lock_irqsave(&pSeries_tlbie_lock, flags);
		_tlbie(va, large);
		spin_unlock_irqrestore(&pSeries_tlbie_lock, flags);
		
		/* Insert the new pp bits into the HPTE */
		dw1 = hptep->dw1.dw1;
		dw1.pp = newpp;
		hptep->dw1.dw1 = dw1;
		
		/* Ensure it is visible before validating */
		__asm__ __volatile__ ("eieio" : : : "memory");
		
		/* Turn the valid bit back on in HPTE */
		dw0.v = 1;
		hptep->dw0.dw0 = dw0;
		
		__asm__ __volatile__ ("ptesync" : : : "memory");
		
		return 0;
	}

	return -1;
}

static long rpa_lpar_hpte_updatepp(unsigned long slot, 
				   unsigned long secondary,
				   unsigned long newpp,
				   unsigned long va, int large)
{
	unsigned long lpar_rc;
	unsigned long flags = (newpp & 7);
	unsigned long avpn = va >> 23;
	HPTE hpte;

	lpar_rc = plpar_pte_read(0, slot, &hpte.dw0.dword0, &hpte.dw1.dword1);

	if ((hpte.dw0.dw0.avpn == avpn) &&
	    (hpte.dw0.dw0.v) && 
	    (hpte.dw0.dw0.h == secondary)) {
		lpar_rc = plpar_pte_protect(flags, slot, 0);
		if (lpar_rc != H_Success)
			panic("bad return code from pte protect rc = %lx\n", 
			      lpar_rc);
		return 0;
	}

	return -1;
}

static long iSeries_hpte_updatepp(unsigned long slot, 
				  unsigned long secondary,
				  unsigned long newpp, 
				  unsigned long va, int large)
{
	unsigned long vpn, avpn;
	HPTE hpte;

	if (large)
		vpn = va >> LARGE_PAGE_SHIFT;
	else
		vpn = va >> PAGE_SHIFT;

	avpn = vpn >> 11;

	HvCallHpt_get(&hpte, slot);
	if ((hpte.dw0.dw0.avpn == avpn) && 
	    (hpte.dw0.dw0.v) &&
	    (hpte.dw0.dw0.h == secondary)) {
		HvCallHpt_setPp(slot, newpp);
		return 0;
	}
	return -1;
}

/*
 * Functions used to update the page protection bits. Intended to be used 
 * to create guard pages for kernel data structures on pages which are bolted
 * in the HPT. Assumes pages being operated on will not be stolen.
 * Does not work on large pages. No need to lock here because we are the 
 * only user.
 * 
 * Input : newpp : page protection flags
 *         ea    : effective kernel address to bolt.
 */
static void hpte_updateboltedpp(unsigned long newpp, unsigned long ea)
{
	unsigned long vsid, va, vpn, flags;
	long slot;
	HPTE *hptep;

	vsid = get_kernel_vsid(ea);
	va = (vsid << 28) | (ea & 0x0fffffff);
	vpn = va >> PAGE_SHIFT;

	slot = hpte_find(vpn);
	if (slot == -1)
		panic("could not find page to bolt\n");
	hptep = htab_data.htab + slot;

	set_pp_bit(newpp, hptep);

	/* Ensure it is out of the tlb too */
	spin_lock_irqsave(&pSeries_tlbie_lock, flags);
	_tlbie(va, 0);
	spin_unlock_irqrestore(&pSeries_tlbie_lock, flags);
}

static void rpa_lpar_hpte_updateboltedpp(unsigned long newpp, unsigned long ea)
{
	unsigned long lpar_rc;
	unsigned long vsid, va, vpn, flags;
	long slot;

	vsid = get_kernel_vsid(ea);
	va = (vsid << 28) | (ea & 0x0fffffff);
	vpn = va >> PAGE_SHIFT;

	slot = rpa_lpar_hpte_find(vpn);
	if (slot == -1)
		panic("updateboltedpp: Could not find page to bolt\n");

	flags = newpp & 3;
	lpar_rc = plpar_pte_protect(flags, slot, 0);

	if (lpar_rc != H_Success)
		panic("Bad return code from pte bolted protect rc = %lx\n",
		      lpar_rc); 
}

void iSeries_hpte_updateboltedpp(unsigned long newpp, unsigned long ea)
{
	unsigned long vsid,va,vpn;
	long slot;

	vsid = get_kernel_vsid( ea );
	va = ( vsid << 28 ) | ( ea & 0x0fffffff );
	vpn = va >> PAGE_SHIFT;

	slot = iSeries_hpte_find(vpn); 
	if (slot == -1)
		panic("updateboltedpp: Could not find page to bolt\n");

	HvCallHpt_setPp(slot, newpp);
}

/*
 * Functions used to insert new hardware page table entries.
 * Will castout non-bolted entries as necessary using a random
 * algorithm.
 *
 * Input : vpn      : virtual page number
 *         prpn     : real page number in absolute space
 *         hpteflags: page protection flags
 *         bolted   : 1 = bolt the page
 *         large    : 1 = large page (16M)
 * Output: hsss, where h = hash group, sss = slot within that group
 */
static long hpte_insert(unsigned long vpn, unsigned long prpn,
			unsigned long hpteflags, int bolted, int large)
{
	HPTE *hptep;
	Hpte_dword0 dw0;
	HPTE lhpte;
	int i, secondary;
	unsigned long hash = hpt_hash(vpn, 0);
	unsigned long avpn = vpn >> 11;
	unsigned long arpn = physRpn_to_absRpn(prpn);
	unsigned long hpte_group;

repeat:
	secondary = 0;
	hpte_group = ((hash & htab_data.htab_hash_mask) *
		      HPTES_PER_GROUP) & ~0x7UL;
	hptep = htab_data.htab + hpte_group;

	for (i = 0; i < HPTES_PER_GROUP; i++) {
		dw0 = hptep->dw0.dw0;
		if (!dw0.v) {
			/* retry with lock held */
			dw0 = hptep->dw0.dw0;
			if (!dw0.v)
				break;
		}
		hptep++;
	}

	if (i == HPTES_PER_GROUP) {
		secondary = 1;
		hpte_group = ((~hash & htab_data.htab_hash_mask) *
			      HPTES_PER_GROUP) & ~0x7UL;
		hptep = htab_data.htab + hpte_group;

		for (i = 0; i < HPTES_PER_GROUP; i++) {
			dw0 = hptep->dw0.dw0;
			if (!dw0.v) {
				/* retry with lock held */
				dw0 = hptep->dw0.dw0;
				if (!dw0.v)
					break;
			}
			hptep++;
		}
		if (i == HPTES_PER_GROUP) {
			if (mftb() & 0x1)
				hpte_group=((hash & htab_data.htab_hash_mask)* 
					    HPTES_PER_GROUP) & ~0x7UL;
			
			hpte_remove(hpte_group);
			goto repeat;
		}
	}

	lhpte.dw1.dword1      = 0;
	lhpte.dw1.dw1.rpn     = arpn;
	lhpte.dw1.flags.flags = hpteflags;

	lhpte.dw0.dword0      = 0;
	lhpte.dw0.dw0.avpn    = avpn;
	lhpte.dw0.dw0.h       = secondary;
	lhpte.dw0.dw0.bolted  = bolted;
	lhpte.dw0.dw0.v       = 1;

	if (large) lhpte.dw0.dw0.l = 1;

	hptep->dw1.dword1 = lhpte.dw1.dword1;

	/* Guarantee the second dword is visible before the valid bit */
	__asm__ __volatile__ ("eieio" : : : "memory");

	/*
	 * Now set the first dword including the valid bit
	 * NOTE: this also unlocks the hpte
	 */
	hptep->dw0.dword0 = lhpte.dw0.dword0;

	__asm__ __volatile__ ("ptesync" : : : "memory");

	return ((secondary << 3) | i);
}

static long rpa_lpar_hpte_insert(unsigned long vpn, unsigned long prpn,
				 unsigned long hpteflags,
				 int bolted, int large)
{
	/* XXX fix for large page */
	unsigned long lpar_rc;
	unsigned long flags;
	unsigned long slot;
	HPTE lhpte;
	int secondary;
	unsigned long hash = hpt_hash(vpn, 0);
	unsigned long avpn = vpn >> 11;
	unsigned long arpn = physRpn_to_absRpn(prpn);
	unsigned long hpte_group;

	/* Fill in the local HPTE with absolute rpn, avpn and flags */
	lhpte.dw1.dword1      = 0;
	lhpte.dw1.dw1.rpn     = arpn;
	lhpte.dw1.flags.flags = hpteflags;

	lhpte.dw0.dword0      = 0;
	lhpte.dw0.dw0.avpn    = avpn;
	lhpte.dw0.dw0.bolted  = bolted;
	lhpte.dw0.dw0.v       = 1;

	if (large) lhpte.dw0.dw0.l = 1;

	/* Now fill in the actual HPTE */
	/* Set CEC cookie to 0         */
	/* Large page = 0              */
	/* Zero page = 0               */
	/* I-cache Invalidate = 0      */
	/* I-cache synchronize = 0     */
	/* Exact = 0                   */
	flags = 0;

	/* XXX why is this here? - Anton */
	/*   -- Because at one point we hit a case where non cachable
	 *      pages where marked coherent & this is rejected by the HV.
	 *      Perhaps it is no longer an issue ... DRENG.
	 */ 
	if (hpteflags & (_PAGE_GUARDED|_PAGE_NO_CACHE))
		lhpte.dw1.flags.flags &= ~_PAGE_COHERENT;

repeat:
	secondary = 0;
	lhpte.dw0.dw0.h = secondary;
	hpte_group = ((hash & htab_data.htab_hash_mask) *
		      HPTES_PER_GROUP) & ~0x7UL;

	__asm__ __volatile__ (
		H_ENTER_r3
		"mr    4, %2\n"
                "mr    5, %3\n"
                "mr    6, %4\n"
                "mr    7, %5\n"
		HVSC
                "mr    %0, 3\n"
                "mr    %1, 4\n"
		: "=r" (lpar_rc), "=r" (slot)
		: "r" (flags), "r" (hpte_group), "r" (lhpte.dw0.dword0),
		"r" (lhpte.dw1.dword1)
		: "r0", "r3", "r4", "r5", "r6", "r7", 
		  "r8", "r9", "r10", "r11", "r12", "cc");

	if (lpar_rc == H_PTEG_Full) {
		secondary = 1;
		lhpte.dw0.dw0.h = secondary;
		hpte_group = ((~hash & htab_data.htab_hash_mask) *
			      HPTES_PER_GROUP) & ~0x7UL;

		__asm__ __volatile__ (
			      H_ENTER_r3
			      "mr    4, %2\n"
			      "mr    5, %3\n"
			      "mr    6, %4\n"
			      "mr    7, %5\n"
			      HVSC
			      "mr    %0, 3\n"
			      "mr    %1, 4\n"
			      : "=r" (lpar_rc), "=r" (slot)
			      : "r" (flags), "r" (hpte_group), "r" (lhpte.dw0.dword0),
			      "r" (lhpte.dw1.dword1)
			      : "r0", "r3", "r4", "r5", "r6", "r7",
			        "r8", "r9", "r10", "r11", "r12", "cc");
		if (lpar_rc == H_PTEG_Full) {
			if (mftb() & 0x1)
				hpte_group=((hash & htab_data.htab_hash_mask)* 
					    HPTES_PER_GROUP) & ~0x7UL;
			
			rpa_lpar_hpte_remove(hpte_group);
			goto repeat;
		}
	}

	if (lpar_rc != H_Success)
		panic("Bad return code from pte enter rc = %lx\n", lpar_rc);

	return ((secondary << 3) | (slot & 0x7));
}

static long iSeries_hpte_insert(unsigned long vpn, unsigned long prpn,
				unsigned long hpteflags,
				int bolted, int large)
{
	HPTE lhpte;
	unsigned long hash, hpte_group;
	unsigned long avpn = vpn >> 11;
	unsigned long arpn = physRpn_to_absRpn( prpn );
	int secondary = 0;
	long slot;

	hash = hpt_hash(vpn, 0);

repeat:
	slot = HvCallHpt_findValid(&lhpte, vpn);
	if (lhpte.dw0.dw0.v) {
		panic("select_hpte_slot found entry already valid\n");
	}

	if (slot == -1) { /* No available entry found in either group */
		if (mftb() & 0x1) {
			hpte_group=((hash & htab_data.htab_hash_mask)* 
				    HPTES_PER_GROUP) & ~0x7UL;
		} else {
			hpte_group=((~hash & htab_data.htab_hash_mask)* 
				    HPTES_PER_GROUP) & ~0x7UL;
		}

		hash = hpt_hash(vpn, 0);
		iSeries_hpte_remove(hpte_group);
		goto repeat;
	} else if (slot < 0) {
		slot &= 0x7fffffffffffffff;
		secondary = 1;
	}

	/* Create the HPTE */
	lhpte.dw1.dword1      = 0;
	lhpte.dw1.dw1.rpn     = arpn;
	lhpte.dw1.flags.flags = hpteflags;

	lhpte.dw0.dword0     = 0;
	lhpte.dw0.dw0.avpn   = avpn;
	lhpte.dw0.dw0.h      = secondary;
	lhpte.dw0.dw0.bolted = bolted;
	lhpte.dw0.dw0.v      = 1;

	/* Now fill in the actual HPTE */
	HvCallHpt_addValidate(slot, secondary, (HPTE *)&lhpte);
	return ((secondary << 3) | (slot & 0x7));
}

/*
 * Functions used to remove hardware page table entries.
 *
 * Input : hpte_group: PTE index of the first entry in a group
 * Output: offset within the group of the entry removed or
 *         -1 on failure
 */
static long hpte_remove(unsigned long hpte_group)
{
	HPTE *hptep;
	Hpte_dword0 dw0;
	int i;
	int slot_offset;
	unsigned long vsid, group, pi, pi_high;
	unsigned long slot;
	unsigned long flags;
	int large;
	unsigned long va;

	/* pick a random slot to start at */
	slot_offset = mftb() & 0x7;

	for (i = 0; i < HPTES_PER_GROUP; i++) {
		hptep = htab_data.htab + hpte_group + slot_offset;
		dw0 = hptep->dw0.dw0;

		if (dw0.v && !dw0.bolted) {
			/* retry with lock held */
			dw0 = hptep->dw0.dw0;
			if (dw0.v && !dw0.bolted)
				break;
		}

		slot_offset++;
		slot_offset &= 0x7;
	}

	if (i == HPTES_PER_GROUP)
		return -1;

	large = dw0.l;

	/* Invalidate the hpte. NOTE: this also unlocks it */
	hptep->dw0.dword0 = 0;

	/* Invalidate the tlb */
	vsid = dw0.avpn >> 5;
	slot = hptep - htab_data.htab;
	group = slot >> 3;
	if (dw0.h)
		group = ~group;
	pi = (vsid ^ group) & 0x7ff;
	pi_high = (dw0.avpn & 0x1f) << 11;
	pi |= pi_high;

	if (large)
		va = pi << LARGE_PAGE_SHIFT;
	else
		va = pi << PAGE_SHIFT;

	spin_lock_irqsave(&pSeries_tlbie_lock, flags);
	_tlbie(va, large);
	spin_unlock_irqrestore(&pSeries_tlbie_lock, flags);

	return i;
}

static long rpa_lpar_hpte_remove(unsigned long hpte_group)
{
	unsigned long slot_offset;
	unsigned long lpar_rc;
	int i;
	unsigned long dummy1, dummy2;

	/* pick a random slot to start at */
	slot_offset = mftb() & 0x7;

	for (i = 0; i < HPTES_PER_GROUP; i++) {

		/* Don't remove a bolted entry */
		lpar_rc = plpar_pte_remove(H_ANDCOND, hpte_group + slot_offset,
					   (0x1UL << 4), &dummy1, &dummy2);

		if (lpar_rc == H_Success)
			return i;

		if (lpar_rc != H_Not_Found)
			panic("Bad return code from pte remove rc = %lx\n",
			      lpar_rc);

		slot_offset++;
		slot_offset &= 0x7;
	}

	return -1;
}

static long iSeries_hpte_remove(unsigned long hpte_group)
{
	unsigned long slot_offset;
	int i;
	HPTE lhpte;

	/* Pick a random slot to start at */
	slot_offset = mftb() & 0x7;

	for (i = 0; i < HPTES_PER_GROUP; i++) {
		lhpte.dw0.dword0 = 
			iSeries_hpte_getword0(hpte_group + slot_offset);

		if (!lhpte.dw0.dw0.bolted) {
			HvCallHpt_invalidateSetSwBitsGet(hpte_group + 
							 slot_offset, 0, 0);
			return i;
		}

		slot_offset++;
		slot_offset &= 0x7;
	}

	return -1;
}

void hpte_init_pSeries(void)
{
	ppc_md.hpte_invalidate     = hpte_invalidate;
	ppc_md.hpte_updatepp       = hpte_updatepp;
	ppc_md.hpte_updateboltedpp = hpte_updateboltedpp;
	ppc_md.hpte_insert	   = hpte_insert;
	ppc_md.hpte_remove	   = hpte_remove;
}

void pSeries_lpar_mm_init(void)
{
	ppc_md.hpte_invalidate     = rpa_lpar_hpte_invalidate;
	ppc_md.hpte_updatepp       = rpa_lpar_hpte_updatepp;
	ppc_md.hpte_updateboltedpp = rpa_lpar_hpte_updateboltedpp;
	ppc_md.hpte_insert         = rpa_lpar_hpte_insert;
	ppc_md.hpte_remove         = rpa_lpar_hpte_remove;
}

void hpte_init_iSeries(void)
{
	ppc_md.hpte_invalidate     = iSeries_hpte_invalidate;
	ppc_md.hpte_updatepp       = iSeries_hpte_updatepp;
	ppc_md.hpte_updateboltedpp = iSeries_hpte_updateboltedpp;
	ppc_md.hpte_insert         = iSeries_hpte_insert;
	ppc_md.hpte_remove         = iSeries_hpte_remove;
}