1effb490f2
used. leave the vax_insque function for no reason except that it maps to an actual instruction.
414 lines
10 KiB
C
414 lines
10 KiB
C
/* $NetBSD: macros.h,v 1.44 2007/11/22 23:02:05 plunky Exp $ */
|
|
|
|
/*
|
|
* Copyright (c) 1994, 1998, 2000 Ludd, University of Lule}, Sweden.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
* must display the following acknowledgement:
|
|
* This product includes software developed at Ludd, University of Lule}.
|
|
* 4. The name of the author may not be used to endorse or promote products
|
|
* derived from this software without specific prior written permission
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
|
|
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
|
|
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
|
|
* IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
|
|
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
|
|
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
|
|
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
/* All bugs are subject to removal without further notice */
|
|
|
|
#if !defined(_VAX_MACROS_H_) && !defined(__lint__)
|
|
#define _VAX_MACROS_H_
|
|
|
|
void __blkset(void *, int, size_t);
|
|
void __blkcpy(const void *, void *, size_t);
|
|
|
|
#if !__GNUC_PREREQ__(4, 1)
|
|
/* Here general macros are supposed to be stored */
|
|
|
|
static __inline int __attribute__((__unused__))
|
|
vax_ffs(int reg)
|
|
{
|
|
register int val;
|
|
|
|
__asm volatile ("ffs $0,$32,%1,%0;"
|
|
"bneq 1f;"
|
|
"mnegl $1,%0;"
|
|
"1:;"
|
|
"incl %0"
|
|
: "=&r" (val)
|
|
: "r" (reg) );
|
|
return val;
|
|
}
|
|
#define ffs vax_ffs
|
|
#endif
|
|
|
|
static __inline void __attribute__((__unused__))
|
|
vax_remque(void *p)
|
|
{
|
|
__asm volatile ("remque (%0),%0;clrl 4(%0)"
|
|
:
|
|
: "r" (p)
|
|
: "memory" );
|
|
}
|
|
|
|
static __inline void __attribute__((__unused__))
|
|
vax_insque(void *p, void *q)
|
|
{
|
|
__asm volatile ("insque (%0),(%1)"
|
|
:
|
|
: "r" (p),"r" (q)
|
|
: "memory" );
|
|
}
|
|
|
|
#if 0
|
|
static __inline void *__attribute__((__unused__))
|
|
vax_memcpy(void *to, const void *from, size_t len)
|
|
{
|
|
if (len > 65535) {
|
|
__blkcpy(from, to, len);
|
|
} else {
|
|
__asm volatile ("movc3 %1,(%2),%0"
|
|
: "=m" (*(char *)to)
|
|
: "g" (len), "r" (*(const char *)from)
|
|
:"r0","r1","r2","r3","r4","r5","memory","cc");
|
|
}
|
|
return to;
|
|
}
|
|
#define memcpy vax_memcpy
|
|
|
|
static __inline void *__attribute__((__unused__))
|
|
vax_memmove(void *to, const void *from, size_t len)
|
|
{
|
|
if (len > 65535) {
|
|
__blkcpy(from, to, len);
|
|
} else {
|
|
__asm __volatile ("movc3 %1,%2,%0"
|
|
: "=m" (*(char *)to)
|
|
: "g" (len), "mo" (*(const char *)from)
|
|
:"r0","r1","r2","r3","r4","r5","memory","cc");
|
|
}
|
|
return to;
|
|
}
|
|
#define memmove vax_memmove
|
|
#endif
|
|
|
|
static __inline void *__attribute__((__unused__))
|
|
vax_memset(void *block, int c, size_t len)
|
|
{
|
|
if (len > 65535) {
|
|
__blkset(block, c, len);
|
|
} else {
|
|
__asm __volatile ("movc5 $0,(%%sp),%2,%1,%0"
|
|
: "=m" (*(char *)block)
|
|
: "g" (len), "g" (c)
|
|
:"r0","r1","r2","r3","r4","r5","memory","cc");
|
|
}
|
|
return block;
|
|
}
|
|
#define memset vax_memset
|
|
|
|
#ifdef notdef
|
|
/* XXX - the return syntax of memcmp is wrong */
|
|
static __inline int __attribute__((__unused__))
|
|
memcmp(const void *b1, const void *b2, size_t len)
|
|
{
|
|
register int ret;
|
|
|
|
__asm volatile("cmpc3 %3,(%1),(%2);"
|
|
"movl %%r0,%0"
|
|
: "=r" (ret)
|
|
: "r" (b1), "r" (b2), "r" (len)
|
|
: "r0","r1","r2","r3" );
|
|
return ret;
|
|
}
|
|
|
|
static __inline int __attribute__((__unused__))
|
|
bcmp(const void *b1, const void *b2, size_t len)
|
|
{
|
|
register int ret;
|
|
|
|
__asm volatile("cmpc3 %3,(%1),(%2);"
|
|
"movl %%r0,%0"
|
|
: "=r" (ret)
|
|
: "r" (b1), "r" (b2), "r" (len)
|
|
: "r0","r1","r2","r3" );
|
|
return ret;
|
|
}
|
|
|
|
/* Begin nya */
|
|
static __inline size_t __attribute__((__unused__))
|
|
strlen(const char *cp)
|
|
{
|
|
register size_t ret;
|
|
|
|
__asm volatile("locc $0,$65535,(%1);"
|
|
"subl3 %%r0,$65535,%0"
|
|
: "=r" (ret)
|
|
: "r" (cp)
|
|
: "r0","r1","cc" );
|
|
return ret;
|
|
}
|
|
|
|
static __inline char * __attribute__((__unused__))
|
|
strcat(char *cp, const char *c2)
|
|
{
|
|
__asm volatile("locc $0,$65535,(%1);"
|
|
"subl3 %%r0,$65535,%%r2;"
|
|
"incl %%r2;"
|
|
"locc $0,$65535,(%0);"
|
|
"movc3 %%r2,(%1),(%%r1)"
|
|
:
|
|
: "r" (cp), "r" (c2)
|
|
: "r0","r1","r2","r3","r4","r5","memory","cc");
|
|
return cp;
|
|
}
|
|
|
|
static __inline char * __attribute__((__unused__))
|
|
strncat(char *cp, const char *c2, size_t count)
|
|
{
|
|
__asm volatile("locc $0,%2,(%1);"
|
|
"subl3 %%r0,%2,%%r2;"
|
|
"locc $0,$65535,(%0);"
|
|
"movc3 %%r2,(%1),(%%r1);"
|
|
"movb $0,(%%r3)"
|
|
:
|
|
: "r" (cp), "r" (c2), "g"(count)
|
|
: "r0","r1","r2","r3","r4","r5","memory","cc");
|
|
return cp;
|
|
}
|
|
|
|
static __inline char * __attribute__((__unused__))
|
|
strcpy(char *cp, const char *c2)
|
|
{
|
|
__asm volatile("locc $0,$65535,(%1);"
|
|
"subl3 %%r0,$65535,%%r2;"
|
|
"movc3 %%r2,(%1),(%0);"
|
|
"movb $0,(%%r3)"
|
|
:
|
|
: "r" (cp), "r" (c2)
|
|
: "r0","r1","r2","r3","r4","r5","memory","cc");
|
|
return cp;
|
|
}
|
|
|
|
static __inline char * __attribute__((__unused__))
|
|
strncpy(char *cp, const char *c2, size_t len)
|
|
{
|
|
__asm volatile("movl %2,%%r2;"
|
|
"locc $0,%%r2,(%1);"
|
|
"beql 1f;"
|
|
"subl3 %%r0,%2,%%r2;"
|
|
"clrb (%0)[%%r2];"
|
|
"1:;"
|
|
"movc3 %%r2,(%1),(%0)"
|
|
:
|
|
: "r" (cp), "r" (c2), "g"(len)
|
|
: "r0","r1","r2","r3","r4","r5","memory","cc");
|
|
return cp;
|
|
}
|
|
|
|
static __inline void *__attribute__((__unused__))
|
|
memchr(const void *cp, int c, size_t len)
|
|
{
|
|
void *ret;
|
|
__asm volatile("locc %2,%3,(%1);"
|
|
"bneq 1f;"
|
|
"clrl %%r1;"
|
|
"1:;"
|
|
"movl %%r1,%0"
|
|
: "=g"(ret)
|
|
: "r" (cp), "r" (c), "g"(len)
|
|
: "r0","r1","cc");
|
|
return ret;
|
|
}
|
|
|
|
static __inline int __attribute__((__unused__))
|
|
strcmp(const char *cp, const char *c2)
|
|
{
|
|
register int ret;
|
|
__asm volatile("locc $0,$65535,(%1);"
|
|
"subl3 %%r0,$65535,%%r0;"
|
|
"incl %%r0;"
|
|
"cmpc3 %%r0,(%1),(%2);"
|
|
"beql 1f;"
|
|
"movl $1,%%r2;"
|
|
"cmpb (%%r1),(%%r3);"
|
|
"bcc 1f;"
|
|
"mnegl $1,%%r2;"
|
|
"1:;"
|
|
"movl %%r2,%0"
|
|
: "=g"(ret)
|
|
: "r" (cp), "r" (c2)
|
|
: "r0","r1","r2","r3","cc");
|
|
return ret;
|
|
}
|
|
#endif
|
|
|
|
#if 0 /* unused, but no point in deleting it since it _is_ an instruction */
|
|
static __inline int __attribute__((__unused__))
|
|
locc(int mask, char *cp, size_t size){
|
|
register ret;
|
|
|
|
__asm volatile("locc %1,%2,(%3);"
|
|
"movl %%r0,%0"
|
|
: "=r" (ret)
|
|
: "r" (mask),"r"(size),"r"(cp)
|
|
: "r0","r1" );
|
|
return ret;
|
|
}
|
|
#endif
|
|
|
|
static __inline int __attribute__((__unused__))
|
|
vax_scanc(u_int size, const u_char *cp, const u_char *table, int mask)
|
|
{
|
|
register int ret;
|
|
|
|
__asm volatile("scanc %1,(%2),(%3),%4;"
|
|
"movl %%r0,%0"
|
|
: "=g"(ret)
|
|
: "r"(size),"r"(cp),"r"(table),"r"(mask)
|
|
: "r0","r1","r2","r3" );
|
|
return ret;
|
|
}
|
|
#define scanc vax_scanc
|
|
|
|
static __inline int __attribute__((__unused__))
|
|
vax_skpc(int mask, size_t size, u_char *cp)
|
|
{
|
|
register int ret;
|
|
|
|
__asm volatile("skpc %1,%2,(%3);"
|
|
"movl %%r0,%0"
|
|
: "=g"(ret)
|
|
: "r"(mask),"r"(size),"r"(cp)
|
|
: "r0","r1" );
|
|
return ret;
|
|
}
|
|
#define skpc vax_skpc
|
|
|
|
/*
|
|
* Set/clear a bit at a memory position; interlocked.
|
|
* Return 0 if already set, 1 otherwise.
|
|
*/
|
|
static __inline int __attribute__((__unused__))
|
|
bbssi(int bitnr, long *addr)
|
|
{
|
|
register int ret;
|
|
|
|
__asm volatile("clrl %%r0;"
|
|
"bbssi %1,%2,1f;"
|
|
"incl %%r0;"
|
|
"1:;"
|
|
"movl %%r0,%0"
|
|
: "=&r"(ret)
|
|
: "g"(bitnr),"m"(*addr)
|
|
: "r0","cc","memory");
|
|
return ret;
|
|
}
|
|
|
|
static __inline int __attribute__((__unused__))
|
|
bbcci(int bitnr, long *addr)
|
|
{
|
|
register int ret;
|
|
|
|
__asm volatile("clrl %%r0;"
|
|
"bbcci %1,%2,1f;"
|
|
"incl %%r0;"
|
|
"1:;"
|
|
"movl %%r0,%0"
|
|
: "=&r"(ret)
|
|
: "g"(bitnr),"m"(*addr)
|
|
: "r0","cc","memory");
|
|
return ret;
|
|
}
|
|
|
|
static inline struct lwp *
|
|
cpu_switchto(struct lwp *oldlwp, struct lwp *newlwp, bool returning)
|
|
{
|
|
struct lwp *prevlwp;
|
|
__asm volatile(
|
|
"movl %1,%%r0;"
|
|
"movl %2,%%r1;"
|
|
"movpsl -(%%sp);"
|
|
"jsb Swtchto;"
|
|
"movl %%r0,%0"
|
|
: "=g"(prevlwp)
|
|
: "g" (oldlwp), "g" (newlwp)
|
|
: "r0", "r1");
|
|
return prevlwp;
|
|
}
|
|
|
|
/*
|
|
* Interlock instructions. Used both in multiprocessor environments to
|
|
* lock between CPUs and in uniprocessor systems when locking is required
|
|
* between I/O devices and the master CPU.
|
|
*/
|
|
/*
|
|
* Insqti() locks and inserts an element into the end of a queue.
|
|
* Returns -1 if interlock failed, 1 if inserted OK and 0 if first in queue.
|
|
*/
|
|
static __inline int __attribute__((__unused__))
|
|
insqti(void *entry, void *header) {
|
|
register int ret;
|
|
|
|
__asm volatile(
|
|
" mnegl $1,%0;"
|
|
" insqti (%1),(%2);"
|
|
" bcs 1f;" /* failed insert */
|
|
" beql 2f;" /* jump if first entry */
|
|
" movl $1,%0;"
|
|
" brb 1f;"
|
|
"2: clrl %0;"
|
|
" 1:;"
|
|
: "=&g"(ret)
|
|
: "r"(entry), "r"(header)
|
|
: "memory");
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Remqhi() removes an element from the head of the queue.
|
|
* Returns -1 if interlock failed, 0 if queue empty, address of the
|
|
* removed element otherwise.
|
|
*/
|
|
static __inline void *__attribute__((__unused__))
|
|
remqhi(void *header) {
|
|
register void *ret;
|
|
|
|
__asm volatile(
|
|
" remqhi (%1),%0;"
|
|
" bcs 1f;" /* failed interlock */
|
|
" bvs 2f;" /* nothing was removed */
|
|
" brb 3f;"
|
|
"1: mnegl $1,%0;"
|
|
" brb 3f;"
|
|
"2: clrl %0;"
|
|
" 3:;"
|
|
: "=&g"(ret)
|
|
: "r"(header)
|
|
: "memory");
|
|
|
|
return ret;
|
|
}
|
|
#define ILCK_FAILED -1 /* Interlock failed */
|
|
#define Q_EMPTY 0 /* Queue is/was empty */
|
|
#define Q_OK 1 /* Inserted OK */
|
|
|
|
#endif /* !_VAX_MACROS_H_ && !__lint__ */
|