/* $NetBSD: macros.h,v 1.30 2003/09/18 18:38:48 ragge Exp $ */ /* * Copyright (c) 1994, 1998, 2000 Ludd, University of Lule}, Sweden. * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. All advertising materials mentioning features or use of this software * must display the following acknowledgement: * This product includes software developed at Ludd, University of Lule}. * 4. The name of the author may not be used to endorse or promote products * derived from this software without specific prior written permission * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ /* All bugs are subject to removal without further notice */ #if !defined(_VAX_MACROS_H_) && !defined(lint) #define _VAX_MACROS_H_ void __blkset(void *, int, size_t); void __blkcpy(const void *, void *, size_t); /* Here general macros are supposed to be stored */ static __inline__ int __attribute__((__unused__)) ffs(int reg) { register int val; __asm__ __volatile ("ffs $0,$32,%1,%0;" "bneq 1f;" "mnegl $1,%0;" "1:;" "incl %0" : "=&r" (val) : "r" (reg) ); return val; } static __inline__ void __attribute__((__unused__)) _remque(void *p) { __asm__ __volatile ("remque (%0),%0;clrl 4(%0)" : : "r" (p) : "memory" ); } static __inline__ void __attribute__((__unused__)) _insque(void *p, void *q) { __asm__ __volatile ("insque (%0),(%1)" : : "r" (p),"r" (q) : "memory" ); } static __inline__ void * __attribute__((__unused__)) memcpy(void *to, const void *from, size_t len) { if (len > 65535) { __blkcpy(from, to, len); } else { __asm__ __volatile ("movc3 %0,%1,%2" : : "g" (len), "m" (*(char *)from), "m" (*(char *)to) :"r0","r1","r2","r3","r4","r5","memory","cc"); } return to; } static __inline__ void * __attribute__((__unused__)) memmove(void *to, const void *from, size_t len) { if (len > 65535) { __blkcpy(from, to, len); } else { __asm__ __volatile ("movc3 %0,%1,%2" : : "g" (len), "m" (*(char *)from), "m" (*(char *)to) :"r0","r1","r2","r3","r4","r5","memory","cc"); } return to; } #ifdef notdef /* bcopy() is obsoleted in kernel */ static __inline__ void __attribute__((__unused__)) bcopy(const void *from, void *to, size_t len) { __asm__ __volatile ("movc3 %0,%1,%2" : : "g" (len), "m" (*(char *)from), "m" (*(char *)to) :"r0","r1","r2","r3","r4","r5","memory","cc"); } #endif static __inline__ void * __attribute__((__unused__)) memset(void *block, int c, size_t len) { if (len > 65535) { __blkset(block, c, len); } else { __asm__ __volatile ("movc5 $0,(%%sp),%2,%1,%0" : : "m" (*(char *)block), "g" (len), "g" (c) :"r0","r1","r2","r3","r4","r5","memory","cc"); } return block; } #ifdef notdef /* bzero() is obsoleted in kernel */ static __inline__ void __attribute__((__unused__)) bzero(void *block, size_t len) { if (len > 65535) __blkset(block, 0, len); else { __asm__ __volatile ("movc5 $0,(%%sp),$0,%1,%0" : : "m" (*(char *)block), "g" (len) :"r0","r1","r2","r3","r4","r5","memory","cc"); } } #endif #ifdef notdef /* XXX - the return syntax of memcmp is wrong */ static __inline__ int __attribute__((__unused__)) memcmp(const void *b1, const void *b2, size_t len) { register int ret; __asm__ __volatile("cmpc3 %3,(%1),(%2);" "movl %%r0,%0" : "=r" (ret) : "r" (b1), "r" (b2), "r" (len) : "r0","r1","r2","r3" ); return ret; } static __inline__ int __attribute__((__unused__)) bcmp(const void *b1, const void *b2, size_t len) { register int ret; __asm__ __volatile("cmpc3 %3,(%1),(%2);" "movl %%r0,%0" : "=r" (ret) : "r" (b1), "r" (b2), "r" (len) : "r0","r1","r2","r3" ); return ret; } /* Begin nya */ static __inline__ size_t __attribute__((__unused__)) strlen(const char *cp) { register size_t ret; __asm__ __volatile("locc $0,$65535,(%1);" "subl3 %%r0,$65535,%0" : "=r" (ret) : "r" (cp) : "r0","r1","cc" ); return ret; } static __inline__ char * __attribute__((__unused__)) strcat(char *cp, const char *c2) { __asm__ __volatile("locc $0,$65535,(%1);" "subl3 %%r0,$65535,%%r2;" "incl %%r2;" "locc $0,$65535,(%0);" "movc3 %%r2,(%1),(%%r1)" : : "r" (cp), "r" (c2) : "r0","r1","r2","r3","r4","r5","memory","cc"); return cp; } static __inline__ char * __attribute__((__unused__)) strncat(char *cp, const char *c2, size_t count) { __asm__ __volatile("locc $0,%2,(%1);" "subl3 %%r0,%2,%%r2;" "locc $0,$65535,(%0);" "movc3 %%r2,(%1),(%%r1);" "movb $0,(%%r3)" : : "r" (cp), "r" (c2), "g"(count) : "r0","r1","r2","r3","r4","r5","memory","cc"); return cp; } static __inline__ char * __attribute__((__unused__)) strcpy(char *cp, const char *c2) { __asm__ __volatile("locc $0,$65535,(%1);" "subl3 %%r0,$65535,%%r2;" "movc3 %%r2,(%1),(%0);" "movb $0,(%%r3)" : : "r" (cp), "r" (c2) : "r0","r1","r2","r3","r4","r5","memory","cc"); return cp; } static __inline__ char * __attribute__((__unused__)) strncpy(char *cp, const char *c2, size_t len) { __asm__ __volatile("movl %2,%%r2;" "locc $0,%%r2,(%1);" "beql 1f;" "subl3 %%r0,%2,%%r2;" "clrb (%0)[%%r2];" "1:;" "movc3 %%r2,(%1),(%0)" : : "r" (cp), "r" (c2), "g"(len) : "r0","r1","r2","r3","r4","r5","memory","cc"); return cp; } static __inline__ void * __attribute__((__unused__)) memchr(const void *cp, int c, size_t len) { void *ret; __asm__ __volatile("locc %2,%3,(%1);" "bneq 1f;" "clrl %%r1;" "1:;" "movl %%r1,%0" : "=g"(ret) : "r" (cp), "r" (c), "g"(len) : "r0","r1","cc"); return ret; } static __inline__ int __attribute__((__unused__)) strcmp(const char *cp, const char *c2) { register int ret; __asm__ __volatile("locc $0,$65535,(%1);" "subl3 %%r0,$65535,%%r0;" "incl %%r0;" "cmpc3 %%r0,(%1),(%2);" "beql 1f;" "movl $1,%%r2;" "cmpb (%%r1),(%%r3);" "bcc 1f;" "mnegl $1,%%r2;" "1:;" "movl %%r2,%0" : "=g"(ret) : "r" (cp), "r" (c2) : "r0","r1","r2","r3","cc"); return ret; } #endif #if 0 /* unused, but no point in deleting it since it _is_ an instruction */ static __inline__ int __attribute__((__unused__)) locc(int mask, char *cp, size_t size){ register ret; __asm__ __volatile("locc %1,%2,(%3);" "movl %%r0,%0" : "=r" (ret) : "r" (mask),"r"(size),"r"(cp) : "r0","r1" ); return ret; } #endif static __inline__ int __attribute__((__unused__)) scanc(u_int size, const u_char *cp, const u_char *table, int mask) { register int ret; __asm__ __volatile("scanc %1,(%2),(%3),%4;" "movl %%r0,%0" : "=g"(ret) : "r"(size),"r"(cp),"r"(table),"r"(mask) : "r0","r1","r2","r3" ); return ret; } static __inline__ int __attribute__((__unused__)) skpc(int mask, size_t size, u_char *cp) { register int ret; __asm__ __volatile("skpc %1,%2,(%3);" "movl %%r0,%0" : "=g"(ret) : "r"(mask),"r"(size),"r"(cp) : "r0","r1" ); return ret; } /* * Set/clear a bit at a memory position; interlocked. * Return 0 if already set, 1 otherwise. */ static __inline__ int __attribute__((__unused__)) bbssi(int bitnr, long *addr) { register int ret; __asm__ __volatile("clrl %%r0;" "bbssi %1,%2,1f;" "incl %%r0;" "1:;" "movl %%r0,%0" : "=&r"(ret) : "g"(bitnr),"m"(*addr) : "r0","cc","memory"); return ret; } static __inline__ int __attribute__((__unused__)) bbcci(int bitnr, long *addr) { register int ret; __asm__ __volatile("clrl %%r0;" "bbcci %1,%2,1f;" "incl %%r0;" "1:;" "movl %%r0,%0" : "=&r"(ret) : "g"(bitnr),"m"(*addr) : "r0","cc","memory"); return ret; } #define setrunqueue(p) \ __asm__ __volatile("movl %0,%%r0;jsb Setrq" :: "g"(p):"r0","r1","r2") #define remrunqueue(p) \ __asm__ __volatile("movl %0,%%r0;jsb Remrq" :: "g"(p):"r0","r1","r2") #define cpu_switch(p, newp) ({ \ register int ret; \ __asm__ __volatile("movpsl -(%%sp);jsb Swtch; movl %%r0,%0" \ : "=g"(ret) ::"r0","r1","r2","r3","r4","r5"); \ ret; }) #define cpu_switchto(p, newp) \ __asm __volatile("movpsl -(%%sp); movl %0,%%r2; jsb Swtchto" \ :: "g" (newp) : "r0", "r1", "r2", "r3", "r4", "r5") /* * Interlock instructions. Used both in multiprocessor environments to * lock between CPUs and in uniprocessor systems when locking is required * between I/O devices and the master CPU. */ /* * Insqti() locks and inserts an element into the end of a queue. * Returns -1 if interlock failed, 1 if inserted OK and 0 if first in queue. */ static __inline__ int __attribute__((__unused__)) insqti(void *entry, void *header) { register int ret; __asm__ __volatile( " mnegl $1,%0;" " insqti (%1),(%2);" " bcs 1f;" /* failed insert */ " beql 2f;" /* jump if first entry */ " movl $1,%0;" " brb 1f;" "2: clrl %0;" " 1:;" : "=&g"(ret) : "r"(entry), "r"(header) : "memory"); return ret; } /* * Remqhi() removes an element from the head of the queue. * Returns -1 if interlock failed, 0 if queue empty, address of the * removed element otherwise. */ static __inline__ void * __attribute__((__unused__)) remqhi(void *header) { register void *ret; __asm__ __volatile( " remqhi (%1),%0;" " bcs 1f;" /* failed interlock */ " bvs 2f;" /* nothing was removed */ " brb 3f;" "1: mnegl $1,%0;" " brb 3f;" "2: clrl %0;" " 3:;" : "=&g"(ret) : "r"(header) : "memory"); return ret; } #define ILCK_FAILED -1 /* Interlock failed */ #define Q_EMPTY 0 /* Queue is/was empty */ #define Q_OK 1 /* Inserted OK */ #endif /* _VAX_MACROS_H_ */