|
System | : | Linux MiraNet 3.0.0-14-generic-pae #23-Ubuntu SMP Mon Nov 21 22:07:10 UTC 2011 i686 |
Software | : | Apache. PHP/5.3.6-13ubuntu3.10 |
ID | : | uid=65534(nobody) gid=65534(nogroup) groups=65534(nogroup)
|
|
Safe Mode | : | OFF |
Open_Basedir | : | OFF |
Freespace | : | 20.45 GB of 70.42 GB (29.04%) |
|
MySQL: ON MSSQL: OFF Oracle: OFF PostgreSQL: OFF Curl: OFF Sockets: ON Fetch: OFF Wget: ON Perl: ON |
Disabled Functions: pcntl_alarm,pcntl_fork,pcntl_waitpid,pcntl_wait,pcntl_wifexited,pcntl_wifstopped,pcntl_wifsignaled,pcntl_wexitstatus,pcntl_wtermsig,pcntl_wstopsig,pcntl_signal,pcntl_signal_dispatch,pcntl_get_last_error,pcntl_strerror,pcntl_sigprocmask,pcntl_sigwaitinfo,pcntl_sigtimedwait,pcntl_exec,pcntl_getpriority,pcntl_setpriority,
|
[ System Info ]
[ Processes ]
[ SQL Manager ]
[ Eval ]
[ Encoder ]
[ Mailer ]
[ Back Connection ]
[ Backdoor Server ]
[ Kernel Exploit Search ]
[ MD5 Decrypter ]
[ Reverse IP ]
[ Kill Shell ]
[ FTP Brute-Force ]
|
|
/
usr/
src/
linux-headers-3.0.0-14/
arch/
tile/
include/
asm/
- drwxr-xr-x
|
Viewing file: spinlock_64.h (4.41 KB) -rw-r--r--Select action/file-type:  ( +) |  ( +) |  ( +) | Code ( +) | Session ( +) |  ( +) | SDB ( +) |  ( +) |  ( +) |  ( +) |  ( +) |  ( +) |
/* * Copyright 2011 Tilera Corporation. All Rights Reserved. * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License * as published by the Free Software Foundation, version 2. * * This program is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or * NON INFRINGEMENT. See the GNU General Public License for * more details. * * 64-bit SMP ticket spinlocks, allowing only a single CPU anywhere * (the type definitions are in asm/spinlock_types.h) */
#ifndef _ASM_TILE_SPINLOCK_64_H #define _ASM_TILE_SPINLOCK_64_H
/* Shifts and masks for the various fields in "lock". */ #define __ARCH_SPIN_CURRENT_SHIFT 17 #define __ARCH_SPIN_NEXT_MASK 0x7fff #define __ARCH_SPIN_NEXT_OVERFLOW 0x8000
/* * Return the "current" portion of a ticket lock value, * i.e. the number that currently owns the lock. */ static inline int arch_spin_current(u32 val) { return val >> __ARCH_SPIN_CURRENT_SHIFT; }
/* * Return the "next" portion of a ticket lock value, * i.e. the number that the next task to try to acquire the lock will get. */ static inline int arch_spin_next(u32 val) { return val & __ARCH_SPIN_NEXT_MASK; }
/* The lock is locked if a task would have to wait to get it. */ static inline int arch_spin_is_locked(arch_spinlock_t *lock) { u32 val = lock->lock; return arch_spin_current(val) != arch_spin_next(val); }
/* Bump the current ticket so the next task owns the lock. */ static inline void arch_spin_unlock(arch_spinlock_t *lock) { wmb(); /* guarantee anything modified under the lock is visible */ __insn_fetchadd4(&lock->lock, 1U << __ARCH_SPIN_CURRENT_SHIFT); }
void arch_spin_unlock_wait(arch_spinlock_t *lock);
void arch_spin_lock_slow(arch_spinlock_t *lock, u32 val);
/* Grab the "next" ticket number and bump it atomically. * If the current ticket is not ours, go to the slow path. * We also take the slow path if the "next" value overflows. */ static inline void arch_spin_lock(arch_spinlock_t *lock) { u32 val = __insn_fetchadd4(&lock->lock, 1); u32 ticket = val & (__ARCH_SPIN_NEXT_MASK | __ARCH_SPIN_NEXT_OVERFLOW); if (unlikely(arch_spin_current(val) != ticket)) arch_spin_lock_slow(lock, ticket); }
/* Try to get the lock, and return whether we succeeded. */ int arch_spin_trylock(arch_spinlock_t *lock);
/* We cannot take an interrupt after getting a ticket, so don't enable them. */ #define arch_spin_lock_flags(lock, flags) arch_spin_lock(lock)
/* * Read-write spinlocks, allowing multiple readers * but only one writer. * * We use fetchadd() for readers, and fetchor() with the sign bit * for writers. */
#define __WRITE_LOCK_BIT (1 << 31)
static inline int arch_write_val_locked(int val) { return val < 0; /* Optimize "val & __WRITE_LOCK_BIT". */ }
/** * read_can_lock - would read_trylock() succeed? * @lock: the rwlock in question. */ static inline int arch_read_can_lock(arch_rwlock_t *rw) { return !arch_write_val_locked(rw->lock); }
/** * write_can_lock - would write_trylock() succeed? * @lock: the rwlock in question. */ static inline int arch_write_can_lock(arch_rwlock_t *rw) { return rw->lock == 0; }
extern void __read_lock_failed(arch_rwlock_t *rw);
static inline void arch_read_lock(arch_rwlock_t *rw) { u32 val = __insn_fetchaddgez4(&rw->lock, 1); if (unlikely(arch_write_val_locked(val))) __read_lock_failed(rw); }
extern void __write_lock_failed(arch_rwlock_t *rw, u32 val);
static inline void arch_write_lock(arch_rwlock_t *rw) { u32 val = __insn_fetchor4(&rw->lock, __WRITE_LOCK_BIT); if (unlikely(val != 0)) __write_lock_failed(rw, val); }
static inline void arch_read_unlock(arch_rwlock_t *rw) { __insn_mf(); __insn_fetchadd4(&rw->lock, -1); }
static inline void arch_write_unlock(arch_rwlock_t *rw) { __insn_mf(); rw->lock = 0; }
static inline int arch_read_trylock(arch_rwlock_t *rw) { return !arch_write_val_locked(__insn_fetchaddgez4(&rw->lock, 1)); }
static inline int arch_write_trylock(arch_rwlock_t *rw) { u32 val = __insn_fetchor4(&rw->lock, __WRITE_LOCK_BIT); if (likely(val == 0)) return 1; if (!arch_write_val_locked(val)) __insn_fetchand4(&rw->lock, ~__WRITE_LOCK_BIT); return 0; }
#define arch_read_lock_flags(lock, flags) arch_read_lock(lock) #define arch_write_lock_flags(lock, flags) arch_write_lock(lock)
#endif /* _ASM_TILE_SPINLOCK_64_H */
|