else
clear_bit(SOCK_PASSSEC, &sock->flags);
break;
+ case SO_MARK:
+ if (!capable(CAP_NET_ADMIN))
+ ret = -EPERM;
+ else {
+ sk->sk_mark = val;
+ }
+ break;
/* We implement the SO_SNDLOWAT etc to
not be settable (1003.1g 5.3) */
case SO_PEERSEC:
return security_socket_getpeersec_stream(sock, optval, optlen, len);
+ case SO_MARK:
+ v.val = sk->sk_mark;
+ break;
+
default:
return -ENOPROTOOPT;
}
{
struct proto *prot = sk->sk_prot;
- atomic_sub(sk->sk_forward_alloc / SK_MEM_QUANTUM,
+ atomic_sub(sk->sk_forward_alloc >> SK_MEM_QUANTUM_SHIFT,
prot->memory_allocated);
sk->sk_forward_alloc &= SK_MEM_QUANTUM - 1;
char *request_sock_slab_name = NULL;
char *timewait_sock_slab_name;
- if (pcounter_alloc(&prot->inuse) != 0) {
+ if (sock_prot_inuse_init(prot) != 0) {
printk(KERN_CRIT "%s: Can't alloc inuse counters!\n", prot->name);
goto out;
}
kmem_cache_destroy(prot->slab);
prot->slab = NULL;
out_free_inuse:
- pcounter_free(&prot->inuse);
+ sock_prot_inuse_free(prot);
out:
return -ENOBUFS;
}
list_del(&prot->node);
write_unlock(&proto_list_lock);
- pcounter_free(&prot->inuse);
+ sock_prot_inuse_free(prot);
if (prot->slab != NULL) {
kmem_cache_destroy(prot->slab);
#ifdef CONFIG_PROC_FS
static void *proto_seq_start(struct seq_file *seq, loff_t *pos)
+ __acquires(proto_list_lock)
{
read_lock(&proto_list_lock);
return seq_list_start_head(&proto_list, *pos);
}
static void proto_seq_stop(struct seq_file *seq, void *v)
+ __releases(proto_list_lock)
{
read_unlock(&proto_list_lock);
}