mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-25 04:20:53 +07:00
3e1d1d28d9
1. Establish a simple API for process freezing defined in linux/include/sched.h: frozen(process) Check for frozen process freezing(process) Check if a process is being frozen freeze(process) Tell a process to freeze (go to refrigerator) thaw_process(process) Restart process frozen_process(process) Process is frozen now 2. Remove all references to PF_FREEZE and PF_FROZEN from all kernel sources except sched.h 3. Fix numerous locations where try_to_freeze is manually done by a driver 4. Remove the argument that is no longer necessary from two function calls. 5. Some whitespace cleanup 6. Clear potential race in refrigerator (provides an open window of PF_FREEZE cleared before setting PF_FROZEN, recalc_sigpending does not check PF_FROZEN). This patch does not address the problem of freeze_processes() violating the rule that a task may only modify its own flags by setting PF_FREEZE. This is not clean in an SMP environment. freeze(process) is therefore not SMP safe! Signed-off-by: Christoph Lameter <christoph@lameter.com> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
262 lines
6.5 KiB
C
262 lines
6.5 KiB
C
/* krxiod.c: Rx I/O daemon
|
|
*
|
|
* Copyright (C) 2002 Red Hat, Inc. All Rights Reserved.
|
|
* Written by David Howells (dhowells@redhat.com)
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation; either version
|
|
* 2 of the License, or (at your option) any later version.
|
|
*/
|
|
|
|
#include <linux/sched.h>
|
|
#include <linux/completion.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/init.h>
|
|
#include <rxrpc/krxiod.h>
|
|
#include <rxrpc/transport.h>
|
|
#include <rxrpc/peer.h>
|
|
#include <rxrpc/call.h>
|
|
#include "internal.h"
|
|
|
|
static DECLARE_WAIT_QUEUE_HEAD(rxrpc_krxiod_sleepq);
|
|
static DECLARE_COMPLETION(rxrpc_krxiod_dead);
|
|
|
|
static atomic_t rxrpc_krxiod_qcount = ATOMIC_INIT(0);
|
|
|
|
static LIST_HEAD(rxrpc_krxiod_transportq);
|
|
static DEFINE_SPINLOCK(rxrpc_krxiod_transportq_lock);
|
|
|
|
static LIST_HEAD(rxrpc_krxiod_callq);
|
|
static DEFINE_SPINLOCK(rxrpc_krxiod_callq_lock);
|
|
|
|
static volatile int rxrpc_krxiod_die;
|
|
|
|
/*****************************************************************************/
|
|
/*
|
|
* Rx I/O daemon
|
|
*/
|
|
static int rxrpc_krxiod(void *arg)
|
|
{
|
|
DECLARE_WAITQUEUE(krxiod,current);
|
|
|
|
printk("Started krxiod %d\n",current->pid);
|
|
|
|
daemonize("krxiod");
|
|
|
|
/* loop around waiting for work to do */
|
|
do {
|
|
/* wait for work or to be told to exit */
|
|
_debug("### Begin Wait");
|
|
if (!atomic_read(&rxrpc_krxiod_qcount)) {
|
|
set_current_state(TASK_INTERRUPTIBLE);
|
|
|
|
add_wait_queue(&rxrpc_krxiod_sleepq, &krxiod);
|
|
|
|
for (;;) {
|
|
set_current_state(TASK_INTERRUPTIBLE);
|
|
if (atomic_read(&rxrpc_krxiod_qcount) ||
|
|
rxrpc_krxiod_die ||
|
|
signal_pending(current))
|
|
break;
|
|
|
|
schedule();
|
|
}
|
|
|
|
remove_wait_queue(&rxrpc_krxiod_sleepq, &krxiod);
|
|
set_current_state(TASK_RUNNING);
|
|
}
|
|
_debug("### End Wait");
|
|
|
|
/* do work if been given some to do */
|
|
_debug("### Begin Work");
|
|
|
|
/* see if there's a transport in need of attention */
|
|
if (!list_empty(&rxrpc_krxiod_transportq)) {
|
|
struct rxrpc_transport *trans = NULL;
|
|
|
|
spin_lock_irq(&rxrpc_krxiod_transportq_lock);
|
|
|
|
if (!list_empty(&rxrpc_krxiod_transportq)) {
|
|
trans = list_entry(
|
|
rxrpc_krxiod_transportq.next,
|
|
struct rxrpc_transport,
|
|
krxiodq_link);
|
|
|
|
list_del_init(&trans->krxiodq_link);
|
|
atomic_dec(&rxrpc_krxiod_qcount);
|
|
|
|
/* make sure it hasn't gone away and doesn't go
|
|
* away */
|
|
if (atomic_read(&trans->usage)>0)
|
|
rxrpc_get_transport(trans);
|
|
else
|
|
trans = NULL;
|
|
}
|
|
|
|
spin_unlock_irq(&rxrpc_krxiod_transportq_lock);
|
|
|
|
if (trans) {
|
|
rxrpc_trans_receive_packet(trans);
|
|
rxrpc_put_transport(trans);
|
|
}
|
|
}
|
|
|
|
/* see if there's a call in need of attention */
|
|
if (!list_empty(&rxrpc_krxiod_callq)) {
|
|
struct rxrpc_call *call = NULL;
|
|
|
|
spin_lock_irq(&rxrpc_krxiod_callq_lock);
|
|
|
|
if (!list_empty(&rxrpc_krxiod_callq)) {
|
|
call = list_entry(rxrpc_krxiod_callq.next,
|
|
struct rxrpc_call,
|
|
rcv_krxiodq_lk);
|
|
list_del_init(&call->rcv_krxiodq_lk);
|
|
atomic_dec(&rxrpc_krxiod_qcount);
|
|
|
|
/* make sure it hasn't gone away and doesn't go
|
|
* away */
|
|
if (atomic_read(&call->usage) > 0) {
|
|
_debug("@@@ KRXIOD"
|
|
" Begin Attend Call %p", call);
|
|
rxrpc_get_call(call);
|
|
}
|
|
else {
|
|
call = NULL;
|
|
}
|
|
}
|
|
|
|
spin_unlock_irq(&rxrpc_krxiod_callq_lock);
|
|
|
|
if (call) {
|
|
rxrpc_call_do_stuff(call);
|
|
rxrpc_put_call(call);
|
|
_debug("@@@ KRXIOD End Attend Call %p", call);
|
|
}
|
|
}
|
|
|
|
_debug("### End Work");
|
|
|
|
try_to_freeze();
|
|
|
|
/* discard pending signals */
|
|
rxrpc_discard_my_signals();
|
|
|
|
} while (!rxrpc_krxiod_die);
|
|
|
|
/* and that's all */
|
|
complete_and_exit(&rxrpc_krxiod_dead, 0);
|
|
|
|
} /* end rxrpc_krxiod() */
|
|
|
|
/*****************************************************************************/
|
|
/*
|
|
* start up a krxiod daemon
|
|
*/
|
|
int __init rxrpc_krxiod_init(void)
|
|
{
|
|
return kernel_thread(rxrpc_krxiod, NULL, 0);
|
|
|
|
} /* end rxrpc_krxiod_init() */
|
|
|
|
/*****************************************************************************/
|
|
/*
|
|
* kill the krxiod daemon and wait for it to complete
|
|
*/
|
|
void rxrpc_krxiod_kill(void)
|
|
{
|
|
rxrpc_krxiod_die = 1;
|
|
wake_up_all(&rxrpc_krxiod_sleepq);
|
|
wait_for_completion(&rxrpc_krxiod_dead);
|
|
|
|
} /* end rxrpc_krxiod_kill() */
|
|
|
|
/*****************************************************************************/
|
|
/*
|
|
* queue a transport for attention by krxiod
|
|
*/
|
|
void rxrpc_krxiod_queue_transport(struct rxrpc_transport *trans)
|
|
{
|
|
unsigned long flags;
|
|
|
|
_enter("");
|
|
|
|
if (list_empty(&trans->krxiodq_link)) {
|
|
spin_lock_irqsave(&rxrpc_krxiod_transportq_lock, flags);
|
|
|
|
if (list_empty(&trans->krxiodq_link)) {
|
|
if (atomic_read(&trans->usage) > 0) {
|
|
list_add_tail(&trans->krxiodq_link,
|
|
&rxrpc_krxiod_transportq);
|
|
atomic_inc(&rxrpc_krxiod_qcount);
|
|
}
|
|
}
|
|
|
|
spin_unlock_irqrestore(&rxrpc_krxiod_transportq_lock, flags);
|
|
wake_up_all(&rxrpc_krxiod_sleepq);
|
|
}
|
|
|
|
_leave("");
|
|
|
|
} /* end rxrpc_krxiod_queue_transport() */
|
|
|
|
/*****************************************************************************/
|
|
/*
|
|
* dequeue a transport from krxiod's attention queue
|
|
*/
|
|
void rxrpc_krxiod_dequeue_transport(struct rxrpc_transport *trans)
|
|
{
|
|
unsigned long flags;
|
|
|
|
_enter("");
|
|
|
|
spin_lock_irqsave(&rxrpc_krxiod_transportq_lock, flags);
|
|
if (!list_empty(&trans->krxiodq_link)) {
|
|
list_del_init(&trans->krxiodq_link);
|
|
atomic_dec(&rxrpc_krxiod_qcount);
|
|
}
|
|
spin_unlock_irqrestore(&rxrpc_krxiod_transportq_lock, flags);
|
|
|
|
_leave("");
|
|
|
|
} /* end rxrpc_krxiod_dequeue_transport() */
|
|
|
|
/*****************************************************************************/
|
|
/*
|
|
* queue a call for attention by krxiod
|
|
*/
|
|
void rxrpc_krxiod_queue_call(struct rxrpc_call *call)
|
|
{
|
|
unsigned long flags;
|
|
|
|
if (list_empty(&call->rcv_krxiodq_lk)) {
|
|
spin_lock_irqsave(&rxrpc_krxiod_callq_lock, flags);
|
|
if (atomic_read(&call->usage) > 0) {
|
|
list_add_tail(&call->rcv_krxiodq_lk,
|
|
&rxrpc_krxiod_callq);
|
|
atomic_inc(&rxrpc_krxiod_qcount);
|
|
}
|
|
spin_unlock_irqrestore(&rxrpc_krxiod_callq_lock, flags);
|
|
}
|
|
wake_up_all(&rxrpc_krxiod_sleepq);
|
|
|
|
} /* end rxrpc_krxiod_queue_call() */
|
|
|
|
/*****************************************************************************/
|
|
/*
|
|
* dequeue a call from krxiod's attention queue
|
|
*/
|
|
void rxrpc_krxiod_dequeue_call(struct rxrpc_call *call)
|
|
{
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&rxrpc_krxiod_callq_lock, flags);
|
|
if (!list_empty(&call->rcv_krxiodq_lk)) {
|
|
list_del_init(&call->rcv_krxiodq_lk);
|
|
atomic_dec(&rxrpc_krxiod_qcount);
|
|
}
|
|
spin_unlock_irqrestore(&rxrpc_krxiod_callq_lock, flags);
|
|
|
|
} /* end rxrpc_krxiod_dequeue_call() */
|