mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-26 03:55:32 +07:00
6240300597
RPC-over-RDMA transport to use the new rdma_rw API. Christoph cleaned the way nfs operations are declared, removing a bunch of function-pointer casts and declaring the operation vectors as const. Christoph's changes touch both client and server, and both client and server pulls this time around should be based on the same commits from Christoph. (Note: Anna and I initially didn't coordinate this well and we realized our pull requests were going to leave you with Christoph's 33 patches duplicated between our two trees. We decided a last-minute rebase was the lesser of two evils, so her pull request will show that last-minute rebase. Yell if that was the wrong choice, and we'll know better for next time....) -----BEGIN PGP SIGNATURE----- iQIcBAABAgAGBQJZZ80JAAoJECebzXlCjuG+PiMP/jmw4IbzY4qt/X8aldVTMPZ8 TkEXuZSrc7FbmroqAR0XN/qJjzENKUcrnlYm7HKVe6iItTZUvJuVThtHQVGzZUZD wP2VRzgkky59aDs9cphfTPGKPKL1MtoC3qQdFmKd/8ZhBDHIq89A2pQJwl7PI4rA IHzvLmZtTKL+xWoypqZQxepONhEY2ZPrffGWL+5OVF/dPmWfJ6m/M6jRTb7zV/YD PZyRqWQ8UY/HwZTwRrxZDCCxUsmRUPZz195iFjM8wvBl7auWNetC22gyyITlvfzf 1m0zJqw3qn09+v2xnAWs/ZVxypg6rsEiIcL2mf0JC/tQh+iIzabc4e/TwDEWqSq+ ocQrvXJuZCjsrMqg4oaIuDFogaZCsGR5wxDAEyfYDS/8fMdiKq8xJzT7v31/2U37 Bsr1hvgAmD4eZWaTrJg11V5RnTzDgns+EtNfISR8t4/k+wehDfyzav8A+j72sqvR JT+7iUEd0QcBwo+MCC7AOnLLsIX45QUjZKKrvZNAC1fmr8RyAF1zo5HHO+NNjLuP J2PUG2GbNxsQkm/JAFKDvyklLpEXZc6uyYAcEefirxYbh1x0GfuetzqtH58DtrQL /1e80MRG9Qgq5S8PvYyvp1bIQPDRaQ188chEvzZy+3QeNXydq2LzDh0bjlM+4A9I DZhP2pNGLh0ImaPtX0q+ =mR/a -----END PGP SIGNATURE----- Merge tag 'nfsd-4.13' of git://linux-nfs.org/~bfields/linux Pull nfsd updates from Bruce Fields: "Chuck's RDMA update overhauls the "call receive" side of the RPC-over-RDMA transport to use the new rdma_rw API. Christoph cleaned the way nfs operations are declared, removing a bunch of function-pointer casts and declaring the operation vectors as const. Christoph's changes touch both client and server, and both client and server pulls this time around should be based on the same commits from Christoph" * tag 'nfsd-4.13' of git://linux-nfs.org/~bfields/linux: (53 commits) svcrdma: fix an incorrect check on -E2BIG and -EINVAL nfsd4: factor ctime into change attribute svcrdma: Remove svc_rdma_chunk_ctxt::cc_dir field svcrdma: use offset_in_page() macro svcrdma: Clean up after converting svc_rdma_recvfrom to rdma_rw API svcrdma: Clean-up svc_rdma_unmap_dma svcrdma: Remove frmr cache svcrdma: Remove unused Read completion handlers svcrdma: Properly compute .len and .buflen for received RPC Calls svcrdma: Use generic RDMA R/W API in RPC Call path svcrdma: Add recvfrom helpers to svc_rdma_rw.c sunrpc: Allocate up to RPCSVC_MAXPAGES per svc_rqst svcrdma: Don't account for Receive queue "starvation" svcrdma: Improve Reply chunk sanity checking svcrdma: Improve Write chunk sanity checking svcrdma: Improve Read chunk sanity checking svcrdma: Remove svc_rdma_marshal.c svcrdma: Avoid Send Queue overflow svcrdma: Squelch disconnection messages sunrpc: Disable splice for krb5i ...
297 lines
9.1 KiB
C
297 lines
9.1 KiB
C
/*
|
|
* linux/include/linux/sunrpc/sched.h
|
|
*
|
|
* Scheduling primitives for kernel Sun RPC.
|
|
*
|
|
* Copyright (C) 1996, Olaf Kirch <okir@monad.swb.de>
|
|
*/
|
|
|
|
#ifndef _LINUX_SUNRPC_SCHED_H_
|
|
#define _LINUX_SUNRPC_SCHED_H_
|
|
|
|
#include <linux/timer.h>
|
|
#include <linux/ktime.h>
|
|
#include <linux/sunrpc/types.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/wait_bit.h>
|
|
#include <linux/workqueue.h>
|
|
#include <linux/sunrpc/xdr.h>
|
|
|
|
/*
|
|
* This is the actual RPC procedure call info.
|
|
*/
|
|
struct rpc_procinfo;
|
|
struct rpc_message {
|
|
const struct rpc_procinfo *rpc_proc; /* Procedure information */
|
|
void * rpc_argp; /* Arguments */
|
|
void * rpc_resp; /* Result */
|
|
struct rpc_cred * rpc_cred; /* Credentials */
|
|
};
|
|
|
|
struct rpc_call_ops;
|
|
struct rpc_wait_queue;
|
|
struct rpc_wait {
|
|
struct list_head list; /* wait queue links */
|
|
struct list_head links; /* Links to related tasks */
|
|
struct list_head timer_list; /* Timer list */
|
|
unsigned long expires;
|
|
};
|
|
|
|
/*
|
|
* This is the RPC task struct
|
|
*/
|
|
struct rpc_task {
|
|
atomic_t tk_count; /* Reference count */
|
|
int tk_status; /* result of last operation */
|
|
struct list_head tk_task; /* global list of tasks */
|
|
|
|
/*
|
|
* callback to be executed after waking up
|
|
* action next procedure for async tasks
|
|
*/
|
|
void (*tk_callback)(struct rpc_task *);
|
|
void (*tk_action)(struct rpc_task *);
|
|
|
|
unsigned long tk_timeout; /* timeout for rpc_sleep() */
|
|
unsigned long tk_runstate; /* Task run status */
|
|
|
|
struct rpc_wait_queue *tk_waitqueue; /* RPC wait queue we're on */
|
|
union {
|
|
struct work_struct tk_work; /* Async task work queue */
|
|
struct rpc_wait tk_wait; /* RPC wait */
|
|
} u;
|
|
|
|
/*
|
|
* RPC call state
|
|
*/
|
|
struct rpc_message tk_msg; /* RPC call info */
|
|
void * tk_calldata; /* Caller private data */
|
|
const struct rpc_call_ops *tk_ops; /* Caller callbacks */
|
|
|
|
struct rpc_clnt * tk_client; /* RPC client */
|
|
struct rpc_xprt * tk_xprt; /* Transport */
|
|
|
|
struct rpc_rqst * tk_rqstp; /* RPC request */
|
|
|
|
struct workqueue_struct *tk_workqueue; /* Normally rpciod, but could
|
|
* be any workqueue
|
|
*/
|
|
ktime_t tk_start; /* RPC task init timestamp */
|
|
|
|
pid_t tk_owner; /* Process id for batching tasks */
|
|
unsigned short tk_flags; /* misc flags */
|
|
unsigned short tk_timeouts; /* maj timeouts */
|
|
|
|
#if IS_ENABLED(CONFIG_SUNRPC_DEBUG) || IS_ENABLED(CONFIG_TRACEPOINTS)
|
|
unsigned short tk_pid; /* debugging aid */
|
|
#endif
|
|
unsigned char tk_priority : 2,/* Task priority */
|
|
tk_garb_retry : 2,
|
|
tk_cred_retry : 2,
|
|
tk_rebind_retry : 2;
|
|
};
|
|
|
|
typedef void (*rpc_action)(struct rpc_task *);
|
|
|
|
struct rpc_call_ops {
|
|
void (*rpc_call_prepare)(struct rpc_task *, void *);
|
|
void (*rpc_call_done)(struct rpc_task *, void *);
|
|
void (*rpc_count_stats)(struct rpc_task *, void *);
|
|
void (*rpc_release)(void *);
|
|
};
|
|
|
|
struct rpc_task_setup {
|
|
struct rpc_task *task;
|
|
struct rpc_clnt *rpc_client;
|
|
struct rpc_xprt *rpc_xprt;
|
|
const struct rpc_message *rpc_message;
|
|
const struct rpc_call_ops *callback_ops;
|
|
void *callback_data;
|
|
struct workqueue_struct *workqueue;
|
|
unsigned short flags;
|
|
signed char priority;
|
|
};
|
|
|
|
/*
|
|
* RPC task flags
|
|
*/
|
|
#define RPC_TASK_ASYNC 0x0001 /* is an async task */
|
|
#define RPC_TASK_SWAPPER 0x0002 /* is swapping in/out */
|
|
#define RPC_CALL_MAJORSEEN 0x0020 /* major timeout seen */
|
|
#define RPC_TASK_ROOTCREDS 0x0040 /* force root creds */
|
|
#define RPC_TASK_DYNAMIC 0x0080 /* task was kmalloc'ed */
|
|
#define RPC_TASK_KILLED 0x0100 /* task was killed */
|
|
#define RPC_TASK_SOFT 0x0200 /* Use soft timeouts */
|
|
#define RPC_TASK_SOFTCONN 0x0400 /* Fail if can't connect */
|
|
#define RPC_TASK_SENT 0x0800 /* message was sent */
|
|
#define RPC_TASK_TIMEOUT 0x1000 /* fail with ETIMEDOUT on timeout */
|
|
#define RPC_TASK_NOCONNECT 0x2000 /* return ENOTCONN if not connected */
|
|
#define RPC_TASK_NO_RETRANS_TIMEOUT 0x4000 /* wait forever for a reply */
|
|
|
|
#define RPC_IS_ASYNC(t) ((t)->tk_flags & RPC_TASK_ASYNC)
|
|
#define RPC_IS_SWAPPER(t) ((t)->tk_flags & RPC_TASK_SWAPPER)
|
|
#define RPC_DO_ROOTOVERRIDE(t) ((t)->tk_flags & RPC_TASK_ROOTCREDS)
|
|
#define RPC_ASSASSINATED(t) ((t)->tk_flags & RPC_TASK_KILLED)
|
|
#define RPC_IS_SOFT(t) ((t)->tk_flags & (RPC_TASK_SOFT|RPC_TASK_TIMEOUT))
|
|
#define RPC_IS_SOFTCONN(t) ((t)->tk_flags & RPC_TASK_SOFTCONN)
|
|
#define RPC_WAS_SENT(t) ((t)->tk_flags & RPC_TASK_SENT)
|
|
|
|
#define RPC_TASK_RUNNING 0
|
|
#define RPC_TASK_QUEUED 1
|
|
#define RPC_TASK_ACTIVE 2
|
|
|
|
#define RPC_IS_RUNNING(t) test_bit(RPC_TASK_RUNNING, &(t)->tk_runstate)
|
|
#define rpc_set_running(t) set_bit(RPC_TASK_RUNNING, &(t)->tk_runstate)
|
|
#define rpc_test_and_set_running(t) \
|
|
test_and_set_bit(RPC_TASK_RUNNING, &(t)->tk_runstate)
|
|
#define rpc_clear_running(t) \
|
|
do { \
|
|
smp_mb__before_atomic(); \
|
|
clear_bit(RPC_TASK_RUNNING, &(t)->tk_runstate); \
|
|
smp_mb__after_atomic(); \
|
|
} while (0)
|
|
|
|
#define RPC_IS_QUEUED(t) test_bit(RPC_TASK_QUEUED, &(t)->tk_runstate)
|
|
#define rpc_set_queued(t) set_bit(RPC_TASK_QUEUED, &(t)->tk_runstate)
|
|
#define rpc_clear_queued(t) \
|
|
do { \
|
|
smp_mb__before_atomic(); \
|
|
clear_bit(RPC_TASK_QUEUED, &(t)->tk_runstate); \
|
|
smp_mb__after_atomic(); \
|
|
} while (0)
|
|
|
|
#define RPC_IS_ACTIVATED(t) test_bit(RPC_TASK_ACTIVE, &(t)->tk_runstate)
|
|
|
|
/*
|
|
* Task priorities.
|
|
* Note: if you change these, you must also change
|
|
* the task initialization definitions below.
|
|
*/
|
|
#define RPC_PRIORITY_LOW (-1)
|
|
#define RPC_PRIORITY_NORMAL (0)
|
|
#define RPC_PRIORITY_HIGH (1)
|
|
#define RPC_PRIORITY_PRIVILEGED (2)
|
|
#define RPC_NR_PRIORITY (1 + RPC_PRIORITY_PRIVILEGED - RPC_PRIORITY_LOW)
|
|
|
|
struct rpc_timer {
|
|
struct timer_list timer;
|
|
struct list_head list;
|
|
unsigned long expires;
|
|
};
|
|
|
|
/*
|
|
* RPC synchronization objects
|
|
*/
|
|
struct rpc_wait_queue {
|
|
spinlock_t lock;
|
|
struct list_head tasks[RPC_NR_PRIORITY]; /* task queue for each priority level */
|
|
pid_t owner; /* process id of last task serviced */
|
|
unsigned char maxpriority; /* maximum priority (0 if queue is not a priority queue) */
|
|
unsigned char priority; /* current priority */
|
|
unsigned char nr; /* # tasks remaining for cookie */
|
|
unsigned short qlen; /* total # tasks waiting in queue */
|
|
struct rpc_timer timer_list;
|
|
#if IS_ENABLED(CONFIG_SUNRPC_DEBUG) || IS_ENABLED(CONFIG_TRACEPOINTS)
|
|
const char * name;
|
|
#endif
|
|
};
|
|
|
|
/*
|
|
* This is the # requests to send consecutively
|
|
* from a single cookie. The aim is to improve
|
|
* performance of NFS operations such as read/write.
|
|
*/
|
|
#define RPC_BATCH_COUNT 16
|
|
#define RPC_IS_PRIORITY(q) ((q)->maxpriority > 0)
|
|
|
|
/*
|
|
* Function prototypes
|
|
*/
|
|
struct rpc_task *rpc_new_task(const struct rpc_task_setup *);
|
|
struct rpc_task *rpc_run_task(const struct rpc_task_setup *);
|
|
struct rpc_task *rpc_run_bc_task(struct rpc_rqst *req);
|
|
void rpc_put_task(struct rpc_task *);
|
|
void rpc_put_task_async(struct rpc_task *);
|
|
void rpc_exit_task(struct rpc_task *);
|
|
void rpc_exit(struct rpc_task *, int);
|
|
void rpc_release_calldata(const struct rpc_call_ops *, void *);
|
|
void rpc_killall_tasks(struct rpc_clnt *);
|
|
void rpc_execute(struct rpc_task *);
|
|
void rpc_init_priority_wait_queue(struct rpc_wait_queue *, const char *);
|
|
void rpc_init_wait_queue(struct rpc_wait_queue *, const char *);
|
|
void rpc_destroy_wait_queue(struct rpc_wait_queue *);
|
|
void rpc_sleep_on(struct rpc_wait_queue *, struct rpc_task *,
|
|
rpc_action action);
|
|
void rpc_sleep_on_priority(struct rpc_wait_queue *,
|
|
struct rpc_task *,
|
|
rpc_action action,
|
|
int priority);
|
|
void rpc_wake_up_queued_task(struct rpc_wait_queue *,
|
|
struct rpc_task *);
|
|
void rpc_wake_up(struct rpc_wait_queue *);
|
|
struct rpc_task *rpc_wake_up_next(struct rpc_wait_queue *);
|
|
struct rpc_task *rpc_wake_up_first_on_wq(struct workqueue_struct *wq,
|
|
struct rpc_wait_queue *,
|
|
bool (*)(struct rpc_task *, void *),
|
|
void *);
|
|
struct rpc_task *rpc_wake_up_first(struct rpc_wait_queue *,
|
|
bool (*)(struct rpc_task *, void *),
|
|
void *);
|
|
void rpc_wake_up_status(struct rpc_wait_queue *, int);
|
|
void rpc_delay(struct rpc_task *, unsigned long);
|
|
int rpc_malloc(struct rpc_task *);
|
|
void rpc_free(struct rpc_task *);
|
|
int rpciod_up(void);
|
|
void rpciod_down(void);
|
|
int __rpc_wait_for_completion_task(struct rpc_task *task, wait_bit_action_f *);
|
|
#if IS_ENABLED(CONFIG_SUNRPC_DEBUG)
|
|
struct net;
|
|
void rpc_show_tasks(struct net *);
|
|
#endif
|
|
int rpc_init_mempool(void);
|
|
void rpc_destroy_mempool(void);
|
|
extern struct workqueue_struct *rpciod_workqueue;
|
|
extern struct workqueue_struct *xprtiod_workqueue;
|
|
void rpc_prepare_task(struct rpc_task *task);
|
|
|
|
static inline int rpc_wait_for_completion_task(struct rpc_task *task)
|
|
{
|
|
return __rpc_wait_for_completion_task(task, NULL);
|
|
}
|
|
|
|
#if IS_ENABLED(CONFIG_SUNRPC_DEBUG) || IS_ENABLED(CONFIG_TRACEPOINTS)
|
|
static inline const char * rpc_qname(const struct rpc_wait_queue *q)
|
|
{
|
|
return ((q && q->name) ? q->name : "unknown");
|
|
}
|
|
|
|
static inline void rpc_assign_waitqueue_name(struct rpc_wait_queue *q,
|
|
const char *name)
|
|
{
|
|
q->name = name;
|
|
}
|
|
#else
|
|
static inline void rpc_assign_waitqueue_name(struct rpc_wait_queue *q,
|
|
const char *name)
|
|
{
|
|
}
|
|
#endif
|
|
|
|
#if IS_ENABLED(CONFIG_SUNRPC_SWAP)
|
|
int rpc_clnt_swap_activate(struct rpc_clnt *clnt);
|
|
void rpc_clnt_swap_deactivate(struct rpc_clnt *clnt);
|
|
#else
|
|
static inline int
|
|
rpc_clnt_swap_activate(struct rpc_clnt *clnt)
|
|
{
|
|
return -EINVAL;
|
|
}
|
|
|
|
static inline void
|
|
rpc_clnt_swap_deactivate(struct rpc_clnt *clnt)
|
|
{
|
|
}
|
|
#endif /* CONFIG_SUNRPC_SWAP */
|
|
|
|
#endif /* _LINUX_SUNRPC_SCHED_H_ */
|