mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-28 11:18:45 +07:00
37db8985b2
As virtio-ccw devices are channel devices, we need to use the dma area within the common I/O layer for any communication with the hypervisor. Note that we do not need to use that area for control blocks directly referenced by instructions, e.g. the orb. It handles neither QDIO in the common code, nor any device type specific stuff (like channel programs constructed by the DASD driver). An interesting side effect is that virtio structures are now going to get allocated in 31 bit addressable storage. Signed-off-by: Halil Pasic <pasic@linux.ibm.com> Reviewed-by: Sebastian Ott <sebott@linux.ibm.com> Reviewed-by: Cornelia Huck <cohuck@redhat.com> Reviewed-by: Michael Mueller <mimu@linux.ibm.com> Tested-by: Michael Mueller <mimu@linux.ibm.com> Signed-off-by: Heiko Carstens <heiko.carstens@de.ibm.com>
186 lines
5.6 KiB
C
186 lines
5.6 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef S390_IO_SCH_H
|
|
#define S390_IO_SCH_H
|
|
|
|
#include <linux/types.h>
|
|
#include <asm/schid.h>
|
|
#include <asm/ccwdev.h>
|
|
#include <asm/irq.h>
|
|
#include "css.h"
|
|
#include "orb.h"
|
|
|
|
struct io_subchannel_dma_area {
|
|
struct ccw1 sense_ccw; /* static ccw for sense command */
|
|
};
|
|
|
|
struct io_subchannel_private {
|
|
union orb orb; /* operation request block */
|
|
struct ccw_device *cdev;/* pointer to the child ccw device */
|
|
struct {
|
|
unsigned int suspend:1; /* allow suspend */
|
|
unsigned int prefetch:1;/* deny prefetch */
|
|
unsigned int inter:1; /* suppress intermediate interrupts */
|
|
} __packed options;
|
|
struct io_subchannel_dma_area *dma_area;
|
|
dma_addr_t dma_area_dma;
|
|
} __aligned(8);
|
|
|
|
#define to_io_private(n) ((struct io_subchannel_private *) \
|
|
dev_get_drvdata(&(n)->dev))
|
|
#define set_io_private(n, p) (dev_set_drvdata(&(n)->dev, p))
|
|
|
|
static inline struct ccw_device *sch_get_cdev(struct subchannel *sch)
|
|
{
|
|
struct io_subchannel_private *priv = to_io_private(sch);
|
|
return priv ? priv->cdev : NULL;
|
|
}
|
|
|
|
static inline void sch_set_cdev(struct subchannel *sch,
|
|
struct ccw_device *cdev)
|
|
{
|
|
struct io_subchannel_private *priv = to_io_private(sch);
|
|
if (priv)
|
|
priv->cdev = cdev;
|
|
}
|
|
|
|
#define MAX_CIWS 8
|
|
|
|
/*
|
|
* Possible status values for a CCW request's I/O.
|
|
*/
|
|
enum io_status {
|
|
IO_DONE,
|
|
IO_RUNNING,
|
|
IO_STATUS_ERROR,
|
|
IO_PATH_ERROR,
|
|
IO_REJECTED,
|
|
IO_KILLED
|
|
};
|
|
|
|
/**
|
|
* ccw_request - Internal CCW request.
|
|
* @cp: channel program to start
|
|
* @timeout: maximum allowable time in jiffies between start I/O and interrupt
|
|
* @maxretries: number of retries per I/O operation and path
|
|
* @lpm: mask of paths to use
|
|
* @check: optional callback that determines if results are final
|
|
* @filter: optional callback to adjust request status based on IRB data
|
|
* @callback: final callback
|
|
* @data: user-defined pointer passed to all callbacks
|
|
* @singlepath: if set, use only one path from @lpm per start I/O
|
|
* @cancel: non-zero if request was cancelled
|
|
* @done: non-zero if request was finished
|
|
* @mask: current path mask
|
|
* @retries: current number of retries
|
|
* @drc: delayed return code
|
|
*/
|
|
struct ccw_request {
|
|
struct ccw1 *cp;
|
|
unsigned long timeout;
|
|
u16 maxretries;
|
|
u8 lpm;
|
|
int (*check)(struct ccw_device *, void *);
|
|
enum io_status (*filter)(struct ccw_device *, void *, struct irb *,
|
|
enum io_status);
|
|
void (*callback)(struct ccw_device *, void *, int);
|
|
void *data;
|
|
unsigned int singlepath:1;
|
|
/* These fields are used internally. */
|
|
unsigned int cancel:1;
|
|
unsigned int done:1;
|
|
u16 mask;
|
|
u16 retries;
|
|
int drc;
|
|
} __attribute__((packed));
|
|
|
|
/*
|
|
* sense-id response buffer layout
|
|
*/
|
|
struct senseid {
|
|
/* common part */
|
|
u8 reserved; /* always 0x'FF' */
|
|
u16 cu_type; /* control unit type */
|
|
u8 cu_model; /* control unit model */
|
|
u16 dev_type; /* device type */
|
|
u8 dev_model; /* device model */
|
|
u8 unused; /* padding byte */
|
|
/* extended part */
|
|
struct ciw ciw[MAX_CIWS]; /* variable # of CIWs */
|
|
} __attribute__ ((packed, aligned(4)));
|
|
|
|
enum cdev_todo {
|
|
CDEV_TODO_NOTHING,
|
|
CDEV_TODO_ENABLE_CMF,
|
|
CDEV_TODO_REBIND,
|
|
CDEV_TODO_REGISTER,
|
|
CDEV_TODO_UNREG,
|
|
CDEV_TODO_UNREG_EVAL,
|
|
};
|
|
|
|
#define FAKE_CMD_IRB 1
|
|
#define FAKE_TM_IRB 2
|
|
|
|
struct ccw_device_dma_area {
|
|
struct senseid senseid; /* SenseID info */
|
|
struct ccw1 iccws[2]; /* ccws for SNID/SID/SPGID commands */
|
|
struct irb irb; /* device status */
|
|
struct pgid pgid[8]; /* path group IDs per chpid*/
|
|
};
|
|
|
|
struct ccw_device_private {
|
|
struct ccw_device *cdev;
|
|
struct subchannel *sch;
|
|
int state; /* device state */
|
|
atomic_t onoff;
|
|
struct ccw_dev_id dev_id; /* device id */
|
|
struct ccw_request req; /* internal I/O request */
|
|
int iretry;
|
|
u8 pgid_valid_mask; /* mask of valid PGIDs */
|
|
u8 pgid_todo_mask; /* mask of PGIDs to be adjusted */
|
|
u8 pgid_reset_mask; /* mask of PGIDs which were reset */
|
|
u8 path_noirq_mask; /* mask of paths for which no irq was
|
|
received */
|
|
u8 path_notoper_mask; /* mask of paths which were found
|
|
not operable */
|
|
u8 path_gone_mask; /* mask of paths, that became unavailable */
|
|
u8 path_new_mask; /* mask of paths, that became available */
|
|
u8 path_broken_mask; /* mask of paths, which were found to be
|
|
unusable */
|
|
struct {
|
|
unsigned int fast:1; /* post with "channel end" */
|
|
unsigned int repall:1; /* report every interrupt status */
|
|
unsigned int pgroup:1; /* do path grouping */
|
|
unsigned int force:1; /* allow forced online */
|
|
unsigned int mpath:1; /* do multipathing */
|
|
} __attribute__ ((packed)) options;
|
|
struct {
|
|
unsigned int esid:1; /* Ext. SenseID supported by HW */
|
|
unsigned int dosense:1; /* delayed SENSE required */
|
|
unsigned int doverify:1; /* delayed path verification */
|
|
unsigned int donotify:1; /* call notify function */
|
|
unsigned int recog_done:1; /* dev. recog. complete */
|
|
unsigned int fake_irb:2; /* deliver faked irb */
|
|
unsigned int resuming:1; /* recognition while resume */
|
|
unsigned int pgroup:1; /* pathgroup is set up */
|
|
unsigned int mpath:1; /* multipathing is set up */
|
|
unsigned int pgid_unknown:1;/* unknown pgid state */
|
|
unsigned int initialized:1; /* set if initial reference held */
|
|
} __attribute__((packed)) flags;
|
|
unsigned long intparm; /* user interruption parameter */
|
|
struct qdio_irq *qdio_data;
|
|
int async_kill_io_rc;
|
|
struct work_struct todo_work;
|
|
enum cdev_todo todo;
|
|
wait_queue_head_t wait_q;
|
|
struct timer_list timer;
|
|
void *cmb; /* measurement information */
|
|
struct list_head cmb_list; /* list of measured devices */
|
|
u64 cmb_start_time; /* clock value of cmb reset */
|
|
void *cmb_wait; /* deferred cmb enable/disable */
|
|
struct gen_pool *dma_pool;
|
|
struct ccw_device_dma_area *dma_area;
|
|
enum interruption_class int_class;
|
|
};
|
|
|
|
#endif
|