mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-21 09:34:51 +07:00
f993cdd99a
Use the more common logging style. Miscellanea: o Convert old style continuation printks without KERN_CONT to pr_cont o Coalesce formats o Realign arguments o Remove unnecessary casts Signed-off-by: Joe Perches <joe@perches.com> Signed-off-by: Corey Minyard <cminyard@mvista.com>
913 lines
19 KiB
C
913 lines
19 KiB
C
// SPDX-License-Identifier: GPL-2.0+
|
|
/*
|
|
* ipmi_devintf.c
|
|
*
|
|
* Linux device interface for the IPMI message handler.
|
|
*
|
|
* Author: MontaVista Software, Inc.
|
|
* Corey Minyard <minyard@mvista.com>
|
|
* source@mvista.com
|
|
*
|
|
* Copyright 2002 MontaVista Software Inc.
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/moduleparam.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/poll.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/ipmi.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/init.h>
|
|
#include <linux/device.h>
|
|
#include <linux/compat.h>
|
|
|
|
struct ipmi_file_private
|
|
{
|
|
struct ipmi_user *user;
|
|
spinlock_t recv_msg_lock;
|
|
struct list_head recv_msgs;
|
|
struct file *file;
|
|
struct fasync_struct *fasync_queue;
|
|
wait_queue_head_t wait;
|
|
struct mutex recv_mutex;
|
|
int default_retries;
|
|
unsigned int default_retry_time_ms;
|
|
};
|
|
|
|
static void file_receive_handler(struct ipmi_recv_msg *msg,
|
|
void *handler_data)
|
|
{
|
|
struct ipmi_file_private *priv = handler_data;
|
|
int was_empty;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&priv->recv_msg_lock, flags);
|
|
was_empty = list_empty(&priv->recv_msgs);
|
|
list_add_tail(&msg->link, &priv->recv_msgs);
|
|
spin_unlock_irqrestore(&priv->recv_msg_lock, flags);
|
|
|
|
if (was_empty) {
|
|
wake_up_interruptible(&priv->wait);
|
|
kill_fasync(&priv->fasync_queue, SIGIO, POLL_IN);
|
|
}
|
|
}
|
|
|
|
static __poll_t ipmi_poll(struct file *file, poll_table *wait)
|
|
{
|
|
struct ipmi_file_private *priv = file->private_data;
|
|
__poll_t mask = 0;
|
|
unsigned long flags;
|
|
|
|
poll_wait(file, &priv->wait, wait);
|
|
|
|
spin_lock_irqsave(&priv->recv_msg_lock, flags);
|
|
|
|
if (!list_empty(&priv->recv_msgs))
|
|
mask |= (EPOLLIN | EPOLLRDNORM);
|
|
|
|
spin_unlock_irqrestore(&priv->recv_msg_lock, flags);
|
|
|
|
return mask;
|
|
}
|
|
|
|
static int ipmi_fasync(int fd, struct file *file, int on)
|
|
{
|
|
struct ipmi_file_private *priv = file->private_data;
|
|
|
|
return fasync_helper(fd, file, on, &priv->fasync_queue);
|
|
}
|
|
|
|
static const struct ipmi_user_hndl ipmi_hndlrs =
|
|
{
|
|
.ipmi_recv_hndl = file_receive_handler,
|
|
};
|
|
|
|
static int ipmi_open(struct inode *inode, struct file *file)
|
|
{
|
|
int if_num = iminor(inode);
|
|
int rv;
|
|
struct ipmi_file_private *priv;
|
|
|
|
priv = kmalloc(sizeof(*priv), GFP_KERNEL);
|
|
if (!priv)
|
|
return -ENOMEM;
|
|
|
|
priv->file = file;
|
|
|
|
rv = ipmi_create_user(if_num,
|
|
&ipmi_hndlrs,
|
|
priv,
|
|
&priv->user);
|
|
if (rv) {
|
|
kfree(priv);
|
|
goto out;
|
|
}
|
|
|
|
file->private_data = priv;
|
|
|
|
spin_lock_init(&priv->recv_msg_lock);
|
|
INIT_LIST_HEAD(&priv->recv_msgs);
|
|
init_waitqueue_head(&priv->wait);
|
|
priv->fasync_queue = NULL;
|
|
mutex_init(&priv->recv_mutex);
|
|
|
|
/* Use the low-level defaults. */
|
|
priv->default_retries = -1;
|
|
priv->default_retry_time_ms = 0;
|
|
|
|
out:
|
|
return rv;
|
|
}
|
|
|
|
static int ipmi_release(struct inode *inode, struct file *file)
|
|
{
|
|
struct ipmi_file_private *priv = file->private_data;
|
|
int rv;
|
|
struct ipmi_recv_msg *msg, *next;
|
|
|
|
rv = ipmi_destroy_user(priv->user);
|
|
if (rv)
|
|
return rv;
|
|
|
|
list_for_each_entry_safe(msg, next, &priv->recv_msgs, link)
|
|
ipmi_free_recv_msg(msg);
|
|
|
|
kfree(priv);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int handle_send_req(struct ipmi_user *user,
|
|
struct ipmi_req *req,
|
|
int retries,
|
|
unsigned int retry_time_ms)
|
|
{
|
|
int rv;
|
|
struct ipmi_addr addr;
|
|
struct kernel_ipmi_msg msg;
|
|
|
|
if (req->addr_len > sizeof(struct ipmi_addr))
|
|
return -EINVAL;
|
|
|
|
if (copy_from_user(&addr, req->addr, req->addr_len))
|
|
return -EFAULT;
|
|
|
|
msg.netfn = req->msg.netfn;
|
|
msg.cmd = req->msg.cmd;
|
|
msg.data_len = req->msg.data_len;
|
|
msg.data = kmalloc(IPMI_MAX_MSG_LENGTH, GFP_KERNEL);
|
|
if (!msg.data)
|
|
return -ENOMEM;
|
|
|
|
/* From here out we cannot return, we must jump to "out" for
|
|
error exits to free msgdata. */
|
|
|
|
rv = ipmi_validate_addr(&addr, req->addr_len);
|
|
if (rv)
|
|
goto out;
|
|
|
|
if (req->msg.data != NULL) {
|
|
if (req->msg.data_len > IPMI_MAX_MSG_LENGTH) {
|
|
rv = -EMSGSIZE;
|
|
goto out;
|
|
}
|
|
|
|
if (copy_from_user(msg.data,
|
|
req->msg.data,
|
|
req->msg.data_len)) {
|
|
rv = -EFAULT;
|
|
goto out;
|
|
}
|
|
} else {
|
|
msg.data_len = 0;
|
|
}
|
|
|
|
rv = ipmi_request_settime(user,
|
|
&addr,
|
|
req->msgid,
|
|
&msg,
|
|
NULL,
|
|
0,
|
|
retries,
|
|
retry_time_ms);
|
|
out:
|
|
kfree(msg.data);
|
|
return rv;
|
|
}
|
|
|
|
static int handle_recv(struct ipmi_file_private *priv,
|
|
bool trunc, struct ipmi_recv *rsp,
|
|
int (*copyout)(struct ipmi_recv *, void __user *),
|
|
void __user *to)
|
|
{
|
|
int addr_len;
|
|
struct list_head *entry;
|
|
struct ipmi_recv_msg *msg;
|
|
unsigned long flags;
|
|
int rv = 0;
|
|
|
|
/* We claim a mutex because we don't want two
|
|
users getting something from the queue at a time.
|
|
Since we have to release the spinlock before we can
|
|
copy the data to the user, it's possible another
|
|
user will grab something from the queue, too. Then
|
|
the messages might get out of order if something
|
|
fails and the message gets put back onto the
|
|
queue. This mutex prevents that problem. */
|
|
mutex_lock(&priv->recv_mutex);
|
|
|
|
/* Grab the message off the list. */
|
|
spin_lock_irqsave(&priv->recv_msg_lock, flags);
|
|
if (list_empty(&(priv->recv_msgs))) {
|
|
spin_unlock_irqrestore(&priv->recv_msg_lock, flags);
|
|
rv = -EAGAIN;
|
|
goto recv_err;
|
|
}
|
|
entry = priv->recv_msgs.next;
|
|
msg = list_entry(entry, struct ipmi_recv_msg, link);
|
|
list_del(entry);
|
|
spin_unlock_irqrestore(&priv->recv_msg_lock, flags);
|
|
|
|
addr_len = ipmi_addr_length(msg->addr.addr_type);
|
|
if (rsp->addr_len < addr_len) {
|
|
rv = -EINVAL;
|
|
goto recv_putback_on_err;
|
|
}
|
|
|
|
if (copy_to_user(rsp->addr, &msg->addr, addr_len)) {
|
|
rv = -EFAULT;
|
|
goto recv_putback_on_err;
|
|
}
|
|
rsp->addr_len = addr_len;
|
|
|
|
rsp->recv_type = msg->recv_type;
|
|
rsp->msgid = msg->msgid;
|
|
rsp->msg.netfn = msg->msg.netfn;
|
|
rsp->msg.cmd = msg->msg.cmd;
|
|
|
|
if (msg->msg.data_len > 0) {
|
|
if (rsp->msg.data_len < msg->msg.data_len) {
|
|
rv = -EMSGSIZE;
|
|
if (trunc)
|
|
msg->msg.data_len = rsp->msg.data_len;
|
|
else
|
|
goto recv_putback_on_err;
|
|
}
|
|
|
|
if (copy_to_user(rsp->msg.data,
|
|
msg->msg.data,
|
|
msg->msg.data_len)) {
|
|
rv = -EFAULT;
|
|
goto recv_putback_on_err;
|
|
}
|
|
rsp->msg.data_len = msg->msg.data_len;
|
|
} else {
|
|
rsp->msg.data_len = 0;
|
|
}
|
|
|
|
rv = copyout(rsp, to);
|
|
if (rv)
|
|
goto recv_putback_on_err;
|
|
|
|
mutex_unlock(&priv->recv_mutex);
|
|
ipmi_free_recv_msg(msg);
|
|
return 0;
|
|
|
|
recv_putback_on_err:
|
|
/* If we got an error, put the message back onto
|
|
the head of the queue. */
|
|
spin_lock_irqsave(&priv->recv_msg_lock, flags);
|
|
list_add(entry, &priv->recv_msgs);
|
|
spin_unlock_irqrestore(&priv->recv_msg_lock, flags);
|
|
recv_err:
|
|
mutex_unlock(&priv->recv_mutex);
|
|
return rv;
|
|
}
|
|
|
|
static int copyout_recv(struct ipmi_recv *rsp, void __user *to)
|
|
{
|
|
return copy_to_user(to, rsp, sizeof(struct ipmi_recv)) ? -EFAULT : 0;
|
|
}
|
|
|
|
static long ipmi_ioctl(struct file *file,
|
|
unsigned int cmd,
|
|
unsigned long data)
|
|
{
|
|
int rv = -EINVAL;
|
|
struct ipmi_file_private *priv = file->private_data;
|
|
void __user *arg = (void __user *)data;
|
|
|
|
switch (cmd)
|
|
{
|
|
case IPMICTL_SEND_COMMAND:
|
|
{
|
|
struct ipmi_req req;
|
|
int retries;
|
|
unsigned int retry_time_ms;
|
|
|
|
if (copy_from_user(&req, arg, sizeof(req))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
mutex_lock(&priv->recv_mutex);
|
|
retries = priv->default_retries;
|
|
retry_time_ms = priv->default_retry_time_ms;
|
|
mutex_unlock(&priv->recv_mutex);
|
|
|
|
rv = handle_send_req(priv->user, &req, retries, retry_time_ms);
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_SEND_COMMAND_SETTIME:
|
|
{
|
|
struct ipmi_req_settime req;
|
|
|
|
if (copy_from_user(&req, arg, sizeof(req))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
rv = handle_send_req(priv->user,
|
|
&req.req,
|
|
req.retries,
|
|
req.retry_time_ms);
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_RECEIVE_MSG:
|
|
case IPMICTL_RECEIVE_MSG_TRUNC:
|
|
{
|
|
struct ipmi_recv rsp;
|
|
|
|
if (copy_from_user(&rsp, arg, sizeof(rsp)))
|
|
rv = -EFAULT;
|
|
else
|
|
rv = handle_recv(priv, cmd == IPMICTL_RECEIVE_MSG_TRUNC,
|
|
&rsp, copyout_recv, arg);
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_REGISTER_FOR_CMD:
|
|
{
|
|
struct ipmi_cmdspec val;
|
|
|
|
if (copy_from_user(&val, arg, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
rv = ipmi_register_for_cmd(priv->user, val.netfn, val.cmd,
|
|
IPMI_CHAN_ALL);
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_UNREGISTER_FOR_CMD:
|
|
{
|
|
struct ipmi_cmdspec val;
|
|
|
|
if (copy_from_user(&val, arg, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
rv = ipmi_unregister_for_cmd(priv->user, val.netfn, val.cmd,
|
|
IPMI_CHAN_ALL);
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_REGISTER_FOR_CMD_CHANS:
|
|
{
|
|
struct ipmi_cmdspec_chans val;
|
|
|
|
if (copy_from_user(&val, arg, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
rv = ipmi_register_for_cmd(priv->user, val.netfn, val.cmd,
|
|
val.chans);
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_UNREGISTER_FOR_CMD_CHANS:
|
|
{
|
|
struct ipmi_cmdspec_chans val;
|
|
|
|
if (copy_from_user(&val, arg, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
rv = ipmi_unregister_for_cmd(priv->user, val.netfn, val.cmd,
|
|
val.chans);
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_SET_GETS_EVENTS_CMD:
|
|
{
|
|
int val;
|
|
|
|
if (copy_from_user(&val, arg, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
rv = ipmi_set_gets_events(priv->user, val);
|
|
break;
|
|
}
|
|
|
|
/* The next four are legacy, not per-channel. */
|
|
case IPMICTL_SET_MY_ADDRESS_CMD:
|
|
{
|
|
unsigned int val;
|
|
|
|
if (copy_from_user(&val, arg, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
rv = ipmi_set_my_address(priv->user, 0, val);
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_GET_MY_ADDRESS_CMD:
|
|
{
|
|
unsigned int val;
|
|
unsigned char rval;
|
|
|
|
rv = ipmi_get_my_address(priv->user, 0, &rval);
|
|
if (rv)
|
|
break;
|
|
|
|
val = rval;
|
|
|
|
if (copy_to_user(arg, &val, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_SET_MY_LUN_CMD:
|
|
{
|
|
unsigned int val;
|
|
|
|
if (copy_from_user(&val, arg, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
rv = ipmi_set_my_LUN(priv->user, 0, val);
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_GET_MY_LUN_CMD:
|
|
{
|
|
unsigned int val;
|
|
unsigned char rval;
|
|
|
|
rv = ipmi_get_my_LUN(priv->user, 0, &rval);
|
|
if (rv)
|
|
break;
|
|
|
|
val = rval;
|
|
|
|
if (copy_to_user(arg, &val, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_SET_MY_CHANNEL_ADDRESS_CMD:
|
|
{
|
|
struct ipmi_channel_lun_address_set val;
|
|
|
|
if (copy_from_user(&val, arg, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
return ipmi_set_my_address(priv->user, val.channel, val.value);
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_GET_MY_CHANNEL_ADDRESS_CMD:
|
|
{
|
|
struct ipmi_channel_lun_address_set val;
|
|
|
|
if (copy_from_user(&val, arg, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
rv = ipmi_get_my_address(priv->user, val.channel, &val.value);
|
|
if (rv)
|
|
break;
|
|
|
|
if (copy_to_user(arg, &val, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_SET_MY_CHANNEL_LUN_CMD:
|
|
{
|
|
struct ipmi_channel_lun_address_set val;
|
|
|
|
if (copy_from_user(&val, arg, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
rv = ipmi_set_my_LUN(priv->user, val.channel, val.value);
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_GET_MY_CHANNEL_LUN_CMD:
|
|
{
|
|
struct ipmi_channel_lun_address_set val;
|
|
|
|
if (copy_from_user(&val, arg, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
rv = ipmi_get_my_LUN(priv->user, val.channel, &val.value);
|
|
if (rv)
|
|
break;
|
|
|
|
if (copy_to_user(arg, &val, sizeof(val))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_SET_TIMING_PARMS_CMD:
|
|
{
|
|
struct ipmi_timing_parms parms;
|
|
|
|
if (copy_from_user(&parms, arg, sizeof(parms))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
mutex_lock(&priv->recv_mutex);
|
|
priv->default_retries = parms.retries;
|
|
priv->default_retry_time_ms = parms.retry_time_ms;
|
|
mutex_unlock(&priv->recv_mutex);
|
|
rv = 0;
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_GET_TIMING_PARMS_CMD:
|
|
{
|
|
struct ipmi_timing_parms parms;
|
|
|
|
mutex_lock(&priv->recv_mutex);
|
|
parms.retries = priv->default_retries;
|
|
parms.retry_time_ms = priv->default_retry_time_ms;
|
|
mutex_unlock(&priv->recv_mutex);
|
|
|
|
if (copy_to_user(arg, &parms, sizeof(parms))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
rv = 0;
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_GET_MAINTENANCE_MODE_CMD:
|
|
{
|
|
int mode;
|
|
|
|
mode = ipmi_get_maintenance_mode(priv->user);
|
|
if (copy_to_user(arg, &mode, sizeof(mode))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
rv = 0;
|
|
break;
|
|
}
|
|
|
|
case IPMICTL_SET_MAINTENANCE_MODE_CMD:
|
|
{
|
|
int mode;
|
|
|
|
if (copy_from_user(&mode, arg, sizeof(mode))) {
|
|
rv = -EFAULT;
|
|
break;
|
|
}
|
|
rv = ipmi_set_maintenance_mode(priv->user, mode);
|
|
break;
|
|
}
|
|
|
|
default:
|
|
rv = -ENOTTY;
|
|
break;
|
|
}
|
|
|
|
return rv;
|
|
}
|
|
|
|
#ifdef CONFIG_COMPAT
|
|
/*
|
|
* The following code contains code for supporting 32-bit compatible
|
|
* ioctls on 64-bit kernels. This allows running 32-bit apps on the
|
|
* 64-bit kernel
|
|
*/
|
|
#define COMPAT_IPMICTL_SEND_COMMAND \
|
|
_IOR(IPMI_IOC_MAGIC, 13, struct compat_ipmi_req)
|
|
#define COMPAT_IPMICTL_SEND_COMMAND_SETTIME \
|
|
_IOR(IPMI_IOC_MAGIC, 21, struct compat_ipmi_req_settime)
|
|
#define COMPAT_IPMICTL_RECEIVE_MSG \
|
|
_IOWR(IPMI_IOC_MAGIC, 12, struct compat_ipmi_recv)
|
|
#define COMPAT_IPMICTL_RECEIVE_MSG_TRUNC \
|
|
_IOWR(IPMI_IOC_MAGIC, 11, struct compat_ipmi_recv)
|
|
|
|
struct compat_ipmi_msg {
|
|
u8 netfn;
|
|
u8 cmd;
|
|
u16 data_len;
|
|
compat_uptr_t data;
|
|
};
|
|
|
|
struct compat_ipmi_req {
|
|
compat_uptr_t addr;
|
|
compat_uint_t addr_len;
|
|
compat_long_t msgid;
|
|
struct compat_ipmi_msg msg;
|
|
};
|
|
|
|
struct compat_ipmi_recv {
|
|
compat_int_t recv_type;
|
|
compat_uptr_t addr;
|
|
compat_uint_t addr_len;
|
|
compat_long_t msgid;
|
|
struct compat_ipmi_msg msg;
|
|
};
|
|
|
|
struct compat_ipmi_req_settime {
|
|
struct compat_ipmi_req req;
|
|
compat_int_t retries;
|
|
compat_uint_t retry_time_ms;
|
|
};
|
|
|
|
/*
|
|
* Define some helper functions for copying IPMI data
|
|
*/
|
|
static void get_compat_ipmi_msg(struct ipmi_msg *p64,
|
|
struct compat_ipmi_msg *p32)
|
|
{
|
|
p64->netfn = p32->netfn;
|
|
p64->cmd = p32->cmd;
|
|
p64->data_len = p32->data_len;
|
|
p64->data = compat_ptr(p32->data);
|
|
}
|
|
|
|
static void get_compat_ipmi_req(struct ipmi_req *p64,
|
|
struct compat_ipmi_req *p32)
|
|
{
|
|
p64->addr = compat_ptr(p32->addr);
|
|
p64->addr_len = p32->addr_len;
|
|
p64->msgid = p32->msgid;
|
|
get_compat_ipmi_msg(&p64->msg, &p32->msg);
|
|
}
|
|
|
|
static void get_compat_ipmi_req_settime(struct ipmi_req_settime *p64,
|
|
struct compat_ipmi_req_settime *p32)
|
|
{
|
|
get_compat_ipmi_req(&p64->req, &p32->req);
|
|
p64->retries = p32->retries;
|
|
p64->retry_time_ms = p32->retry_time_ms;
|
|
}
|
|
|
|
static void get_compat_ipmi_recv(struct ipmi_recv *p64,
|
|
struct compat_ipmi_recv *p32)
|
|
{
|
|
memset(p64, 0, sizeof(struct ipmi_recv));
|
|
p64->recv_type = p32->recv_type;
|
|
p64->addr = compat_ptr(p32->addr);
|
|
p64->addr_len = p32->addr_len;
|
|
p64->msgid = p32->msgid;
|
|
get_compat_ipmi_msg(&p64->msg, &p32->msg);
|
|
}
|
|
|
|
static int copyout_recv32(struct ipmi_recv *p64, void __user *to)
|
|
{
|
|
struct compat_ipmi_recv v32;
|
|
memset(&v32, 0, sizeof(struct compat_ipmi_recv));
|
|
v32.recv_type = p64->recv_type;
|
|
v32.addr = ptr_to_compat(p64->addr);
|
|
v32.addr_len = p64->addr_len;
|
|
v32.msgid = p64->msgid;
|
|
v32.msg.netfn = p64->msg.netfn;
|
|
v32.msg.cmd = p64->msg.cmd;
|
|
v32.msg.data_len = p64->msg.data_len;
|
|
v32.msg.data = ptr_to_compat(p64->msg.data);
|
|
return copy_to_user(to, &v32, sizeof(v32)) ? -EFAULT : 0;
|
|
}
|
|
|
|
/*
|
|
* Handle compatibility ioctls
|
|
*/
|
|
static long compat_ipmi_ioctl(struct file *filep, unsigned int cmd,
|
|
unsigned long arg)
|
|
{
|
|
struct ipmi_file_private *priv = filep->private_data;
|
|
|
|
switch(cmd) {
|
|
case COMPAT_IPMICTL_SEND_COMMAND:
|
|
{
|
|
struct ipmi_req rp;
|
|
struct compat_ipmi_req r32;
|
|
int retries;
|
|
unsigned int retry_time_ms;
|
|
|
|
if (copy_from_user(&r32, compat_ptr(arg), sizeof(r32)))
|
|
return -EFAULT;
|
|
|
|
get_compat_ipmi_req(&rp, &r32);
|
|
|
|
mutex_lock(&priv->recv_mutex);
|
|
retries = priv->default_retries;
|
|
retry_time_ms = priv->default_retry_time_ms;
|
|
mutex_unlock(&priv->recv_mutex);
|
|
|
|
return handle_send_req(priv->user, &rp,
|
|
retries, retry_time_ms);
|
|
}
|
|
case COMPAT_IPMICTL_SEND_COMMAND_SETTIME:
|
|
{
|
|
struct ipmi_req_settime sp;
|
|
struct compat_ipmi_req_settime sp32;
|
|
|
|
if (copy_from_user(&sp32, compat_ptr(arg), sizeof(sp32)))
|
|
return -EFAULT;
|
|
|
|
get_compat_ipmi_req_settime(&sp, &sp32);
|
|
|
|
return handle_send_req(priv->user, &sp.req,
|
|
sp.retries, sp.retry_time_ms);
|
|
}
|
|
case COMPAT_IPMICTL_RECEIVE_MSG:
|
|
case COMPAT_IPMICTL_RECEIVE_MSG_TRUNC:
|
|
{
|
|
struct ipmi_recv recv64;
|
|
struct compat_ipmi_recv recv32;
|
|
|
|
if (copy_from_user(&recv32, compat_ptr(arg), sizeof(recv32)))
|
|
return -EFAULT;
|
|
|
|
get_compat_ipmi_recv(&recv64, &recv32);
|
|
|
|
return handle_recv(priv,
|
|
cmd == COMPAT_IPMICTL_RECEIVE_MSG_TRUNC,
|
|
&recv64, copyout_recv32, compat_ptr(arg));
|
|
}
|
|
default:
|
|
return ipmi_ioctl(filep, cmd, arg);
|
|
}
|
|
}
|
|
#endif
|
|
|
|
static const struct file_operations ipmi_fops = {
|
|
.owner = THIS_MODULE,
|
|
.unlocked_ioctl = ipmi_ioctl,
|
|
#ifdef CONFIG_COMPAT
|
|
.compat_ioctl = compat_ipmi_ioctl,
|
|
#endif
|
|
.open = ipmi_open,
|
|
.release = ipmi_release,
|
|
.fasync = ipmi_fasync,
|
|
.poll = ipmi_poll,
|
|
.llseek = noop_llseek,
|
|
};
|
|
|
|
#define DEVICE_NAME "ipmidev"
|
|
|
|
static int ipmi_major;
|
|
module_param(ipmi_major, int, 0);
|
|
MODULE_PARM_DESC(ipmi_major, "Sets the major number of the IPMI device. By"
|
|
" default, or if you set it to zero, it will choose the next"
|
|
" available device. Setting it to -1 will disable the"
|
|
" interface. Other values will set the major device number"
|
|
" to that value.");
|
|
|
|
/* Keep track of the devices that are registered. */
|
|
struct ipmi_reg_list {
|
|
dev_t dev;
|
|
struct list_head link;
|
|
};
|
|
static LIST_HEAD(reg_list);
|
|
static DEFINE_MUTEX(reg_list_mutex);
|
|
|
|
static struct class *ipmi_class;
|
|
|
|
static void ipmi_new_smi(int if_num, struct device *device)
|
|
{
|
|
dev_t dev = MKDEV(ipmi_major, if_num);
|
|
struct ipmi_reg_list *entry;
|
|
|
|
entry = kmalloc(sizeof(*entry), GFP_KERNEL);
|
|
if (!entry) {
|
|
pr_err("ipmi_devintf: Unable to create the ipmi class device link\n");
|
|
return;
|
|
}
|
|
entry->dev = dev;
|
|
|
|
mutex_lock(®_list_mutex);
|
|
device_create(ipmi_class, device, dev, NULL, "ipmi%d", if_num);
|
|
list_add(&entry->link, ®_list);
|
|
mutex_unlock(®_list_mutex);
|
|
}
|
|
|
|
static void ipmi_smi_gone(int if_num)
|
|
{
|
|
dev_t dev = MKDEV(ipmi_major, if_num);
|
|
struct ipmi_reg_list *entry;
|
|
|
|
mutex_lock(®_list_mutex);
|
|
list_for_each_entry(entry, ®_list, link) {
|
|
if (entry->dev == dev) {
|
|
list_del(&entry->link);
|
|
kfree(entry);
|
|
break;
|
|
}
|
|
}
|
|
device_destroy(ipmi_class, dev);
|
|
mutex_unlock(®_list_mutex);
|
|
}
|
|
|
|
static struct ipmi_smi_watcher smi_watcher =
|
|
{
|
|
.owner = THIS_MODULE,
|
|
.new_smi = ipmi_new_smi,
|
|
.smi_gone = ipmi_smi_gone,
|
|
};
|
|
|
|
static int __init init_ipmi_devintf(void)
|
|
{
|
|
int rv;
|
|
|
|
if (ipmi_major < 0)
|
|
return -EINVAL;
|
|
|
|
pr_info("ipmi device interface\n");
|
|
|
|
ipmi_class = class_create(THIS_MODULE, "ipmi");
|
|
if (IS_ERR(ipmi_class)) {
|
|
pr_err("ipmi: can't register device class\n");
|
|
return PTR_ERR(ipmi_class);
|
|
}
|
|
|
|
rv = register_chrdev(ipmi_major, DEVICE_NAME, &ipmi_fops);
|
|
if (rv < 0) {
|
|
class_destroy(ipmi_class);
|
|
pr_err("ipmi: can't get major %d\n", ipmi_major);
|
|
return rv;
|
|
}
|
|
|
|
if (ipmi_major == 0) {
|
|
ipmi_major = rv;
|
|
}
|
|
|
|
rv = ipmi_smi_watcher_register(&smi_watcher);
|
|
if (rv) {
|
|
unregister_chrdev(ipmi_major, DEVICE_NAME);
|
|
class_destroy(ipmi_class);
|
|
pr_warn("ipmi: can't register smi watcher\n");
|
|
return rv;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
module_init(init_ipmi_devintf);
|
|
|
|
static void __exit cleanup_ipmi(void)
|
|
{
|
|
struct ipmi_reg_list *entry, *entry2;
|
|
mutex_lock(®_list_mutex);
|
|
list_for_each_entry_safe(entry, entry2, ®_list, link) {
|
|
list_del(&entry->link);
|
|
device_destroy(ipmi_class, entry->dev);
|
|
kfree(entry);
|
|
}
|
|
mutex_unlock(®_list_mutex);
|
|
class_destroy(ipmi_class);
|
|
ipmi_smi_watcher_unregister(&smi_watcher);
|
|
unregister_chrdev(ipmi_major, DEVICE_NAME);
|
|
}
|
|
module_exit(cleanup_ipmi);
|
|
|
|
MODULE_LICENSE("GPL");
|
|
MODULE_AUTHOR("Corey Minyard <minyard@mvista.com>");
|
|
MODULE_DESCRIPTION("Linux device interface for the IPMI message handler.");
|