mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-24 23:30:53 +07:00
c90fca951e
Notable changes: - Support for split PMD page table lock on 64-bit Book3S (Power8/9). - Add support for HAVE_RELIABLE_STACKTRACE, so we properly support live patching again. - Add support for patching barrier_nospec in copy_from_user() and syscall entry. - A couple of fixes for our data breakpoints on Book3S. - A series from Nick optimising TLB/mm handling with the Radix MMU. - Numerous small cleanups to squash sparse/gcc warnings from Mathieu Malaterre. - Several series optimising various parts of the 32-bit code from Christophe Leroy. - Removal of support for two old machines, "SBC834xE" and "C2K" ("GEFanuc,C2K"), which is why the diffstat has so many deletions. And many other small improvements & fixes. There's a few out-of-area changes. Some minor ftrace changes OK'ed by Steve, and a fix to our powernv cpuidle driver. Then there's a series touching mm, x86 and fs/proc/task_mmu.c, which cleans up some details around pkey support. It was ack'ed/reviewed by Ingo & Dave and has been in next for several weeks. Thanks to: Akshay Adiga, Alastair D'Silva, Alexey Kardashevskiy, Al Viro, Andrew Donnellan, Aneesh Kumar K.V, Anju T Sudhakar, Arnd Bergmann, Balbir Singh, Cédric Le Goater, Christophe Leroy, Christophe Lombard, Colin Ian King, Dave Hansen, Fabio Estevam, Finn Thain, Frederic Barrat, Gautham R. Shenoy, Haren Myneni, Hari Bathini, Ingo Molnar, Jonathan Neuschäfer, Josh Poimboeuf, Kamalesh Babulal, Madhavan Srinivasan, Mahesh Salgaonkar, Mark Greer, Mathieu Malaterre, Matthew Wilcox, Michael Neuling, Michal Suchanek, Naveen N. Rao, Nicholas Piggin, Nicolai Stange, Olof Johansson, Paul Gortmaker, Paul Mackerras, Peter Rosin, Pridhiviraj Paidipeddi, Ram Pai, Rashmica Gupta, Ravi Bangoria, Russell Currey, Sam Bobroff, Samuel Mendoza-Jonas, Segher Boessenkool, Shilpasri G Bhat, Simon Guo, Souptick Joarder, Stewart Smith, Thiago Jung Bauermann, Torsten Duwe, Vaibhav Jain, Wei Yongjun, Wolfram Sang, Yisheng Xie, YueHaibing. -----BEGIN PGP SIGNATURE----- iQIwBAABCAAaBQJbGQKBExxtcGVAZWxsZXJtYW4uaWQuYXUACgkQUevqPMjhpYBq TRAAioK7rz5xYMkxaM3Ng3ybobEeNAwQqOolz98xvmnB9SfDWNuc99vf8cGu0/fQ zc8AKZ5RcnwipOjyGlxW9oa1ZhVq0xtYnQPiYLEKMdLQmh5D+C7+KpvAd1UElweg ub40/xDySWfMujfuMSF9JDCWPIXyojt4Xg5nJKIVRrAm/3YMe/+i5Am7NWHuMCEb aQmZtlYW5Mz81XY0968hjpUO6eKFRmsaM7yFAhGTXx6+oLRpGj1PZB4AwdRIKS2L Ak7q/VgxtE4W+s3a0GK2s+eXIhGKeFuX9AVnx3nti+8/K1OqrqhDcLMUC/9JpCpv EvOtO7dxPnZujHjdu4Eai/xNoo4h6zRy7bWqve9LoBM40CP5jljKzu1lwqqb5yO0 jC7/aXhgiSIxxcRJLjoI/TYpZPu40MifrkydmczykdPyPCnMIWEJDcj4KsRL/9Y8 9SSbJzRNC/SgQNTbUYPZFFi6G0QaMmlcbCb628k8QT+Gn3Xkdf/ZtxzqEyoF4Irq 46kFBsiSSK4Bu0rVlcUtJQLgdqytWULO6NKEYnD67laxYcgQd8pGFQ8SjZhRZLgU q5LA3HIWhoAI4M0wZhOnKXO6JfiQ1UbO8gUJLsWsfF0Fk5KAcdm+4kb4jbI1H4Qk Vol9WNRZwEllyaiqScZN9RuVVuH0GPOZeEH1dtWK+uWi0lM= =ZlBf -----END PGP SIGNATURE----- Merge tag 'powerpc-4.18-1' of git://git.kernel.org/pub/scm/linux/kernel/git/powerpc/linux Pull powerpc updates from Michael Ellerman: "Notable changes: - Support for split PMD page table lock on 64-bit Book3S (Power8/9). - Add support for HAVE_RELIABLE_STACKTRACE, so we properly support live patching again. - Add support for patching barrier_nospec in copy_from_user() and syscall entry. - A couple of fixes for our data breakpoints on Book3S. - A series from Nick optimising TLB/mm handling with the Radix MMU. - Numerous small cleanups to squash sparse/gcc warnings from Mathieu Malaterre. - Several series optimising various parts of the 32-bit code from Christophe Leroy. - Removal of support for two old machines, "SBC834xE" and "C2K" ("GEFanuc,C2K"), which is why the diffstat has so many deletions. And many other small improvements & fixes. There's a few out-of-area changes. Some minor ftrace changes OK'ed by Steve, and a fix to our powernv cpuidle driver. Then there's a series touching mm, x86 and fs/proc/task_mmu.c, which cleans up some details around pkey support. It was ack'ed/reviewed by Ingo & Dave and has been in next for several weeks. Thanks to: Akshay Adiga, Alastair D'Silva, Alexey Kardashevskiy, Al Viro, Andrew Donnellan, Aneesh Kumar K.V, Anju T Sudhakar, Arnd Bergmann, Balbir Singh, Cédric Le Goater, Christophe Leroy, Christophe Lombard, Colin Ian King, Dave Hansen, Fabio Estevam, Finn Thain, Frederic Barrat, Gautham R. Shenoy, Haren Myneni, Hari Bathini, Ingo Molnar, Jonathan Neuschäfer, Josh Poimboeuf, Kamalesh Babulal, Madhavan Srinivasan, Mahesh Salgaonkar, Mark Greer, Mathieu Malaterre, Matthew Wilcox, Michael Neuling, Michal Suchanek, Naveen N. Rao, Nicholas Piggin, Nicolai Stange, Olof Johansson, Paul Gortmaker, Paul Mackerras, Peter Rosin, Pridhiviraj Paidipeddi, Ram Pai, Rashmica Gupta, Ravi Bangoria, Russell Currey, Sam Bobroff, Samuel Mendoza-Jonas, Segher Boessenkool, Shilpasri G Bhat, Simon Guo, Souptick Joarder, Stewart Smith, Thiago Jung Bauermann, Torsten Duwe, Vaibhav Jain, Wei Yongjun, Wolfram Sang, Yisheng Xie, YueHaibing" * tag 'powerpc-4.18-1' of git://git.kernel.org/pub/scm/linux/kernel/git/powerpc/linux: (251 commits) powerpc/64s/radix: Fix missing ptesync in flush_cache_vmap cpuidle: powernv: Fix promotion from snooze if next state disabled powerpc: fix build failure by disabling attribute-alias warning in pci_32 ocxl: Fix missing unlock on error in afu_ioctl_enable_p9_wait() powerpc-opal: fix spelling mistake "Uniterrupted" -> "Uninterrupted" powerpc: fix spelling mistake: "Usupported" -> "Unsupported" powerpc/pkeys: Detach execute_only key on !PROT_EXEC powerpc/powernv: copy/paste - Mask SO bit in CR powerpc: Remove core support for Marvell mv64x60 hostbridges powerpc/boot: Remove core support for Marvell mv64x60 hostbridges powerpc/boot: Remove support for Marvell mv64x60 i2c controller powerpc/boot: Remove support for Marvell MPSC serial controller powerpc/embedded6xx: Remove C2K board support powerpc/lib: optimise PPC32 memcmp powerpc/lib: optimise 32 bits __clear_user() powerpc/time: inline arch_vtime_task_switch() powerpc/Makefile: set -mcpu=860 flag for the 8xx powerpc: Implement csum_ipv6_magic in assembly powerpc/32: Optimise __csum_partial() powerpc/lib: Adjust .balign inside string functions for PPC32 ...
2563 lines
61 KiB
C
2563 lines
61 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Device driver for the via-pmu on Apple Powermacs.
|
|
*
|
|
* The VIA (versatile interface adapter) interfaces to the PMU,
|
|
* a 6805 microprocessor core whose primary function is to control
|
|
* battery charging and system power on the PowerBook 3400 and 2400.
|
|
* The PMU also controls the ADB (Apple Desktop Bus) which connects
|
|
* to the keyboard and mouse, as well as the non-volatile RAM
|
|
* and the RTC (real time clock) chip.
|
|
*
|
|
* Copyright (C) 1998 Paul Mackerras and Fabio Riccardi.
|
|
* Copyright (C) 2001-2002 Benjamin Herrenschmidt
|
|
* Copyright (C) 2006-2007 Johannes Berg
|
|
*
|
|
* THIS DRIVER IS BECOMING A TOTAL MESS !
|
|
* - Cleanup atomically disabling reply to PMU events after
|
|
* a sleep or a freq. switch
|
|
*
|
|
*/
|
|
#include <stdarg.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/types.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/sched/signal.h>
|
|
#include <linux/miscdevice.h>
|
|
#include <linux/blkdev.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/poll.h>
|
|
#include <linux/adb.h>
|
|
#include <linux/pmu.h>
|
|
#include <linux/cuda.h>
|
|
#include <linux/module.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/pm.h>
|
|
#include <linux/proc_fs.h>
|
|
#include <linux/seq_file.h>
|
|
#include <linux/init.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/device.h>
|
|
#include <linux/syscore_ops.h>
|
|
#include <linux/freezer.h>
|
|
#include <linux/syscalls.h>
|
|
#include <linux/suspend.h>
|
|
#include <linux/cpu.h>
|
|
#include <linux/compat.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/of_irq.h>
|
|
#include <asm/prom.h>
|
|
#include <asm/machdep.h>
|
|
#include <asm/io.h>
|
|
#include <asm/pgtable.h>
|
|
#include <asm/sections.h>
|
|
#include <asm/irq.h>
|
|
#include <asm/pmac_feature.h>
|
|
#include <asm/pmac_pfunc.h>
|
|
#include <asm/pmac_low_i2c.h>
|
|
#include <linux/uaccess.h>
|
|
#include <asm/mmu_context.h>
|
|
#include <asm/cputable.h>
|
|
#include <asm/time.h>
|
|
#include <asm/backlight.h>
|
|
|
|
#include "via-pmu-event.h"
|
|
|
|
/* Some compile options */
|
|
#undef DEBUG_SLEEP
|
|
|
|
/* Misc minor number allocated for /dev/pmu */
|
|
#define PMU_MINOR 154
|
|
|
|
/* How many iterations between battery polls */
|
|
#define BATTERY_POLLING_COUNT 2
|
|
|
|
static DEFINE_MUTEX(pmu_info_proc_mutex);
|
|
static volatile unsigned char __iomem *via;
|
|
|
|
/* VIA registers - spaced 0x200 bytes apart */
|
|
#define RS 0x200 /* skip between registers */
|
|
#define B 0 /* B-side data */
|
|
#define A RS /* A-side data */
|
|
#define DIRB (2*RS) /* B-side direction (1=output) */
|
|
#define DIRA (3*RS) /* A-side direction (1=output) */
|
|
#define T1CL (4*RS) /* Timer 1 ctr/latch (low 8 bits) */
|
|
#define T1CH (5*RS) /* Timer 1 counter (high 8 bits) */
|
|
#define T1LL (6*RS) /* Timer 1 latch (low 8 bits) */
|
|
#define T1LH (7*RS) /* Timer 1 latch (high 8 bits) */
|
|
#define T2CL (8*RS) /* Timer 2 ctr/latch (low 8 bits) */
|
|
#define T2CH (9*RS) /* Timer 2 counter (high 8 bits) */
|
|
#define SR (10*RS) /* Shift register */
|
|
#define ACR (11*RS) /* Auxiliary control register */
|
|
#define PCR (12*RS) /* Peripheral control register */
|
|
#define IFR (13*RS) /* Interrupt flag register */
|
|
#define IER (14*RS) /* Interrupt enable register */
|
|
#define ANH (15*RS) /* A-side data, no handshake */
|
|
|
|
/* Bits in B data register: both active low */
|
|
#define TACK 0x08 /* Transfer acknowledge (input) */
|
|
#define TREQ 0x10 /* Transfer request (output) */
|
|
|
|
/* Bits in ACR */
|
|
#define SR_CTRL 0x1c /* Shift register control bits */
|
|
#define SR_EXT 0x0c /* Shift on external clock */
|
|
#define SR_OUT 0x10 /* Shift out if 1 */
|
|
|
|
/* Bits in IFR and IER */
|
|
#define IER_SET 0x80 /* set bits in IER */
|
|
#define IER_CLR 0 /* clear bits in IER */
|
|
#define SR_INT 0x04 /* Shift register full/empty */
|
|
#define CB2_INT 0x08
|
|
#define CB1_INT 0x10 /* transition on CB1 input */
|
|
|
|
static volatile enum pmu_state {
|
|
idle,
|
|
sending,
|
|
intack,
|
|
reading,
|
|
reading_intr,
|
|
locked,
|
|
} pmu_state;
|
|
|
|
static volatile enum int_data_state {
|
|
int_data_empty,
|
|
int_data_fill,
|
|
int_data_ready,
|
|
int_data_flush
|
|
} int_data_state[2] = { int_data_empty, int_data_empty };
|
|
|
|
static struct adb_request *current_req;
|
|
static struct adb_request *last_req;
|
|
static struct adb_request *req_awaiting_reply;
|
|
static unsigned char interrupt_data[2][32];
|
|
static int interrupt_data_len[2];
|
|
static int int_data_last;
|
|
static unsigned char *reply_ptr;
|
|
static int data_index;
|
|
static int data_len;
|
|
static volatile int adb_int_pending;
|
|
static volatile int disable_poll;
|
|
static struct device_node *vias;
|
|
static int pmu_kind = PMU_UNKNOWN;
|
|
static int pmu_fully_inited;
|
|
static int pmu_has_adb;
|
|
static struct device_node *gpio_node;
|
|
static unsigned char __iomem *gpio_reg;
|
|
static int gpio_irq = 0;
|
|
static int gpio_irq_enabled = -1;
|
|
static volatile int pmu_suspended;
|
|
static spinlock_t pmu_lock;
|
|
static u8 pmu_intr_mask;
|
|
static int pmu_version;
|
|
static int drop_interrupts;
|
|
#if defined(CONFIG_SUSPEND) && defined(CONFIG_PPC32)
|
|
static int option_lid_wakeup = 1;
|
|
#endif /* CONFIG_SUSPEND && CONFIG_PPC32 */
|
|
static unsigned long async_req_locks;
|
|
static unsigned int pmu_irq_stats[11];
|
|
|
|
static struct proc_dir_entry *proc_pmu_root;
|
|
static struct proc_dir_entry *proc_pmu_info;
|
|
static struct proc_dir_entry *proc_pmu_irqstats;
|
|
static struct proc_dir_entry *proc_pmu_options;
|
|
static int option_server_mode;
|
|
|
|
int pmu_battery_count;
|
|
int pmu_cur_battery;
|
|
unsigned int pmu_power_flags = PMU_PWR_AC_PRESENT;
|
|
struct pmu_battery_info pmu_batteries[PMU_MAX_BATTERIES];
|
|
static int query_batt_timer = BATTERY_POLLING_COUNT;
|
|
static struct adb_request batt_req;
|
|
static struct proc_dir_entry *proc_pmu_batt[PMU_MAX_BATTERIES];
|
|
|
|
int __fake_sleep;
|
|
int asleep;
|
|
|
|
#ifdef CONFIG_ADB
|
|
static int adb_dev_map;
|
|
static int pmu_adb_flags;
|
|
|
|
static int pmu_probe(void);
|
|
static int pmu_init(void);
|
|
static int pmu_send_request(struct adb_request *req, int sync);
|
|
static int pmu_adb_autopoll(int devs);
|
|
static int pmu_adb_reset_bus(void);
|
|
#endif /* CONFIG_ADB */
|
|
|
|
static int init_pmu(void);
|
|
static void pmu_start(void);
|
|
static irqreturn_t via_pmu_interrupt(int irq, void *arg);
|
|
static irqreturn_t gpio1_interrupt(int irq, void *arg);
|
|
static int pmu_info_proc_show(struct seq_file *m, void *v);
|
|
static int pmu_irqstats_proc_show(struct seq_file *m, void *v);
|
|
static int pmu_battery_proc_show(struct seq_file *m, void *v);
|
|
static void pmu_pass_intr(unsigned char *data, int len);
|
|
static const struct file_operations pmu_options_proc_fops;
|
|
|
|
#ifdef CONFIG_ADB
|
|
const struct adb_driver via_pmu_driver = {
|
|
.name = "PMU",
|
|
.probe = pmu_probe,
|
|
.init = pmu_init,
|
|
.send_request = pmu_send_request,
|
|
.autopoll = pmu_adb_autopoll,
|
|
.poll = pmu_poll_adb,
|
|
.reset_bus = pmu_adb_reset_bus,
|
|
};
|
|
#endif /* CONFIG_ADB */
|
|
|
|
extern void low_sleep_handler(void);
|
|
extern void enable_kernel_altivec(void);
|
|
extern void enable_kernel_fp(void);
|
|
|
|
#ifdef DEBUG_SLEEP
|
|
int pmu_polled_request(struct adb_request *req);
|
|
void pmu_blink(int n);
|
|
#endif
|
|
|
|
/*
|
|
* This table indicates for each PMU opcode:
|
|
* - the number of data bytes to be sent with the command, or -1
|
|
* if a length byte should be sent,
|
|
* - the number of response bytes which the PMU will return, or
|
|
* -1 if it will send a length byte.
|
|
*/
|
|
static const s8 pmu_data_len[256][2] = {
|
|
/* 0 1 2 3 4 5 6 7 */
|
|
/*00*/ {-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},
|
|
/*08*/ {-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},
|
|
/*10*/ { 1, 0},{ 1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},
|
|
/*18*/ { 0, 1},{ 0, 1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{ 0, 0},
|
|
/*20*/ {-1, 0},{ 0, 0},{ 2, 0},{ 1, 0},{ 1, 0},{-1, 0},{-1, 0},{-1, 0},
|
|
/*28*/ { 0,-1},{ 0,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{ 0,-1},
|
|
/*30*/ { 4, 0},{20, 0},{-1, 0},{ 3, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},
|
|
/*38*/ { 0, 4},{ 0,20},{ 2,-1},{ 2, 1},{ 3,-1},{-1,-1},{-1,-1},{ 4, 0},
|
|
/*40*/ { 1, 0},{ 1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},
|
|
/*48*/ { 0, 1},{ 0, 1},{-1,-1},{ 1, 0},{ 1, 0},{-1,-1},{-1,-1},{-1,-1},
|
|
/*50*/ { 1, 0},{ 0, 0},{ 2, 0},{ 2, 0},{-1, 0},{ 1, 0},{ 3, 0},{ 1, 0},
|
|
/*58*/ { 0, 1},{ 1, 0},{ 0, 2},{ 0, 2},{ 0,-1},{-1,-1},{-1,-1},{-1,-1},
|
|
/*60*/ { 2, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},
|
|
/*68*/ { 0, 3},{ 0, 3},{ 0, 2},{ 0, 8},{ 0,-1},{ 0,-1},{-1,-1},{-1,-1},
|
|
/*70*/ { 1, 0},{ 1, 0},{ 1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},
|
|
/*78*/ { 0,-1},{ 0,-1},{-1,-1},{-1,-1},{-1,-1},{ 5, 1},{ 4, 1},{ 4, 1},
|
|
/*80*/ { 4, 0},{-1, 0},{ 0, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},
|
|
/*88*/ { 0, 5},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},
|
|
/*90*/ { 1, 0},{ 2, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},
|
|
/*98*/ { 0, 1},{ 0, 1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},
|
|
/*a0*/ { 2, 0},{ 2, 0},{ 2, 0},{ 4, 0},{-1, 0},{ 0, 0},{-1, 0},{-1, 0},
|
|
/*a8*/ { 1, 1},{ 1, 0},{ 3, 0},{ 2, 0},{-1,-1},{-1,-1},{-1,-1},{-1,-1},
|
|
/*b0*/ {-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},
|
|
/*b8*/ {-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},
|
|
/*c0*/ {-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},
|
|
/*c8*/ {-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},
|
|
/*d0*/ { 0, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},
|
|
/*d8*/ { 1, 1},{ 1, 1},{-1,-1},{-1,-1},{ 0, 1},{ 0,-1},{-1,-1},{-1,-1},
|
|
/*e0*/ {-1, 0},{ 4, 0},{ 0, 1},{-1, 0},{-1, 0},{ 4, 0},{-1, 0},{-1, 0},
|
|
/*e8*/ { 3,-1},{-1,-1},{ 0, 1},{-1,-1},{ 0,-1},{-1,-1},{-1,-1},{ 0, 0},
|
|
/*f0*/ {-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},{-1, 0},
|
|
/*f8*/ {-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},{-1,-1},
|
|
};
|
|
|
|
static char *pbook_type[] = {
|
|
"Unknown PowerBook",
|
|
"PowerBook 2400/3400/3500(G3)",
|
|
"PowerBook G3 Series",
|
|
"1999 PowerBook G3",
|
|
"Core99"
|
|
};
|
|
|
|
int __init find_via_pmu(void)
|
|
{
|
|
u64 taddr;
|
|
const u32 *reg;
|
|
|
|
if (via)
|
|
return 1;
|
|
vias = of_find_node_by_name(NULL, "via-pmu");
|
|
if (vias == NULL)
|
|
return 0;
|
|
|
|
reg = of_get_property(vias, "reg", NULL);
|
|
if (reg == NULL) {
|
|
printk(KERN_ERR "via-pmu: No \"reg\" property !\n");
|
|
goto fail;
|
|
}
|
|
taddr = of_translate_address(vias, reg);
|
|
if (taddr == OF_BAD_ADDR) {
|
|
printk(KERN_ERR "via-pmu: Can't translate address !\n");
|
|
goto fail;
|
|
}
|
|
|
|
spin_lock_init(&pmu_lock);
|
|
|
|
pmu_has_adb = 1;
|
|
|
|
pmu_intr_mask = PMU_INT_PCEJECT |
|
|
PMU_INT_SNDBRT |
|
|
PMU_INT_ADB |
|
|
PMU_INT_TICK;
|
|
|
|
if (vias->parent->name && ((strcmp(vias->parent->name, "ohare") == 0)
|
|
|| of_device_is_compatible(vias->parent, "ohare")))
|
|
pmu_kind = PMU_OHARE_BASED;
|
|
else if (of_device_is_compatible(vias->parent, "paddington"))
|
|
pmu_kind = PMU_PADDINGTON_BASED;
|
|
else if (of_device_is_compatible(vias->parent, "heathrow"))
|
|
pmu_kind = PMU_HEATHROW_BASED;
|
|
else if (of_device_is_compatible(vias->parent, "Keylargo")
|
|
|| of_device_is_compatible(vias->parent, "K2-Keylargo")) {
|
|
struct device_node *gpiop;
|
|
struct device_node *adbp;
|
|
u64 gaddr = OF_BAD_ADDR;
|
|
|
|
pmu_kind = PMU_KEYLARGO_BASED;
|
|
adbp = of_find_node_by_type(NULL, "adb");
|
|
pmu_has_adb = (adbp != NULL);
|
|
of_node_put(adbp);
|
|
pmu_intr_mask = PMU_INT_PCEJECT |
|
|
PMU_INT_SNDBRT |
|
|
PMU_INT_ADB |
|
|
PMU_INT_TICK |
|
|
PMU_INT_ENVIRONMENT;
|
|
|
|
gpiop = of_find_node_by_name(NULL, "gpio");
|
|
if (gpiop) {
|
|
reg = of_get_property(gpiop, "reg", NULL);
|
|
if (reg)
|
|
gaddr = of_translate_address(gpiop, reg);
|
|
if (gaddr != OF_BAD_ADDR)
|
|
gpio_reg = ioremap(gaddr, 0x10);
|
|
of_node_put(gpiop);
|
|
}
|
|
if (gpio_reg == NULL) {
|
|
printk(KERN_ERR "via-pmu: Can't find GPIO reg !\n");
|
|
goto fail;
|
|
}
|
|
} else
|
|
pmu_kind = PMU_UNKNOWN;
|
|
|
|
via = ioremap(taddr, 0x2000);
|
|
if (via == NULL) {
|
|
printk(KERN_ERR "via-pmu: Can't map address !\n");
|
|
goto fail_via_remap;
|
|
}
|
|
|
|
out_8(&via[IER], IER_CLR | 0x7f); /* disable all intrs */
|
|
out_8(&via[IFR], 0x7f); /* clear IFR */
|
|
|
|
pmu_state = idle;
|
|
|
|
if (!init_pmu())
|
|
goto fail_init;
|
|
|
|
printk(KERN_INFO "PMU driver v%d initialized for %s, firmware: %02x\n",
|
|
PMU_DRIVER_VERSION, pbook_type[pmu_kind], pmu_version);
|
|
|
|
sys_ctrler = SYS_CTRLER_PMU;
|
|
|
|
return 1;
|
|
|
|
fail_init:
|
|
iounmap(via);
|
|
via = NULL;
|
|
fail_via_remap:
|
|
iounmap(gpio_reg);
|
|
gpio_reg = NULL;
|
|
fail:
|
|
of_node_put(vias);
|
|
vias = NULL;
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_ADB
|
|
static int pmu_probe(void)
|
|
{
|
|
return vias == NULL? -ENODEV: 0;
|
|
}
|
|
|
|
static int __init pmu_init(void)
|
|
{
|
|
if (vias == NULL)
|
|
return -ENODEV;
|
|
return 0;
|
|
}
|
|
#endif /* CONFIG_ADB */
|
|
|
|
/*
|
|
* We can't wait until pmu_init gets called, that happens too late.
|
|
* It happens after IDE and SCSI initialization, which can take a few
|
|
* seconds, and by that time the PMU could have given up on us and
|
|
* turned us off.
|
|
* Thus this is called with arch_initcall rather than device_initcall.
|
|
*/
|
|
static int __init via_pmu_start(void)
|
|
{
|
|
unsigned int irq;
|
|
|
|
if (vias == NULL)
|
|
return -ENODEV;
|
|
|
|
batt_req.complete = 1;
|
|
|
|
irq = irq_of_parse_and_map(vias, 0);
|
|
if (!irq) {
|
|
printk(KERN_ERR "via-pmu: can't map interrupt\n");
|
|
return -ENODEV;
|
|
}
|
|
/* We set IRQF_NO_SUSPEND because we don't want the interrupt
|
|
* to be disabled between the 2 passes of driver suspend, we
|
|
* control our own disabling for that one
|
|
*/
|
|
if (request_irq(irq, via_pmu_interrupt, IRQF_NO_SUSPEND,
|
|
"VIA-PMU", (void *)0)) {
|
|
printk(KERN_ERR "via-pmu: can't request irq %d\n", irq);
|
|
return -ENODEV;
|
|
}
|
|
|
|
if (pmu_kind == PMU_KEYLARGO_BASED) {
|
|
gpio_node = of_find_node_by_name(NULL, "extint-gpio1");
|
|
if (gpio_node == NULL)
|
|
gpio_node = of_find_node_by_name(NULL,
|
|
"pmu-interrupt");
|
|
if (gpio_node)
|
|
gpio_irq = irq_of_parse_and_map(gpio_node, 0);
|
|
|
|
if (gpio_irq) {
|
|
if (request_irq(gpio_irq, gpio1_interrupt,
|
|
IRQF_NO_SUSPEND, "GPIO1 ADB",
|
|
(void *)0))
|
|
printk(KERN_ERR "pmu: can't get irq %d"
|
|
" (GPIO1)\n", gpio_irq);
|
|
else
|
|
gpio_irq_enabled = 1;
|
|
}
|
|
}
|
|
|
|
/* Enable interrupts */
|
|
out_8(&via[IER], IER_SET | SR_INT | CB1_INT);
|
|
|
|
pmu_fully_inited = 1;
|
|
|
|
/* Make sure PMU settle down before continuing. This is _very_ important
|
|
* since the IDE probe may shut interrupts down for quite a bit of time. If
|
|
* a PMU communication is pending while this happens, the PMU may timeout
|
|
* Not that on Core99 machines, the PMU keeps sending us environement
|
|
* messages, we should find a way to either fix IDE or make it call
|
|
* pmu_suspend() before masking interrupts. This can also happens while
|
|
* scolling with some fbdevs.
|
|
*/
|
|
do {
|
|
pmu_poll();
|
|
} while (pmu_state != idle);
|
|
|
|
return 0;
|
|
}
|
|
|
|
arch_initcall(via_pmu_start);
|
|
|
|
/*
|
|
* This has to be done after pci_init, which is a subsys_initcall.
|
|
*/
|
|
static int __init via_pmu_dev_init(void)
|
|
{
|
|
if (vias == NULL)
|
|
return -ENODEV;
|
|
|
|
#ifdef CONFIG_PMAC_BACKLIGHT
|
|
/* Initialize backlight */
|
|
pmu_backlight_init();
|
|
#endif
|
|
|
|
#ifdef CONFIG_PPC32
|
|
if (of_machine_is_compatible("AAPL,3400/2400") ||
|
|
of_machine_is_compatible("AAPL,3500")) {
|
|
int mb = pmac_call_feature(PMAC_FTR_GET_MB_INFO,
|
|
NULL, PMAC_MB_INFO_MODEL, 0);
|
|
pmu_battery_count = 1;
|
|
if (mb == PMAC_TYPE_COMET)
|
|
pmu_batteries[0].flags |= PMU_BATT_TYPE_COMET;
|
|
else
|
|
pmu_batteries[0].flags |= PMU_BATT_TYPE_HOOPER;
|
|
} else if (of_machine_is_compatible("AAPL,PowerBook1998") ||
|
|
of_machine_is_compatible("PowerBook1,1")) {
|
|
pmu_battery_count = 2;
|
|
pmu_batteries[0].flags |= PMU_BATT_TYPE_SMART;
|
|
pmu_batteries[1].flags |= PMU_BATT_TYPE_SMART;
|
|
} else {
|
|
struct device_node* prim =
|
|
of_find_node_by_name(NULL, "power-mgt");
|
|
const u32 *prim_info = NULL;
|
|
if (prim)
|
|
prim_info = of_get_property(prim, "prim-info", NULL);
|
|
if (prim_info) {
|
|
/* Other stuffs here yet unknown */
|
|
pmu_battery_count = (prim_info[6] >> 16) & 0xff;
|
|
pmu_batteries[0].flags |= PMU_BATT_TYPE_SMART;
|
|
if (pmu_battery_count > 1)
|
|
pmu_batteries[1].flags |= PMU_BATT_TYPE_SMART;
|
|
}
|
|
of_node_put(prim);
|
|
}
|
|
#endif /* CONFIG_PPC32 */
|
|
|
|
/* Create /proc/pmu */
|
|
proc_pmu_root = proc_mkdir("pmu", NULL);
|
|
if (proc_pmu_root) {
|
|
long i;
|
|
|
|
for (i=0; i<pmu_battery_count; i++) {
|
|
char title[16];
|
|
sprintf(title, "battery_%ld", i);
|
|
proc_pmu_batt[i] = proc_create_single_data(title, 0,
|
|
proc_pmu_root, pmu_battery_proc_show,
|
|
(void *)i);
|
|
}
|
|
|
|
proc_pmu_info = proc_create_single("info", 0, proc_pmu_root,
|
|
pmu_info_proc_show);
|
|
proc_pmu_irqstats = proc_create_single("interrupts", 0,
|
|
proc_pmu_root, pmu_irqstats_proc_show);
|
|
proc_pmu_options = proc_create("options", 0600, proc_pmu_root,
|
|
&pmu_options_proc_fops);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
device_initcall(via_pmu_dev_init);
|
|
|
|
static int
|
|
init_pmu(void)
|
|
{
|
|
int timeout;
|
|
struct adb_request req;
|
|
|
|
out_8(&via[B], via[B] | TREQ); /* negate TREQ */
|
|
out_8(&via[DIRB], (via[DIRB] | TREQ) & ~TACK); /* TACK in, TREQ out */
|
|
|
|
pmu_request(&req, NULL, 2, PMU_SET_INTR_MASK, pmu_intr_mask);
|
|
timeout = 100000;
|
|
while (!req.complete) {
|
|
if (--timeout < 0) {
|
|
printk(KERN_ERR "init_pmu: no response from PMU\n");
|
|
return 0;
|
|
}
|
|
udelay(10);
|
|
pmu_poll();
|
|
}
|
|
|
|
/* ack all pending interrupts */
|
|
timeout = 100000;
|
|
interrupt_data[0][0] = 1;
|
|
while (interrupt_data[0][0] || pmu_state != idle) {
|
|
if (--timeout < 0) {
|
|
printk(KERN_ERR "init_pmu: timed out acking intrs\n");
|
|
return 0;
|
|
}
|
|
if (pmu_state == idle)
|
|
adb_int_pending = 1;
|
|
via_pmu_interrupt(0, NULL);
|
|
udelay(10);
|
|
}
|
|
|
|
/* Tell PMU we are ready. */
|
|
if (pmu_kind == PMU_KEYLARGO_BASED) {
|
|
pmu_request(&req, NULL, 2, PMU_SYSTEM_READY, 2);
|
|
while (!req.complete)
|
|
pmu_poll();
|
|
}
|
|
|
|
/* Read PMU version */
|
|
pmu_request(&req, NULL, 1, PMU_GET_VERSION);
|
|
pmu_wait_complete(&req);
|
|
if (req.reply_len > 0)
|
|
pmu_version = req.reply[0];
|
|
|
|
/* Read server mode setting */
|
|
if (pmu_kind == PMU_KEYLARGO_BASED) {
|
|
pmu_request(&req, NULL, 2, PMU_POWER_EVENTS,
|
|
PMU_PWR_GET_POWERUP_EVENTS);
|
|
pmu_wait_complete(&req);
|
|
if (req.reply_len == 2) {
|
|
if (req.reply[1] & PMU_PWR_WAKEUP_AC_INSERT)
|
|
option_server_mode = 1;
|
|
printk(KERN_INFO "via-pmu: Server Mode is %s\n",
|
|
option_server_mode ? "enabled" : "disabled");
|
|
}
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
int
|
|
pmu_get_model(void)
|
|
{
|
|
return pmu_kind;
|
|
}
|
|
|
|
static void pmu_set_server_mode(int server_mode)
|
|
{
|
|
struct adb_request req;
|
|
|
|
if (pmu_kind != PMU_KEYLARGO_BASED)
|
|
return;
|
|
|
|
option_server_mode = server_mode;
|
|
pmu_request(&req, NULL, 2, PMU_POWER_EVENTS, PMU_PWR_GET_POWERUP_EVENTS);
|
|
pmu_wait_complete(&req);
|
|
if (req.reply_len < 2)
|
|
return;
|
|
if (server_mode)
|
|
pmu_request(&req, NULL, 4, PMU_POWER_EVENTS,
|
|
PMU_PWR_SET_POWERUP_EVENTS,
|
|
req.reply[0], PMU_PWR_WAKEUP_AC_INSERT);
|
|
else
|
|
pmu_request(&req, NULL, 4, PMU_POWER_EVENTS,
|
|
PMU_PWR_CLR_POWERUP_EVENTS,
|
|
req.reply[0], PMU_PWR_WAKEUP_AC_INSERT);
|
|
pmu_wait_complete(&req);
|
|
}
|
|
|
|
/* This new version of the code for 2400/3400/3500 powerbooks
|
|
* is inspired from the implementation in gkrellm-pmu
|
|
*/
|
|
static void
|
|
done_battery_state_ohare(struct adb_request* req)
|
|
{
|
|
/* format:
|
|
* [0] : flags
|
|
* 0x01 : AC indicator
|
|
* 0x02 : charging
|
|
* 0x04 : battery exist
|
|
* 0x08 :
|
|
* 0x10 :
|
|
* 0x20 : full charged
|
|
* 0x40 : pcharge reset
|
|
* 0x80 : battery exist
|
|
*
|
|
* [1][2] : battery voltage
|
|
* [3] : CPU temperature
|
|
* [4] : battery temperature
|
|
* [5] : current
|
|
* [6][7] : pcharge
|
|
* --tkoba
|
|
*/
|
|
unsigned int bat_flags = PMU_BATT_TYPE_HOOPER;
|
|
long pcharge, charge, vb, vmax, lmax;
|
|
long vmax_charging, vmax_charged;
|
|
long amperage, voltage, time, max;
|
|
int mb = pmac_call_feature(PMAC_FTR_GET_MB_INFO,
|
|
NULL, PMAC_MB_INFO_MODEL, 0);
|
|
|
|
if (req->reply[0] & 0x01)
|
|
pmu_power_flags |= PMU_PWR_AC_PRESENT;
|
|
else
|
|
pmu_power_flags &= ~PMU_PWR_AC_PRESENT;
|
|
|
|
if (mb == PMAC_TYPE_COMET) {
|
|
vmax_charged = 189;
|
|
vmax_charging = 213;
|
|
lmax = 6500;
|
|
} else {
|
|
vmax_charged = 330;
|
|
vmax_charging = 330;
|
|
lmax = 6500;
|
|
}
|
|
vmax = vmax_charged;
|
|
|
|
/* If battery installed */
|
|
if (req->reply[0] & 0x04) {
|
|
bat_flags |= PMU_BATT_PRESENT;
|
|
if (req->reply[0] & 0x02)
|
|
bat_flags |= PMU_BATT_CHARGING;
|
|
vb = (req->reply[1] << 8) | req->reply[2];
|
|
voltage = (vb * 265 + 72665) / 10;
|
|
amperage = req->reply[5];
|
|
if ((req->reply[0] & 0x01) == 0) {
|
|
if (amperage > 200)
|
|
vb += ((amperage - 200) * 15)/100;
|
|
} else if (req->reply[0] & 0x02) {
|
|
vb = (vb * 97) / 100;
|
|
vmax = vmax_charging;
|
|
}
|
|
charge = (100 * vb) / vmax;
|
|
if (req->reply[0] & 0x40) {
|
|
pcharge = (req->reply[6] << 8) + req->reply[7];
|
|
if (pcharge > lmax)
|
|
pcharge = lmax;
|
|
pcharge *= 100;
|
|
pcharge = 100 - pcharge / lmax;
|
|
if (pcharge < charge)
|
|
charge = pcharge;
|
|
}
|
|
if (amperage > 0)
|
|
time = (charge * 16440) / amperage;
|
|
else
|
|
time = 0;
|
|
max = 100;
|
|
amperage = -amperage;
|
|
} else
|
|
charge = max = amperage = voltage = time = 0;
|
|
|
|
pmu_batteries[pmu_cur_battery].flags = bat_flags;
|
|
pmu_batteries[pmu_cur_battery].charge = charge;
|
|
pmu_batteries[pmu_cur_battery].max_charge = max;
|
|
pmu_batteries[pmu_cur_battery].amperage = amperage;
|
|
pmu_batteries[pmu_cur_battery].voltage = voltage;
|
|
pmu_batteries[pmu_cur_battery].time_remaining = time;
|
|
|
|
clear_bit(0, &async_req_locks);
|
|
}
|
|
|
|
static void
|
|
done_battery_state_smart(struct adb_request* req)
|
|
{
|
|
/* format:
|
|
* [0] : format of this structure (known: 3,4,5)
|
|
* [1] : flags
|
|
*
|
|
* format 3 & 4:
|
|
*
|
|
* [2] : charge
|
|
* [3] : max charge
|
|
* [4] : current
|
|
* [5] : voltage
|
|
*
|
|
* format 5:
|
|
*
|
|
* [2][3] : charge
|
|
* [4][5] : max charge
|
|
* [6][7] : current
|
|
* [8][9] : voltage
|
|
*/
|
|
|
|
unsigned int bat_flags = PMU_BATT_TYPE_SMART;
|
|
int amperage;
|
|
unsigned int capa, max, voltage;
|
|
|
|
if (req->reply[1] & 0x01)
|
|
pmu_power_flags |= PMU_PWR_AC_PRESENT;
|
|
else
|
|
pmu_power_flags &= ~PMU_PWR_AC_PRESENT;
|
|
|
|
|
|
capa = max = amperage = voltage = 0;
|
|
|
|
if (req->reply[1] & 0x04) {
|
|
bat_flags |= PMU_BATT_PRESENT;
|
|
switch(req->reply[0]) {
|
|
case 3:
|
|
case 4: capa = req->reply[2];
|
|
max = req->reply[3];
|
|
amperage = *((signed char *)&req->reply[4]);
|
|
voltage = req->reply[5];
|
|
break;
|
|
case 5: capa = (req->reply[2] << 8) | req->reply[3];
|
|
max = (req->reply[4] << 8) | req->reply[5];
|
|
amperage = *((signed short *)&req->reply[6]);
|
|
voltage = (req->reply[8] << 8) | req->reply[9];
|
|
break;
|
|
default:
|
|
pr_warn("pmu.c: unrecognized battery info, "
|
|
"len: %d, %4ph\n", req->reply_len,
|
|
req->reply);
|
|
break;
|
|
}
|
|
}
|
|
|
|
if ((req->reply[1] & 0x01) && (amperage > 0))
|
|
bat_flags |= PMU_BATT_CHARGING;
|
|
|
|
pmu_batteries[pmu_cur_battery].flags = bat_flags;
|
|
pmu_batteries[pmu_cur_battery].charge = capa;
|
|
pmu_batteries[pmu_cur_battery].max_charge = max;
|
|
pmu_batteries[pmu_cur_battery].amperage = amperage;
|
|
pmu_batteries[pmu_cur_battery].voltage = voltage;
|
|
if (amperage) {
|
|
if ((req->reply[1] & 0x01) && (amperage > 0))
|
|
pmu_batteries[pmu_cur_battery].time_remaining
|
|
= ((max-capa) * 3600) / amperage;
|
|
else
|
|
pmu_batteries[pmu_cur_battery].time_remaining
|
|
= (capa * 3600) / (-amperage);
|
|
} else
|
|
pmu_batteries[pmu_cur_battery].time_remaining = 0;
|
|
|
|
pmu_cur_battery = (pmu_cur_battery + 1) % pmu_battery_count;
|
|
|
|
clear_bit(0, &async_req_locks);
|
|
}
|
|
|
|
static void
|
|
query_battery_state(void)
|
|
{
|
|
if (test_and_set_bit(0, &async_req_locks))
|
|
return;
|
|
if (pmu_kind == PMU_OHARE_BASED)
|
|
pmu_request(&batt_req, done_battery_state_ohare,
|
|
1, PMU_BATTERY_STATE);
|
|
else
|
|
pmu_request(&batt_req, done_battery_state_smart,
|
|
2, PMU_SMART_BATTERY_STATE, pmu_cur_battery+1);
|
|
}
|
|
|
|
static int pmu_info_proc_show(struct seq_file *m, void *v)
|
|
{
|
|
seq_printf(m, "PMU driver version : %d\n", PMU_DRIVER_VERSION);
|
|
seq_printf(m, "PMU firmware version : %02x\n", pmu_version);
|
|
seq_printf(m, "AC Power : %d\n",
|
|
((pmu_power_flags & PMU_PWR_AC_PRESENT) != 0) || pmu_battery_count == 0);
|
|
seq_printf(m, "Battery count : %d\n", pmu_battery_count);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int pmu_irqstats_proc_show(struct seq_file *m, void *v)
|
|
{
|
|
int i;
|
|
static const char *irq_names[] = {
|
|
"Total CB1 triggered events",
|
|
"Total GPIO1 triggered events",
|
|
"PC-Card eject button",
|
|
"Sound/Brightness button",
|
|
"ADB message",
|
|
"Battery state change",
|
|
"Environment interrupt",
|
|
"Tick timer",
|
|
"Ghost interrupt (zero len)",
|
|
"Empty interrupt (empty mask)",
|
|
"Max irqs in a row"
|
|
};
|
|
|
|
for (i=0; i<11; i++) {
|
|
seq_printf(m, " %2u: %10u (%s)\n",
|
|
i, pmu_irq_stats[i], irq_names[i]);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int pmu_battery_proc_show(struct seq_file *m, void *v)
|
|
{
|
|
long batnum = (long)m->private;
|
|
|
|
seq_putc(m, '\n');
|
|
seq_printf(m, "flags : %08x\n", pmu_batteries[batnum].flags);
|
|
seq_printf(m, "charge : %d\n", pmu_batteries[batnum].charge);
|
|
seq_printf(m, "max_charge : %d\n", pmu_batteries[batnum].max_charge);
|
|
seq_printf(m, "current : %d\n", pmu_batteries[batnum].amperage);
|
|
seq_printf(m, "voltage : %d\n", pmu_batteries[batnum].voltage);
|
|
seq_printf(m, "time rem. : %d\n", pmu_batteries[batnum].time_remaining);
|
|
return 0;
|
|
}
|
|
|
|
static int pmu_options_proc_show(struct seq_file *m, void *v)
|
|
{
|
|
#if defined(CONFIG_SUSPEND) && defined(CONFIG_PPC32)
|
|
if (pmu_kind == PMU_KEYLARGO_BASED &&
|
|
pmac_call_feature(PMAC_FTR_SLEEP_STATE,NULL,0,-1) >= 0)
|
|
seq_printf(m, "lid_wakeup=%d\n", option_lid_wakeup);
|
|
#endif
|
|
if (pmu_kind == PMU_KEYLARGO_BASED)
|
|
seq_printf(m, "server_mode=%d\n", option_server_mode);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int pmu_options_proc_open(struct inode *inode, struct file *file)
|
|
{
|
|
return single_open(file, pmu_options_proc_show, NULL);
|
|
}
|
|
|
|
static ssize_t pmu_options_proc_write(struct file *file,
|
|
const char __user *buffer, size_t count, loff_t *pos)
|
|
{
|
|
char tmp[33];
|
|
char *label, *val;
|
|
size_t fcount = count;
|
|
|
|
if (!count)
|
|
return -EINVAL;
|
|
if (count > 32)
|
|
count = 32;
|
|
if (copy_from_user(tmp, buffer, count))
|
|
return -EFAULT;
|
|
tmp[count] = 0;
|
|
|
|
label = tmp;
|
|
while(*label == ' ')
|
|
label++;
|
|
val = label;
|
|
while(*val && (*val != '=')) {
|
|
if (*val == ' ')
|
|
*val = 0;
|
|
val++;
|
|
}
|
|
if ((*val) == 0)
|
|
return -EINVAL;
|
|
*(val++) = 0;
|
|
while(*val == ' ')
|
|
val++;
|
|
#if defined(CONFIG_SUSPEND) && defined(CONFIG_PPC32)
|
|
if (pmu_kind == PMU_KEYLARGO_BASED &&
|
|
pmac_call_feature(PMAC_FTR_SLEEP_STATE,NULL,0,-1) >= 0)
|
|
if (!strcmp(label, "lid_wakeup"))
|
|
option_lid_wakeup = ((*val) == '1');
|
|
#endif
|
|
if (pmu_kind == PMU_KEYLARGO_BASED && !strcmp(label, "server_mode")) {
|
|
int new_value;
|
|
new_value = ((*val) == '1');
|
|
if (new_value != option_server_mode)
|
|
pmu_set_server_mode(new_value);
|
|
}
|
|
return fcount;
|
|
}
|
|
|
|
static const struct file_operations pmu_options_proc_fops = {
|
|
.owner = THIS_MODULE,
|
|
.open = pmu_options_proc_open,
|
|
.read = seq_read,
|
|
.llseek = seq_lseek,
|
|
.release = single_release,
|
|
.write = pmu_options_proc_write,
|
|
};
|
|
|
|
#ifdef CONFIG_ADB
|
|
/* Send an ADB command */
|
|
static int pmu_send_request(struct adb_request *req, int sync)
|
|
{
|
|
int i, ret;
|
|
|
|
if ((vias == NULL) || (!pmu_fully_inited)) {
|
|
req->complete = 1;
|
|
return -ENXIO;
|
|
}
|
|
|
|
ret = -EINVAL;
|
|
|
|
switch (req->data[0]) {
|
|
case PMU_PACKET:
|
|
for (i = 0; i < req->nbytes - 1; ++i)
|
|
req->data[i] = req->data[i+1];
|
|
--req->nbytes;
|
|
if (pmu_data_len[req->data[0]][1] != 0) {
|
|
req->reply[0] = ADB_RET_OK;
|
|
req->reply_len = 1;
|
|
} else
|
|
req->reply_len = 0;
|
|
ret = pmu_queue_request(req);
|
|
break;
|
|
case CUDA_PACKET:
|
|
switch (req->data[1]) {
|
|
case CUDA_GET_TIME:
|
|
if (req->nbytes != 2)
|
|
break;
|
|
req->data[0] = PMU_READ_RTC;
|
|
req->nbytes = 1;
|
|
req->reply_len = 3;
|
|
req->reply[0] = CUDA_PACKET;
|
|
req->reply[1] = 0;
|
|
req->reply[2] = CUDA_GET_TIME;
|
|
ret = pmu_queue_request(req);
|
|
break;
|
|
case CUDA_SET_TIME:
|
|
if (req->nbytes != 6)
|
|
break;
|
|
req->data[0] = PMU_SET_RTC;
|
|
req->nbytes = 5;
|
|
for (i = 1; i <= 4; ++i)
|
|
req->data[i] = req->data[i+1];
|
|
req->reply_len = 3;
|
|
req->reply[0] = CUDA_PACKET;
|
|
req->reply[1] = 0;
|
|
req->reply[2] = CUDA_SET_TIME;
|
|
ret = pmu_queue_request(req);
|
|
break;
|
|
}
|
|
break;
|
|
case ADB_PACKET:
|
|
if (!pmu_has_adb)
|
|
return -ENXIO;
|
|
for (i = req->nbytes - 1; i > 1; --i)
|
|
req->data[i+2] = req->data[i];
|
|
req->data[3] = req->nbytes - 2;
|
|
req->data[2] = pmu_adb_flags;
|
|
/*req->data[1] = req->data[1];*/
|
|
req->data[0] = PMU_ADB_CMD;
|
|
req->nbytes += 2;
|
|
req->reply_expected = 1;
|
|
req->reply_len = 0;
|
|
ret = pmu_queue_request(req);
|
|
break;
|
|
}
|
|
if (ret) {
|
|
req->complete = 1;
|
|
return ret;
|
|
}
|
|
|
|
if (sync)
|
|
while (!req->complete)
|
|
pmu_poll();
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Enable/disable autopolling */
|
|
static int __pmu_adb_autopoll(int devs)
|
|
{
|
|
struct adb_request req;
|
|
|
|
if (devs) {
|
|
pmu_request(&req, NULL, 5, PMU_ADB_CMD, 0, 0x86,
|
|
adb_dev_map >> 8, adb_dev_map);
|
|
pmu_adb_flags = 2;
|
|
} else {
|
|
pmu_request(&req, NULL, 1, PMU_ADB_POLL_OFF);
|
|
pmu_adb_flags = 0;
|
|
}
|
|
while (!req.complete)
|
|
pmu_poll();
|
|
return 0;
|
|
}
|
|
|
|
static int pmu_adb_autopoll(int devs)
|
|
{
|
|
if ((vias == NULL) || (!pmu_fully_inited) || !pmu_has_adb)
|
|
return -ENXIO;
|
|
|
|
adb_dev_map = devs;
|
|
return __pmu_adb_autopoll(devs);
|
|
}
|
|
|
|
/* Reset the ADB bus */
|
|
static int pmu_adb_reset_bus(void)
|
|
{
|
|
struct adb_request req;
|
|
int save_autopoll = adb_dev_map;
|
|
|
|
if ((vias == NULL) || (!pmu_fully_inited) || !pmu_has_adb)
|
|
return -ENXIO;
|
|
|
|
/* anyone got a better idea?? */
|
|
__pmu_adb_autopoll(0);
|
|
|
|
req.nbytes = 4;
|
|
req.done = NULL;
|
|
req.data[0] = PMU_ADB_CMD;
|
|
req.data[1] = ADB_BUSRESET;
|
|
req.data[2] = 0;
|
|
req.data[3] = 0;
|
|
req.data[4] = 0;
|
|
req.reply_len = 0;
|
|
req.reply_expected = 1;
|
|
if (pmu_queue_request(&req) != 0) {
|
|
printk(KERN_ERR "pmu_adb_reset_bus: pmu_queue_request failed\n");
|
|
return -EIO;
|
|
}
|
|
pmu_wait_complete(&req);
|
|
|
|
if (save_autopoll != 0)
|
|
__pmu_adb_autopoll(save_autopoll);
|
|
|
|
return 0;
|
|
}
|
|
#endif /* CONFIG_ADB */
|
|
|
|
/* Construct and send a pmu request */
|
|
int
|
|
pmu_request(struct adb_request *req, void (*done)(struct adb_request *),
|
|
int nbytes, ...)
|
|
{
|
|
va_list list;
|
|
int i;
|
|
|
|
if (vias == NULL)
|
|
return -ENXIO;
|
|
|
|
if (nbytes < 0 || nbytes > 32) {
|
|
printk(KERN_ERR "pmu_request: bad nbytes (%d)\n", nbytes);
|
|
req->complete = 1;
|
|
return -EINVAL;
|
|
}
|
|
req->nbytes = nbytes;
|
|
req->done = done;
|
|
va_start(list, nbytes);
|
|
for (i = 0; i < nbytes; ++i)
|
|
req->data[i] = va_arg(list, int);
|
|
va_end(list);
|
|
req->reply_len = 0;
|
|
req->reply_expected = 0;
|
|
return pmu_queue_request(req);
|
|
}
|
|
|
|
int
|
|
pmu_queue_request(struct adb_request *req)
|
|
{
|
|
unsigned long flags;
|
|
int nsend;
|
|
|
|
if (via == NULL) {
|
|
req->complete = 1;
|
|
return -ENXIO;
|
|
}
|
|
if (req->nbytes <= 0) {
|
|
req->complete = 1;
|
|
return 0;
|
|
}
|
|
nsend = pmu_data_len[req->data[0]][0];
|
|
if (nsend >= 0 && req->nbytes != nsend + 1) {
|
|
req->complete = 1;
|
|
return -EINVAL;
|
|
}
|
|
|
|
req->next = NULL;
|
|
req->sent = 0;
|
|
req->complete = 0;
|
|
|
|
spin_lock_irqsave(&pmu_lock, flags);
|
|
if (current_req) {
|
|
last_req->next = req;
|
|
last_req = req;
|
|
} else {
|
|
current_req = req;
|
|
last_req = req;
|
|
if (pmu_state == idle)
|
|
pmu_start();
|
|
}
|
|
spin_unlock_irqrestore(&pmu_lock, flags);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline void
|
|
wait_for_ack(void)
|
|
{
|
|
/* Sightly increased the delay, I had one occurrence of the message
|
|
* reported
|
|
*/
|
|
int timeout = 4000;
|
|
while ((in_8(&via[B]) & TACK) == 0) {
|
|
if (--timeout < 0) {
|
|
printk(KERN_ERR "PMU not responding (!ack)\n");
|
|
return;
|
|
}
|
|
udelay(10);
|
|
}
|
|
}
|
|
|
|
/* New PMU seems to be very sensitive to those timings, so we make sure
|
|
* PCI is flushed immediately */
|
|
static inline void
|
|
send_byte(int x)
|
|
{
|
|
volatile unsigned char __iomem *v = via;
|
|
|
|
out_8(&v[ACR], in_8(&v[ACR]) | SR_OUT | SR_EXT);
|
|
out_8(&v[SR], x);
|
|
out_8(&v[B], in_8(&v[B]) & ~TREQ); /* assert TREQ */
|
|
(void)in_8(&v[B]);
|
|
}
|
|
|
|
static inline void
|
|
recv_byte(void)
|
|
{
|
|
volatile unsigned char __iomem *v = via;
|
|
|
|
out_8(&v[ACR], (in_8(&v[ACR]) & ~SR_OUT) | SR_EXT);
|
|
in_8(&v[SR]); /* resets SR */
|
|
out_8(&v[B], in_8(&v[B]) & ~TREQ);
|
|
(void)in_8(&v[B]);
|
|
}
|
|
|
|
static inline void
|
|
pmu_done(struct adb_request *req)
|
|
{
|
|
void (*done)(struct adb_request *) = req->done;
|
|
mb();
|
|
req->complete = 1;
|
|
/* Here, we assume that if the request has a done member, the
|
|
* struct request will survive to setting req->complete to 1
|
|
*/
|
|
if (done)
|
|
(*done)(req);
|
|
}
|
|
|
|
static void
|
|
pmu_start(void)
|
|
{
|
|
struct adb_request *req;
|
|
|
|
/* assert pmu_state == idle */
|
|
/* get the packet to send */
|
|
req = current_req;
|
|
if (!req || pmu_state != idle
|
|
|| (/*req->reply_expected && */req_awaiting_reply))
|
|
return;
|
|
|
|
pmu_state = sending;
|
|
data_index = 1;
|
|
data_len = pmu_data_len[req->data[0]][0];
|
|
|
|
/* Sounds safer to make sure ACK is high before writing. This helped
|
|
* kill a problem with ADB and some iBooks
|
|
*/
|
|
wait_for_ack();
|
|
/* set the shift register to shift out and send a byte */
|
|
send_byte(req->data[0]);
|
|
}
|
|
|
|
void
|
|
pmu_poll(void)
|
|
{
|
|
if (!via)
|
|
return;
|
|
if (disable_poll)
|
|
return;
|
|
via_pmu_interrupt(0, NULL);
|
|
}
|
|
|
|
void
|
|
pmu_poll_adb(void)
|
|
{
|
|
if (!via)
|
|
return;
|
|
if (disable_poll)
|
|
return;
|
|
/* Kicks ADB read when PMU is suspended */
|
|
adb_int_pending = 1;
|
|
do {
|
|
via_pmu_interrupt(0, NULL);
|
|
} while (pmu_suspended && (adb_int_pending || pmu_state != idle
|
|
|| req_awaiting_reply));
|
|
}
|
|
|
|
void
|
|
pmu_wait_complete(struct adb_request *req)
|
|
{
|
|
if (!via)
|
|
return;
|
|
while((pmu_state != idle && pmu_state != locked) || !req->complete)
|
|
via_pmu_interrupt(0, NULL);
|
|
}
|
|
|
|
/* This function loops until the PMU is idle and prevents it from
|
|
* anwsering to ADB interrupts. pmu_request can still be called.
|
|
* This is done to avoid spurrious shutdowns when we know we'll have
|
|
* interrupts switched off for a long time
|
|
*/
|
|
void
|
|
pmu_suspend(void)
|
|
{
|
|
unsigned long flags;
|
|
|
|
if (!via)
|
|
return;
|
|
|
|
spin_lock_irqsave(&pmu_lock, flags);
|
|
pmu_suspended++;
|
|
if (pmu_suspended > 1) {
|
|
spin_unlock_irqrestore(&pmu_lock, flags);
|
|
return;
|
|
}
|
|
|
|
do {
|
|
spin_unlock_irqrestore(&pmu_lock, flags);
|
|
if (req_awaiting_reply)
|
|
adb_int_pending = 1;
|
|
via_pmu_interrupt(0, NULL);
|
|
spin_lock_irqsave(&pmu_lock, flags);
|
|
if (!adb_int_pending && pmu_state == idle && !req_awaiting_reply) {
|
|
if (gpio_irq >= 0)
|
|
disable_irq_nosync(gpio_irq);
|
|
out_8(&via[IER], CB1_INT | IER_CLR);
|
|
spin_unlock_irqrestore(&pmu_lock, flags);
|
|
break;
|
|
}
|
|
} while (1);
|
|
}
|
|
|
|
void
|
|
pmu_resume(void)
|
|
{
|
|
unsigned long flags;
|
|
|
|
if (!via || (pmu_suspended < 1))
|
|
return;
|
|
|
|
spin_lock_irqsave(&pmu_lock, flags);
|
|
pmu_suspended--;
|
|
if (pmu_suspended > 0) {
|
|
spin_unlock_irqrestore(&pmu_lock, flags);
|
|
return;
|
|
}
|
|
adb_int_pending = 1;
|
|
if (gpio_irq >= 0)
|
|
enable_irq(gpio_irq);
|
|
out_8(&via[IER], CB1_INT | IER_SET);
|
|
spin_unlock_irqrestore(&pmu_lock, flags);
|
|
pmu_poll();
|
|
}
|
|
|
|
/* Interrupt data could be the result data from an ADB cmd */
|
|
static void
|
|
pmu_handle_data(unsigned char *data, int len)
|
|
{
|
|
unsigned char ints, pirq;
|
|
int i = 0;
|
|
|
|
asleep = 0;
|
|
if (drop_interrupts || len < 1) {
|
|
adb_int_pending = 0;
|
|
pmu_irq_stats[8]++;
|
|
return;
|
|
}
|
|
|
|
/* Get PMU interrupt mask */
|
|
ints = data[0];
|
|
|
|
/* Record zero interrupts for stats */
|
|
if (ints == 0)
|
|
pmu_irq_stats[9]++;
|
|
|
|
/* Hack to deal with ADB autopoll flag */
|
|
if (ints & PMU_INT_ADB)
|
|
ints &= ~(PMU_INT_ADB_AUTO | PMU_INT_AUTO_SRQ_POLL);
|
|
|
|
next:
|
|
|
|
if (ints == 0) {
|
|
if (i > pmu_irq_stats[10])
|
|
pmu_irq_stats[10] = i;
|
|
return;
|
|
}
|
|
|
|
for (pirq = 0; pirq < 8; pirq++)
|
|
if (ints & (1 << pirq))
|
|
break;
|
|
pmu_irq_stats[pirq]++;
|
|
i++;
|
|
ints &= ~(1 << pirq);
|
|
|
|
/* Note: for some reason, we get an interrupt with len=1,
|
|
* data[0]==0 after each normal ADB interrupt, at least
|
|
* on the Pismo. Still investigating... --BenH
|
|
*/
|
|
if ((1 << pirq) & PMU_INT_ADB) {
|
|
if ((data[0] & PMU_INT_ADB_AUTO) == 0) {
|
|
struct adb_request *req = req_awaiting_reply;
|
|
if (!req) {
|
|
printk(KERN_ERR "PMU: extra ADB reply\n");
|
|
return;
|
|
}
|
|
req_awaiting_reply = NULL;
|
|
if (len <= 2)
|
|
req->reply_len = 0;
|
|
else {
|
|
memcpy(req->reply, data + 1, len - 1);
|
|
req->reply_len = len - 1;
|
|
}
|
|
pmu_done(req);
|
|
} else {
|
|
if (len == 4 && data[1] == 0x2c) {
|
|
extern int xmon_wants_key, xmon_adb_keycode;
|
|
if (xmon_wants_key) {
|
|
xmon_adb_keycode = data[2];
|
|
return;
|
|
}
|
|
}
|
|
#ifdef CONFIG_ADB
|
|
/*
|
|
* XXX On the [23]400 the PMU gives us an up
|
|
* event for keycodes 0x74 or 0x75 when the PC
|
|
* card eject buttons are released, so we
|
|
* ignore those events.
|
|
*/
|
|
if (!(pmu_kind == PMU_OHARE_BASED && len == 4
|
|
&& data[1] == 0x2c && data[3] == 0xff
|
|
&& (data[2] & ~1) == 0xf4))
|
|
adb_input(data+1, len-1, 1);
|
|
#endif /* CONFIG_ADB */
|
|
}
|
|
}
|
|
/* Sound/brightness button pressed */
|
|
else if ((1 << pirq) & PMU_INT_SNDBRT) {
|
|
#ifdef CONFIG_PMAC_BACKLIGHT
|
|
if (len == 3)
|
|
pmac_backlight_set_legacy_brightness_pmu(data[1] >> 4);
|
|
#endif
|
|
}
|
|
/* Tick interrupt */
|
|
else if ((1 << pirq) & PMU_INT_TICK) {
|
|
/* Environement or tick interrupt, query batteries */
|
|
if (pmu_battery_count) {
|
|
if ((--query_batt_timer) == 0) {
|
|
query_battery_state();
|
|
query_batt_timer = BATTERY_POLLING_COUNT;
|
|
}
|
|
}
|
|
}
|
|
else if ((1 << pirq) & PMU_INT_ENVIRONMENT) {
|
|
if (pmu_battery_count)
|
|
query_battery_state();
|
|
pmu_pass_intr(data, len);
|
|
/* len == 6 is probably a bad check. But how do I
|
|
* know what PMU versions send what events here? */
|
|
if (len == 6) {
|
|
via_pmu_event(PMU_EVT_POWER, !!(data[1]&8));
|
|
via_pmu_event(PMU_EVT_LID, data[1]&1);
|
|
}
|
|
} else {
|
|
pmu_pass_intr(data, len);
|
|
}
|
|
goto next;
|
|
}
|
|
|
|
static struct adb_request*
|
|
pmu_sr_intr(void)
|
|
{
|
|
struct adb_request *req;
|
|
int bite = 0;
|
|
|
|
if (via[B] & TREQ) {
|
|
printk(KERN_ERR "PMU: spurious SR intr (%x)\n", via[B]);
|
|
out_8(&via[IFR], SR_INT);
|
|
return NULL;
|
|
}
|
|
/* The ack may not yet be low when we get the interrupt */
|
|
while ((in_8(&via[B]) & TACK) != 0)
|
|
;
|
|
|
|
/* if reading grab the byte, and reset the interrupt */
|
|
if (pmu_state == reading || pmu_state == reading_intr)
|
|
bite = in_8(&via[SR]);
|
|
|
|
/* reset TREQ and wait for TACK to go high */
|
|
out_8(&via[B], in_8(&via[B]) | TREQ);
|
|
wait_for_ack();
|
|
|
|
switch (pmu_state) {
|
|
case sending:
|
|
req = current_req;
|
|
if (data_len < 0) {
|
|
data_len = req->nbytes - 1;
|
|
send_byte(data_len);
|
|
break;
|
|
}
|
|
if (data_index <= data_len) {
|
|
send_byte(req->data[data_index++]);
|
|
break;
|
|
}
|
|
req->sent = 1;
|
|
data_len = pmu_data_len[req->data[0]][1];
|
|
if (data_len == 0) {
|
|
pmu_state = idle;
|
|
current_req = req->next;
|
|
if (req->reply_expected)
|
|
req_awaiting_reply = req;
|
|
else
|
|
return req;
|
|
} else {
|
|
pmu_state = reading;
|
|
data_index = 0;
|
|
reply_ptr = req->reply + req->reply_len;
|
|
recv_byte();
|
|
}
|
|
break;
|
|
|
|
case intack:
|
|
data_index = 0;
|
|
data_len = -1;
|
|
pmu_state = reading_intr;
|
|
reply_ptr = interrupt_data[int_data_last];
|
|
recv_byte();
|
|
if (gpio_irq >= 0 && !gpio_irq_enabled) {
|
|
enable_irq(gpio_irq);
|
|
gpio_irq_enabled = 1;
|
|
}
|
|
break;
|
|
|
|
case reading:
|
|
case reading_intr:
|
|
if (data_len == -1) {
|
|
data_len = bite;
|
|
if (bite > 32)
|
|
printk(KERN_ERR "PMU: bad reply len %d\n", bite);
|
|
} else if (data_index < 32) {
|
|
reply_ptr[data_index++] = bite;
|
|
}
|
|
if (data_index < data_len) {
|
|
recv_byte();
|
|
break;
|
|
}
|
|
|
|
if (pmu_state == reading_intr) {
|
|
pmu_state = idle;
|
|
int_data_state[int_data_last] = int_data_ready;
|
|
interrupt_data_len[int_data_last] = data_len;
|
|
} else {
|
|
req = current_req;
|
|
/*
|
|
* For PMU sleep and freq change requests, we lock the
|
|
* PMU until it's explicitly unlocked. This avoids any
|
|
* spurrious event polling getting in
|
|
*/
|
|
current_req = req->next;
|
|
req->reply_len += data_index;
|
|
if (req->data[0] == PMU_SLEEP || req->data[0] == PMU_CPU_SPEED)
|
|
pmu_state = locked;
|
|
else
|
|
pmu_state = idle;
|
|
return req;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
printk(KERN_ERR "via_pmu_interrupt: unknown state %d?\n",
|
|
pmu_state);
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static irqreturn_t
|
|
via_pmu_interrupt(int irq, void *arg)
|
|
{
|
|
unsigned long flags;
|
|
int intr;
|
|
int nloop = 0;
|
|
int int_data = -1;
|
|
struct adb_request *req = NULL;
|
|
int handled = 0;
|
|
|
|
/* This is a bit brutal, we can probably do better */
|
|
spin_lock_irqsave(&pmu_lock, flags);
|
|
++disable_poll;
|
|
|
|
for (;;) {
|
|
intr = in_8(&via[IFR]) & (SR_INT | CB1_INT);
|
|
if (intr == 0)
|
|
break;
|
|
handled = 1;
|
|
if (++nloop > 1000) {
|
|
printk(KERN_DEBUG "PMU: stuck in intr loop, "
|
|
"intr=%x, ier=%x pmu_state=%d\n",
|
|
intr, in_8(&via[IER]), pmu_state);
|
|
break;
|
|
}
|
|
out_8(&via[IFR], intr);
|
|
if (intr & CB1_INT) {
|
|
adb_int_pending = 1;
|
|
pmu_irq_stats[0]++;
|
|
}
|
|
if (intr & SR_INT) {
|
|
req = pmu_sr_intr();
|
|
if (req)
|
|
break;
|
|
}
|
|
}
|
|
|
|
recheck:
|
|
if (pmu_state == idle) {
|
|
if (adb_int_pending) {
|
|
if (int_data_state[0] == int_data_empty)
|
|
int_data_last = 0;
|
|
else if (int_data_state[1] == int_data_empty)
|
|
int_data_last = 1;
|
|
else
|
|
goto no_free_slot;
|
|
pmu_state = intack;
|
|
int_data_state[int_data_last] = int_data_fill;
|
|
/* Sounds safer to make sure ACK is high before writing.
|
|
* This helped kill a problem with ADB and some iBooks
|
|
*/
|
|
wait_for_ack();
|
|
send_byte(PMU_INT_ACK);
|
|
adb_int_pending = 0;
|
|
} else if (current_req)
|
|
pmu_start();
|
|
}
|
|
no_free_slot:
|
|
/* Mark the oldest buffer for flushing */
|
|
if (int_data_state[!int_data_last] == int_data_ready) {
|
|
int_data_state[!int_data_last] = int_data_flush;
|
|
int_data = !int_data_last;
|
|
} else if (int_data_state[int_data_last] == int_data_ready) {
|
|
int_data_state[int_data_last] = int_data_flush;
|
|
int_data = int_data_last;
|
|
}
|
|
--disable_poll;
|
|
spin_unlock_irqrestore(&pmu_lock, flags);
|
|
|
|
/* Deal with completed PMU requests outside of the lock */
|
|
if (req) {
|
|
pmu_done(req);
|
|
req = NULL;
|
|
}
|
|
|
|
/* Deal with interrupt datas outside of the lock */
|
|
if (int_data >= 0) {
|
|
pmu_handle_data(interrupt_data[int_data], interrupt_data_len[int_data]);
|
|
spin_lock_irqsave(&pmu_lock, flags);
|
|
++disable_poll;
|
|
int_data_state[int_data] = int_data_empty;
|
|
int_data = -1;
|
|
goto recheck;
|
|
}
|
|
|
|
return IRQ_RETVAL(handled);
|
|
}
|
|
|
|
void
|
|
pmu_unlock(void)
|
|
{
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&pmu_lock, flags);
|
|
if (pmu_state == locked)
|
|
pmu_state = idle;
|
|
adb_int_pending = 1;
|
|
spin_unlock_irqrestore(&pmu_lock, flags);
|
|
}
|
|
|
|
|
|
static irqreturn_t
|
|
gpio1_interrupt(int irq, void *arg)
|
|
{
|
|
unsigned long flags;
|
|
|
|
if ((in_8(gpio_reg + 0x9) & 0x02) == 0) {
|
|
spin_lock_irqsave(&pmu_lock, flags);
|
|
if (gpio_irq_enabled > 0) {
|
|
disable_irq_nosync(gpio_irq);
|
|
gpio_irq_enabled = 0;
|
|
}
|
|
pmu_irq_stats[1]++;
|
|
adb_int_pending = 1;
|
|
spin_unlock_irqrestore(&pmu_lock, flags);
|
|
via_pmu_interrupt(0, NULL);
|
|
return IRQ_HANDLED;
|
|
}
|
|
return IRQ_NONE;
|
|
}
|
|
|
|
void
|
|
pmu_enable_irled(int on)
|
|
{
|
|
struct adb_request req;
|
|
|
|
if (vias == NULL)
|
|
return ;
|
|
if (pmu_kind == PMU_KEYLARGO_BASED)
|
|
return ;
|
|
|
|
pmu_request(&req, NULL, 2, PMU_POWER_CTRL, PMU_POW_IRLED |
|
|
(on ? PMU_POW_ON : PMU_POW_OFF));
|
|
pmu_wait_complete(&req);
|
|
}
|
|
|
|
void
|
|
pmu_restart(void)
|
|
{
|
|
struct adb_request req;
|
|
|
|
if (via == NULL)
|
|
return;
|
|
|
|
local_irq_disable();
|
|
|
|
drop_interrupts = 1;
|
|
|
|
if (pmu_kind != PMU_KEYLARGO_BASED) {
|
|
pmu_request(&req, NULL, 2, PMU_SET_INTR_MASK, PMU_INT_ADB |
|
|
PMU_INT_TICK );
|
|
while(!req.complete)
|
|
pmu_poll();
|
|
}
|
|
|
|
pmu_request(&req, NULL, 1, PMU_RESET);
|
|
pmu_wait_complete(&req);
|
|
for (;;)
|
|
;
|
|
}
|
|
|
|
void
|
|
pmu_shutdown(void)
|
|
{
|
|
struct adb_request req;
|
|
|
|
if (via == NULL)
|
|
return;
|
|
|
|
local_irq_disable();
|
|
|
|
drop_interrupts = 1;
|
|
|
|
if (pmu_kind != PMU_KEYLARGO_BASED) {
|
|
pmu_request(&req, NULL, 2, PMU_SET_INTR_MASK, PMU_INT_ADB |
|
|
PMU_INT_TICK );
|
|
pmu_wait_complete(&req);
|
|
} else {
|
|
/* Disable server mode on shutdown or we'll just
|
|
* wake up again
|
|
*/
|
|
pmu_set_server_mode(0);
|
|
}
|
|
|
|
pmu_request(&req, NULL, 5, PMU_SHUTDOWN,
|
|
'M', 'A', 'T', 'T');
|
|
pmu_wait_complete(&req);
|
|
for (;;)
|
|
;
|
|
}
|
|
|
|
int
|
|
pmu_present(void)
|
|
{
|
|
return via != NULL;
|
|
}
|
|
|
|
#if defined(CONFIG_SUSPEND) && defined(CONFIG_PPC32)
|
|
/*
|
|
* Put the powerbook to sleep.
|
|
*/
|
|
|
|
static u32 save_via[8];
|
|
|
|
static void
|
|
save_via_state(void)
|
|
{
|
|
save_via[0] = in_8(&via[ANH]);
|
|
save_via[1] = in_8(&via[DIRA]);
|
|
save_via[2] = in_8(&via[B]);
|
|
save_via[3] = in_8(&via[DIRB]);
|
|
save_via[4] = in_8(&via[PCR]);
|
|
save_via[5] = in_8(&via[ACR]);
|
|
save_via[6] = in_8(&via[T1CL]);
|
|
save_via[7] = in_8(&via[T1CH]);
|
|
}
|
|
static void
|
|
restore_via_state(void)
|
|
{
|
|
out_8(&via[ANH], save_via[0]);
|
|
out_8(&via[DIRA], save_via[1]);
|
|
out_8(&via[B], save_via[2]);
|
|
out_8(&via[DIRB], save_via[3]);
|
|
out_8(&via[PCR], save_via[4]);
|
|
out_8(&via[ACR], save_via[5]);
|
|
out_8(&via[T1CL], save_via[6]);
|
|
out_8(&via[T1CH], save_via[7]);
|
|
out_8(&via[IER], IER_CLR | 0x7f); /* disable all intrs */
|
|
out_8(&via[IFR], 0x7f); /* clear IFR */
|
|
out_8(&via[IER], IER_SET | SR_INT | CB1_INT);
|
|
}
|
|
|
|
#define GRACKLE_PM (1<<7)
|
|
#define GRACKLE_DOZE (1<<5)
|
|
#define GRACKLE_NAP (1<<4)
|
|
#define GRACKLE_SLEEP (1<<3)
|
|
|
|
static int powerbook_sleep_grackle(void)
|
|
{
|
|
unsigned long save_l2cr;
|
|
unsigned short pmcr1;
|
|
struct adb_request req;
|
|
struct pci_dev *grackle;
|
|
|
|
grackle = pci_get_domain_bus_and_slot(0, 0, 0);
|
|
if (!grackle)
|
|
return -ENODEV;
|
|
|
|
/* Turn off various things. Darwin does some retry tests here... */
|
|
pmu_request(&req, NULL, 2, PMU_POWER_CTRL0, PMU_POW0_OFF|PMU_POW0_HARD_DRIVE);
|
|
pmu_wait_complete(&req);
|
|
pmu_request(&req, NULL, 2, PMU_POWER_CTRL,
|
|
PMU_POW_OFF|PMU_POW_BACKLIGHT|PMU_POW_IRLED|PMU_POW_MEDIABAY);
|
|
pmu_wait_complete(&req);
|
|
|
|
/* For 750, save backside cache setting and disable it */
|
|
save_l2cr = _get_L2CR(); /* (returns -1 if not available) */
|
|
|
|
if (!__fake_sleep) {
|
|
/* Ask the PMU to put us to sleep */
|
|
pmu_request(&req, NULL, 5, PMU_SLEEP, 'M', 'A', 'T', 'T');
|
|
pmu_wait_complete(&req);
|
|
}
|
|
|
|
/* The VIA is supposed not to be restored correctly*/
|
|
save_via_state();
|
|
/* We shut down some HW */
|
|
pmac_call_feature(PMAC_FTR_SLEEP_STATE,NULL,0,1);
|
|
|
|
pci_read_config_word(grackle, 0x70, &pmcr1);
|
|
/* Apparently, MacOS uses NAP mode for Grackle ??? */
|
|
pmcr1 &= ~(GRACKLE_DOZE|GRACKLE_SLEEP);
|
|
pmcr1 |= GRACKLE_PM|GRACKLE_NAP;
|
|
pci_write_config_word(grackle, 0x70, pmcr1);
|
|
|
|
/* Call low-level ASM sleep handler */
|
|
if (__fake_sleep)
|
|
mdelay(5000);
|
|
else
|
|
low_sleep_handler();
|
|
|
|
/* We're awake again, stop grackle PM */
|
|
pci_read_config_word(grackle, 0x70, &pmcr1);
|
|
pmcr1 &= ~(GRACKLE_PM|GRACKLE_DOZE|GRACKLE_SLEEP|GRACKLE_NAP);
|
|
pci_write_config_word(grackle, 0x70, pmcr1);
|
|
|
|
pci_dev_put(grackle);
|
|
|
|
/* Make sure the PMU is idle */
|
|
pmac_call_feature(PMAC_FTR_SLEEP_STATE,NULL,0,0);
|
|
restore_via_state();
|
|
|
|
/* Restore L2 cache */
|
|
if (save_l2cr != 0xffffffff && (save_l2cr & L2CR_L2E) != 0)
|
|
_set_L2CR(save_l2cr);
|
|
|
|
/* Restore userland MMU context */
|
|
switch_mmu_context(NULL, current->active_mm, NULL);
|
|
|
|
/* Power things up */
|
|
pmu_unlock();
|
|
pmu_request(&req, NULL, 2, PMU_SET_INTR_MASK, pmu_intr_mask);
|
|
pmu_wait_complete(&req);
|
|
pmu_request(&req, NULL, 2, PMU_POWER_CTRL0,
|
|
PMU_POW0_ON|PMU_POW0_HARD_DRIVE);
|
|
pmu_wait_complete(&req);
|
|
pmu_request(&req, NULL, 2, PMU_POWER_CTRL,
|
|
PMU_POW_ON|PMU_POW_BACKLIGHT|PMU_POW_CHARGER|PMU_POW_IRLED|PMU_POW_MEDIABAY);
|
|
pmu_wait_complete(&req);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
powerbook_sleep_Core99(void)
|
|
{
|
|
unsigned long save_l2cr;
|
|
unsigned long save_l3cr;
|
|
struct adb_request req;
|
|
|
|
if (pmac_call_feature(PMAC_FTR_SLEEP_STATE,NULL,0,-1) < 0) {
|
|
printk(KERN_ERR "Sleep mode not supported on this machine\n");
|
|
return -ENOSYS;
|
|
}
|
|
|
|
if (num_online_cpus() > 1 || cpu_is_offline(0))
|
|
return -EAGAIN;
|
|
|
|
/* Stop environment and ADB interrupts */
|
|
pmu_request(&req, NULL, 2, PMU_SET_INTR_MASK, 0);
|
|
pmu_wait_complete(&req);
|
|
|
|
/* Tell PMU what events will wake us up */
|
|
pmu_request(&req, NULL, 4, PMU_POWER_EVENTS, PMU_PWR_CLR_WAKEUP_EVENTS,
|
|
0xff, 0xff);
|
|
pmu_wait_complete(&req);
|
|
pmu_request(&req, NULL, 4, PMU_POWER_EVENTS, PMU_PWR_SET_WAKEUP_EVENTS,
|
|
0, PMU_PWR_WAKEUP_KEY |
|
|
(option_lid_wakeup ? PMU_PWR_WAKEUP_LID_OPEN : 0));
|
|
pmu_wait_complete(&req);
|
|
|
|
/* Save the state of the L2 and L3 caches */
|
|
save_l3cr = _get_L3CR(); /* (returns -1 if not available) */
|
|
save_l2cr = _get_L2CR(); /* (returns -1 if not available) */
|
|
|
|
if (!__fake_sleep) {
|
|
/* Ask the PMU to put us to sleep */
|
|
pmu_request(&req, NULL, 5, PMU_SLEEP, 'M', 'A', 'T', 'T');
|
|
pmu_wait_complete(&req);
|
|
}
|
|
|
|
/* The VIA is supposed not to be restored correctly*/
|
|
save_via_state();
|
|
|
|
/* Shut down various ASICs. There's a chance that we can no longer
|
|
* talk to the PMU after this, so I moved it to _after_ sending the
|
|
* sleep command to it. Still need to be checked.
|
|
*/
|
|
pmac_call_feature(PMAC_FTR_SLEEP_STATE, NULL, 0, 1);
|
|
|
|
/* Call low-level ASM sleep handler */
|
|
if (__fake_sleep)
|
|
mdelay(5000);
|
|
else
|
|
low_sleep_handler();
|
|
|
|
/* Restore Apple core ASICs state */
|
|
pmac_call_feature(PMAC_FTR_SLEEP_STATE, NULL, 0, 0);
|
|
|
|
/* Restore VIA */
|
|
restore_via_state();
|
|
|
|
/* tweak LPJ before cpufreq is there */
|
|
loops_per_jiffy *= 2;
|
|
|
|
/* Restore video */
|
|
pmac_call_early_video_resume();
|
|
|
|
/* Restore L2 cache */
|
|
if (save_l2cr != 0xffffffff && (save_l2cr & L2CR_L2E) != 0)
|
|
_set_L2CR(save_l2cr);
|
|
/* Restore L3 cache */
|
|
if (save_l3cr != 0xffffffff && (save_l3cr & L3CR_L3E) != 0)
|
|
_set_L3CR(save_l3cr);
|
|
|
|
/* Restore userland MMU context */
|
|
switch_mmu_context(NULL, current->active_mm, NULL);
|
|
|
|
/* Tell PMU we are ready */
|
|
pmu_unlock();
|
|
pmu_request(&req, NULL, 2, PMU_SYSTEM_READY, 2);
|
|
pmu_wait_complete(&req);
|
|
pmu_request(&req, NULL, 2, PMU_SET_INTR_MASK, pmu_intr_mask);
|
|
pmu_wait_complete(&req);
|
|
|
|
/* Restore LPJ, cpufreq will adjust the cpu frequency */
|
|
loops_per_jiffy /= 2;
|
|
|
|
return 0;
|
|
}
|
|
|
|
#define PB3400_MEM_CTRL 0xf8000000
|
|
#define PB3400_MEM_CTRL_SLEEP 0x70
|
|
|
|
static void __iomem *pb3400_mem_ctrl;
|
|
|
|
static void powerbook_sleep_init_3400(void)
|
|
{
|
|
/* map in the memory controller registers */
|
|
pb3400_mem_ctrl = ioremap(PB3400_MEM_CTRL, 0x100);
|
|
if (pb3400_mem_ctrl == NULL)
|
|
printk(KERN_WARNING "ioremap failed: sleep won't be possible");
|
|
}
|
|
|
|
static int powerbook_sleep_3400(void)
|
|
{
|
|
int i, x;
|
|
unsigned int hid0;
|
|
unsigned long msr;
|
|
struct adb_request sleep_req;
|
|
unsigned int __iomem *mem_ctrl_sleep;
|
|
|
|
if (pb3400_mem_ctrl == NULL)
|
|
return -ENOMEM;
|
|
mem_ctrl_sleep = pb3400_mem_ctrl + PB3400_MEM_CTRL_SLEEP;
|
|
|
|
/* Set the memory controller to keep the memory refreshed
|
|
while we're asleep */
|
|
for (i = 0x403f; i >= 0x4000; --i) {
|
|
out_be32(mem_ctrl_sleep, i);
|
|
do {
|
|
x = (in_be32(mem_ctrl_sleep) >> 16) & 0x3ff;
|
|
} while (x == 0);
|
|
if (x >= 0x100)
|
|
break;
|
|
}
|
|
|
|
/* Ask the PMU to put us to sleep */
|
|
pmu_request(&sleep_req, NULL, 5, PMU_SLEEP, 'M', 'A', 'T', 'T');
|
|
pmu_wait_complete(&sleep_req);
|
|
pmu_unlock();
|
|
|
|
pmac_call_feature(PMAC_FTR_SLEEP_STATE, NULL, 0, 1);
|
|
|
|
asleep = 1;
|
|
|
|
/* Put the CPU into sleep mode */
|
|
hid0 = mfspr(SPRN_HID0);
|
|
hid0 = (hid0 & ~(HID0_NAP | HID0_DOZE)) | HID0_SLEEP;
|
|
mtspr(SPRN_HID0, hid0);
|
|
local_irq_enable();
|
|
msr = mfmsr() | MSR_POW;
|
|
while (asleep) {
|
|
mb();
|
|
mtmsr(msr);
|
|
isync();
|
|
}
|
|
local_irq_disable();
|
|
|
|
/* OK, we're awake again, start restoring things */
|
|
out_be32(mem_ctrl_sleep, 0x3f);
|
|
pmac_call_feature(PMAC_FTR_SLEEP_STATE, NULL, 0, 0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
#endif /* CONFIG_SUSPEND && CONFIG_PPC32 */
|
|
|
|
/*
|
|
* Support for /dev/pmu device
|
|
*/
|
|
#define RB_SIZE 0x10
|
|
struct pmu_private {
|
|
struct list_head list;
|
|
int rb_get;
|
|
int rb_put;
|
|
struct rb_entry {
|
|
unsigned short len;
|
|
unsigned char data[16];
|
|
} rb_buf[RB_SIZE];
|
|
wait_queue_head_t wait;
|
|
spinlock_t lock;
|
|
#if defined(CONFIG_INPUT_ADBHID) && defined(CONFIG_PMAC_BACKLIGHT)
|
|
int backlight_locker;
|
|
#endif
|
|
};
|
|
|
|
static LIST_HEAD(all_pmu_pvt);
|
|
static DEFINE_SPINLOCK(all_pvt_lock);
|
|
|
|
static void
|
|
pmu_pass_intr(unsigned char *data, int len)
|
|
{
|
|
struct pmu_private *pp;
|
|
struct list_head *list;
|
|
int i;
|
|
unsigned long flags;
|
|
|
|
if (len > sizeof(pp->rb_buf[0].data))
|
|
len = sizeof(pp->rb_buf[0].data);
|
|
spin_lock_irqsave(&all_pvt_lock, flags);
|
|
for (list = &all_pmu_pvt; (list = list->next) != &all_pmu_pvt; ) {
|
|
pp = list_entry(list, struct pmu_private, list);
|
|
spin_lock(&pp->lock);
|
|
i = pp->rb_put + 1;
|
|
if (i >= RB_SIZE)
|
|
i = 0;
|
|
if (i != pp->rb_get) {
|
|
struct rb_entry *rp = &pp->rb_buf[pp->rb_put];
|
|
rp->len = len;
|
|
memcpy(rp->data, data, len);
|
|
pp->rb_put = i;
|
|
wake_up_interruptible(&pp->wait);
|
|
}
|
|
spin_unlock(&pp->lock);
|
|
}
|
|
spin_unlock_irqrestore(&all_pvt_lock, flags);
|
|
}
|
|
|
|
static int
|
|
pmu_open(struct inode *inode, struct file *file)
|
|
{
|
|
struct pmu_private *pp;
|
|
unsigned long flags;
|
|
|
|
pp = kmalloc(sizeof(struct pmu_private), GFP_KERNEL);
|
|
if (!pp)
|
|
return -ENOMEM;
|
|
pp->rb_get = pp->rb_put = 0;
|
|
spin_lock_init(&pp->lock);
|
|
init_waitqueue_head(&pp->wait);
|
|
mutex_lock(&pmu_info_proc_mutex);
|
|
spin_lock_irqsave(&all_pvt_lock, flags);
|
|
#if defined(CONFIG_INPUT_ADBHID) && defined(CONFIG_PMAC_BACKLIGHT)
|
|
pp->backlight_locker = 0;
|
|
#endif
|
|
list_add(&pp->list, &all_pmu_pvt);
|
|
spin_unlock_irqrestore(&all_pvt_lock, flags);
|
|
file->private_data = pp;
|
|
mutex_unlock(&pmu_info_proc_mutex);
|
|
return 0;
|
|
}
|
|
|
|
static ssize_t
|
|
pmu_read(struct file *file, char __user *buf,
|
|
size_t count, loff_t *ppos)
|
|
{
|
|
struct pmu_private *pp = file->private_data;
|
|
DECLARE_WAITQUEUE(wait, current);
|
|
unsigned long flags;
|
|
int ret = 0;
|
|
|
|
if (count < 1 || !pp)
|
|
return -EINVAL;
|
|
if (!access_ok(VERIFY_WRITE, buf, count))
|
|
return -EFAULT;
|
|
|
|
spin_lock_irqsave(&pp->lock, flags);
|
|
add_wait_queue(&pp->wait, &wait);
|
|
set_current_state(TASK_INTERRUPTIBLE);
|
|
|
|
for (;;) {
|
|
ret = -EAGAIN;
|
|
if (pp->rb_get != pp->rb_put) {
|
|
int i = pp->rb_get;
|
|
struct rb_entry *rp = &pp->rb_buf[i];
|
|
ret = rp->len;
|
|
spin_unlock_irqrestore(&pp->lock, flags);
|
|
if (ret > count)
|
|
ret = count;
|
|
if (ret > 0 && copy_to_user(buf, rp->data, ret))
|
|
ret = -EFAULT;
|
|
if (++i >= RB_SIZE)
|
|
i = 0;
|
|
spin_lock_irqsave(&pp->lock, flags);
|
|
pp->rb_get = i;
|
|
}
|
|
if (ret >= 0)
|
|
break;
|
|
if (file->f_flags & O_NONBLOCK)
|
|
break;
|
|
ret = -ERESTARTSYS;
|
|
if (signal_pending(current))
|
|
break;
|
|
spin_unlock_irqrestore(&pp->lock, flags);
|
|
schedule();
|
|
spin_lock_irqsave(&pp->lock, flags);
|
|
}
|
|
__set_current_state(TASK_RUNNING);
|
|
remove_wait_queue(&pp->wait, &wait);
|
|
spin_unlock_irqrestore(&pp->lock, flags);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static ssize_t
|
|
pmu_write(struct file *file, const char __user *buf,
|
|
size_t count, loff_t *ppos)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static __poll_t
|
|
pmu_fpoll(struct file *filp, poll_table *wait)
|
|
{
|
|
struct pmu_private *pp = filp->private_data;
|
|
__poll_t mask = 0;
|
|
unsigned long flags;
|
|
|
|
if (!pp)
|
|
return 0;
|
|
poll_wait(filp, &pp->wait, wait);
|
|
spin_lock_irqsave(&pp->lock, flags);
|
|
if (pp->rb_get != pp->rb_put)
|
|
mask |= EPOLLIN;
|
|
spin_unlock_irqrestore(&pp->lock, flags);
|
|
return mask;
|
|
}
|
|
|
|
static int
|
|
pmu_release(struct inode *inode, struct file *file)
|
|
{
|
|
struct pmu_private *pp = file->private_data;
|
|
unsigned long flags;
|
|
|
|
if (pp) {
|
|
file->private_data = NULL;
|
|
spin_lock_irqsave(&all_pvt_lock, flags);
|
|
list_del(&pp->list);
|
|
spin_unlock_irqrestore(&all_pvt_lock, flags);
|
|
|
|
#if defined(CONFIG_INPUT_ADBHID) && defined(CONFIG_PMAC_BACKLIGHT)
|
|
if (pp->backlight_locker)
|
|
pmac_backlight_enable();
|
|
#endif
|
|
|
|
kfree(pp);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
#if defined(CONFIG_SUSPEND) && defined(CONFIG_PPC32)
|
|
static void pmac_suspend_disable_irqs(void)
|
|
{
|
|
/* Call platform functions marked "on sleep" */
|
|
pmac_pfunc_i2c_suspend();
|
|
pmac_pfunc_base_suspend();
|
|
}
|
|
|
|
static int powerbook_sleep(suspend_state_t state)
|
|
{
|
|
int error = 0;
|
|
|
|
/* Wait for completion of async requests */
|
|
while (!batt_req.complete)
|
|
pmu_poll();
|
|
|
|
/* Giveup the lazy FPU & vec so we don't have to back them
|
|
* up from the low level code
|
|
*/
|
|
enable_kernel_fp();
|
|
|
|
#ifdef CONFIG_ALTIVEC
|
|
if (cpu_has_feature(CPU_FTR_ALTIVEC))
|
|
enable_kernel_altivec();
|
|
#endif /* CONFIG_ALTIVEC */
|
|
|
|
switch (pmu_kind) {
|
|
case PMU_OHARE_BASED:
|
|
error = powerbook_sleep_3400();
|
|
break;
|
|
case PMU_HEATHROW_BASED:
|
|
case PMU_PADDINGTON_BASED:
|
|
error = powerbook_sleep_grackle();
|
|
break;
|
|
case PMU_KEYLARGO_BASED:
|
|
error = powerbook_sleep_Core99();
|
|
break;
|
|
default:
|
|
return -ENOSYS;
|
|
}
|
|
|
|
if (error)
|
|
return error;
|
|
|
|
mdelay(100);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void pmac_suspend_enable_irqs(void)
|
|
{
|
|
/* Force a poll of ADB interrupts */
|
|
adb_int_pending = 1;
|
|
via_pmu_interrupt(0, NULL);
|
|
|
|
mdelay(10);
|
|
|
|
/* Call platform functions marked "on wake" */
|
|
pmac_pfunc_base_resume();
|
|
pmac_pfunc_i2c_resume();
|
|
}
|
|
|
|
static int pmu_sleep_valid(suspend_state_t state)
|
|
{
|
|
return state == PM_SUSPEND_MEM
|
|
&& (pmac_call_feature(PMAC_FTR_SLEEP_STATE, NULL, 0, -1) >= 0);
|
|
}
|
|
|
|
static const struct platform_suspend_ops pmu_pm_ops = {
|
|
.enter = powerbook_sleep,
|
|
.valid = pmu_sleep_valid,
|
|
};
|
|
|
|
static int register_pmu_pm_ops(void)
|
|
{
|
|
if (pmu_kind == PMU_OHARE_BASED)
|
|
powerbook_sleep_init_3400();
|
|
ppc_md.suspend_disable_irqs = pmac_suspend_disable_irqs;
|
|
ppc_md.suspend_enable_irqs = pmac_suspend_enable_irqs;
|
|
suspend_set_ops(&pmu_pm_ops);
|
|
|
|
return 0;
|
|
}
|
|
|
|
device_initcall(register_pmu_pm_ops);
|
|
#endif
|
|
|
|
static int pmu_ioctl(struct file *filp,
|
|
u_int cmd, u_long arg)
|
|
{
|
|
__u32 __user *argp = (__u32 __user *)arg;
|
|
int error = -EINVAL;
|
|
|
|
switch (cmd) {
|
|
case PMU_IOC_SLEEP:
|
|
if (!capable(CAP_SYS_ADMIN))
|
|
return -EACCES;
|
|
return pm_suspend(PM_SUSPEND_MEM);
|
|
case PMU_IOC_CAN_SLEEP:
|
|
if (pmac_call_feature(PMAC_FTR_SLEEP_STATE, NULL, 0, -1) < 0)
|
|
return put_user(0, argp);
|
|
else
|
|
return put_user(1, argp);
|
|
|
|
#ifdef CONFIG_PMAC_BACKLIGHT_LEGACY
|
|
/* Compatibility ioctl's for backlight */
|
|
case PMU_IOC_GET_BACKLIGHT:
|
|
{
|
|
int brightness;
|
|
|
|
brightness = pmac_backlight_get_legacy_brightness();
|
|
if (brightness < 0)
|
|
return brightness;
|
|
else
|
|
return put_user(brightness, argp);
|
|
|
|
}
|
|
case PMU_IOC_SET_BACKLIGHT:
|
|
{
|
|
int brightness;
|
|
|
|
error = get_user(brightness, argp);
|
|
if (error)
|
|
return error;
|
|
|
|
return pmac_backlight_set_legacy_brightness(brightness);
|
|
}
|
|
#ifdef CONFIG_INPUT_ADBHID
|
|
case PMU_IOC_GRAB_BACKLIGHT: {
|
|
struct pmu_private *pp = filp->private_data;
|
|
|
|
if (pp->backlight_locker)
|
|
return 0;
|
|
|
|
pp->backlight_locker = 1;
|
|
pmac_backlight_disable();
|
|
|
|
return 0;
|
|
}
|
|
#endif /* CONFIG_INPUT_ADBHID */
|
|
#endif /* CONFIG_PMAC_BACKLIGHT_LEGACY */
|
|
|
|
case PMU_IOC_GET_MODEL:
|
|
return put_user(pmu_kind, argp);
|
|
case PMU_IOC_HAS_ADB:
|
|
return put_user(pmu_has_adb, argp);
|
|
}
|
|
return error;
|
|
}
|
|
|
|
static long pmu_unlocked_ioctl(struct file *filp,
|
|
u_int cmd, u_long arg)
|
|
{
|
|
int ret;
|
|
|
|
mutex_lock(&pmu_info_proc_mutex);
|
|
ret = pmu_ioctl(filp, cmd, arg);
|
|
mutex_unlock(&pmu_info_proc_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
#ifdef CONFIG_COMPAT
|
|
#define PMU_IOC_GET_BACKLIGHT32 _IOR('B', 1, compat_size_t)
|
|
#define PMU_IOC_SET_BACKLIGHT32 _IOW('B', 2, compat_size_t)
|
|
#define PMU_IOC_GET_MODEL32 _IOR('B', 3, compat_size_t)
|
|
#define PMU_IOC_HAS_ADB32 _IOR('B', 4, compat_size_t)
|
|
#define PMU_IOC_CAN_SLEEP32 _IOR('B', 5, compat_size_t)
|
|
#define PMU_IOC_GRAB_BACKLIGHT32 _IOR('B', 6, compat_size_t)
|
|
|
|
static long compat_pmu_ioctl (struct file *filp, u_int cmd, u_long arg)
|
|
{
|
|
switch (cmd) {
|
|
case PMU_IOC_SLEEP:
|
|
break;
|
|
case PMU_IOC_GET_BACKLIGHT32:
|
|
cmd = PMU_IOC_GET_BACKLIGHT;
|
|
break;
|
|
case PMU_IOC_SET_BACKLIGHT32:
|
|
cmd = PMU_IOC_SET_BACKLIGHT;
|
|
break;
|
|
case PMU_IOC_GET_MODEL32:
|
|
cmd = PMU_IOC_GET_MODEL;
|
|
break;
|
|
case PMU_IOC_HAS_ADB32:
|
|
cmd = PMU_IOC_HAS_ADB;
|
|
break;
|
|
case PMU_IOC_CAN_SLEEP32:
|
|
cmd = PMU_IOC_CAN_SLEEP;
|
|
break;
|
|
case PMU_IOC_GRAB_BACKLIGHT32:
|
|
cmd = PMU_IOC_GRAB_BACKLIGHT;
|
|
break;
|
|
default:
|
|
return -ENOIOCTLCMD;
|
|
}
|
|
return pmu_unlocked_ioctl(filp, cmd, (unsigned long)compat_ptr(arg));
|
|
}
|
|
#endif
|
|
|
|
static const struct file_operations pmu_device_fops = {
|
|
.read = pmu_read,
|
|
.write = pmu_write,
|
|
.poll = pmu_fpoll,
|
|
.unlocked_ioctl = pmu_unlocked_ioctl,
|
|
#ifdef CONFIG_COMPAT
|
|
.compat_ioctl = compat_pmu_ioctl,
|
|
#endif
|
|
.open = pmu_open,
|
|
.release = pmu_release,
|
|
.llseek = noop_llseek,
|
|
};
|
|
|
|
static struct miscdevice pmu_device = {
|
|
PMU_MINOR, "pmu", &pmu_device_fops
|
|
};
|
|
|
|
static int pmu_device_init(void)
|
|
{
|
|
if (!via)
|
|
return 0;
|
|
if (misc_register(&pmu_device) < 0)
|
|
printk(KERN_ERR "via-pmu: cannot register misc device.\n");
|
|
return 0;
|
|
}
|
|
device_initcall(pmu_device_init);
|
|
|
|
|
|
#ifdef DEBUG_SLEEP
|
|
static inline void
|
|
polled_handshake(volatile unsigned char __iomem *via)
|
|
{
|
|
via[B] &= ~TREQ; eieio();
|
|
while ((via[B] & TACK) != 0)
|
|
;
|
|
via[B] |= TREQ; eieio();
|
|
while ((via[B] & TACK) == 0)
|
|
;
|
|
}
|
|
|
|
static inline void
|
|
polled_send_byte(volatile unsigned char __iomem *via, int x)
|
|
{
|
|
via[ACR] |= SR_OUT | SR_EXT; eieio();
|
|
via[SR] = x; eieio();
|
|
polled_handshake(via);
|
|
}
|
|
|
|
static inline int
|
|
polled_recv_byte(volatile unsigned char __iomem *via)
|
|
{
|
|
int x;
|
|
|
|
via[ACR] = (via[ACR] & ~SR_OUT) | SR_EXT; eieio();
|
|
x = via[SR]; eieio();
|
|
polled_handshake(via);
|
|
x = via[SR]; eieio();
|
|
return x;
|
|
}
|
|
|
|
int
|
|
pmu_polled_request(struct adb_request *req)
|
|
{
|
|
unsigned long flags;
|
|
int i, l, c;
|
|
volatile unsigned char __iomem *v = via;
|
|
|
|
req->complete = 1;
|
|
c = req->data[0];
|
|
l = pmu_data_len[c][0];
|
|
if (l >= 0 && req->nbytes != l + 1)
|
|
return -EINVAL;
|
|
|
|
local_irq_save(flags);
|
|
while (pmu_state != idle)
|
|
pmu_poll();
|
|
|
|
while ((via[B] & TACK) == 0)
|
|
;
|
|
polled_send_byte(v, c);
|
|
if (l < 0) {
|
|
l = req->nbytes - 1;
|
|
polled_send_byte(v, l);
|
|
}
|
|
for (i = 1; i <= l; ++i)
|
|
polled_send_byte(v, req->data[i]);
|
|
|
|
l = pmu_data_len[c][1];
|
|
if (l < 0)
|
|
l = polled_recv_byte(v);
|
|
for (i = 0; i < l; ++i)
|
|
req->reply[i + req->reply_len] = polled_recv_byte(v);
|
|
|
|
if (req->done)
|
|
(*req->done)(req);
|
|
|
|
local_irq_restore(flags);
|
|
return 0;
|
|
}
|
|
|
|
/* N.B. This doesn't work on the 3400 */
|
|
void pmu_blink(int n)
|
|
{
|
|
struct adb_request req;
|
|
|
|
memset(&req, 0, sizeof(req));
|
|
|
|
for (; n > 0; --n) {
|
|
req.nbytes = 4;
|
|
req.done = NULL;
|
|
req.data[0] = 0xee;
|
|
req.data[1] = 4;
|
|
req.data[2] = 0;
|
|
req.data[3] = 1;
|
|
req.reply[0] = ADB_RET_OK;
|
|
req.reply_len = 1;
|
|
req.reply_expected = 0;
|
|
pmu_polled_request(&req);
|
|
mdelay(50);
|
|
req.nbytes = 4;
|
|
req.done = NULL;
|
|
req.data[0] = 0xee;
|
|
req.data[1] = 4;
|
|
req.data[2] = 0;
|
|
req.data[3] = 0;
|
|
req.reply[0] = ADB_RET_OK;
|
|
req.reply_len = 1;
|
|
req.reply_expected = 0;
|
|
pmu_polled_request(&req);
|
|
mdelay(50);
|
|
}
|
|
mdelay(50);
|
|
}
|
|
#endif /* DEBUG_SLEEP */
|
|
|
|
#if defined(CONFIG_SUSPEND) && defined(CONFIG_PPC32)
|
|
int pmu_sys_suspended;
|
|
|
|
static int pmu_syscore_suspend(void)
|
|
{
|
|
/* Suspend PMU event interrupts */
|
|
pmu_suspend();
|
|
pmu_sys_suspended = 1;
|
|
|
|
#ifdef CONFIG_PMAC_BACKLIGHT
|
|
/* Tell backlight code not to muck around with the chip anymore */
|
|
pmu_backlight_set_sleep(1);
|
|
#endif
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void pmu_syscore_resume(void)
|
|
{
|
|
struct adb_request req;
|
|
|
|
if (!pmu_sys_suspended)
|
|
return;
|
|
|
|
/* Tell PMU we are ready */
|
|
pmu_request(&req, NULL, 2, PMU_SYSTEM_READY, 2);
|
|
pmu_wait_complete(&req);
|
|
|
|
#ifdef CONFIG_PMAC_BACKLIGHT
|
|
/* Tell backlight code it can use the chip again */
|
|
pmu_backlight_set_sleep(0);
|
|
#endif
|
|
/* Resume PMU event interrupts */
|
|
pmu_resume();
|
|
pmu_sys_suspended = 0;
|
|
}
|
|
|
|
static struct syscore_ops pmu_syscore_ops = {
|
|
.suspend = pmu_syscore_suspend,
|
|
.resume = pmu_syscore_resume,
|
|
};
|
|
|
|
static int pmu_syscore_register(void)
|
|
{
|
|
register_syscore_ops(&pmu_syscore_ops);
|
|
|
|
return 0;
|
|
}
|
|
subsys_initcall(pmu_syscore_register);
|
|
#endif /* CONFIG_SUSPEND && CONFIG_PPC32 */
|
|
|
|
EXPORT_SYMBOL(pmu_request);
|
|
EXPORT_SYMBOL(pmu_queue_request);
|
|
EXPORT_SYMBOL(pmu_poll);
|
|
EXPORT_SYMBOL(pmu_poll_adb);
|
|
EXPORT_SYMBOL(pmu_wait_complete);
|
|
EXPORT_SYMBOL(pmu_suspend);
|
|
EXPORT_SYMBOL(pmu_resume);
|
|
EXPORT_SYMBOL(pmu_unlock);
|
|
#if defined(CONFIG_PPC32)
|
|
EXPORT_SYMBOL(pmu_enable_irled);
|
|
EXPORT_SYMBOL(pmu_battery_count);
|
|
EXPORT_SYMBOL(pmu_batteries);
|
|
EXPORT_SYMBOL(pmu_power_flags);
|
|
#endif /* CONFIG_SUSPEND && CONFIG_PPC32 */
|
|
|