forked from luck/tmp_suning_uos_patched
696ac2e3bf
Similar to commit 0266d81e9b
("acpi/processor: Prevent cpu hotplug
deadlock") except this is for acpi_processor_ffh_cstate_probe():
"The problem is that the work is scheduled on the current CPU from the
hotplug thread associated with that CPU.
It's not required to invoke these functions via the workqueue because
the hotplug thread runs on the target CPU already.
Check whether current is a per cpu thread pinned on the target CPU and
invoke the function directly to avoid the workqueue."
WARNING: possible circular locking dependency detected
------------------------------------------------------
cpuhp/1/15 is trying to acquire lock:
ffffc90003447a28 ((work_completion)(&wfc.work)){+.+.}-{0:0}, at: __flush_work+0x4c6/0x630
but task is already holding lock:
ffffffffafa1c0e8 (cpuidle_lock){+.+.}-{3:3}, at: cpuidle_pause_and_lock+0x17/0x20
which lock already depends on the new lock.
the existing dependency chain (in reverse order) is:
-> #1 (cpu_hotplug_lock){++++}-{0:0}:
cpus_read_lock+0x3e/0xc0
irq_calc_affinity_vectors+0x5f/0x91
__pci_enable_msix_range+0x10f/0x9a0
pci_alloc_irq_vectors_affinity+0x13e/0x1f0
pci_alloc_irq_vectors_affinity at drivers/pci/msi.c:1208
pqi_ctrl_init+0x72f/0x1618 [smartpqi]
pqi_pci_probe.cold.63+0x882/0x892 [smartpqi]
local_pci_probe+0x7a/0xc0
work_for_cpu_fn+0x2e/0x50
process_one_work+0x57e/0xb90
worker_thread+0x363/0x5b0
kthread+0x1f4/0x220
ret_from_fork+0x27/0x50
-> #0 ((work_completion)(&wfc.work)){+.+.}-{0:0}:
__lock_acquire+0x2244/0x32a0
lock_acquire+0x1a2/0x680
__flush_work+0x4e6/0x630
work_on_cpu+0x114/0x160
acpi_processor_ffh_cstate_probe+0x129/0x250
acpi_processor_evaluate_cst+0x4c8/0x580
acpi_processor_get_power_info+0x86/0x740
acpi_processor_hotplug+0xc3/0x140
acpi_soft_cpu_online+0x102/0x1d0
cpuhp_invoke_callback+0x197/0x1120
cpuhp_thread_fun+0x252/0x2f0
smpboot_thread_fn+0x255/0x440
kthread+0x1f4/0x220
ret_from_fork+0x27/0x50
other info that might help us debug this:
Chain exists of:
(work_completion)(&wfc.work) --> cpuhp_state-up --> cpuidle_lock
Possible unsafe locking scenario:
CPU0 CPU1
---- ----
lock(cpuidle_lock);
lock(cpuhp_state-up);
lock(cpuidle_lock);
lock((work_completion)(&wfc.work));
*** DEADLOCK ***
3 locks held by cpuhp/1/15:
#0: ffffffffaf51ab10 (cpu_hotplug_lock){++++}-{0:0}, at: cpuhp_thread_fun+0x69/0x2f0
#1: ffffffffaf51ad40 (cpuhp_state-up){+.+.}-{0:0}, at: cpuhp_thread_fun+0x69/0x2f0
#2: ffffffffafa1c0e8 (cpuidle_lock){+.+.}-{3:3}, at: cpuidle_pause_and_lock+0x17/0x20
Call Trace:
dump_stack+0xa0/0xea
print_circular_bug.cold.52+0x147/0x14c
check_noncircular+0x295/0x2d0
__lock_acquire+0x2244/0x32a0
lock_acquire+0x1a2/0x680
__flush_work+0x4e6/0x630
work_on_cpu+0x114/0x160
acpi_processor_ffh_cstate_probe+0x129/0x250
acpi_processor_evaluate_cst+0x4c8/0x580
acpi_processor_get_power_info+0x86/0x740
acpi_processor_hotplug+0xc3/0x140
acpi_soft_cpu_online+0x102/0x1d0
cpuhp_invoke_callback+0x197/0x1120
cpuhp_thread_fun+0x252/0x2f0
smpboot_thread_fn+0x255/0x440
kthread+0x1f4/0x220
ret_from_fork+0x27/0x50
Signed-off-by: Qian Cai <cai@lca.pw>
Tested-by: Borislav Petkov <bp@suse.de>
[ rjw: Subject ]
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
456 lines
12 KiB
C
456 lines
12 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef __ACPI_PROCESSOR_H
|
|
#define __ACPI_PROCESSOR_H
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/cpu.h>
|
|
#include <linux/cpufreq.h>
|
|
#include <linux/pm_qos.h>
|
|
#include <linux/thermal.h>
|
|
#include <asm/acpi.h>
|
|
|
|
#define ACPI_PROCESSOR_CLASS "processor"
|
|
#define ACPI_PROCESSOR_DEVICE_NAME "Processor"
|
|
#define ACPI_PROCESSOR_DEVICE_HID "ACPI0007"
|
|
#define ACPI_PROCESSOR_CONTAINER_HID "ACPI0010"
|
|
|
|
#define ACPI_PROCESSOR_BUSY_METRIC 10
|
|
|
|
#define ACPI_PROCESSOR_MAX_POWER 8
|
|
#define ACPI_PROCESSOR_MAX_C2_LATENCY 100
|
|
#define ACPI_PROCESSOR_MAX_C3_LATENCY 1000
|
|
|
|
#define ACPI_PROCESSOR_MAX_THROTTLING 16
|
|
#define ACPI_PROCESSOR_MAX_THROTTLE 250 /* 25% */
|
|
#define ACPI_PROCESSOR_MAX_DUTY_WIDTH 4
|
|
|
|
#define ACPI_PDC_REVISION_ID 0x1
|
|
|
|
#define ACPI_PSD_REV0_REVISION 0 /* Support for _PSD as in ACPI 3.0 */
|
|
#define ACPI_PSD_REV0_ENTRIES 5
|
|
|
|
#define ACPI_TSD_REV0_REVISION 0 /* Support for _PSD as in ACPI 3.0 */
|
|
#define ACPI_TSD_REV0_ENTRIES 5
|
|
/*
|
|
* Types of coordination defined in ACPI 3.0. Same macros can be used across
|
|
* P, C and T states
|
|
*/
|
|
#define DOMAIN_COORD_TYPE_SW_ALL 0xfc
|
|
#define DOMAIN_COORD_TYPE_SW_ANY 0xfd
|
|
#define DOMAIN_COORD_TYPE_HW_ALL 0xfe
|
|
|
|
#define ACPI_CSTATE_SYSTEMIO 0
|
|
#define ACPI_CSTATE_FFH 1
|
|
#define ACPI_CSTATE_HALT 2
|
|
#define ACPI_CSTATE_INTEGER 3
|
|
|
|
#define ACPI_CX_DESC_LEN 32
|
|
|
|
/* Power Management */
|
|
|
|
struct acpi_processor_cx;
|
|
|
|
struct acpi_power_register {
|
|
u8 descriptor;
|
|
u16 length;
|
|
u8 space_id;
|
|
u8 bit_width;
|
|
u8 bit_offset;
|
|
u8 access_size;
|
|
u64 address;
|
|
} __packed;
|
|
|
|
struct acpi_processor_cx {
|
|
u8 valid;
|
|
u8 type;
|
|
u32 address;
|
|
u8 entry_method;
|
|
u8 index;
|
|
u32 latency;
|
|
u8 bm_sts_skip;
|
|
char desc[ACPI_CX_DESC_LEN];
|
|
};
|
|
|
|
struct acpi_lpi_state {
|
|
u32 min_residency;
|
|
u32 wake_latency; /* worst case */
|
|
u32 flags;
|
|
u32 arch_flags;
|
|
u32 res_cnt_freq;
|
|
u32 enable_parent_state;
|
|
u64 address;
|
|
u8 index;
|
|
u8 entry_method;
|
|
char desc[ACPI_CX_DESC_LEN];
|
|
};
|
|
|
|
struct acpi_processor_power {
|
|
int count;
|
|
union {
|
|
struct acpi_processor_cx states[ACPI_PROCESSOR_MAX_POWER];
|
|
struct acpi_lpi_state lpi_states[ACPI_PROCESSOR_MAX_POWER];
|
|
};
|
|
int timer_broadcast_on_state;
|
|
};
|
|
|
|
/* Performance Management */
|
|
|
|
struct acpi_psd_package {
|
|
u64 num_entries;
|
|
u64 revision;
|
|
u64 domain;
|
|
u64 coord_type;
|
|
u64 num_processors;
|
|
} __packed;
|
|
|
|
struct acpi_pct_register {
|
|
u8 descriptor;
|
|
u16 length;
|
|
u8 space_id;
|
|
u8 bit_width;
|
|
u8 bit_offset;
|
|
u8 reserved;
|
|
u64 address;
|
|
} __packed;
|
|
|
|
struct acpi_processor_px {
|
|
u64 core_frequency; /* megahertz */
|
|
u64 power; /* milliWatts */
|
|
u64 transition_latency; /* microseconds */
|
|
u64 bus_master_latency; /* microseconds */
|
|
u64 control; /* control value */
|
|
u64 status; /* success indicator */
|
|
};
|
|
|
|
struct acpi_processor_performance {
|
|
unsigned int state;
|
|
unsigned int platform_limit;
|
|
struct acpi_pct_register control_register;
|
|
struct acpi_pct_register status_register;
|
|
unsigned int state_count;
|
|
struct acpi_processor_px *states;
|
|
struct acpi_psd_package domain_info;
|
|
cpumask_var_t shared_cpu_map;
|
|
unsigned int shared_type;
|
|
};
|
|
|
|
/* Throttling Control */
|
|
|
|
struct acpi_tsd_package {
|
|
u64 num_entries;
|
|
u64 revision;
|
|
u64 domain;
|
|
u64 coord_type;
|
|
u64 num_processors;
|
|
} __packed;
|
|
|
|
struct acpi_ptc_register {
|
|
u8 descriptor;
|
|
u16 length;
|
|
u8 space_id;
|
|
u8 bit_width;
|
|
u8 bit_offset;
|
|
u8 reserved;
|
|
u64 address;
|
|
} __packed;
|
|
|
|
struct acpi_processor_tx_tss {
|
|
u64 freqpercentage; /* */
|
|
u64 power; /* milliWatts */
|
|
u64 transition_latency; /* microseconds */
|
|
u64 control; /* control value */
|
|
u64 status; /* success indicator */
|
|
};
|
|
struct acpi_processor_tx {
|
|
u16 power;
|
|
u16 performance;
|
|
};
|
|
|
|
struct acpi_processor;
|
|
struct acpi_processor_throttling {
|
|
unsigned int state;
|
|
unsigned int platform_limit;
|
|
struct acpi_pct_register control_register;
|
|
struct acpi_pct_register status_register;
|
|
unsigned int state_count;
|
|
struct acpi_processor_tx_tss *states_tss;
|
|
struct acpi_tsd_package domain_info;
|
|
cpumask_var_t shared_cpu_map;
|
|
int (*acpi_processor_get_throttling) (struct acpi_processor * pr);
|
|
int (*acpi_processor_set_throttling) (struct acpi_processor * pr,
|
|
int state, bool force);
|
|
|
|
u32 address;
|
|
u8 duty_offset;
|
|
u8 duty_width;
|
|
u8 tsd_valid_flag;
|
|
unsigned int shared_type;
|
|
struct acpi_processor_tx states[ACPI_PROCESSOR_MAX_THROTTLING];
|
|
};
|
|
|
|
/* Limit Interface */
|
|
|
|
struct acpi_processor_lx {
|
|
int px; /* performance state */
|
|
int tx; /* throttle level */
|
|
};
|
|
|
|
struct acpi_processor_limit {
|
|
struct acpi_processor_lx state; /* current limit */
|
|
struct acpi_processor_lx thermal; /* thermal limit */
|
|
struct acpi_processor_lx user; /* user limit */
|
|
};
|
|
|
|
struct acpi_processor_flags {
|
|
u8 power:1;
|
|
u8 performance:1;
|
|
u8 throttling:1;
|
|
u8 limit:1;
|
|
u8 bm_control:1;
|
|
u8 bm_check:1;
|
|
u8 has_cst:1;
|
|
u8 has_lpi:1;
|
|
u8 power_setup_done:1;
|
|
u8 bm_rld_set:1;
|
|
u8 need_hotplug_init:1;
|
|
};
|
|
|
|
struct acpi_processor {
|
|
acpi_handle handle;
|
|
u32 acpi_id;
|
|
phys_cpuid_t phys_id; /* CPU hardware ID such as APIC ID for x86 */
|
|
u32 id; /* CPU logical ID allocated by OS */
|
|
u32 pblk;
|
|
int performance_platform_limit;
|
|
int throttling_platform_limit;
|
|
/* 0 - states 0..n-th state available */
|
|
|
|
struct acpi_processor_flags flags;
|
|
struct acpi_processor_power power;
|
|
struct acpi_processor_performance *performance;
|
|
struct acpi_processor_throttling throttling;
|
|
struct acpi_processor_limit limit;
|
|
struct thermal_cooling_device *cdev;
|
|
struct device *dev; /* Processor device. */
|
|
struct freq_qos_request perflib_req;
|
|
struct freq_qos_request thermal_req;
|
|
};
|
|
|
|
struct acpi_processor_errata {
|
|
u8 smp;
|
|
struct {
|
|
u8 throttle:1;
|
|
u8 fdma:1;
|
|
u8 reserved:6;
|
|
u32 bmisx;
|
|
} piix4;
|
|
};
|
|
|
|
extern int acpi_processor_preregister_performance(struct
|
|
acpi_processor_performance
|
|
__percpu *performance);
|
|
|
|
extern int acpi_processor_register_performance(struct acpi_processor_performance
|
|
*performance, unsigned int cpu);
|
|
extern void acpi_processor_unregister_performance(unsigned int cpu);
|
|
|
|
int acpi_processor_pstate_control(void);
|
|
/* note: this locks both the calling module and the processor module
|
|
if a _PPC object exists, rmmod is disallowed then */
|
|
int acpi_processor_notify_smm(struct module *calling_module);
|
|
int acpi_processor_get_psd(acpi_handle handle,
|
|
struct acpi_psd_package *pdomain);
|
|
|
|
/* parsing the _P* objects. */
|
|
extern int acpi_processor_get_performance_info(struct acpi_processor *pr);
|
|
|
|
/* for communication between multiple parts of the processor kernel module */
|
|
DECLARE_PER_CPU(struct acpi_processor *, processors);
|
|
extern struct acpi_processor_errata errata;
|
|
|
|
#if defined(ARCH_HAS_POWER_INIT) && defined(CONFIG_ACPI_PROCESSOR_CSTATE)
|
|
void acpi_processor_power_init_bm_check(struct acpi_processor_flags *flags,
|
|
unsigned int cpu);
|
|
int acpi_processor_ffh_cstate_probe(unsigned int cpu,
|
|
struct acpi_processor_cx *cx,
|
|
struct acpi_power_register *reg);
|
|
void acpi_processor_ffh_cstate_enter(struct acpi_processor_cx *cstate);
|
|
#else
|
|
static inline void acpi_processor_power_init_bm_check(struct
|
|
acpi_processor_flags
|
|
*flags, unsigned int cpu)
|
|
{
|
|
flags->bm_check = 1;
|
|
return;
|
|
}
|
|
static inline int acpi_processor_ffh_cstate_probe(unsigned int cpu,
|
|
struct acpi_processor_cx *cx,
|
|
struct acpi_power_register
|
|
*reg)
|
|
{
|
|
return -1;
|
|
}
|
|
static inline void acpi_processor_ffh_cstate_enter(struct acpi_processor_cx
|
|
*cstate)
|
|
{
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
static inline int call_on_cpu(int cpu, long (*fn)(void *), void *arg,
|
|
bool direct)
|
|
{
|
|
if (direct || (is_percpu_thread() && cpu == smp_processor_id()))
|
|
return fn(arg);
|
|
return work_on_cpu(cpu, fn, arg);
|
|
}
|
|
|
|
/* in processor_perflib.c */
|
|
|
|
#ifdef CONFIG_CPU_FREQ
|
|
extern bool acpi_processor_cpufreq_init;
|
|
void acpi_processor_ignore_ppc_init(void);
|
|
void acpi_processor_ppc_init(struct cpufreq_policy *policy);
|
|
void acpi_processor_ppc_exit(struct cpufreq_policy *policy);
|
|
void acpi_processor_ppc_has_changed(struct acpi_processor *pr, int event_flag);
|
|
extern int acpi_processor_get_bios_limit(int cpu, unsigned int *limit);
|
|
#else
|
|
static inline void acpi_processor_ignore_ppc_init(void)
|
|
{
|
|
return;
|
|
}
|
|
static inline void acpi_processor_ppc_init(struct cpufreq_policy *policy)
|
|
{
|
|
return;
|
|
}
|
|
static inline void acpi_processor_ppc_exit(struct cpufreq_policy *policy)
|
|
{
|
|
return;
|
|
}
|
|
static inline void acpi_processor_ppc_has_changed(struct acpi_processor *pr,
|
|
int event_flag)
|
|
{
|
|
static unsigned int printout = 1;
|
|
if (printout) {
|
|
printk(KERN_WARNING
|
|
"Warning: Processor Platform Limit event detected, but not handled.\n");
|
|
printk(KERN_WARNING
|
|
"Consider compiling CPUfreq support into your kernel.\n");
|
|
printout = 0;
|
|
}
|
|
}
|
|
static inline int acpi_processor_get_bios_limit(int cpu, unsigned int *limit)
|
|
{
|
|
return -ENODEV;
|
|
}
|
|
|
|
#endif /* CONFIG_CPU_FREQ */
|
|
|
|
/* in processor_core.c */
|
|
phys_cpuid_t acpi_get_phys_id(acpi_handle, int type, u32 acpi_id);
|
|
phys_cpuid_t acpi_map_madt_entry(u32 acpi_id);
|
|
int acpi_map_cpuid(phys_cpuid_t phys_id, u32 acpi_id);
|
|
int acpi_get_cpuid(acpi_handle, int type, u32 acpi_id);
|
|
|
|
#ifdef CONFIG_ACPI_CPPC_LIB
|
|
extern int acpi_cppc_processor_probe(struct acpi_processor *pr);
|
|
extern void acpi_cppc_processor_exit(struct acpi_processor *pr);
|
|
#else
|
|
static inline int acpi_cppc_processor_probe(struct acpi_processor *pr)
|
|
{
|
|
return 0;
|
|
}
|
|
static inline void acpi_cppc_processor_exit(struct acpi_processor *pr)
|
|
{
|
|
return;
|
|
}
|
|
#endif /* CONFIG_ACPI_CPPC_LIB */
|
|
|
|
/* in processor_pdc.c */
|
|
void acpi_processor_set_pdc(acpi_handle handle);
|
|
|
|
/* in processor_throttling.c */
|
|
#ifdef CONFIG_ACPI_CPU_FREQ_PSS
|
|
int acpi_processor_tstate_has_changed(struct acpi_processor *pr);
|
|
int acpi_processor_get_throttling_info(struct acpi_processor *pr);
|
|
extern int acpi_processor_set_throttling(struct acpi_processor *pr,
|
|
int state, bool force);
|
|
/*
|
|
* Reevaluate whether the T-state is invalid after one cpu is
|
|
* onlined/offlined. In such case the flags.throttling will be updated.
|
|
*/
|
|
extern void acpi_processor_reevaluate_tstate(struct acpi_processor *pr,
|
|
bool is_dead);
|
|
extern const struct file_operations acpi_processor_throttling_fops;
|
|
extern void acpi_processor_throttling_init(void);
|
|
#else
|
|
static inline int acpi_processor_tstate_has_changed(struct acpi_processor *pr)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline int acpi_processor_get_throttling_info(struct acpi_processor *pr)
|
|
{
|
|
return -ENODEV;
|
|
}
|
|
|
|
static inline int acpi_processor_set_throttling(struct acpi_processor *pr,
|
|
int state, bool force)
|
|
{
|
|
return -ENODEV;
|
|
}
|
|
|
|
static inline void acpi_processor_reevaluate_tstate(struct acpi_processor *pr,
|
|
bool is_dead) {}
|
|
|
|
static inline void acpi_processor_throttling_init(void) {}
|
|
#endif /* CONFIG_ACPI_CPU_FREQ_PSS */
|
|
|
|
/* in processor_idle.c */
|
|
extern struct cpuidle_driver acpi_idle_driver;
|
|
#ifdef CONFIG_ACPI_PROCESSOR_IDLE
|
|
int acpi_processor_power_init(struct acpi_processor *pr);
|
|
int acpi_processor_power_exit(struct acpi_processor *pr);
|
|
int acpi_processor_power_state_has_changed(struct acpi_processor *pr);
|
|
int acpi_processor_hotplug(struct acpi_processor *pr);
|
|
#else
|
|
static inline int acpi_processor_power_init(struct acpi_processor *pr)
|
|
{
|
|
return -ENODEV;
|
|
}
|
|
|
|
static inline int acpi_processor_power_exit(struct acpi_processor *pr)
|
|
{
|
|
return -ENODEV;
|
|
}
|
|
|
|
static inline int acpi_processor_power_state_has_changed(struct acpi_processor *pr)
|
|
{
|
|
return -ENODEV;
|
|
}
|
|
|
|
static inline int acpi_processor_hotplug(struct acpi_processor *pr)
|
|
{
|
|
return -ENODEV;
|
|
}
|
|
#endif /* CONFIG_ACPI_PROCESSOR_IDLE */
|
|
|
|
/* in processor_thermal.c */
|
|
int acpi_processor_get_limit_info(struct acpi_processor *pr);
|
|
extern const struct thermal_cooling_device_ops processor_cooling_ops;
|
|
#if defined(CONFIG_ACPI_CPU_FREQ_PSS) & defined(CONFIG_CPU_FREQ)
|
|
void acpi_thermal_cpufreq_init(struct cpufreq_policy *policy);
|
|
void acpi_thermal_cpufreq_exit(struct cpufreq_policy *policy);
|
|
#else
|
|
static inline void acpi_thermal_cpufreq_init(struct cpufreq_policy *policy)
|
|
{
|
|
return;
|
|
}
|
|
static inline void acpi_thermal_cpufreq_exit(struct cpufreq_policy *policy)
|
|
{
|
|
return;
|
|
}
|
|
#endif /* CONFIG_ACPI_CPU_FREQ_PSS */
|
|
|
|
#endif
|