Miscellaneous ia64 changes for 3.11 merge window
-----BEGIN PGP SIGNATURE----- Version: GnuPG v1.4.12 (GNU/Linux) iQIcBAABAgAGBQJRzg4kAAoJEKurIx+X31iBGWkP/RflCjUdJ11l8wj/KTpbiP1R jW0Jx3T8SIqMEtDRVgXRT9Q51lpz1rmHmFBd2+Gw8H0FkoOT/8glyU3rBNW+gSK0 +14M/ebfiBNj4X6pIzKEGbVr8TgpajjRXvjjQCgyQIxBgvvv+hZ2Ui8XIcRiAk6d Ikme6YICti8ikgY18onsY6t7kI9zWclj0opt4uPKn4fkulyFWQJszLgTOx1EbFRp WltW/coCzigCbZRYOhdJpszwII73ItfyFCvk8CCuQDzqQ0PwdSE6lbIJRyT/UqHi f7HUxb6j6Nai5gAhXJ3OeJ8PlxBRNwiYFvgRCKLzsiX+74vdeIkCV0scwwbAYORO 8qE+btgw7npQ8zlRuE6mOIJ9/zIkBm7S0aWDZmPv5fMGMYAL0D/ZYMMvXl/IYdPu NovS86unw7c/lz1+VEfzFeQb3K8aIYpApup7pVxAZ70+dvCyM7IfHjngQNsCbkU9 u8n/yuhuZXzTDo/gq/ctV/A10DFGXoJVHxMyjsaN8bhvrLSb/OBaJL/hljB1VEsF +IJsnpi/V7VS917mM2F3tfar5khlINLcoYFrOkkYGJY+Zr28llqozsJ3rQQXKF2g LdFQM0Cl7UQaQtQtV47n+9F0AsExjJWn0G5asn5Mh29y7RmGU+iEQ8QDbhRNYCO4 51+FOWHIArCrA2lchCz9 =hiGp -----END PGP SIGNATURE----- Merge tag 'please-pull-misc-3.11' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux Pull misc ia64 updates from Tony Luck: "Miscellaneous ia64 changes for 3.11 merge window" * tag 'please-pull-misc-3.11' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux: [IA64] Delete __cpuinit usage from all ia64 users [IA64] hpsim: Fix check for overlong simscsi prefix. [IA64] pci: Remove unused fallback_dev [IA64] perfmon: Use %*phD specifier to dump small buffers [IA64] Fix trap #45 handlinghifive-unleashed-5.1
commit
e39dfe52f8
|
@ -88,8 +88,8 @@ simscsi_setup (char *s)
|
||||||
if (strlen(s) > MAX_ROOT_LEN) {
|
if (strlen(s) > MAX_ROOT_LEN) {
|
||||||
printk(KERN_ERR "simscsi_setup: prefix too long---using default %s\n",
|
printk(KERN_ERR "simscsi_setup: prefix too long---using default %s\n",
|
||||||
simscsi_root);
|
simscsi_root);
|
||||||
}
|
} else
|
||||||
simscsi_root = s;
|
simscsi_root = s;
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -807,7 +807,7 @@ int acpi_isa_irq_to_gsi(unsigned isa_irq, u32 *gsi)
|
||||||
* ACPI based hotplug CPU support
|
* ACPI based hotplug CPU support
|
||||||
*/
|
*/
|
||||||
#ifdef CONFIG_ACPI_HOTPLUG_CPU
|
#ifdef CONFIG_ACPI_HOTPLUG_CPU
|
||||||
static __cpuinit
|
static
|
||||||
int acpi_map_cpu2node(acpi_handle handle, int cpu, int physid)
|
int acpi_map_cpu2node(acpi_handle handle, int cpu, int physid)
|
||||||
{
|
{
|
||||||
#ifdef CONFIG_ACPI_NUMA
|
#ifdef CONFIG_ACPI_NUMA
|
||||||
|
@ -882,7 +882,7 @@ __init void prefill_possible_map(void)
|
||||||
set_cpu_possible(i, true);
|
set_cpu_possible(i, true);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int __cpuinit _acpi_map_lsapic(acpi_handle handle, int *pcpu)
|
static int _acpi_map_lsapic(acpi_handle handle, int *pcpu)
|
||||||
{
|
{
|
||||||
struct acpi_buffer buffer = { ACPI_ALLOCATE_BUFFER, NULL };
|
struct acpi_buffer buffer = { ACPI_ALLOCATE_BUFFER, NULL };
|
||||||
union acpi_object *obj;
|
union acpi_object *obj;
|
||||||
|
|
|
@ -225,17 +225,17 @@ static struct attribute_group err_inject_attr_group = {
|
||||||
.name = "err_inject"
|
.name = "err_inject"
|
||||||
};
|
};
|
||||||
/* Add/Remove err_inject interface for CPU device */
|
/* Add/Remove err_inject interface for CPU device */
|
||||||
static int __cpuinit err_inject_add_dev(struct device * sys_dev)
|
static int err_inject_add_dev(struct device *sys_dev)
|
||||||
{
|
{
|
||||||
return sysfs_create_group(&sys_dev->kobj, &err_inject_attr_group);
|
return sysfs_create_group(&sys_dev->kobj, &err_inject_attr_group);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int __cpuinit err_inject_remove_dev(struct device * sys_dev)
|
static int err_inject_remove_dev(struct device *sys_dev)
|
||||||
{
|
{
|
||||||
sysfs_remove_group(&sys_dev->kobj, &err_inject_attr_group);
|
sysfs_remove_group(&sys_dev->kobj, &err_inject_attr_group);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
static int __cpuinit err_inject_cpu_callback(struct notifier_block *nfb,
|
static int err_inject_cpu_callback(struct notifier_block *nfb,
|
||||||
unsigned long action, void *hcpu)
|
unsigned long action, void *hcpu)
|
||||||
{
|
{
|
||||||
unsigned int cpu = (unsigned long)hcpu;
|
unsigned int cpu = (unsigned long)hcpu;
|
||||||
|
@ -256,7 +256,7 @@ static int __cpuinit err_inject_cpu_callback(struct notifier_block *nfb,
|
||||||
return NOTIFY_OK;
|
return NOTIFY_OK;
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct notifier_block __cpuinitdata err_inject_cpu_notifier =
|
static struct notifier_block err_inject_cpu_notifier =
|
||||||
{
|
{
|
||||||
.notifier_call = err_inject_cpu_callback,
|
.notifier_call = err_inject_cpu_callback,
|
||||||
};
|
};
|
||||||
|
|
|
@ -631,7 +631,7 @@ ia64_mca_register_cpev (int cpev)
|
||||||
* Outputs
|
* Outputs
|
||||||
* None
|
* None
|
||||||
*/
|
*/
|
||||||
void __cpuinit
|
void
|
||||||
ia64_mca_cmc_vector_setup (void)
|
ia64_mca_cmc_vector_setup (void)
|
||||||
{
|
{
|
||||||
cmcv_reg_t cmcv;
|
cmcv_reg_t cmcv;
|
||||||
|
@ -1814,7 +1814,7 @@ static struct irqaction mca_cpep_irqaction = {
|
||||||
* format most of the fields.
|
* format most of the fields.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
static void __cpuinit
|
static void
|
||||||
format_mca_init_stack(void *mca_data, unsigned long offset,
|
format_mca_init_stack(void *mca_data, unsigned long offset,
|
||||||
const char *type, int cpu)
|
const char *type, int cpu)
|
||||||
{
|
{
|
||||||
|
@ -1844,7 +1844,7 @@ static void * __init_refok mca_bootmem(void)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Do per-CPU MCA-related initialization. */
|
/* Do per-CPU MCA-related initialization. */
|
||||||
void __cpuinit
|
void
|
||||||
ia64_mca_cpu_init(void *cpu_data)
|
ia64_mca_cpu_init(void *cpu_data)
|
||||||
{
|
{
|
||||||
void *pal_vaddr;
|
void *pal_vaddr;
|
||||||
|
@ -1896,7 +1896,7 @@ ia64_mca_cpu_init(void *cpu_data)
|
||||||
PAGE_KERNEL));
|
PAGE_KERNEL));
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __cpuinit ia64_mca_cmc_vector_adjust(void *dummy)
|
static void ia64_mca_cmc_vector_adjust(void *dummy)
|
||||||
{
|
{
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
|
||||||
|
@ -1906,7 +1906,7 @@ static void __cpuinit ia64_mca_cmc_vector_adjust(void *dummy)
|
||||||
local_irq_restore(flags);
|
local_irq_restore(flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int __cpuinit mca_cpu_callback(struct notifier_block *nfb,
|
static int mca_cpu_callback(struct notifier_block *nfb,
|
||||||
unsigned long action,
|
unsigned long action,
|
||||||
void *hcpu)
|
void *hcpu)
|
||||||
{
|
{
|
||||||
|
@ -1922,7 +1922,7 @@ static int __cpuinit mca_cpu_callback(struct notifier_block *nfb,
|
||||||
return NOTIFY_OK;
|
return NOTIFY_OK;
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct notifier_block mca_cpu_notifier __cpuinitdata = {
|
static struct notifier_block mca_cpu_notifier = {
|
||||||
.notifier_call = mca_cpu_callback
|
.notifier_call = mca_cpu_callback
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -30,7 +30,7 @@ EXPORT_SYMBOL(cpu_to_node_map);
|
||||||
cpumask_t node_to_cpu_mask[MAX_NUMNODES] __cacheline_aligned;
|
cpumask_t node_to_cpu_mask[MAX_NUMNODES] __cacheline_aligned;
|
||||||
EXPORT_SYMBOL(node_to_cpu_mask);
|
EXPORT_SYMBOL(node_to_cpu_mask);
|
||||||
|
|
||||||
void __cpuinit map_cpu_to_node(int cpu, int nid)
|
void map_cpu_to_node(int cpu, int nid)
|
||||||
{
|
{
|
||||||
int oldnid;
|
int oldnid;
|
||||||
if (nid < 0) { /* just initialize by zero */
|
if (nid < 0) { /* just initialize by zero */
|
||||||
|
@ -51,7 +51,7 @@ void __cpuinit map_cpu_to_node(int cpu, int nid)
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
void __cpuinit unmap_cpu_from_node(int cpu, int nid)
|
void unmap_cpu_from_node(int cpu, int nid)
|
||||||
{
|
{
|
||||||
WARN_ON(!cpu_isset(cpu, node_to_cpu_mask[nid]));
|
WARN_ON(!cpu_isset(cpu, node_to_cpu_mask[nid]));
|
||||||
WARN_ON(cpu_to_node_map[cpu] != nid);
|
WARN_ON(cpu_to_node_map[cpu] != nid);
|
||||||
|
|
|
@ -932,7 +932,7 @@ static const struct file_operations proc_palinfo_fops = {
|
||||||
.release = single_release,
|
.release = single_release,
|
||||||
};
|
};
|
||||||
|
|
||||||
static void __cpuinit
|
static void
|
||||||
create_palinfo_proc_entries(unsigned int cpu)
|
create_palinfo_proc_entries(unsigned int cpu)
|
||||||
{
|
{
|
||||||
pal_func_cpu_u_t f;
|
pal_func_cpu_u_t f;
|
||||||
|
@ -962,7 +962,7 @@ remove_palinfo_proc_entries(unsigned int hcpu)
|
||||||
remove_proc_subtree(cpustr, palinfo_dir);
|
remove_proc_subtree(cpustr, palinfo_dir);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int __cpuinit palinfo_cpu_callback(struct notifier_block *nfb,
|
static int palinfo_cpu_callback(struct notifier_block *nfb,
|
||||||
unsigned long action, void *hcpu)
|
unsigned long action, void *hcpu)
|
||||||
{
|
{
|
||||||
unsigned int hotcpu = (unsigned long)hcpu;
|
unsigned int hotcpu = (unsigned long)hcpu;
|
||||||
|
|
|
@ -33,15 +33,6 @@ int force_iommu __read_mostly;
|
||||||
|
|
||||||
int iommu_pass_through;
|
int iommu_pass_through;
|
||||||
|
|
||||||
/* Dummy device used for NULL arguments (normally ISA). Better would
|
|
||||||
be probably a smaller DMA mask, but this is bug-to-bug compatible
|
|
||||||
to i386. */
|
|
||||||
struct device fallback_dev = {
|
|
||||||
.init_name = "fallback device",
|
|
||||||
.coherent_dma_mask = DMA_BIT_MASK(32),
|
|
||||||
.dma_mask = &fallback_dev.coherent_dma_mask,
|
|
||||||
};
|
|
||||||
|
|
||||||
extern struct dma_map_ops intel_dma_ops;
|
extern struct dma_map_ops intel_dma_ops;
|
||||||
|
|
||||||
static int __init pci_iommu_init(void)
|
static int __init pci_iommu_init(void)
|
||||||
|
|
|
@ -5647,24 +5647,8 @@ pfm_proc_show_header(struct seq_file *m)
|
||||||
|
|
||||||
list_for_each(pos, &pfm_buffer_fmt_list) {
|
list_for_each(pos, &pfm_buffer_fmt_list) {
|
||||||
entry = list_entry(pos, pfm_buffer_fmt_t, fmt_list);
|
entry = list_entry(pos, pfm_buffer_fmt_t, fmt_list);
|
||||||
seq_printf(m, "format : %02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x %s\n",
|
seq_printf(m, "format : %16phD %s\n",
|
||||||
entry->fmt_uuid[0],
|
entry->fmt_uuid, entry->fmt_name);
|
||||||
entry->fmt_uuid[1],
|
|
||||||
entry->fmt_uuid[2],
|
|
||||||
entry->fmt_uuid[3],
|
|
||||||
entry->fmt_uuid[4],
|
|
||||||
entry->fmt_uuid[5],
|
|
||||||
entry->fmt_uuid[6],
|
|
||||||
entry->fmt_uuid[7],
|
|
||||||
entry->fmt_uuid[8],
|
|
||||||
entry->fmt_uuid[9],
|
|
||||||
entry->fmt_uuid[10],
|
|
||||||
entry->fmt_uuid[11],
|
|
||||||
entry->fmt_uuid[12],
|
|
||||||
entry->fmt_uuid[13],
|
|
||||||
entry->fmt_uuid[14],
|
|
||||||
entry->fmt_uuid[15],
|
|
||||||
entry->fmt_name);
|
|
||||||
}
|
}
|
||||||
spin_unlock(&pfm_buffer_fmt_lock);
|
spin_unlock(&pfm_buffer_fmt_lock);
|
||||||
|
|
||||||
|
|
|
@ -568,7 +568,7 @@ static const struct file_operations salinfo_data_fops = {
|
||||||
.llseek = default_llseek,
|
.llseek = default_llseek,
|
||||||
};
|
};
|
||||||
|
|
||||||
static int __cpuinit
|
static int
|
||||||
salinfo_cpu_callback(struct notifier_block *nb, unsigned long action, void *hcpu)
|
salinfo_cpu_callback(struct notifier_block *nb, unsigned long action, void *hcpu)
|
||||||
{
|
{
|
||||||
unsigned int i, cpu = (unsigned long)hcpu;
|
unsigned int i, cpu = (unsigned long)hcpu;
|
||||||
|
@ -609,7 +609,7 @@ salinfo_cpu_callback(struct notifier_block *nb, unsigned long action, void *hcpu
|
||||||
return NOTIFY_OK;
|
return NOTIFY_OK;
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct notifier_block salinfo_cpu_notifier __cpuinitdata =
|
static struct notifier_block salinfo_cpu_notifier =
|
||||||
{
|
{
|
||||||
.notifier_call = salinfo_cpu_callback,
|
.notifier_call = salinfo_cpu_callback,
|
||||||
.priority = 0,
|
.priority = 0,
|
||||||
|
|
|
@ -748,7 +748,7 @@ const struct seq_operations cpuinfo_op = {
|
||||||
#define MAX_BRANDS 8
|
#define MAX_BRANDS 8
|
||||||
static char brandname[MAX_BRANDS][128];
|
static char brandname[MAX_BRANDS][128];
|
||||||
|
|
||||||
static char * __cpuinit
|
static char *
|
||||||
get_model_name(__u8 family, __u8 model)
|
get_model_name(__u8 family, __u8 model)
|
||||||
{
|
{
|
||||||
static int overflow;
|
static int overflow;
|
||||||
|
@ -778,7 +778,7 @@ get_model_name(__u8 family, __u8 model)
|
||||||
return "Unknown";
|
return "Unknown";
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __cpuinit
|
static void
|
||||||
identify_cpu (struct cpuinfo_ia64 *c)
|
identify_cpu (struct cpuinfo_ia64 *c)
|
||||||
{
|
{
|
||||||
union {
|
union {
|
||||||
|
@ -850,7 +850,7 @@ identify_cpu (struct cpuinfo_ia64 *c)
|
||||||
* 2. the minimum of the i-cache stride sizes for "flush_icache_range()".
|
* 2. the minimum of the i-cache stride sizes for "flush_icache_range()".
|
||||||
* 3. the minimum of the cache stride sizes for "clflush_cache_range()".
|
* 3. the minimum of the cache stride sizes for "clflush_cache_range()".
|
||||||
*/
|
*/
|
||||||
static void __cpuinit
|
static void
|
||||||
get_cache_info(void)
|
get_cache_info(void)
|
||||||
{
|
{
|
||||||
unsigned long line_size, max = 1;
|
unsigned long line_size, max = 1;
|
||||||
|
@ -915,10 +915,10 @@ get_cache_info(void)
|
||||||
* cpu_init() initializes state that is per-CPU. This function acts
|
* cpu_init() initializes state that is per-CPU. This function acts
|
||||||
* as a 'CPU state barrier', nothing should get across.
|
* as a 'CPU state barrier', nothing should get across.
|
||||||
*/
|
*/
|
||||||
void __cpuinit
|
void
|
||||||
cpu_init (void)
|
cpu_init (void)
|
||||||
{
|
{
|
||||||
extern void __cpuinit ia64_mmu_init (void *);
|
extern void ia64_mmu_init(void *);
|
||||||
static unsigned long max_num_phys_stacked = IA64_NUM_PHYS_STACK_REG;
|
static unsigned long max_num_phys_stacked = IA64_NUM_PHYS_STACK_REG;
|
||||||
unsigned long num_phys_stacked;
|
unsigned long num_phys_stacked;
|
||||||
pal_vm_info_2_u_t vmi;
|
pal_vm_info_2_u_t vmi;
|
||||||
|
|
|
@ -351,7 +351,7 @@ static inline void smp_setup_percpu_timer(void)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __cpuinit
|
static void
|
||||||
smp_callin (void)
|
smp_callin (void)
|
||||||
{
|
{
|
||||||
int cpuid, phys_id, itc_master;
|
int cpuid, phys_id, itc_master;
|
||||||
|
@ -442,7 +442,7 @@ smp_callin (void)
|
||||||
/*
|
/*
|
||||||
* Activate a secondary processor. head.S calls this.
|
* Activate a secondary processor. head.S calls this.
|
||||||
*/
|
*/
|
||||||
int __cpuinit
|
int
|
||||||
start_secondary (void *unused)
|
start_secondary (void *unused)
|
||||||
{
|
{
|
||||||
/* Early console may use I/O ports */
|
/* Early console may use I/O ports */
|
||||||
|
@ -459,7 +459,7 @@ start_secondary (void *unused)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int __cpuinit
|
static int
|
||||||
do_boot_cpu (int sapicid, int cpu, struct task_struct *idle)
|
do_boot_cpu (int sapicid, int cpu, struct task_struct *idle)
|
||||||
{
|
{
|
||||||
int timeout;
|
int timeout;
|
||||||
|
@ -728,7 +728,7 @@ static inline void set_cpu_sibling_map(int cpu)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
int __cpuinit
|
int
|
||||||
__cpu_up(unsigned int cpu, struct task_struct *tidle)
|
__cpu_up(unsigned int cpu, struct task_struct *tidle)
|
||||||
{
|
{
|
||||||
int ret;
|
int ret;
|
||||||
|
|
|
@ -135,11 +135,11 @@ struct cpu_cache_info {
|
||||||
struct kobject kobj;
|
struct kobject kobj;
|
||||||
};
|
};
|
||||||
|
|
||||||
static struct cpu_cache_info all_cpu_cache_info[NR_CPUS] __cpuinitdata;
|
static struct cpu_cache_info all_cpu_cache_info[NR_CPUS];
|
||||||
#define LEAF_KOBJECT_PTR(x,y) (&all_cpu_cache_info[x].cache_leaves[y])
|
#define LEAF_KOBJECT_PTR(x,y) (&all_cpu_cache_info[x].cache_leaves[y])
|
||||||
|
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
static void __cpuinit cache_shared_cpu_map_setup( unsigned int cpu,
|
static void cache_shared_cpu_map_setup(unsigned int cpu,
|
||||||
struct cache_info * this_leaf)
|
struct cache_info * this_leaf)
|
||||||
{
|
{
|
||||||
pal_cache_shared_info_t csi;
|
pal_cache_shared_info_t csi;
|
||||||
|
@ -174,7 +174,7 @@ static void __cpuinit cache_shared_cpu_map_setup( unsigned int cpu,
|
||||||
&csi) == PAL_STATUS_SUCCESS);
|
&csi) == PAL_STATUS_SUCCESS);
|
||||||
}
|
}
|
||||||
#else
|
#else
|
||||||
static void __cpuinit cache_shared_cpu_map_setup(unsigned int cpu,
|
static void cache_shared_cpu_map_setup(unsigned int cpu,
|
||||||
struct cache_info * this_leaf)
|
struct cache_info * this_leaf)
|
||||||
{
|
{
|
||||||
cpu_set(cpu, this_leaf->shared_cpu_map);
|
cpu_set(cpu, this_leaf->shared_cpu_map);
|
||||||
|
@ -298,7 +298,7 @@ static struct kobj_type cache_ktype_percpu_entry = {
|
||||||
.sysfs_ops = &cache_sysfs_ops,
|
.sysfs_ops = &cache_sysfs_ops,
|
||||||
};
|
};
|
||||||
|
|
||||||
static void __cpuinit cpu_cache_sysfs_exit(unsigned int cpu)
|
static void cpu_cache_sysfs_exit(unsigned int cpu)
|
||||||
{
|
{
|
||||||
kfree(all_cpu_cache_info[cpu].cache_leaves);
|
kfree(all_cpu_cache_info[cpu].cache_leaves);
|
||||||
all_cpu_cache_info[cpu].cache_leaves = NULL;
|
all_cpu_cache_info[cpu].cache_leaves = NULL;
|
||||||
|
@ -307,7 +307,7 @@ static void __cpuinit cpu_cache_sysfs_exit(unsigned int cpu)
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int __cpuinit cpu_cache_sysfs_init(unsigned int cpu)
|
static int cpu_cache_sysfs_init(unsigned int cpu)
|
||||||
{
|
{
|
||||||
unsigned long i, levels, unique_caches;
|
unsigned long i, levels, unique_caches;
|
||||||
pal_cache_config_info_t cci;
|
pal_cache_config_info_t cci;
|
||||||
|
@ -351,7 +351,7 @@ static int __cpuinit cpu_cache_sysfs_init(unsigned int cpu)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Add cache interface for CPU device */
|
/* Add cache interface for CPU device */
|
||||||
static int __cpuinit cache_add_dev(struct device * sys_dev)
|
static int cache_add_dev(struct device *sys_dev)
|
||||||
{
|
{
|
||||||
unsigned int cpu = sys_dev->id;
|
unsigned int cpu = sys_dev->id;
|
||||||
unsigned long i, j;
|
unsigned long i, j;
|
||||||
|
@ -401,7 +401,7 @@ static int __cpuinit cache_add_dev(struct device * sys_dev)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Remove cache interface for CPU device */
|
/* Remove cache interface for CPU device */
|
||||||
static int __cpuinit cache_remove_dev(struct device * sys_dev)
|
static int cache_remove_dev(struct device *sys_dev)
|
||||||
{
|
{
|
||||||
unsigned int cpu = sys_dev->id;
|
unsigned int cpu = sys_dev->id;
|
||||||
unsigned long i;
|
unsigned long i;
|
||||||
|
@ -425,7 +425,7 @@ static int __cpuinit cache_remove_dev(struct device * sys_dev)
|
||||||
* When a cpu is hot-plugged, do a check and initiate
|
* When a cpu is hot-plugged, do a check and initiate
|
||||||
* cache kobject if necessary
|
* cache kobject if necessary
|
||||||
*/
|
*/
|
||||||
static int __cpuinit cache_cpu_callback(struct notifier_block *nfb,
|
static int cache_cpu_callback(struct notifier_block *nfb,
|
||||||
unsigned long action, void *hcpu)
|
unsigned long action, void *hcpu)
|
||||||
{
|
{
|
||||||
unsigned int cpu = (unsigned long)hcpu;
|
unsigned int cpu = (unsigned long)hcpu;
|
||||||
|
@ -445,7 +445,7 @@ static int __cpuinit cache_cpu_callback(struct notifier_block *nfb,
|
||||||
return NOTIFY_OK;
|
return NOTIFY_OK;
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct notifier_block __cpuinitdata cache_cpu_notifier =
|
static struct notifier_block cache_cpu_notifier =
|
||||||
{
|
{
|
||||||
.notifier_call = cache_cpu_callback
|
.notifier_call = cache_cpu_callback
|
||||||
};
|
};
|
||||||
|
|
|
@ -630,7 +630,7 @@ ia64_fault (unsigned long vector, unsigned long isr, unsigned long ifa,
|
||||||
printk(KERN_ERR " iip - 0x%lx, ifa - 0x%lx, isr - 0x%lx\n",
|
printk(KERN_ERR " iip - 0x%lx, ifa - 0x%lx, isr - 0x%lx\n",
|
||||||
iip, ifa, isr);
|
iip, ifa, isr);
|
||||||
force_sig(SIGSEGV, current);
|
force_sig(SIGSEGV, current);
|
||||||
break;
|
return;
|
||||||
|
|
||||||
case 46:
|
case 46:
|
||||||
printk(KERN_ERR "Unexpected IA-32 intercept trap (Trap 46)\n");
|
printk(KERN_ERR "Unexpected IA-32 intercept trap (Trap 46)\n");
|
||||||
|
|
|
@ -156,8 +156,7 @@ static void *cpu_data;
|
||||||
*
|
*
|
||||||
* Allocate and setup per-cpu data areas.
|
* Allocate and setup per-cpu data areas.
|
||||||
*/
|
*/
|
||||||
void * __cpuinit
|
void *per_cpu_init(void)
|
||||||
per_cpu_init (void)
|
|
||||||
{
|
{
|
||||||
static bool first_time = true;
|
static bool first_time = true;
|
||||||
void *cpu0_data = __cpu0_per_cpu;
|
void *cpu0_data = __cpu0_per_cpu;
|
||||||
|
|
|
@ -593,7 +593,7 @@ void __init find_memory(void)
|
||||||
* find_pernode_space() does most of this already, we just need to set
|
* find_pernode_space() does most of this already, we just need to set
|
||||||
* local_per_cpu_offset
|
* local_per_cpu_offset
|
||||||
*/
|
*/
|
||||||
void __cpuinit *per_cpu_init(void)
|
void *per_cpu_init(void)
|
||||||
{
|
{
|
||||||
int cpu;
|
int cpu;
|
||||||
static int first_time = 1;
|
static int first_time = 1;
|
||||||
|
|
|
@ -86,7 +86,7 @@ int __meminit __early_pfn_to_nid(unsigned long pfn)
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
void __cpuinit numa_clear_node(int cpu)
|
void numa_clear_node(int cpu)
|
||||||
{
|
{
|
||||||
unmap_cpu_from_node(cpu, NUMA_NO_NODE);
|
unmap_cpu_from_node(cpu, NUMA_NO_NODE);
|
||||||
}
|
}
|
||||||
|
|
|
@ -192,7 +192,7 @@ void __init early_sn_setup(void)
|
||||||
}
|
}
|
||||||
|
|
||||||
extern int platform_intr_list[];
|
extern int platform_intr_list[];
|
||||||
static int __cpuinitdata shub_1_1_found;
|
static int shub_1_1_found;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* sn_check_for_wars
|
* sn_check_for_wars
|
||||||
|
@ -200,7 +200,7 @@ static int __cpuinitdata shub_1_1_found;
|
||||||
* Set flag for enabling shub specific wars
|
* Set flag for enabling shub specific wars
|
||||||
*/
|
*/
|
||||||
|
|
||||||
static inline int __cpuinit is_shub_1_1(int nasid)
|
static inline int is_shub_1_1(int nasid)
|
||||||
{
|
{
|
||||||
unsigned long id;
|
unsigned long id;
|
||||||
int rev;
|
int rev;
|
||||||
|
@ -212,7 +212,7 @@ static inline int __cpuinit is_shub_1_1(int nasid)
|
||||||
return rev <= 2;
|
return rev <= 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __cpuinit sn_check_for_wars(void)
|
static void sn_check_for_wars(void)
|
||||||
{
|
{
|
||||||
int cnode;
|
int cnode;
|
||||||
|
|
||||||
|
@ -558,7 +558,7 @@ static void __init sn_init_pdas(char **cmdline_p)
|
||||||
* Also sets up a few fields in the nodepda. Also known as
|
* Also sets up a few fields in the nodepda. Also known as
|
||||||
* platform_cpu_init() by the ia64 machvec code.
|
* platform_cpu_init() by the ia64 machvec code.
|
||||||
*/
|
*/
|
||||||
void __cpuinit sn_cpu_init(void)
|
void sn_cpu_init(void)
|
||||||
{
|
{
|
||||||
int cpuid;
|
int cpuid;
|
||||||
int cpuphyid;
|
int cpuphyid;
|
||||||
|
|
|
@ -74,7 +74,7 @@ void __init xen_setup_vcpu_info_placement(void)
|
||||||
xen_vcpu_setup(cpu);
|
xen_vcpu_setup(cpu);
|
||||||
}
|
}
|
||||||
|
|
||||||
void __cpuinit
|
void
|
||||||
xen_cpu_init(void)
|
xen_cpu_init(void)
|
||||||
{
|
{
|
||||||
xen_smp_intr_init();
|
xen_smp_intr_init();
|
||||||
|
|
Loading…
Reference in New Issue