Merge commit 'v3.4' into android-3.4
This commit is contained in:
commit
5500e4fab2
|
@ -539,3 +539,13 @@ When: 3.6
|
||||||
Why: setitimer is not returning -EFAULT if user pointer is NULL. This
|
Why: setitimer is not returning -EFAULT if user pointer is NULL. This
|
||||||
violates the spec.
|
violates the spec.
|
||||||
Who: Sasikantha Babu <sasikanth.v19@gmail.com>
|
Who: Sasikantha Babu <sasikanth.v19@gmail.com>
|
||||||
|
|
||||||
|
----------------------------
|
||||||
|
|
||||||
|
What: V4L2_CID_HCENTER, V4L2_CID_VCENTER V4L2 controls
|
||||||
|
When: 3.7
|
||||||
|
Why: The V4L2_CID_VCENTER, V4L2_CID_HCENTER controls have been deprecated
|
||||||
|
for about 4 years and they are not used by any mainline driver.
|
||||||
|
There are newer controls (V4L2_CID_PAN*, V4L2_CID_TILT*) that provide
|
||||||
|
similar functionality.
|
||||||
|
Who: Sylwester Nawrocki <sylvester.nawrocki@gmail.com>
|
||||||
|
|
11
Makefile
11
Makefile
|
@ -1,7 +1,7 @@
|
||||||
VERSION = 3
|
VERSION = 3
|
||||||
PATCHLEVEL = 4
|
PATCHLEVEL = 4
|
||||||
SUBLEVEL = 0
|
SUBLEVEL = 0
|
||||||
EXTRAVERSION = -rc7
|
EXTRAVERSION =
|
||||||
NAME = Saber-toothed Squirrel
|
NAME = Saber-toothed Squirrel
|
||||||
|
|
||||||
# *DOCUMENTATION*
|
# *DOCUMENTATION*
|
||||||
|
@ -442,7 +442,7 @@ asm-generic:
|
||||||
|
|
||||||
no-dot-config-targets := clean mrproper distclean \
|
no-dot-config-targets := clean mrproper distclean \
|
||||||
cscope gtags TAGS tags help %docs check% coccicheck \
|
cscope gtags TAGS tags help %docs check% coccicheck \
|
||||||
include/linux/version.h headers_% archheaders \
|
include/linux/version.h headers_% archheaders archscripts \
|
||||||
kernelversion %src-pkg
|
kernelversion %src-pkg
|
||||||
|
|
||||||
config-targets := 0
|
config-targets := 0
|
||||||
|
@ -979,7 +979,7 @@ prepare1: prepare2 include/linux/version.h include/generated/utsrelease.h \
|
||||||
include/config/auto.conf
|
include/config/auto.conf
|
||||||
$(cmd_crmodverdir)
|
$(cmd_crmodverdir)
|
||||||
|
|
||||||
archprepare: archheaders prepare1 scripts_basic
|
archprepare: archheaders archscripts prepare1 scripts_basic
|
||||||
|
|
||||||
prepare0: archprepare FORCE
|
prepare0: archprepare FORCE
|
||||||
$(Q)$(MAKE) $(build)=.
|
$(Q)$(MAKE) $(build)=.
|
||||||
|
@ -1049,8 +1049,11 @@ hdr-dst = $(if $(KBUILD_HEADERS), dst=include/asm-$(hdr-arch), dst=include/asm)
|
||||||
PHONY += archheaders
|
PHONY += archheaders
|
||||||
archheaders:
|
archheaders:
|
||||||
|
|
||||||
|
PHONY += archscripts
|
||||||
|
archscripts:
|
||||||
|
|
||||||
PHONY += __headers
|
PHONY += __headers
|
||||||
__headers: include/linux/version.h scripts_basic asm-generic archheaders FORCE
|
__headers: include/linux/version.h scripts_basic asm-generic archheaders archscripts FORCE
|
||||||
$(Q)$(MAKE) $(build)=scripts build_unifdef
|
$(Q)$(MAKE) $(build)=scripts build_unifdef
|
||||||
|
|
||||||
PHONY += headers_install_all
|
PHONY += headers_install_all
|
||||||
|
|
|
@ -42,7 +42,8 @@ sirfsoc_alloc_gc(void __iomem *base, unsigned int irq_start, unsigned int num)
|
||||||
static __init void sirfsoc_irq_init(void)
|
static __init void sirfsoc_irq_init(void)
|
||||||
{
|
{
|
||||||
sirfsoc_alloc_gc(sirfsoc_intc_base, 0, 32);
|
sirfsoc_alloc_gc(sirfsoc_intc_base, 0, 32);
|
||||||
sirfsoc_alloc_gc(sirfsoc_intc_base + 4, 32, SIRFSOC_INTENAL_IRQ_END - 32);
|
sirfsoc_alloc_gc(sirfsoc_intc_base + 4, 32,
|
||||||
|
SIRFSOC_INTENAL_IRQ_END + 1 - 32);
|
||||||
|
|
||||||
writel_relaxed(0, sirfsoc_intc_base + SIRFSOC_INT_RISC_LEVEL0);
|
writel_relaxed(0, sirfsoc_intc_base + SIRFSOC_INT_RISC_LEVEL0);
|
||||||
writel_relaxed(0, sirfsoc_intc_base + SIRFSOC_INT_RISC_LEVEL1);
|
writel_relaxed(0, sirfsoc_intc_base + SIRFSOC_INT_RISC_LEVEL1);
|
||||||
|
@ -68,7 +69,8 @@ void __init sirfsoc_of_irq_init(void)
|
||||||
if (!sirfsoc_intc_base)
|
if (!sirfsoc_intc_base)
|
||||||
panic("unable to map intc cpu registers\n");
|
panic("unable to map intc cpu registers\n");
|
||||||
|
|
||||||
irq_domain_add_legacy(np, 32, 0, 0, &irq_domain_simple_ops, NULL);
|
irq_domain_add_legacy(np, SIRFSOC_INTENAL_IRQ_END + 1, 0, 0,
|
||||||
|
&irq_domain_simple_ops, NULL);
|
||||||
|
|
||||||
of_node_put(np);
|
of_node_put(np);
|
||||||
|
|
||||||
|
|
|
@ -53,10 +53,10 @@ static void flowctrl_update(u8 offset, u32 value)
|
||||||
|
|
||||||
void flowctrl_write_cpu_csr(unsigned int cpuid, u32 value)
|
void flowctrl_write_cpu_csr(unsigned int cpuid, u32 value)
|
||||||
{
|
{
|
||||||
return flowctrl_update(flowctrl_offset_halt_cpu[cpuid], value);
|
return flowctrl_update(flowctrl_offset_cpu_csr[cpuid], value);
|
||||||
}
|
}
|
||||||
|
|
||||||
void flowctrl_write_cpu_halt(unsigned int cpuid, u32 value)
|
void flowctrl_write_cpu_halt(unsigned int cpuid, u32 value)
|
||||||
{
|
{
|
||||||
return flowctrl_update(flowctrl_offset_cpu_csr[cpuid], value);
|
return flowctrl_update(flowctrl_offset_halt_cpu[cpuid], value);
|
||||||
}
|
}
|
||||||
|
|
|
@ -247,7 +247,9 @@ good_area:
|
||||||
return handle_mm_fault(mm, vma, addr & PAGE_MASK, flags);
|
return handle_mm_fault(mm, vma, addr & PAGE_MASK, flags);
|
||||||
|
|
||||||
check_stack:
|
check_stack:
|
||||||
if (vma->vm_flags & VM_GROWSDOWN && !expand_stack(vma, addr))
|
/* Don't allow expansion below FIRST_USER_ADDRESS */
|
||||||
|
if (vma->vm_flags & VM_GROWSDOWN &&
|
||||||
|
addr >= FIRST_USER_ADDRESS && !expand_stack(vma, addr))
|
||||||
goto good_area;
|
goto good_area;
|
||||||
out:
|
out:
|
||||||
return fault;
|
return fault;
|
||||||
|
|
|
@ -489,6 +489,7 @@ static void __init build_mem_type_table(void)
|
||||||
*/
|
*/
|
||||||
for (i = 0; i < ARRAY_SIZE(mem_types); i++) {
|
for (i = 0; i < ARRAY_SIZE(mem_types); i++) {
|
||||||
mem_types[i].prot_pte |= PTE_EXT_AF;
|
mem_types[i].prot_pte |= PTE_EXT_AF;
|
||||||
|
if (mem_types[i].prot_sect)
|
||||||
mem_types[i].prot_sect |= PMD_SECT_AF;
|
mem_types[i].prot_sect |= PMD_SECT_AF;
|
||||||
}
|
}
|
||||||
kern_pgprot |= PTE_EXT_AF;
|
kern_pgprot |= PTE_EXT_AF;
|
||||||
|
|
|
@ -11,6 +11,7 @@
|
||||||
#include <linux/types.h>
|
#include <linux/types.h>
|
||||||
#include <linux/cpu.h>
|
#include <linux/cpu.h>
|
||||||
#include <linux/cpu_pm.h>
|
#include <linux/cpu_pm.h>
|
||||||
|
#include <linux/hardirq.h>
|
||||||
#include <linux/kernel.h>
|
#include <linux/kernel.h>
|
||||||
#include <linux/notifier.h>
|
#include <linux/notifier.h>
|
||||||
#include <linux/signal.h>
|
#include <linux/signal.h>
|
||||||
|
@ -432,7 +433,10 @@ void VFP_bounce(u32 trigger, u32 fpexc, struct pt_regs *regs)
|
||||||
|
|
||||||
static void vfp_enable(void *unused)
|
static void vfp_enable(void *unused)
|
||||||
{
|
{
|
||||||
u32 access = get_copro_access();
|
u32 access;
|
||||||
|
|
||||||
|
BUG_ON(preemptible());
|
||||||
|
access = get_copro_access();
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Enable full access to VFP (cp10 and cp11)
|
* Enable full access to VFP (cp10 and cp11)
|
||||||
|
@ -579,12 +583,6 @@ int vfp_preserve_user_clear_hwstate(struct user_vfp __user *ufp,
|
||||||
* entry.
|
* entry.
|
||||||
*/
|
*/
|
||||||
hwstate->fpscr &= ~(FPSCR_LENGTH_MASK | FPSCR_STRIDE_MASK);
|
hwstate->fpscr &= ~(FPSCR_LENGTH_MASK | FPSCR_STRIDE_MASK);
|
||||||
|
|
||||||
/*
|
|
||||||
* Disable VFP in the hwstate so that we can detect if it gets
|
|
||||||
* used.
|
|
||||||
*/
|
|
||||||
hwstate->fpexc &= ~FPEXC_EN;
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -597,11 +595,7 @@ int vfp_restore_user_hwstate(struct user_vfp __user *ufp,
|
||||||
unsigned long fpexc;
|
unsigned long fpexc;
|
||||||
int err = 0;
|
int err = 0;
|
||||||
|
|
||||||
/*
|
/* Disable VFP to avoid corrupting the new thread state. */
|
||||||
* If VFP has been used, then disable it to avoid corrupting
|
|
||||||
* the new thread state.
|
|
||||||
*/
|
|
||||||
if (hwstate->fpexc & FPEXC_EN)
|
|
||||||
vfp_flush_hwstate(thread);
|
vfp_flush_hwstate(thread);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -663,7 +657,7 @@ static int __init vfp_init(void)
|
||||||
unsigned int cpu_arch = cpu_architecture();
|
unsigned int cpu_arch = cpu_architecture();
|
||||||
|
|
||||||
if (cpu_arch >= CPU_ARCH_ARMv6)
|
if (cpu_arch >= CPU_ARCH_ARMv6)
|
||||||
vfp_enable(NULL);
|
on_each_cpu(vfp_enable, NULL, 1);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* First check that there is a VFP that we can use.
|
* First check that there is a VFP that we can use.
|
||||||
|
@ -684,8 +678,6 @@ static int __init vfp_init(void)
|
||||||
} else {
|
} else {
|
||||||
hotcpu_notifier(vfp_hotplug, 0);
|
hotcpu_notifier(vfp_hotplug, 0);
|
||||||
|
|
||||||
smp_call_function(vfp_enable, NULL, 1);
|
|
||||||
|
|
||||||
VFP_arch = (vfpsid & FPSID_ARCH_MASK) >> FPSID_ARCH_BIT; /* Extract the architecture version */
|
VFP_arch = (vfpsid & FPSID_ARCH_MASK) >> FPSID_ARCH_BIT; /* Extract the architecture version */
|
||||||
printk("implementor %02x architecture %d part %02x variant %x rev %x\n",
|
printk("implementor %02x architecture %d part %02x variant %x rev %x\n",
|
||||||
(vfpsid & FPSID_IMPLEMENTER_MASK) >> FPSID_IMPLEMENTER_BIT,
|
(vfpsid & FPSID_IMPLEMENTER_MASK) >> FPSID_IMPLEMENTER_BIT,
|
||||||
|
|
|
@ -135,10 +135,6 @@ unsigned long get_wchan(struct task_struct *p);
|
||||||
#define KSTK_EIP(tsk) ((tsk)->thread.frame0->pc)
|
#define KSTK_EIP(tsk) ((tsk)->thread.frame0->pc)
|
||||||
#define KSTK_ESP(tsk) ((tsk)->thread.frame0->sp)
|
#define KSTK_ESP(tsk) ((tsk)->thread.frame0->sp)
|
||||||
|
|
||||||
/* Allocation and freeing of basic task resources. */
|
|
||||||
extern struct task_struct *alloc_task_struct_node(int node);
|
|
||||||
extern void free_task_struct(struct task_struct *p);
|
|
||||||
|
|
||||||
#define cpu_relax() barrier()
|
#define cpu_relax() barrier()
|
||||||
|
|
||||||
/* data cache prefetch */
|
/* data cache prefetch */
|
||||||
|
|
|
@ -24,6 +24,7 @@
|
||||||
#include <linux/sched.h>
|
#include <linux/sched.h>
|
||||||
#include <linux/profile.h>
|
#include <linux/profile.h>
|
||||||
#include <linux/smp.h>
|
#include <linux/smp.h>
|
||||||
|
#include <linux/cpu.h>
|
||||||
#include <asm/tlbflush.h>
|
#include <asm/tlbflush.h>
|
||||||
#include <asm/bitops.h>
|
#include <asm/bitops.h>
|
||||||
#include <asm/processor.h>
|
#include <asm/processor.h>
|
||||||
|
@ -38,7 +39,6 @@
|
||||||
#include "internal.h"
|
#include "internal.h"
|
||||||
|
|
||||||
#ifdef CONFIG_HOTPLUG_CPU
|
#ifdef CONFIG_HOTPLUG_CPU
|
||||||
#include <linux/cpu.h>
|
|
||||||
#include <asm/cacheflush.h>
|
#include <asm/cacheflush.h>
|
||||||
|
|
||||||
static unsigned long sleep_mode[NR_CPUS];
|
static unsigned long sleep_mode[NR_CPUS];
|
||||||
|
@ -874,10 +874,13 @@ static void __init smp_online(void)
|
||||||
|
|
||||||
cpu = smp_processor_id();
|
cpu = smp_processor_id();
|
||||||
|
|
||||||
local_irq_enable();
|
notify_cpu_starting(cpu);
|
||||||
|
|
||||||
|
ipi_call_lock();
|
||||||
set_cpu_online(cpu, true);
|
set_cpu_online(cpu, true);
|
||||||
smp_wmb();
|
ipi_call_unlock();
|
||||||
|
|
||||||
|
local_irq_enable();
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -21,7 +21,12 @@
|
||||||
#define ARCH_HAS_PREFETCH
|
#define ARCH_HAS_PREFETCH
|
||||||
static inline void prefetch(const void *addr)
|
static inline void prefetch(const void *addr)
|
||||||
{
|
{
|
||||||
__asm__("ldw 0(%0), %%r0" : : "r" (addr));
|
__asm__(
|
||||||
|
#ifndef CONFIG_PA20
|
||||||
|
/* Need to avoid prefetch of NULL on PA7300LC */
|
||||||
|
" extrw,u,= %0,31,32,%%r0\n"
|
||||||
|
#endif
|
||||||
|
" ldw 0(%0), %%r0" : : "r" (addr));
|
||||||
}
|
}
|
||||||
|
|
||||||
/* LDD is a PA2.0 addition. */
|
/* LDD is a PA2.0 addition. */
|
||||||
|
|
|
@ -581,7 +581,11 @@
|
||||||
*/
|
*/
|
||||||
cmpiclr,= 0x01,\tmp,%r0
|
cmpiclr,= 0x01,\tmp,%r0
|
||||||
ldi (_PAGE_DIRTY|_PAGE_READ|_PAGE_WRITE),\prot
|
ldi (_PAGE_DIRTY|_PAGE_READ|_PAGE_WRITE),\prot
|
||||||
|
#ifdef CONFIG_64BIT
|
||||||
depd,z \prot,8,7,\prot
|
depd,z \prot,8,7,\prot
|
||||||
|
#else
|
||||||
|
depw,z \prot,8,7,\prot
|
||||||
|
#endif
|
||||||
/*
|
/*
|
||||||
* OK, it is in the temp alias region, check whether "from" or "to".
|
* OK, it is in the temp alias region, check whether "from" or "to".
|
||||||
* Check "subtle" note in pacache.S re: r23/r26.
|
* Check "subtle" note in pacache.S re: r23/r26.
|
||||||
|
|
|
@ -692,7 +692,7 @@ ENTRY(flush_icache_page_asm)
|
||||||
|
|
||||||
/* Purge any old translation */
|
/* Purge any old translation */
|
||||||
|
|
||||||
pitlb (%sr0,%r28)
|
pitlb (%sr4,%r28)
|
||||||
|
|
||||||
ldil L%icache_stride, %r1
|
ldil L%icache_stride, %r1
|
||||||
ldw R%icache_stride(%r1), %r1
|
ldw R%icache_stride(%r1), %r1
|
||||||
|
@ -706,27 +706,29 @@ ENTRY(flush_icache_page_asm)
|
||||||
sub %r25, %r1, %r25
|
sub %r25, %r1, %r25
|
||||||
|
|
||||||
|
|
||||||
1: fic,m %r1(%r28)
|
/* fic only has the type 26 form on PA1.1, requiring an
|
||||||
fic,m %r1(%r28)
|
* explicit space specification, so use %sr4 */
|
||||||
fic,m %r1(%r28)
|
1: fic,m %r1(%sr4,%r28)
|
||||||
fic,m %r1(%r28)
|
fic,m %r1(%sr4,%r28)
|
||||||
fic,m %r1(%r28)
|
fic,m %r1(%sr4,%r28)
|
||||||
fic,m %r1(%r28)
|
fic,m %r1(%sr4,%r28)
|
||||||
fic,m %r1(%r28)
|
fic,m %r1(%sr4,%r28)
|
||||||
fic,m %r1(%r28)
|
fic,m %r1(%sr4,%r28)
|
||||||
fic,m %r1(%r28)
|
fic,m %r1(%sr4,%r28)
|
||||||
fic,m %r1(%r28)
|
fic,m %r1(%sr4,%r28)
|
||||||
fic,m %r1(%r28)
|
fic,m %r1(%sr4,%r28)
|
||||||
fic,m %r1(%r28)
|
fic,m %r1(%sr4,%r28)
|
||||||
fic,m %r1(%r28)
|
fic,m %r1(%sr4,%r28)
|
||||||
fic,m %r1(%r28)
|
fic,m %r1(%sr4,%r28)
|
||||||
fic,m %r1(%r28)
|
fic,m %r1(%sr4,%r28)
|
||||||
|
fic,m %r1(%sr4,%r28)
|
||||||
|
fic,m %r1(%sr4,%r28)
|
||||||
cmpb,COND(<<) %r28, %r25,1b
|
cmpb,COND(<<) %r28, %r25,1b
|
||||||
fic,m %r1(%r28)
|
fic,m %r1(%sr4,%r28)
|
||||||
|
|
||||||
sync
|
sync
|
||||||
bv %r0(%r2)
|
bv %r0(%r2)
|
||||||
pitlb (%sr0,%r25)
|
pitlb (%sr4,%r25)
|
||||||
.exit
|
.exit
|
||||||
|
|
||||||
.procend
|
.procend
|
||||||
|
|
|
@ -31,6 +31,7 @@
|
||||||
#include <linux/delay.h>
|
#include <linux/delay.h>
|
||||||
#include <linux/bitops.h>
|
#include <linux/bitops.h>
|
||||||
#include <linux/ftrace.h>
|
#include <linux/ftrace.h>
|
||||||
|
#include <linux/cpu.h>
|
||||||
|
|
||||||
#include <linux/atomic.h>
|
#include <linux/atomic.h>
|
||||||
#include <asm/current.h>
|
#include <asm/current.h>
|
||||||
|
@ -296,7 +297,12 @@ smp_cpu_init(int cpunum)
|
||||||
printk(KERN_CRIT "CPU#%d already initialized!\n", cpunum);
|
printk(KERN_CRIT "CPU#%d already initialized!\n", cpunum);
|
||||||
machine_halt();
|
machine_halt();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
notify_cpu_starting(cpunum);
|
||||||
|
|
||||||
|
ipi_call_lock();
|
||||||
set_cpu_online(cpunum, true);
|
set_cpu_online(cpunum, true);
|
||||||
|
ipi_call_unlock();
|
||||||
|
|
||||||
/* Initialise the idle task for this CPU */
|
/* Initialise the idle task for this CPU */
|
||||||
atomic_inc(&init_mm.mm_count);
|
atomic_inc(&init_mm.mm_count);
|
||||||
|
|
|
@ -81,12 +81,13 @@ struct kvmppc_vcpu_book3s {
|
||||||
u64 sdr1;
|
u64 sdr1;
|
||||||
u64 hior;
|
u64 hior;
|
||||||
u64 msr_mask;
|
u64 msr_mask;
|
||||||
u64 vsid_next;
|
|
||||||
#ifdef CONFIG_PPC_BOOK3S_32
|
#ifdef CONFIG_PPC_BOOK3S_32
|
||||||
u32 vsid_pool[VSID_POOL_SIZE];
|
u32 vsid_pool[VSID_POOL_SIZE];
|
||||||
|
u32 vsid_next;
|
||||||
#else
|
#else
|
||||||
u64 vsid_first;
|
u64 proto_vsid_first;
|
||||||
u64 vsid_max;
|
u64 proto_vsid_max;
|
||||||
|
u64 proto_vsid_next;
|
||||||
#endif
|
#endif
|
||||||
int context_id[SID_CONTEXTS];
|
int context_id[SID_CONTEXTS];
|
||||||
|
|
||||||
|
|
|
@ -194,14 +194,14 @@ static struct kvmppc_sid_map *create_sid_map(struct kvm_vcpu *vcpu, u64 gvsid)
|
||||||
backwards_map = !backwards_map;
|
backwards_map = !backwards_map;
|
||||||
|
|
||||||
/* Uh-oh ... out of mappings. Let's flush! */
|
/* Uh-oh ... out of mappings. Let's flush! */
|
||||||
if (vcpu_book3s->vsid_next == vcpu_book3s->vsid_max) {
|
if (vcpu_book3s->proto_vsid_next == vcpu_book3s->proto_vsid_max) {
|
||||||
vcpu_book3s->vsid_next = vcpu_book3s->vsid_first;
|
vcpu_book3s->proto_vsid_next = vcpu_book3s->proto_vsid_first;
|
||||||
memset(vcpu_book3s->sid_map, 0,
|
memset(vcpu_book3s->sid_map, 0,
|
||||||
sizeof(struct kvmppc_sid_map) * SID_MAP_NUM);
|
sizeof(struct kvmppc_sid_map) * SID_MAP_NUM);
|
||||||
kvmppc_mmu_pte_flush(vcpu, 0, 0);
|
kvmppc_mmu_pte_flush(vcpu, 0, 0);
|
||||||
kvmppc_mmu_flush_segments(vcpu);
|
kvmppc_mmu_flush_segments(vcpu);
|
||||||
}
|
}
|
||||||
map->host_vsid = vcpu_book3s->vsid_next++;
|
map->host_vsid = vsid_scramble(vcpu_book3s->proto_vsid_next++, 256M);
|
||||||
|
|
||||||
map->guest_vsid = gvsid;
|
map->guest_vsid = gvsid;
|
||||||
map->valid = true;
|
map->valid = true;
|
||||||
|
@ -319,9 +319,10 @@ int kvmppc_mmu_init(struct kvm_vcpu *vcpu)
|
||||||
return -1;
|
return -1;
|
||||||
vcpu3s->context_id[0] = err;
|
vcpu3s->context_id[0] = err;
|
||||||
|
|
||||||
vcpu3s->vsid_max = ((vcpu3s->context_id[0] + 1) << USER_ESID_BITS) - 1;
|
vcpu3s->proto_vsid_max = ((vcpu3s->context_id[0] + 1)
|
||||||
vcpu3s->vsid_first = vcpu3s->context_id[0] << USER_ESID_BITS;
|
<< USER_ESID_BITS) - 1;
|
||||||
vcpu3s->vsid_next = vcpu3s->vsid_first;
|
vcpu3s->proto_vsid_first = vcpu3s->context_id[0] << USER_ESID_BITS;
|
||||||
|
vcpu3s->proto_vsid_next = vcpu3s->proto_vsid_first;
|
||||||
|
|
||||||
kvmppc_mmu_hpte_init(vcpu);
|
kvmppc_mmu_hpte_init(vcpu);
|
||||||
|
|
||||||
|
|
|
@ -463,6 +463,7 @@ long kvmppc_h_bulk_remove(struct kvm_vcpu *vcpu)
|
||||||
/* insert R and C bits from PTE */
|
/* insert R and C bits from PTE */
|
||||||
rcbits = rev->guest_rpte & (HPTE_R_R|HPTE_R_C);
|
rcbits = rev->guest_rpte & (HPTE_R_R|HPTE_R_C);
|
||||||
args[j] |= rcbits << (56 - 5);
|
args[j] |= rcbits << (56 - 5);
|
||||||
|
hp[0] = 0;
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -197,7 +197,8 @@ kvmppc_interrupt:
|
||||||
/* Save guest PC and MSR */
|
/* Save guest PC and MSR */
|
||||||
#ifdef CONFIG_PPC64
|
#ifdef CONFIG_PPC64
|
||||||
BEGIN_FTR_SECTION
|
BEGIN_FTR_SECTION
|
||||||
andi. r0,r12,0x2
|
andi. r0, r12, 0x2
|
||||||
|
cmpwi cr1, r0, 0
|
||||||
beq 1f
|
beq 1f
|
||||||
mfspr r3,SPRN_HSRR0
|
mfspr r3,SPRN_HSRR0
|
||||||
mfspr r4,SPRN_HSRR1
|
mfspr r4,SPRN_HSRR1
|
||||||
|
@ -250,6 +251,12 @@ END_FTR_SECTION_IFSET(CPU_FTR_HVMODE)
|
||||||
beq ld_last_prev_inst
|
beq ld_last_prev_inst
|
||||||
cmpwi r12, BOOK3S_INTERRUPT_ALIGNMENT
|
cmpwi r12, BOOK3S_INTERRUPT_ALIGNMENT
|
||||||
beq- ld_last_inst
|
beq- ld_last_inst
|
||||||
|
#ifdef CONFIG_PPC64
|
||||||
|
BEGIN_FTR_SECTION
|
||||||
|
cmpwi r12, BOOK3S_INTERRUPT_H_EMUL_ASSIST
|
||||||
|
beq- ld_last_inst
|
||||||
|
END_FTR_SECTION_IFSET(CPU_FTR_HVMODE)
|
||||||
|
#endif
|
||||||
|
|
||||||
b no_ld_last_inst
|
b no_ld_last_inst
|
||||||
|
|
||||||
|
@ -316,23 +323,17 @@ no_dcbz32_off:
|
||||||
* Having set up SRR0/1 with the address where we want
|
* Having set up SRR0/1 with the address where we want
|
||||||
* to continue with relocation on (potentially in module
|
* to continue with relocation on (potentially in module
|
||||||
* space), we either just go straight there with rfi[d],
|
* space), we either just go straight there with rfi[d],
|
||||||
* or we jump to an interrupt handler with bctr if there
|
* or we jump to an interrupt handler if there is an
|
||||||
* is an interrupt to be handled first. In the latter
|
* interrupt to be handled first. In the latter case,
|
||||||
* case, the rfi[d] at the end of the interrupt handler
|
* the rfi[d] at the end of the interrupt handler will
|
||||||
* will get us back to where we want to continue.
|
* get us back to where we want to continue.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
cmpwi r12, BOOK3S_INTERRUPT_EXTERNAL
|
|
||||||
beq 1f
|
|
||||||
cmpwi r12, BOOK3S_INTERRUPT_DECREMENTER
|
|
||||||
beq 1f
|
|
||||||
cmpwi r12, BOOK3S_INTERRUPT_PERFMON
|
|
||||||
1: mtctr r12
|
|
||||||
|
|
||||||
/* Register usage at this point:
|
/* Register usage at this point:
|
||||||
*
|
*
|
||||||
* R1 = host R1
|
* R1 = host R1
|
||||||
* R2 = host R2
|
* R2 = host R2
|
||||||
|
* R10 = raw exit handler id
|
||||||
* R12 = exit handler id
|
* R12 = exit handler id
|
||||||
* R13 = shadow vcpu (32-bit) or PACA (64-bit)
|
* R13 = shadow vcpu (32-bit) or PACA (64-bit)
|
||||||
* SVCPU.* = guest *
|
* SVCPU.* = guest *
|
||||||
|
@ -342,12 +343,25 @@ no_dcbz32_off:
|
||||||
PPC_LL r6, HSTATE_HOST_MSR(r13)
|
PPC_LL r6, HSTATE_HOST_MSR(r13)
|
||||||
PPC_LL r8, HSTATE_VMHANDLER(r13)
|
PPC_LL r8, HSTATE_VMHANDLER(r13)
|
||||||
|
|
||||||
/* Restore host msr -> SRR1 */
|
#ifdef CONFIG_PPC64
|
||||||
|
BEGIN_FTR_SECTION
|
||||||
|
beq cr1, 1f
|
||||||
|
mtspr SPRN_HSRR1, r6
|
||||||
|
mtspr SPRN_HSRR0, r8
|
||||||
|
END_FTR_SECTION_IFSET(CPU_FTR_HVMODE)
|
||||||
|
#endif
|
||||||
|
1: /* Restore host msr -> SRR1 */
|
||||||
mtsrr1 r6
|
mtsrr1 r6
|
||||||
/* Load highmem handler address */
|
/* Load highmem handler address */
|
||||||
mtsrr0 r8
|
mtsrr0 r8
|
||||||
|
|
||||||
/* RFI into the highmem handler, or jump to interrupt handler */
|
/* RFI into the highmem handler, or jump to interrupt handler */
|
||||||
beqctr
|
cmpwi r12, BOOK3S_INTERRUPT_EXTERNAL
|
||||||
|
beqa BOOK3S_INTERRUPT_EXTERNAL
|
||||||
|
cmpwi r12, BOOK3S_INTERRUPT_DECREMENTER
|
||||||
|
beqa BOOK3S_INTERRUPT_DECREMENTER
|
||||||
|
cmpwi r12, BOOK3S_INTERRUPT_PERFMON
|
||||||
|
beqa BOOK3S_INTERRUPT_PERFMON
|
||||||
|
|
||||||
RFI
|
RFI
|
||||||
kvmppc_handler_trampoline_exit_end:
|
kvmppc_handler_trampoline_exit_end:
|
||||||
|
|
|
@ -11,6 +11,7 @@ config TILE
|
||||||
select GENERIC_IRQ_PROBE
|
select GENERIC_IRQ_PROBE
|
||||||
select GENERIC_PENDING_IRQ if SMP
|
select GENERIC_PENDING_IRQ if SMP
|
||||||
select GENERIC_IRQ_SHOW
|
select GENERIC_IRQ_SHOW
|
||||||
|
select HAVE_SYSCALL_WRAPPERS if TILEGX
|
||||||
select SYS_HYPERVISOR
|
select SYS_HYPERVISOR
|
||||||
select ARCH_HAVE_NMI_SAFE_CMPXCHG
|
select ARCH_HAVE_NMI_SAFE_CMPXCHG
|
||||||
|
|
||||||
|
|
|
@ -100,9 +100,14 @@ extern void cpu_idle_on_new_stack(struct thread_info *old_ti,
|
||||||
|
|
||||||
#else /* __ASSEMBLY__ */
|
#else /* __ASSEMBLY__ */
|
||||||
|
|
||||||
/* how to get the thread information struct from ASM */
|
/*
|
||||||
|
* How to get the thread information struct from assembly.
|
||||||
|
* Note that we use different macros since different architectures
|
||||||
|
* have different semantics in their "mm" instruction and we would
|
||||||
|
* like to guarantee that the macro expands to exactly one instruction.
|
||||||
|
*/
|
||||||
#ifdef __tilegx__
|
#ifdef __tilegx__
|
||||||
#define GET_THREAD_INFO(reg) move reg, sp; mm reg, zero, LOG2_THREAD_SIZE, 63
|
#define EXTRACT_THREAD_INFO(reg) mm reg, zero, LOG2_THREAD_SIZE, 63
|
||||||
#else
|
#else
|
||||||
#define GET_THREAD_INFO(reg) mm reg, sp, zero, LOG2_THREAD_SIZE, 31
|
#define GET_THREAD_INFO(reg) mm reg, sp, zero, LOG2_THREAD_SIZE, 31
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -403,19 +403,17 @@ int compat_setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info,
|
||||||
* Set up registers for signal handler.
|
* Set up registers for signal handler.
|
||||||
* Registers that we don't modify keep the value they had from
|
* Registers that we don't modify keep the value they had from
|
||||||
* user-space at the time we took the signal.
|
* user-space at the time we took the signal.
|
||||||
|
* We always pass siginfo and mcontext, regardless of SA_SIGINFO,
|
||||||
|
* since some things rely on this (e.g. glibc's debug/segfault.c).
|
||||||
*/
|
*/
|
||||||
regs->pc = ptr_to_compat_reg(ka->sa.sa_handler);
|
regs->pc = ptr_to_compat_reg(ka->sa.sa_handler);
|
||||||
regs->ex1 = PL_ICS_EX1(USER_PL, 1); /* set crit sec in handler */
|
regs->ex1 = PL_ICS_EX1(USER_PL, 1); /* set crit sec in handler */
|
||||||
regs->sp = ptr_to_compat_reg(frame);
|
regs->sp = ptr_to_compat_reg(frame);
|
||||||
regs->lr = restorer;
|
regs->lr = restorer;
|
||||||
regs->regs[0] = (unsigned long) usig;
|
regs->regs[0] = (unsigned long) usig;
|
||||||
|
|
||||||
if (ka->sa.sa_flags & SA_SIGINFO) {
|
|
||||||
/* Need extra arguments, so mark to restore caller-saves. */
|
|
||||||
regs->regs[1] = ptr_to_compat_reg(&frame->info);
|
regs->regs[1] = ptr_to_compat_reg(&frame->info);
|
||||||
regs->regs[2] = ptr_to_compat_reg(&frame->uc);
|
regs->regs[2] = ptr_to_compat_reg(&frame->uc);
|
||||||
regs->flags |= PT_FLAGS_CALLER_SAVES;
|
regs->flags |= PT_FLAGS_CALLER_SAVES;
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Notify any tracer that was single-stepping it.
|
* Notify any tracer that was single-stepping it.
|
||||||
|
|
|
@ -838,6 +838,18 @@ STD_ENTRY(interrupt_return)
|
||||||
.Lresume_userspace:
|
.Lresume_userspace:
|
||||||
FEEDBACK_REENTER(interrupt_return)
|
FEEDBACK_REENTER(interrupt_return)
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Use r33 to hold whether we have already loaded the callee-saves
|
||||||
|
* into ptregs. We don't want to do it twice in this loop, since
|
||||||
|
* then we'd clobber whatever changes are made by ptrace, etc.
|
||||||
|
* Get base of stack in r32.
|
||||||
|
*/
|
||||||
|
{
|
||||||
|
GET_THREAD_INFO(r32)
|
||||||
|
movei r33, 0
|
||||||
|
}
|
||||||
|
|
||||||
|
.Lretry_work_pending:
|
||||||
/*
|
/*
|
||||||
* Disable interrupts so as to make sure we don't
|
* Disable interrupts so as to make sure we don't
|
||||||
* miss an interrupt that sets any of the thread flags (like
|
* miss an interrupt that sets any of the thread flags (like
|
||||||
|
@ -848,9 +860,6 @@ STD_ENTRY(interrupt_return)
|
||||||
IRQ_DISABLE(r20, r21)
|
IRQ_DISABLE(r20, r21)
|
||||||
TRACE_IRQS_OFF /* Note: clobbers registers r0-r29 */
|
TRACE_IRQS_OFF /* Note: clobbers registers r0-r29 */
|
||||||
|
|
||||||
/* Get base of stack in r32; note r30/31 are used as arguments here. */
|
|
||||||
GET_THREAD_INFO(r32)
|
|
||||||
|
|
||||||
|
|
||||||
/* Check to see if there is any work to do before returning to user. */
|
/* Check to see if there is any work to do before returning to user. */
|
||||||
{
|
{
|
||||||
|
@ -866,16 +875,18 @@ STD_ENTRY(interrupt_return)
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Make sure we have all the registers saved for signal
|
* Make sure we have all the registers saved for signal
|
||||||
* handling or single-step. Call out to C code to figure out
|
* handling, notify-resume, or single-step. Call out to C
|
||||||
* exactly what we need to do for each flag bit, then if
|
* code to figure out exactly what we need to do for each flag bit,
|
||||||
* necessary, reload the flags and recheck.
|
* then if necessary, reload the flags and recheck.
|
||||||
*/
|
*/
|
||||||
push_extra_callee_saves r0
|
|
||||||
{
|
{
|
||||||
PTREGS_PTR(r0, PTREGS_OFFSET_BASE)
|
PTREGS_PTR(r0, PTREGS_OFFSET_BASE)
|
||||||
jal do_work_pending
|
bnz r33, 1f
|
||||||
}
|
}
|
||||||
bnz r0, .Lresume_userspace
|
push_extra_callee_saves r0
|
||||||
|
movei r33, 1
|
||||||
|
1: jal do_work_pending
|
||||||
|
bnz r0, .Lretry_work_pending
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* In the NMI case we
|
* In the NMI case we
|
||||||
|
@ -1180,10 +1191,12 @@ handle_syscall:
|
||||||
add r20, r20, tp
|
add r20, r20, tp
|
||||||
lw r21, r20
|
lw r21, r20
|
||||||
addi r21, r21, 1
|
addi r21, r21, 1
|
||||||
|
{
|
||||||
sw r20, r21
|
sw r20, r21
|
||||||
|
GET_THREAD_INFO(r31)
|
||||||
|
}
|
||||||
|
|
||||||
/* Trace syscalls, if requested. */
|
/* Trace syscalls, if requested. */
|
||||||
GET_THREAD_INFO(r31)
|
|
||||||
addi r31, r31, THREAD_INFO_FLAGS_OFFSET
|
addi r31, r31, THREAD_INFO_FLAGS_OFFSET
|
||||||
lw r30, r31
|
lw r30, r31
|
||||||
andi r30, r30, _TIF_SYSCALL_TRACE
|
andi r30, r30, _TIF_SYSCALL_TRACE
|
||||||
|
@ -1362,7 +1375,10 @@ handle_ill:
|
||||||
3:
|
3:
|
||||||
/* set PC and continue */
|
/* set PC and continue */
|
||||||
lw r26, r24
|
lw r26, r24
|
||||||
|
{
|
||||||
sw r28, r26
|
sw r28, r26
|
||||||
|
GET_THREAD_INFO(r0)
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Clear TIF_SINGLESTEP to prevent recursion if we execute an ill.
|
* Clear TIF_SINGLESTEP to prevent recursion if we execute an ill.
|
||||||
|
@ -1370,7 +1386,6 @@ handle_ill:
|
||||||
* need to clear it here and can't really impose on all other arches.
|
* need to clear it here and can't really impose on all other arches.
|
||||||
* So what's another write between friends?
|
* So what's another write between friends?
|
||||||
*/
|
*/
|
||||||
GET_THREAD_INFO(r0)
|
|
||||||
|
|
||||||
addi r1, r0, THREAD_INFO_FLAGS_OFFSET
|
addi r1, r0, THREAD_INFO_FLAGS_OFFSET
|
||||||
{
|
{
|
||||||
|
|
|
@ -646,6 +646,20 @@ STD_ENTRY(interrupt_return)
|
||||||
.Lresume_userspace:
|
.Lresume_userspace:
|
||||||
FEEDBACK_REENTER(interrupt_return)
|
FEEDBACK_REENTER(interrupt_return)
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Use r33 to hold whether we have already loaded the callee-saves
|
||||||
|
* into ptregs. We don't want to do it twice in this loop, since
|
||||||
|
* then we'd clobber whatever changes are made by ptrace, etc.
|
||||||
|
*/
|
||||||
|
{
|
||||||
|
movei r33, 0
|
||||||
|
move r32, sp
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Get base of stack in r32. */
|
||||||
|
EXTRACT_THREAD_INFO(r32)
|
||||||
|
|
||||||
|
.Lretry_work_pending:
|
||||||
/*
|
/*
|
||||||
* Disable interrupts so as to make sure we don't
|
* Disable interrupts so as to make sure we don't
|
||||||
* miss an interrupt that sets any of the thread flags (like
|
* miss an interrupt that sets any of the thread flags (like
|
||||||
|
@ -656,9 +670,6 @@ STD_ENTRY(interrupt_return)
|
||||||
IRQ_DISABLE(r20, r21)
|
IRQ_DISABLE(r20, r21)
|
||||||
TRACE_IRQS_OFF /* Note: clobbers registers r0-r29 */
|
TRACE_IRQS_OFF /* Note: clobbers registers r0-r29 */
|
||||||
|
|
||||||
/* Get base of stack in r32; note r30/31 are used as arguments here. */
|
|
||||||
GET_THREAD_INFO(r32)
|
|
||||||
|
|
||||||
|
|
||||||
/* Check to see if there is any work to do before returning to user. */
|
/* Check to see if there is any work to do before returning to user. */
|
||||||
{
|
{
|
||||||
|
@ -674,16 +685,18 @@ STD_ENTRY(interrupt_return)
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Make sure we have all the registers saved for signal
|
* Make sure we have all the registers saved for signal
|
||||||
* handling or single-step. Call out to C code to figure out
|
* handling or notify-resume. Call out to C code to figure out
|
||||||
* exactly what we need to do for each flag bit, then if
|
* exactly what we need to do for each flag bit, then if
|
||||||
* necessary, reload the flags and recheck.
|
* necessary, reload the flags and recheck.
|
||||||
*/
|
*/
|
||||||
push_extra_callee_saves r0
|
|
||||||
{
|
{
|
||||||
PTREGS_PTR(r0, PTREGS_OFFSET_BASE)
|
PTREGS_PTR(r0, PTREGS_OFFSET_BASE)
|
||||||
jal do_work_pending
|
bnez r33, 1f
|
||||||
}
|
}
|
||||||
bnez r0, .Lresume_userspace
|
push_extra_callee_saves r0
|
||||||
|
movei r33, 1
|
||||||
|
1: jal do_work_pending
|
||||||
|
bnez r0, .Lretry_work_pending
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* In the NMI case we
|
* In the NMI case we
|
||||||
|
@ -968,11 +981,16 @@ handle_syscall:
|
||||||
shl16insli r20, r20, hw0(irq_stat + IRQ_CPUSTAT_SYSCALL_COUNT_OFFSET)
|
shl16insli r20, r20, hw0(irq_stat + IRQ_CPUSTAT_SYSCALL_COUNT_OFFSET)
|
||||||
add r20, r20, tp
|
add r20, r20, tp
|
||||||
ld4s r21, r20
|
ld4s r21, r20
|
||||||
|
{
|
||||||
addi r21, r21, 1
|
addi r21, r21, 1
|
||||||
|
move r31, sp
|
||||||
|
}
|
||||||
|
{
|
||||||
st4 r20, r21
|
st4 r20, r21
|
||||||
|
EXTRACT_THREAD_INFO(r31)
|
||||||
|
}
|
||||||
|
|
||||||
/* Trace syscalls, if requested. */
|
/* Trace syscalls, if requested. */
|
||||||
GET_THREAD_INFO(r31)
|
|
||||||
addi r31, r31, THREAD_INFO_FLAGS_OFFSET
|
addi r31, r31, THREAD_INFO_FLAGS_OFFSET
|
||||||
ld r30, r31
|
ld r30, r31
|
||||||
andi r30, r30, _TIF_SYSCALL_TRACE
|
andi r30, r30, _TIF_SYSCALL_TRACE
|
||||||
|
|
|
@ -567,6 +567,10 @@ struct task_struct *__sched _switch_to(struct task_struct *prev,
|
||||||
*/
|
*/
|
||||||
int do_work_pending(struct pt_regs *regs, u32 thread_info_flags)
|
int do_work_pending(struct pt_regs *regs, u32 thread_info_flags)
|
||||||
{
|
{
|
||||||
|
/* If we enter in kernel mode, do nothing and exit the caller loop. */
|
||||||
|
if (!user_mode(regs))
|
||||||
|
return 0;
|
||||||
|
|
||||||
if (thread_info_flags & _TIF_NEED_RESCHED) {
|
if (thread_info_flags & _TIF_NEED_RESCHED) {
|
||||||
schedule();
|
schedule();
|
||||||
return 1;
|
return 1;
|
||||||
|
@ -589,7 +593,6 @@ int do_work_pending(struct pt_regs *regs, u32 thread_info_flags)
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
if (thread_info_flags & _TIF_SINGLESTEP) {
|
if (thread_info_flags & _TIF_SINGLESTEP) {
|
||||||
if ((regs->ex1 & SPR_EX_CONTEXT_1_1__PL_MASK) == 0)
|
|
||||||
single_step_once(regs);
|
single_step_once(regs);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
|
@ -134,6 +134,9 @@ KBUILD_CFLAGS += $(call cc-option,-mno-avx,)
|
||||||
KBUILD_CFLAGS += $(mflags-y)
|
KBUILD_CFLAGS += $(mflags-y)
|
||||||
KBUILD_AFLAGS += $(mflags-y)
|
KBUILD_AFLAGS += $(mflags-y)
|
||||||
|
|
||||||
|
archscripts:
|
||||||
|
$(Q)$(MAKE) $(build)=arch/x86/tools relocs
|
||||||
|
|
||||||
###
|
###
|
||||||
# Syscall table generation
|
# Syscall table generation
|
||||||
|
|
||||||
|
|
|
@ -40,13 +40,12 @@ OBJCOPYFLAGS_vmlinux.bin := -R .comment -S
|
||||||
$(obj)/vmlinux.bin: vmlinux FORCE
|
$(obj)/vmlinux.bin: vmlinux FORCE
|
||||||
$(call if_changed,objcopy)
|
$(call if_changed,objcopy)
|
||||||
|
|
||||||
|
targets += vmlinux.bin.all vmlinux.relocs
|
||||||
|
|
||||||
targets += vmlinux.bin.all vmlinux.relocs relocs
|
CMD_RELOCS = arch/x86/tools/relocs
|
||||||
hostprogs-$(CONFIG_X86_NEED_RELOCS) += relocs
|
|
||||||
|
|
||||||
quiet_cmd_relocs = RELOCS $@
|
quiet_cmd_relocs = RELOCS $@
|
||||||
cmd_relocs = $(obj)/relocs $< > $@;$(obj)/relocs --abs-relocs $<
|
cmd_relocs = $(CMD_RELOCS) $< > $@;$(CMD_RELOCS) --abs-relocs $<
|
||||||
$(obj)/vmlinux.relocs: vmlinux $(obj)/relocs FORCE
|
$(obj)/vmlinux.relocs: vmlinux FORCE
|
||||||
$(call if_changed,relocs)
|
$(call if_changed,relocs)
|
||||||
|
|
||||||
vmlinux.bin.all-y := $(obj)/vmlinux.bin
|
vmlinux.bin.all-y := $(obj)/vmlinux.bin
|
||||||
|
|
|
@ -170,6 +170,9 @@ static inline int kvm_para_available(void)
|
||||||
unsigned int eax, ebx, ecx, edx;
|
unsigned int eax, ebx, ecx, edx;
|
||||||
char signature[13];
|
char signature[13];
|
||||||
|
|
||||||
|
if (boot_cpu_data.cpuid_level < 0)
|
||||||
|
return 0; /* So we don't blow up on old processors */
|
||||||
|
|
||||||
cpuid(KVM_CPUID_SIGNATURE, &eax, &ebx, &ecx, &edx);
|
cpuid(KVM_CPUID_SIGNATURE, &eax, &ebx, &ecx, &edx);
|
||||||
memcpy(signature + 0, &ebx, 4);
|
memcpy(signature + 0, &ebx, 4);
|
||||||
memcpy(signature + 4, &ecx, 4);
|
memcpy(signature + 4, &ecx, 4);
|
||||||
|
|
|
@ -593,7 +593,7 @@ void __init acpi_set_irq_model_ioapic(void)
|
||||||
#ifdef CONFIG_ACPI_HOTPLUG_CPU
|
#ifdef CONFIG_ACPI_HOTPLUG_CPU
|
||||||
#include <acpi/processor.h>
|
#include <acpi/processor.h>
|
||||||
|
|
||||||
static void __cpuinitdata acpi_map_cpu2node(acpi_handle handle, int cpu, int physid)
|
static void __cpuinit acpi_map_cpu2node(acpi_handle handle, int cpu, int physid)
|
||||||
{
|
{
|
||||||
#ifdef CONFIG_ACPI_NUMA
|
#ifdef CONFIG_ACPI_NUMA
|
||||||
int nid;
|
int nid;
|
||||||
|
|
|
@ -945,9 +945,10 @@ struct mce_info {
|
||||||
atomic_t inuse;
|
atomic_t inuse;
|
||||||
struct task_struct *t;
|
struct task_struct *t;
|
||||||
__u64 paddr;
|
__u64 paddr;
|
||||||
|
int restartable;
|
||||||
} mce_info[MCE_INFO_MAX];
|
} mce_info[MCE_INFO_MAX];
|
||||||
|
|
||||||
static void mce_save_info(__u64 addr)
|
static void mce_save_info(__u64 addr, int c)
|
||||||
{
|
{
|
||||||
struct mce_info *mi;
|
struct mce_info *mi;
|
||||||
|
|
||||||
|
@ -955,6 +956,7 @@ static void mce_save_info(__u64 addr)
|
||||||
if (atomic_cmpxchg(&mi->inuse, 0, 1) == 0) {
|
if (atomic_cmpxchg(&mi->inuse, 0, 1) == 0) {
|
||||||
mi->t = current;
|
mi->t = current;
|
||||||
mi->paddr = addr;
|
mi->paddr = addr;
|
||||||
|
mi->restartable = c;
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1130,7 +1132,7 @@ void do_machine_check(struct pt_regs *regs, long error_code)
|
||||||
mce_panic("Fatal machine check on current CPU", &m, msg);
|
mce_panic("Fatal machine check on current CPU", &m, msg);
|
||||||
if (worst == MCE_AR_SEVERITY) {
|
if (worst == MCE_AR_SEVERITY) {
|
||||||
/* schedule action before return to userland */
|
/* schedule action before return to userland */
|
||||||
mce_save_info(m.addr);
|
mce_save_info(m.addr, m.mcgstatus & MCG_STATUS_RIPV);
|
||||||
set_thread_flag(TIF_MCE_NOTIFY);
|
set_thread_flag(TIF_MCE_NOTIFY);
|
||||||
} else if (kill_it) {
|
} else if (kill_it) {
|
||||||
force_sig(SIGBUS, current);
|
force_sig(SIGBUS, current);
|
||||||
|
@ -1179,7 +1181,13 @@ void mce_notify_process(void)
|
||||||
|
|
||||||
pr_err("Uncorrected hardware memory error in user-access at %llx",
|
pr_err("Uncorrected hardware memory error in user-access at %llx",
|
||||||
mi->paddr);
|
mi->paddr);
|
||||||
if (memory_failure(pfn, MCE_VECTOR, MF_ACTION_REQUIRED) < 0) {
|
/*
|
||||||
|
* We must call memory_failure() here even if the current process is
|
||||||
|
* doomed. We still need to mark the page as poisoned and alert any
|
||||||
|
* other users of the page.
|
||||||
|
*/
|
||||||
|
if (memory_failure(pfn, MCE_VECTOR, MF_ACTION_REQUIRED) < 0 ||
|
||||||
|
mi->restartable == 0) {
|
||||||
pr_err("Memory error not recovered");
|
pr_err("Memory error not recovered");
|
||||||
force_sig(SIGBUS, current);
|
force_sig(SIGBUS, current);
|
||||||
}
|
}
|
||||||
|
|
|
@ -147,12 +147,6 @@ static int collect_cpu_info(int cpu_num, struct cpu_signature *csig)
|
||||||
|
|
||||||
memset(csig, 0, sizeof(*csig));
|
memset(csig, 0, sizeof(*csig));
|
||||||
|
|
||||||
if (c->x86_vendor != X86_VENDOR_INTEL || c->x86 < 6 ||
|
|
||||||
cpu_has(c, X86_FEATURE_IA64)) {
|
|
||||||
pr_err("CPU%d not a capable Intel processor\n", cpu_num);
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
csig->sig = cpuid_eax(0x00000001);
|
csig->sig = cpuid_eax(0x00000001);
|
||||||
|
|
||||||
if ((c->x86_model >= 5) || (c->x86 > 6)) {
|
if ((c->x86_model >= 5) || (c->x86 > 6)) {
|
||||||
|
@ -463,6 +457,14 @@ static struct microcode_ops microcode_intel_ops = {
|
||||||
|
|
||||||
struct microcode_ops * __init init_intel_microcode(void)
|
struct microcode_ops * __init init_intel_microcode(void)
|
||||||
{
|
{
|
||||||
|
struct cpuinfo_x86 *c = &cpu_data(0);
|
||||||
|
|
||||||
|
if (c->x86_vendor != X86_VENDOR_INTEL || c->x86 < 6 ||
|
||||||
|
cpu_has(c, X86_FEATURE_IA64)) {
|
||||||
|
pr_err("Intel CPU family 0x%x not supported\n", c->x86);
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
return µcode_intel_ops;
|
return µcode_intel_ops;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -0,0 +1 @@
|
||||||
|
relocs
|
|
@ -36,3 +36,7 @@ HOSTCFLAGS_insn_sanity.o := -Wall -I$(objtree)/arch/x86/lib/ -I$(srctree)/arch/x
|
||||||
$(obj)/test_get_len.o: $(srctree)/arch/x86/lib/insn.c $(srctree)/arch/x86/lib/inat.c $(srctree)/arch/x86/include/asm/inat_types.h $(srctree)/arch/x86/include/asm/inat.h $(srctree)/arch/x86/include/asm/insn.h $(objtree)/arch/x86/lib/inat-tables.c
|
$(obj)/test_get_len.o: $(srctree)/arch/x86/lib/insn.c $(srctree)/arch/x86/lib/inat.c $(srctree)/arch/x86/include/asm/inat_types.h $(srctree)/arch/x86/include/asm/inat.h $(srctree)/arch/x86/include/asm/insn.h $(objtree)/arch/x86/lib/inat-tables.c
|
||||||
|
|
||||||
$(obj)/insn_sanity.o: $(srctree)/arch/x86/lib/insn.c $(srctree)/arch/x86/lib/inat.c $(srctree)/arch/x86/include/asm/inat_types.h $(srctree)/arch/x86/include/asm/inat.h $(srctree)/arch/x86/include/asm/insn.h $(objtree)/arch/x86/lib/inat-tables.c
|
$(obj)/insn_sanity.o: $(srctree)/arch/x86/lib/insn.c $(srctree)/arch/x86/lib/inat.c $(srctree)/arch/x86/include/asm/inat_types.h $(srctree)/arch/x86/include/asm/inat.h $(srctree)/arch/x86/include/asm/insn.h $(objtree)/arch/x86/lib/inat-tables.c
|
||||||
|
|
||||||
|
HOST_EXTRACFLAGS += -I$(srctree)/tools/include
|
||||||
|
hostprogs-y += relocs
|
||||||
|
relocs: $(obj)/relocs
|
||||||
|
|
|
@ -18,6 +18,8 @@ static void die(char *fmt, ...);
|
||||||
static Elf32_Ehdr ehdr;
|
static Elf32_Ehdr ehdr;
|
||||||
static unsigned long reloc_count, reloc_idx;
|
static unsigned long reloc_count, reloc_idx;
|
||||||
static unsigned long *relocs;
|
static unsigned long *relocs;
|
||||||
|
static unsigned long reloc16_count, reloc16_idx;
|
||||||
|
static unsigned long *relocs16;
|
||||||
|
|
||||||
struct section {
|
struct section {
|
||||||
Elf32_Shdr shdr;
|
Elf32_Shdr shdr;
|
||||||
|
@ -28,52 +30,86 @@ struct section {
|
||||||
};
|
};
|
||||||
static struct section *secs;
|
static struct section *secs;
|
||||||
|
|
||||||
|
enum symtype {
|
||||||
|
S_ABS,
|
||||||
|
S_REL,
|
||||||
|
S_SEG,
|
||||||
|
S_LIN,
|
||||||
|
S_NSYMTYPES
|
||||||
|
};
|
||||||
|
|
||||||
|
static const char * const sym_regex_kernel[S_NSYMTYPES] = {
|
||||||
/*
|
/*
|
||||||
* Following symbols have been audited. There values are constant and do
|
* Following symbols have been audited. There values are constant and do
|
||||||
* not change if bzImage is loaded at a different physical address than
|
* not change if bzImage is loaded at a different physical address than
|
||||||
* the address for which it has been compiled. Don't warn user about
|
* the address for which it has been compiled. Don't warn user about
|
||||||
* absolute relocations present w.r.t these symbols.
|
* absolute relocations present w.r.t these symbols.
|
||||||
*/
|
*/
|
||||||
static const char abs_sym_regex[] =
|
[S_ABS] =
|
||||||
"^(xen_irq_disable_direct_reloc$|"
|
"^(xen_irq_disable_direct_reloc$|"
|
||||||
"xen_save_fl_direct_reloc$|"
|
"xen_save_fl_direct_reloc$|"
|
||||||
"VDSO|"
|
"VDSO|"
|
||||||
"__crc_)";
|
"__crc_)",
|
||||||
static regex_t abs_sym_regex_c;
|
|
||||||
static int is_abs_reloc(const char *sym_name)
|
|
||||||
{
|
|
||||||
return !regexec(&abs_sym_regex_c, sym_name, 0, NULL, 0);
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* These symbols are known to be relative, even if the linker marks them
|
* These symbols are known to be relative, even if the linker marks them
|
||||||
* as absolute (typically defined outside any section in the linker script.)
|
* as absolute (typically defined outside any section in the linker script.)
|
||||||
*/
|
*/
|
||||||
static const char rel_sym_regex[] =
|
[S_REL] =
|
||||||
"^_end$";
|
"^(__init_(begin|end)|"
|
||||||
static regex_t rel_sym_regex_c;
|
"__x86_cpu_dev_(start|end)|"
|
||||||
static int is_rel_reloc(const char *sym_name)
|
"(__parainstructions|__alt_instructions)(|_end)|"
|
||||||
|
"(__iommu_table|__apicdrivers|__smp_locks)(|_end)|"
|
||||||
|
"_end)$"
|
||||||
|
};
|
||||||
|
|
||||||
|
|
||||||
|
static const char * const sym_regex_realmode[S_NSYMTYPES] = {
|
||||||
|
/*
|
||||||
|
* These are 16-bit segment symbols when compiling 16-bit code.
|
||||||
|
*/
|
||||||
|
[S_SEG] =
|
||||||
|
"^real_mode_seg$",
|
||||||
|
|
||||||
|
/*
|
||||||
|
* These are offsets belonging to segments, as opposed to linear addresses,
|
||||||
|
* when compiling 16-bit code.
|
||||||
|
*/
|
||||||
|
[S_LIN] =
|
||||||
|
"^pa_",
|
||||||
|
};
|
||||||
|
|
||||||
|
static const char * const *sym_regex;
|
||||||
|
|
||||||
|
static regex_t sym_regex_c[S_NSYMTYPES];
|
||||||
|
static int is_reloc(enum symtype type, const char *sym_name)
|
||||||
{
|
{
|
||||||
return !regexec(&rel_sym_regex_c, sym_name, 0, NULL, 0);
|
return sym_regex[type] &&
|
||||||
|
!regexec(&sym_regex_c[type], sym_name, 0, NULL, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void regex_init(void)
|
static void regex_init(int use_real_mode)
|
||||||
{
|
{
|
||||||
char errbuf[128];
|
char errbuf[128];
|
||||||
int err;
|
int err;
|
||||||
|
int i;
|
||||||
|
|
||||||
err = regcomp(&abs_sym_regex_c, abs_sym_regex,
|
if (use_real_mode)
|
||||||
|
sym_regex = sym_regex_realmode;
|
||||||
|
else
|
||||||
|
sym_regex = sym_regex_kernel;
|
||||||
|
|
||||||
|
for (i = 0; i < S_NSYMTYPES; i++) {
|
||||||
|
if (!sym_regex[i])
|
||||||
|
continue;
|
||||||
|
|
||||||
|
err = regcomp(&sym_regex_c[i], sym_regex[i],
|
||||||
REG_EXTENDED|REG_NOSUB);
|
REG_EXTENDED|REG_NOSUB);
|
||||||
|
|
||||||
if (err) {
|
if (err) {
|
||||||
regerror(err, &abs_sym_regex_c, errbuf, sizeof errbuf);
|
regerror(err, &sym_regex_c[i], errbuf, sizeof errbuf);
|
||||||
die("%s", errbuf);
|
die("%s", errbuf);
|
||||||
}
|
}
|
||||||
|
|
||||||
err = regcomp(&rel_sym_regex_c, rel_sym_regex,
|
|
||||||
REG_EXTENDED|REG_NOSUB);
|
|
||||||
if (err) {
|
|
||||||
regerror(err, &rel_sym_regex_c, errbuf, sizeof errbuf);
|
|
||||||
die("%s", errbuf);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -154,6 +190,10 @@ static const char *rel_type(unsigned type)
|
||||||
REL_TYPE(R_386_RELATIVE),
|
REL_TYPE(R_386_RELATIVE),
|
||||||
REL_TYPE(R_386_GOTOFF),
|
REL_TYPE(R_386_GOTOFF),
|
||||||
REL_TYPE(R_386_GOTPC),
|
REL_TYPE(R_386_GOTPC),
|
||||||
|
REL_TYPE(R_386_8),
|
||||||
|
REL_TYPE(R_386_PC8),
|
||||||
|
REL_TYPE(R_386_16),
|
||||||
|
REL_TYPE(R_386_PC16),
|
||||||
#undef REL_TYPE
|
#undef REL_TYPE
|
||||||
};
|
};
|
||||||
const char *name = "unknown type rel type name";
|
const char *name = "unknown type rel type name";
|
||||||
|
@ -189,7 +229,7 @@ static const char *sym_name(const char *sym_strtab, Elf32_Sym *sym)
|
||||||
name = sym_strtab + sym->st_name;
|
name = sym_strtab + sym->st_name;
|
||||||
}
|
}
|
||||||
else {
|
else {
|
||||||
name = sec_name(secs[sym->st_shndx].shdr.sh_name);
|
name = sec_name(sym->st_shndx);
|
||||||
}
|
}
|
||||||
return name;
|
return name;
|
||||||
}
|
}
|
||||||
|
@ -472,7 +512,7 @@ static void print_absolute_relocs(void)
|
||||||
* Before warning check if this absolute symbol
|
* Before warning check if this absolute symbol
|
||||||
* relocation is harmless.
|
* relocation is harmless.
|
||||||
*/
|
*/
|
||||||
if (is_abs_reloc(name) || is_rel_reloc(name))
|
if (is_reloc(S_ABS, name) || is_reloc(S_REL, name))
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
if (!printed) {
|
if (!printed) {
|
||||||
|
@ -496,7 +536,8 @@ static void print_absolute_relocs(void)
|
||||||
printf("\n");
|
printf("\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
static void walk_relocs(void (*visit)(Elf32_Rel *rel, Elf32_Sym *sym))
|
static void walk_relocs(void (*visit)(Elf32_Rel *rel, Elf32_Sym *sym),
|
||||||
|
int use_real_mode)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
/* Walk through the relocations */
|
/* Walk through the relocations */
|
||||||
|
@ -521,30 +562,67 @@ static void walk_relocs(void (*visit)(Elf32_Rel *rel, Elf32_Sym *sym))
|
||||||
Elf32_Rel *rel;
|
Elf32_Rel *rel;
|
||||||
Elf32_Sym *sym;
|
Elf32_Sym *sym;
|
||||||
unsigned r_type;
|
unsigned r_type;
|
||||||
|
const char *symname;
|
||||||
|
int shn_abs;
|
||||||
|
|
||||||
rel = &sec->reltab[j];
|
rel = &sec->reltab[j];
|
||||||
sym = &sh_symtab[ELF32_R_SYM(rel->r_info)];
|
sym = &sh_symtab[ELF32_R_SYM(rel->r_info)];
|
||||||
r_type = ELF32_R_TYPE(rel->r_info);
|
r_type = ELF32_R_TYPE(rel->r_info);
|
||||||
/* Don't visit relocations to absolute symbols */
|
|
||||||
if (sym->st_shndx == SHN_ABS &&
|
shn_abs = sym->st_shndx == SHN_ABS;
|
||||||
!is_rel_reloc(sym_name(sym_strtab, sym))) {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
switch (r_type) {
|
switch (r_type) {
|
||||||
case R_386_NONE:
|
case R_386_NONE:
|
||||||
case R_386_PC32:
|
case R_386_PC32:
|
||||||
|
case R_386_PC16:
|
||||||
|
case R_386_PC8:
|
||||||
/*
|
/*
|
||||||
* NONE can be ignored and and PC relative
|
* NONE can be ignored and and PC relative
|
||||||
* relocations don't need to be adjusted.
|
* relocations don't need to be adjusted.
|
||||||
*/
|
*/
|
||||||
break;
|
break;
|
||||||
|
|
||||||
|
case R_386_16:
|
||||||
|
symname = sym_name(sym_strtab, sym);
|
||||||
|
if (!use_real_mode)
|
||||||
|
goto bad;
|
||||||
|
if (shn_abs) {
|
||||||
|
if (is_reloc(S_ABS, symname))
|
||||||
|
break;
|
||||||
|
else if (!is_reloc(S_SEG, symname))
|
||||||
|
goto bad;
|
||||||
|
} else {
|
||||||
|
if (is_reloc(S_LIN, symname))
|
||||||
|
goto bad;
|
||||||
|
else
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
visit(rel, sym);
|
||||||
|
break;
|
||||||
|
|
||||||
case R_386_32:
|
case R_386_32:
|
||||||
/* Visit relocations that need to be adjusted */
|
symname = sym_name(sym_strtab, sym);
|
||||||
|
if (shn_abs) {
|
||||||
|
if (is_reloc(S_ABS, symname))
|
||||||
|
break;
|
||||||
|
else if (!is_reloc(S_REL, symname))
|
||||||
|
goto bad;
|
||||||
|
} else {
|
||||||
|
if (use_real_mode &&
|
||||||
|
!is_reloc(S_LIN, symname))
|
||||||
|
break;
|
||||||
|
}
|
||||||
visit(rel, sym);
|
visit(rel, sym);
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
die("Unsupported relocation type: %s (%d)\n",
|
die("Unsupported relocation type: %s (%d)\n",
|
||||||
rel_type(r_type), r_type);
|
rel_type(r_type), r_type);
|
||||||
break;
|
break;
|
||||||
|
bad:
|
||||||
|
symname = sym_name(sym_strtab, sym);
|
||||||
|
die("Invalid %s %s relocation: %s\n",
|
||||||
|
shn_abs ? "absolute" : "relative",
|
||||||
|
rel_type(r_type), symname);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -552,12 +630,18 @@ static void walk_relocs(void (*visit)(Elf32_Rel *rel, Elf32_Sym *sym))
|
||||||
|
|
||||||
static void count_reloc(Elf32_Rel *rel, Elf32_Sym *sym)
|
static void count_reloc(Elf32_Rel *rel, Elf32_Sym *sym)
|
||||||
{
|
{
|
||||||
reloc_count += 1;
|
if (ELF32_R_TYPE(rel->r_info) == R_386_16)
|
||||||
|
reloc16_count++;
|
||||||
|
else
|
||||||
|
reloc_count++;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void collect_reloc(Elf32_Rel *rel, Elf32_Sym *sym)
|
static void collect_reloc(Elf32_Rel *rel, Elf32_Sym *sym)
|
||||||
{
|
{
|
||||||
/* Remember the address that needs to be adjusted. */
|
/* Remember the address that needs to be adjusted. */
|
||||||
|
if (ELF32_R_TYPE(rel->r_info) == R_386_16)
|
||||||
|
relocs16[reloc16_idx++] = rel->r_offset;
|
||||||
|
else
|
||||||
relocs[reloc_idx++] = rel->r_offset;
|
relocs[reloc_idx++] = rel->r_offset;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -568,23 +652,41 @@ static int cmp_relocs(const void *va, const void *vb)
|
||||||
return (*a == *b)? 0 : (*a > *b)? 1 : -1;
|
return (*a == *b)? 0 : (*a > *b)? 1 : -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void emit_relocs(int as_text)
|
static int write32(unsigned int v, FILE *f)
|
||||||
|
{
|
||||||
|
unsigned char buf[4];
|
||||||
|
|
||||||
|
put_unaligned_le32(v, buf);
|
||||||
|
return fwrite(buf, 1, 4, f) == 4 ? 0 : -1;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void emit_relocs(int as_text, int use_real_mode)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
/* Count how many relocations I have and allocate space for them. */
|
/* Count how many relocations I have and allocate space for them. */
|
||||||
reloc_count = 0;
|
reloc_count = 0;
|
||||||
walk_relocs(count_reloc);
|
walk_relocs(count_reloc, use_real_mode);
|
||||||
relocs = malloc(reloc_count * sizeof(relocs[0]));
|
relocs = malloc(reloc_count * sizeof(relocs[0]));
|
||||||
if (!relocs) {
|
if (!relocs) {
|
||||||
die("malloc of %d entries for relocs failed\n",
|
die("malloc of %d entries for relocs failed\n",
|
||||||
reloc_count);
|
reloc_count);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
relocs16 = malloc(reloc16_count * sizeof(relocs[0]));
|
||||||
|
if (!relocs16) {
|
||||||
|
die("malloc of %d entries for relocs16 failed\n",
|
||||||
|
reloc16_count);
|
||||||
|
}
|
||||||
/* Collect up the relocations */
|
/* Collect up the relocations */
|
||||||
reloc_idx = 0;
|
reloc_idx = 0;
|
||||||
walk_relocs(collect_reloc);
|
walk_relocs(collect_reloc, use_real_mode);
|
||||||
|
|
||||||
|
if (reloc16_count && !use_real_mode)
|
||||||
|
die("Segment relocations found but --realmode not specified\n");
|
||||||
|
|
||||||
/* Order the relocations for more efficient processing */
|
/* Order the relocations for more efficient processing */
|
||||||
qsort(relocs, reloc_count, sizeof(relocs[0]), cmp_relocs);
|
qsort(relocs, reloc_count, sizeof(relocs[0]), cmp_relocs);
|
||||||
|
qsort(relocs16, reloc16_count, sizeof(relocs16[0]), cmp_relocs);
|
||||||
|
|
||||||
/* Print the relocations */
|
/* Print the relocations */
|
||||||
if (as_text) {
|
if (as_text) {
|
||||||
|
@ -593,58 +695,83 @@ static void emit_relocs(int as_text)
|
||||||
*/
|
*/
|
||||||
printf(".section \".data.reloc\",\"a\"\n");
|
printf(".section \".data.reloc\",\"a\"\n");
|
||||||
printf(".balign 4\n");
|
printf(".balign 4\n");
|
||||||
|
if (use_real_mode) {
|
||||||
|
printf("\t.long %lu\n", reloc16_count);
|
||||||
|
for (i = 0; i < reloc16_count; i++)
|
||||||
|
printf("\t.long 0x%08lx\n", relocs16[i]);
|
||||||
|
printf("\t.long %lu\n", reloc_count);
|
||||||
for (i = 0; i < reloc_count; i++) {
|
for (i = 0; i < reloc_count; i++) {
|
||||||
printf("\t .long 0x%08lx\n", relocs[i]);
|
printf("\t.long 0x%08lx\n", relocs[i]);
|
||||||
}
|
}
|
||||||
|
} else {
|
||||||
|
/* Print a stop */
|
||||||
|
printf("\t.long 0x%08lx\n", (unsigned long)0);
|
||||||
|
for (i = 0; i < reloc_count; i++) {
|
||||||
|
printf("\t.long 0x%08lx\n", relocs[i]);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
printf("\n");
|
printf("\n");
|
||||||
}
|
}
|
||||||
else {
|
else {
|
||||||
unsigned char buf[4];
|
if (use_real_mode) {
|
||||||
|
write32(reloc16_count, stdout);
|
||||||
|
for (i = 0; i < reloc16_count; i++)
|
||||||
|
write32(relocs16[i], stdout);
|
||||||
|
write32(reloc_count, stdout);
|
||||||
|
|
||||||
|
/* Now print each relocation */
|
||||||
|
for (i = 0; i < reloc_count; i++)
|
||||||
|
write32(relocs[i], stdout);
|
||||||
|
} else {
|
||||||
/* Print a stop */
|
/* Print a stop */
|
||||||
fwrite("\0\0\0\0", 4, 1, stdout);
|
write32(0, stdout);
|
||||||
|
|
||||||
/* Now print each relocation */
|
/* Now print each relocation */
|
||||||
for (i = 0; i < reloc_count; i++) {
|
for (i = 0; i < reloc_count; i++) {
|
||||||
put_unaligned_le32(relocs[i], buf);
|
write32(relocs[i], stdout);
|
||||||
fwrite(buf, 4, 1, stdout);
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void usage(void)
|
static void usage(void)
|
||||||
{
|
{
|
||||||
die("relocs [--abs-syms |--abs-relocs | --text] vmlinux\n");
|
die("relocs [--abs-syms|--abs-relocs|--text|--realmode] vmlinux\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
int main(int argc, char **argv)
|
int main(int argc, char **argv)
|
||||||
{
|
{
|
||||||
int show_absolute_syms, show_absolute_relocs;
|
int show_absolute_syms, show_absolute_relocs;
|
||||||
int as_text;
|
int as_text, use_real_mode;
|
||||||
const char *fname;
|
const char *fname;
|
||||||
FILE *fp;
|
FILE *fp;
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
regex_init();
|
|
||||||
|
|
||||||
show_absolute_syms = 0;
|
show_absolute_syms = 0;
|
||||||
show_absolute_relocs = 0;
|
show_absolute_relocs = 0;
|
||||||
as_text = 0;
|
as_text = 0;
|
||||||
|
use_real_mode = 0;
|
||||||
fname = NULL;
|
fname = NULL;
|
||||||
for (i = 1; i < argc; i++) {
|
for (i = 1; i < argc; i++) {
|
||||||
char *arg = argv[i];
|
char *arg = argv[i];
|
||||||
if (*arg == '-') {
|
if (*arg == '-') {
|
||||||
if (strcmp(argv[1], "--abs-syms") == 0) {
|
if (strcmp(arg, "--abs-syms") == 0) {
|
||||||
show_absolute_syms = 1;
|
show_absolute_syms = 1;
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
if (strcmp(arg, "--abs-relocs") == 0) {
|
||||||
if (strcmp(argv[1], "--abs-relocs") == 0) {
|
|
||||||
show_absolute_relocs = 1;
|
show_absolute_relocs = 1;
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
else if (strcmp(argv[1], "--text") == 0) {
|
if (strcmp(arg, "--text") == 0) {
|
||||||
as_text = 1;
|
as_text = 1;
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
if (strcmp(arg, "--realmode") == 0) {
|
||||||
|
use_real_mode = 1;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
else if (!fname) {
|
else if (!fname) {
|
||||||
fname = arg;
|
fname = arg;
|
||||||
|
@ -655,6 +782,7 @@ int main(int argc, char **argv)
|
||||||
if (!fname) {
|
if (!fname) {
|
||||||
usage();
|
usage();
|
||||||
}
|
}
|
||||||
|
regex_init(use_real_mode);
|
||||||
fp = fopen(fname, "r");
|
fp = fopen(fname, "r");
|
||||||
if (!fp) {
|
if (!fp) {
|
||||||
die("Cannot open %s: %s\n",
|
die("Cannot open %s: %s\n",
|
||||||
|
@ -673,6 +801,6 @@ int main(int argc, char **argv)
|
||||||
print_absolute_relocs();
|
print_absolute_relocs();
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
emit_relocs(as_text);
|
emit_relocs(as_text, use_real_mode);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
|
@ -743,7 +743,7 @@ void __init printk_all_partitions(void)
|
||||||
struct hd_struct *part;
|
struct hd_struct *part;
|
||||||
char name_buf[BDEVNAME_SIZE];
|
char name_buf[BDEVNAME_SIZE];
|
||||||
char devt_buf[BDEVT_SIZE];
|
char devt_buf[BDEVT_SIZE];
|
||||||
u8 uuid[PARTITION_META_INFO_UUIDLTH * 2 + 1];
|
char uuid_buf[PARTITION_META_INFO_UUIDLTH * 2 + 5];
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Don't show empty devices or things that have been
|
* Don't show empty devices or things that have been
|
||||||
|
@ -762,14 +762,16 @@ void __init printk_all_partitions(void)
|
||||||
while ((part = disk_part_iter_next(&piter))) {
|
while ((part = disk_part_iter_next(&piter))) {
|
||||||
bool is_part0 = part == &disk->part0;
|
bool is_part0 = part == &disk->part0;
|
||||||
|
|
||||||
uuid[0] = 0;
|
uuid_buf[0] = '\0';
|
||||||
if (part->info)
|
if (part->info)
|
||||||
part_unpack_uuid(part->info->uuid, uuid);
|
snprintf(uuid_buf, sizeof(uuid_buf), "%pU",
|
||||||
|
part->info->uuid);
|
||||||
|
|
||||||
printk("%s%s %10llu %s %s", is_part0 ? "" : " ",
|
printk("%s%s %10llu %s %s", is_part0 ? "" : " ",
|
||||||
bdevt_str(part_devt(part), devt_buf),
|
bdevt_str(part_devt(part), devt_buf),
|
||||||
(unsigned long long)part->nr_sects >> 1,
|
(unsigned long long)part->nr_sects >> 1,
|
||||||
disk_name(disk, part->partno, name_buf), uuid);
|
disk_name(disk, part->partno, name_buf),
|
||||||
|
uuid_buf);
|
||||||
if (is_part0) {
|
if (is_part0) {
|
||||||
if (disk->driverfs_dev != NULL &&
|
if (disk->driverfs_dev != NULL &&
|
||||||
disk->driverfs_dev->driver != NULL)
|
disk->driverfs_dev->driver != NULL)
|
||||||
|
|
|
@ -250,6 +250,10 @@ static int __acpi_bus_set_power(struct acpi_device *device, int state)
|
||||||
return -ENODEV;
|
return -ENODEV;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* For D3cold we should execute _PS3, not _PS4. */
|
||||||
|
if (state == ACPI_STATE_D3_COLD)
|
||||||
|
object_name[3] = '3';
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Transition Power
|
* Transition Power
|
||||||
* ----------------
|
* ----------------
|
||||||
|
|
|
@ -660,7 +660,7 @@ int acpi_power_on_resources(struct acpi_device *device, int state)
|
||||||
|
|
||||||
int acpi_power_transition(struct acpi_device *device, int state)
|
int acpi_power_transition(struct acpi_device *device, int state)
|
||||||
{
|
{
|
||||||
int result;
|
int result = 0;
|
||||||
|
|
||||||
if (!device || (state < ACPI_STATE_D0) || (state > ACPI_STATE_D3_COLD))
|
if (!device || (state < ACPI_STATE_D0) || (state > ACPI_STATE_D3_COLD))
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
@ -679,8 +679,11 @@ int acpi_power_transition(struct acpi_device *device, int state)
|
||||||
* (e.g. so the device doesn't lose power while transitioning). Then,
|
* (e.g. so the device doesn't lose power while transitioning). Then,
|
||||||
* we dereference all power resources used in the current list.
|
* we dereference all power resources used in the current list.
|
||||||
*/
|
*/
|
||||||
result = acpi_power_on_list(&device->power.states[state].resources);
|
if (state < ACPI_STATE_D3_COLD)
|
||||||
if (!result)
|
result = acpi_power_on_list(
|
||||||
|
&device->power.states[state].resources);
|
||||||
|
|
||||||
|
if (!result && device->power.state < ACPI_STATE_D3_COLD)
|
||||||
acpi_power_off_list(
|
acpi_power_off_list(
|
||||||
&device->power.states[device->power.state].resources);
|
&device->power.states[device->power.state].resources);
|
||||||
|
|
||||||
|
|
|
@ -908,6 +908,10 @@ static int acpi_bus_get_power_flags(struct acpi_device *device)
|
||||||
device->power.states[ACPI_STATE_D3].flags.valid = 1;
|
device->power.states[ACPI_STATE_D3].flags.valid = 1;
|
||||||
device->power.states[ACPI_STATE_D3].power = 0;
|
device->power.states[ACPI_STATE_D3].power = 0;
|
||||||
|
|
||||||
|
/* Set D3cold's explicit_set flag if _PS3 exists. */
|
||||||
|
if (device->power.states[ACPI_STATE_D3_HOT].flags.explicit_set)
|
||||||
|
device->power.states[ACPI_STATE_D3_COLD].flags.explicit_set = 1;
|
||||||
|
|
||||||
acpi_bus_init_power(device);
|
acpi_bus_init_power(device);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
|
@ -6580,14 +6580,11 @@ static const struct file_operations dac960_user_command_proc_fops = {
|
||||||
|
|
||||||
static void DAC960_CreateProcEntries(DAC960_Controller_T *Controller)
|
static void DAC960_CreateProcEntries(DAC960_Controller_T *Controller)
|
||||||
{
|
{
|
||||||
struct proc_dir_entry *StatusProcEntry;
|
|
||||||
struct proc_dir_entry *ControllerProcEntry;
|
struct proc_dir_entry *ControllerProcEntry;
|
||||||
struct proc_dir_entry *UserCommandProcEntry;
|
|
||||||
|
|
||||||
if (DAC960_ProcDirectoryEntry == NULL) {
|
if (DAC960_ProcDirectoryEntry == NULL) {
|
||||||
DAC960_ProcDirectoryEntry = proc_mkdir("rd", NULL);
|
DAC960_ProcDirectoryEntry = proc_mkdir("rd", NULL);
|
||||||
StatusProcEntry = proc_create("status", 0,
|
proc_create("status", 0, DAC960_ProcDirectoryEntry,
|
||||||
DAC960_ProcDirectoryEntry,
|
|
||||||
&dac960_proc_fops);
|
&dac960_proc_fops);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -6596,7 +6593,7 @@ static void DAC960_CreateProcEntries(DAC960_Controller_T *Controller)
|
||||||
DAC960_ProcDirectoryEntry);
|
DAC960_ProcDirectoryEntry);
|
||||||
proc_create_data("initial_status", 0, ControllerProcEntry, &dac960_initial_status_proc_fops, Controller);
|
proc_create_data("initial_status", 0, ControllerProcEntry, &dac960_initial_status_proc_fops, Controller);
|
||||||
proc_create_data("current_status", 0, ControllerProcEntry, &dac960_current_status_proc_fops, Controller);
|
proc_create_data("current_status", 0, ControllerProcEntry, &dac960_current_status_proc_fops, Controller);
|
||||||
UserCommandProcEntry = proc_create_data("user_command", S_IWUSR | S_IRUSR, ControllerProcEntry, &dac960_user_command_proc_fops, Controller);
|
proc_create_data("user_command", S_IWUSR | S_IRUSR, ControllerProcEntry, &dac960_user_command_proc_fops, Controller);
|
||||||
Controller->ControllerProcEntry = ControllerProcEntry;
|
Controller->ControllerProcEntry = ControllerProcEntry;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -2510,8 +2510,10 @@ static struct scatterlist *mtip_hw_get_scatterlist(struct driver_data *dd,
|
||||||
up(&dd->port->cmd_slot);
|
up(&dd->port->cmd_slot);
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
if (unlikely(*tag < 0))
|
if (unlikely(*tag < 0)) {
|
||||||
|
up(&dd->port->cmd_slot);
|
||||||
return NULL;
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
return dd->port->commands[*tag].sg;
|
return dd->port->commands[*tag].sg;
|
||||||
}
|
}
|
||||||
|
|
|
@ -1895,6 +1895,13 @@ static int virtcons_restore(struct virtio_device *vdev)
|
||||||
|
|
||||||
/* Get port open/close status on the host */
|
/* Get port open/close status on the host */
|
||||||
send_control_msg(port, VIRTIO_CONSOLE_PORT_READY, 1);
|
send_control_msg(port, VIRTIO_CONSOLE_PORT_READY, 1);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* If a port was open at the time of suspending, we
|
||||||
|
* have to let the host know that it's still open.
|
||||||
|
*/
|
||||||
|
if (port->guest_connected)
|
||||||
|
send_control_msg(port, VIRTIO_CONSOLE_PORT_OPEN, 1);
|
||||||
}
|
}
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
|
@ -164,6 +164,7 @@ config CRYPTO_DEV_MV_CESA
|
||||||
select CRYPTO_ALGAPI
|
select CRYPTO_ALGAPI
|
||||||
select CRYPTO_AES
|
select CRYPTO_AES
|
||||||
select CRYPTO_BLKCIPHER2
|
select CRYPTO_BLKCIPHER2
|
||||||
|
select CRYPTO_HASH
|
||||||
help
|
help
|
||||||
This driver allows you to utilize the Cryptographic Engines and
|
This driver allows you to utilize the Cryptographic Engines and
|
||||||
Security Accelerator (CESA) which can be found on the Marvell Orion
|
Security Accelerator (CESA) which can be found on the Marvell Orion
|
||||||
|
|
|
@ -245,6 +245,8 @@ atc_chain_complete(struct at_dma_chan *atchan, struct at_desc *desc)
|
||||||
dev_vdbg(chan2dev(&atchan->chan_common),
|
dev_vdbg(chan2dev(&atchan->chan_common),
|
||||||
"descriptor %u complete\n", txd->cookie);
|
"descriptor %u complete\n", txd->cookie);
|
||||||
|
|
||||||
|
/* mark the descriptor as complete for non cyclic cases only */
|
||||||
|
if (!atc_chan_is_cyclic(atchan))
|
||||||
dma_cookie_complete(txd);
|
dma_cookie_complete(txd);
|
||||||
|
|
||||||
/* move children to free_list */
|
/* move children to free_list */
|
||||||
|
|
|
@ -703,6 +703,8 @@ static void ep93xx_dma_tasklet(unsigned long data)
|
||||||
desc = ep93xx_dma_get_active(edmac);
|
desc = ep93xx_dma_get_active(edmac);
|
||||||
if (desc) {
|
if (desc) {
|
||||||
if (desc->complete) {
|
if (desc->complete) {
|
||||||
|
/* mark descriptor complete for non cyclic case only */
|
||||||
|
if (!test_bit(EP93XX_DMA_IS_CYCLIC, &edmac->flags))
|
||||||
dma_cookie_complete(&desc->txd);
|
dma_cookie_complete(&desc->txd);
|
||||||
list_splice_init(&edmac->active, &list);
|
list_splice_init(&edmac->active, &list);
|
||||||
}
|
}
|
||||||
|
|
|
@ -2322,6 +2322,7 @@ static void pl330_tasklet(unsigned long data)
|
||||||
/* Pick up ripe tomatoes */
|
/* Pick up ripe tomatoes */
|
||||||
list_for_each_entry_safe(desc, _dt, &pch->work_list, node)
|
list_for_each_entry_safe(desc, _dt, &pch->work_list, node)
|
||||||
if (desc->status == DONE) {
|
if (desc->status == DONE) {
|
||||||
|
if (pch->cyclic)
|
||||||
dma_cookie_complete(&desc->txd);
|
dma_cookie_complete(&desc->txd);
|
||||||
list_move_tail(&desc->node, &list);
|
list_move_tail(&desc->node, &list);
|
||||||
}
|
}
|
||||||
|
|
|
@ -1632,6 +1632,21 @@ static int bind_control_target(struct pool *pool, struct dm_target *ti)
|
||||||
pool->low_water_blocks = pt->low_water_blocks;
|
pool->low_water_blocks = pt->low_water_blocks;
|
||||||
pool->pf = pt->pf;
|
pool->pf = pt->pf;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* If discard_passdown was enabled verify that the data device
|
||||||
|
* supports discards. Disable discard_passdown if not; otherwise
|
||||||
|
* -EOPNOTSUPP will be returned.
|
||||||
|
*/
|
||||||
|
if (pt->pf.discard_passdown) {
|
||||||
|
struct request_queue *q = bdev_get_queue(pt->data_dev->bdev);
|
||||||
|
if (!q || !blk_queue_discard(q)) {
|
||||||
|
char buf[BDEVNAME_SIZE];
|
||||||
|
DMWARN("Discard unsupported by data device (%s): Disabling discard passdown.",
|
||||||
|
bdevname(pt->data_dev->bdev, buf));
|
||||||
|
pool->pf.discard_passdown = 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1988,19 +2003,6 @@ static int pool_ctr(struct dm_target *ti, unsigned argc, char **argv)
|
||||||
goto out_flags_changed;
|
goto out_flags_changed;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
|
||||||
* If discard_passdown was enabled verify that the data device
|
|
||||||
* supports discards. Disable discard_passdown if not; otherwise
|
|
||||||
* -EOPNOTSUPP will be returned.
|
|
||||||
*/
|
|
||||||
if (pf.discard_passdown) {
|
|
||||||
struct request_queue *q = bdev_get_queue(data_dev->bdev);
|
|
||||||
if (!q || !blk_queue_discard(q)) {
|
|
||||||
DMWARN("Discard unsupported by data device: Disabling discard passdown.");
|
|
||||||
pf.discard_passdown = 0;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
pt->pool = pool;
|
pt->pool = pool;
|
||||||
pt->ti = ti;
|
pt->ti = ti;
|
||||||
pt->metadata_dev = metadata_dev;
|
pt->metadata_dev = metadata_dev;
|
||||||
|
@ -2385,7 +2387,7 @@ static int pool_status(struct dm_target *ti, status_type_t type,
|
||||||
(unsigned long long)pt->low_water_blocks);
|
(unsigned long long)pt->low_water_blocks);
|
||||||
|
|
||||||
count = !pool->pf.zero_new_blocks + !pool->pf.discard_enabled +
|
count = !pool->pf.zero_new_blocks + !pool->pf.discard_enabled +
|
||||||
!pool->pf.discard_passdown;
|
!pt->pf.discard_passdown;
|
||||||
DMEMIT("%u ", count);
|
DMEMIT("%u ", count);
|
||||||
|
|
||||||
if (!pool->pf.zero_new_blocks)
|
if (!pool->pf.zero_new_blocks)
|
||||||
|
@ -2394,7 +2396,7 @@ static int pool_status(struct dm_target *ti, status_type_t type,
|
||||||
if (!pool->pf.discard_enabled)
|
if (!pool->pf.discard_enabled)
|
||||||
DMEMIT("ignore_discard ");
|
DMEMIT("ignore_discard ");
|
||||||
|
|
||||||
if (!pool->pf.discard_passdown)
|
if (!pt->pf.discard_passdown)
|
||||||
DMEMIT("no_discard_passdown ");
|
DMEMIT("no_discard_passdown ");
|
||||||
|
|
||||||
break;
|
break;
|
||||||
|
|
|
@ -391,6 +391,8 @@ void mddev_suspend(struct mddev *mddev)
|
||||||
synchronize_rcu();
|
synchronize_rcu();
|
||||||
wait_event(mddev->sb_wait, atomic_read(&mddev->active_io) == 0);
|
wait_event(mddev->sb_wait, atomic_read(&mddev->active_io) == 0);
|
||||||
mddev->pers->quiesce(mddev, 1);
|
mddev->pers->quiesce(mddev, 1);
|
||||||
|
|
||||||
|
del_timer_sync(&mddev->safemode_timer);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(mddev_suspend);
|
EXPORT_SYMBOL_GPL(mddev_suspend);
|
||||||
|
|
||||||
|
|
|
@ -3164,12 +3164,40 @@ raid10_size(struct mddev *mddev, sector_t sectors, int raid_disks)
|
||||||
return size << conf->chunk_shift;
|
return size << conf->chunk_shift;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void calc_sectors(struct r10conf *conf, sector_t size)
|
||||||
|
{
|
||||||
|
/* Calculate the number of sectors-per-device that will
|
||||||
|
* actually be used, and set conf->dev_sectors and
|
||||||
|
* conf->stride
|
||||||
|
*/
|
||||||
|
|
||||||
|
size = size >> conf->chunk_shift;
|
||||||
|
sector_div(size, conf->far_copies);
|
||||||
|
size = size * conf->raid_disks;
|
||||||
|
sector_div(size, conf->near_copies);
|
||||||
|
/* 'size' is now the number of chunks in the array */
|
||||||
|
/* calculate "used chunks per device" */
|
||||||
|
size = size * conf->copies;
|
||||||
|
|
||||||
|
/* We need to round up when dividing by raid_disks to
|
||||||
|
* get the stride size.
|
||||||
|
*/
|
||||||
|
size = DIV_ROUND_UP_SECTOR_T(size, conf->raid_disks);
|
||||||
|
|
||||||
|
conf->dev_sectors = size << conf->chunk_shift;
|
||||||
|
|
||||||
|
if (conf->far_offset)
|
||||||
|
conf->stride = 1 << conf->chunk_shift;
|
||||||
|
else {
|
||||||
|
sector_div(size, conf->far_copies);
|
||||||
|
conf->stride = size << conf->chunk_shift;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
static struct r10conf *setup_conf(struct mddev *mddev)
|
static struct r10conf *setup_conf(struct mddev *mddev)
|
||||||
{
|
{
|
||||||
struct r10conf *conf = NULL;
|
struct r10conf *conf = NULL;
|
||||||
int nc, fc, fo;
|
int nc, fc, fo;
|
||||||
sector_t stride, size;
|
|
||||||
int err = -EINVAL;
|
int err = -EINVAL;
|
||||||
|
|
||||||
if (mddev->new_chunk_sectors < (PAGE_SIZE >> 9) ||
|
if (mddev->new_chunk_sectors < (PAGE_SIZE >> 9) ||
|
||||||
|
@ -3219,28 +3247,7 @@ static struct r10conf *setup_conf(struct mddev *mddev)
|
||||||
if (!conf->r10bio_pool)
|
if (!conf->r10bio_pool)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
size = mddev->dev_sectors >> conf->chunk_shift;
|
calc_sectors(conf, mddev->dev_sectors);
|
||||||
sector_div(size, fc);
|
|
||||||
size = size * conf->raid_disks;
|
|
||||||
sector_div(size, nc);
|
|
||||||
/* 'size' is now the number of chunks in the array */
|
|
||||||
/* calculate "used chunks per device" in 'stride' */
|
|
||||||
stride = size * conf->copies;
|
|
||||||
|
|
||||||
/* We need to round up when dividing by raid_disks to
|
|
||||||
* get the stride size.
|
|
||||||
*/
|
|
||||||
stride += conf->raid_disks - 1;
|
|
||||||
sector_div(stride, conf->raid_disks);
|
|
||||||
|
|
||||||
conf->dev_sectors = stride << conf->chunk_shift;
|
|
||||||
|
|
||||||
if (fo)
|
|
||||||
stride = 1;
|
|
||||||
else
|
|
||||||
sector_div(stride, fc);
|
|
||||||
conf->stride = stride << conf->chunk_shift;
|
|
||||||
|
|
||||||
|
|
||||||
spin_lock_init(&conf->device_lock);
|
spin_lock_init(&conf->device_lock);
|
||||||
INIT_LIST_HEAD(&conf->retry_list);
|
INIT_LIST_HEAD(&conf->retry_list);
|
||||||
|
@ -3468,7 +3475,8 @@ static int raid10_resize(struct mddev *mddev, sector_t sectors)
|
||||||
mddev->recovery_cp = oldsize;
|
mddev->recovery_cp = oldsize;
|
||||||
set_bit(MD_RECOVERY_NEEDED, &mddev->recovery);
|
set_bit(MD_RECOVERY_NEEDED, &mddev->recovery);
|
||||||
}
|
}
|
||||||
mddev->dev_sectors = sectors;
|
calc_sectors(conf, sectors);
|
||||||
|
mddev->dev_sectors = conf->dev_sectors;
|
||||||
mddev->resync_max_sectors = size;
|
mddev->resync_max_sectors = size;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
|
@ -1921,6 +1921,10 @@ static int dtv_set_frontend(struct dvb_frontend *fe)
|
||||||
} else {
|
} else {
|
||||||
/* default values */
|
/* default values */
|
||||||
switch (c->delivery_system) {
|
switch (c->delivery_system) {
|
||||||
|
case SYS_DVBS:
|
||||||
|
case SYS_DVBS2:
|
||||||
|
case SYS_ISDBS:
|
||||||
|
case SYS_TURBO:
|
||||||
case SYS_DVBC_ANNEX_A:
|
case SYS_DVBC_ANNEX_A:
|
||||||
case SYS_DVBC_ANNEX_C:
|
case SYS_DVBC_ANNEX_C:
|
||||||
fepriv->min_delay = HZ / 20;
|
fepriv->min_delay = HZ / 20;
|
||||||
|
|
|
@ -1018,22 +1018,6 @@ static int ene_probe(struct pnp_dev *pnp_dev, const struct pnp_device_id *id)
|
||||||
|
|
||||||
spin_lock_init(&dev->hw_lock);
|
spin_lock_init(&dev->hw_lock);
|
||||||
|
|
||||||
/* claim the resources */
|
|
||||||
error = -EBUSY;
|
|
||||||
dev->hw_io = pnp_port_start(pnp_dev, 0);
|
|
||||||
if (!request_region(dev->hw_io, ENE_IO_SIZE, ENE_DRIVER_NAME)) {
|
|
||||||
dev->hw_io = -1;
|
|
||||||
dev->irq = -1;
|
|
||||||
goto error;
|
|
||||||
}
|
|
||||||
|
|
||||||
dev->irq = pnp_irq(pnp_dev, 0);
|
|
||||||
if (request_irq(dev->irq, ene_isr,
|
|
||||||
IRQF_SHARED, ENE_DRIVER_NAME, (void *)dev)) {
|
|
||||||
dev->irq = -1;
|
|
||||||
goto error;
|
|
||||||
}
|
|
||||||
|
|
||||||
pnp_set_drvdata(pnp_dev, dev);
|
pnp_set_drvdata(pnp_dev, dev);
|
||||||
dev->pnp_dev = pnp_dev;
|
dev->pnp_dev = pnp_dev;
|
||||||
|
|
||||||
|
@ -1086,6 +1070,22 @@ static int ene_probe(struct pnp_dev *pnp_dev, const struct pnp_device_id *id)
|
||||||
device_set_wakeup_capable(&pnp_dev->dev, true);
|
device_set_wakeup_capable(&pnp_dev->dev, true);
|
||||||
device_set_wakeup_enable(&pnp_dev->dev, true);
|
device_set_wakeup_enable(&pnp_dev->dev, true);
|
||||||
|
|
||||||
|
/* claim the resources */
|
||||||
|
error = -EBUSY;
|
||||||
|
dev->hw_io = pnp_port_start(pnp_dev, 0);
|
||||||
|
if (!request_region(dev->hw_io, ENE_IO_SIZE, ENE_DRIVER_NAME)) {
|
||||||
|
dev->hw_io = -1;
|
||||||
|
dev->irq = -1;
|
||||||
|
goto error;
|
||||||
|
}
|
||||||
|
|
||||||
|
dev->irq = pnp_irq(pnp_dev, 0);
|
||||||
|
if (request_irq(dev->irq, ene_isr,
|
||||||
|
IRQF_SHARED, ENE_DRIVER_NAME, (void *)dev)) {
|
||||||
|
dev->irq = -1;
|
||||||
|
goto error;
|
||||||
|
}
|
||||||
|
|
||||||
error = rc_register_device(rdev);
|
error = rc_register_device(rdev);
|
||||||
if (error < 0)
|
if (error < 0)
|
||||||
goto error;
|
goto error;
|
||||||
|
|
|
@ -197,7 +197,7 @@ static int fintek_hw_detect(struct fintek_dev *fintek)
|
||||||
/*
|
/*
|
||||||
* Newer reviews of this chipset uses port 8 instead of 5
|
* Newer reviews of this chipset uses port 8 instead of 5
|
||||||
*/
|
*/
|
||||||
if ((chip != 0x0408) || (chip != 0x0804))
|
if ((chip != 0x0408) && (chip != 0x0804))
|
||||||
fintek->logical_dev_cir = LOGICAL_DEV_CIR_REV2;
|
fintek->logical_dev_cir = LOGICAL_DEV_CIR_REV2;
|
||||||
else
|
else
|
||||||
fintek->logical_dev_cir = LOGICAL_DEV_CIR_REV1;
|
fintek->logical_dev_cir = LOGICAL_DEV_CIR_REV1;
|
||||||
|
@ -514,16 +514,6 @@ static int fintek_probe(struct pnp_dev *pdev, const struct pnp_device_id *dev_id
|
||||||
|
|
||||||
spin_lock_init(&fintek->fintek_lock);
|
spin_lock_init(&fintek->fintek_lock);
|
||||||
|
|
||||||
ret = -EBUSY;
|
|
||||||
/* now claim resources */
|
|
||||||
if (!request_region(fintek->cir_addr,
|
|
||||||
fintek->cir_port_len, FINTEK_DRIVER_NAME))
|
|
||||||
goto failure;
|
|
||||||
|
|
||||||
if (request_irq(fintek->cir_irq, fintek_cir_isr, IRQF_SHARED,
|
|
||||||
FINTEK_DRIVER_NAME, (void *)fintek))
|
|
||||||
goto failure;
|
|
||||||
|
|
||||||
pnp_set_drvdata(pdev, fintek);
|
pnp_set_drvdata(pdev, fintek);
|
||||||
fintek->pdev = pdev;
|
fintek->pdev = pdev;
|
||||||
|
|
||||||
|
@ -558,6 +548,16 @@ static int fintek_probe(struct pnp_dev *pdev, const struct pnp_device_id *dev_id
|
||||||
/* rx resolution is hardwired to 50us atm, 1, 25, 100 also possible */
|
/* rx resolution is hardwired to 50us atm, 1, 25, 100 also possible */
|
||||||
rdev->rx_resolution = US_TO_NS(CIR_SAMPLE_PERIOD);
|
rdev->rx_resolution = US_TO_NS(CIR_SAMPLE_PERIOD);
|
||||||
|
|
||||||
|
ret = -EBUSY;
|
||||||
|
/* now claim resources */
|
||||||
|
if (!request_region(fintek->cir_addr,
|
||||||
|
fintek->cir_port_len, FINTEK_DRIVER_NAME))
|
||||||
|
goto failure;
|
||||||
|
|
||||||
|
if (request_irq(fintek->cir_irq, fintek_cir_isr, IRQF_SHARED,
|
||||||
|
FINTEK_DRIVER_NAME, (void *)fintek))
|
||||||
|
goto failure;
|
||||||
|
|
||||||
ret = rc_register_device(rdev);
|
ret = rc_register_device(rdev);
|
||||||
if (ret)
|
if (ret)
|
||||||
goto failure;
|
goto failure;
|
||||||
|
|
|
@ -1515,16 +1515,6 @@ static int ite_probe(struct pnp_dev *pdev, const struct pnp_device_id
|
||||||
/* initialize raw event */
|
/* initialize raw event */
|
||||||
init_ir_raw_event(&itdev->rawir);
|
init_ir_raw_event(&itdev->rawir);
|
||||||
|
|
||||||
ret = -EBUSY;
|
|
||||||
/* now claim resources */
|
|
||||||
if (!request_region(itdev->cir_addr,
|
|
||||||
dev_desc->io_region_size, ITE_DRIVER_NAME))
|
|
||||||
goto failure;
|
|
||||||
|
|
||||||
if (request_irq(itdev->cir_irq, ite_cir_isr, IRQF_SHARED,
|
|
||||||
ITE_DRIVER_NAME, (void *)itdev))
|
|
||||||
goto failure;
|
|
||||||
|
|
||||||
/* set driver data into the pnp device */
|
/* set driver data into the pnp device */
|
||||||
pnp_set_drvdata(pdev, itdev);
|
pnp_set_drvdata(pdev, itdev);
|
||||||
itdev->pdev = pdev;
|
itdev->pdev = pdev;
|
||||||
|
@ -1600,6 +1590,16 @@ static int ite_probe(struct pnp_dev *pdev, const struct pnp_device_id
|
||||||
rdev->driver_name = ITE_DRIVER_NAME;
|
rdev->driver_name = ITE_DRIVER_NAME;
|
||||||
rdev->map_name = RC_MAP_RC6_MCE;
|
rdev->map_name = RC_MAP_RC6_MCE;
|
||||||
|
|
||||||
|
ret = -EBUSY;
|
||||||
|
/* now claim resources */
|
||||||
|
if (!request_region(itdev->cir_addr,
|
||||||
|
dev_desc->io_region_size, ITE_DRIVER_NAME))
|
||||||
|
goto failure;
|
||||||
|
|
||||||
|
if (request_irq(itdev->cir_irq, ite_cir_isr, IRQF_SHARED,
|
||||||
|
ITE_DRIVER_NAME, (void *)itdev))
|
||||||
|
goto failure;
|
||||||
|
|
||||||
ret = rc_register_device(rdev);
|
ret = rc_register_device(rdev);
|
||||||
if (ret)
|
if (ret)
|
||||||
goto failure;
|
goto failure;
|
||||||
|
|
|
@ -1021,24 +1021,6 @@ static int nvt_probe(struct pnp_dev *pdev, const struct pnp_device_id *dev_id)
|
||||||
spin_lock_init(&nvt->nvt_lock);
|
spin_lock_init(&nvt->nvt_lock);
|
||||||
spin_lock_init(&nvt->tx.lock);
|
spin_lock_init(&nvt->tx.lock);
|
||||||
|
|
||||||
ret = -EBUSY;
|
|
||||||
/* now claim resources */
|
|
||||||
if (!request_region(nvt->cir_addr,
|
|
||||||
CIR_IOREG_LENGTH, NVT_DRIVER_NAME))
|
|
||||||
goto failure;
|
|
||||||
|
|
||||||
if (request_irq(nvt->cir_irq, nvt_cir_isr, IRQF_SHARED,
|
|
||||||
NVT_DRIVER_NAME, (void *)nvt))
|
|
||||||
goto failure;
|
|
||||||
|
|
||||||
if (!request_region(nvt->cir_wake_addr,
|
|
||||||
CIR_IOREG_LENGTH, NVT_DRIVER_NAME))
|
|
||||||
goto failure;
|
|
||||||
|
|
||||||
if (request_irq(nvt->cir_wake_irq, nvt_cir_wake_isr, IRQF_SHARED,
|
|
||||||
NVT_DRIVER_NAME, (void *)nvt))
|
|
||||||
goto failure;
|
|
||||||
|
|
||||||
pnp_set_drvdata(pdev, nvt);
|
pnp_set_drvdata(pdev, nvt);
|
||||||
nvt->pdev = pdev;
|
nvt->pdev = pdev;
|
||||||
|
|
||||||
|
@ -1085,6 +1067,24 @@ static int nvt_probe(struct pnp_dev *pdev, const struct pnp_device_id *dev_id)
|
||||||
rdev->tx_resolution = XYZ;
|
rdev->tx_resolution = XYZ;
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
ret = -EBUSY;
|
||||||
|
/* now claim resources */
|
||||||
|
if (!request_region(nvt->cir_addr,
|
||||||
|
CIR_IOREG_LENGTH, NVT_DRIVER_NAME))
|
||||||
|
goto failure;
|
||||||
|
|
||||||
|
if (request_irq(nvt->cir_irq, nvt_cir_isr, IRQF_SHARED,
|
||||||
|
NVT_DRIVER_NAME, (void *)nvt))
|
||||||
|
goto failure;
|
||||||
|
|
||||||
|
if (!request_region(nvt->cir_wake_addr,
|
||||||
|
CIR_IOREG_LENGTH, NVT_DRIVER_NAME))
|
||||||
|
goto failure;
|
||||||
|
|
||||||
|
if (request_irq(nvt->cir_wake_irq, nvt_cir_wake_isr, IRQF_SHARED,
|
||||||
|
NVT_DRIVER_NAME, (void *)nvt))
|
||||||
|
goto failure;
|
||||||
|
|
||||||
ret = rc_register_device(rdev);
|
ret = rc_register_device(rdev);
|
||||||
if (ret)
|
if (ret)
|
||||||
goto failure;
|
goto failure;
|
||||||
|
|
|
@ -991,39 +991,10 @@ wbcir_probe(struct pnp_dev *device, const struct pnp_device_id *dev_id)
|
||||||
"(w: 0x%lX, e: 0x%lX, s: 0x%lX, i: %u)\n",
|
"(w: 0x%lX, e: 0x%lX, s: 0x%lX, i: %u)\n",
|
||||||
data->wbase, data->ebase, data->sbase, data->irq);
|
data->wbase, data->ebase, data->sbase, data->irq);
|
||||||
|
|
||||||
if (!request_region(data->wbase, WAKEUP_IOMEM_LEN, DRVNAME)) {
|
|
||||||
dev_err(dev, "Region 0x%lx-0x%lx already in use!\n",
|
|
||||||
data->wbase, data->wbase + WAKEUP_IOMEM_LEN - 1);
|
|
||||||
err = -EBUSY;
|
|
||||||
goto exit_free_data;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!request_region(data->ebase, EHFUNC_IOMEM_LEN, DRVNAME)) {
|
|
||||||
dev_err(dev, "Region 0x%lx-0x%lx already in use!\n",
|
|
||||||
data->ebase, data->ebase + EHFUNC_IOMEM_LEN - 1);
|
|
||||||
err = -EBUSY;
|
|
||||||
goto exit_release_wbase;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!request_region(data->sbase, SP_IOMEM_LEN, DRVNAME)) {
|
|
||||||
dev_err(dev, "Region 0x%lx-0x%lx already in use!\n",
|
|
||||||
data->sbase, data->sbase + SP_IOMEM_LEN - 1);
|
|
||||||
err = -EBUSY;
|
|
||||||
goto exit_release_ebase;
|
|
||||||
}
|
|
||||||
|
|
||||||
err = request_irq(data->irq, wbcir_irq_handler,
|
|
||||||
IRQF_DISABLED, DRVNAME, device);
|
|
||||||
if (err) {
|
|
||||||
dev_err(dev, "Failed to claim IRQ %u\n", data->irq);
|
|
||||||
err = -EBUSY;
|
|
||||||
goto exit_release_sbase;
|
|
||||||
}
|
|
||||||
|
|
||||||
led_trigger_register_simple("cir-tx", &data->txtrigger);
|
led_trigger_register_simple("cir-tx", &data->txtrigger);
|
||||||
if (!data->txtrigger) {
|
if (!data->txtrigger) {
|
||||||
err = -ENOMEM;
|
err = -ENOMEM;
|
||||||
goto exit_free_irq;
|
goto exit_free_data;
|
||||||
}
|
}
|
||||||
|
|
||||||
led_trigger_register_simple("cir-rx", &data->rxtrigger);
|
led_trigger_register_simple("cir-rx", &data->rxtrigger);
|
||||||
|
@ -1062,9 +1033,38 @@ wbcir_probe(struct pnp_dev *device, const struct pnp_device_id *dev_id)
|
||||||
data->dev->priv = data;
|
data->dev->priv = data;
|
||||||
data->dev->dev.parent = &device->dev;
|
data->dev->dev.parent = &device->dev;
|
||||||
|
|
||||||
|
if (!request_region(data->wbase, WAKEUP_IOMEM_LEN, DRVNAME)) {
|
||||||
|
dev_err(dev, "Region 0x%lx-0x%lx already in use!\n",
|
||||||
|
data->wbase, data->wbase + WAKEUP_IOMEM_LEN - 1);
|
||||||
|
err = -EBUSY;
|
||||||
|
goto exit_free_rc;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!request_region(data->ebase, EHFUNC_IOMEM_LEN, DRVNAME)) {
|
||||||
|
dev_err(dev, "Region 0x%lx-0x%lx already in use!\n",
|
||||||
|
data->ebase, data->ebase + EHFUNC_IOMEM_LEN - 1);
|
||||||
|
err = -EBUSY;
|
||||||
|
goto exit_release_wbase;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!request_region(data->sbase, SP_IOMEM_LEN, DRVNAME)) {
|
||||||
|
dev_err(dev, "Region 0x%lx-0x%lx already in use!\n",
|
||||||
|
data->sbase, data->sbase + SP_IOMEM_LEN - 1);
|
||||||
|
err = -EBUSY;
|
||||||
|
goto exit_release_ebase;
|
||||||
|
}
|
||||||
|
|
||||||
|
err = request_irq(data->irq, wbcir_irq_handler,
|
||||||
|
IRQF_DISABLED, DRVNAME, device);
|
||||||
|
if (err) {
|
||||||
|
dev_err(dev, "Failed to claim IRQ %u\n", data->irq);
|
||||||
|
err = -EBUSY;
|
||||||
|
goto exit_release_sbase;
|
||||||
|
}
|
||||||
|
|
||||||
err = rc_register_device(data->dev);
|
err = rc_register_device(data->dev);
|
||||||
if (err)
|
if (err)
|
||||||
goto exit_free_rc;
|
goto exit_free_irq;
|
||||||
|
|
||||||
device_init_wakeup(&device->dev, 1);
|
device_init_wakeup(&device->dev, 1);
|
||||||
|
|
||||||
|
@ -1072,14 +1072,6 @@ wbcir_probe(struct pnp_dev *device, const struct pnp_device_id *dev_id)
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
exit_free_rc:
|
|
||||||
rc_free_device(data->dev);
|
|
||||||
exit_unregister_led:
|
|
||||||
led_classdev_unregister(&data->led);
|
|
||||||
exit_unregister_rxtrigger:
|
|
||||||
led_trigger_unregister_simple(data->rxtrigger);
|
|
||||||
exit_unregister_txtrigger:
|
|
||||||
led_trigger_unregister_simple(data->txtrigger);
|
|
||||||
exit_free_irq:
|
exit_free_irq:
|
||||||
free_irq(data->irq, device);
|
free_irq(data->irq, device);
|
||||||
exit_release_sbase:
|
exit_release_sbase:
|
||||||
|
@ -1088,6 +1080,14 @@ exit_release_ebase:
|
||||||
release_region(data->ebase, EHFUNC_IOMEM_LEN);
|
release_region(data->ebase, EHFUNC_IOMEM_LEN);
|
||||||
exit_release_wbase:
|
exit_release_wbase:
|
||||||
release_region(data->wbase, WAKEUP_IOMEM_LEN);
|
release_region(data->wbase, WAKEUP_IOMEM_LEN);
|
||||||
|
exit_free_rc:
|
||||||
|
rc_free_device(data->dev);
|
||||||
|
exit_unregister_led:
|
||||||
|
led_classdev_unregister(&data->led);
|
||||||
|
exit_unregister_rxtrigger:
|
||||||
|
led_trigger_unregister_simple(data->rxtrigger);
|
||||||
|
exit_unregister_txtrigger:
|
||||||
|
led_trigger_unregister_simple(data->txtrigger);
|
||||||
exit_free_data:
|
exit_free_data:
|
||||||
kfree(data);
|
kfree(data);
|
||||||
pnp_set_drvdata(device, NULL);
|
pnp_set_drvdata(device, NULL);
|
||||||
|
|
|
@ -2923,6 +2923,10 @@ static void sd_pkt_scan(struct gspca_dev *gspca_dev,
|
||||||
* not the JPEG end of frame ('ff d9').
|
* not the JPEG end of frame ('ff d9').
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
/* count the packets and their size */
|
||||||
|
sd->npkt++;
|
||||||
|
sd->pktsz += len;
|
||||||
|
|
||||||
/*fixme: assumption about the following code:
|
/*fixme: assumption about the following code:
|
||||||
* - there can be only one marker in a packet
|
* - there can be only one marker in a packet
|
||||||
*/
|
*/
|
||||||
|
@ -2945,10 +2949,6 @@ static void sd_pkt_scan(struct gspca_dev *gspca_dev,
|
||||||
data += i;
|
data += i;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* count the packets and their size */
|
|
||||||
sd->npkt++;
|
|
||||||
sd->pktsz += len;
|
|
||||||
|
|
||||||
/* search backwards if there is a marker in the packet */
|
/* search backwards if there is a marker in the packet */
|
||||||
for (i = len - 1; --i >= 0; ) {
|
for (i = len - 1; --i >= 0; ) {
|
||||||
if (data[i] != 0xff) {
|
if (data[i] != 0xff) {
|
||||||
|
|
|
@ -181,7 +181,6 @@ static int mmpcam_probe(struct platform_device *pdev)
|
||||||
INIT_LIST_HEAD(&cam->devlist);
|
INIT_LIST_HEAD(&cam->devlist);
|
||||||
|
|
||||||
mcam = &cam->mcam;
|
mcam = &cam->mcam;
|
||||||
mcam->platform = MHP_Armada610;
|
|
||||||
mcam->plat_power_up = mmpcam_power_up;
|
mcam->plat_power_up = mmpcam_power_up;
|
||||||
mcam->plat_power_down = mmpcam_power_down;
|
mcam->plat_power_down = mmpcam_power_down;
|
||||||
mcam->dev = &pdev->dev;
|
mcam->dev = &pdev->dev;
|
||||||
|
|
|
@ -246,28 +246,37 @@ int fimc_capture_resume(struct fimc_dev *fimc)
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static unsigned int get_plane_size(struct fimc_frame *fr, unsigned int plane)
|
|
||||||
{
|
|
||||||
if (!fr || plane >= fr->fmt->memplanes)
|
|
||||||
return 0;
|
|
||||||
return fr->f_width * fr->f_height * fr->fmt->depth[plane] / 8;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int queue_setup(struct vb2_queue *vq, const struct v4l2_format *pfmt,
|
static int queue_setup(struct vb2_queue *vq, const struct v4l2_format *pfmt,
|
||||||
unsigned int *num_buffers, unsigned int *num_planes,
|
unsigned int *num_buffers, unsigned int *num_planes,
|
||||||
unsigned int sizes[], void *allocators[])
|
unsigned int sizes[], void *allocators[])
|
||||||
{
|
{
|
||||||
|
const struct v4l2_pix_format_mplane *pixm = NULL;
|
||||||
struct fimc_ctx *ctx = vq->drv_priv;
|
struct fimc_ctx *ctx = vq->drv_priv;
|
||||||
struct fimc_fmt *fmt = ctx->d_frame.fmt;
|
struct fimc_frame *frame = &ctx->d_frame;
|
||||||
|
struct fimc_fmt *fmt = frame->fmt;
|
||||||
|
unsigned long wh;
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
if (!fmt)
|
if (pfmt) {
|
||||||
|
pixm = &pfmt->fmt.pix_mp;
|
||||||
|
fmt = fimc_find_format(&pixm->pixelformat, NULL,
|
||||||
|
FMT_FLAGS_CAM | FMT_FLAGS_M2M, -1);
|
||||||
|
wh = pixm->width * pixm->height;
|
||||||
|
} else {
|
||||||
|
wh = frame->f_width * frame->f_height;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (fmt == NULL)
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
|
||||||
*num_planes = fmt->memplanes;
|
*num_planes = fmt->memplanes;
|
||||||
|
|
||||||
for (i = 0; i < fmt->memplanes; i++) {
|
for (i = 0; i < fmt->memplanes; i++) {
|
||||||
sizes[i] = get_plane_size(&ctx->d_frame, i);
|
unsigned int size = (wh * fmt->depth[i]) / 8;
|
||||||
|
if (pixm)
|
||||||
|
sizes[i] = max(size, pixm->plane_fmt[i].sizeimage);
|
||||||
|
else
|
||||||
|
sizes[i] = size;
|
||||||
allocators[i] = ctx->fimc_dev->alloc_ctx;
|
allocators[i] = ctx->fimc_dev->alloc_ctx;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1383,7 +1392,7 @@ static int fimc_subdev_set_crop(struct v4l2_subdev *sd,
|
||||||
fimc_capture_try_crop(ctx, r, crop->pad);
|
fimc_capture_try_crop(ctx, r, crop->pad);
|
||||||
|
|
||||||
if (crop->which == V4L2_SUBDEV_FORMAT_TRY) {
|
if (crop->which == V4L2_SUBDEV_FORMAT_TRY) {
|
||||||
mutex_lock(&fimc->lock);
|
mutex_unlock(&fimc->lock);
|
||||||
*v4l2_subdev_get_try_crop(fh, crop->pad) = *r;
|
*v4l2_subdev_get_try_crop(fh, crop->pad) = *r;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
|
@ -1048,14 +1048,14 @@ static int fimc_m2m_g_fmt_mplane(struct file *file, void *fh,
|
||||||
* @mask: the color flags to match
|
* @mask: the color flags to match
|
||||||
* @index: offset in the fimc_formats array, ignored if negative
|
* @index: offset in the fimc_formats array, ignored if negative
|
||||||
*/
|
*/
|
||||||
struct fimc_fmt *fimc_find_format(u32 *pixelformat, u32 *mbus_code,
|
struct fimc_fmt *fimc_find_format(const u32 *pixelformat, const u32 *mbus_code,
|
||||||
unsigned int mask, int index)
|
unsigned int mask, int index)
|
||||||
{
|
{
|
||||||
struct fimc_fmt *fmt, *def_fmt = NULL;
|
struct fimc_fmt *fmt, *def_fmt = NULL;
|
||||||
unsigned int i;
|
unsigned int i;
|
||||||
int id = 0;
|
int id = 0;
|
||||||
|
|
||||||
if (index >= ARRAY_SIZE(fimc_formats))
|
if (index >= (int)ARRAY_SIZE(fimc_formats))
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
for (i = 0; i < ARRAY_SIZE(fimc_formats); ++i) {
|
for (i = 0; i < ARRAY_SIZE(fimc_formats); ++i) {
|
||||||
|
|
|
@ -718,7 +718,7 @@ void fimc_alpha_ctrl_update(struct fimc_ctx *ctx);
|
||||||
int fimc_fill_format(struct fimc_frame *frame, struct v4l2_format *f);
|
int fimc_fill_format(struct fimc_frame *frame, struct v4l2_format *f);
|
||||||
void fimc_adjust_mplane_format(struct fimc_fmt *fmt, u32 width, u32 height,
|
void fimc_adjust_mplane_format(struct fimc_fmt *fmt, u32 width, u32 height,
|
||||||
struct v4l2_pix_format_mplane *pix);
|
struct v4l2_pix_format_mplane *pix);
|
||||||
struct fimc_fmt *fimc_find_format(u32 *pixelformat, u32 *mbus_code,
|
struct fimc_fmt *fimc_find_format(const u32 *pixelformat, const u32 *mbus_code,
|
||||||
unsigned int mask, int index);
|
unsigned int mask, int index);
|
||||||
|
|
||||||
int fimc_check_scaler_ratio(struct fimc_ctx *ctx, int sw, int sh,
|
int fimc_check_scaler_ratio(struct fimc_ctx *ctx, int sw, int sh,
|
||||||
|
|
|
@ -530,7 +530,10 @@ static int soc_camera_open(struct file *file)
|
||||||
if (icl->reset)
|
if (icl->reset)
|
||||||
icl->reset(icd->pdev);
|
icl->reset(icd->pdev);
|
||||||
|
|
||||||
|
/* Don't mess with the host during probe */
|
||||||
|
mutex_lock(&ici->host_lock);
|
||||||
ret = ici->ops->add(icd);
|
ret = ici->ops->add(icd);
|
||||||
|
mutex_unlock(&ici->host_lock);
|
||||||
if (ret < 0) {
|
if (ret < 0) {
|
||||||
dev_err(icd->pdev, "Couldn't activate the camera: %d\n", ret);
|
dev_err(icd->pdev, "Couldn't activate the camera: %d\n", ret);
|
||||||
goto eiciadd;
|
goto eiciadd;
|
||||||
|
@ -956,7 +959,7 @@ static void scan_add_host(struct soc_camera_host *ici)
|
||||||
{
|
{
|
||||||
struct soc_camera_device *icd;
|
struct soc_camera_device *icd;
|
||||||
|
|
||||||
mutex_lock(&list_lock);
|
mutex_lock(&ici->host_lock);
|
||||||
|
|
||||||
list_for_each_entry(icd, &devices, list) {
|
list_for_each_entry(icd, &devices, list) {
|
||||||
if (icd->iface == ici->nr) {
|
if (icd->iface == ici->nr) {
|
||||||
|
@ -967,7 +970,7 @@ static void scan_add_host(struct soc_camera_host *ici)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
mutex_unlock(&list_lock);
|
mutex_unlock(&ici->host_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_I2C_BOARDINFO
|
#ifdef CONFIG_I2C_BOARDINFO
|
||||||
|
@ -1313,6 +1316,7 @@ int soc_camera_host_register(struct soc_camera_host *ici)
|
||||||
list_add_tail(&ici->list, &hosts);
|
list_add_tail(&ici->list, &hosts);
|
||||||
mutex_unlock(&list_lock);
|
mutex_unlock(&list_lock);
|
||||||
|
|
||||||
|
mutex_init(&ici->host_lock);
|
||||||
scan_add_host(ici);
|
scan_add_host(ici);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
|
@ -15,6 +15,7 @@
|
||||||
#include <linux/dma-mapping.h>
|
#include <linux/dma-mapping.h>
|
||||||
|
|
||||||
#include <media/videobuf2-core.h>
|
#include <media/videobuf2-core.h>
|
||||||
|
#include <media/videobuf2-dma-contig.h>
|
||||||
#include <media/videobuf2-memops.h>
|
#include <media/videobuf2-memops.h>
|
||||||
|
|
||||||
struct vb2_dc_conf {
|
struct vb2_dc_conf {
|
||||||
|
@ -85,7 +86,7 @@ static void *vb2_dma_contig_vaddr(void *buf_priv)
|
||||||
{
|
{
|
||||||
struct vb2_dc_buf *buf = buf_priv;
|
struct vb2_dc_buf *buf = buf_priv;
|
||||||
if (!buf)
|
if (!buf)
|
||||||
return 0;
|
return NULL;
|
||||||
|
|
||||||
return buf->vaddr;
|
return buf->vaddr;
|
||||||
}
|
}
|
||||||
|
|
|
@ -55,6 +55,7 @@ struct vm_area_struct *vb2_get_vma(struct vm_area_struct *vma)
|
||||||
|
|
||||||
return vma_copy;
|
return vma_copy;
|
||||||
}
|
}
|
||||||
|
EXPORT_SYMBOL_GPL(vb2_get_vma);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* vb2_put_userptr() - release a userspace virtual memory area
|
* vb2_put_userptr() - release a userspace virtual memory area
|
||||||
|
|
|
@ -376,7 +376,7 @@ static int otp_select_filemode(struct mtd_file_info *mfi, int mode)
|
||||||
* Make a fake call to mtd_read_fact_prot_reg() to check if OTP
|
* Make a fake call to mtd_read_fact_prot_reg() to check if OTP
|
||||||
* operations are supported.
|
* operations are supported.
|
||||||
*/
|
*/
|
||||||
if (mtd_read_fact_prot_reg(mtd, -1, -1, &retlen, NULL) == -EOPNOTSUPP)
|
if (mtd_read_fact_prot_reg(mtd, -1, 0, &retlen, NULL) == -EOPNOTSUPP)
|
||||||
return -EOPNOTSUPP;
|
return -EOPNOTSUPP;
|
||||||
|
|
||||||
switch (mode) {
|
switch (mode) {
|
||||||
|
|
|
@ -212,18 +212,17 @@ static int __devinit ams_delta_init(struct platform_device *pdev)
|
||||||
/* Link the private data with the MTD structure */
|
/* Link the private data with the MTD structure */
|
||||||
ams_delta_mtd->priv = this;
|
ams_delta_mtd->priv = this;
|
||||||
|
|
||||||
if (!request_mem_region(res->start, resource_size(res),
|
/*
|
||||||
dev_name(&pdev->dev))) {
|
* Don't try to request the memory region from here,
|
||||||
dev_err(&pdev->dev, "request_mem_region failed\n");
|
* it should have been already requested from the
|
||||||
err = -EBUSY;
|
* gpio-omap driver and requesting it again would fail.
|
||||||
goto out_free;
|
*/
|
||||||
}
|
|
||||||
|
|
||||||
io_base = ioremap(res->start, resource_size(res));
|
io_base = ioremap(res->start, resource_size(res));
|
||||||
if (io_base == NULL) {
|
if (io_base == NULL) {
|
||||||
dev_err(&pdev->dev, "ioremap failed\n");
|
dev_err(&pdev->dev, "ioremap failed\n");
|
||||||
err = -EIO;
|
err = -EIO;
|
||||||
goto out_release_io;
|
goto out_free;
|
||||||
}
|
}
|
||||||
|
|
||||||
this->priv = io_base;
|
this->priv = io_base;
|
||||||
|
@ -271,8 +270,6 @@ out_gpio:
|
||||||
platform_set_drvdata(pdev, NULL);
|
platform_set_drvdata(pdev, NULL);
|
||||||
gpio_free(AMS_DELTA_GPIO_PIN_NAND_RB);
|
gpio_free(AMS_DELTA_GPIO_PIN_NAND_RB);
|
||||||
iounmap(io_base);
|
iounmap(io_base);
|
||||||
out_release_io:
|
|
||||||
release_mem_region(res->start, resource_size(res));
|
|
||||||
out_free:
|
out_free:
|
||||||
kfree(ams_delta_mtd);
|
kfree(ams_delta_mtd);
|
||||||
out:
|
out:
|
||||||
|
@ -285,7 +282,6 @@ out_free:
|
||||||
static int __devexit ams_delta_cleanup(struct platform_device *pdev)
|
static int __devexit ams_delta_cleanup(struct platform_device *pdev)
|
||||||
{
|
{
|
||||||
void __iomem *io_base = platform_get_drvdata(pdev);
|
void __iomem *io_base = platform_get_drvdata(pdev);
|
||||||
struct resource *res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
|
|
||||||
|
|
||||||
/* Release resources, unregister device */
|
/* Release resources, unregister device */
|
||||||
nand_release(ams_delta_mtd);
|
nand_release(ams_delta_mtd);
|
||||||
|
@ -293,7 +289,6 @@ static int __devexit ams_delta_cleanup(struct platform_device *pdev)
|
||||||
gpio_free_array(_mandatory_gpio, ARRAY_SIZE(_mandatory_gpio));
|
gpio_free_array(_mandatory_gpio, ARRAY_SIZE(_mandatory_gpio));
|
||||||
gpio_free(AMS_DELTA_GPIO_PIN_NAND_RB);
|
gpio_free(AMS_DELTA_GPIO_PIN_NAND_RB);
|
||||||
iounmap(io_base);
|
iounmap(io_base);
|
||||||
release_mem_region(res->start, resource_size(res));
|
|
||||||
|
|
||||||
/* Free the MTD device structure */
|
/* Free the MTD device structure */
|
||||||
kfree(ams_delta_mtd);
|
kfree(ams_delta_mtd);
|
||||||
|
|
|
@ -342,26 +342,26 @@ static void rlb_update_entry_from_arp(struct bonding *bond, struct arp_pkt *arp)
|
||||||
_unlock_rx_hashtbl_bh(bond);
|
_unlock_rx_hashtbl_bh(bond);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void rlb_arp_recv(struct sk_buff *skb, struct bonding *bond,
|
static int rlb_arp_recv(struct sk_buff *skb, struct bonding *bond,
|
||||||
struct slave *slave)
|
struct slave *slave)
|
||||||
{
|
{
|
||||||
struct arp_pkt *arp;
|
struct arp_pkt *arp;
|
||||||
|
|
||||||
if (skb->protocol != cpu_to_be16(ETH_P_ARP))
|
if (skb->protocol != cpu_to_be16(ETH_P_ARP))
|
||||||
return;
|
goto out;
|
||||||
|
|
||||||
arp = (struct arp_pkt *) skb->data;
|
arp = (struct arp_pkt *) skb->data;
|
||||||
if (!arp) {
|
if (!arp) {
|
||||||
pr_debug("Packet has no ARP data\n");
|
pr_debug("Packet has no ARP data\n");
|
||||||
return;
|
goto out;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!pskb_may_pull(skb, arp_hdr_len(bond->dev)))
|
if (!pskb_may_pull(skb, arp_hdr_len(bond->dev)))
|
||||||
return;
|
goto out;
|
||||||
|
|
||||||
if (skb->len < sizeof(struct arp_pkt)) {
|
if (skb->len < sizeof(struct arp_pkt)) {
|
||||||
pr_debug("Packet is too small to be an ARP\n");
|
pr_debug("Packet is too small to be an ARP\n");
|
||||||
return;
|
goto out;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (arp->op_code == htons(ARPOP_REPLY)) {
|
if (arp->op_code == htons(ARPOP_REPLY)) {
|
||||||
|
@ -369,6 +369,8 @@ static void rlb_arp_recv(struct sk_buff *skb, struct bonding *bond,
|
||||||
rlb_update_entry_from_arp(bond, arp);
|
rlb_update_entry_from_arp(bond, arp);
|
||||||
pr_debug("Server received an ARP Reply from client\n");
|
pr_debug("Server received an ARP Reply from client\n");
|
||||||
}
|
}
|
||||||
|
out:
|
||||||
|
return RX_HANDLER_ANOTHER;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Caller must hold bond lock for read */
|
/* Caller must hold bond lock for read */
|
||||||
|
|
|
@ -218,7 +218,7 @@ struct bonding {
|
||||||
struct slave *primary_slave;
|
struct slave *primary_slave;
|
||||||
bool force_primary;
|
bool force_primary;
|
||||||
s32 slave_cnt; /* never change this value outside the attach/detach wrappers */
|
s32 slave_cnt; /* never change this value outside the attach/detach wrappers */
|
||||||
void (*recv_probe)(struct sk_buff *, struct bonding *,
|
int (*recv_probe)(struct sk_buff *, struct bonding *,
|
||||||
struct slave *);
|
struct slave *);
|
||||||
rwlock_t lock;
|
rwlock_t lock;
|
||||||
rwlock_t curr_slave_lock;
|
rwlock_t curr_slave_lock;
|
||||||
|
|
|
@ -493,7 +493,11 @@ out:
|
||||||
static void e1000_down_and_stop(struct e1000_adapter *adapter)
|
static void e1000_down_and_stop(struct e1000_adapter *adapter)
|
||||||
{
|
{
|
||||||
set_bit(__E1000_DOWN, &adapter->flags);
|
set_bit(__E1000_DOWN, &adapter->flags);
|
||||||
|
|
||||||
|
/* Only kill reset task if adapter is not resetting */
|
||||||
|
if (!test_bit(__E1000_RESETTING, &adapter->flags))
|
||||||
cancel_work_sync(&adapter->reset_task);
|
cancel_work_sync(&adapter->reset_task);
|
||||||
|
|
||||||
cancel_delayed_work_sync(&adapter->watchdog_task);
|
cancel_delayed_work_sync(&adapter->watchdog_task);
|
||||||
cancel_delayed_work_sync(&adapter->phy_info_task);
|
cancel_delayed_work_sync(&adapter->phy_info_task);
|
||||||
cancel_delayed_work_sync(&adapter->fifo_stall_task);
|
cancel_delayed_work_sync(&adapter->fifo_stall_task);
|
||||||
|
|
|
@ -584,7 +584,6 @@ struct pch_gbe_hw_stats {
|
||||||
/**
|
/**
|
||||||
* struct pch_gbe_adapter - board specific private data structure
|
* struct pch_gbe_adapter - board specific private data structure
|
||||||
* @stats_lock: Spinlock structure for status
|
* @stats_lock: Spinlock structure for status
|
||||||
* @tx_queue_lock: Spinlock structure for transmit
|
|
||||||
* @ethtool_lock: Spinlock structure for ethtool
|
* @ethtool_lock: Spinlock structure for ethtool
|
||||||
* @irq_sem: Semaphore for interrupt
|
* @irq_sem: Semaphore for interrupt
|
||||||
* @netdev: Pointer of network device structure
|
* @netdev: Pointer of network device structure
|
||||||
|
@ -609,7 +608,6 @@ struct pch_gbe_hw_stats {
|
||||||
|
|
||||||
struct pch_gbe_adapter {
|
struct pch_gbe_adapter {
|
||||||
spinlock_t stats_lock;
|
spinlock_t stats_lock;
|
||||||
spinlock_t tx_queue_lock;
|
|
||||||
spinlock_t ethtool_lock;
|
spinlock_t ethtool_lock;
|
||||||
atomic_t irq_sem;
|
atomic_t irq_sem;
|
||||||
struct net_device *netdev;
|
struct net_device *netdev;
|
||||||
|
|
|
@ -640,14 +640,11 @@ static void pch_gbe_mac_set_pause_packet(struct pch_gbe_hw *hw)
|
||||||
*/
|
*/
|
||||||
static int pch_gbe_alloc_queues(struct pch_gbe_adapter *adapter)
|
static int pch_gbe_alloc_queues(struct pch_gbe_adapter *adapter)
|
||||||
{
|
{
|
||||||
int size;
|
adapter->tx_ring = kzalloc(sizeof(*adapter->tx_ring), GFP_KERNEL);
|
||||||
|
|
||||||
size = (int)sizeof(struct pch_gbe_tx_ring);
|
|
||||||
adapter->tx_ring = kzalloc(size, GFP_KERNEL);
|
|
||||||
if (!adapter->tx_ring)
|
if (!adapter->tx_ring)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
size = (int)sizeof(struct pch_gbe_rx_ring);
|
|
||||||
adapter->rx_ring = kzalloc(size, GFP_KERNEL);
|
adapter->rx_ring = kzalloc(sizeof(*adapter->rx_ring), GFP_KERNEL);
|
||||||
if (!adapter->rx_ring) {
|
if (!adapter->rx_ring) {
|
||||||
kfree(adapter->tx_ring);
|
kfree(adapter->tx_ring);
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
@ -1162,7 +1159,6 @@ static void pch_gbe_tx_queue(struct pch_gbe_adapter *adapter,
|
||||||
struct sk_buff *tmp_skb;
|
struct sk_buff *tmp_skb;
|
||||||
unsigned int frame_ctrl;
|
unsigned int frame_ctrl;
|
||||||
unsigned int ring_num;
|
unsigned int ring_num;
|
||||||
unsigned long flags;
|
|
||||||
|
|
||||||
/*-- Set frame control --*/
|
/*-- Set frame control --*/
|
||||||
frame_ctrl = 0;
|
frame_ctrl = 0;
|
||||||
|
@ -1211,14 +1207,14 @@ static void pch_gbe_tx_queue(struct pch_gbe_adapter *adapter,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
spin_lock_irqsave(&tx_ring->tx_lock, flags);
|
|
||||||
ring_num = tx_ring->next_to_use;
|
ring_num = tx_ring->next_to_use;
|
||||||
if (unlikely((ring_num + 1) == tx_ring->count))
|
if (unlikely((ring_num + 1) == tx_ring->count))
|
||||||
tx_ring->next_to_use = 0;
|
tx_ring->next_to_use = 0;
|
||||||
else
|
else
|
||||||
tx_ring->next_to_use = ring_num + 1;
|
tx_ring->next_to_use = ring_num + 1;
|
||||||
|
|
||||||
spin_unlock_irqrestore(&tx_ring->tx_lock, flags);
|
|
||||||
buffer_info = &tx_ring->buffer_info[ring_num];
|
buffer_info = &tx_ring->buffer_info[ring_num];
|
||||||
tmp_skb = buffer_info->skb;
|
tmp_skb = buffer_info->skb;
|
||||||
|
|
||||||
|
@ -1518,7 +1514,7 @@ pch_gbe_alloc_rx_buffers_pool(struct pch_gbe_adapter *adapter,
|
||||||
&rx_ring->rx_buff_pool_logic,
|
&rx_ring->rx_buff_pool_logic,
|
||||||
GFP_KERNEL);
|
GFP_KERNEL);
|
||||||
if (!rx_ring->rx_buff_pool) {
|
if (!rx_ring->rx_buff_pool) {
|
||||||
pr_err("Unable to allocate memory for the receive poll buffer\n");
|
pr_err("Unable to allocate memory for the receive pool buffer\n");
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
}
|
}
|
||||||
memset(rx_ring->rx_buff_pool, 0, size);
|
memset(rx_ring->rx_buff_pool, 0, size);
|
||||||
|
@ -1637,15 +1633,17 @@ pch_gbe_clean_tx(struct pch_gbe_adapter *adapter,
|
||||||
pr_debug("called pch_gbe_unmap_and_free_tx_resource() %d count\n",
|
pr_debug("called pch_gbe_unmap_and_free_tx_resource() %d count\n",
|
||||||
cleaned_count);
|
cleaned_count);
|
||||||
/* Recover from running out of Tx resources in xmit_frame */
|
/* Recover from running out of Tx resources in xmit_frame */
|
||||||
|
spin_lock(&tx_ring->tx_lock);
|
||||||
if (unlikely(cleaned && (netif_queue_stopped(adapter->netdev)))) {
|
if (unlikely(cleaned && (netif_queue_stopped(adapter->netdev)))) {
|
||||||
netif_wake_queue(adapter->netdev);
|
netif_wake_queue(adapter->netdev);
|
||||||
adapter->stats.tx_restart_count++;
|
adapter->stats.tx_restart_count++;
|
||||||
pr_debug("Tx wake queue\n");
|
pr_debug("Tx wake queue\n");
|
||||||
}
|
}
|
||||||
spin_lock(&adapter->tx_queue_lock);
|
|
||||||
tx_ring->next_to_clean = i;
|
tx_ring->next_to_clean = i;
|
||||||
spin_unlock(&adapter->tx_queue_lock);
|
|
||||||
pr_debug("next_to_clean : %d\n", tx_ring->next_to_clean);
|
pr_debug("next_to_clean : %d\n", tx_ring->next_to_clean);
|
||||||
|
spin_unlock(&tx_ring->tx_lock);
|
||||||
return cleaned;
|
return cleaned;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2037,7 +2035,6 @@ static int pch_gbe_sw_init(struct pch_gbe_adapter *adapter)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
}
|
}
|
||||||
spin_lock_init(&adapter->hw.miim_lock);
|
spin_lock_init(&adapter->hw.miim_lock);
|
||||||
spin_lock_init(&adapter->tx_queue_lock);
|
|
||||||
spin_lock_init(&adapter->stats_lock);
|
spin_lock_init(&adapter->stats_lock);
|
||||||
spin_lock_init(&adapter->ethtool_lock);
|
spin_lock_init(&adapter->ethtool_lock);
|
||||||
atomic_set(&adapter->irq_sem, 0);
|
atomic_set(&adapter->irq_sem, 0);
|
||||||
|
@ -2142,10 +2139,10 @@ static int pch_gbe_xmit_frame(struct sk_buff *skb, struct net_device *netdev)
|
||||||
tx_ring->next_to_use, tx_ring->next_to_clean);
|
tx_ring->next_to_use, tx_ring->next_to_clean);
|
||||||
return NETDEV_TX_BUSY;
|
return NETDEV_TX_BUSY;
|
||||||
}
|
}
|
||||||
spin_unlock_irqrestore(&tx_ring->tx_lock, flags);
|
|
||||||
|
|
||||||
/* CRC,ITAG no support */
|
/* CRC,ITAG no support */
|
||||||
pch_gbe_tx_queue(adapter, tx_ring, skb);
|
pch_gbe_tx_queue(adapter, tx_ring, skb);
|
||||||
|
spin_unlock_irqrestore(&tx_ring->tx_lock, flags);
|
||||||
return NETDEV_TX_OK;
|
return NETDEV_TX_OK;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -485,6 +485,7 @@ static const struct driver_info wwan_info = {
|
||||||
/*-------------------------------------------------------------------------*/
|
/*-------------------------------------------------------------------------*/
|
||||||
|
|
||||||
#define HUAWEI_VENDOR_ID 0x12D1
|
#define HUAWEI_VENDOR_ID 0x12D1
|
||||||
|
#define NOVATEL_VENDOR_ID 0x1410
|
||||||
|
|
||||||
static const struct usb_device_id products [] = {
|
static const struct usb_device_id products [] = {
|
||||||
/*
|
/*
|
||||||
|
@ -602,6 +603,21 @@ static const struct usb_device_id products [] = {
|
||||||
* because of bugs/quirks in a given product (like Zaurus, above).
|
* because of bugs/quirks in a given product (like Zaurus, above).
|
||||||
*/
|
*/
|
||||||
{
|
{
|
||||||
|
/* Novatel USB551L */
|
||||||
|
/* This match must come *before* the generic CDC-ETHER match so that
|
||||||
|
* we get FLAG_WWAN set on the device, since it's descriptors are
|
||||||
|
* generic CDC-ETHER.
|
||||||
|
*/
|
||||||
|
.match_flags = USB_DEVICE_ID_MATCH_VENDOR
|
||||||
|
| USB_DEVICE_ID_MATCH_PRODUCT
|
||||||
|
| USB_DEVICE_ID_MATCH_INT_INFO,
|
||||||
|
.idVendor = NOVATEL_VENDOR_ID,
|
||||||
|
.idProduct = 0xB001,
|
||||||
|
.bInterfaceClass = USB_CLASS_COMM,
|
||||||
|
.bInterfaceSubClass = USB_CDC_SUBCLASS_ETHERNET,
|
||||||
|
.bInterfaceProtocol = USB_CDC_PROTO_NONE,
|
||||||
|
.driver_info = (unsigned long)&wwan_info,
|
||||||
|
}, {
|
||||||
USB_INTERFACE_INFO(USB_CLASS_COMM, USB_CDC_SUBCLASS_ETHERNET,
|
USB_INTERFACE_INFO(USB_CLASS_COMM, USB_CDC_SUBCLASS_ETHERNET,
|
||||||
USB_CDC_PROTO_NONE),
|
USB_CDC_PROTO_NONE),
|
||||||
.driver_info = (unsigned long) &cdc_info,
|
.driver_info = (unsigned long) &cdc_info,
|
||||||
|
|
|
@ -282,17 +282,32 @@ int usbnet_change_mtu (struct net_device *net, int new_mtu)
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(usbnet_change_mtu);
|
EXPORT_SYMBOL_GPL(usbnet_change_mtu);
|
||||||
|
|
||||||
|
/* The caller must hold list->lock */
|
||||||
|
static void __usbnet_queue_skb(struct sk_buff_head *list,
|
||||||
|
struct sk_buff *newsk, enum skb_state state)
|
||||||
|
{
|
||||||
|
struct skb_data *entry = (struct skb_data *) newsk->cb;
|
||||||
|
|
||||||
|
__skb_queue_tail(list, newsk);
|
||||||
|
entry->state = state;
|
||||||
|
}
|
||||||
|
|
||||||
/*-------------------------------------------------------------------------*/
|
/*-------------------------------------------------------------------------*/
|
||||||
|
|
||||||
/* some LK 2.4 HCDs oopsed if we freed or resubmitted urbs from
|
/* some LK 2.4 HCDs oopsed if we freed or resubmitted urbs from
|
||||||
* completion callbacks. 2.5 should have fixed those bugs...
|
* completion callbacks. 2.5 should have fixed those bugs...
|
||||||
*/
|
*/
|
||||||
|
|
||||||
static void defer_bh(struct usbnet *dev, struct sk_buff *skb, struct sk_buff_head *list)
|
static enum skb_state defer_bh(struct usbnet *dev, struct sk_buff *skb,
|
||||||
|
struct sk_buff_head *list, enum skb_state state)
|
||||||
{
|
{
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
enum skb_state old_state;
|
||||||
|
struct skb_data *entry = (struct skb_data *) skb->cb;
|
||||||
|
|
||||||
spin_lock_irqsave(&list->lock, flags);
|
spin_lock_irqsave(&list->lock, flags);
|
||||||
|
old_state = entry->state;
|
||||||
|
entry->state = state;
|
||||||
__skb_unlink(skb, list);
|
__skb_unlink(skb, list);
|
||||||
spin_unlock(&list->lock);
|
spin_unlock(&list->lock);
|
||||||
spin_lock(&dev->done.lock);
|
spin_lock(&dev->done.lock);
|
||||||
|
@ -300,6 +315,7 @@ static void defer_bh(struct usbnet *dev, struct sk_buff *skb, struct sk_buff_hea
|
||||||
if (dev->done.qlen == 1)
|
if (dev->done.qlen == 1)
|
||||||
tasklet_schedule(&dev->bh);
|
tasklet_schedule(&dev->bh);
|
||||||
spin_unlock_irqrestore(&dev->done.lock, flags);
|
spin_unlock_irqrestore(&dev->done.lock, flags);
|
||||||
|
return old_state;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* some work can't be done in tasklets, so we use keventd
|
/* some work can't be done in tasklets, so we use keventd
|
||||||
|
@ -340,7 +356,6 @@ static int rx_submit (struct usbnet *dev, struct urb *urb, gfp_t flags)
|
||||||
entry = (struct skb_data *) skb->cb;
|
entry = (struct skb_data *) skb->cb;
|
||||||
entry->urb = urb;
|
entry->urb = urb;
|
||||||
entry->dev = dev;
|
entry->dev = dev;
|
||||||
entry->state = rx_start;
|
|
||||||
entry->length = 0;
|
entry->length = 0;
|
||||||
|
|
||||||
usb_fill_bulk_urb (urb, dev->udev, dev->in,
|
usb_fill_bulk_urb (urb, dev->udev, dev->in,
|
||||||
|
@ -372,7 +387,7 @@ static int rx_submit (struct usbnet *dev, struct urb *urb, gfp_t flags)
|
||||||
tasklet_schedule (&dev->bh);
|
tasklet_schedule (&dev->bh);
|
||||||
break;
|
break;
|
||||||
case 0:
|
case 0:
|
||||||
__skb_queue_tail (&dev->rxq, skb);
|
__usbnet_queue_skb(&dev->rxq, skb, rx_start);
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
netif_dbg(dev, ifdown, dev->net, "rx: stopped\n");
|
netif_dbg(dev, ifdown, dev->net, "rx: stopped\n");
|
||||||
|
@ -423,16 +438,17 @@ static void rx_complete (struct urb *urb)
|
||||||
struct skb_data *entry = (struct skb_data *) skb->cb;
|
struct skb_data *entry = (struct skb_data *) skb->cb;
|
||||||
struct usbnet *dev = entry->dev;
|
struct usbnet *dev = entry->dev;
|
||||||
int urb_status = urb->status;
|
int urb_status = urb->status;
|
||||||
|
enum skb_state state;
|
||||||
|
|
||||||
skb_put (skb, urb->actual_length);
|
skb_put (skb, urb->actual_length);
|
||||||
entry->state = rx_done;
|
state = rx_done;
|
||||||
entry->urb = NULL;
|
entry->urb = NULL;
|
||||||
|
|
||||||
switch (urb_status) {
|
switch (urb_status) {
|
||||||
/* success */
|
/* success */
|
||||||
case 0:
|
case 0:
|
||||||
if (skb->len < dev->net->hard_header_len) {
|
if (skb->len < dev->net->hard_header_len) {
|
||||||
entry->state = rx_cleanup;
|
state = rx_cleanup;
|
||||||
dev->net->stats.rx_errors++;
|
dev->net->stats.rx_errors++;
|
||||||
dev->net->stats.rx_length_errors++;
|
dev->net->stats.rx_length_errors++;
|
||||||
netif_dbg(dev, rx_err, dev->net,
|
netif_dbg(dev, rx_err, dev->net,
|
||||||
|
@ -471,7 +487,7 @@ static void rx_complete (struct urb *urb)
|
||||||
"rx throttle %d\n", urb_status);
|
"rx throttle %d\n", urb_status);
|
||||||
}
|
}
|
||||||
block:
|
block:
|
||||||
entry->state = rx_cleanup;
|
state = rx_cleanup;
|
||||||
entry->urb = urb;
|
entry->urb = urb;
|
||||||
urb = NULL;
|
urb = NULL;
|
||||||
break;
|
break;
|
||||||
|
@ -482,17 +498,18 @@ block:
|
||||||
// FALLTHROUGH
|
// FALLTHROUGH
|
||||||
|
|
||||||
default:
|
default:
|
||||||
entry->state = rx_cleanup;
|
state = rx_cleanup;
|
||||||
dev->net->stats.rx_errors++;
|
dev->net->stats.rx_errors++;
|
||||||
netif_dbg(dev, rx_err, dev->net, "rx status %d\n", urb_status);
|
netif_dbg(dev, rx_err, dev->net, "rx status %d\n", urb_status);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
defer_bh(dev, skb, &dev->rxq);
|
state = defer_bh(dev, skb, &dev->rxq, state);
|
||||||
|
|
||||||
if (urb) {
|
if (urb) {
|
||||||
if (netif_running (dev->net) &&
|
if (netif_running (dev->net) &&
|
||||||
!test_bit (EVENT_RX_HALT, &dev->flags)) {
|
!test_bit (EVENT_RX_HALT, &dev->flags) &&
|
||||||
|
state != unlink_start) {
|
||||||
rx_submit (dev, urb, GFP_ATOMIC);
|
rx_submit (dev, urb, GFP_ATOMIC);
|
||||||
usb_mark_last_busy(dev->udev);
|
usb_mark_last_busy(dev->udev);
|
||||||
return;
|
return;
|
||||||
|
@ -579,16 +596,23 @@ EXPORT_SYMBOL_GPL(usbnet_purge_paused_rxq);
|
||||||
static int unlink_urbs (struct usbnet *dev, struct sk_buff_head *q)
|
static int unlink_urbs (struct usbnet *dev, struct sk_buff_head *q)
|
||||||
{
|
{
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
struct sk_buff *skb, *skbnext;
|
struct sk_buff *skb;
|
||||||
int count = 0;
|
int count = 0;
|
||||||
|
|
||||||
spin_lock_irqsave (&q->lock, flags);
|
spin_lock_irqsave (&q->lock, flags);
|
||||||
skb_queue_walk_safe(q, skb, skbnext) {
|
while (!skb_queue_empty(q)) {
|
||||||
struct skb_data *entry;
|
struct skb_data *entry;
|
||||||
struct urb *urb;
|
struct urb *urb;
|
||||||
int retval;
|
int retval;
|
||||||
|
|
||||||
|
skb_queue_walk(q, skb) {
|
||||||
entry = (struct skb_data *) skb->cb;
|
entry = (struct skb_data *) skb->cb;
|
||||||
|
if (entry->state != unlink_start)
|
||||||
|
goto found;
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
found:
|
||||||
|
entry->state = unlink_start;
|
||||||
urb = entry->urb;
|
urb = entry->urb;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -1039,8 +1063,7 @@ static void tx_complete (struct urb *urb)
|
||||||
}
|
}
|
||||||
|
|
||||||
usb_autopm_put_interface_async(dev->intf);
|
usb_autopm_put_interface_async(dev->intf);
|
||||||
entry->state = tx_done;
|
(void) defer_bh(dev, skb, &dev->txq, tx_done);
|
||||||
defer_bh(dev, skb, &dev->txq);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*-------------------------------------------------------------------------*/
|
/*-------------------------------------------------------------------------*/
|
||||||
|
@ -1096,7 +1119,6 @@ netdev_tx_t usbnet_start_xmit (struct sk_buff *skb,
|
||||||
entry = (struct skb_data *) skb->cb;
|
entry = (struct skb_data *) skb->cb;
|
||||||
entry->urb = urb;
|
entry->urb = urb;
|
||||||
entry->dev = dev;
|
entry->dev = dev;
|
||||||
entry->state = tx_start;
|
|
||||||
entry->length = length;
|
entry->length = length;
|
||||||
|
|
||||||
usb_fill_bulk_urb (urb, dev->udev, dev->out,
|
usb_fill_bulk_urb (urb, dev->udev, dev->out,
|
||||||
|
@ -1155,7 +1177,7 @@ netdev_tx_t usbnet_start_xmit (struct sk_buff *skb,
|
||||||
break;
|
break;
|
||||||
case 0:
|
case 0:
|
||||||
net->trans_start = jiffies;
|
net->trans_start = jiffies;
|
||||||
__skb_queue_tail (&dev->txq, skb);
|
__usbnet_queue_skb(&dev->txq, skb, tx_start);
|
||||||
if (dev->txq.qlen >= TX_QLEN (dev))
|
if (dev->txq.qlen >= TX_QLEN (dev))
|
||||||
netif_stop_queue (net);
|
netif_stop_queue (net);
|
||||||
}
|
}
|
||||||
|
|
|
@ -492,7 +492,9 @@ static void virtnet_napi_enable(struct virtnet_info *vi)
|
||||||
* We synchronize against interrupts via NAPI_STATE_SCHED */
|
* We synchronize against interrupts via NAPI_STATE_SCHED */
|
||||||
if (napi_schedule_prep(&vi->napi)) {
|
if (napi_schedule_prep(&vi->napi)) {
|
||||||
virtqueue_disable_cb(vi->rvq);
|
virtqueue_disable_cb(vi->rvq);
|
||||||
|
local_bh_disable();
|
||||||
__napi_schedule(&vi->napi);
|
__napi_schedule(&vi->napi);
|
||||||
|
local_bh_enable();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1851,14 +1851,6 @@ int __devinit rtl_pci_probe(struct pci_dev *pdev,
|
||||||
/*like read eeprom and so on */
|
/*like read eeprom and so on */
|
||||||
rtlpriv->cfg->ops->read_eeprom_info(hw);
|
rtlpriv->cfg->ops->read_eeprom_info(hw);
|
||||||
|
|
||||||
if (rtlpriv->cfg->ops->init_sw_vars(hw)) {
|
|
||||||
RT_TRACE(rtlpriv, COMP_ERR, DBG_EMERG, "Can't init_sw_vars\n");
|
|
||||||
err = -ENODEV;
|
|
||||||
goto fail3;
|
|
||||||
}
|
|
||||||
|
|
||||||
rtlpriv->cfg->ops->init_sw_leds(hw);
|
|
||||||
|
|
||||||
/*aspm */
|
/*aspm */
|
||||||
rtl_pci_init_aspm(hw);
|
rtl_pci_init_aspm(hw);
|
||||||
|
|
||||||
|
@ -1877,6 +1869,14 @@ int __devinit rtl_pci_probe(struct pci_dev *pdev,
|
||||||
goto fail3;
|
goto fail3;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (rtlpriv->cfg->ops->init_sw_vars(hw)) {
|
||||||
|
RT_TRACE(rtlpriv, COMP_ERR, DBG_EMERG, "Can't init_sw_vars\n");
|
||||||
|
err = -ENODEV;
|
||||||
|
goto fail3;
|
||||||
|
}
|
||||||
|
|
||||||
|
rtlpriv->cfg->ops->init_sw_leds(hw);
|
||||||
|
|
||||||
err = sysfs_create_group(&pdev->dev.kobj, &rtl_attribute_group);
|
err = sysfs_create_group(&pdev->dev.kobj, &rtl_attribute_group);
|
||||||
if (err) {
|
if (err) {
|
||||||
RT_TRACE(rtlpriv, COMP_ERR, DBG_EMERG,
|
RT_TRACE(rtlpriv, COMP_ERR, DBG_EMERG,
|
||||||
|
|
|
@ -971,11 +971,6 @@ int __devinit rtl_usb_probe(struct usb_interface *intf,
|
||||||
rtlpriv->cfg->ops->read_chip_version(hw);
|
rtlpriv->cfg->ops->read_chip_version(hw);
|
||||||
/*like read eeprom and so on */
|
/*like read eeprom and so on */
|
||||||
rtlpriv->cfg->ops->read_eeprom_info(hw);
|
rtlpriv->cfg->ops->read_eeprom_info(hw);
|
||||||
if (rtlpriv->cfg->ops->init_sw_vars(hw)) {
|
|
||||||
RT_TRACE(rtlpriv, COMP_ERR, DBG_EMERG, "Can't init_sw_vars\n");
|
|
||||||
goto error_out;
|
|
||||||
}
|
|
||||||
rtlpriv->cfg->ops->init_sw_leds(hw);
|
|
||||||
err = _rtl_usb_init(hw);
|
err = _rtl_usb_init(hw);
|
||||||
if (err)
|
if (err)
|
||||||
goto error_out;
|
goto error_out;
|
||||||
|
@ -987,6 +982,11 @@ int __devinit rtl_usb_probe(struct usb_interface *intf,
|
||||||
"Can't allocate sw for mac80211\n");
|
"Can't allocate sw for mac80211\n");
|
||||||
goto error_out;
|
goto error_out;
|
||||||
}
|
}
|
||||||
|
if (rtlpriv->cfg->ops->init_sw_vars(hw)) {
|
||||||
|
RT_TRACE(rtlpriv, COMP_ERR, DBG_EMERG, "Can't init_sw_vars\n");
|
||||||
|
goto error_out;
|
||||||
|
}
|
||||||
|
rtlpriv->cfg->ops->init_sw_leds(hw);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
error_out:
|
error_out:
|
||||||
|
|
|
@ -223,7 +223,7 @@ static int acpi_pci_set_power_state(struct pci_dev *dev, pci_power_t state)
|
||||||
[PCI_D0] = ACPI_STATE_D0,
|
[PCI_D0] = ACPI_STATE_D0,
|
||||||
[PCI_D1] = ACPI_STATE_D1,
|
[PCI_D1] = ACPI_STATE_D1,
|
||||||
[PCI_D2] = ACPI_STATE_D2,
|
[PCI_D2] = ACPI_STATE_D2,
|
||||||
[PCI_D3hot] = ACPI_STATE_D3_HOT,
|
[PCI_D3hot] = ACPI_STATE_D3,
|
||||||
[PCI_D3cold] = ACPI_STATE_D3
|
[PCI_D3cold] = ACPI_STATE_D3
|
||||||
};
|
};
|
||||||
int error = -EINVAL;
|
int error = -EINVAL;
|
||||||
|
|
|
@ -30,6 +30,7 @@
|
||||||
#include <linux/module.h>
|
#include <linux/module.h>
|
||||||
#include <linux/pci.h>
|
#include <linux/pci.h>
|
||||||
#include <linux/ptp_clock_kernel.h>
|
#include <linux/ptp_clock_kernel.h>
|
||||||
|
#include <linux/slab.h>
|
||||||
|
|
||||||
#define STATION_ADDR_LEN 20
|
#define STATION_ADDR_LEN 20
|
||||||
#define PCI_DEVICE_ID_PCH_1588 0x8819
|
#define PCI_DEVICE_ID_PCH_1588 0x8819
|
||||||
|
|
|
@ -354,7 +354,7 @@ static void __rproc_free_vrings(struct rproc_vdev *rvdev, int i)
|
||||||
{
|
{
|
||||||
struct rproc *rproc = rvdev->rproc;
|
struct rproc *rproc = rvdev->rproc;
|
||||||
|
|
||||||
for (i--; i > 0; i--) {
|
for (i--; i >= 0; i--) {
|
||||||
struct rproc_vring *rvring = &rvdev->vring[i];
|
struct rproc_vring *rvring = &rvdev->vring[i];
|
||||||
int size = PAGE_ALIGN(vring_size(rvring->len, rvring->align));
|
int size = PAGE_ALIGN(vring_size(rvring->len, rvring->align));
|
||||||
|
|
||||||
|
|
|
@ -312,6 +312,7 @@ static int pl031_probe(struct amba_device *adev, const struct amba_id *id)
|
||||||
int ret;
|
int ret;
|
||||||
struct pl031_local *ldata;
|
struct pl031_local *ldata;
|
||||||
struct rtc_class_ops *ops = id->data;
|
struct rtc_class_ops *ops = id->data;
|
||||||
|
unsigned long time;
|
||||||
|
|
||||||
ret = amba_request_regions(adev, NULL);
|
ret = amba_request_regions(adev, NULL);
|
||||||
if (ret)
|
if (ret)
|
||||||
|
@ -343,6 +344,23 @@ static int pl031_probe(struct amba_device *adev, const struct amba_id *id)
|
||||||
writel(readl(ldata->base + RTC_CR) | RTC_CR_CWEN,
|
writel(readl(ldata->base + RTC_CR) | RTC_CR_CWEN,
|
||||||
ldata->base + RTC_CR);
|
ldata->base + RTC_CR);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* On ST PL031 variants, the RTC reset value does not provide correct
|
||||||
|
* weekday for 2000-01-01. Correct the erroneous sunday to saturday.
|
||||||
|
*/
|
||||||
|
if (ldata->hw_designer == AMBA_VENDOR_ST) {
|
||||||
|
if (readl(ldata->base + RTC_YDR) == 0x2000) {
|
||||||
|
time = readl(ldata->base + RTC_DR);
|
||||||
|
if ((time &
|
||||||
|
(RTC_MON_MASK | RTC_MDAY_MASK | RTC_WDAY_MASK))
|
||||||
|
== 0x02120000) {
|
||||||
|
time = time | (0x7 << RTC_WDAY_SHIFT);
|
||||||
|
writel(0x2000, ldata->base + RTC_YLR);
|
||||||
|
writel(time, ldata->base + RTC_LR);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
ldata->rtc = rtc_device_register("pl031", &adev->dev, ops,
|
ldata->rtc = rtc_device_register("pl031", &adev->dev, ops,
|
||||||
THIS_MODULE);
|
THIS_MODULE);
|
||||||
if (IS_ERR(ldata->rtc)) {
|
if (IS_ERR(ldata->rtc)) {
|
||||||
|
|
|
@ -169,6 +169,7 @@ static struct se_device *fd_create_virtdevice(
|
||||||
inode = file->f_mapping->host;
|
inode = file->f_mapping->host;
|
||||||
if (S_ISBLK(inode->i_mode)) {
|
if (S_ISBLK(inode->i_mode)) {
|
||||||
struct request_queue *q;
|
struct request_queue *q;
|
||||||
|
unsigned long long dev_size;
|
||||||
/*
|
/*
|
||||||
* Setup the local scope queue_limits from struct request_queue->limits
|
* Setup the local scope queue_limits from struct request_queue->limits
|
||||||
* to pass into transport_add_device_to_core_hba() as struct se_dev_limits.
|
* to pass into transport_add_device_to_core_hba() as struct se_dev_limits.
|
||||||
|
@ -183,13 +184,12 @@ static struct se_device *fd_create_virtdevice(
|
||||||
* one (1) logical sector from underlying struct block_device
|
* one (1) logical sector from underlying struct block_device
|
||||||
*/
|
*/
|
||||||
fd_dev->fd_block_size = bdev_logical_block_size(inode->i_bdev);
|
fd_dev->fd_block_size = bdev_logical_block_size(inode->i_bdev);
|
||||||
fd_dev->fd_dev_size = (i_size_read(file->f_mapping->host) -
|
dev_size = (i_size_read(file->f_mapping->host) -
|
||||||
fd_dev->fd_block_size);
|
fd_dev->fd_block_size);
|
||||||
|
|
||||||
pr_debug("FILEIO: Using size: %llu bytes from struct"
|
pr_debug("FILEIO: Using size: %llu bytes from struct"
|
||||||
" block_device blocks: %llu logical_block_size: %d\n",
|
" block_device blocks: %llu logical_block_size: %d\n",
|
||||||
fd_dev->fd_dev_size,
|
dev_size, div_u64(dev_size, fd_dev->fd_block_size),
|
||||||
div_u64(fd_dev->fd_dev_size, fd_dev->fd_block_size),
|
|
||||||
fd_dev->fd_block_size);
|
fd_dev->fd_block_size);
|
||||||
} else {
|
} else {
|
||||||
if (!(fd_dev->fbd_flags & FBDF_HAS_SIZE)) {
|
if (!(fd_dev->fbd_flags & FBDF_HAS_SIZE)) {
|
||||||
|
@ -605,10 +605,20 @@ static u32 fd_get_device_type(struct se_device *dev)
|
||||||
static sector_t fd_get_blocks(struct se_device *dev)
|
static sector_t fd_get_blocks(struct se_device *dev)
|
||||||
{
|
{
|
||||||
struct fd_dev *fd_dev = dev->dev_ptr;
|
struct fd_dev *fd_dev = dev->dev_ptr;
|
||||||
unsigned long long blocks_long = div_u64(fd_dev->fd_dev_size,
|
struct file *f = fd_dev->fd_file;
|
||||||
dev->se_sub_dev->se_dev_attrib.block_size);
|
struct inode *i = f->f_mapping->host;
|
||||||
|
unsigned long long dev_size;
|
||||||
|
/*
|
||||||
|
* When using a file that references an underlying struct block_device,
|
||||||
|
* ensure dev_size is always based on the current inode size in order
|
||||||
|
* to handle underlying block_device resize operations.
|
||||||
|
*/
|
||||||
|
if (S_ISBLK(i->i_mode))
|
||||||
|
dev_size = (i_size_read(i) - fd_dev->fd_block_size);
|
||||||
|
else
|
||||||
|
dev_size = fd_dev->fd_dev_size;
|
||||||
|
|
||||||
return blocks_long;
|
return div_u64(dev_size, dev->se_sub_dev->se_dev_attrib.block_size);
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct se_subsystem_api fileio_template = {
|
static struct se_subsystem_api fileio_template = {
|
||||||
|
|
|
@ -220,6 +220,9 @@ int target_scsi2_reservation_release(struct se_task *task)
|
||||||
if (dev->dev_reserved_node_acl != sess->se_node_acl)
|
if (dev->dev_reserved_node_acl != sess->se_node_acl)
|
||||||
goto out_unlock;
|
goto out_unlock;
|
||||||
|
|
||||||
|
if (dev->dev_res_bin_isid != sess->sess_bin_isid)
|
||||||
|
goto out_unlock;
|
||||||
|
|
||||||
dev->dev_reserved_node_acl = NULL;
|
dev->dev_reserved_node_acl = NULL;
|
||||||
dev->dev_flags &= ~DF_SPC2_RESERVATIONS;
|
dev->dev_flags &= ~DF_SPC2_RESERVATIONS;
|
||||||
if (dev->dev_flags & DF_SPC2_RESERVATIONS_WITH_ISID) {
|
if (dev->dev_flags & DF_SPC2_RESERVATIONS_WITH_ISID) {
|
||||||
|
|
|
@ -2044,7 +2044,7 @@ int vt_do_kdskled(int console, int cmd, unsigned long arg, int perm)
|
||||||
kbd->default_ledflagstate = ((arg >> 4) & 7);
|
kbd->default_ledflagstate = ((arg >> 4) & 7);
|
||||||
set_leds();
|
set_leds();
|
||||||
spin_unlock_irqrestore(&kbd_event_lock, flags);
|
spin_unlock_irqrestore(&kbd_event_lock, flags);
|
||||||
break;
|
return 0;
|
||||||
|
|
||||||
/* the ioctls below only set the lights, not the functions */
|
/* the ioctls below only set the lights, not the functions */
|
||||||
/* for those, see KDGKBLED and KDSKBLED above */
|
/* for those, see KDGKBLED and KDSKBLED above */
|
||||||
|
|
|
@ -390,6 +390,7 @@ static void __devexit virtballoon_remove(struct virtio_device *vdev)
|
||||||
/* There might be pages left in the balloon: free them. */
|
/* There might be pages left in the balloon: free them. */
|
||||||
while (vb->num_pages)
|
while (vb->num_pages)
|
||||||
leak_balloon(vb, vb->num_pages);
|
leak_balloon(vb, vb->num_pages);
|
||||||
|
update_balloon_size(vb);
|
||||||
|
|
||||||
/* Now we reset the device so we can clean up the queues. */
|
/* Now we reset the device so we can clean up the queues. */
|
||||||
vdev->config->reset(vdev);
|
vdev->config->reset(vdev);
|
||||||
|
|
7
fs/bio.c
7
fs/bio.c
|
@ -505,9 +505,14 @@ EXPORT_SYMBOL(bio_clone);
|
||||||
int bio_get_nr_vecs(struct block_device *bdev)
|
int bio_get_nr_vecs(struct block_device *bdev)
|
||||||
{
|
{
|
||||||
struct request_queue *q = bdev_get_queue(bdev);
|
struct request_queue *q = bdev_get_queue(bdev);
|
||||||
return min_t(unsigned,
|
int nr_pages;
|
||||||
|
|
||||||
|
nr_pages = min_t(unsigned,
|
||||||
queue_max_segments(q),
|
queue_max_segments(q),
|
||||||
queue_max_sectors(q) / (PAGE_SIZE >> 9) + 1);
|
queue_max_sectors(q) / (PAGE_SIZE >> 9) + 1);
|
||||||
|
|
||||||
|
return min_t(unsigned, nr_pages, BIO_MAX_PAGES);
|
||||||
|
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(bio_get_nr_vecs);
|
EXPORT_SYMBOL(bio_get_nr_vecs);
|
||||||
|
|
||||||
|
|
|
@ -70,7 +70,7 @@ static void bdev_inode_switch_bdi(struct inode *inode,
|
||||||
spin_unlock(&dst->wb.list_lock);
|
spin_unlock(&dst->wb.list_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
static sector_t max_block(struct block_device *bdev)
|
sector_t blkdev_max_block(struct block_device *bdev)
|
||||||
{
|
{
|
||||||
sector_t retval = ~((sector_t)0);
|
sector_t retval = ~((sector_t)0);
|
||||||
loff_t sz = i_size_read(bdev->bd_inode);
|
loff_t sz = i_size_read(bdev->bd_inode);
|
||||||
|
@ -163,7 +163,7 @@ static int
|
||||||
blkdev_get_block(struct inode *inode, sector_t iblock,
|
blkdev_get_block(struct inode *inode, sector_t iblock,
|
||||||
struct buffer_head *bh, int create)
|
struct buffer_head *bh, int create)
|
||||||
{
|
{
|
||||||
if (iblock >= max_block(I_BDEV(inode))) {
|
if (iblock >= blkdev_max_block(I_BDEV(inode))) {
|
||||||
if (create)
|
if (create)
|
||||||
return -EIO;
|
return -EIO;
|
||||||
|
|
||||||
|
@ -185,7 +185,7 @@ static int
|
||||||
blkdev_get_blocks(struct inode *inode, sector_t iblock,
|
blkdev_get_blocks(struct inode *inode, sector_t iblock,
|
||||||
struct buffer_head *bh, int create)
|
struct buffer_head *bh, int create)
|
||||||
{
|
{
|
||||||
sector_t end_block = max_block(I_BDEV(inode));
|
sector_t end_block = blkdev_max_block(I_BDEV(inode));
|
||||||
unsigned long max_blocks = bh->b_size >> inode->i_blkbits;
|
unsigned long max_blocks = bh->b_size >> inode->i_blkbits;
|
||||||
|
|
||||||
if ((iblock + max_blocks) > end_block) {
|
if ((iblock + max_blocks) > end_block) {
|
||||||
|
|
|
@ -921,6 +921,7 @@ init_page_buffers(struct page *page, struct block_device *bdev,
|
||||||
struct buffer_head *head = page_buffers(page);
|
struct buffer_head *head = page_buffers(page);
|
||||||
struct buffer_head *bh = head;
|
struct buffer_head *bh = head;
|
||||||
int uptodate = PageUptodate(page);
|
int uptodate = PageUptodate(page);
|
||||||
|
sector_t end_block = blkdev_max_block(I_BDEV(bdev->bd_inode));
|
||||||
|
|
||||||
do {
|
do {
|
||||||
if (!buffer_mapped(bh)) {
|
if (!buffer_mapped(bh)) {
|
||||||
|
@ -929,6 +930,7 @@ init_page_buffers(struct page *page, struct block_device *bdev,
|
||||||
bh->b_blocknr = block;
|
bh->b_blocknr = block;
|
||||||
if (uptodate)
|
if (uptodate)
|
||||||
set_buffer_uptodate(bh);
|
set_buffer_uptodate(bh);
|
||||||
|
if (block < end_block)
|
||||||
set_buffer_mapped(bh);
|
set_buffer_mapped(bh);
|
||||||
}
|
}
|
||||||
block++;
|
block++;
|
||||||
|
|
|
@ -164,7 +164,8 @@ static const match_table_t cifs_mount_option_tokens = {
|
||||||
{ Opt_sign, "sign" },
|
{ Opt_sign, "sign" },
|
||||||
{ Opt_seal, "seal" },
|
{ Opt_seal, "seal" },
|
||||||
{ Opt_direct, "direct" },
|
{ Opt_direct, "direct" },
|
||||||
{ Opt_direct, "forceddirectio" },
|
{ Opt_direct, "directio" },
|
||||||
|
{ Opt_direct, "forcedirectio" },
|
||||||
{ Opt_strictcache, "strictcache" },
|
{ Opt_strictcache, "strictcache" },
|
||||||
{ Opt_noac, "noac" },
|
{ Opt_noac, "noac" },
|
||||||
{ Opt_fsc, "fsc" },
|
{ Opt_fsc, "fsc" },
|
||||||
|
|
|
@ -234,8 +234,8 @@ int jffs2_garbage_collect_pass(struct jffs2_sb_info *c)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
jffs2_dbg(1, "No progress from erasing block; doing GC anyway\n");
|
jffs2_dbg(1, "No progress from erasing block; doing GC anyway\n");
|
||||||
spin_lock(&c->erase_completion_lock);
|
|
||||||
mutex_lock(&c->alloc_sem);
|
mutex_lock(&c->alloc_sem);
|
||||||
|
spin_lock(&c->erase_completion_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* First, work out which block we're garbage-collecting */
|
/* First, work out which block we're garbage-collecting */
|
||||||
|
|
|
@ -1834,10 +1834,15 @@ static int tid_fd_revalidate(struct dentry *dentry, struct nameidata *nd)
|
||||||
if (task) {
|
if (task) {
|
||||||
files = get_files_struct(task);
|
files = get_files_struct(task);
|
||||||
if (files) {
|
if (files) {
|
||||||
|
struct file *file;
|
||||||
rcu_read_lock();
|
rcu_read_lock();
|
||||||
if (fcheck_files(files, fd)) {
|
file = fcheck_files(files, fd);
|
||||||
|
if (file) {
|
||||||
|
unsigned i_mode, f_mode = file->f_mode;
|
||||||
|
|
||||||
rcu_read_unlock();
|
rcu_read_unlock();
|
||||||
put_files_struct(files);
|
put_files_struct(files);
|
||||||
|
|
||||||
if (task_dumpable(task)) {
|
if (task_dumpable(task)) {
|
||||||
rcu_read_lock();
|
rcu_read_lock();
|
||||||
cred = __task_cred(task);
|
cred = __task_cred(task);
|
||||||
|
@ -1848,7 +1853,14 @@ static int tid_fd_revalidate(struct dentry *dentry, struct nameidata *nd)
|
||||||
inode->i_uid = 0;
|
inode->i_uid = 0;
|
||||||
inode->i_gid = 0;
|
inode->i_gid = 0;
|
||||||
}
|
}
|
||||||
inode->i_mode &= ~(S_ISUID | S_ISGID);
|
|
||||||
|
i_mode = S_IFLNK;
|
||||||
|
if (f_mode & FMODE_READ)
|
||||||
|
i_mode |= S_IRUSR | S_IXUSR;
|
||||||
|
if (f_mode & FMODE_WRITE)
|
||||||
|
i_mode |= S_IWUSR | S_IXUSR;
|
||||||
|
inode->i_mode = i_mode;
|
||||||
|
|
||||||
security_task_to_inode(task, inode);
|
security_task_to_inode(task, inode);
|
||||||
put_task_struct(task);
|
put_task_struct(task);
|
||||||
return 1;
|
return 1;
|
||||||
|
@ -1872,8 +1884,6 @@ static struct dentry *proc_fd_instantiate(struct inode *dir,
|
||||||
struct dentry *dentry, struct task_struct *task, const void *ptr)
|
struct dentry *dentry, struct task_struct *task, const void *ptr)
|
||||||
{
|
{
|
||||||
unsigned fd = *(const unsigned *)ptr;
|
unsigned fd = *(const unsigned *)ptr;
|
||||||
struct file *file;
|
|
||||||
struct files_struct *files;
|
|
||||||
struct inode *inode;
|
struct inode *inode;
|
||||||
struct proc_inode *ei;
|
struct proc_inode *ei;
|
||||||
struct dentry *error = ERR_PTR(-ENOENT);
|
struct dentry *error = ERR_PTR(-ENOENT);
|
||||||
|
@ -1883,25 +1893,6 @@ static struct dentry *proc_fd_instantiate(struct inode *dir,
|
||||||
goto out;
|
goto out;
|
||||||
ei = PROC_I(inode);
|
ei = PROC_I(inode);
|
||||||
ei->fd = fd;
|
ei->fd = fd;
|
||||||
files = get_files_struct(task);
|
|
||||||
if (!files)
|
|
||||||
goto out_iput;
|
|
||||||
inode->i_mode = S_IFLNK;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* We are not taking a ref to the file structure, so we must
|
|
||||||
* hold ->file_lock.
|
|
||||||
*/
|
|
||||||
spin_lock(&files->file_lock);
|
|
||||||
file = fcheck_files(files, fd);
|
|
||||||
if (!file)
|
|
||||||
goto out_unlock;
|
|
||||||
if (file->f_mode & FMODE_READ)
|
|
||||||
inode->i_mode |= S_IRUSR | S_IXUSR;
|
|
||||||
if (file->f_mode & FMODE_WRITE)
|
|
||||||
inode->i_mode |= S_IWUSR | S_IXUSR;
|
|
||||||
spin_unlock(&files->file_lock);
|
|
||||||
put_files_struct(files);
|
|
||||||
|
|
||||||
inode->i_op = &proc_pid_link_inode_operations;
|
inode->i_op = &proc_pid_link_inode_operations;
|
||||||
inode->i_size = 64;
|
inode->i_size = 64;
|
||||||
|
@ -1914,12 +1905,6 @@ static struct dentry *proc_fd_instantiate(struct inode *dir,
|
||||||
|
|
||||||
out:
|
out:
|
||||||
return error;
|
return error;
|
||||||
out_unlock:
|
|
||||||
spin_unlock(&files->file_lock);
|
|
||||||
put_files_struct(files);
|
|
||||||
out_iput:
|
|
||||||
iput(inode);
|
|
||||||
goto out;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct dentry *proc_lookupfd_common(struct inode *dir,
|
static struct dentry *proc_lookupfd_common(struct inode *dir,
|
||||||
|
@ -2212,16 +2197,16 @@ static struct dentry *proc_map_files_lookup(struct inode *dir,
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
result = ERR_PTR(-EACCES);
|
result = ERR_PTR(-EACCES);
|
||||||
if (lock_trace(task))
|
if (!ptrace_may_access(task, PTRACE_MODE_READ))
|
||||||
goto out_put_task;
|
goto out_put_task;
|
||||||
|
|
||||||
result = ERR_PTR(-ENOENT);
|
result = ERR_PTR(-ENOENT);
|
||||||
if (dname_to_vma_addr(dentry, &vm_start, &vm_end))
|
if (dname_to_vma_addr(dentry, &vm_start, &vm_end))
|
||||||
goto out_unlock;
|
goto out_put_task;
|
||||||
|
|
||||||
mm = get_task_mm(task);
|
mm = get_task_mm(task);
|
||||||
if (!mm)
|
if (!mm)
|
||||||
goto out_unlock;
|
goto out_put_task;
|
||||||
|
|
||||||
down_read(&mm->mmap_sem);
|
down_read(&mm->mmap_sem);
|
||||||
vma = find_exact_vma(mm, vm_start, vm_end);
|
vma = find_exact_vma(mm, vm_start, vm_end);
|
||||||
|
@ -2233,8 +2218,6 @@ static struct dentry *proc_map_files_lookup(struct inode *dir,
|
||||||
out_no_vma:
|
out_no_vma:
|
||||||
up_read(&mm->mmap_sem);
|
up_read(&mm->mmap_sem);
|
||||||
mmput(mm);
|
mmput(mm);
|
||||||
out_unlock:
|
|
||||||
unlock_trace(task);
|
|
||||||
out_put_task:
|
out_put_task:
|
||||||
put_task_struct(task);
|
put_task_struct(task);
|
||||||
out:
|
out:
|
||||||
|
@ -2268,7 +2251,7 @@ proc_map_files_readdir(struct file *filp, void *dirent, filldir_t filldir)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
ret = -EACCES;
|
ret = -EACCES;
|
||||||
if (lock_trace(task))
|
if (!ptrace_may_access(task, PTRACE_MODE_READ))
|
||||||
goto out_put_task;
|
goto out_put_task;
|
||||||
|
|
||||||
ret = 0;
|
ret = 0;
|
||||||
|
@ -2276,12 +2259,12 @@ proc_map_files_readdir(struct file *filp, void *dirent, filldir_t filldir)
|
||||||
case 0:
|
case 0:
|
||||||
ino = inode->i_ino;
|
ino = inode->i_ino;
|
||||||
if (filldir(dirent, ".", 1, 0, ino, DT_DIR) < 0)
|
if (filldir(dirent, ".", 1, 0, ino, DT_DIR) < 0)
|
||||||
goto out_unlock;
|
goto out_put_task;
|
||||||
filp->f_pos++;
|
filp->f_pos++;
|
||||||
case 1:
|
case 1:
|
||||||
ino = parent_ino(dentry);
|
ino = parent_ino(dentry);
|
||||||
if (filldir(dirent, "..", 2, 1, ino, DT_DIR) < 0)
|
if (filldir(dirent, "..", 2, 1, ino, DT_DIR) < 0)
|
||||||
goto out_unlock;
|
goto out_put_task;
|
||||||
filp->f_pos++;
|
filp->f_pos++;
|
||||||
default:
|
default:
|
||||||
{
|
{
|
||||||
|
@ -2292,7 +2275,7 @@ proc_map_files_readdir(struct file *filp, void *dirent, filldir_t filldir)
|
||||||
|
|
||||||
mm = get_task_mm(task);
|
mm = get_task_mm(task);
|
||||||
if (!mm)
|
if (!mm)
|
||||||
goto out_unlock;
|
goto out_put_task;
|
||||||
down_read(&mm->mmap_sem);
|
down_read(&mm->mmap_sem);
|
||||||
|
|
||||||
nr_files = 0;
|
nr_files = 0;
|
||||||
|
@ -2322,7 +2305,7 @@ proc_map_files_readdir(struct file *filp, void *dirent, filldir_t filldir)
|
||||||
flex_array_free(fa);
|
flex_array_free(fa);
|
||||||
up_read(&mm->mmap_sem);
|
up_read(&mm->mmap_sem);
|
||||||
mmput(mm);
|
mmput(mm);
|
||||||
goto out_unlock;
|
goto out_put_task;
|
||||||
}
|
}
|
||||||
for (i = 0, vma = mm->mmap, pos = 2; vma;
|
for (i = 0, vma = mm->mmap, pos = 2; vma;
|
||||||
vma = vma->vm_next) {
|
vma = vma->vm_next) {
|
||||||
|
@ -2367,8 +2350,6 @@ proc_map_files_readdir(struct file *filp, void *dirent, filldir_t filldir)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
out_unlock:
|
|
||||||
unlock_trace(task);
|
|
||||||
out_put_task:
|
out_put_task:
|
||||||
put_task_struct(task);
|
put_task_struct(task);
|
||||||
out:
|
out:
|
||||||
|
|
|
@ -1,9 +1,10 @@
|
||||||
#ifndef _LINUX_BLKDEV_H
|
#ifndef _LINUX_BLKDEV_H
|
||||||
#define _LINUX_BLKDEV_H
|
#define _LINUX_BLKDEV_H
|
||||||
|
|
||||||
|
#include <linux/sched.h>
|
||||||
|
|
||||||
#ifdef CONFIG_BLOCK
|
#ifdef CONFIG_BLOCK
|
||||||
|
|
||||||
#include <linux/sched.h>
|
|
||||||
#include <linux/major.h>
|
#include <linux/major.h>
|
||||||
#include <linux/genhd.h>
|
#include <linux/genhd.h>
|
||||||
#include <linux/list.h>
|
#include <linux/list.h>
|
||||||
|
|
|
@ -2051,6 +2051,7 @@ extern void unregister_blkdev(unsigned int, const char *);
|
||||||
extern struct block_device *bdget(dev_t);
|
extern struct block_device *bdget(dev_t);
|
||||||
extern struct block_device *bdgrab(struct block_device *bdev);
|
extern struct block_device *bdgrab(struct block_device *bdev);
|
||||||
extern void bd_set_size(struct block_device *, loff_t size);
|
extern void bd_set_size(struct block_device *, loff_t size);
|
||||||
|
extern sector_t blkdev_max_block(struct block_device *bdev);
|
||||||
extern void bd_forget(struct inode *inode);
|
extern void bd_forget(struct inode *inode);
|
||||||
extern void bdput(struct block_device *);
|
extern void bdput(struct block_device *);
|
||||||
extern void invalidate_bdev(struct block_device *);
|
extern void invalidate_bdev(struct block_device *);
|
||||||
|
|
|
@ -179,6 +179,7 @@ enum {
|
||||||
TRACE_EVENT_FL_RECORDED_CMD_BIT,
|
TRACE_EVENT_FL_RECORDED_CMD_BIT,
|
||||||
TRACE_EVENT_FL_CAP_ANY_BIT,
|
TRACE_EVENT_FL_CAP_ANY_BIT,
|
||||||
TRACE_EVENT_FL_NO_SET_FILTER_BIT,
|
TRACE_EVENT_FL_NO_SET_FILTER_BIT,
|
||||||
|
TRACE_EVENT_FL_IGNORE_ENABLE_BIT,
|
||||||
};
|
};
|
||||||
|
|
||||||
enum {
|
enum {
|
||||||
|
@ -187,6 +188,7 @@ enum {
|
||||||
TRACE_EVENT_FL_RECORDED_CMD = (1 << TRACE_EVENT_FL_RECORDED_CMD_BIT),
|
TRACE_EVENT_FL_RECORDED_CMD = (1 << TRACE_EVENT_FL_RECORDED_CMD_BIT),
|
||||||
TRACE_EVENT_FL_CAP_ANY = (1 << TRACE_EVENT_FL_CAP_ANY_BIT),
|
TRACE_EVENT_FL_CAP_ANY = (1 << TRACE_EVENT_FL_CAP_ANY_BIT),
|
||||||
TRACE_EVENT_FL_NO_SET_FILTER = (1 << TRACE_EVENT_FL_NO_SET_FILTER_BIT),
|
TRACE_EVENT_FL_NO_SET_FILTER = (1 << TRACE_EVENT_FL_NO_SET_FILTER_BIT),
|
||||||
|
TRACE_EVENT_FL_IGNORE_ENABLE = (1 << TRACE_EVENT_FL_IGNORE_ENABLE_BIT),
|
||||||
};
|
};
|
||||||
|
|
||||||
struct ftrace_event_call {
|
struct ftrace_event_call {
|
||||||
|
|
|
@ -222,12 +222,6 @@ static inline void part_pack_uuid(const u8 *uuid_str, u8 *to)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline char *part_unpack_uuid(const u8 *uuid, char *out)
|
|
||||||
{
|
|
||||||
sprintf(out, "%pU", uuid);
|
|
||||||
return out;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline int disk_max_parts(struct gendisk *disk)
|
static inline int disk_max_parts(struct gendisk *disk)
|
||||||
{
|
{
|
||||||
if (disk->flags & GENHD_FL_EXT_DEVT)
|
if (disk->flags & GENHD_FL_EXT_DEVT)
|
||||||
|
|
|
@ -99,6 +99,22 @@ struct ip_set_hash {
|
||||||
#endif
|
#endif
|
||||||
};
|
};
|
||||||
|
|
||||||
|
static size_t
|
||||||
|
htable_size(u8 hbits)
|
||||||
|
{
|
||||||
|
size_t hsize;
|
||||||
|
|
||||||
|
/* We must fit both into u32 in jhash and size_t */
|
||||||
|
if (hbits > 31)
|
||||||
|
return 0;
|
||||||
|
hsize = jhash_size(hbits);
|
||||||
|
if ((((size_t)-1) - sizeof(struct htable))/sizeof(struct hbucket)
|
||||||
|
< hsize)
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
return hsize * sizeof(struct hbucket) + sizeof(struct htable);
|
||||||
|
}
|
||||||
|
|
||||||
/* Compute htable_bits from the user input parameter hashsize */
|
/* Compute htable_bits from the user input parameter hashsize */
|
||||||
static u8
|
static u8
|
||||||
htable_bits(u32 hashsize)
|
htable_bits(u32 hashsize)
|
||||||
|
|
|
@ -191,7 +191,8 @@ extern void usbnet_cdc_status(struct usbnet *, struct urb *);
|
||||||
enum skb_state {
|
enum skb_state {
|
||||||
illegal = 0,
|
illegal = 0,
|
||||||
tx_start, tx_done,
|
tx_start, tx_done,
|
||||||
rx_start, rx_done, rx_cleanup
|
rx_start, rx_done, rx_cleanup,
|
||||||
|
unlink_start
|
||||||
};
|
};
|
||||||
|
|
||||||
struct skb_data { /* skb->cb is one of these */
|
struct skb_data { /* skb->cb is one of these */
|
||||||
|
|
|
@ -59,6 +59,7 @@ struct soc_camera_device {
|
||||||
struct soc_camera_host {
|
struct soc_camera_host {
|
||||||
struct v4l2_device v4l2_dev;
|
struct v4l2_device v4l2_dev;
|
||||||
struct list_head list;
|
struct list_head list;
|
||||||
|
struct mutex host_lock; /* Protect during probing */
|
||||||
unsigned char nr; /* Host number */
|
unsigned char nr; /* Host number */
|
||||||
void *priv;
|
void *priv;
|
||||||
const char *drv_name;
|
const char *drv_name;
|
||||||
|
|
|
@ -191,6 +191,7 @@ struct bt_sock {
|
||||||
struct list_head accept_q;
|
struct list_head accept_q;
|
||||||
struct sock *parent;
|
struct sock *parent;
|
||||||
u32 defer_setup;
|
u32 defer_setup;
|
||||||
|
bool suspended;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct bt_sock_list {
|
struct bt_sock_list {
|
||||||
|
|
|
@ -518,6 +518,7 @@ handle_edge_irq(unsigned int irq, struct irq_desc *desc)
|
||||||
out_unlock:
|
out_unlock:
|
||||||
raw_spin_unlock(&desc->lock);
|
raw_spin_unlock(&desc->lock);
|
||||||
}
|
}
|
||||||
|
EXPORT_SYMBOL(handle_edge_irq);
|
||||||
|
|
||||||
#ifdef CONFIG_IRQ_EDGE_EOI_HANDLER
|
#ifdef CONFIG_IRQ_EDGE_EOI_HANDLER
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -112,6 +112,7 @@ struct irq_desc *irq_to_desc(unsigned int irq)
|
||||||
{
|
{
|
||||||
return radix_tree_lookup(&irq_desc_tree, irq);
|
return radix_tree_lookup(&irq_desc_tree, irq);
|
||||||
}
|
}
|
||||||
|
EXPORT_SYMBOL(irq_to_desc);
|
||||||
|
|
||||||
static void delete_irq_desc(unsigned int irq)
|
static void delete_irq_desc(unsigned int irq)
|
||||||
{
|
{
|
||||||
|
|
|
@ -6382,6 +6382,8 @@ static int __sdt_alloc(const struct cpumask *cpu_map)
|
||||||
if (!sg)
|
if (!sg)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
sg->next = sg;
|
||||||
|
|
||||||
*per_cpu_ptr(sdd->sg, j) = sg;
|
*per_cpu_ptr(sdd->sg, j) = sg;
|
||||||
|
|
||||||
sgp = kzalloc_node(sizeof(struct sched_group_power),
|
sgp = kzalloc_node(sizeof(struct sched_group_power),
|
||||||
|
|
|
@ -294,6 +294,9 @@ static int __ftrace_set_clr_event(const char *match, const char *sub,
|
||||||
if (!call->name || !call->class || !call->class->reg)
|
if (!call->name || !call->class || !call->class->reg)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
|
if (call->flags & TRACE_EVENT_FL_IGNORE_ENABLE)
|
||||||
|
continue;
|
||||||
|
|
||||||
if (match &&
|
if (match &&
|
||||||
strcmp(match, call->name) != 0 &&
|
strcmp(match, call->name) != 0 &&
|
||||||
strcmp(match, call->class->system) != 0)
|
strcmp(match, call->class->system) != 0)
|
||||||
|
@ -1164,7 +1167,7 @@ event_create_dir(struct ftrace_event_call *call, struct dentry *d_events,
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (call->class->reg)
|
if (call->class->reg && !(call->flags & TRACE_EVENT_FL_IGNORE_ENABLE))
|
||||||
trace_create_file("enable", 0644, call->dir, call,
|
trace_create_file("enable", 0644, call->dir, call,
|
||||||
enable);
|
enable);
|
||||||
|
|
||||||
|
|
|
@ -180,6 +180,7 @@ struct ftrace_event_call __used event_##call = { \
|
||||||
.event.type = etype, \
|
.event.type = etype, \
|
||||||
.class = &event_class_ftrace_##call, \
|
.class = &event_class_ftrace_##call, \
|
||||||
.print_fmt = print, \
|
.print_fmt = print, \
|
||||||
|
.flags = TRACE_EVENT_FL_IGNORE_ENABLE, \
|
||||||
}; \
|
}; \
|
||||||
struct ftrace_event_call __used \
|
struct ftrace_event_call __used \
|
||||||
__attribute__((section("_ftrace_events"))) *__event_##call = &event_##call;
|
__attribute__((section("_ftrace_events"))) *__event_##call = &event_##call;
|
||||||
|
|
|
@ -5481,7 +5481,7 @@ static int mem_cgroup_move_charge_pte_range(pmd_t *pmd,
|
||||||
* part of thp split is not executed yet.
|
* part of thp split is not executed yet.
|
||||||
*/
|
*/
|
||||||
if (pmd_trans_huge_lock(pmd, vma) == 1) {
|
if (pmd_trans_huge_lock(pmd, vma) == 1) {
|
||||||
if (!mc.precharge) {
|
if (mc.precharge < HPAGE_PMD_NR) {
|
||||||
spin_unlock(&vma->vm_mm->page_table_lock);
|
spin_unlock(&vma->vm_mm->page_table_lock);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue