Merge linux 6.6.39
This commit is contained in:
commit
1ece9da2fd
2
Makefile
2
Makefile
|
@ -8,7 +8,7 @@ else
|
|||
# SPDX-License-Identifier: GPL-2.0
|
||||
VERSION = 6
|
||||
PATCHLEVEL = 6
|
||||
SUBLEVEL = 38
|
||||
SUBLEVEL = 39
|
||||
EXTRAVERSION =
|
||||
NAME = Hurr durr I'ma ninja sloth
|
||||
|
||||
|
|
|
@ -289,7 +289,7 @@
|
|||
regulator-name = "vdd_gpu";
|
||||
regulator-always-on;
|
||||
regulator-boot-on;
|
||||
regulator-min-microvolt = <900000>;
|
||||
regulator-min-microvolt = <500000>;
|
||||
regulator-max-microvolt = <1350000>;
|
||||
regulator-ramp-delay = <6001>;
|
||||
|
||||
|
|
|
@ -336,6 +336,14 @@ static inline void interrupt_nmi_enter_prepare(struct pt_regs *regs, struct inte
|
|||
if (IS_ENABLED(CONFIG_KASAN))
|
||||
return;
|
||||
|
||||
/*
|
||||
* Likewise, do not use it in real mode if percpu first chunk is not
|
||||
* embedded. With CONFIG_NEED_PER_CPU_PAGE_FIRST_CHUNK enabled there
|
||||
* are chances where percpu allocation can come from vmalloc area.
|
||||
*/
|
||||
if (percpu_first_chunk_is_paged)
|
||||
return;
|
||||
|
||||
/* Otherwise, it should be safe to call it */
|
||||
nmi_enter();
|
||||
}
|
||||
|
@ -351,6 +359,8 @@ static inline void interrupt_nmi_exit_prepare(struct pt_regs *regs, struct inter
|
|||
// no nmi_exit for a pseries hash guest taking a real mode exception
|
||||
} else if (IS_ENABLED(CONFIG_KASAN)) {
|
||||
// no nmi_exit for KASAN in real mode
|
||||
} else if (percpu_first_chunk_is_paged) {
|
||||
// no nmi_exit if percpu first chunk is not embedded
|
||||
} else {
|
||||
nmi_exit();
|
||||
}
|
||||
|
|
|
@ -37,7 +37,7 @@ extern struct pci_dev *isa_bridge_pcidev;
|
|||
* define properly based on the platform
|
||||
*/
|
||||
#ifndef CONFIG_PCI
|
||||
#define _IO_BASE 0
|
||||
#define _IO_BASE POISON_POINTER_DELTA
|
||||
#define _ISA_MEM_BASE 0
|
||||
#define PCI_DRAM_OFFSET 0
|
||||
#elif defined(CONFIG_PPC32)
|
||||
|
|
|
@ -15,6 +15,16 @@
|
|||
#endif /* CONFIG_SMP */
|
||||
#endif /* __powerpc64__ */
|
||||
|
||||
#if defined(CONFIG_NEED_PER_CPU_PAGE_FIRST_CHUNK) && defined(CONFIG_SMP)
|
||||
#include <linux/jump_label.h>
|
||||
DECLARE_STATIC_KEY_FALSE(__percpu_first_chunk_is_paged);
|
||||
|
||||
#define percpu_first_chunk_is_paged \
|
||||
(static_key_enabled(&__percpu_first_chunk_is_paged.key))
|
||||
#else
|
||||
#define percpu_first_chunk_is_paged false
|
||||
#endif /* CONFIG_PPC64 && CONFIG_SMP */
|
||||
|
||||
#include <asm-generic/percpu.h>
|
||||
|
||||
#include <asm/paca.h>
|
||||
|
|
|
@ -647,8 +647,9 @@ __after_prom_start:
|
|||
* Note: This process overwrites the OF exception vectors.
|
||||
*/
|
||||
LOAD_REG_IMMEDIATE(r3, PAGE_OFFSET)
|
||||
mr. r4,r26 /* In some cases the loader may */
|
||||
beq 9f /* have already put us at zero */
|
||||
mr r4,r26 /* Load the virtual source address into r4 */
|
||||
cmpld r3,r4 /* Check if source == dest */
|
||||
beq 9f /* If so skip the copy */
|
||||
li r6,0x100 /* Start offset, the first 0x100 */
|
||||
/* bytes were copied earlier. */
|
||||
|
||||
|
|
|
@ -834,6 +834,7 @@ static __init int pcpu_cpu_to_node(int cpu)
|
|||
|
||||
unsigned long __per_cpu_offset[NR_CPUS] __read_mostly;
|
||||
EXPORT_SYMBOL(__per_cpu_offset);
|
||||
DEFINE_STATIC_KEY_FALSE(__percpu_first_chunk_is_paged);
|
||||
|
||||
void __init setup_per_cpu_areas(void)
|
||||
{
|
||||
|
@ -876,6 +877,7 @@ void __init setup_per_cpu_areas(void)
|
|||
if (rc < 0)
|
||||
panic("cannot initialize percpu area (err=%d)", rc);
|
||||
|
||||
static_key_enable(&__percpu_first_chunk_is_paged.key);
|
||||
delta = (unsigned long)pcpu_base_addr - (unsigned long)__per_cpu_start;
|
||||
for_each_possible_cpu(cpu) {
|
||||
__per_cpu_offset[cpu] = delta + pcpu_unit_offsets[cpu];
|
||||
|
|
|
@ -26,6 +26,7 @@
|
|||
#include <asm/paca.h>
|
||||
#include <asm/mmu.h>
|
||||
#include <asm/sections.h> /* _end */
|
||||
#include <asm/setup.h>
|
||||
#include <asm/smp.h>
|
||||
#include <asm/hw_breakpoint.h>
|
||||
#include <asm/svm.h>
|
||||
|
@ -316,6 +317,16 @@ void default_machine_kexec(struct kimage *image)
|
|||
if (!kdump_in_progress())
|
||||
kexec_prepare_cpus();
|
||||
|
||||
#ifdef CONFIG_PPC_PSERIES
|
||||
/*
|
||||
* This must be done after other CPUs have shut down, otherwise they
|
||||
* could execute the 'scv' instruction, which is not supported with
|
||||
* reloc disabled (see configure_exceptions()).
|
||||
*/
|
||||
if (firmware_has_feature(FW_FEATURE_SET_MODE))
|
||||
pseries_disable_reloc_on_exc();
|
||||
#endif
|
||||
|
||||
printk("kexec: Starting switchover sequence.\n");
|
||||
|
||||
/* switch to a staticly allocated stack. Based on irq stack code.
|
||||
|
|
|
@ -61,11 +61,3 @@ void pseries_kexec_cpu_down(int crash_shutdown, int secondary)
|
|||
} else
|
||||
xics_kexec_teardown_cpu(secondary);
|
||||
}
|
||||
|
||||
void pseries_machine_kexec(struct kimage *image)
|
||||
{
|
||||
if (firmware_has_feature(FW_FEATURE_SET_MODE))
|
||||
pseries_disable_reloc_on_exc();
|
||||
|
||||
default_machine_kexec(image);
|
||||
}
|
||||
|
|
|
@ -38,7 +38,6 @@ static inline void smp_init_pseries(void) { }
|
|||
#endif
|
||||
|
||||
extern void pseries_kexec_cpu_down(int crash_shutdown, int secondary);
|
||||
void pseries_machine_kexec(struct kimage *image);
|
||||
|
||||
extern void pSeries_final_fixup(void);
|
||||
|
||||
|
|
|
@ -1153,7 +1153,6 @@ define_machine(pseries) {
|
|||
.machine_check_exception = pSeries_machine_check_exception,
|
||||
.machine_check_log_err = pSeries_machine_check_log_err,
|
||||
#ifdef CONFIG_KEXEC_CORE
|
||||
.machine_kexec = pseries_machine_kexec,
|
||||
.kexec_cpu_down = pseries_kexec_cpu_down,
|
||||
#endif
|
||||
#ifdef CONFIG_MEMORY_HOTPLUG
|
||||
|
|
|
@ -1352,7 +1352,7 @@ static int cpu_cmd(void)
|
|||
}
|
||||
termch = cpu;
|
||||
|
||||
if (!scanhex(&cpu)) {
|
||||
if (!scanhex(&cpu) || cpu >= num_possible_cpus()) {
|
||||
/* print cpus waiting or in xmon */
|
||||
printf("cpus stopped:");
|
||||
last_cpu = first_cpu = NR_CPUS;
|
||||
|
@ -2774,7 +2774,7 @@ static void dump_pacas(void)
|
|||
|
||||
termch = c; /* Put c back, it wasn't 'a' */
|
||||
|
||||
if (scanhex(&num))
|
||||
if (scanhex(&num) && num < num_possible_cpus())
|
||||
dump_one_paca(num);
|
||||
else
|
||||
dump_one_paca(xmon_owner);
|
||||
|
@ -2847,7 +2847,7 @@ static void dump_xives(void)
|
|||
|
||||
termch = c; /* Put c back, it wasn't 'a' */
|
||||
|
||||
if (scanhex(&num))
|
||||
if (scanhex(&num) && num < num_possible_cpus())
|
||||
dump_one_xive(num);
|
||||
else
|
||||
dump_one_xive(xmon_owner);
|
||||
|
|
|
@ -44,11 +44,21 @@ ALTERNATIVE(__stringify(RISCV_PTR do_page_fault), \
|
|||
CONFIG_ERRATA_SIFIVE_CIP_453)
|
||||
#else /* !__ASSEMBLY__ */
|
||||
|
||||
#define ALT_FLUSH_TLB_PAGE(x) \
|
||||
#define ALT_SFENCE_VMA_ASID(asid) \
|
||||
asm(ALTERNATIVE("sfence.vma x0, %0", "sfence.vma", SIFIVE_VENDOR_ID, \
|
||||
ERRATA_SIFIVE_CIP_1200, CONFIG_ERRATA_SIFIVE_CIP_1200) \
|
||||
: : "r" (asid) : "memory")
|
||||
|
||||
#define ALT_SFENCE_VMA_ADDR(addr) \
|
||||
asm(ALTERNATIVE("sfence.vma %0", "sfence.vma", SIFIVE_VENDOR_ID, \
|
||||
ERRATA_SIFIVE_CIP_1200, CONFIG_ERRATA_SIFIVE_CIP_1200) \
|
||||
: : "r" (addr) : "memory")
|
||||
|
||||
#define ALT_SFENCE_VMA_ADDR_ASID(addr, asid) \
|
||||
asm(ALTERNATIVE("sfence.vma %0, %1", "sfence.vma", SIFIVE_VENDOR_ID, \
|
||||
ERRATA_SIFIVE_CIP_1200, CONFIG_ERRATA_SIFIVE_CIP_1200) \
|
||||
: : "r" (addr), "r" (asid) : "memory")
|
||||
|
||||
/*
|
||||
* _val is marked as "will be overwritten", so need to set it to 0
|
||||
* in the default case.
|
||||
|
|
|
@ -22,10 +22,27 @@ static inline void local_flush_tlb_all(void)
|
|||
__asm__ __volatile__ ("sfence.vma" : : : "memory");
|
||||
}
|
||||
|
||||
static inline void local_flush_tlb_all_asid(unsigned long asid)
|
||||
{
|
||||
if (asid != FLUSH_TLB_NO_ASID)
|
||||
ALT_SFENCE_VMA_ASID(asid);
|
||||
else
|
||||
local_flush_tlb_all();
|
||||
}
|
||||
|
||||
/* Flush one page from local TLB */
|
||||
static inline void local_flush_tlb_page(unsigned long addr)
|
||||
{
|
||||
ALT_FLUSH_TLB_PAGE(__asm__ __volatile__ ("sfence.vma %0" : : "r" (addr) : "memory"));
|
||||
ALT_SFENCE_VMA_ADDR(addr);
|
||||
}
|
||||
|
||||
static inline void local_flush_tlb_page_asid(unsigned long addr,
|
||||
unsigned long asid)
|
||||
{
|
||||
if (asid != FLUSH_TLB_NO_ASID)
|
||||
ALT_SFENCE_VMA_ADDR_ASID(addr, asid);
|
||||
else
|
||||
local_flush_tlb_page(addr);
|
||||
}
|
||||
#else /* CONFIG_MMU */
|
||||
#define local_flush_tlb_all() do { } while (0)
|
||||
|
|
|
@ -147,20 +147,12 @@ static void machine_kexec_mask_interrupts(void)
|
|||
|
||||
for_each_irq_desc(i, desc) {
|
||||
struct irq_chip *chip;
|
||||
int ret;
|
||||
|
||||
chip = irq_desc_get_chip(desc);
|
||||
if (!chip)
|
||||
continue;
|
||||
|
||||
/*
|
||||
* First try to remove the active state. If this
|
||||
* fails, try to EOI the interrupt.
|
||||
*/
|
||||
ret = irq_set_irqchip_state(i, IRQCHIP_STATE_ACTIVE, false);
|
||||
|
||||
if (ret && irqd_irq_inprogress(&desc->irq_data) &&
|
||||
chip->irq_eoi)
|
||||
if (chip->irq_eoi && irqd_irq_inprogress(&desc->irq_data))
|
||||
chip->irq_eoi(&desc->irq_data);
|
||||
|
||||
if (chip->irq_mask)
|
||||
|
|
|
@ -39,7 +39,7 @@ static u64 kvm_pmu_get_sample_period(struct kvm_pmc *pmc)
|
|||
u64 sample_period;
|
||||
|
||||
if (!pmc->counter_val)
|
||||
sample_period = counter_val_mask + 1;
|
||||
sample_period = counter_val_mask;
|
||||
else
|
||||
sample_period = (-pmc->counter_val) & counter_val_mask;
|
||||
|
||||
|
|
|
@ -6,29 +6,6 @@
|
|||
#include <asm/sbi.h>
|
||||
#include <asm/mmu_context.h>
|
||||
|
||||
static inline void local_flush_tlb_all_asid(unsigned long asid)
|
||||
{
|
||||
if (asid != FLUSH_TLB_NO_ASID)
|
||||
__asm__ __volatile__ ("sfence.vma x0, %0"
|
||||
:
|
||||
: "r" (asid)
|
||||
: "memory");
|
||||
else
|
||||
local_flush_tlb_all();
|
||||
}
|
||||
|
||||
static inline void local_flush_tlb_page_asid(unsigned long addr,
|
||||
unsigned long asid)
|
||||
{
|
||||
if (asid != FLUSH_TLB_NO_ASID)
|
||||
__asm__ __volatile__ ("sfence.vma %0, %1"
|
||||
:
|
||||
: "r" (addr), "r" (asid)
|
||||
: "memory");
|
||||
else
|
||||
local_flush_tlb_page(addr);
|
||||
}
|
||||
|
||||
/*
|
||||
* Flush entire TLB if number of entries to be flushed is greater
|
||||
* than the threshold below.
|
||||
|
|
|
@ -427,6 +427,7 @@ struct kvm_vcpu_stat {
|
|||
u64 instruction_io_other;
|
||||
u64 instruction_lpsw;
|
||||
u64 instruction_lpswe;
|
||||
u64 instruction_lpswey;
|
||||
u64 instruction_pfmf;
|
||||
u64 instruction_ptff;
|
||||
u64 instruction_sck;
|
||||
|
|
|
@ -132,6 +132,7 @@ const struct _kvm_stats_desc kvm_vcpu_stats_desc[] = {
|
|||
STATS_DESC_COUNTER(VCPU, instruction_io_other),
|
||||
STATS_DESC_COUNTER(VCPU, instruction_lpsw),
|
||||
STATS_DESC_COUNTER(VCPU, instruction_lpswe),
|
||||
STATS_DESC_COUNTER(VCPU, instruction_lpswey),
|
||||
STATS_DESC_COUNTER(VCPU, instruction_pfmf),
|
||||
STATS_DESC_COUNTER(VCPU, instruction_ptff),
|
||||
STATS_DESC_COUNTER(VCPU, instruction_sck),
|
||||
|
|
|
@ -120,6 +120,21 @@ static inline u64 kvm_s390_get_base_disp_s(struct kvm_vcpu *vcpu, u8 *ar)
|
|||
return (base2 ? vcpu->run->s.regs.gprs[base2] : 0) + disp2;
|
||||
}
|
||||
|
||||
static inline u64 kvm_s390_get_base_disp_siy(struct kvm_vcpu *vcpu, u8 *ar)
|
||||
{
|
||||
u32 base1 = vcpu->arch.sie_block->ipb >> 28;
|
||||
s64 disp1;
|
||||
|
||||
/* The displacement is a 20bit _SIGNED_ value */
|
||||
disp1 = sign_extend64(((vcpu->arch.sie_block->ipb & 0x0fff0000) >> 16) +
|
||||
((vcpu->arch.sie_block->ipb & 0xff00) << 4), 19);
|
||||
|
||||
if (ar)
|
||||
*ar = base1;
|
||||
|
||||
return (base1 ? vcpu->run->s.regs.gprs[base1] : 0) + disp1;
|
||||
}
|
||||
|
||||
static inline void kvm_s390_get_base_disp_sse(struct kvm_vcpu *vcpu,
|
||||
u64 *address1, u64 *address2,
|
||||
u8 *ar_b1, u8 *ar_b2)
|
||||
|
|
|
@ -793,6 +793,36 @@ static int handle_lpswe(struct kvm_vcpu *vcpu)
|
|||
return 0;
|
||||
}
|
||||
|
||||
static int handle_lpswey(struct kvm_vcpu *vcpu)
|
||||
{
|
||||
psw_t new_psw;
|
||||
u64 addr;
|
||||
int rc;
|
||||
u8 ar;
|
||||
|
||||
vcpu->stat.instruction_lpswey++;
|
||||
|
||||
if (!test_kvm_facility(vcpu->kvm, 193))
|
||||
return kvm_s390_inject_program_int(vcpu, PGM_OPERATION);
|
||||
|
||||
if (vcpu->arch.sie_block->gpsw.mask & PSW_MASK_PSTATE)
|
||||
return kvm_s390_inject_program_int(vcpu, PGM_PRIVILEGED_OP);
|
||||
|
||||
addr = kvm_s390_get_base_disp_siy(vcpu, &ar);
|
||||
if (addr & 7)
|
||||
return kvm_s390_inject_program_int(vcpu, PGM_SPECIFICATION);
|
||||
|
||||
rc = read_guest(vcpu, addr, ar, &new_psw, sizeof(new_psw));
|
||||
if (rc)
|
||||
return kvm_s390_inject_prog_cond(vcpu, rc);
|
||||
|
||||
vcpu->arch.sie_block->gpsw = new_psw;
|
||||
if (!is_valid_psw(&vcpu->arch.sie_block->gpsw))
|
||||
return kvm_s390_inject_program_int(vcpu, PGM_SPECIFICATION);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int handle_stidp(struct kvm_vcpu *vcpu)
|
||||
{
|
||||
u64 stidp_data = vcpu->kvm->arch.model.cpuid;
|
||||
|
@ -1458,6 +1488,8 @@ int kvm_s390_handle_eb(struct kvm_vcpu *vcpu)
|
|||
case 0x61:
|
||||
case 0x62:
|
||||
return handle_ri(vcpu);
|
||||
case 0x71:
|
||||
return handle_lpswey(vcpu);
|
||||
default:
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
|
|
|
@ -45,8 +45,7 @@ static int setkey_unaligned(struct crypto_aead *tfm, const u8 *key,
|
|||
alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1);
|
||||
memcpy(alignbuffer, key, keylen);
|
||||
ret = crypto_aead_alg(tfm)->setkey(tfm, alignbuffer, keylen);
|
||||
memset(alignbuffer, 0, keylen);
|
||||
kfree(buffer);
|
||||
kfree_sensitive(buffer);
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
|
|
@ -34,8 +34,7 @@ static int setkey_unaligned(struct crypto_cipher *tfm, const u8 *key,
|
|||
alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1);
|
||||
memcpy(alignbuffer, key, keylen);
|
||||
ret = cia->cia_setkey(crypto_cipher_tfm(tfm), alignbuffer, keylen);
|
||||
memset(alignbuffer, 0, keylen);
|
||||
kfree(buffer);
|
||||
kfree_sensitive(buffer);
|
||||
return ret;
|
||||
|
||||
}
|
||||
|
|
|
@ -350,7 +350,8 @@ static const struct regmap_bus *regmap_get_i2c_bus(struct i2c_client *i2c,
|
|||
|
||||
if (quirks->max_write_len &&
|
||||
(bus->max_raw_write == 0 || bus->max_raw_write > quirks->max_write_len))
|
||||
max_write = quirks->max_write_len;
|
||||
max_write = quirks->max_write_len -
|
||||
(config->reg_bits + config->pad_bits) / BITS_PER_BYTE;
|
||||
|
||||
if (max_read || max_write) {
|
||||
ret_bus = kmemdup(bus, sizeof(*bus), GFP_KERNEL);
|
||||
|
|
|
@ -83,6 +83,17 @@ int null_init_zoned_dev(struct nullb_device *dev, struct request_queue *q)
|
|||
return -EINVAL;
|
||||
}
|
||||
|
||||
/*
|
||||
* If a smaller zone capacity was requested, do not allow a smaller last
|
||||
* zone at the same time as such zone configuration does not correspond
|
||||
* to any real zoned device.
|
||||
*/
|
||||
if (dev->zone_capacity != dev->zone_size &&
|
||||
dev->size & (dev->zone_size - 1)) {
|
||||
pr_err("A smaller last zone is not allowed with zone capacity smaller than zone size.\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
zone_capacity_sects = mb_to_sects(dev->zone_capacity);
|
||||
dev_capacity_sects = mb_to_sects(dev->size);
|
||||
dev->zone_size_sects = mb_to_sects(dev->zone_size);
|
||||
|
|
|
@ -716,7 +716,7 @@ static void bcm4377_handle_ack(struct bcm4377_data *bcm4377,
|
|||
ring->events[msgid] = NULL;
|
||||
}
|
||||
|
||||
bitmap_release_region(ring->msgids, msgid, ring->n_entries);
|
||||
bitmap_release_region(ring->msgids, msgid, 0);
|
||||
|
||||
unlock:
|
||||
spin_unlock_irqrestore(&ring->lock, flags);
|
||||
|
|
|
@ -2428,15 +2428,27 @@ static void qca_serdev_shutdown(struct device *dev)
|
|||
struct qca_serdev *qcadev = serdev_device_get_drvdata(serdev);
|
||||
struct hci_uart *hu = &qcadev->serdev_hu;
|
||||
struct hci_dev *hdev = hu->hdev;
|
||||
struct qca_data *qca = hu->priv;
|
||||
const u8 ibs_wake_cmd[] = { 0xFD };
|
||||
const u8 edl_reset_soc_cmd[] = { 0x01, 0x00, 0xFC, 0x01, 0x05 };
|
||||
|
||||
if (qcadev->btsoc_type == QCA_QCA6390) {
|
||||
if (test_bit(QCA_BT_OFF, &qca->flags) ||
|
||||
!test_bit(HCI_RUNNING, &hdev->flags))
|
||||
/* The purpose of sending the VSC is to reset SOC into a initial
|
||||
* state and the state will ensure next hdev->setup() success.
|
||||
* if HCI_QUIRK_NON_PERSISTENT_SETUP is set, it means that
|
||||
* hdev->setup() can do its job regardless of SoC state, so
|
||||
* don't need to send the VSC.
|
||||
* if HCI_SETUP is set, it means that hdev->setup() was never
|
||||
* invoked and the SOC is already in the initial state, so
|
||||
* don't also need to send the VSC.
|
||||
*/
|
||||
if (test_bit(HCI_QUIRK_NON_PERSISTENT_SETUP, &hdev->quirks) ||
|
||||
hci_dev_test_flag(hdev, HCI_SETUP))
|
||||
return;
|
||||
|
||||
/* The serdev must be in open state when conrol logic arrives
|
||||
* here, so also fix the use-after-free issue caused by that
|
||||
* the serdev is flushed or wrote after it is closed.
|
||||
*/
|
||||
serdev_device_write_flush(serdev);
|
||||
ret = serdev_device_write_buf(serdev, ibs_wake_cmd,
|
||||
sizeof(ibs_wake_cmd));
|
||||
|
|
|
@ -2358,7 +2358,7 @@ static int cdrom_ioctl_timed_media_change(struct cdrom_device_info *cdi,
|
|||
return -EFAULT;
|
||||
|
||||
tmp_info.media_flags = 0;
|
||||
if (tmp_info.last_media_change - cdi->last_media_change_ms < 0)
|
||||
if (cdi->last_media_change_ms > tmp_info.last_media_change)
|
||||
tmp_info.media_flags |= MEDIA_CHANGED_FLAG;
|
||||
|
||||
tmp_info.last_media_change = cdi->last_media_change_ms;
|
||||
|
|
|
@ -29,6 +29,7 @@ static const struct mtk_gate mfg_clks[] = {
|
|||
static const struct mtk_clk_desc mfg_desc = {
|
||||
.clks = mfg_clks,
|
||||
.num_clks = ARRAY_SIZE(mfg_clks),
|
||||
.need_runtime_pm = true,
|
||||
};
|
||||
|
||||
static const struct of_device_id of_match_clk_mt8183_mfg[] = {
|
||||
|
|
|
@ -496,14 +496,16 @@ static int __mtk_clk_simple_probe(struct platform_device *pdev,
|
|||
}
|
||||
|
||||
|
||||
devm_pm_runtime_enable(&pdev->dev);
|
||||
/*
|
||||
* Do a pm_runtime_resume_and_get() to workaround a possible
|
||||
* deadlock between clk_register() and the genpd framework.
|
||||
*/
|
||||
r = pm_runtime_resume_and_get(&pdev->dev);
|
||||
if (r)
|
||||
return r;
|
||||
if (mcd->need_runtime_pm) {
|
||||
devm_pm_runtime_enable(&pdev->dev);
|
||||
/*
|
||||
* Do a pm_runtime_resume_and_get() to workaround a possible
|
||||
* deadlock between clk_register() and the genpd framework.
|
||||
*/
|
||||
r = pm_runtime_resume_and_get(&pdev->dev);
|
||||
if (r)
|
||||
return r;
|
||||
}
|
||||
|
||||
/* Calculate how many clk_hw_onecell_data entries to allocate */
|
||||
num_clks = mcd->num_clks + mcd->num_composite_clks;
|
||||
|
@ -585,7 +587,8 @@ static int __mtk_clk_simple_probe(struct platform_device *pdev,
|
|||
goto unregister_clks;
|
||||
}
|
||||
|
||||
pm_runtime_put(&pdev->dev);
|
||||
if (mcd->need_runtime_pm)
|
||||
pm_runtime_put(&pdev->dev);
|
||||
|
||||
return r;
|
||||
|
||||
|
@ -618,7 +621,8 @@ free_base:
|
|||
if (mcd->shared_io && base)
|
||||
iounmap(base);
|
||||
|
||||
pm_runtime_put(&pdev->dev);
|
||||
if (mcd->need_runtime_pm)
|
||||
pm_runtime_put(&pdev->dev);
|
||||
return r;
|
||||
}
|
||||
|
||||
|
|
|
@ -237,6 +237,8 @@ struct mtk_clk_desc {
|
|||
|
||||
int (*clk_notifier_func)(struct device *dev, struct clk *clk);
|
||||
unsigned int mfg_clk_idx;
|
||||
|
||||
bool need_runtime_pm;
|
||||
};
|
||||
|
||||
int mtk_clk_pdev_probe(struct platform_device *pdev);
|
||||
|
|
|
@ -2510,6 +2510,9 @@ static int clk_alpha_pll_stromer_plus_set_rate(struct clk_hw *hw,
|
|||
regmap_write(pll->clkr.regmap, PLL_ALPHA_VAL_U(pll),
|
||||
a >> ALPHA_BITWIDTH);
|
||||
|
||||
regmap_update_bits(pll->clkr.regmap, PLL_USER_CTL(pll),
|
||||
PLL_ALPHA_EN, PLL_ALPHA_EN);
|
||||
|
||||
regmap_write(pll->clkr.regmap, PLL_MODE(pll), PLL_BYPASSNL);
|
||||
|
||||
/* Wait five micro seconds or more */
|
||||
|
|
|
@ -2140,9 +2140,10 @@ static struct clk_rcg2 pcnoc_bfdcd_clk_src = {
|
|||
|
||||
static struct clk_branch gcc_crypto_axi_clk = {
|
||||
.halt_reg = 0x16010,
|
||||
.halt_check = BRANCH_HALT_VOTED,
|
||||
.clkr = {
|
||||
.enable_reg = 0x16010,
|
||||
.enable_mask = BIT(0),
|
||||
.enable_reg = 0xb004,
|
||||
.enable_mask = BIT(15),
|
||||
.hw.init = &(const struct clk_init_data) {
|
||||
.name = "gcc_crypto_axi_clk",
|
||||
.parent_hws = (const struct clk_hw *[]) {
|
||||
|
@ -2156,9 +2157,10 @@ static struct clk_branch gcc_crypto_axi_clk = {
|
|||
|
||||
static struct clk_branch gcc_crypto_ahb_clk = {
|
||||
.halt_reg = 0x16014,
|
||||
.halt_check = BRANCH_HALT_VOTED,
|
||||
.clkr = {
|
||||
.enable_reg = 0x16014,
|
||||
.enable_mask = BIT(0),
|
||||
.enable_reg = 0xb004,
|
||||
.enable_mask = BIT(16),
|
||||
.hw.init = &(const struct clk_init_data) {
|
||||
.name = "gcc_crypto_ahb_clk",
|
||||
.parent_hws = (const struct clk_hw *[]) {
|
||||
|
|
|
@ -100,8 +100,8 @@ static struct clk_alpha_pll gpll6 = {
|
|||
.enable_mask = BIT(6),
|
||||
.hw.init = &(struct clk_init_data){
|
||||
.name = "gpll6",
|
||||
.parent_hws = (const struct clk_hw*[]){
|
||||
&gpll0.clkr.hw,
|
||||
.parent_data = &(const struct clk_parent_data){
|
||||
.fw_name = "bi_tcxo",
|
||||
},
|
||||
.num_parents = 1,
|
||||
.ops = &clk_alpha_pll_fixed_fabia_ops,
|
||||
|
@ -124,7 +124,7 @@ static struct clk_alpha_pll_postdiv gpll6_out_even = {
|
|||
.clkr.hw.init = &(struct clk_init_data){
|
||||
.name = "gpll6_out_even",
|
||||
.parent_hws = (const struct clk_hw*[]){
|
||||
&gpll0.clkr.hw,
|
||||
&gpll6.clkr.hw,
|
||||
},
|
||||
.num_parents = 1,
|
||||
.ops = &clk_alpha_pll_postdiv_fabia_ops,
|
||||
|
@ -139,8 +139,8 @@ static struct clk_alpha_pll gpll7 = {
|
|||
.enable_mask = BIT(7),
|
||||
.hw.init = &(struct clk_init_data){
|
||||
.name = "gpll7",
|
||||
.parent_hws = (const struct clk_hw*[]){
|
||||
&gpll0.clkr.hw,
|
||||
.parent_data = &(const struct clk_parent_data){
|
||||
.fw_name = "bi_tcxo",
|
||||
},
|
||||
.num_parents = 1,
|
||||
.ops = &clk_alpha_pll_fixed_fabia_ops,
|
||||
|
|
|
@ -132,7 +132,6 @@ static int sunxi_ccu_probe(struct sunxi_ccu *ccu, struct device *dev,
|
|||
|
||||
for (i = 0; i < desc->hw_clks->num ; i++) {
|
||||
struct clk_hw *hw = desc->hw_clks->hws[i];
|
||||
struct ccu_common *common = hw_to_ccu_common(hw);
|
||||
const char *name;
|
||||
|
||||
if (!hw)
|
||||
|
@ -147,14 +146,21 @@ static int sunxi_ccu_probe(struct sunxi_ccu *ccu, struct device *dev,
|
|||
pr_err("Couldn't register clock %d - %s\n", i, name);
|
||||
goto err_clk_unreg;
|
||||
}
|
||||
}
|
||||
|
||||
if (common->max_rate)
|
||||
clk_hw_set_rate_range(hw, common->min_rate,
|
||||
common->max_rate);
|
||||
for (i = 0; i < desc->num_ccu_clks; i++) {
|
||||
struct ccu_common *cclk = desc->ccu_clks[i];
|
||||
|
||||
if (!cclk)
|
||||
continue;
|
||||
|
||||
if (cclk->max_rate)
|
||||
clk_hw_set_rate_range(&cclk->hw, cclk->min_rate,
|
||||
cclk->max_rate);
|
||||
else
|
||||
WARN(common->min_rate,
|
||||
WARN(cclk->min_rate,
|
||||
"No max_rate, ignoring min_rate of clock %d - %s\n",
|
||||
i, name);
|
||||
i, clk_hw_get_name(&cclk->hw));
|
||||
}
|
||||
|
||||
ret = of_clk_add_hw_provider(node, of_clk_hw_onecell_get,
|
||||
|
|
|
@ -794,8 +794,14 @@ static void dfx_regs_uninit(struct hisi_qm *qm,
|
|||
{
|
||||
int i;
|
||||
|
||||
if (!dregs)
|
||||
return;
|
||||
|
||||
/* Setting the pointer is NULL to prevent double free */
|
||||
for (i = 0; i < reg_len; i++) {
|
||||
if (!dregs[i].regs)
|
||||
continue;
|
||||
|
||||
kfree(dregs[i].regs);
|
||||
dregs[i].regs = NULL;
|
||||
}
|
||||
|
@ -845,14 +851,21 @@ alloc_error:
|
|||
static int qm_diff_regs_init(struct hisi_qm *qm,
|
||||
struct dfx_diff_registers *dregs, u32 reg_len)
|
||||
{
|
||||
int ret;
|
||||
|
||||
qm->debug.qm_diff_regs = dfx_regs_init(qm, qm_diff_regs, ARRAY_SIZE(qm_diff_regs));
|
||||
if (IS_ERR(qm->debug.qm_diff_regs))
|
||||
return PTR_ERR(qm->debug.qm_diff_regs);
|
||||
if (IS_ERR(qm->debug.qm_diff_regs)) {
|
||||
ret = PTR_ERR(qm->debug.qm_diff_regs);
|
||||
qm->debug.qm_diff_regs = NULL;
|
||||
return ret;
|
||||
}
|
||||
|
||||
qm->debug.acc_diff_regs = dfx_regs_init(qm, dregs, reg_len);
|
||||
if (IS_ERR(qm->debug.acc_diff_regs)) {
|
||||
dfx_regs_uninit(qm, qm->debug.qm_diff_regs, ARRAY_SIZE(qm_diff_regs));
|
||||
return PTR_ERR(qm->debug.acc_diff_regs);
|
||||
ret = PTR_ERR(qm->debug.acc_diff_regs);
|
||||
qm->debug.acc_diff_regs = NULL;
|
||||
return ret;
|
||||
}
|
||||
|
||||
return 0;
|
||||
|
@ -893,7 +906,9 @@ static int qm_last_regs_init(struct hisi_qm *qm)
|
|||
static void qm_diff_regs_uninit(struct hisi_qm *qm, u32 reg_len)
|
||||
{
|
||||
dfx_regs_uninit(qm, qm->debug.acc_diff_regs, reg_len);
|
||||
qm->debug.acc_diff_regs = NULL;
|
||||
dfx_regs_uninit(qm, qm->debug.qm_diff_regs, ARRAY_SIZE(qm_diff_regs));
|
||||
qm->debug.qm_diff_regs = NULL;
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -101,6 +101,17 @@ static void dmi_decode_table(u8 *buf,
|
|||
(data - buf + sizeof(struct dmi_header)) <= dmi_len) {
|
||||
const struct dmi_header *dm = (const struct dmi_header *)data;
|
||||
|
||||
/*
|
||||
* If a short entry is found (less than 4 bytes), not only it
|
||||
* is invalid, but we cannot reliably locate the next entry.
|
||||
*/
|
||||
if (dm->length < sizeof(struct dmi_header)) {
|
||||
pr_warn(FW_BUG
|
||||
"Corrupted DMI table, offset %zd (only %d entries processed)\n",
|
||||
data - buf, i);
|
||||
break;
|
||||
}
|
||||
|
||||
/*
|
||||
* We want to know the total length (formatted area and
|
||||
* strings) before decoding to make sure we won't run off the
|
||||
|
|
|
@ -622,8 +622,6 @@ int bgpio_init(struct gpio_chip *gc, struct device *dev,
|
|||
ret = gpiochip_get_ngpios(gc, dev);
|
||||
if (ret)
|
||||
gc->ngpio = gc->bgpio_bits;
|
||||
else
|
||||
gc->bgpio_bits = roundup_pow_of_two(round_up(gc->ngpio, 8));
|
||||
|
||||
ret = bgpio_setup_io(gc, dat, set, clr, flags);
|
||||
if (ret)
|
||||
|
|
|
@ -192,6 +192,24 @@ static void of_gpio_try_fixup_polarity(const struct device_node *np,
|
|||
*/
|
||||
{ "himax,hx8357", "gpios-reset", false },
|
||||
{ "himax,hx8369", "gpios-reset", false },
|
||||
#endif
|
||||
#if IS_ENABLED(CONFIG_PCI_LANTIQ)
|
||||
/*
|
||||
* According to the PCI specification, the RST# pin is an
|
||||
* active-low signal. However, most of the device trees that
|
||||
* have been widely used for a long time incorrectly describe
|
||||
* reset GPIO as active-high, and were also using wrong name
|
||||
* for the property.
|
||||
*/
|
||||
{ "lantiq,pci-xway", "gpio-reset", false },
|
||||
#endif
|
||||
#if IS_ENABLED(CONFIG_TOUCHSCREEN_TSC2005)
|
||||
/*
|
||||
* DTS for Nokia N900 incorrectly specified "active high"
|
||||
* polarity for the reset line, while the chip actually
|
||||
* treats it as "active low".
|
||||
*/
|
||||
{ "ti,tsc2005", "reset-gpios", false },
|
||||
#endif
|
||||
};
|
||||
unsigned int i;
|
||||
|
@ -491,9 +509,9 @@ static struct gpio_desc *of_find_gpio_rename(struct device_node *np,
|
|||
{ "reset", "reset-n-io", "marvell,nfc-uart" },
|
||||
{ "reset", "reset-n-io", "mrvl,nfc-uart" },
|
||||
#endif
|
||||
#if !IS_ENABLED(CONFIG_PCI_LANTIQ)
|
||||
#if IS_ENABLED(CONFIG_PCI_LANTIQ)
|
||||
/* MIPS Lantiq PCI */
|
||||
{ "reset", "gpios-reset", "lantiq,pci-xway" },
|
||||
{ "reset", "gpio-reset", "lantiq,pci-xway" },
|
||||
#endif
|
||||
|
||||
/*
|
||||
|
|
|
@ -100,7 +100,7 @@ static int aldebaran_mode2_suspend_ip(struct amdgpu_device *adev)
|
|||
adev->ip_blocks[i].status.hw = false;
|
||||
}
|
||||
|
||||
return r;
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
|
|
|
@ -2052,12 +2052,13 @@ static ssize_t amdgpu_reset_dump_register_list_write(struct file *f,
|
|||
struct amdgpu_device *adev = (struct amdgpu_device *)file_inode(f)->i_private;
|
||||
char reg_offset[11];
|
||||
uint32_t *new = NULL, *tmp = NULL;
|
||||
int ret, i = 0, len = 0;
|
||||
unsigned int len = 0;
|
||||
int ret, i = 0;
|
||||
|
||||
do {
|
||||
memset(reg_offset, 0, 11);
|
||||
if (copy_from_user(reg_offset, buf + len,
|
||||
min(10, ((int)size-len)))) {
|
||||
min(10, (size-len)))) {
|
||||
ret = -EFAULT;
|
||||
goto error_free;
|
||||
}
|
||||
|
|
|
@ -1184,7 +1184,8 @@ void amdgpu_gfx_cp_init_microcode(struct amdgpu_device *adev,
|
|||
fw_size = le32_to_cpu(cp_hdr_v2_0->data_size_bytes);
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
dev_err(adev->dev, "Invalid ucode id %u\n", ucode_id);
|
||||
return;
|
||||
}
|
||||
|
||||
if (adev->firmware.load_type == AMDGPU_FW_LOAD_PSP) {
|
||||
|
|
|
@ -438,6 +438,14 @@ void amdgpu_irq_dispatch(struct amdgpu_device *adev,
|
|||
|
||||
entry.ih = ih;
|
||||
entry.iv_entry = (const uint32_t *)&ih->ring[ring_index];
|
||||
|
||||
/*
|
||||
* timestamp is not supported on some legacy SOCs (cik, cz, iceland,
|
||||
* si and tonga), so initialize timestamp and timestamp_src to 0
|
||||
*/
|
||||
entry.timestamp = 0;
|
||||
entry.timestamp_src = 0;
|
||||
|
||||
amdgpu_ih_decode_iv(adev, &entry);
|
||||
|
||||
trace_amdgpu_iv(ih - &adev->irq.ih, &entry);
|
||||
|
|
|
@ -742,7 +742,8 @@ int amdgpu_vce_ring_parse_cs(struct amdgpu_cs_parser *p,
|
|||
uint32_t created = 0;
|
||||
uint32_t allocated = 0;
|
||||
uint32_t tmp, handle = 0;
|
||||
uint32_t *size = &tmp;
|
||||
uint32_t dummy = 0xffffffff;
|
||||
uint32_t *size = &dummy;
|
||||
unsigned int idx;
|
||||
int i, r = 0;
|
||||
|
||||
|
|
|
@ -93,7 +93,7 @@ static int sienna_cichlid_mode2_suspend_ip(struct amdgpu_device *adev)
|
|||
adev->ip_blocks[i].status.hw = false;
|
||||
}
|
||||
|
||||
return r;
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
|
|
|
@ -264,7 +264,7 @@ static u32 dm_vblank_get_counter(struct amdgpu_device *adev, int crtc)
|
|||
static int dm_crtc_get_scanoutpos(struct amdgpu_device *adev, int crtc,
|
||||
u32 *vbl, u32 *position)
|
||||
{
|
||||
u32 v_blank_start, v_blank_end, h_position, v_position;
|
||||
u32 v_blank_start = 0, v_blank_end = 0, h_position = 0, v_position = 0;
|
||||
struct amdgpu_crtc *acrtc = NULL;
|
||||
|
||||
if ((crtc < 0) || (crtc >= adev->mode_info.num_crtc))
|
||||
|
@ -801,7 +801,7 @@ static void dm_handle_hpd_work(struct work_struct *work)
|
|||
*/
|
||||
static void dm_dmub_outbox1_low_irq(void *interrupt_params)
|
||||
{
|
||||
struct dmub_notification notify;
|
||||
struct dmub_notification notify = {0};
|
||||
struct common_irq_params *irq_params = interrupt_params;
|
||||
struct amdgpu_device *adev = irq_params->adev;
|
||||
struct amdgpu_display_manager *dm = &adev->dm;
|
||||
|
@ -6895,7 +6895,7 @@ static int dm_update_mst_vcpi_slots_for_dsc(struct drm_atomic_state *state,
|
|||
struct amdgpu_dm_connector *aconnector;
|
||||
struct dm_connector_state *dm_conn_state;
|
||||
int i, j, ret;
|
||||
int vcpi, pbn_div, pbn, slot_num = 0;
|
||||
int vcpi, pbn_div, pbn = 0, slot_num = 0;
|
||||
|
||||
for_each_new_connector_in_state(state, connector, new_con_state, i) {
|
||||
|
||||
|
@ -10064,7 +10064,7 @@ static int amdgpu_dm_atomic_check(struct drm_device *dev,
|
|||
struct dm_crtc_state *dm_old_crtc_state, *dm_new_crtc_state;
|
||||
struct drm_dp_mst_topology_mgr *mgr;
|
||||
struct drm_dp_mst_topology_state *mst_state;
|
||||
struct dsc_mst_fairness_vars vars[MAX_PIPES];
|
||||
struct dsc_mst_fairness_vars vars[MAX_PIPES] = {0};
|
||||
|
||||
trace_amdgpu_dm_atomic_check_begin(state);
|
||||
|
||||
|
|
|
@ -1219,7 +1219,7 @@ static ssize_t dp_sdp_message_debugfs_write(struct file *f, const char __user *b
|
|||
size_t size, loff_t *pos)
|
||||
{
|
||||
int r;
|
||||
uint8_t data[36];
|
||||
uint8_t data[36] = {0};
|
||||
struct amdgpu_dm_connector *connector = file_inode(f)->i_private;
|
||||
struct dm_crtc_state *acrtc_state;
|
||||
uint32_t write_size = 36;
|
||||
|
@ -2929,7 +2929,7 @@ static int psr_read_residency(void *data, u64 *val)
|
|||
{
|
||||
struct amdgpu_dm_connector *connector = data;
|
||||
struct dc_link *link = connector->dc_link;
|
||||
u32 residency;
|
||||
u32 residency = 0;
|
||||
|
||||
link->dc->link_srv->edp_get_psr_residency(link, &residency);
|
||||
|
||||
|
|
|
@ -2385,6 +2385,9 @@ static struct audio *find_first_free_audio(
|
|||
{
|
||||
int i, available_audio_count;
|
||||
|
||||
if (id == ENGINE_ID_UNKNOWN)
|
||||
return NULL;
|
||||
|
||||
available_audio_count = pool->audio_count;
|
||||
|
||||
for (i = 0; i < available_audio_count; i++) {
|
||||
|
|
|
@ -211,8 +211,12 @@ bool dce110_vblank_set(struct irq_service *irq_service,
|
|||
info->ext_id);
|
||||
uint8_t pipe_offset = dal_irq_src - IRQ_TYPE_VBLANK;
|
||||
|
||||
struct timing_generator *tg =
|
||||
dc->current_state->res_ctx.pipe_ctx[pipe_offset].stream_res.tg;
|
||||
struct timing_generator *tg;
|
||||
|
||||
if (pipe_offset >= MAX_PIPES)
|
||||
return false;
|
||||
|
||||
tg = dc->current_state->res_ctx.pipe_ctx[pipe_offset].stream_res.tg;
|
||||
|
||||
if (enable) {
|
||||
if (!tg || !tg->funcs->arm_vert_intr(tg, 2)) {
|
||||
|
|
|
@ -156,6 +156,10 @@ static enum mod_hdcp_status read(struct mod_hdcp *hdcp,
|
|||
uint32_t cur_size = 0;
|
||||
uint32_t data_offset = 0;
|
||||
|
||||
if (msg_id == MOD_HDCP_MESSAGE_ID_INVALID) {
|
||||
return MOD_HDCP_STATUS_DDC_FAILURE;
|
||||
}
|
||||
|
||||
if (is_dp_hdcp(hdcp)) {
|
||||
while (buf_len > 0) {
|
||||
cur_size = MIN(buf_len, HDCP_MAX_AUX_TRANSACTION_SIZE);
|
||||
|
@ -215,6 +219,10 @@ static enum mod_hdcp_status write(struct mod_hdcp *hdcp,
|
|||
uint32_t cur_size = 0;
|
||||
uint32_t data_offset = 0;
|
||||
|
||||
if (msg_id == MOD_HDCP_MESSAGE_ID_INVALID) {
|
||||
return MOD_HDCP_STATUS_DDC_FAILURE;
|
||||
}
|
||||
|
||||
if (is_dp_hdcp(hdcp)) {
|
||||
while (buf_len > 0) {
|
||||
cur_size = MIN(buf_len, HDCP_MAX_AUX_TRANSACTION_SIZE);
|
||||
|
|
|
@ -702,7 +702,7 @@ struct atom_gpio_pin_lut_v2_1
|
|||
{
|
||||
struct atom_common_table_header table_header;
|
||||
/*the real number of this included in the structure is calcualted by using the (whole structure size - the header size)/size of atom_gpio_pin_lut */
|
||||
struct atom_gpio_pin_assignment gpio_pin[8];
|
||||
struct atom_gpio_pin_assignment gpio_pin[];
|
||||
};
|
||||
|
||||
|
||||
|
@ -3551,7 +3551,7 @@ struct atom_gpio_voltage_object_v4
|
|||
uint8_t phase_delay_us; // phase delay in unit of micro second
|
||||
uint8_t reserved;
|
||||
uint32_t gpio_mask_val; // GPIO Mask value
|
||||
struct atom_voltage_gpio_map_lut voltage_gpio_lut[1];
|
||||
struct atom_voltage_gpio_map_lut voltage_gpio_lut[] __counted_by(gpio_entry_num);
|
||||
};
|
||||
|
||||
struct atom_svid2_voltage_object_v4
|
||||
|
|
|
@ -421,6 +421,13 @@ static const struct dmi_system_id orientation_data[] = {
|
|||
DMI_EXACT_MATCH(DMI_PRODUCT_VERSION, "1"),
|
||||
},
|
||||
.driver_data = (void *)&lcd800x1280_rightside_up,
|
||||
}, { /* Valve Steam Deck */
|
||||
.matches = {
|
||||
DMI_EXACT_MATCH(DMI_SYS_VENDOR, "Valve"),
|
||||
DMI_EXACT_MATCH(DMI_PRODUCT_NAME, "Galileo"),
|
||||
DMI_EXACT_MATCH(DMI_PRODUCT_VERSION, "1"),
|
||||
},
|
||||
.driver_data = (void *)&lcd800x1280_rightside_up,
|
||||
}, { /* VIOS LTH17 */
|
||||
.matches = {
|
||||
DMI_EXACT_MATCH(DMI_SYS_VENDOR, "VIOS"),
|
||||
|
|
|
@ -324,7 +324,9 @@ int lima_gp_init(struct lima_ip *ip)
|
|||
|
||||
void lima_gp_fini(struct lima_ip *ip)
|
||||
{
|
||||
struct lima_device *dev = ip->dev;
|
||||
|
||||
devm_free_irq(dev->dev, ip->irq, ip);
|
||||
}
|
||||
|
||||
int lima_gp_pipe_init(struct lima_device *dev)
|
||||
|
|
|
@ -118,7 +118,12 @@ int lima_mmu_init(struct lima_ip *ip)
|
|||
|
||||
void lima_mmu_fini(struct lima_ip *ip)
|
||||
{
|
||||
struct lima_device *dev = ip->dev;
|
||||
|
||||
if (ip->id == lima_ip_ppmmu_bcast)
|
||||
return;
|
||||
|
||||
devm_free_irq(dev->dev, ip->irq, ip);
|
||||
}
|
||||
|
||||
void lima_mmu_flush_tlb(struct lima_ip *ip)
|
||||
|
|
|
@ -266,7 +266,9 @@ int lima_pp_init(struct lima_ip *ip)
|
|||
|
||||
void lima_pp_fini(struct lima_ip *ip)
|
||||
{
|
||||
struct lima_device *dev = ip->dev;
|
||||
|
||||
devm_free_irq(dev->dev, ip->irq, ip);
|
||||
}
|
||||
|
||||
int lima_pp_bcast_resume(struct lima_ip *ip)
|
||||
|
@ -299,7 +301,9 @@ int lima_pp_bcast_init(struct lima_ip *ip)
|
|||
|
||||
void lima_pp_bcast_fini(struct lima_ip *ip)
|
||||
{
|
||||
struct lima_device *dev = ip->dev;
|
||||
|
||||
devm_free_irq(dev->dev, ip->irq, ip);
|
||||
}
|
||||
|
||||
static int lima_pp_task_validate(struct lima_sched_pipe *pipe,
|
||||
|
|
|
@ -983,6 +983,9 @@ nouveau_connector_get_modes(struct drm_connector *connector)
|
|||
struct drm_display_mode *mode;
|
||||
|
||||
mode = drm_mode_duplicate(dev, nv_connector->native_mode);
|
||||
if (!mode)
|
||||
return 0;
|
||||
|
||||
drm_mode_probed_add(connector, mode);
|
||||
ret = 1;
|
||||
}
|
||||
|
|
|
@ -1052,7 +1052,7 @@ static const struct pci_device_id i801_ids[] = {
|
|||
MODULE_DEVICE_TABLE(pci, i801_ids);
|
||||
|
||||
#if defined CONFIG_X86 && defined CONFIG_DMI
|
||||
static unsigned char apanel_addr;
|
||||
static unsigned char apanel_addr __ro_after_init;
|
||||
|
||||
/* Scan the system ROM for the signature "FJKEYINF" */
|
||||
static __init const void __iomem *bios_signature(const void __iomem *bios)
|
||||
|
|
|
@ -15,7 +15,6 @@
|
|||
#include <linux/ioport.h>
|
||||
#include <linux/delay.h>
|
||||
#include <linux/i2c.h>
|
||||
#include <linux/timer.h>
|
||||
#include <linux/completion.h>
|
||||
#include <linux/platform_device.h>
|
||||
#include <linux/io.h>
|
||||
|
@ -32,7 +31,6 @@ struct i2c_pnx_mif {
|
|||
int ret; /* Return value */
|
||||
int mode; /* Interface mode */
|
||||
struct completion complete; /* I/O completion */
|
||||
struct timer_list timer; /* Timeout */
|
||||
u8 * buf; /* Data buffer */
|
||||
int len; /* Length of data buffer */
|
||||
int order; /* RX Bytes to order via TX */
|
||||
|
@ -117,24 +115,6 @@ static inline int wait_reset(struct i2c_pnx_algo_data *data)
|
|||
return (timeout <= 0);
|
||||
}
|
||||
|
||||
static inline void i2c_pnx_arm_timer(struct i2c_pnx_algo_data *alg_data)
|
||||
{
|
||||
struct timer_list *timer = &alg_data->mif.timer;
|
||||
unsigned long expires = msecs_to_jiffies(alg_data->timeout);
|
||||
|
||||
if (expires <= 1)
|
||||
expires = 2;
|
||||
|
||||
del_timer_sync(timer);
|
||||
|
||||
dev_dbg(&alg_data->adapter.dev, "Timer armed at %lu plus %lu jiffies.\n",
|
||||
jiffies, expires);
|
||||
|
||||
timer->expires = jiffies + expires;
|
||||
|
||||
add_timer(timer);
|
||||
}
|
||||
|
||||
/**
|
||||
* i2c_pnx_start - start a device
|
||||
* @slave_addr: slave address
|
||||
|
@ -259,8 +239,6 @@ static int i2c_pnx_master_xmit(struct i2c_pnx_algo_data *alg_data)
|
|||
~(mcntrl_afie | mcntrl_naie | mcntrl_drmie),
|
||||
I2C_REG_CTL(alg_data));
|
||||
|
||||
del_timer_sync(&alg_data->mif.timer);
|
||||
|
||||
dev_dbg(&alg_data->adapter.dev,
|
||||
"%s(): Waking up xfer routine.\n",
|
||||
__func__);
|
||||
|
@ -276,8 +254,6 @@ static int i2c_pnx_master_xmit(struct i2c_pnx_algo_data *alg_data)
|
|||
~(mcntrl_afie | mcntrl_naie | mcntrl_drmie),
|
||||
I2C_REG_CTL(alg_data));
|
||||
|
||||
/* Stop timer. */
|
||||
del_timer_sync(&alg_data->mif.timer);
|
||||
dev_dbg(&alg_data->adapter.dev,
|
||||
"%s(): Waking up xfer routine after zero-xfer.\n",
|
||||
__func__);
|
||||
|
@ -364,8 +340,6 @@ static int i2c_pnx_master_rcv(struct i2c_pnx_algo_data *alg_data)
|
|||
mcntrl_drmie | mcntrl_daie);
|
||||
iowrite32(ctl, I2C_REG_CTL(alg_data));
|
||||
|
||||
/* Kill timer. */
|
||||
del_timer_sync(&alg_data->mif.timer);
|
||||
complete(&alg_data->mif.complete);
|
||||
}
|
||||
}
|
||||
|
@ -400,8 +374,6 @@ static irqreturn_t i2c_pnx_interrupt(int irq, void *dev_id)
|
|||
mcntrl_drmie);
|
||||
iowrite32(ctl, I2C_REG_CTL(alg_data));
|
||||
|
||||
/* Stop timer, to prevent timeout. */
|
||||
del_timer_sync(&alg_data->mif.timer);
|
||||
complete(&alg_data->mif.complete);
|
||||
} else if (stat & mstatus_nai) {
|
||||
/* Slave did not acknowledge, generate a STOP */
|
||||
|
@ -419,8 +391,6 @@ static irqreturn_t i2c_pnx_interrupt(int irq, void *dev_id)
|
|||
/* Our return value. */
|
||||
alg_data->mif.ret = -EIO;
|
||||
|
||||
/* Stop timer, to prevent timeout. */
|
||||
del_timer_sync(&alg_data->mif.timer);
|
||||
complete(&alg_data->mif.complete);
|
||||
} else {
|
||||
/*
|
||||
|
@ -453,9 +423,8 @@ static irqreturn_t i2c_pnx_interrupt(int irq, void *dev_id)
|
|||
return IRQ_HANDLED;
|
||||
}
|
||||
|
||||
static void i2c_pnx_timeout(struct timer_list *t)
|
||||
static void i2c_pnx_timeout(struct i2c_pnx_algo_data *alg_data)
|
||||
{
|
||||
struct i2c_pnx_algo_data *alg_data = from_timer(alg_data, t, mif.timer);
|
||||
u32 ctl;
|
||||
|
||||
dev_err(&alg_data->adapter.dev,
|
||||
|
@ -472,7 +441,6 @@ static void i2c_pnx_timeout(struct timer_list *t)
|
|||
iowrite32(ctl, I2C_REG_CTL(alg_data));
|
||||
wait_reset(alg_data);
|
||||
alg_data->mif.ret = -EIO;
|
||||
complete(&alg_data->mif.complete);
|
||||
}
|
||||
|
||||
static inline void bus_reset_if_active(struct i2c_pnx_algo_data *alg_data)
|
||||
|
@ -514,6 +482,7 @@ i2c_pnx_xfer(struct i2c_adapter *adap, struct i2c_msg *msgs, int num)
|
|||
struct i2c_msg *pmsg;
|
||||
int rc = 0, completed = 0, i;
|
||||
struct i2c_pnx_algo_data *alg_data = adap->algo_data;
|
||||
unsigned long time_left;
|
||||
u32 stat;
|
||||
|
||||
dev_dbg(&alg_data->adapter.dev,
|
||||
|
@ -548,7 +517,6 @@ i2c_pnx_xfer(struct i2c_adapter *adap, struct i2c_msg *msgs, int num)
|
|||
dev_dbg(&alg_data->adapter.dev, "%s(): mode %d, %d bytes\n",
|
||||
__func__, alg_data->mif.mode, alg_data->mif.len);
|
||||
|
||||
i2c_pnx_arm_timer(alg_data);
|
||||
|
||||
/* initialize the completion var */
|
||||
init_completion(&alg_data->mif.complete);
|
||||
|
@ -564,7 +532,10 @@ i2c_pnx_xfer(struct i2c_adapter *adap, struct i2c_msg *msgs, int num)
|
|||
break;
|
||||
|
||||
/* Wait for completion */
|
||||
wait_for_completion(&alg_data->mif.complete);
|
||||
time_left = wait_for_completion_timeout(&alg_data->mif.complete,
|
||||
alg_data->timeout);
|
||||
if (time_left == 0)
|
||||
i2c_pnx_timeout(alg_data);
|
||||
|
||||
if (!(rc = alg_data->mif.ret))
|
||||
completed++;
|
||||
|
@ -653,7 +624,10 @@ static int i2c_pnx_probe(struct platform_device *pdev)
|
|||
alg_data->adapter.algo_data = alg_data;
|
||||
alg_data->adapter.nr = pdev->id;
|
||||
|
||||
alg_data->timeout = I2C_PNX_TIMEOUT_DEFAULT;
|
||||
alg_data->timeout = msecs_to_jiffies(I2C_PNX_TIMEOUT_DEFAULT);
|
||||
if (alg_data->timeout <= 1)
|
||||
alg_data->timeout = 2;
|
||||
|
||||
#ifdef CONFIG_OF
|
||||
alg_data->adapter.dev.of_node = of_node_get(pdev->dev.of_node);
|
||||
if (pdev->dev.of_node) {
|
||||
|
@ -673,8 +647,6 @@ static int i2c_pnx_probe(struct platform_device *pdev)
|
|||
if (IS_ERR(alg_data->clk))
|
||||
return PTR_ERR(alg_data->clk);
|
||||
|
||||
timer_setup(&alg_data->mif.timer, i2c_pnx_timeout, 0);
|
||||
|
||||
snprintf(alg_data->adapter.name, sizeof(alg_data->adapter.name),
|
||||
"%s", pdev->name);
|
||||
|
||||
|
|
|
@ -63,6 +63,8 @@ MODULE_AUTHOR("Roland Dreier");
|
|||
MODULE_DESCRIPTION("InfiniBand userspace MAD packet access");
|
||||
MODULE_LICENSE("Dual BSD/GPL");
|
||||
|
||||
#define MAX_UMAD_RECV_LIST_SIZE 200000
|
||||
|
||||
enum {
|
||||
IB_UMAD_MAX_PORTS = RDMA_MAX_PORTS,
|
||||
IB_UMAD_MAX_AGENTS = 32,
|
||||
|
@ -113,6 +115,7 @@ struct ib_umad_file {
|
|||
struct mutex mutex;
|
||||
struct ib_umad_port *port;
|
||||
struct list_head recv_list;
|
||||
atomic_t recv_list_size;
|
||||
struct list_head send_list;
|
||||
struct list_head port_list;
|
||||
spinlock_t send_lock;
|
||||
|
@ -180,24 +183,28 @@ static struct ib_mad_agent *__get_agent(struct ib_umad_file *file, int id)
|
|||
return file->agents_dead ? NULL : file->agent[id];
|
||||
}
|
||||
|
||||
static int queue_packet(struct ib_umad_file *file,
|
||||
struct ib_mad_agent *agent,
|
||||
struct ib_umad_packet *packet)
|
||||
static int queue_packet(struct ib_umad_file *file, struct ib_mad_agent *agent,
|
||||
struct ib_umad_packet *packet, bool is_recv_mad)
|
||||
{
|
||||
int ret = 1;
|
||||
|
||||
mutex_lock(&file->mutex);
|
||||
|
||||
if (is_recv_mad &&
|
||||
atomic_read(&file->recv_list_size) > MAX_UMAD_RECV_LIST_SIZE)
|
||||
goto unlock;
|
||||
|
||||
for (packet->mad.hdr.id = 0;
|
||||
packet->mad.hdr.id < IB_UMAD_MAX_AGENTS;
|
||||
packet->mad.hdr.id++)
|
||||
if (agent == __get_agent(file, packet->mad.hdr.id)) {
|
||||
list_add_tail(&packet->list, &file->recv_list);
|
||||
atomic_inc(&file->recv_list_size);
|
||||
wake_up_interruptible(&file->recv_wait);
|
||||
ret = 0;
|
||||
break;
|
||||
}
|
||||
|
||||
unlock:
|
||||
mutex_unlock(&file->mutex);
|
||||
|
||||
return ret;
|
||||
|
@ -224,7 +231,7 @@ static void send_handler(struct ib_mad_agent *agent,
|
|||
if (send_wc->status == IB_WC_RESP_TIMEOUT_ERR) {
|
||||
packet->length = IB_MGMT_MAD_HDR;
|
||||
packet->mad.hdr.status = ETIMEDOUT;
|
||||
if (!queue_packet(file, agent, packet))
|
||||
if (!queue_packet(file, agent, packet, false))
|
||||
return;
|
||||
}
|
||||
kfree(packet);
|
||||
|
@ -284,7 +291,7 @@ static void recv_handler(struct ib_mad_agent *agent,
|
|||
rdma_destroy_ah_attr(&ah_attr);
|
||||
}
|
||||
|
||||
if (queue_packet(file, agent, packet))
|
||||
if (queue_packet(file, agent, packet, true))
|
||||
goto err2;
|
||||
return;
|
||||
|
||||
|
@ -409,6 +416,7 @@ static ssize_t ib_umad_read(struct file *filp, char __user *buf,
|
|||
|
||||
packet = list_entry(file->recv_list.next, struct ib_umad_packet, list);
|
||||
list_del(&packet->list);
|
||||
atomic_dec(&file->recv_list_size);
|
||||
|
||||
mutex_unlock(&file->mutex);
|
||||
|
||||
|
@ -421,6 +429,7 @@ static ssize_t ib_umad_read(struct file *filp, char __user *buf,
|
|||
/* Requeue packet */
|
||||
mutex_lock(&file->mutex);
|
||||
list_add(&packet->list, &file->recv_list);
|
||||
atomic_inc(&file->recv_list_size);
|
||||
mutex_unlock(&file->mutex);
|
||||
} else {
|
||||
if (packet->recv_wc)
|
||||
|
|
|
@ -9,8 +9,10 @@
|
|||
/* #define DEBUG */
|
||||
|
||||
#include <linux/input.h>
|
||||
#include <linux/limits.h>
|
||||
#include <linux/module.h>
|
||||
#include <linux/mutex.h>
|
||||
#include <linux/overflow.h>
|
||||
#include <linux/sched.h>
|
||||
#include <linux/slab.h>
|
||||
|
||||
|
@ -315,9 +317,8 @@ int input_ff_create(struct input_dev *dev, unsigned int max_effects)
|
|||
return -EINVAL;
|
||||
}
|
||||
|
||||
ff_dev_size = sizeof(struct ff_device) +
|
||||
max_effects * sizeof(struct file *);
|
||||
if (ff_dev_size < max_effects) /* overflow */
|
||||
ff_dev_size = struct_size(ff, effect_owners, max_effects);
|
||||
if (ff_dev_size == SIZE_MAX) /* overflow */
|
||||
return -EINVAL;
|
||||
|
||||
ff = kzalloc(ff_dev_size, GFP_KERNEL);
|
||||
|
|
|
@ -283,7 +283,10 @@ static int an30259a_probe(struct i2c_client *client)
|
|||
if (err < 0)
|
||||
return err;
|
||||
|
||||
mutex_init(&chip->mutex);
|
||||
err = devm_mutex_init(&client->dev, &chip->mutex);
|
||||
if (err)
|
||||
return err;
|
||||
|
||||
chip->client = client;
|
||||
i2c_set_clientdata(client, chip);
|
||||
|
||||
|
@ -317,17 +320,9 @@ static int an30259a_probe(struct i2c_client *client)
|
|||
return 0;
|
||||
|
||||
exit:
|
||||
mutex_destroy(&chip->mutex);
|
||||
return err;
|
||||
}
|
||||
|
||||
static void an30259a_remove(struct i2c_client *client)
|
||||
{
|
||||
struct an30259a *chip = i2c_get_clientdata(client);
|
||||
|
||||
mutex_destroy(&chip->mutex);
|
||||
}
|
||||
|
||||
static const struct of_device_id an30259a_match_table[] = {
|
||||
{ .compatible = "panasonic,an30259a", },
|
||||
{ /* sentinel */ },
|
||||
|
@ -347,7 +342,6 @@ static struct i2c_driver an30259a_driver = {
|
|||
.of_match_table = an30259a_match_table,
|
||||
},
|
||||
.probe = an30259a_probe,
|
||||
.remove = an30259a_remove,
|
||||
.id_table = an30259a_id,
|
||||
};
|
||||
|
||||
|
|
|
@ -174,6 +174,6 @@ struct as10x_register_addr {
|
|||
uint32_t addr;
|
||||
/* register mode access */
|
||||
uint8_t mode;
|
||||
};
|
||||
} __packed;
|
||||
|
||||
#endif
|
||||
|
|
|
@ -410,6 +410,7 @@ static int tda10048_set_if(struct dvb_frontend *fe, u32 bw)
|
|||
struct tda10048_config *config = &state->config;
|
||||
int i;
|
||||
u32 if_freq_khz;
|
||||
u64 sample_freq;
|
||||
|
||||
dprintk(1, "%s(bw = %d)\n", __func__, bw);
|
||||
|
||||
|
@ -451,9 +452,11 @@ static int tda10048_set_if(struct dvb_frontend *fe, u32 bw)
|
|||
dprintk(1, "- pll_pfactor = %d\n", state->pll_pfactor);
|
||||
|
||||
/* Calculate the sample frequency */
|
||||
state->sample_freq = state->xtal_hz * (state->pll_mfactor + 45);
|
||||
state->sample_freq /= (state->pll_nfactor + 1);
|
||||
state->sample_freq /= (state->pll_pfactor + 4);
|
||||
sample_freq = state->xtal_hz;
|
||||
sample_freq *= state->pll_mfactor + 45;
|
||||
do_div(sample_freq, state->pll_nfactor + 1);
|
||||
do_div(sample_freq, state->pll_pfactor + 4);
|
||||
state->sample_freq = sample_freq;
|
||||
dprintk(1, "- sample_freq = %d\n", state->sample_freq);
|
||||
|
||||
/* Update the I/F */
|
||||
|
|
|
@ -328,7 +328,7 @@ static int CalcMainPLL(struct tda_state *state, u32 freq)
|
|||
|
||||
OscFreq = (u64) freq * (u64) Div;
|
||||
OscFreq *= (u64) 16384;
|
||||
do_div(OscFreq, (u64)16000000);
|
||||
do_div(OscFreq, 16000000);
|
||||
MainDiv = OscFreq;
|
||||
|
||||
state->m_Regs[MPD] = PostDiv & 0x77;
|
||||
|
@ -352,7 +352,7 @@ static int CalcCalPLL(struct tda_state *state, u32 freq)
|
|||
OscFreq = (u64)freq * (u64)Div;
|
||||
/* CalDiv = u32( OscFreq * 16384 / 16000000 ); */
|
||||
OscFreq *= (u64)16384;
|
||||
do_div(OscFreq, (u64)16000000);
|
||||
do_div(OscFreq, 16000000);
|
||||
CalDiv = OscFreq;
|
||||
|
||||
state->m_Regs[CPD] = PostDiv;
|
||||
|
|
|
@ -1023,18 +1023,26 @@ static void vdec_av1_slice_free_working_buffer(struct vdec_av1_slice_instance *i
|
|||
int i;
|
||||
|
||||
for (i = 0; i < ARRAY_SIZE(instance->mv); i++)
|
||||
mtk_vcodec_mem_free(ctx, &instance->mv[i]);
|
||||
if (instance->mv[i].va)
|
||||
mtk_vcodec_mem_free(ctx, &instance->mv[i]);
|
||||
|
||||
for (i = 0; i < ARRAY_SIZE(instance->seg); i++)
|
||||
mtk_vcodec_mem_free(ctx, &instance->seg[i]);
|
||||
if (instance->seg[i].va)
|
||||
mtk_vcodec_mem_free(ctx, &instance->seg[i]);
|
||||
|
||||
for (i = 0; i < ARRAY_SIZE(instance->cdf); i++)
|
||||
mtk_vcodec_mem_free(ctx, &instance->cdf[i]);
|
||||
if (instance->cdf[i].va)
|
||||
mtk_vcodec_mem_free(ctx, &instance->cdf[i]);
|
||||
|
||||
mtk_vcodec_mem_free(ctx, &instance->tile);
|
||||
mtk_vcodec_mem_free(ctx, &instance->cdf_temp);
|
||||
mtk_vcodec_mem_free(ctx, &instance->cdf_table);
|
||||
mtk_vcodec_mem_free(ctx, &instance->iq_table);
|
||||
|
||||
if (instance->tile.va)
|
||||
mtk_vcodec_mem_free(ctx, &instance->tile);
|
||||
if (instance->cdf_temp.va)
|
||||
mtk_vcodec_mem_free(ctx, &instance->cdf_temp);
|
||||
if (instance->cdf_table.va)
|
||||
mtk_vcodec_mem_free(ctx, &instance->cdf_table);
|
||||
if (instance->iq_table.va)
|
||||
mtk_vcodec_mem_free(ctx, &instance->iq_table);
|
||||
|
||||
instance->level = AV1_RES_NONE;
|
||||
}
|
||||
|
|
|
@ -301,11 +301,12 @@ static void h264_enc_free_work_buf(struct venc_h264_inst *inst)
|
|||
* other buffers need to be freed by AP.
|
||||
*/
|
||||
for (i = 0; i < VENC_H264_VPU_WORK_BUF_MAX; i++) {
|
||||
if (i != VENC_H264_VPU_WORK_BUF_SKIP_FRAME)
|
||||
if (i != VENC_H264_VPU_WORK_BUF_SKIP_FRAME && inst->work_bufs[i].va)
|
||||
mtk_vcodec_mem_free(inst->ctx, &inst->work_bufs[i]);
|
||||
}
|
||||
|
||||
mtk_vcodec_mem_free(inst->ctx, &inst->pps_buf);
|
||||
if (inst->pps_buf.va)
|
||||
mtk_vcodec_mem_free(inst->ctx, &inst->pps_buf);
|
||||
}
|
||||
|
||||
static int h264_enc_alloc_work_buf(struct venc_h264_inst *inst, bool is_34bit)
|
||||
|
|
|
@ -2412,7 +2412,12 @@ static int stk9090m_frontend_attach(struct dvb_usb_adapter *adap)
|
|||
|
||||
adap->fe_adap[0].fe = dvb_attach(dib9000_attach, &adap->dev->i2c_adap, 0x80, &stk9090m_config);
|
||||
|
||||
return adap->fe_adap[0].fe == NULL ? -ENODEV : 0;
|
||||
if (!adap->fe_adap[0].fe) {
|
||||
release_firmware(state->frontend_firmware);
|
||||
return -ENODEV;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int dib9090_tuner_attach(struct dvb_usb_adapter *adap)
|
||||
|
@ -2485,8 +2490,10 @@ static int nim9090md_frontend_attach(struct dvb_usb_adapter *adap)
|
|||
dib9000_i2c_enumeration(&adap->dev->i2c_adap, 1, 0x20, 0x80);
|
||||
adap->fe_adap[0].fe = dvb_attach(dib9000_attach, &adap->dev->i2c_adap, 0x80, &nim9090md_config[0]);
|
||||
|
||||
if (adap->fe_adap[0].fe == NULL)
|
||||
if (!adap->fe_adap[0].fe) {
|
||||
release_firmware(state->frontend_firmware);
|
||||
return -ENODEV;
|
||||
}
|
||||
|
||||
i2c = dib9000_get_i2c_master(adap->fe_adap[0].fe, DIBX000_I2C_INTERFACE_GPIO_3_4, 0);
|
||||
dib9000_i2c_enumeration(i2c, 1, 0x12, 0x82);
|
||||
|
@ -2494,7 +2501,12 @@ static int nim9090md_frontend_attach(struct dvb_usb_adapter *adap)
|
|||
fe_slave = dvb_attach(dib9000_attach, i2c, 0x82, &nim9090md_config[1]);
|
||||
dib9000_set_slave_frontend(adap->fe_adap[0].fe, fe_slave);
|
||||
|
||||
return fe_slave == NULL ? -ENODEV : 0;
|
||||
if (!fe_slave) {
|
||||
release_firmware(state->frontend_firmware);
|
||||
return -ENODEV;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int nim9090md_tuner_attach(struct dvb_usb_adapter *adap)
|
||||
|
|
|
@ -716,6 +716,7 @@ static int su3000_i2c_transfer(struct i2c_adapter *adap, struct i2c_msg msg[],
|
|||
{
|
||||
struct dvb_usb_device *d = i2c_get_adapdata(adap);
|
||||
struct dw2102_state *state;
|
||||
int j;
|
||||
|
||||
if (!d)
|
||||
return -ENODEV;
|
||||
|
@ -729,11 +730,11 @@ static int su3000_i2c_transfer(struct i2c_adapter *adap, struct i2c_msg msg[],
|
|||
return -EAGAIN;
|
||||
}
|
||||
|
||||
switch (num) {
|
||||
case 1:
|
||||
switch (msg[0].addr) {
|
||||
j = 0;
|
||||
while (j < num) {
|
||||
switch (msg[j].addr) {
|
||||
case SU3000_STREAM_CTRL:
|
||||
state->data[0] = msg[0].buf[0] + 0x36;
|
||||
state->data[0] = msg[j].buf[0] + 0x36;
|
||||
state->data[1] = 3;
|
||||
state->data[2] = 0;
|
||||
if (dvb_usb_generic_rw(d, state->data, 3,
|
||||
|
@ -745,61 +746,86 @@ static int su3000_i2c_transfer(struct i2c_adapter *adap, struct i2c_msg msg[],
|
|||
if (dvb_usb_generic_rw(d, state->data, 1,
|
||||
state->data, 2, 0) < 0)
|
||||
err("i2c transfer failed.");
|
||||
msg[0].buf[1] = state->data[0];
|
||||
msg[0].buf[0] = state->data[1];
|
||||
msg[j].buf[1] = state->data[0];
|
||||
msg[j].buf[0] = state->data[1];
|
||||
break;
|
||||
default:
|
||||
if (3 + msg[0].len > sizeof(state->data)) {
|
||||
warn("i2c wr: len=%d is too big!\n",
|
||||
msg[0].len);
|
||||
/* if the current write msg is followed by a another
|
||||
* read msg to/from the same address
|
||||
*/
|
||||
if ((j+1 < num) && (msg[j+1].flags & I2C_M_RD) &&
|
||||
(msg[j].addr == msg[j+1].addr)) {
|
||||
/* join both i2c msgs to one usb read command */
|
||||
if (4 + msg[j].len > sizeof(state->data)) {
|
||||
warn("i2c combined wr/rd: write len=%d is too big!\n",
|
||||
msg[j].len);
|
||||
num = -EOPNOTSUPP;
|
||||
break;
|
||||
}
|
||||
if (1 + msg[j+1].len > sizeof(state->data)) {
|
||||
warn("i2c combined wr/rd: read len=%d is too big!\n",
|
||||
msg[j+1].len);
|
||||
num = -EOPNOTSUPP;
|
||||
break;
|
||||
}
|
||||
|
||||
state->data[0] = 0x09;
|
||||
state->data[1] = msg[j].len;
|
||||
state->data[2] = msg[j+1].len;
|
||||
state->data[3] = msg[j].addr;
|
||||
memcpy(&state->data[4], msg[j].buf, msg[j].len);
|
||||
|
||||
if (dvb_usb_generic_rw(d, state->data, msg[j].len + 4,
|
||||
state->data, msg[j+1].len + 1, 0) < 0)
|
||||
err("i2c transfer failed.");
|
||||
|
||||
memcpy(msg[j+1].buf, &state->data[1], msg[j+1].len);
|
||||
j++;
|
||||
break;
|
||||
}
|
||||
|
||||
if (msg[j].flags & I2C_M_RD) {
|
||||
/* single read */
|
||||
if (4 + msg[j].len > sizeof(state->data)) {
|
||||
warn("i2c rd: len=%d is too big!\n", msg[j].len);
|
||||
num = -EOPNOTSUPP;
|
||||
break;
|
||||
}
|
||||
|
||||
state->data[0] = 0x09;
|
||||
state->data[1] = 0;
|
||||
state->data[2] = msg[j].len;
|
||||
state->data[3] = msg[j].addr;
|
||||
memcpy(&state->data[4], msg[j].buf, msg[j].len);
|
||||
|
||||
if (dvb_usb_generic_rw(d, state->data, 4,
|
||||
state->data, msg[j].len + 1, 0) < 0)
|
||||
err("i2c transfer failed.");
|
||||
|
||||
memcpy(msg[j].buf, &state->data[1], msg[j].len);
|
||||
break;
|
||||
}
|
||||
|
||||
/* single write */
|
||||
if (3 + msg[j].len > sizeof(state->data)) {
|
||||
warn("i2c wr: len=%d is too big!\n", msg[j].len);
|
||||
num = -EOPNOTSUPP;
|
||||
break;
|
||||
}
|
||||
|
||||
/* always i2c write*/
|
||||
state->data[0] = 0x08;
|
||||
state->data[1] = msg[0].addr;
|
||||
state->data[2] = msg[0].len;
|
||||
state->data[1] = msg[j].addr;
|
||||
state->data[2] = msg[j].len;
|
||||
|
||||
memcpy(&state->data[3], msg[0].buf, msg[0].len);
|
||||
memcpy(&state->data[3], msg[j].buf, msg[j].len);
|
||||
|
||||
if (dvb_usb_generic_rw(d, state->data, msg[0].len + 3,
|
||||
if (dvb_usb_generic_rw(d, state->data, msg[j].len + 3,
|
||||
state->data, 1, 0) < 0)
|
||||
err("i2c transfer failed.");
|
||||
} // switch
|
||||
j++;
|
||||
|
||||
}
|
||||
break;
|
||||
case 2:
|
||||
/* always i2c read */
|
||||
if (4 + msg[0].len > sizeof(state->data)) {
|
||||
warn("i2c rd: len=%d is too big!\n",
|
||||
msg[0].len);
|
||||
num = -EOPNOTSUPP;
|
||||
break;
|
||||
}
|
||||
if (1 + msg[1].len > sizeof(state->data)) {
|
||||
warn("i2c rd: len=%d is too big!\n",
|
||||
msg[1].len);
|
||||
num = -EOPNOTSUPP;
|
||||
break;
|
||||
}
|
||||
|
||||
state->data[0] = 0x09;
|
||||
state->data[1] = msg[0].len;
|
||||
state->data[2] = msg[1].len;
|
||||
state->data[3] = msg[0].addr;
|
||||
memcpy(&state->data[4], msg[0].buf, msg[0].len);
|
||||
|
||||
if (dvb_usb_generic_rw(d, state->data, msg[0].len + 4,
|
||||
state->data, msg[1].len + 1, 0) < 0)
|
||||
err("i2c transfer failed.");
|
||||
|
||||
memcpy(msg[1].buf, &state->data[1], msg[1].len);
|
||||
break;
|
||||
default:
|
||||
warn("more than 2 i2c messages at a time is not handled yet.");
|
||||
break;
|
||||
}
|
||||
} // while
|
||||
mutex_unlock(&d->data_mutex);
|
||||
mutex_unlock(&d->i2c_mutex);
|
||||
return num;
|
||||
|
|
|
@ -247,7 +247,7 @@ struct s2255_vc {
|
|||
struct s2255_dev {
|
||||
struct s2255_vc vc[MAX_CHANNELS];
|
||||
struct v4l2_device v4l2_dev;
|
||||
atomic_t num_channels;
|
||||
refcount_t num_channels;
|
||||
int frames;
|
||||
struct mutex lock; /* channels[].vdev.lock */
|
||||
struct mutex cmdlock; /* protects cmdbuf */
|
||||
|
@ -1550,11 +1550,11 @@ static void s2255_video_device_release(struct video_device *vdev)
|
|||
container_of(vdev, struct s2255_vc, vdev);
|
||||
|
||||
dprintk(dev, 4, "%s, chnls: %d\n", __func__,
|
||||
atomic_read(&dev->num_channels));
|
||||
refcount_read(&dev->num_channels));
|
||||
|
||||
v4l2_ctrl_handler_free(&vc->hdl);
|
||||
|
||||
if (atomic_dec_and_test(&dev->num_channels))
|
||||
if (refcount_dec_and_test(&dev->num_channels))
|
||||
s2255_destroy(dev);
|
||||
return;
|
||||
}
|
||||
|
@ -1659,7 +1659,7 @@ static int s2255_probe_v4l(struct s2255_dev *dev)
|
|||
"failed to register video device!\n");
|
||||
break;
|
||||
}
|
||||
atomic_inc(&dev->num_channels);
|
||||
refcount_inc(&dev->num_channels);
|
||||
v4l2_info(&dev->v4l2_dev, "V4L2 device registered as %s\n",
|
||||
video_device_node_name(&vc->vdev));
|
||||
|
||||
|
@ -1667,11 +1667,11 @@ static int s2255_probe_v4l(struct s2255_dev *dev)
|
|||
pr_info("Sensoray 2255 V4L driver Revision: %s\n",
|
||||
S2255_VERSION);
|
||||
/* if no channels registered, return error and probe will fail*/
|
||||
if (atomic_read(&dev->num_channels) == 0) {
|
||||
if (refcount_read(&dev->num_channels) == 0) {
|
||||
v4l2_device_unregister(&dev->v4l2_dev);
|
||||
return ret;
|
||||
}
|
||||
if (atomic_read(&dev->num_channels) != MAX_CHANNELS)
|
||||
if (refcount_read(&dev->num_channels) != MAX_CHANNELS)
|
||||
pr_warn("s2255: Not all channels available.\n");
|
||||
return 0;
|
||||
}
|
||||
|
@ -2220,7 +2220,7 @@ static int s2255_probe(struct usb_interface *interface,
|
|||
goto errorFWDATA1;
|
||||
}
|
||||
|
||||
atomic_set(&dev->num_channels, 0);
|
||||
refcount_set(&dev->num_channels, 0);
|
||||
dev->pid = id->idProduct;
|
||||
dev->fw_data = kzalloc(sizeof(struct s2255_fw), GFP_KERNEL);
|
||||
if (!dev->fw_data)
|
||||
|
@ -2340,12 +2340,12 @@ static void s2255_disconnect(struct usb_interface *interface)
|
|||
{
|
||||
struct s2255_dev *dev = to_s2255_dev(usb_get_intfdata(interface));
|
||||
int i;
|
||||
int channels = atomic_read(&dev->num_channels);
|
||||
int channels = refcount_read(&dev->num_channels);
|
||||
mutex_lock(&dev->lock);
|
||||
v4l2_device_disconnect(&dev->v4l2_dev);
|
||||
mutex_unlock(&dev->lock);
|
||||
/*see comments in the uvc_driver.c usb disconnect function */
|
||||
atomic_inc(&dev->num_channels);
|
||||
refcount_inc(&dev->num_channels);
|
||||
/* unregister each video device. */
|
||||
for (i = 0; i < channels; i++)
|
||||
video_unregister_device(&dev->vc[i].vdev);
|
||||
|
@ -2358,7 +2358,7 @@ static void s2255_disconnect(struct usb_interface *interface)
|
|||
dev->vc[i].vidstatus_ready = 1;
|
||||
wake_up(&dev->vc[i].wait_vidstatus);
|
||||
}
|
||||
if (atomic_dec_and_test(&dev->num_channels))
|
||||
if (refcount_dec_and_test(&dev->num_channels))
|
||||
s2255_destroy(dev);
|
||||
dev_info(&interface->dev, "%s\n", __func__);
|
||||
}
|
||||
|
|
|
@ -1090,28 +1090,32 @@ static int nand_fill_column_cycles(struct nand_chip *chip, u8 *addrs,
|
|||
unsigned int offset_in_page)
|
||||
{
|
||||
struct mtd_info *mtd = nand_to_mtd(chip);
|
||||
bool ident_stage = !mtd->writesize;
|
||||
|
||||
/* Make sure the offset is less than the actual page size. */
|
||||
if (offset_in_page > mtd->writesize + mtd->oobsize)
|
||||
return -EINVAL;
|
||||
|
||||
/*
|
||||
* On small page NANDs, there's a dedicated command to access the OOB
|
||||
* area, and the column address is relative to the start of the OOB
|
||||
* area, not the start of the page. Asjust the address accordingly.
|
||||
*/
|
||||
if (mtd->writesize <= 512 && offset_in_page >= mtd->writesize)
|
||||
offset_in_page -= mtd->writesize;
|
||||
|
||||
/*
|
||||
* The offset in page is expressed in bytes, if the NAND bus is 16-bit
|
||||
* wide, then it must be divided by 2.
|
||||
*/
|
||||
if (chip->options & NAND_BUSWIDTH_16) {
|
||||
if (WARN_ON(offset_in_page % 2))
|
||||
/* Bypass all checks during NAND identification */
|
||||
if (likely(!ident_stage)) {
|
||||
/* Make sure the offset is less than the actual page size. */
|
||||
if (offset_in_page > mtd->writesize + mtd->oobsize)
|
||||
return -EINVAL;
|
||||
|
||||
offset_in_page /= 2;
|
||||
/*
|
||||
* On small page NANDs, there's a dedicated command to access the OOB
|
||||
* area, and the column address is relative to the start of the OOB
|
||||
* area, not the start of the page. Asjust the address accordingly.
|
||||
*/
|
||||
if (mtd->writesize <= 512 && offset_in_page >= mtd->writesize)
|
||||
offset_in_page -= mtd->writesize;
|
||||
|
||||
/*
|
||||
* The offset in page is expressed in bytes, if the NAND bus is 16-bit
|
||||
* wide, then it must be divided by 2.
|
||||
*/
|
||||
if (chip->options & NAND_BUSWIDTH_16) {
|
||||
if (WARN_ON(offset_in_page % 2))
|
||||
return -EINVAL;
|
||||
|
||||
offset_in_page /= 2;
|
||||
}
|
||||
}
|
||||
|
||||
addrs[0] = offset_in_page;
|
||||
|
@ -1120,7 +1124,7 @@ static int nand_fill_column_cycles(struct nand_chip *chip, u8 *addrs,
|
|||
* Small page NANDs use 1 cycle for the columns, while large page NANDs
|
||||
* need 2
|
||||
*/
|
||||
if (mtd->writesize <= 512)
|
||||
if (!ident_stage && mtd->writesize <= 512)
|
||||
return 1;
|
||||
|
||||
addrs[1] = offset_in_page >> 8;
|
||||
|
@ -1419,16 +1423,19 @@ int nand_change_read_column_op(struct nand_chip *chip,
|
|||
unsigned int len, bool force_8bit)
|
||||
{
|
||||
struct mtd_info *mtd = nand_to_mtd(chip);
|
||||
bool ident_stage = !mtd->writesize;
|
||||
|
||||
if (len && !buf)
|
||||
return -EINVAL;
|
||||
|
||||
if (offset_in_page + len > mtd->writesize + mtd->oobsize)
|
||||
return -EINVAL;
|
||||
if (!ident_stage) {
|
||||
if (offset_in_page + len > mtd->writesize + mtd->oobsize)
|
||||
return -EINVAL;
|
||||
|
||||
/* Small page NANDs do not support column change. */
|
||||
if (mtd->writesize <= 512)
|
||||
return -ENOTSUPP;
|
||||
/* Small page NANDs do not support column change. */
|
||||
if (mtd->writesize <= 512)
|
||||
return -ENOTSUPP;
|
||||
}
|
||||
|
||||
if (nand_has_exec_op(chip)) {
|
||||
const struct nand_interface_config *conf =
|
||||
|
@ -2154,7 +2161,7 @@ EXPORT_SYMBOL_GPL(nand_reset_op);
|
|||
int nand_read_data_op(struct nand_chip *chip, void *buf, unsigned int len,
|
||||
bool force_8bit, bool check_only)
|
||||
{
|
||||
if (!len || !buf)
|
||||
if (!len || (!check_only && !buf))
|
||||
return -EINVAL;
|
||||
|
||||
if (nand_has_exec_op(chip)) {
|
||||
|
@ -6282,6 +6289,7 @@ static const struct nand_ops rawnand_ops = {
|
|||
static int nand_scan_tail(struct nand_chip *chip)
|
||||
{
|
||||
struct mtd_info *mtd = nand_to_mtd(chip);
|
||||
struct nand_device *base = &chip->base;
|
||||
struct nand_ecc_ctrl *ecc = &chip->ecc;
|
||||
int ret, i;
|
||||
|
||||
|
@ -6426,9 +6434,13 @@ static int nand_scan_tail(struct nand_chip *chip)
|
|||
if (!ecc->write_oob_raw)
|
||||
ecc->write_oob_raw = ecc->write_oob;
|
||||
|
||||
/* propagate ecc info to mtd_info */
|
||||
/* Propagate ECC info to the generic NAND and MTD layers */
|
||||
mtd->ecc_strength = ecc->strength;
|
||||
if (!base->ecc.ctx.conf.strength)
|
||||
base->ecc.ctx.conf.strength = ecc->strength;
|
||||
mtd->ecc_step_size = ecc->size;
|
||||
if (!base->ecc.ctx.conf.step_size)
|
||||
base->ecc.ctx.conf.step_size = ecc->size;
|
||||
|
||||
/*
|
||||
* Set the number of read / write steps for one page depending on ECC
|
||||
|
@ -6436,6 +6448,8 @@ static int nand_scan_tail(struct nand_chip *chip)
|
|||
*/
|
||||
if (!ecc->steps)
|
||||
ecc->steps = mtd->writesize / ecc->size;
|
||||
if (!base->ecc.ctx.nsteps)
|
||||
base->ecc.ctx.nsteps = ecc->steps;
|
||||
if (ecc->steps * ecc->size != mtd->writesize) {
|
||||
WARN(1, "Invalid ECC parameters\n");
|
||||
ret = -EINVAL;
|
||||
|
|
|
@ -420,13 +420,13 @@ static int rk_nfc_setup_interface(struct nand_chip *chip, int target,
|
|||
u32 rate, tc2rw, trwpw, trw2c;
|
||||
u32 temp;
|
||||
|
||||
if (target < 0)
|
||||
return 0;
|
||||
|
||||
timings = nand_get_sdr_timings(conf);
|
||||
if (IS_ERR(timings))
|
||||
return -EOPNOTSUPP;
|
||||
|
||||
if (target < 0)
|
||||
return 0;
|
||||
|
||||
if (IS_ERR(nfc->nfc_clk))
|
||||
rate = clk_get_rate(nfc->ahb_clk);
|
||||
else
|
||||
|
|
|
@ -1198,9 +1198,9 @@ static int bond_option_arp_ip_targets_set(struct bonding *bond,
|
|||
__be32 target;
|
||||
|
||||
if (newval->string) {
|
||||
if (!in4_pton(newval->string+1, -1, (u8 *)&target, -1, NULL)) {
|
||||
netdev_err(bond->dev, "invalid ARP target %pI4 specified\n",
|
||||
&target);
|
||||
if (strlen(newval->string) < 1 ||
|
||||
!in4_pton(newval->string + 1, -1, (u8 *)&target, -1, NULL)) {
|
||||
netdev_err(bond->dev, "invalid ARP target specified\n");
|
||||
return ret;
|
||||
}
|
||||
if (newval->string[0] == '+')
|
||||
|
|
|
@ -124,6 +124,7 @@ static const struct kvaser_usb_driver_info kvaser_usb_driver_info_leaf_err_liste
|
|||
|
||||
static const struct kvaser_usb_driver_info kvaser_usb_driver_info_leafimx = {
|
||||
.quirks = 0,
|
||||
.family = KVASER_LEAF,
|
||||
.ops = &kvaser_usb_leaf_dev_ops,
|
||||
};
|
||||
|
||||
|
|
|
@ -131,8 +131,8 @@ struct mii_bus *mv88e6xxx_default_mdio_bus(struct mv88e6xxx_chip *chip)
|
|||
{
|
||||
struct mv88e6xxx_mdio_bus *mdio_bus;
|
||||
|
||||
mdio_bus = list_first_entry(&chip->mdios, struct mv88e6xxx_mdio_bus,
|
||||
list);
|
||||
mdio_bus = list_first_entry_or_null(&chip->mdios,
|
||||
struct mv88e6xxx_mdio_bus, list);
|
||||
if (!mdio_bus)
|
||||
return NULL;
|
||||
|
||||
|
|
|
@ -1262,7 +1262,7 @@ enum {
|
|||
|
||||
struct bnx2x_fw_stats_req {
|
||||
struct stats_query_header hdr;
|
||||
struct stats_query_entry query[FP_SB_MAX_E1x+
|
||||
struct stats_query_entry query[FP_SB_MAX_E2 +
|
||||
BNX2X_FIRST_QUEUE_QUERY_IDX];
|
||||
};
|
||||
|
||||
|
|
|
@ -6363,49 +6363,49 @@ static void e1000e_s0ix_entry_flow(struct e1000_adapter *adapter)
|
|||
mac_data |= E1000_EXTCNF_CTRL_GATE_PHY_CFG;
|
||||
ew32(EXTCNF_CTRL, mac_data);
|
||||
|
||||
/* Enable the Dynamic Power Gating in the MAC */
|
||||
mac_data = er32(FEXTNVM7);
|
||||
mac_data |= BIT(22);
|
||||
ew32(FEXTNVM7, mac_data);
|
||||
|
||||
/* Disable disconnected cable conditioning for Power Gating */
|
||||
mac_data = er32(DPGFR);
|
||||
mac_data |= BIT(2);
|
||||
ew32(DPGFR, mac_data);
|
||||
|
||||
/* Don't wake from dynamic Power Gating with clock request */
|
||||
mac_data = er32(FEXTNVM12);
|
||||
mac_data |= BIT(12);
|
||||
ew32(FEXTNVM12, mac_data);
|
||||
|
||||
/* Ungate PGCB clock */
|
||||
mac_data = er32(FEXTNVM9);
|
||||
mac_data &= ~BIT(28);
|
||||
ew32(FEXTNVM9, mac_data);
|
||||
|
||||
/* Enable K1 off to enable mPHY Power Gating */
|
||||
mac_data = er32(FEXTNVM6);
|
||||
mac_data |= BIT(31);
|
||||
ew32(FEXTNVM6, mac_data);
|
||||
|
||||
/* Enable mPHY power gating for any link and speed */
|
||||
mac_data = er32(FEXTNVM8);
|
||||
mac_data |= BIT(9);
|
||||
ew32(FEXTNVM8, mac_data);
|
||||
|
||||
/* Enable the Dynamic Clock Gating in the DMA and MAC */
|
||||
mac_data = er32(CTRL_EXT);
|
||||
mac_data |= E1000_CTRL_EXT_DMA_DYN_CLK_EN;
|
||||
ew32(CTRL_EXT, mac_data);
|
||||
|
||||
/* No MAC DPG gating SLP_S0 in modern standby
|
||||
* Switch the logic of the lanphypc to use PMC counter
|
||||
*/
|
||||
mac_data = er32(FEXTNVM5);
|
||||
mac_data |= BIT(7);
|
||||
ew32(FEXTNVM5, mac_data);
|
||||
}
|
||||
|
||||
/* Enable the Dynamic Power Gating in the MAC */
|
||||
mac_data = er32(FEXTNVM7);
|
||||
mac_data |= BIT(22);
|
||||
ew32(FEXTNVM7, mac_data);
|
||||
|
||||
/* Don't wake from dynamic Power Gating with clock request */
|
||||
mac_data = er32(FEXTNVM12);
|
||||
mac_data |= BIT(12);
|
||||
ew32(FEXTNVM12, mac_data);
|
||||
|
||||
/* Ungate PGCB clock */
|
||||
mac_data = er32(FEXTNVM9);
|
||||
mac_data &= ~BIT(28);
|
||||
ew32(FEXTNVM9, mac_data);
|
||||
|
||||
/* Enable K1 off to enable mPHY Power Gating */
|
||||
mac_data = er32(FEXTNVM6);
|
||||
mac_data |= BIT(31);
|
||||
ew32(FEXTNVM6, mac_data);
|
||||
|
||||
/* Enable mPHY power gating for any link and speed */
|
||||
mac_data = er32(FEXTNVM8);
|
||||
mac_data |= BIT(9);
|
||||
ew32(FEXTNVM8, mac_data);
|
||||
|
||||
/* No MAC DPG gating SLP_S0 in modern standby
|
||||
* Switch the logic of the lanphypc to use PMC counter
|
||||
*/
|
||||
mac_data = er32(FEXTNVM5);
|
||||
mac_data |= BIT(7);
|
||||
ew32(FEXTNVM5, mac_data);
|
||||
|
||||
/* Disable the time synchronization clock */
|
||||
mac_data = er32(FEXTNVM7);
|
||||
mac_data |= BIT(31);
|
||||
|
@ -6498,33 +6498,6 @@ static void e1000e_s0ix_exit_flow(struct e1000_adapter *adapter)
|
|||
} else {
|
||||
/* Request driver unconfigure the device from S0ix */
|
||||
|
||||
/* Disable the Dynamic Power Gating in the MAC */
|
||||
mac_data = er32(FEXTNVM7);
|
||||
mac_data &= 0xFFBFFFFF;
|
||||
ew32(FEXTNVM7, mac_data);
|
||||
|
||||
/* Disable mPHY power gating for any link and speed */
|
||||
mac_data = er32(FEXTNVM8);
|
||||
mac_data &= ~BIT(9);
|
||||
ew32(FEXTNVM8, mac_data);
|
||||
|
||||
/* Disable K1 off */
|
||||
mac_data = er32(FEXTNVM6);
|
||||
mac_data &= ~BIT(31);
|
||||
ew32(FEXTNVM6, mac_data);
|
||||
|
||||
/* Disable Ungate PGCB clock */
|
||||
mac_data = er32(FEXTNVM9);
|
||||
mac_data |= BIT(28);
|
||||
ew32(FEXTNVM9, mac_data);
|
||||
|
||||
/* Cancel not waking from dynamic
|
||||
* Power Gating with clock request
|
||||
*/
|
||||
mac_data = er32(FEXTNVM12);
|
||||
mac_data &= ~BIT(12);
|
||||
ew32(FEXTNVM12, mac_data);
|
||||
|
||||
/* Cancel disable disconnected cable conditioning
|
||||
* for Power Gating
|
||||
*/
|
||||
|
@ -6537,13 +6510,6 @@ static void e1000e_s0ix_exit_flow(struct e1000_adapter *adapter)
|
|||
mac_data &= 0xFFF7FFFF;
|
||||
ew32(CTRL_EXT, mac_data);
|
||||
|
||||
/* Revert the lanphypc logic to use the internal Gbe counter
|
||||
* and not the PMC counter
|
||||
*/
|
||||
mac_data = er32(FEXTNVM5);
|
||||
mac_data &= 0xFFFFFF7F;
|
||||
ew32(FEXTNVM5, mac_data);
|
||||
|
||||
/* Enable the periodic inband message,
|
||||
* Request PCIe clock in K1 page770_17[10:9] =01b
|
||||
*/
|
||||
|
@ -6581,6 +6547,40 @@ static void e1000e_s0ix_exit_flow(struct e1000_adapter *adapter)
|
|||
mac_data &= ~BIT(31);
|
||||
mac_data |= BIT(0);
|
||||
ew32(FEXTNVM7, mac_data);
|
||||
|
||||
/* Disable the Dynamic Power Gating in the MAC */
|
||||
mac_data = er32(FEXTNVM7);
|
||||
mac_data &= 0xFFBFFFFF;
|
||||
ew32(FEXTNVM7, mac_data);
|
||||
|
||||
/* Disable mPHY power gating for any link and speed */
|
||||
mac_data = er32(FEXTNVM8);
|
||||
mac_data &= ~BIT(9);
|
||||
ew32(FEXTNVM8, mac_data);
|
||||
|
||||
/* Disable K1 off */
|
||||
mac_data = er32(FEXTNVM6);
|
||||
mac_data &= ~BIT(31);
|
||||
ew32(FEXTNVM6, mac_data);
|
||||
|
||||
/* Disable Ungate PGCB clock */
|
||||
mac_data = er32(FEXTNVM9);
|
||||
mac_data |= BIT(28);
|
||||
ew32(FEXTNVM9, mac_data);
|
||||
|
||||
/* Cancel not waking from dynamic
|
||||
* Power Gating with clock request
|
||||
*/
|
||||
mac_data = er32(FEXTNVM12);
|
||||
mac_data &= ~BIT(12);
|
||||
ew32(FEXTNVM12, mac_data);
|
||||
|
||||
/* Revert the lanphypc logic to use the internal Gbe counter
|
||||
* and not the PMC counter
|
||||
*/
|
||||
mac_data = er32(FEXTNVM5);
|
||||
mac_data &= 0xFFFFFF7F;
|
||||
ew32(FEXTNVM5, mac_data);
|
||||
}
|
||||
|
||||
static int e1000e_pm_freeze(struct device *dev)
|
||||
|
|
|
@ -5700,6 +5700,11 @@ void mlx5e_priv_cleanup(struct mlx5e_priv *priv)
|
|||
kfree(priv->htb_qos_sq_stats[i]);
|
||||
kvfree(priv->htb_qos_sq_stats);
|
||||
|
||||
if (priv->mqprio_rl) {
|
||||
mlx5e_mqprio_rl_cleanup(priv->mqprio_rl);
|
||||
mlx5e_mqprio_rl_free(priv->mqprio_rl);
|
||||
}
|
||||
|
||||
memset(priv, 0, sizeof(*priv));
|
||||
}
|
||||
|
||||
|
|
|
@ -6,6 +6,9 @@
|
|||
#include "helper.h"
|
||||
#include "ofld.h"
|
||||
|
||||
static int
|
||||
acl_ingress_ofld_setup(struct mlx5_eswitch *esw, struct mlx5_vport *vport);
|
||||
|
||||
static bool
|
||||
esw_acl_ingress_prio_tag_enabled(struct mlx5_eswitch *esw,
|
||||
const struct mlx5_vport *vport)
|
||||
|
@ -123,18 +126,31 @@ static int esw_acl_ingress_src_port_drop_create(struct mlx5_eswitch *esw,
|
|||
{
|
||||
struct mlx5_flow_act flow_act = {};
|
||||
struct mlx5_flow_handle *flow_rule;
|
||||
bool created = false;
|
||||
int err = 0;
|
||||
|
||||
if (!vport->ingress.acl) {
|
||||
err = acl_ingress_ofld_setup(esw, vport);
|
||||
if (err)
|
||||
return err;
|
||||
created = true;
|
||||
}
|
||||
|
||||
flow_act.action = MLX5_FLOW_CONTEXT_ACTION_DROP;
|
||||
flow_act.fg = vport->ingress.offloads.drop_grp;
|
||||
flow_rule = mlx5_add_flow_rules(vport->ingress.acl, NULL, &flow_act, NULL, 0);
|
||||
if (IS_ERR(flow_rule)) {
|
||||
err = PTR_ERR(flow_rule);
|
||||
goto out;
|
||||
goto err_out;
|
||||
}
|
||||
|
||||
vport->ingress.offloads.drop_rule = flow_rule;
|
||||
out:
|
||||
|
||||
return 0;
|
||||
err_out:
|
||||
/* Only destroy ingress acl created in this function. */
|
||||
if (created)
|
||||
esw_acl_ingress_ofld_cleanup(esw, vport);
|
||||
return err;
|
||||
}
|
||||
|
||||
|
@ -299,16 +315,12 @@ static void esw_acl_ingress_ofld_groups_destroy(struct mlx5_vport *vport)
|
|||
}
|
||||
}
|
||||
|
||||
int esw_acl_ingress_ofld_setup(struct mlx5_eswitch *esw,
|
||||
struct mlx5_vport *vport)
|
||||
static int
|
||||
acl_ingress_ofld_setup(struct mlx5_eswitch *esw, struct mlx5_vport *vport)
|
||||
{
|
||||
int num_ftes = 0;
|
||||
int err;
|
||||
|
||||
if (!mlx5_eswitch_vport_match_metadata_enabled(esw) &&
|
||||
!esw_acl_ingress_prio_tag_enabled(esw, vport))
|
||||
return 0;
|
||||
|
||||
esw_acl_ingress_allow_rule_destroy(vport);
|
||||
|
||||
if (mlx5_eswitch_vport_match_metadata_enabled(esw))
|
||||
|
@ -347,6 +359,15 @@ group_err:
|
|||
return err;
|
||||
}
|
||||
|
||||
int esw_acl_ingress_ofld_setup(struct mlx5_eswitch *esw, struct mlx5_vport *vport)
|
||||
{
|
||||
if (!mlx5_eswitch_vport_match_metadata_enabled(esw) &&
|
||||
!esw_acl_ingress_prio_tag_enabled(esw, vport))
|
||||
return 0;
|
||||
|
||||
return acl_ingress_ofld_setup(esw, vport);
|
||||
}
|
||||
|
||||
void esw_acl_ingress_ofld_cleanup(struct mlx5_eswitch *esw,
|
||||
struct mlx5_vport *vport)
|
||||
{
|
||||
|
|
|
@ -1484,6 +1484,7 @@ err_type_file_file_validate:
|
|||
vfree(types_info->data);
|
||||
err_data_alloc:
|
||||
kfree(types_info);
|
||||
linecards->types_info = NULL;
|
||||
return err;
|
||||
}
|
||||
|
||||
|
|
|
@ -268,7 +268,7 @@ static const struct ethqos_emac_por emac_v4_0_0_por[] = {
|
|||
|
||||
static const struct ethqos_emac_driver_data emac_v4_0_0_data = {
|
||||
.por = emac_v4_0_0_por,
|
||||
.num_por = ARRAY_SIZE(emac_v3_0_0_por),
|
||||
.num_por = ARRAY_SIZE(emac_v4_0_0_por),
|
||||
.rgmii_config_loopback_en = false,
|
||||
.has_emac_ge_3 = true,
|
||||
.link_clk_name = "phyaux",
|
||||
|
|
|
@ -1657,6 +1657,7 @@ static int wx_set_interrupt_capability(struct wx *wx)
|
|||
}
|
||||
|
||||
pdev->irq = pci_irq_vector(pdev, 0);
|
||||
wx->num_q_vectors = 1;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
|
|
@ -119,7 +119,7 @@ static void ntb_netdev_rx_handler(struct ntb_transport_qp *qp, void *qp_data,
|
|||
skb->protocol = eth_type_trans(skb, ndev);
|
||||
skb->ip_summed = CHECKSUM_NONE;
|
||||
|
||||
if (__netif_rx(skb) == NET_RX_DROP) {
|
||||
if (netif_rx(skb) == NET_RX_DROP) {
|
||||
ndev->stats.rx_errors++;
|
||||
ndev->stats.rx_dropped++;
|
||||
} else {
|
||||
|
|
|
@ -255,7 +255,7 @@ mt76_connac_mcu_add_nested_tlv(struct sk_buff *skb, int tag, int len,
|
|||
};
|
||||
u16 ntlv;
|
||||
|
||||
ptlv = skb_put(skb, len);
|
||||
ptlv = skb_put_zero(skb, len);
|
||||
memcpy(ptlv, &tlv, sizeof(tlv));
|
||||
|
||||
ntlv = le16_to_cpu(ntlv_hdr->tlv_num);
|
||||
|
@ -1654,7 +1654,7 @@ int mt76_connac_mcu_hw_scan(struct mt76_phy *phy, struct ieee80211_vif *vif,
|
|||
set_bit(MT76_HW_SCANNING, &phy->state);
|
||||
mvif->scan_seq_num = (mvif->scan_seq_num + 1) & 0x7f;
|
||||
|
||||
req = (struct mt76_connac_hw_scan_req *)skb_put(skb, sizeof(*req));
|
||||
req = (struct mt76_connac_hw_scan_req *)skb_put_zero(skb, sizeof(*req));
|
||||
|
||||
req->seq_num = mvif->scan_seq_num | mvif->band_idx << 7;
|
||||
req->bss_idx = mvif->idx;
|
||||
|
@ -1782,7 +1782,7 @@ int mt76_connac_mcu_sched_scan_req(struct mt76_phy *phy,
|
|||
|
||||
mvif->scan_seq_num = (mvif->scan_seq_num + 1) & 0x7f;
|
||||
|
||||
req = (struct mt76_connac_sched_scan_req *)skb_put(skb, sizeof(*req));
|
||||
req = (struct mt76_connac_sched_scan_req *)skb_put_zero(skb, sizeof(*req));
|
||||
req->version = 1;
|
||||
req->seq_num = mvif->scan_seq_num | mvif->band_idx << 7;
|
||||
|
||||
|
@ -2416,7 +2416,7 @@ int mt76_connac_mcu_update_gtk_rekey(struct ieee80211_hw *hw,
|
|||
return -ENOMEM;
|
||||
|
||||
skb_put_data(skb, &hdr, sizeof(hdr));
|
||||
gtk_tlv = (struct mt76_connac_gtk_rekey_tlv *)skb_put(skb,
|
||||
gtk_tlv = (struct mt76_connac_gtk_rekey_tlv *)skb_put_zero(skb,
|
||||
sizeof(*gtk_tlv));
|
||||
gtk_tlv->tag = cpu_to_le16(UNI_OFFLOAD_OFFLOAD_GTK_REKEY);
|
||||
gtk_tlv->len = cpu_to_le16(sizeof(*gtk_tlv));
|
||||
|
@ -2539,7 +2539,7 @@ mt76_connac_mcu_set_wow_pattern(struct mt76_dev *dev,
|
|||
return -ENOMEM;
|
||||
|
||||
skb_put_data(skb, &hdr, sizeof(hdr));
|
||||
ptlv = (struct mt76_connac_wow_pattern_tlv *)skb_put(skb, sizeof(*ptlv));
|
||||
ptlv = (struct mt76_connac_wow_pattern_tlv *)skb_put_zero(skb, sizeof(*ptlv));
|
||||
ptlv->tag = cpu_to_le16(UNI_SUSPEND_WOW_PATTERN);
|
||||
ptlv->len = cpu_to_le16(sizeof(*ptlv));
|
||||
ptlv->data_len = pattern->pattern_len;
|
||||
|
|
|
@ -422,7 +422,7 @@ mt7915_mcu_add_nested_subtlv(struct sk_buff *skb, int sub_tag, int sub_len,
|
|||
.len = cpu_to_le16(sub_len),
|
||||
};
|
||||
|
||||
ptlv = skb_put(skb, sub_len);
|
||||
ptlv = skb_put_zero(skb, sub_len);
|
||||
memcpy(ptlv, &tlv, sizeof(tlv));
|
||||
|
||||
le16_add_cpu(sub_ntlv, 1);
|
||||
|
|
|
@ -225,6 +225,11 @@ mt7996_radar_trigger(void *data, u64 val)
|
|||
if (val > MT_RX_SEL2)
|
||||
return -EINVAL;
|
||||
|
||||
if (val == MT_RX_SEL2 && !dev->rdd2_phy) {
|
||||
dev_err(dev->mt76.dev, "Background radar is not enabled\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
return mt7996_mcu_rdd_cmd(dev, RDD_RADAR_EMULATE,
|
||||
val, 0, 0);
|
||||
}
|
||||
|
|
|
@ -339,7 +339,10 @@ mt7996_mcu_rx_radar_detected(struct mt7996_dev *dev, struct sk_buff *skb)
|
|||
if (r->band_idx >= ARRAY_SIZE(dev->mt76.phys))
|
||||
return;
|
||||
|
||||
if (dev->rdd2_phy && r->band_idx == MT_RX_SEL2)
|
||||
if (r->band_idx == MT_RX_SEL2 && !dev->rdd2_phy)
|
||||
return;
|
||||
|
||||
if (r->band_idx == MT_RX_SEL2)
|
||||
mphy = dev->rdd2_phy->mt76;
|
||||
else
|
||||
mphy = dev->mt76.phys[r->band_idx];
|
||||
|
|
|
@ -379,7 +379,8 @@ void *wilc_parse_join_bss_param(struct cfg80211_bss *bss,
|
|||
struct ieee80211_p2p_noa_attr noa_attr;
|
||||
const struct cfg80211_bss_ies *ies;
|
||||
struct wilc_join_bss_param *param;
|
||||
u8 rates_len = 0, ies_len;
|
||||
u8 rates_len = 0;
|
||||
int ies_len;
|
||||
int ret;
|
||||
|
||||
param = kzalloc(sizeof(*param), GFP_KERNEL);
|
||||
|
|
|
@ -125,6 +125,10 @@ static ssize_t virtual_ncidev_write(struct file *file,
|
|||
kfree_skb(skb);
|
||||
return -EFAULT;
|
||||
}
|
||||
if (strnlen(skb->data, count) != count) {
|
||||
kfree_skb(skb);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
nci_recv_frame(vdev->ndev, skb);
|
||||
return count;
|
||||
|
|
|
@ -594,7 +594,7 @@ static void nvme_mpath_set_live(struct nvme_ns *ns)
|
|||
int node, srcu_idx;
|
||||
|
||||
srcu_idx = srcu_read_lock(&head->srcu);
|
||||
for_each_node(node)
|
||||
for_each_online_node(node)
|
||||
__nvme_find_path(head, node);
|
||||
srcu_read_unlock(&head->srcu, srcu_idx);
|
||||
}
|
||||
|
|
|
@ -778,7 +778,8 @@ static blk_status_t nvme_map_data(struct nvme_dev *dev, struct request *req,
|
|||
struct bio_vec bv = req_bvec(req);
|
||||
|
||||
if (!is_pci_p2pdma_page(bv.bv_page)) {
|
||||
if (bv.bv_offset + bv.bv_len <= NVME_CTRL_PAGE_SIZE * 2)
|
||||
if ((bv.bv_offset & (NVME_CTRL_PAGE_SIZE - 1)) +
|
||||
bv.bv_len <= NVME_CTRL_PAGE_SIZE * 2)
|
||||
return nvme_setup_prp_simple(dev, req,
|
||||
&cmnd->rw, &bv);
|
||||
|
||||
|
|
|
@ -806,6 +806,15 @@ void nvmet_sq_destroy(struct nvmet_sq *sq)
|
|||
percpu_ref_exit(&sq->ref);
|
||||
nvmet_auth_sq_free(sq);
|
||||
|
||||
/*
|
||||
* we must reference the ctrl again after waiting for inflight IO
|
||||
* to complete. Because admin connect may have sneaked in after we
|
||||
* store sq->ctrl locally, but before we killed the percpu_ref. the
|
||||
* admin connect allocates and assigns sq->ctrl, which now needs a
|
||||
* final ref put, as this ctrl is going away.
|
||||
*/
|
||||
ctrl = sq->ctrl;
|
||||
|
||||
if (ctrl) {
|
||||
/*
|
||||
* The teardown flow may take some time, and the host may not
|
||||
|
|
|
@ -3275,7 +3275,7 @@ static const char *find_hci_method(acpi_handle handle)
|
|||
*/
|
||||
#define QUIRK_HCI_HOTKEY_QUICKSTART BIT(1)
|
||||
|
||||
static const struct dmi_system_id toshiba_dmi_quirks[] = {
|
||||
static const struct dmi_system_id toshiba_dmi_quirks[] __initconst = {
|
||||
{
|
||||
/* Toshiba Portégé R700 */
|
||||
/* https://bugzilla.kernel.org/show_bug.cgi?id=21012 */
|
||||
|
@ -3310,8 +3310,6 @@ static int toshiba_acpi_add(struct acpi_device *acpi_dev)
|
|||
struct toshiba_acpi_dev *dev;
|
||||
const char *hci_method;
|
||||
u32 dummy;
|
||||
const struct dmi_system_id *dmi_id;
|
||||
long quirks = 0;
|
||||
int ret = 0;
|
||||
|
||||
if (toshiba_acpi)
|
||||
|
@ -3464,16 +3462,6 @@ iio_error:
|
|||
}
|
||||
#endif
|
||||
|
||||
dmi_id = dmi_first_match(toshiba_dmi_quirks);
|
||||
if (dmi_id)
|
||||
quirks = (long)dmi_id->driver_data;
|
||||
|
||||
if (turn_on_panel_on_resume == -1)
|
||||
turn_on_panel_on_resume = !!(quirks & QUIRK_TURN_ON_PANEL_ON_RESUME);
|
||||
|
||||
if (hci_hotkey_quickstart == -1)
|
||||
hci_hotkey_quickstart = !!(quirks & QUIRK_HCI_HOTKEY_QUICKSTART);
|
||||
|
||||
toshiba_wwan_available(dev);
|
||||
if (dev->wwan_supported)
|
||||
toshiba_acpi_setup_wwan_rfkill(dev);
|
||||
|
@ -3622,10 +3610,27 @@ static struct acpi_driver toshiba_acpi_driver = {
|
|||
.drv.pm = &toshiba_acpi_pm,
|
||||
};
|
||||
|
||||
static void __init toshiba_dmi_init(void)
|
||||
{
|
||||
const struct dmi_system_id *dmi_id;
|
||||
long quirks = 0;
|
||||
|
||||
dmi_id = dmi_first_match(toshiba_dmi_quirks);
|
||||
if (dmi_id)
|
||||
quirks = (long)dmi_id->driver_data;
|
||||
|
||||
if (turn_on_panel_on_resume == -1)
|
||||
turn_on_panel_on_resume = !!(quirks & QUIRK_TURN_ON_PANEL_ON_RESUME);
|
||||
|
||||
if (hci_hotkey_quickstart == -1)
|
||||
hci_hotkey_quickstart = !!(quirks & QUIRK_HCI_HOTKEY_QUICKSTART);
|
||||
}
|
||||
|
||||
static int __init toshiba_acpi_init(void)
|
||||
{
|
||||
int ret;
|
||||
|
||||
toshiba_dmi_init();
|
||||
toshiba_proc_dir = proc_mkdir(PROC_TOSHIBA, acpi_root_dir);
|
||||
if (!toshiba_proc_dir) {
|
||||
pr_err("Unable to create proc dir " PROC_TOSHIBA "\n");
|
||||
|
|
|
@ -902,6 +902,22 @@ static const struct ts_dmi_data schneider_sct101ctm_data = {
|
|||
.properties = schneider_sct101ctm_props,
|
||||
};
|
||||
|
||||
static const struct property_entry globalspace_solt_ivw116_props[] = {
|
||||
PROPERTY_ENTRY_U32("touchscreen-min-x", 7),
|
||||
PROPERTY_ENTRY_U32("touchscreen-min-y", 22),
|
||||
PROPERTY_ENTRY_U32("touchscreen-size-x", 1723),
|
||||
PROPERTY_ENTRY_U32("touchscreen-size-y", 1077),
|
||||
PROPERTY_ENTRY_STRING("firmware-name", "gsl1680-globalspace-solt-ivw116.fw"),
|
||||
PROPERTY_ENTRY_U32("silead,max-fingers", 10),
|
||||
PROPERTY_ENTRY_BOOL("silead,home-button"),
|
||||
{ }
|
||||
};
|
||||
|
||||
static const struct ts_dmi_data globalspace_solt_ivw116_data = {
|
||||
.acpi_name = "MSSL1680:00",
|
||||
.properties = globalspace_solt_ivw116_props,
|
||||
};
|
||||
|
||||
static const struct property_entry techbite_arc_11_6_props[] = {
|
||||
PROPERTY_ENTRY_U32("touchscreen-min-x", 5),
|
||||
PROPERTY_ENTRY_U32("touchscreen-min-y", 7),
|
||||
|
@ -1390,6 +1406,17 @@ const struct dmi_system_id touchscreen_dmi_table[] = {
|
|||
DMI_MATCH(DMI_BIOS_DATE, "04/24/2018"),
|
||||
},
|
||||
},
|
||||
{
|
||||
/* Jumper EZpad 6s Pro */
|
||||
.driver_data = (void *)&jumper_ezpad_6_pro_b_data,
|
||||
.matches = {
|
||||
DMI_MATCH(DMI_SYS_VENDOR, "Jumper"),
|
||||
DMI_MATCH(DMI_PRODUCT_NAME, "Ezpad"),
|
||||
/* Above matches are too generic, add bios match */
|
||||
DMI_MATCH(DMI_BIOS_VERSION, "E.WSA116_8.E1.042.bin"),
|
||||
DMI_MATCH(DMI_BIOS_DATE, "01/08/2020"),
|
||||
},
|
||||
},
|
||||
{
|
||||
/* Jumper EZpad 6 m4 */
|
||||
.driver_data = (void *)&jumper_ezpad_6_m4_data,
|
||||
|
@ -1629,6 +1656,15 @@ const struct dmi_system_id touchscreen_dmi_table[] = {
|
|||
DMI_MATCH(DMI_PRODUCT_NAME, "SCT101CTM"),
|
||||
},
|
||||
},
|
||||
{
|
||||
/* GlobalSpace SoLT IVW 11.6" */
|
||||
.driver_data = (void *)&globalspace_solt_ivw116_data,
|
||||
.matches = {
|
||||
DMI_MATCH(DMI_SYS_VENDOR, "Globalspace Tech Pvt Ltd"),
|
||||
DMI_MATCH(DMI_PRODUCT_NAME, "SolTIVW"),
|
||||
DMI_MATCH(DMI_PRODUCT_SKU, "PN20170413488"),
|
||||
},
|
||||
},
|
||||
{
|
||||
/* Techbite Arc 11.6 */
|
||||
.driver_data = (void *)&techbite_arc_11_6_data,
|
||||
|
|
|
@ -1369,7 +1369,7 @@ static long pkey_unlocked_ioctl(struct file *filp, unsigned int cmd,
|
|||
if (rc)
|
||||
break;
|
||||
if (copy_to_user(ucs, &kcs, sizeof(kcs)))
|
||||
return -EFAULT;
|
||||
rc = -EFAULT;
|
||||
memzero_explicit(&kcs, sizeof(kcs));
|
||||
break;
|
||||
}
|
||||
|
@ -1404,7 +1404,7 @@ static long pkey_unlocked_ioctl(struct file *filp, unsigned int cmd,
|
|||
if (rc)
|
||||
break;
|
||||
if (copy_to_user(ucp, &kcp, sizeof(kcp)))
|
||||
return -EFAULT;
|
||||
rc = -EFAULT;
|
||||
memzero_explicit(&kcp, sizeof(kcp));
|
||||
break;
|
||||
}
|
||||
|
|
|
@ -1355,11 +1355,21 @@ static struct mpi3mr_sas_port *mpi3mr_sas_port_add(struct mpi3mr_ioc *mrioc,
|
|||
mpi3mr_sas_port_sanity_check(mrioc, mr_sas_node,
|
||||
mr_sas_port->remote_identify.sas_address, hba_port);
|
||||
|
||||
if (mr_sas_node->num_phys > sizeof(mr_sas_port->phy_mask) * 8)
|
||||
ioc_info(mrioc, "max port count %u could be too high\n",
|
||||
mr_sas_node->num_phys);
|
||||
|
||||
for (i = 0; i < mr_sas_node->num_phys; i++) {
|
||||
if ((mr_sas_node->phy[i].remote_identify.sas_address !=
|
||||
mr_sas_port->remote_identify.sas_address) ||
|
||||
(mr_sas_node->phy[i].hba_port != hba_port))
|
||||
continue;
|
||||
|
||||
if (i > sizeof(mr_sas_port->phy_mask) * 8) {
|
||||
ioc_warn(mrioc, "skipping port %u, max allowed value is %zu\n",
|
||||
i, sizeof(mr_sas_port->phy_mask) * 8);
|
||||
goto out_fail;
|
||||
}
|
||||
list_add_tail(&mr_sas_node->phy[i].port_siblings,
|
||||
&mr_sas_port->phy_list);
|
||||
mr_sas_port->num_phys++;
|
||||
|
|
|
@ -2331,9 +2331,6 @@ static int qedf_execute_tmf(struct qedf_rport *fcport, struct scsi_cmnd *sc_cmd,
|
|||
io_req->fcport = fcport;
|
||||
io_req->cmd_type = QEDF_TASK_MGMT_CMD;
|
||||
|
||||
/* Record which cpu this request is associated with */
|
||||
io_req->cpu = smp_processor_id();
|
||||
|
||||
/* Set TM flags */
|
||||
io_req->io_req_flags = QEDF_READ;
|
||||
io_req->data_xfer_len = 0;
|
||||
|
@ -2355,6 +2352,9 @@ static int qedf_execute_tmf(struct qedf_rport *fcport, struct scsi_cmnd *sc_cmd,
|
|||
|
||||
spin_lock_irqsave(&fcport->rport_lock, flags);
|
||||
|
||||
/* Record which cpu this request is associated with */
|
||||
io_req->cpu = smp_processor_id();
|
||||
|
||||
sqe_idx = qedf_get_sqe_idx(fcport);
|
||||
sqe = &fcport->sq[sqe_idx];
|
||||
memset(sqe, 0, sizeof(struct fcoe_wqe));
|
||||
|
|
|
@ -145,6 +145,9 @@
|
|||
#define CDNS_XSPI_STIG_DONE_FLAG BIT(0)
|
||||
#define CDNS_XSPI_TRD_STATUS 0x0104
|
||||
|
||||
#define MODE_NO_OF_BYTES GENMASK(25, 24)
|
||||
#define MODEBYTES_COUNT 1
|
||||
|
||||
/* Helper macros for filling command registers */
|
||||
#define CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_1(op, data_phase) ( \
|
||||
FIELD_PREP(CDNS_XSPI_CMD_INSTR_TYPE, (data_phase) ? \
|
||||
|
@ -157,9 +160,10 @@
|
|||
FIELD_PREP(CDNS_XSPI_CMD_P1_R2_ADDR3, ((op)->addr.val >> 24) & 0xFF) | \
|
||||
FIELD_PREP(CDNS_XSPI_CMD_P1_R2_ADDR4, ((op)->addr.val >> 32) & 0xFF))
|
||||
|
||||
#define CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_3(op) ( \
|
||||
#define CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_3(op, modebytes) ( \
|
||||
FIELD_PREP(CDNS_XSPI_CMD_P1_R3_ADDR5, ((op)->addr.val >> 40) & 0xFF) | \
|
||||
FIELD_PREP(CDNS_XSPI_CMD_P1_R3_CMD, (op)->cmd.opcode) | \
|
||||
FIELD_PREP(MODE_NO_OF_BYTES, modebytes) | \
|
||||
FIELD_PREP(CDNS_XSPI_CMD_P1_R3_NUM_ADDR_BYTES, (op)->addr.nbytes))
|
||||
|
||||
#define CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_4(op, chipsel) ( \
|
||||
|
@ -173,12 +177,12 @@
|
|||
#define CDNS_XSPI_CMD_FLD_DSEQ_CMD_2(op) \
|
||||
FIELD_PREP(CDNS_XSPI_CMD_DSEQ_R2_DCNT_L, (op)->data.nbytes & 0xFFFF)
|
||||
|
||||
#define CDNS_XSPI_CMD_FLD_DSEQ_CMD_3(op) ( \
|
||||
#define CDNS_XSPI_CMD_FLD_DSEQ_CMD_3(op, dummybytes) ( \
|
||||
FIELD_PREP(CDNS_XSPI_CMD_DSEQ_R3_DCNT_H, \
|
||||
((op)->data.nbytes >> 16) & 0xffff) | \
|
||||
FIELD_PREP(CDNS_XSPI_CMD_DSEQ_R3_NUM_OF_DUMMY, \
|
||||
(op)->dummy.buswidth != 0 ? \
|
||||
(((op)->dummy.nbytes * 8) / (op)->dummy.buswidth) : \
|
||||
(((dummybytes) * 8) / (op)->dummy.buswidth) : \
|
||||
0))
|
||||
|
||||
#define CDNS_XSPI_CMD_FLD_DSEQ_CMD_4(op, chipsel) ( \
|
||||
|
@ -351,6 +355,7 @@ static int cdns_xspi_send_stig_command(struct cdns_xspi_dev *cdns_xspi,
|
|||
u32 cmd_regs[6];
|
||||
u32 cmd_status;
|
||||
int ret;
|
||||
int dummybytes = op->dummy.nbytes;
|
||||
|
||||
ret = cdns_xspi_wait_for_controller_idle(cdns_xspi);
|
||||
if (ret < 0)
|
||||
|
@ -365,7 +370,12 @@ static int cdns_xspi_send_stig_command(struct cdns_xspi_dev *cdns_xspi,
|
|||
memset(cmd_regs, 0, sizeof(cmd_regs));
|
||||
cmd_regs[1] = CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_1(op, data_phase);
|
||||
cmd_regs[2] = CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_2(op);
|
||||
cmd_regs[3] = CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_3(op);
|
||||
if (dummybytes != 0) {
|
||||
cmd_regs[3] = CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_3(op, 1);
|
||||
dummybytes--;
|
||||
} else {
|
||||
cmd_regs[3] = CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_3(op, 0);
|
||||
}
|
||||
cmd_regs[4] = CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_4(op,
|
||||
cdns_xspi->cur_cs);
|
||||
|
||||
|
@ -375,7 +385,7 @@ static int cdns_xspi_send_stig_command(struct cdns_xspi_dev *cdns_xspi,
|
|||
cmd_regs[0] = CDNS_XSPI_STIG_DONE_FLAG;
|
||||
cmd_regs[1] = CDNS_XSPI_CMD_FLD_DSEQ_CMD_1(op);
|
||||
cmd_regs[2] = CDNS_XSPI_CMD_FLD_DSEQ_CMD_2(op);
|
||||
cmd_regs[3] = CDNS_XSPI_CMD_FLD_DSEQ_CMD_3(op);
|
||||
cmd_regs[3] = CDNS_XSPI_CMD_FLD_DSEQ_CMD_3(op, dummybytes);
|
||||
cmd_regs[4] = CDNS_XSPI_CMD_FLD_DSEQ_CMD_4(op,
|
||||
cdns_xspi->cur_cs);
|
||||
|
||||
|
|
|
@ -1208,6 +1208,8 @@ static int lvts_probe(struct platform_device *pdev)
|
|||
return -ENOMEM;
|
||||
|
||||
lvts_data = of_device_get_match_data(dev);
|
||||
if (!lvts_data)
|
||||
return -ENODEV;
|
||||
|
||||
lvts_td->clk = devm_clk_get_enabled(dev, NULL);
|
||||
if (IS_ERR(lvts_td->clk))
|
||||
|
|
|
@ -1320,7 +1320,7 @@ static void imx_uart_clear_rx_errors(struct imx_port *sport)
|
|||
|
||||
}
|
||||
|
||||
#define TXTL_DEFAULT 2 /* reset default */
|
||||
#define TXTL_DEFAULT 8
|
||||
#define RXTL_DEFAULT 8 /* 8 characters or aging timer */
|
||||
#define TXTL_DMA 8 /* DMA burst setting */
|
||||
#define RXTL_DMA 9 /* DMA burst setting */
|
||||
|
|
|
@ -2651,16 +2651,17 @@ static int handle_tx_event(struct xhci_hcd *xhci,
|
|||
else
|
||||
xhci_handle_halted_endpoint(xhci, ep, NULL,
|
||||
EP_SOFT_RESET);
|
||||
goto cleanup;
|
||||
break;
|
||||
case COMP_RING_UNDERRUN:
|
||||
case COMP_RING_OVERRUN:
|
||||
case COMP_STOPPED_LENGTH_INVALID:
|
||||
goto cleanup;
|
||||
break;
|
||||
default:
|
||||
xhci_err(xhci, "ERROR Transfer event for unknown stream ring slot %u ep %u\n",
|
||||
slot_id, ep_index);
|
||||
goto err_out;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
/* Count current td numbers if ep->skip is set */
|
||||
|
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue