ASoC: Updates for v4.1
More updates for v4.1, pretty much all drivers: - Lots of cleanups from Lars, mainly moving things from the CODEC level to the card level. - Continuing improvements to rcar from Morimoto-san, pcm512x from Howard and Peter, the Intel platforms from Vinod, Jie, Jin and Han, and to rt5670 from Bard. - Support for some non-DSP Qualcomm platforms, Google's Storm platform, Maxmim MAX98925 CODECs and the Ingenic JZ4780 SoC. -----BEGIN PGP SIGNATURE----- Version: GnuPG v1 iQEcBAABAgAGBQJVK7E7AAoJECTWi3JdVIfQeRwH+wZHxh5yXBwSysga6ITc8GzZ Swy9LCdVr4uDXBrzn6qusfpgyiy6i09aB3tlr02j1gnhaA6tZ52Xi5S5RGds2nQL Qi+Nmt/7Expys09mJrE1Z8ZBRXnSbKw36odNxHiVVPVSfBGEXeQErDmLzsQ3ccqA 8HDC2TGRjjal9ZVW9kNsi5EkR9z8dRlkymAvzlpozs4aLwaOsH/xiF+4xI3zh0xZ +rG8HH9w3/yePVKiKZGjToNgzZ2ATLB5s+JZyFiDn0uXMo3UZWnQItPv8KJE0/FR A7D7XyMN66WSTnWHMIPetrJbgyNP9cM/Wk+prn/PKvObsqYiP7lJikzEAlklfJE= =LjN/ -----END PGP SIGNATURE----- Merge tag 'asoc-v4.1-2' of git://git.kernel.org/pub/scm/linux/kernel/git/broonie/sound into for-linus ASoC: Updates for v4.1 More updates for v4.1, pretty much all drivers: - Lots of cleanups from Lars, mainly moving things from the CODEC level to the card level. - Continuing improvements to rcar from Morimoto-san, pcm512x from Howard and Peter, the Intel platforms from Vinod, Jie, Jin and Han, and to rt5670 from Bard. - Support for some non-DSP Qualcomm platforms, Google's Storm platform, Maxmim MAX98925 CODECs and the Ingenic JZ4780 SoC.
This commit is contained in:
commit
ce4524e5a7
|
@ -19,7 +19,9 @@ the parent DSA node. The maximum number of allowed child nodes is 4
|
||||||
(DSA_MAX_SWITCHES).
|
(DSA_MAX_SWITCHES).
|
||||||
Each of these switch child nodes should have the following required properties:
|
Each of these switch child nodes should have the following required properties:
|
||||||
|
|
||||||
- reg : Describes the switch address on the MII bus
|
- reg : Contains two fields. The first one describes the
|
||||||
|
address on the MII bus. The second is the switch
|
||||||
|
number that must be unique in cascaded configurations
|
||||||
- #address-cells : Must be 1
|
- #address-cells : Must be 1
|
||||||
- #size-cells : Must be 0
|
- #size-cells : Must be 0
|
||||||
|
|
||||||
|
|
|
@ -1,7 +1,7 @@
|
||||||
Ingenic JZ4740 I2S controller
|
Ingenic JZ4740 I2S controller
|
||||||
|
|
||||||
Required properties:
|
Required properties:
|
||||||
- compatible : "ingenic,jz4740-i2s"
|
- compatible : "ingenic,jz4740-i2s" or "ingenic,jz4780-i2s"
|
||||||
- reg : I2S registers location and length
|
- reg : I2S registers location and length
|
||||||
- clocks : AIC and I2S PLL clock specifiers.
|
- clocks : AIC and I2S PLL clock specifiers.
|
||||||
- clock-names: "aic" and "i2s"
|
- clock-names: "aic" and "i2s"
|
||||||
|
|
|
@ -0,0 +1,22 @@
|
||||||
|
max98925 audio CODEC
|
||||||
|
|
||||||
|
This device supports I2C.
|
||||||
|
|
||||||
|
Required properties:
|
||||||
|
|
||||||
|
- compatible : "maxim,max98925"
|
||||||
|
|
||||||
|
- vmon-slot-no : slot number used to send voltage information
|
||||||
|
|
||||||
|
- imon-slot-no : slot number used to send current information
|
||||||
|
|
||||||
|
- reg : the I2C address of the device for I2C
|
||||||
|
|
||||||
|
Example:
|
||||||
|
|
||||||
|
codec: max98925@1a {
|
||||||
|
compatible = "maxim,max98925";
|
||||||
|
vmon-slot-no = <0>;
|
||||||
|
imon-slot-no = <2>;
|
||||||
|
reg = <0x1a>;
|
||||||
|
};
|
|
@ -0,0 +1,43 @@
|
||||||
|
* Qualcomm Technologies LPASS CPU DAI
|
||||||
|
|
||||||
|
This node models the Qualcomm Technologies Low-Power Audio SubSystem (LPASS).
|
||||||
|
|
||||||
|
Required properties:
|
||||||
|
|
||||||
|
- compatible : "qcom,lpass-cpu"
|
||||||
|
- clocks : Must contain an entry for each entry in clock-names.
|
||||||
|
- clock-names : A list which must include the following entries:
|
||||||
|
* "ahbix-clk"
|
||||||
|
* "mi2s-osr-clk"
|
||||||
|
* "mi2s-bit-clk"
|
||||||
|
- interrupts : Must contain an entry for each entry in
|
||||||
|
interrupt-names.
|
||||||
|
- interrupt-names : A list which must include the following entries:
|
||||||
|
* "lpass-irq-lpaif"
|
||||||
|
- pinctrl-N : One property must exist for each entry in
|
||||||
|
pinctrl-names. See ../pinctrl/pinctrl-bindings.txt
|
||||||
|
for details of the property values.
|
||||||
|
- pinctrl-names : Must contain a "default" entry.
|
||||||
|
- reg : Must contain an address for each entry in reg-names.
|
||||||
|
- reg-names : A list which must include the following entries:
|
||||||
|
* "lpass-lpaif"
|
||||||
|
|
||||||
|
Optional properties:
|
||||||
|
|
||||||
|
- qcom,adsp : Phandle for the audio DSP node
|
||||||
|
|
||||||
|
Example:
|
||||||
|
|
||||||
|
lpass@28100000 {
|
||||||
|
compatible = "qcom,lpass-cpu";
|
||||||
|
clocks = <&lcc AHBIX_CLK>, <&lcc MI2S_OSR_CLK>, <&lcc MI2S_BIT_CLK>;
|
||||||
|
clock-names = "ahbix-clk", "mi2s-osr-clk", "mi2s-bit-clk";
|
||||||
|
interrupts = <0 85 1>;
|
||||||
|
interrupt-names = "lpass-irq-lpaif";
|
||||||
|
pinctrl-names = "default", "idle";
|
||||||
|
pinctrl-0 = <&mi2s_default>;
|
||||||
|
pinctrl-1 = <&mi2s_idle>;
|
||||||
|
reg = <0x28100000 0x10000>;
|
||||||
|
reg-names = "lpass-lpaif";
|
||||||
|
qcom,adsp = <&adsp>;
|
||||||
|
};
|
|
@ -29,9 +29,17 @@ SSI subnode properties:
|
||||||
- shared-pin : if shared clock pin
|
- shared-pin : if shared clock pin
|
||||||
- pio-transfer : use PIO transfer mode
|
- pio-transfer : use PIO transfer mode
|
||||||
- no-busif : BUSIF is not ussed when [mem -> SSI] via DMA case
|
- no-busif : BUSIF is not ussed when [mem -> SSI] via DMA case
|
||||||
|
- dma : Should contain Audio DMAC entry
|
||||||
|
- dma-names : SSI case "rx" (=playback), "tx" (=capture)
|
||||||
|
SSIU case "rxu" (=playback), "txu" (=capture)
|
||||||
|
|
||||||
SRC subnode properties:
|
SRC subnode properties:
|
||||||
no properties at this point
|
- dma : Should contain Audio DMAC entry
|
||||||
|
- dma-names : "rx" (=playback), "tx" (=capture)
|
||||||
|
|
||||||
|
DVC subnode properties:
|
||||||
|
- dma : Should contain Audio DMAC entry
|
||||||
|
- dma-names : "tx" (=playback/capture)
|
||||||
|
|
||||||
DAI subnode properties:
|
DAI subnode properties:
|
||||||
- playback : list of playback modules
|
- playback : list of playback modules
|
||||||
|
@ -45,56 +53,145 @@ rcar_sound: rcar_sound@ec500000 {
|
||||||
reg = <0 0xec500000 0 0x1000>, /* SCU */
|
reg = <0 0xec500000 0 0x1000>, /* SCU */
|
||||||
<0 0xec5a0000 0 0x100>, /* ADG */
|
<0 0xec5a0000 0 0x100>, /* ADG */
|
||||||
<0 0xec540000 0 0x1000>, /* SSIU */
|
<0 0xec540000 0 0x1000>, /* SSIU */
|
||||||
<0 0xec541000 0 0x1280>; /* SSI */
|
<0 0xec541000 0 0x1280>, /* SSI */
|
||||||
|
<0 0xec740000 0 0x200>; /* Audio DMAC peri peri*/
|
||||||
|
reg-names = "scu", "adg", "ssiu", "ssi", "audmapp";
|
||||||
|
|
||||||
|
clocks = <&mstp10_clks R8A7790_CLK_SSI_ALL>,
|
||||||
|
<&mstp10_clks R8A7790_CLK_SSI9>, <&mstp10_clks R8A7790_CLK_SSI8>,
|
||||||
|
<&mstp10_clks R8A7790_CLK_SSI7>, <&mstp10_clks R8A7790_CLK_SSI6>,
|
||||||
|
<&mstp10_clks R8A7790_CLK_SSI5>, <&mstp10_clks R8A7790_CLK_SSI4>,
|
||||||
|
<&mstp10_clks R8A7790_CLK_SSI3>, <&mstp10_clks R8A7790_CLK_SSI2>,
|
||||||
|
<&mstp10_clks R8A7790_CLK_SSI1>, <&mstp10_clks R8A7790_CLK_SSI0>,
|
||||||
|
<&mstp10_clks R8A7790_CLK_SCU_SRC9>, <&mstp10_clks R8A7790_CLK_SCU_SRC8>,
|
||||||
|
<&mstp10_clks R8A7790_CLK_SCU_SRC7>, <&mstp10_clks R8A7790_CLK_SCU_SRC6>,
|
||||||
|
<&mstp10_clks R8A7790_CLK_SCU_SRC5>, <&mstp10_clks R8A7790_CLK_SCU_SRC4>,
|
||||||
|
<&mstp10_clks R8A7790_CLK_SCU_SRC3>, <&mstp10_clks R8A7790_CLK_SCU_SRC2>,
|
||||||
|
<&mstp10_clks R8A7790_CLK_SCU_SRC1>, <&mstp10_clks R8A7790_CLK_SCU_SRC0>,
|
||||||
|
<&mstp10_clks R8A7790_CLK_SCU_DVC0>, <&mstp10_clks R8A7790_CLK_SCU_DVC1>,
|
||||||
|
<&audio_clk_a>, <&audio_clk_b>, <&audio_clk_c>, <&m2_clk>;
|
||||||
|
clock-names = "ssi-all",
|
||||||
|
"ssi.9", "ssi.8", "ssi.7", "ssi.6", "ssi.5",
|
||||||
|
"ssi.4", "ssi.3", "ssi.2", "ssi.1", "ssi.0",
|
||||||
|
"src.9", "src.8", "src.7", "src.6", "src.5",
|
||||||
|
"src.4", "src.3", "src.2", "src.1", "src.0",
|
||||||
|
"dvc.0", "dvc.1",
|
||||||
|
"clk_a", "clk_b", "clk_c", "clk_i";
|
||||||
|
|
||||||
rcar_sound,dvc {
|
rcar_sound,dvc {
|
||||||
dvc0: dvc@0 { };
|
dvc0: dvc@0 {
|
||||||
dvc1: dvc@1 { };
|
dmas = <&audma0 0xbc>;
|
||||||
|
dma-names = "tx";
|
||||||
|
};
|
||||||
|
dvc1: dvc@1 {
|
||||||
|
dmas = <&audma0 0xbe>;
|
||||||
|
dma-names = "tx";
|
||||||
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
rcar_sound,src {
|
rcar_sound,src {
|
||||||
src0: src@0 { };
|
src0: src@0 {
|
||||||
src1: src@1 { };
|
interrupts = <0 352 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
src2: src@2 { };
|
dmas = <&audma0 0x85>, <&audma1 0x9a>;
|
||||||
src3: src@3 { };
|
dma-names = "rx", "tx";
|
||||||
src4: src@4 { };
|
};
|
||||||
src5: src@5 { };
|
src1: src@1 {
|
||||||
src6: src@6 { };
|
interrupts = <0 353 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
src7: src@7 { };
|
dmas = <&audma0 0x87>, <&audma1 0x9c>;
|
||||||
src8: src@8 { };
|
dma-names = "rx", "tx";
|
||||||
src9: src@9 { };
|
};
|
||||||
|
src2: src@2 {
|
||||||
|
interrupts = <0 354 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x89>, <&audma1 0x9e>;
|
||||||
|
dma-names = "rx", "tx";
|
||||||
|
};
|
||||||
|
src3: src@3 {
|
||||||
|
interrupts = <0 355 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x8b>, <&audma1 0xa0>;
|
||||||
|
dma-names = "rx", "tx";
|
||||||
|
};
|
||||||
|
src4: src@4 {
|
||||||
|
interrupts = <0 356 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x8d>, <&audma1 0xb0>;
|
||||||
|
dma-names = "rx", "tx";
|
||||||
|
};
|
||||||
|
src5: src@5 {
|
||||||
|
interrupts = <0 357 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x8f>, <&audma1 0xb2>;
|
||||||
|
dma-names = "rx", "tx";
|
||||||
|
};
|
||||||
|
src6: src@6 {
|
||||||
|
interrupts = <0 358 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x91>, <&audma1 0xb4>;
|
||||||
|
dma-names = "rx", "tx";
|
||||||
|
};
|
||||||
|
src7: src@7 {
|
||||||
|
interrupts = <0 359 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x93>, <&audma1 0xb6>;
|
||||||
|
dma-names = "rx", "tx";
|
||||||
|
};
|
||||||
|
src8: src@8 {
|
||||||
|
interrupts = <0 360 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x95>, <&audma1 0xb8>;
|
||||||
|
dma-names = "rx", "tx";
|
||||||
|
};
|
||||||
|
src9: src@9 {
|
||||||
|
interrupts = <0 361 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x97>, <&audma1 0xba>;
|
||||||
|
dma-names = "rx", "tx";
|
||||||
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
rcar_sound,ssi {
|
rcar_sound,ssi {
|
||||||
ssi0: ssi@0 {
|
ssi0: ssi@0 {
|
||||||
interrupts = <0 370 IRQ_TYPE_LEVEL_HIGH>;
|
interrupts = <0 370 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x01>, <&audma1 0x02>, <&audma0 0x15>, <&audma1 0x16>;
|
||||||
|
dma-names = "rx", "tx", "rxu", "txu";
|
||||||
};
|
};
|
||||||
ssi1: ssi@1 {
|
ssi1: ssi@1 {
|
||||||
interrupts = <0 371 IRQ_TYPE_LEVEL_HIGH>;
|
interrupts = <0 371 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x03>, <&audma1 0x04>, <&audma0 0x49>, <&audma1 0x4a>;
|
||||||
|
dma-names = "rx", "tx", "rxu", "txu";
|
||||||
};
|
};
|
||||||
ssi2: ssi@2 {
|
ssi2: ssi@2 {
|
||||||
interrupts = <0 372 IRQ_TYPE_LEVEL_HIGH>;
|
interrupts = <0 372 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x05>, <&audma1 0x06>, <&audma0 0x63>, <&audma1 0x64>;
|
||||||
|
dma-names = "rx", "tx", "rxu", "txu";
|
||||||
};
|
};
|
||||||
ssi3: ssi@3 {
|
ssi3: ssi@3 {
|
||||||
interrupts = <0 373 IRQ_TYPE_LEVEL_HIGH>;
|
interrupts = <0 373 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x07>, <&audma1 0x08>, <&audma0 0x6f>, <&audma1 0x70>;
|
||||||
|
dma-names = "rx", "tx", "rxu", "txu";
|
||||||
};
|
};
|
||||||
ssi4: ssi@4 {
|
ssi4: ssi@4 {
|
||||||
interrupts = <0 374 IRQ_TYPE_LEVEL_HIGH>;
|
interrupts = <0 374 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x09>, <&audma1 0x0a>, <&audma0 0x71>, <&audma1 0x72>;
|
||||||
|
dma-names = "rx", "tx", "rxu", "txu";
|
||||||
};
|
};
|
||||||
ssi5: ssi@5 {
|
ssi5: ssi@5 {
|
||||||
interrupts = <0 375 IRQ_TYPE_LEVEL_HIGH>;
|
interrupts = <0 375 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x0b>, <&audma1 0x0c>, <&audma0 0x73>, <&audma1 0x74>;
|
||||||
|
dma-names = "rx", "tx", "rxu", "txu";
|
||||||
};
|
};
|
||||||
ssi6: ssi@6 {
|
ssi6: ssi@6 {
|
||||||
interrupts = <0 376 IRQ_TYPE_LEVEL_HIGH>;
|
interrupts = <0 376 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x0d>, <&audma1 0x0e>, <&audma0 0x75>, <&audma1 0x76>;
|
||||||
|
dma-names = "rx", "tx", "rxu", "txu";
|
||||||
};
|
};
|
||||||
ssi7: ssi@7 {
|
ssi7: ssi@7 {
|
||||||
interrupts = <0 377 IRQ_TYPE_LEVEL_HIGH>;
|
interrupts = <0 377 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x0f>, <&audma1 0x10>, <&audma0 0x79>, <&audma1 0x7a>;
|
||||||
|
dma-names = "rx", "tx", "rxu", "txu";
|
||||||
};
|
};
|
||||||
ssi8: ssi@8 {
|
ssi8: ssi@8 {
|
||||||
interrupts = <0 378 IRQ_TYPE_LEVEL_HIGH>;
|
interrupts = <0 378 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x11>, <&audma1 0x12>, <&audma0 0x7b>, <&audma1 0x7c>;
|
||||||
|
dma-names = "rx", "tx", "rxu", "txu";
|
||||||
};
|
};
|
||||||
ssi9: ssi@9 {
|
ssi9: ssi@9 {
|
||||||
interrupts = <0 379 IRQ_TYPE_LEVEL_HIGH>;
|
interrupts = <0 379 IRQ_TYPE_LEVEL_HIGH>;
|
||||||
|
dmas = <&audma0 0x13>, <&audma1 0x14>, <&audma0 0x7d>, <&audma1 0x7e>;
|
||||||
|
dma-names = "rx", "tx", "rxu", "txu";
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -0,0 +1,67 @@
|
||||||
|
Renesas Sampling Rate Convert Sound Card:
|
||||||
|
|
||||||
|
Renesas Sampling Rate Convert Sound Card specifies audio DAI connections of SoC <-> codec.
|
||||||
|
|
||||||
|
Required properties:
|
||||||
|
|
||||||
|
- compatible : "renesas,rsrc-card,<board>"
|
||||||
|
Examples with soctypes are:
|
||||||
|
- "renesas,rsrc-card,lager"
|
||||||
|
- "renesas,rsrc-card,koelsch"
|
||||||
|
Optional properties:
|
||||||
|
|
||||||
|
- card_name : User specified audio sound card name, one string
|
||||||
|
property.
|
||||||
|
- cpu : CPU sub-node
|
||||||
|
- codec : CODEC sub-node
|
||||||
|
|
||||||
|
Optional subnode properties:
|
||||||
|
|
||||||
|
- format : CPU/CODEC common audio format.
|
||||||
|
"i2s", "right_j", "left_j" , "dsp_a"
|
||||||
|
"dsp_b", "ac97", "pdm", "msb", "lsb"
|
||||||
|
- frame-master : Indicates dai-link frame master.
|
||||||
|
phandle to a cpu or codec subnode.
|
||||||
|
- bitclock-master : Indicates dai-link bit clock master.
|
||||||
|
phandle to a cpu or codec subnode.
|
||||||
|
- bitclock-inversion : bool property. Add this if the
|
||||||
|
dai-link uses bit clock inversion.
|
||||||
|
- frame-inversion : bool property. Add this if the
|
||||||
|
dai-link uses frame clock inversion.
|
||||||
|
- convert-rate : platform specified sampling rate convert
|
||||||
|
|
||||||
|
Required CPU/CODEC subnodes properties:
|
||||||
|
|
||||||
|
- sound-dai : phandle and port of CPU/CODEC
|
||||||
|
|
||||||
|
Optional CPU/CODEC subnodes properties:
|
||||||
|
|
||||||
|
- clocks / system-clock-frequency : specify subnode's clock if needed.
|
||||||
|
it can be specified via "clocks" if system has
|
||||||
|
clock node (= common clock), or "system-clock-frequency"
|
||||||
|
(if system doens't support common clock)
|
||||||
|
If a clock is specified, it is
|
||||||
|
enabled with clk_prepare_enable()
|
||||||
|
in dai startup() and disabled with
|
||||||
|
clk_disable_unprepare() in dai
|
||||||
|
shutdown().
|
||||||
|
|
||||||
|
Example
|
||||||
|
|
||||||
|
sound {
|
||||||
|
compatible = "renesas,rsrc-card,lager";
|
||||||
|
|
||||||
|
card-name = "rsnd-ak4643";
|
||||||
|
format = "left_j";
|
||||||
|
bitclock-master = <&sndcodec>;
|
||||||
|
frame-master = <&sndcodec>;
|
||||||
|
|
||||||
|
sndcpu: cpu {
|
||||||
|
sound-dai = <&rcar_sound>;
|
||||||
|
};
|
||||||
|
|
||||||
|
sndcodec: codec {
|
||||||
|
sound-dai = <&ak4643>;
|
||||||
|
system-clock-frequency = <11289600>;
|
||||||
|
};
|
||||||
|
};
|
|
@ -0,0 +1,23 @@
|
||||||
|
* Sound complex for Storm boards
|
||||||
|
|
||||||
|
Models a soundcard for Storm boards with the Qualcomm Technologies IPQ806x SOC
|
||||||
|
connected to a MAX98357A DAC via I2S.
|
||||||
|
|
||||||
|
Required properties:
|
||||||
|
|
||||||
|
- compatible : "google,storm-audio"
|
||||||
|
- cpu : Phandle of the CPU DAI
|
||||||
|
- codec : Phandle of the codec DAI
|
||||||
|
|
||||||
|
Optional properties:
|
||||||
|
|
||||||
|
- qcom,model : The user-visible name of this sound card.
|
||||||
|
|
||||||
|
Example:
|
||||||
|
|
||||||
|
sound {
|
||||||
|
compatible = "google,storm-audio";
|
||||||
|
qcom,model = "ipq806x-storm";
|
||||||
|
cpu = <&lpass_cpu>;
|
||||||
|
codec = <&max98357a>;
|
||||||
|
};
|
|
@ -10,6 +10,13 @@ Required properties:
|
||||||
- reg : the I2C address of the device for I2C, the chip select
|
- reg : the I2C address of the device for I2C, the chip select
|
||||||
number for SPI.
|
number for SPI.
|
||||||
|
|
||||||
|
- PVDD-supply, DVDD-supply : Power supplies for the device, as covered
|
||||||
|
in Documentation/devicetree/bindings/regulator/regulator.txt
|
||||||
|
|
||||||
|
Optional properties:
|
||||||
|
|
||||||
|
- wlf,reset-gpio: A GPIO specifier for the GPIO controlling the reset pin
|
||||||
|
|
||||||
Example:
|
Example:
|
||||||
|
|
||||||
codec: wm8804@1a {
|
codec: wm8804@1a {
|
||||||
|
|
|
@ -114,6 +114,9 @@ ALPS Absolute Mode - Protocol Version 2
|
||||||
byte 4: 0 y6 y5 y4 y3 y2 y1 y0
|
byte 4: 0 y6 y5 y4 y3 y2 y1 y0
|
||||||
byte 5: 0 z6 z5 z4 z3 z2 z1 z0
|
byte 5: 0 z6 z5 z4 z3 z2 z1 z0
|
||||||
|
|
||||||
|
Protocol Version 2 DualPoint devices send standard PS/2 mouse packets for
|
||||||
|
the DualPoint Stick.
|
||||||
|
|
||||||
Dualpoint device -- interleaved packet format
|
Dualpoint device -- interleaved packet format
|
||||||
---------------------------------------------
|
---------------------------------------------
|
||||||
|
|
||||||
|
@ -127,6 +130,11 @@ Dualpoint device -- interleaved packet format
|
||||||
byte 7: 0 y6 y5 y4 y3 y2 y1 y0
|
byte 7: 0 y6 y5 y4 y3 y2 y1 y0
|
||||||
byte 8: 0 z6 z5 z4 z3 z2 z1 z0
|
byte 8: 0 z6 z5 z4 z3 z2 z1 z0
|
||||||
|
|
||||||
|
Devices which use the interleaving format normally send standard PS/2 mouse
|
||||||
|
packets for the DualPoint Stick + ALPS Absolute Mode packets for the
|
||||||
|
touchpad, switching to the interleaved packet format when both the stick and
|
||||||
|
the touchpad are used at the same time.
|
||||||
|
|
||||||
ALPS Absolute Mode - Protocol Version 3
|
ALPS Absolute Mode - Protocol Version 3
|
||||||
---------------------------------------
|
---------------------------------------
|
||||||
|
|
||||||
|
|
|
@ -294,6 +294,12 @@ accordingly. This property does not affect kernel behavior.
|
||||||
The kernel does not provide button emulation for such devices but treats
|
The kernel does not provide button emulation for such devices but treats
|
||||||
them as any other INPUT_PROP_BUTTONPAD device.
|
them as any other INPUT_PROP_BUTTONPAD device.
|
||||||
|
|
||||||
|
INPUT_PROP_ACCELEROMETER
|
||||||
|
-------------------------
|
||||||
|
Directional axes on this device (absolute and/or relative x, y, z) represent
|
||||||
|
accelerometer data. All other axes retain their meaning. A device must not mix
|
||||||
|
regular directional axes and accelerometer axes on the same event node.
|
||||||
|
|
||||||
Guidelines:
|
Guidelines:
|
||||||
==========
|
==========
|
||||||
The guidelines below ensure proper single-touch and multi-finger functionality.
|
The guidelines below ensure proper single-touch and multi-finger functionality.
|
||||||
|
|
|
@ -312,9 +312,12 @@ ABS_MT_TOOL_TYPE
|
||||||
|
|
||||||
The type of approaching tool. A lot of kernel drivers cannot distinguish
|
The type of approaching tool. A lot of kernel drivers cannot distinguish
|
||||||
between different tool types, such as a finger or a pen. In such cases, the
|
between different tool types, such as a finger or a pen. In such cases, the
|
||||||
event should be omitted. The protocol currently supports MT_TOOL_FINGER and
|
event should be omitted. The protocol currently supports MT_TOOL_FINGER,
|
||||||
MT_TOOL_PEN [2]. For type B devices, this event is handled by input core;
|
MT_TOOL_PEN, and MT_TOOL_PALM [2]. For type B devices, this event is handled
|
||||||
drivers should instead use input_mt_report_slot_state().
|
by input core; drivers should instead use input_mt_report_slot_state().
|
||||||
|
A contact's ABS_MT_TOOL_TYPE may change over time while still touching the
|
||||||
|
device, because the firmware may not be able to determine which tool is being
|
||||||
|
used when it first appears.
|
||||||
|
|
||||||
ABS_MT_BLOB_ID
|
ABS_MT_BLOB_ID
|
||||||
|
|
||||||
|
|
|
@ -71,11 +71,11 @@ SOURCE:
|
||||||
HDMI/DP (either HDMI or DisplayPort)
|
HDMI/DP (either HDMI or DisplayPort)
|
||||||
|
|
||||||
Exceptions (deprecated):
|
Exceptions (deprecated):
|
||||||
[Digital] Capture Source
|
[Analogue|Digital] Capture Source
|
||||||
[Digital] Capture Switch (aka input gain switch)
|
[Analogue|Digital] Capture Switch (aka input gain switch)
|
||||||
[Digital] Capture Volume (aka input gain volume)
|
[Analogue|Digital] Capture Volume (aka input gain volume)
|
||||||
[Digital] Playback Switch (aka output gain switch)
|
[Analogue|Digital] Playback Switch (aka output gain switch)
|
||||||
[Digital] Playback Volume (aka output gain volume)
|
[Analogue|Digital] Playback Volume (aka output gain volume)
|
||||||
Tone Control - Switch
|
Tone Control - Switch
|
||||||
Tone Control - Bass
|
Tone Control - Bass
|
||||||
Tone Control - Treble
|
Tone Control - Treble
|
||||||
|
|
39
MAINTAINERS
39
MAINTAINERS
|
@ -637,8 +637,7 @@ F: drivers/gpu/drm/radeon/radeon_kfd.h
|
||||||
F: include/uapi/linux/kfd_ioctl.h
|
F: include/uapi/linux/kfd_ioctl.h
|
||||||
|
|
||||||
AMD MICROCODE UPDATE SUPPORT
|
AMD MICROCODE UPDATE SUPPORT
|
||||||
M: Andreas Herrmann <herrmann.der.user@googlemail.com>
|
M: Borislav Petkov <bp@alien8.de>
|
||||||
L: amd64-microcode@amd64.org
|
|
||||||
S: Maintained
|
S: Maintained
|
||||||
F: arch/x86/kernel/cpu/microcode/amd*
|
F: arch/x86/kernel/cpu/microcode/amd*
|
||||||
|
|
||||||
|
@ -1763,7 +1762,7 @@ S: Supported
|
||||||
F: drivers/tty/serial/atmel_serial.c
|
F: drivers/tty/serial/atmel_serial.c
|
||||||
|
|
||||||
ATMEL Audio ALSA driver
|
ATMEL Audio ALSA driver
|
||||||
M: Bo Shen <voice.shen@atmel.com>
|
M: Nicolas Ferre <nicolas.ferre@atmel.com>
|
||||||
L: alsa-devel@alsa-project.org (moderated for non-subscribers)
|
L: alsa-devel@alsa-project.org (moderated for non-subscribers)
|
||||||
S: Supported
|
S: Supported
|
||||||
F: sound/soc/atmel
|
F: sound/soc/atmel
|
||||||
|
@ -5095,7 +5094,7 @@ S: Supported
|
||||||
F: drivers/platform/x86/intel_menlow.c
|
F: drivers/platform/x86/intel_menlow.c
|
||||||
|
|
||||||
INTEL IA32 MICROCODE UPDATE SUPPORT
|
INTEL IA32 MICROCODE UPDATE SUPPORT
|
||||||
M: Tigran Aivazian <tigran@aivazian.fsnet.co.uk>
|
M: Borislav Petkov <bp@alien8.de>
|
||||||
S: Maintained
|
S: Maintained
|
||||||
F: arch/x86/kernel/cpu/microcode/core*
|
F: arch/x86/kernel/cpu/microcode/core*
|
||||||
F: arch/x86/kernel/cpu/microcode/intel*
|
F: arch/x86/kernel/cpu/microcode/intel*
|
||||||
|
@ -5136,22 +5135,21 @@ M: Deepak Saxena <dsaxena@plexity.net>
|
||||||
S: Maintained
|
S: Maintained
|
||||||
F: drivers/char/hw_random/ixp4xx-rng.c
|
F: drivers/char/hw_random/ixp4xx-rng.c
|
||||||
|
|
||||||
INTEL ETHERNET DRIVERS (e100/e1000/e1000e/fm10k/igb/igbvf/ixgb/ixgbe/ixgbevf/i40e/i40evf)
|
INTEL ETHERNET DRIVERS
|
||||||
M: Jeff Kirsher <jeffrey.t.kirsher@intel.com>
|
M: Jeff Kirsher <jeffrey.t.kirsher@intel.com>
|
||||||
M: Jesse Brandeburg <jesse.brandeburg@intel.com>
|
R: Jesse Brandeburg <jesse.brandeburg@intel.com>
|
||||||
M: Bruce Allan <bruce.w.allan@intel.com>
|
R: Shannon Nelson <shannon.nelson@intel.com>
|
||||||
M: Carolyn Wyborny <carolyn.wyborny@intel.com>
|
R: Carolyn Wyborny <carolyn.wyborny@intel.com>
|
||||||
M: Don Skidmore <donald.c.skidmore@intel.com>
|
R: Don Skidmore <donald.c.skidmore@intel.com>
|
||||||
M: Greg Rose <gregory.v.rose@intel.com>
|
R: Matthew Vick <matthew.vick@intel.com>
|
||||||
M: Matthew Vick <matthew.vick@intel.com>
|
R: John Ronciak <john.ronciak@intel.com>
|
||||||
M: John Ronciak <john.ronciak@intel.com>
|
R: Mitch Williams <mitch.a.williams@intel.com>
|
||||||
M: Mitch Williams <mitch.a.williams@intel.com>
|
L: intel-wired-lan@lists.osuosl.org
|
||||||
M: Linux NICS <linux.nics@intel.com>
|
|
||||||
L: e1000-devel@lists.sourceforge.net
|
|
||||||
W: http://www.intel.com/support/feedback.htm
|
W: http://www.intel.com/support/feedback.htm
|
||||||
W: http://e1000.sourceforge.net/
|
W: http://e1000.sourceforge.net/
|
||||||
T: git git://git.kernel.org/pub/scm/linux/kernel/git/jkirsher/net.git
|
Q: http://patchwork.ozlabs.org/project/intel-wired-lan/list/
|
||||||
T: git git://git.kernel.org/pub/scm/linux/kernel/git/jkirsher/net-next.git
|
T: git git://git.kernel.org/pub/scm/linux/kernel/git/jkirsher/net-queue.git
|
||||||
|
T: git git://git.kernel.org/pub/scm/linux/kernel/git/jkirsher/next-queue.git
|
||||||
S: Supported
|
S: Supported
|
||||||
F: Documentation/networking/e100.txt
|
F: Documentation/networking/e100.txt
|
||||||
F: Documentation/networking/e1000.txt
|
F: Documentation/networking/e1000.txt
|
||||||
|
@ -5273,6 +5271,13 @@ F: drivers/char/ipmi/
|
||||||
F: include/linux/ipmi*
|
F: include/linux/ipmi*
|
||||||
F: include/uapi/linux/ipmi*
|
F: include/uapi/linux/ipmi*
|
||||||
|
|
||||||
|
QCOM AUDIO (ASoC) DRIVERS
|
||||||
|
M: Patrick Lai <plai@codeaurora.org>
|
||||||
|
M: Banajit Goswami <bgoswami@codeaurora.org>
|
||||||
|
L: alsa-devel@alsa-project.org (moderated for non-subscribers)
|
||||||
|
S: Supported
|
||||||
|
F: sound/soc/qcom/
|
||||||
|
|
||||||
IPS SCSI RAID DRIVER
|
IPS SCSI RAID DRIVER
|
||||||
M: Adaptec OEM Raid Solutions <aacraid@adaptec.com>
|
M: Adaptec OEM Raid Solutions <aacraid@adaptec.com>
|
||||||
L: linux-scsi@vger.kernel.org
|
L: linux-scsi@vger.kernel.org
|
||||||
|
|
2
Makefile
2
Makefile
|
@ -1,7 +1,7 @@
|
||||||
VERSION = 4
|
VERSION = 4
|
||||||
PATCHLEVEL = 0
|
PATCHLEVEL = 0
|
||||||
SUBLEVEL = 0
|
SUBLEVEL = 0
|
||||||
EXTRAVERSION = -rc6
|
EXTRAVERSION = -rc7
|
||||||
NAME = Hurr durr I'ma sheep
|
NAME = Hurr durr I'ma sheep
|
||||||
|
|
||||||
# *DOCUMENTATION*
|
# *DOCUMENTATION*
|
||||||
|
|
|
@ -1015,7 +1015,6 @@ static struct asoc_simple_card_info fsi_wm8978_info = {
|
||||||
.platform = "sh_fsi2",
|
.platform = "sh_fsi2",
|
||||||
.daifmt = SND_SOC_DAIFMT_I2S | SND_SOC_DAIFMT_CBM_CFM,
|
.daifmt = SND_SOC_DAIFMT_I2S | SND_SOC_DAIFMT_CBM_CFM,
|
||||||
.cpu_dai = {
|
.cpu_dai = {
|
||||||
.fmt = SND_SOC_DAIFMT_IB_NF,
|
|
||||||
.name = "fsia-dai",
|
.name = "fsia-dai",
|
||||||
},
|
},
|
||||||
.codec_dai = {
|
.codec_dai = {
|
||||||
|
@ -1040,9 +1039,9 @@ static struct asoc_simple_card_info fsi2_hdmi_info = {
|
||||||
.card = "FSI2B-HDMI",
|
.card = "FSI2B-HDMI",
|
||||||
.codec = "sh-mobile-hdmi",
|
.codec = "sh-mobile-hdmi",
|
||||||
.platform = "sh_fsi2",
|
.platform = "sh_fsi2",
|
||||||
|
.daifmt = SND_SOC_DAIFMT_CBS_CFS,
|
||||||
.cpu_dai = {
|
.cpu_dai = {
|
||||||
.name = "fsib-dai",
|
.name = "fsib-dai",
|
||||||
.fmt = SND_SOC_DAIFMT_CBS_CFS,
|
|
||||||
},
|
},
|
||||||
.codec_dai = {
|
.codec_dai = {
|
||||||
.name = "sh_mobile_hdmi-hifi",
|
.name = "sh_mobile_hdmi-hifi",
|
||||||
|
|
|
@ -55,7 +55,7 @@ static inline cpumask_t cpu_thread_mask_to_cores(const struct cpumask *threads)
|
||||||
|
|
||||||
static inline int cpu_nr_cores(void)
|
static inline int cpu_nr_cores(void)
|
||||||
{
|
{
|
||||||
return NR_CPUS >> threads_shift;
|
return nr_cpu_ids >> threads_shift;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline cpumask_t cpu_online_cores_map(void)
|
static inline cpumask_t cpu_online_cores_map(void)
|
||||||
|
|
|
@ -212,11 +212,11 @@ static struct event_constraint intel_hsw_event_constraints[] = {
|
||||||
INTEL_UEVENT_CONSTRAINT(0x01c0, 0x2), /* INST_RETIRED.PREC_DIST */
|
INTEL_UEVENT_CONSTRAINT(0x01c0, 0x2), /* INST_RETIRED.PREC_DIST */
|
||||||
INTEL_EVENT_CONSTRAINT(0xcd, 0x8), /* MEM_TRANS_RETIRED.LOAD_LATENCY */
|
INTEL_EVENT_CONSTRAINT(0xcd, 0x8), /* MEM_TRANS_RETIRED.LOAD_LATENCY */
|
||||||
/* CYCLE_ACTIVITY.CYCLES_L1D_PENDING */
|
/* CYCLE_ACTIVITY.CYCLES_L1D_PENDING */
|
||||||
INTEL_EVENT_CONSTRAINT(0x08a3, 0x4),
|
INTEL_UEVENT_CONSTRAINT(0x08a3, 0x4),
|
||||||
/* CYCLE_ACTIVITY.STALLS_L1D_PENDING */
|
/* CYCLE_ACTIVITY.STALLS_L1D_PENDING */
|
||||||
INTEL_EVENT_CONSTRAINT(0x0ca3, 0x4),
|
INTEL_UEVENT_CONSTRAINT(0x0ca3, 0x4),
|
||||||
/* CYCLE_ACTIVITY.CYCLES_NO_EXECUTE */
|
/* CYCLE_ACTIVITY.CYCLES_NO_EXECUTE */
|
||||||
INTEL_EVENT_CONSTRAINT(0x04a3, 0xf),
|
INTEL_UEVENT_CONSTRAINT(0x04a3, 0xf),
|
||||||
EVENT_CONSTRAINT_END
|
EVENT_CONSTRAINT_END
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -1649,11 +1649,11 @@ intel_get_event_constraints(struct cpu_hw_events *cpuc, struct perf_event *event
|
||||||
if (c)
|
if (c)
|
||||||
return c;
|
return c;
|
||||||
|
|
||||||
c = intel_pebs_constraints(event);
|
c = intel_shared_regs_constraints(cpuc, event);
|
||||||
if (c)
|
if (c)
|
||||||
return c;
|
return c;
|
||||||
|
|
||||||
c = intel_shared_regs_constraints(cpuc, event);
|
c = intel_pebs_constraints(event);
|
||||||
if (c)
|
if (c)
|
||||||
return c;
|
return c;
|
||||||
|
|
||||||
|
|
|
@ -799,7 +799,21 @@ retint_swapgs: /* return to user-space */
|
||||||
cmpq %r11,(EFLAGS-ARGOFFSET)(%rsp) /* R11 == RFLAGS */
|
cmpq %r11,(EFLAGS-ARGOFFSET)(%rsp) /* R11 == RFLAGS */
|
||||||
jne opportunistic_sysret_failed
|
jne opportunistic_sysret_failed
|
||||||
|
|
||||||
testq $X86_EFLAGS_RF,%r11 /* sysret can't restore RF */
|
/*
|
||||||
|
* SYSRET can't restore RF. SYSRET can restore TF, but unlike IRET,
|
||||||
|
* restoring TF results in a trap from userspace immediately after
|
||||||
|
* SYSRET. This would cause an infinite loop whenever #DB happens
|
||||||
|
* with register state that satisfies the opportunistic SYSRET
|
||||||
|
* conditions. For example, single-stepping this user code:
|
||||||
|
*
|
||||||
|
* movq $stuck_here,%rcx
|
||||||
|
* pushfq
|
||||||
|
* popq %r11
|
||||||
|
* stuck_here:
|
||||||
|
*
|
||||||
|
* would never get past 'stuck_here'.
|
||||||
|
*/
|
||||||
|
testq $(X86_EFLAGS_RF|X86_EFLAGS_TF), %r11
|
||||||
jnz opportunistic_sysret_failed
|
jnz opportunistic_sysret_failed
|
||||||
|
|
||||||
/* nothing to check for RSP */
|
/* nothing to check for RSP */
|
||||||
|
|
|
@ -72,7 +72,7 @@ struct dbg_reg_def_t dbg_reg_def[DBG_MAX_REG_NUM] =
|
||||||
{ "bx", 8, offsetof(struct pt_regs, bx) },
|
{ "bx", 8, offsetof(struct pt_regs, bx) },
|
||||||
{ "cx", 8, offsetof(struct pt_regs, cx) },
|
{ "cx", 8, offsetof(struct pt_regs, cx) },
|
||||||
{ "dx", 8, offsetof(struct pt_regs, dx) },
|
{ "dx", 8, offsetof(struct pt_regs, dx) },
|
||||||
{ "si", 8, offsetof(struct pt_regs, dx) },
|
{ "si", 8, offsetof(struct pt_regs, si) },
|
||||||
{ "di", 8, offsetof(struct pt_regs, di) },
|
{ "di", 8, offsetof(struct pt_regs, di) },
|
||||||
{ "bp", 8, offsetof(struct pt_regs, bp) },
|
{ "bp", 8, offsetof(struct pt_regs, bp) },
|
||||||
{ "sp", 8, offsetof(struct pt_regs, sp) },
|
{ "sp", 8, offsetof(struct pt_regs, sp) },
|
||||||
|
|
|
@ -183,6 +183,16 @@ static struct dmi_system_id __initdata reboot_dmi_table[] = {
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
|
/* ASRock */
|
||||||
|
{ /* Handle problems with rebooting on ASRock Q1900DC-ITX */
|
||||||
|
.callback = set_pci_reboot,
|
||||||
|
.ident = "ASRock Q1900DC-ITX",
|
||||||
|
.matches = {
|
||||||
|
DMI_MATCH(DMI_BOARD_VENDOR, "ASRock"),
|
||||||
|
DMI_MATCH(DMI_BOARD_NAME, "Q1900DC-ITX"),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
|
||||||
/* ASUS */
|
/* ASUS */
|
||||||
{ /* Handle problems with rebooting on ASUS P4S800 */
|
{ /* Handle problems with rebooting on ASUS P4S800 */
|
||||||
.callback = set_bios_reboot,
|
.callback = set_bios_reboot,
|
||||||
|
|
|
@ -91,6 +91,12 @@ EXPORT_SYMBOL_GPL(xen_p2m_size);
|
||||||
unsigned long xen_max_p2m_pfn __read_mostly;
|
unsigned long xen_max_p2m_pfn __read_mostly;
|
||||||
EXPORT_SYMBOL_GPL(xen_max_p2m_pfn);
|
EXPORT_SYMBOL_GPL(xen_max_p2m_pfn);
|
||||||
|
|
||||||
|
#ifdef CONFIG_XEN_BALLOON_MEMORY_HOTPLUG_LIMIT
|
||||||
|
#define P2M_LIMIT CONFIG_XEN_BALLOON_MEMORY_HOTPLUG_LIMIT
|
||||||
|
#else
|
||||||
|
#define P2M_LIMIT 0
|
||||||
|
#endif
|
||||||
|
|
||||||
static DEFINE_SPINLOCK(p2m_update_lock);
|
static DEFINE_SPINLOCK(p2m_update_lock);
|
||||||
|
|
||||||
static unsigned long *p2m_mid_missing_mfn;
|
static unsigned long *p2m_mid_missing_mfn;
|
||||||
|
@ -385,9 +391,11 @@ static void __init xen_rebuild_p2m_list(unsigned long *p2m)
|
||||||
void __init xen_vmalloc_p2m_tree(void)
|
void __init xen_vmalloc_p2m_tree(void)
|
||||||
{
|
{
|
||||||
static struct vm_struct vm;
|
static struct vm_struct vm;
|
||||||
|
unsigned long p2m_limit;
|
||||||
|
|
||||||
|
p2m_limit = (phys_addr_t)P2M_LIMIT * 1024 * 1024 * 1024 / PAGE_SIZE;
|
||||||
vm.flags = VM_ALLOC;
|
vm.flags = VM_ALLOC;
|
||||||
vm.size = ALIGN(sizeof(unsigned long) * xen_max_p2m_pfn,
|
vm.size = ALIGN(sizeof(unsigned long) * max(xen_max_p2m_pfn, p2m_limit),
|
||||||
PMD_SIZE * PMDS_PER_MID_PAGE);
|
PMD_SIZE * PMDS_PER_MID_PAGE);
|
||||||
vm_area_register_early(&vm, PMD_SIZE * PMDS_PER_MID_PAGE);
|
vm_area_register_early(&vm, PMD_SIZE * PMDS_PER_MID_PAGE);
|
||||||
pr_notice("p2m virtual area at %p, size is %lx\n", vm.addr, vm.size);
|
pr_notice("p2m virtual area at %p, size is %lx\n", vm.addr, vm.size);
|
||||||
|
|
|
@ -585,7 +585,7 @@ int blk_stack_limits(struct queue_limits *t, struct queue_limits *b,
|
||||||
b->physical_block_size);
|
b->physical_block_size);
|
||||||
|
|
||||||
t->io_min = max(t->io_min, b->io_min);
|
t->io_min = max(t->io_min, b->io_min);
|
||||||
t->io_opt = lcm(t->io_opt, b->io_opt);
|
t->io_opt = lcm_not_zero(t->io_opt, b->io_opt);
|
||||||
|
|
||||||
t->cluster &= b->cluster;
|
t->cluster &= b->cluster;
|
||||||
t->discard_zeroes_data &= b->discard_zeroes_data;
|
t->discard_zeroes_data &= b->discard_zeroes_data;
|
||||||
|
@ -616,7 +616,7 @@ int blk_stack_limits(struct queue_limits *t, struct queue_limits *b,
|
||||||
b->raid_partial_stripes_expensive);
|
b->raid_partial_stripes_expensive);
|
||||||
|
|
||||||
/* Find lowest common alignment_offset */
|
/* Find lowest common alignment_offset */
|
||||||
t->alignment_offset = lcm(t->alignment_offset, alignment)
|
t->alignment_offset = lcm_not_zero(t->alignment_offset, alignment)
|
||||||
% max(t->physical_block_size, t->io_min);
|
% max(t->physical_block_size, t->io_min);
|
||||||
|
|
||||||
/* Verify that new alignment_offset is on a logical block boundary */
|
/* Verify that new alignment_offset is on a logical block boundary */
|
||||||
|
@ -643,7 +643,7 @@ int blk_stack_limits(struct queue_limits *t, struct queue_limits *b,
|
||||||
b->max_discard_sectors);
|
b->max_discard_sectors);
|
||||||
t->discard_granularity = max(t->discard_granularity,
|
t->discard_granularity = max(t->discard_granularity,
|
||||||
b->discard_granularity);
|
b->discard_granularity);
|
||||||
t->discard_alignment = lcm(t->discard_alignment, alignment) %
|
t->discard_alignment = lcm_not_zero(t->discard_alignment, alignment) %
|
||||||
t->discard_granularity;
|
t->discard_granularity;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -4204,9 +4204,18 @@ static const struct ata_blacklist_entry ata_device_blacklist [] = {
|
||||||
{ "PIONEER DVD-RW DVR-216D", NULL, ATA_HORKAGE_NOSETXFER },
|
{ "PIONEER DVD-RW DVR-216D", NULL, ATA_HORKAGE_NOSETXFER },
|
||||||
|
|
||||||
/* devices that don't properly handle queued TRIM commands */
|
/* devices that don't properly handle queued TRIM commands */
|
||||||
{ "Micron_M[56]*", NULL, ATA_HORKAGE_NO_NCQ_TRIM |
|
{ "Micron_M500*", NULL, ATA_HORKAGE_NO_NCQ_TRIM |
|
||||||
|
ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
||||||
|
{ "Crucial_CT*M500*", NULL, ATA_HORKAGE_NO_NCQ_TRIM |
|
||||||
|
ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
||||||
|
{ "Micron_M5[15]0*", "MU01", ATA_HORKAGE_NO_NCQ_TRIM |
|
||||||
|
ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
||||||
|
{ "Crucial_CT*M550*", "MU01", ATA_HORKAGE_NO_NCQ_TRIM |
|
||||||
|
ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
||||||
|
{ "Crucial_CT*MX100*", "MU01", ATA_HORKAGE_NO_NCQ_TRIM |
|
||||||
|
ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
||||||
|
{ "Samsung SSD 850 PRO*", NULL, ATA_HORKAGE_NO_NCQ_TRIM |
|
||||||
ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
||||||
{ "Crucial_CT*SSD*", NULL, ATA_HORKAGE_NO_NCQ_TRIM, },
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* As defined, the DRAT (Deterministic Read After Trim) and RZAT
|
* As defined, the DRAT (Deterministic Read After Trim) and RZAT
|
||||||
|
@ -4226,6 +4235,8 @@ static const struct ata_blacklist_entry ata_device_blacklist [] = {
|
||||||
*/
|
*/
|
||||||
{ "INTEL*SSDSC2MH*", NULL, 0, },
|
{ "INTEL*SSDSC2MH*", NULL, 0, },
|
||||||
|
|
||||||
|
{ "Micron*", NULL, ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
||||||
|
{ "Crucial*", NULL, ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
||||||
{ "INTEL*SSD*", NULL, ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
{ "INTEL*SSD*", NULL, ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
||||||
{ "SSD*INTEL*", NULL, ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
{ "SSD*INTEL*", NULL, ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
||||||
{ "Samsung*SSD*", NULL, ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
{ "Samsung*SSD*", NULL, ATA_HORKAGE_ZERO_AFTER_TRIM, },
|
||||||
|
|
|
@ -475,6 +475,7 @@ static int bcm2835_dma_terminate_all(struct dma_chan *chan)
|
||||||
* c->desc is NULL and exit.)
|
* c->desc is NULL and exit.)
|
||||||
*/
|
*/
|
||||||
if (c->desc) {
|
if (c->desc) {
|
||||||
|
bcm2835_dma_desc_free(&c->desc->vd);
|
||||||
c->desc = NULL;
|
c->desc = NULL;
|
||||||
bcm2835_dma_abort(c->chan_base);
|
bcm2835_dma_abort(c->chan_base);
|
||||||
|
|
||||||
|
|
|
@ -511,6 +511,9 @@ static void jz4740_dma_desc_free(struct virt_dma_desc *vdesc)
|
||||||
kfree(container_of(vdesc, struct jz4740_dma_desc, vdesc));
|
kfree(container_of(vdesc, struct jz4740_dma_desc, vdesc));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#define JZ4740_DMA_BUSWIDTHS (BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) | \
|
||||||
|
BIT(DMA_SLAVE_BUSWIDTH_2_BYTES) | BIT(DMA_SLAVE_BUSWIDTH_4_BYTES))
|
||||||
|
|
||||||
static int jz4740_dma_probe(struct platform_device *pdev)
|
static int jz4740_dma_probe(struct platform_device *pdev)
|
||||||
{
|
{
|
||||||
struct jz4740_dmaengine_chan *chan;
|
struct jz4740_dmaengine_chan *chan;
|
||||||
|
@ -548,6 +551,10 @@ static int jz4740_dma_probe(struct platform_device *pdev)
|
||||||
dd->device_prep_dma_cyclic = jz4740_dma_prep_dma_cyclic;
|
dd->device_prep_dma_cyclic = jz4740_dma_prep_dma_cyclic;
|
||||||
dd->device_config = jz4740_dma_slave_config;
|
dd->device_config = jz4740_dma_slave_config;
|
||||||
dd->device_terminate_all = jz4740_dma_terminate_all;
|
dd->device_terminate_all = jz4740_dma_terminate_all;
|
||||||
|
dd->src_addr_widths = JZ4740_DMA_BUSWIDTHS;
|
||||||
|
dd->dst_addr_widths = JZ4740_DMA_BUSWIDTHS;
|
||||||
|
dd->directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
|
||||||
|
dd->residue_granularity = DMA_RESIDUE_GRANULARITY_BURST;
|
||||||
dd->dev = &pdev->dev;
|
dd->dev = &pdev->dev;
|
||||||
INIT_LIST_HEAD(&dd->channels);
|
INIT_LIST_HEAD(&dd->channels);
|
||||||
|
|
||||||
|
|
|
@ -260,6 +260,13 @@ static int edma_terminate_all(struct dma_chan *chan)
|
||||||
*/
|
*/
|
||||||
if (echan->edesc) {
|
if (echan->edesc) {
|
||||||
int cyclic = echan->edesc->cyclic;
|
int cyclic = echan->edesc->cyclic;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* free the running request descriptor
|
||||||
|
* since it is not in any of the vdesc lists
|
||||||
|
*/
|
||||||
|
edma_desc_free(&echan->edesc->vdesc);
|
||||||
|
|
||||||
echan->edesc = NULL;
|
echan->edesc = NULL;
|
||||||
edma_stop(echan->ch_num);
|
edma_stop(echan->ch_num);
|
||||||
/* Move the cyclic channel back to default queue */
|
/* Move the cyclic channel back to default queue */
|
||||||
|
|
|
@ -193,8 +193,10 @@ static int moxart_terminate_all(struct dma_chan *chan)
|
||||||
|
|
||||||
spin_lock_irqsave(&ch->vc.lock, flags);
|
spin_lock_irqsave(&ch->vc.lock, flags);
|
||||||
|
|
||||||
if (ch->desc)
|
if (ch->desc) {
|
||||||
|
moxart_dma_desc_free(&ch->desc->vd);
|
||||||
ch->desc = NULL;
|
ch->desc = NULL;
|
||||||
|
}
|
||||||
|
|
||||||
ctrl = readl(ch->base + REG_OFF_CTRL);
|
ctrl = readl(ch->base + REG_OFF_CTRL);
|
||||||
ctrl &= ~(APB_DMA_ENABLE | APB_DMA_FIN_INT_EN | APB_DMA_ERR_INT_EN);
|
ctrl &= ~(APB_DMA_ENABLE | APB_DMA_FIN_INT_EN | APB_DMA_ERR_INT_EN);
|
||||||
|
|
|
@ -194,6 +194,7 @@ struct dma_chan *of_dma_request_slave_channel(struct device_node *np,
|
||||||
|
|
||||||
return ERR_PTR(ret_no_channel);
|
return ERR_PTR(ret_no_channel);
|
||||||
}
|
}
|
||||||
|
EXPORT_SYMBOL_GPL(of_dma_request_slave_channel);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* of_dma_simple_xlate - Simple DMA engine translation function
|
* of_dma_simple_xlate - Simple DMA engine translation function
|
||||||
|
|
|
@ -981,6 +981,7 @@ static int omap_dma_terminate_all(struct dma_chan *chan)
|
||||||
* c->desc is NULL and exit.)
|
* c->desc is NULL and exit.)
|
||||||
*/
|
*/
|
||||||
if (c->desc) {
|
if (c->desc) {
|
||||||
|
omap_dma_desc_free(&c->desc->vd);
|
||||||
c->desc = NULL;
|
c->desc = NULL;
|
||||||
/* Avoid stopping the dma twice */
|
/* Avoid stopping the dma twice */
|
||||||
if (!c->paused)
|
if (!c->paused)
|
||||||
|
|
|
@ -86,10 +86,13 @@ static void dmi_table(u8 *buf, u32 len, int num,
|
||||||
int i = 0;
|
int i = 0;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Stop when we see all the items the table claimed to have
|
* Stop when we have seen all the items the table claimed to have
|
||||||
* OR we run off the end of the table (also happens)
|
* (SMBIOS < 3.0 only) OR we reach an end-of-table marker OR we run
|
||||||
|
* off the end of the table (should never happen but sometimes does
|
||||||
|
* on bogus implementations.)
|
||||||
*/
|
*/
|
||||||
while ((i < num) && (data - buf + sizeof(struct dmi_header)) <= len) {
|
while ((!num || i < num) &&
|
||||||
|
(data - buf + sizeof(struct dmi_header)) <= len) {
|
||||||
const struct dmi_header *dm = (const struct dmi_header *)data;
|
const struct dmi_header *dm = (const struct dmi_header *)data;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -529,21 +532,10 @@ static int __init dmi_smbios3_present(const u8 *buf)
|
||||||
if (memcmp(buf, "_SM3_", 5) == 0 &&
|
if (memcmp(buf, "_SM3_", 5) == 0 &&
|
||||||
buf[6] < 32 && dmi_checksum(buf, buf[6])) {
|
buf[6] < 32 && dmi_checksum(buf, buf[6])) {
|
||||||
dmi_ver = get_unaligned_be16(buf + 7);
|
dmi_ver = get_unaligned_be16(buf + 7);
|
||||||
|
dmi_num = 0; /* No longer specified */
|
||||||
dmi_len = get_unaligned_le32(buf + 12);
|
dmi_len = get_unaligned_le32(buf + 12);
|
||||||
dmi_base = get_unaligned_le64(buf + 16);
|
dmi_base = get_unaligned_le64(buf + 16);
|
||||||
|
|
||||||
/*
|
|
||||||
* The 64-bit SMBIOS 3.0 entry point no longer has a field
|
|
||||||
* containing the number of structures present in the table.
|
|
||||||
* Instead, it defines the table size as a maximum size, and
|
|
||||||
* relies on the end-of-table structure type (#127) to be used
|
|
||||||
* to signal the end of the table.
|
|
||||||
* So let's define dmi_num as an upper bound as well: each
|
|
||||||
* structure has a 4 byte header, so dmi_len / 4 is an upper
|
|
||||||
* bound for the number of structures in the table.
|
|
||||||
*/
|
|
||||||
dmi_num = dmi_len / 4;
|
|
||||||
|
|
||||||
if (dmi_walk_early(dmi_decode) == 0) {
|
if (dmi_walk_early(dmi_decode) == 0) {
|
||||||
pr_info("SMBIOS %d.%d present.\n",
|
pr_info("SMBIOS %d.%d present.\n",
|
||||||
dmi_ver >> 8, dmi_ver & 0xFF);
|
dmi_ver >> 8, dmi_ver & 0xFF);
|
||||||
|
|
|
@ -334,7 +334,7 @@ static struct irq_domain_ops mpc8xxx_gpio_irq_ops = {
|
||||||
.xlate = irq_domain_xlate_twocell,
|
.xlate = irq_domain_xlate_twocell,
|
||||||
};
|
};
|
||||||
|
|
||||||
static struct of_device_id mpc8xxx_gpio_ids[] __initdata = {
|
static struct of_device_id mpc8xxx_gpio_ids[] = {
|
||||||
{ .compatible = "fsl,mpc8349-gpio", },
|
{ .compatible = "fsl,mpc8349-gpio", },
|
||||||
{ .compatible = "fsl,mpc8572-gpio", },
|
{ .compatible = "fsl,mpc8572-gpio", },
|
||||||
{ .compatible = "fsl,mpc8610-gpio", },
|
{ .compatible = "fsl,mpc8610-gpio", },
|
||||||
|
|
|
@ -219,7 +219,7 @@ static int syscon_gpio_probe(struct platform_device *pdev)
|
||||||
ret = of_property_read_u32_index(np, "gpio,syscon-dev", 2,
|
ret = of_property_read_u32_index(np, "gpio,syscon-dev", 2,
|
||||||
&priv->dir_reg_offset);
|
&priv->dir_reg_offset);
|
||||||
if (ret)
|
if (ret)
|
||||||
dev_err(dev, "can't read the dir register offset!\n");
|
dev_dbg(dev, "can't read the dir register offset!\n");
|
||||||
|
|
||||||
priv->dir_reg_offset <<= 3;
|
priv->dir_reg_offset <<= 3;
|
||||||
}
|
}
|
||||||
|
|
|
@ -201,6 +201,10 @@ static acpi_status acpi_gpiochip_request_interrupt(struct acpi_resource *ares,
|
||||||
if (!handler)
|
if (!handler)
|
||||||
return AE_BAD_PARAMETER;
|
return AE_BAD_PARAMETER;
|
||||||
|
|
||||||
|
pin = acpi_gpiochip_pin_to_gpio_offset(chip, pin);
|
||||||
|
if (pin < 0)
|
||||||
|
return AE_BAD_PARAMETER;
|
||||||
|
|
||||||
desc = gpiochip_request_own_desc(chip, pin, "ACPI:Event");
|
desc = gpiochip_request_own_desc(chip, pin, "ACPI:Event");
|
||||||
if (IS_ERR(desc)) {
|
if (IS_ERR(desc)) {
|
||||||
dev_err(chip->dev, "Failed to request GPIO\n");
|
dev_err(chip->dev, "Failed to request GPIO\n");
|
||||||
|
@ -551,6 +555,12 @@ acpi_gpio_adr_space_handler(u32 function, acpi_physical_address address,
|
||||||
struct gpio_desc *desc;
|
struct gpio_desc *desc;
|
||||||
bool found;
|
bool found;
|
||||||
|
|
||||||
|
pin = acpi_gpiochip_pin_to_gpio_offset(chip, pin);
|
||||||
|
if (pin < 0) {
|
||||||
|
status = AE_BAD_PARAMETER;
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
mutex_lock(&achip->conn_lock);
|
mutex_lock(&achip->conn_lock);
|
||||||
|
|
||||||
found = false;
|
found = false;
|
||||||
|
|
|
@ -287,6 +287,7 @@ int drm_load_edid_firmware(struct drm_connector *connector)
|
||||||
|
|
||||||
drm_mode_connector_update_edid_property(connector, edid);
|
drm_mode_connector_update_edid_property(connector, edid);
|
||||||
ret = drm_add_edid_modes(connector, edid);
|
ret = drm_add_edid_modes(connector, edid);
|
||||||
|
drm_edid_to_eld(connector, edid);
|
||||||
kfree(edid);
|
kfree(edid);
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
|
|
|
@ -174,6 +174,7 @@ static int drm_helper_probe_single_connector_modes_merge_bits(struct drm_connect
|
||||||
struct edid *edid = (struct edid *) connector->edid_blob_ptr->data;
|
struct edid *edid = (struct edid *) connector->edid_blob_ptr->data;
|
||||||
|
|
||||||
count = drm_add_edid_modes(connector, edid);
|
count = drm_add_edid_modes(connector, edid);
|
||||||
|
drm_edid_to_eld(connector, edid);
|
||||||
} else
|
} else
|
||||||
count = (*connector_funcs->get_modes)(connector);
|
count = (*connector_funcs->get_modes)(connector);
|
||||||
}
|
}
|
||||||
|
|
|
@ -147,6 +147,7 @@ struct fimd_win_data {
|
||||||
unsigned int ovl_height;
|
unsigned int ovl_height;
|
||||||
unsigned int fb_width;
|
unsigned int fb_width;
|
||||||
unsigned int fb_height;
|
unsigned int fb_height;
|
||||||
|
unsigned int fb_pitch;
|
||||||
unsigned int bpp;
|
unsigned int bpp;
|
||||||
unsigned int pixel_format;
|
unsigned int pixel_format;
|
||||||
dma_addr_t dma_addr;
|
dma_addr_t dma_addr;
|
||||||
|
@ -532,13 +533,14 @@ static void fimd_win_mode_set(struct exynos_drm_crtc *crtc,
|
||||||
win_data->offset_y = plane->crtc_y;
|
win_data->offset_y = plane->crtc_y;
|
||||||
win_data->ovl_width = plane->crtc_width;
|
win_data->ovl_width = plane->crtc_width;
|
||||||
win_data->ovl_height = plane->crtc_height;
|
win_data->ovl_height = plane->crtc_height;
|
||||||
|
win_data->fb_pitch = plane->pitch;
|
||||||
win_data->fb_width = plane->fb_width;
|
win_data->fb_width = plane->fb_width;
|
||||||
win_data->fb_height = plane->fb_height;
|
win_data->fb_height = plane->fb_height;
|
||||||
win_data->dma_addr = plane->dma_addr[0] + offset;
|
win_data->dma_addr = plane->dma_addr[0] + offset;
|
||||||
win_data->bpp = plane->bpp;
|
win_data->bpp = plane->bpp;
|
||||||
win_data->pixel_format = plane->pixel_format;
|
win_data->pixel_format = plane->pixel_format;
|
||||||
win_data->buf_offsize = (plane->fb_width - plane->crtc_width) *
|
win_data->buf_offsize =
|
||||||
(plane->bpp >> 3);
|
plane->pitch - (plane->crtc_width * (plane->bpp >> 3));
|
||||||
win_data->line_size = plane->crtc_width * (plane->bpp >> 3);
|
win_data->line_size = plane->crtc_width * (plane->bpp >> 3);
|
||||||
|
|
||||||
DRM_DEBUG_KMS("offset_x = %d, offset_y = %d\n",
|
DRM_DEBUG_KMS("offset_x = %d, offset_y = %d\n",
|
||||||
|
@ -704,7 +706,7 @@ static void fimd_win_commit(struct exynos_drm_crtc *crtc, int zpos)
|
||||||
writel(val, ctx->regs + VIDWx_BUF_START(win, 0));
|
writel(val, ctx->regs + VIDWx_BUF_START(win, 0));
|
||||||
|
|
||||||
/* buffer end address */
|
/* buffer end address */
|
||||||
size = win_data->fb_width * win_data->ovl_height * (win_data->bpp >> 3);
|
size = win_data->fb_pitch * win_data->ovl_height * (win_data->bpp >> 3);
|
||||||
val = (unsigned long)(win_data->dma_addr + size);
|
val = (unsigned long)(win_data->dma_addr + size);
|
||||||
writel(val, ctx->regs + VIDWx_BUF_END(win, 0));
|
writel(val, ctx->regs + VIDWx_BUF_END(win, 0));
|
||||||
|
|
||||||
|
|
|
@ -55,6 +55,7 @@ struct hdmi_win_data {
|
||||||
unsigned int fb_x;
|
unsigned int fb_x;
|
||||||
unsigned int fb_y;
|
unsigned int fb_y;
|
||||||
unsigned int fb_width;
|
unsigned int fb_width;
|
||||||
|
unsigned int fb_pitch;
|
||||||
unsigned int fb_height;
|
unsigned int fb_height;
|
||||||
unsigned int src_width;
|
unsigned int src_width;
|
||||||
unsigned int src_height;
|
unsigned int src_height;
|
||||||
|
@ -438,7 +439,7 @@ static void vp_video_buffer(struct mixer_context *ctx, int win)
|
||||||
} else {
|
} else {
|
||||||
luma_addr[0] = win_data->dma_addr;
|
luma_addr[0] = win_data->dma_addr;
|
||||||
chroma_addr[0] = win_data->dma_addr
|
chroma_addr[0] = win_data->dma_addr
|
||||||
+ (win_data->fb_width * win_data->fb_height);
|
+ (win_data->fb_pitch * win_data->fb_height);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (win_data->scan_flags & DRM_MODE_FLAG_INTERLACE) {
|
if (win_data->scan_flags & DRM_MODE_FLAG_INTERLACE) {
|
||||||
|
@ -447,8 +448,8 @@ static void vp_video_buffer(struct mixer_context *ctx, int win)
|
||||||
luma_addr[1] = luma_addr[0] + 0x40;
|
luma_addr[1] = luma_addr[0] + 0x40;
|
||||||
chroma_addr[1] = chroma_addr[0] + 0x40;
|
chroma_addr[1] = chroma_addr[0] + 0x40;
|
||||||
} else {
|
} else {
|
||||||
luma_addr[1] = luma_addr[0] + win_data->fb_width;
|
luma_addr[1] = luma_addr[0] + win_data->fb_pitch;
|
||||||
chroma_addr[1] = chroma_addr[0] + win_data->fb_width;
|
chroma_addr[1] = chroma_addr[0] + win_data->fb_pitch;
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
ctx->interlace = false;
|
ctx->interlace = false;
|
||||||
|
@ -469,10 +470,10 @@ static void vp_video_buffer(struct mixer_context *ctx, int win)
|
||||||
vp_reg_writemask(res, VP_MODE, val, VP_MODE_FMT_MASK);
|
vp_reg_writemask(res, VP_MODE, val, VP_MODE_FMT_MASK);
|
||||||
|
|
||||||
/* setting size of input image */
|
/* setting size of input image */
|
||||||
vp_reg_write(res, VP_IMG_SIZE_Y, VP_IMG_HSIZE(win_data->fb_width) |
|
vp_reg_write(res, VP_IMG_SIZE_Y, VP_IMG_HSIZE(win_data->fb_pitch) |
|
||||||
VP_IMG_VSIZE(win_data->fb_height));
|
VP_IMG_VSIZE(win_data->fb_height));
|
||||||
/* chroma height has to reduced by 2 to avoid chroma distorions */
|
/* chroma height has to reduced by 2 to avoid chroma distorions */
|
||||||
vp_reg_write(res, VP_IMG_SIZE_C, VP_IMG_HSIZE(win_data->fb_width) |
|
vp_reg_write(res, VP_IMG_SIZE_C, VP_IMG_HSIZE(win_data->fb_pitch) |
|
||||||
VP_IMG_VSIZE(win_data->fb_height / 2));
|
VP_IMG_VSIZE(win_data->fb_height / 2));
|
||||||
|
|
||||||
vp_reg_write(res, VP_SRC_WIDTH, win_data->src_width);
|
vp_reg_write(res, VP_SRC_WIDTH, win_data->src_width);
|
||||||
|
@ -559,7 +560,7 @@ static void mixer_graph_buffer(struct mixer_context *ctx, int win)
|
||||||
/* converting dma address base and source offset */
|
/* converting dma address base and source offset */
|
||||||
dma_addr = win_data->dma_addr
|
dma_addr = win_data->dma_addr
|
||||||
+ (win_data->fb_x * win_data->bpp >> 3)
|
+ (win_data->fb_x * win_data->bpp >> 3)
|
||||||
+ (win_data->fb_y * win_data->fb_width * win_data->bpp >> 3);
|
+ (win_data->fb_y * win_data->fb_pitch);
|
||||||
src_x_offset = 0;
|
src_x_offset = 0;
|
||||||
src_y_offset = 0;
|
src_y_offset = 0;
|
||||||
|
|
||||||
|
@ -576,7 +577,8 @@ static void mixer_graph_buffer(struct mixer_context *ctx, int win)
|
||||||
MXR_GRP_CFG_FORMAT_VAL(fmt), MXR_GRP_CFG_FORMAT_MASK);
|
MXR_GRP_CFG_FORMAT_VAL(fmt), MXR_GRP_CFG_FORMAT_MASK);
|
||||||
|
|
||||||
/* setup geometry */
|
/* setup geometry */
|
||||||
mixer_reg_write(res, MXR_GRAPHIC_SPAN(win), win_data->fb_width);
|
mixer_reg_write(res, MXR_GRAPHIC_SPAN(win),
|
||||||
|
win_data->fb_pitch / (win_data->bpp >> 3));
|
||||||
|
|
||||||
/* setup display size */
|
/* setup display size */
|
||||||
if (ctx->mxr_ver == MXR_VER_128_0_0_184 &&
|
if (ctx->mxr_ver == MXR_VER_128_0_0_184 &&
|
||||||
|
@ -961,6 +963,7 @@ static void mixer_win_mode_set(struct exynos_drm_crtc *crtc,
|
||||||
win_data->fb_y = plane->fb_y;
|
win_data->fb_y = plane->fb_y;
|
||||||
win_data->fb_width = plane->fb_width;
|
win_data->fb_width = plane->fb_width;
|
||||||
win_data->fb_height = plane->fb_height;
|
win_data->fb_height = plane->fb_height;
|
||||||
|
win_data->fb_pitch = plane->pitch;
|
||||||
win_data->src_width = plane->src_width;
|
win_data->src_width = plane->src_width;
|
||||||
win_data->src_height = plane->src_height;
|
win_data->src_height = plane->src_height;
|
||||||
|
|
||||||
|
|
|
@ -1487,7 +1487,7 @@ i915_gem_do_execbuffer(struct drm_device *dev, void *data,
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (i915_needs_cmd_parser(ring)) {
|
if (i915_needs_cmd_parser(ring) && args->batch_len) {
|
||||||
batch_obj = i915_gem_execbuffer_parse(ring,
|
batch_obj = i915_gem_execbuffer_parse(ring,
|
||||||
&shadow_exec_entry,
|
&shadow_exec_entry,
|
||||||
eb,
|
eb,
|
||||||
|
|
|
@ -1322,7 +1322,7 @@ int intel_sprite_set_colorkey(struct drm_device *dev, void *data,
|
||||||
drm_modeset_lock_all(dev);
|
drm_modeset_lock_all(dev);
|
||||||
|
|
||||||
plane = drm_plane_find(dev, set->plane_id);
|
plane = drm_plane_find(dev, set->plane_id);
|
||||||
if (!plane) {
|
if (!plane || plane->type != DRM_PLANE_TYPE_OVERLAY) {
|
||||||
ret = -ENOENT;
|
ret = -ENOENT;
|
||||||
goto out_unlock;
|
goto out_unlock;
|
||||||
}
|
}
|
||||||
|
@ -1349,7 +1349,7 @@ int intel_sprite_get_colorkey(struct drm_device *dev, void *data,
|
||||||
drm_modeset_lock_all(dev);
|
drm_modeset_lock_all(dev);
|
||||||
|
|
||||||
plane = drm_plane_find(dev, get->plane_id);
|
plane = drm_plane_find(dev, get->plane_id);
|
||||||
if (!plane) {
|
if (!plane || plane->type != DRM_PLANE_TYPE_OVERLAY) {
|
||||||
ret = -ENOENT;
|
ret = -ENOENT;
|
||||||
goto out_unlock;
|
goto out_unlock;
|
||||||
}
|
}
|
||||||
|
|
|
@ -2129,6 +2129,7 @@
|
||||||
#define VCE_UENC_REG_CLOCK_GATING 0x207c0
|
#define VCE_UENC_REG_CLOCK_GATING 0x207c0
|
||||||
#define VCE_SYS_INT_EN 0x21300
|
#define VCE_SYS_INT_EN 0x21300
|
||||||
# define VCE_SYS_INT_TRAP_INTERRUPT_EN (1 << 3)
|
# define VCE_SYS_INT_TRAP_INTERRUPT_EN (1 << 3)
|
||||||
|
#define VCE_LMI_VCPU_CACHE_40BIT_BAR 0x2145c
|
||||||
#define VCE_LMI_CTRL2 0x21474
|
#define VCE_LMI_CTRL2 0x21474
|
||||||
#define VCE_LMI_CTRL 0x21498
|
#define VCE_LMI_CTRL 0x21498
|
||||||
#define VCE_LMI_VM_CTRL 0x214a0
|
#define VCE_LMI_VM_CTRL 0x214a0
|
||||||
|
|
|
@ -1565,6 +1565,7 @@ struct radeon_dpm {
|
||||||
int new_active_crtc_count;
|
int new_active_crtc_count;
|
||||||
u32 current_active_crtcs;
|
u32 current_active_crtcs;
|
||||||
int current_active_crtc_count;
|
int current_active_crtc_count;
|
||||||
|
bool single_display;
|
||||||
struct radeon_dpm_dynamic_state dyn_state;
|
struct radeon_dpm_dynamic_state dyn_state;
|
||||||
struct radeon_dpm_fan fan;
|
struct radeon_dpm_fan fan;
|
||||||
u32 tdp_limit;
|
u32 tdp_limit;
|
||||||
|
|
|
@ -76,7 +76,7 @@ static bool igp_read_bios_from_vram(struct radeon_device *rdev)
|
||||||
|
|
||||||
static bool radeon_read_bios(struct radeon_device *rdev)
|
static bool radeon_read_bios(struct radeon_device *rdev)
|
||||||
{
|
{
|
||||||
uint8_t __iomem *bios;
|
uint8_t __iomem *bios, val1, val2;
|
||||||
size_t size;
|
size_t size;
|
||||||
|
|
||||||
rdev->bios = NULL;
|
rdev->bios = NULL;
|
||||||
|
@ -86,15 +86,19 @@ static bool radeon_read_bios(struct radeon_device *rdev)
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (size == 0 || bios[0] != 0x55 || bios[1] != 0xaa) {
|
val1 = readb(&bios[0]);
|
||||||
|
val2 = readb(&bios[1]);
|
||||||
|
|
||||||
|
if (size == 0 || val1 != 0x55 || val2 != 0xaa) {
|
||||||
pci_unmap_rom(rdev->pdev, bios);
|
pci_unmap_rom(rdev->pdev, bios);
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
rdev->bios = kmemdup(bios, size, GFP_KERNEL);
|
rdev->bios = kzalloc(size, GFP_KERNEL);
|
||||||
if (rdev->bios == NULL) {
|
if (rdev->bios == NULL) {
|
||||||
pci_unmap_rom(rdev->pdev, bios);
|
pci_unmap_rom(rdev->pdev, bios);
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
memcpy_fromio(rdev->bios, bios, size);
|
||||||
pci_unmap_rom(rdev->pdev, bios);
|
pci_unmap_rom(rdev->pdev, bios);
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
|
@ -122,7 +122,6 @@ static void radeon_mn_invalidate_range_start(struct mmu_notifier *mn,
|
||||||
it = interval_tree_iter_first(&rmn->objects, start, end);
|
it = interval_tree_iter_first(&rmn->objects, start, end);
|
||||||
while (it) {
|
while (it) {
|
||||||
struct radeon_bo *bo;
|
struct radeon_bo *bo;
|
||||||
struct fence *fence;
|
|
||||||
int r;
|
int r;
|
||||||
|
|
||||||
bo = container_of(it, struct radeon_bo, mn_it);
|
bo = container_of(it, struct radeon_bo, mn_it);
|
||||||
|
@ -134,12 +133,10 @@ static void radeon_mn_invalidate_range_start(struct mmu_notifier *mn,
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
fence = reservation_object_get_excl(bo->tbo.resv);
|
r = reservation_object_wait_timeout_rcu(bo->tbo.resv, true,
|
||||||
if (fence) {
|
false, MAX_SCHEDULE_TIMEOUT);
|
||||||
r = radeon_fence_wait((struct radeon_fence *)fence, false);
|
if (r)
|
||||||
if (r)
|
DRM_ERROR("(%d) failed to wait for user bo\n", r);
|
||||||
DRM_ERROR("(%d) failed to wait for user bo\n", r);
|
|
||||||
}
|
|
||||||
|
|
||||||
radeon_ttm_placement_from_domain(bo, RADEON_GEM_DOMAIN_CPU);
|
radeon_ttm_placement_from_domain(bo, RADEON_GEM_DOMAIN_CPU);
|
||||||
r = ttm_bo_validate(&bo->tbo, &bo->placement, false, false);
|
r = ttm_bo_validate(&bo->tbo, &bo->placement, false, false);
|
||||||
|
|
|
@ -837,12 +837,8 @@ static void radeon_dpm_thermal_work_handler(struct work_struct *work)
|
||||||
radeon_pm_compute_clocks(rdev);
|
radeon_pm_compute_clocks(rdev);
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct radeon_ps *radeon_dpm_pick_power_state(struct radeon_device *rdev,
|
static bool radeon_dpm_single_display(struct radeon_device *rdev)
|
||||||
enum radeon_pm_state_type dpm_state)
|
|
||||||
{
|
{
|
||||||
int i;
|
|
||||||
struct radeon_ps *ps;
|
|
||||||
u32 ui_class;
|
|
||||||
bool single_display = (rdev->pm.dpm.new_active_crtc_count < 2) ?
|
bool single_display = (rdev->pm.dpm.new_active_crtc_count < 2) ?
|
||||||
true : false;
|
true : false;
|
||||||
|
|
||||||
|
@ -858,6 +854,17 @@ static struct radeon_ps *radeon_dpm_pick_power_state(struct radeon_device *rdev,
|
||||||
if (single_display && (r600_dpm_get_vrefresh(rdev) >= 120))
|
if (single_display && (r600_dpm_get_vrefresh(rdev) >= 120))
|
||||||
single_display = false;
|
single_display = false;
|
||||||
|
|
||||||
|
return single_display;
|
||||||
|
}
|
||||||
|
|
||||||
|
static struct radeon_ps *radeon_dpm_pick_power_state(struct radeon_device *rdev,
|
||||||
|
enum radeon_pm_state_type dpm_state)
|
||||||
|
{
|
||||||
|
int i;
|
||||||
|
struct radeon_ps *ps;
|
||||||
|
u32 ui_class;
|
||||||
|
bool single_display = radeon_dpm_single_display(rdev);
|
||||||
|
|
||||||
/* certain older asics have a separare 3D performance state,
|
/* certain older asics have a separare 3D performance state,
|
||||||
* so try that first if the user selected performance
|
* so try that first if the user selected performance
|
||||||
*/
|
*/
|
||||||
|
@ -983,6 +990,7 @@ static void radeon_dpm_change_power_state_locked(struct radeon_device *rdev)
|
||||||
struct radeon_ps *ps;
|
struct radeon_ps *ps;
|
||||||
enum radeon_pm_state_type dpm_state;
|
enum radeon_pm_state_type dpm_state;
|
||||||
int ret;
|
int ret;
|
||||||
|
bool single_display = radeon_dpm_single_display(rdev);
|
||||||
|
|
||||||
/* if dpm init failed */
|
/* if dpm init failed */
|
||||||
if (!rdev->pm.dpm_enabled)
|
if (!rdev->pm.dpm_enabled)
|
||||||
|
@ -1007,6 +1015,9 @@ static void radeon_dpm_change_power_state_locked(struct radeon_device *rdev)
|
||||||
/* vce just modifies an existing state so force a change */
|
/* vce just modifies an existing state so force a change */
|
||||||
if (ps->vce_active != rdev->pm.dpm.vce_active)
|
if (ps->vce_active != rdev->pm.dpm.vce_active)
|
||||||
goto force;
|
goto force;
|
||||||
|
/* user has made a display change (such as timing) */
|
||||||
|
if (rdev->pm.dpm.single_display != single_display)
|
||||||
|
goto force;
|
||||||
if ((rdev->family < CHIP_BARTS) || (rdev->flags & RADEON_IS_IGP)) {
|
if ((rdev->family < CHIP_BARTS) || (rdev->flags & RADEON_IS_IGP)) {
|
||||||
/* for pre-BTC and APUs if the num crtcs changed but state is the same,
|
/* for pre-BTC and APUs if the num crtcs changed but state is the same,
|
||||||
* all we need to do is update the display configuration.
|
* all we need to do is update the display configuration.
|
||||||
|
@ -1069,6 +1080,7 @@ force:
|
||||||
|
|
||||||
rdev->pm.dpm.current_active_crtcs = rdev->pm.dpm.new_active_crtcs;
|
rdev->pm.dpm.current_active_crtcs = rdev->pm.dpm.new_active_crtcs;
|
||||||
rdev->pm.dpm.current_active_crtc_count = rdev->pm.dpm.new_active_crtc_count;
|
rdev->pm.dpm.current_active_crtc_count = rdev->pm.dpm.new_active_crtc_count;
|
||||||
|
rdev->pm.dpm.single_display = single_display;
|
||||||
|
|
||||||
/* wait for the rings to drain */
|
/* wait for the rings to drain */
|
||||||
for (i = 0; i < RADEON_NUM_RINGS; i++) {
|
for (i = 0; i < RADEON_NUM_RINGS; i++) {
|
||||||
|
|
|
@ -495,7 +495,7 @@ static int radeon_debugfs_ring_info(struct seq_file *m, void *data)
|
||||||
seq_printf(m, "%u free dwords in ring\n", ring->ring_free_dw);
|
seq_printf(m, "%u free dwords in ring\n", ring->ring_free_dw);
|
||||||
seq_printf(m, "%u dwords in ring\n", count);
|
seq_printf(m, "%u dwords in ring\n", count);
|
||||||
|
|
||||||
if (!ring->ready)
|
if (!ring->ring)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
/* print 8 dw before current rptr as often it's the last executed
|
/* print 8 dw before current rptr as often it's the last executed
|
||||||
|
|
|
@ -598,6 +598,10 @@ static void radeon_ttm_tt_unpin_userptr(struct ttm_tt *ttm)
|
||||||
enum dma_data_direction direction = write ?
|
enum dma_data_direction direction = write ?
|
||||||
DMA_BIDIRECTIONAL : DMA_TO_DEVICE;
|
DMA_BIDIRECTIONAL : DMA_TO_DEVICE;
|
||||||
|
|
||||||
|
/* double check that we don't free the table twice */
|
||||||
|
if (!ttm->sg->sgl)
|
||||||
|
return;
|
||||||
|
|
||||||
/* free the sg table and pages again */
|
/* free the sg table and pages again */
|
||||||
dma_unmap_sg(rdev->dev, ttm->sg->sgl, ttm->sg->nents, direction);
|
dma_unmap_sg(rdev->dev, ttm->sg->sgl, ttm->sg->nents, direction);
|
||||||
|
|
||||||
|
|
|
@ -156,6 +156,9 @@ int vce_v2_0_resume(struct radeon_device *rdev)
|
||||||
WREG32(VCE_LMI_SWAP_CNTL1, 0);
|
WREG32(VCE_LMI_SWAP_CNTL1, 0);
|
||||||
WREG32(VCE_LMI_VM_CTRL, 0);
|
WREG32(VCE_LMI_VM_CTRL, 0);
|
||||||
|
|
||||||
|
WREG32(VCE_LMI_VCPU_CACHE_40BIT_BAR, addr >> 8);
|
||||||
|
|
||||||
|
addr &= 0xff;
|
||||||
size = RADEON_GPU_PAGE_ALIGN(rdev->vce_fw->size);
|
size = RADEON_GPU_PAGE_ALIGN(rdev->vce_fw->size);
|
||||||
WREG32(VCE_VCPU_CACHE_OFFSET0, addr & 0x7fffffff);
|
WREG32(VCE_VCPU_CACHE_OFFSET0, addr & 0x7fffffff);
|
||||||
WREG32(VCE_VCPU_CACHE_SIZE0, size);
|
WREG32(VCE_VCPU_CACHE_SIZE0, size);
|
||||||
|
|
|
@ -659,7 +659,7 @@ static irqreturn_t bma180_trigger_handler(int irq, void *p)
|
||||||
|
|
||||||
mutex_lock(&data->mutex);
|
mutex_lock(&data->mutex);
|
||||||
|
|
||||||
for_each_set_bit(bit, indio_dev->buffer->scan_mask,
|
for_each_set_bit(bit, indio_dev->active_scan_mask,
|
||||||
indio_dev->masklength) {
|
indio_dev->masklength) {
|
||||||
ret = bma180_get_data_reg(data, bit);
|
ret = bma180_get_data_reg(data, bit);
|
||||||
if (ret < 0) {
|
if (ret < 0) {
|
||||||
|
|
|
@ -168,14 +168,14 @@ static const struct {
|
||||||
int val;
|
int val;
|
||||||
int val2;
|
int val2;
|
||||||
u8 bw_bits;
|
u8 bw_bits;
|
||||||
} bmc150_accel_samp_freq_table[] = { {7, 810000, 0x08},
|
} bmc150_accel_samp_freq_table[] = { {15, 620000, 0x08},
|
||||||
{15, 630000, 0x09},
|
{31, 260000, 0x09},
|
||||||
{31, 250000, 0x0A},
|
{62, 500000, 0x0A},
|
||||||
{62, 500000, 0x0B},
|
{125, 0, 0x0B},
|
||||||
{125, 0, 0x0C},
|
{250, 0, 0x0C},
|
||||||
{250, 0, 0x0D},
|
{500, 0, 0x0D},
|
||||||
{500, 0, 0x0E},
|
{1000, 0, 0x0E},
|
||||||
{1000, 0, 0x0F} };
|
{2000, 0, 0x0F} };
|
||||||
|
|
||||||
static const struct {
|
static const struct {
|
||||||
int bw_bits;
|
int bw_bits;
|
||||||
|
@ -840,7 +840,7 @@ static int bmc150_accel_validate_trigger(struct iio_dev *indio_dev,
|
||||||
}
|
}
|
||||||
|
|
||||||
static IIO_CONST_ATTR_SAMP_FREQ_AVAIL(
|
static IIO_CONST_ATTR_SAMP_FREQ_AVAIL(
|
||||||
"7.810000 15.630000 31.250000 62.500000 125 250 500 1000");
|
"15.620000 31.260000 62.50000 125 250 500 1000 2000");
|
||||||
|
|
||||||
static struct attribute *bmc150_accel_attributes[] = {
|
static struct attribute *bmc150_accel_attributes[] = {
|
||||||
&iio_const_attr_sampling_frequency_available.dev_attr.attr,
|
&iio_const_attr_sampling_frequency_available.dev_attr.attr,
|
||||||
|
@ -986,7 +986,7 @@ static irqreturn_t bmc150_accel_trigger_handler(int irq, void *p)
|
||||||
int bit, ret, i = 0;
|
int bit, ret, i = 0;
|
||||||
|
|
||||||
mutex_lock(&data->mutex);
|
mutex_lock(&data->mutex);
|
||||||
for_each_set_bit(bit, indio_dev->buffer->scan_mask,
|
for_each_set_bit(bit, indio_dev->active_scan_mask,
|
||||||
indio_dev->masklength) {
|
indio_dev->masklength) {
|
||||||
ret = i2c_smbus_read_word_data(data->client,
|
ret = i2c_smbus_read_word_data(data->client,
|
||||||
BMC150_ACCEL_AXIS_TO_REG(bit));
|
BMC150_ACCEL_AXIS_TO_REG(bit));
|
||||||
|
|
|
@ -956,7 +956,7 @@ static irqreturn_t kxcjk1013_trigger_handler(int irq, void *p)
|
||||||
|
|
||||||
mutex_lock(&data->mutex);
|
mutex_lock(&data->mutex);
|
||||||
|
|
||||||
for_each_set_bit(bit, indio_dev->buffer->scan_mask,
|
for_each_set_bit(bit, indio_dev->active_scan_mask,
|
||||||
indio_dev->masklength) {
|
indio_dev->masklength) {
|
||||||
ret = kxcjk1013_get_acc_reg(data, bit);
|
ret = kxcjk1013_get_acc_reg(data, bit);
|
||||||
if (ret < 0) {
|
if (ret < 0) {
|
||||||
|
|
|
@ -137,7 +137,8 @@ config AXP288_ADC
|
||||||
|
|
||||||
config CC10001_ADC
|
config CC10001_ADC
|
||||||
tristate "Cosmic Circuits 10001 ADC driver"
|
tristate "Cosmic Circuits 10001 ADC driver"
|
||||||
depends on HAS_IOMEM || HAVE_CLK || REGULATOR
|
depends on HAVE_CLK || REGULATOR
|
||||||
|
depends on HAS_IOMEM
|
||||||
select IIO_BUFFER
|
select IIO_BUFFER
|
||||||
select IIO_TRIGGERED_BUFFER
|
select IIO_TRIGGERED_BUFFER
|
||||||
help
|
help
|
||||||
|
|
|
@ -544,7 +544,6 @@ static int at91_adc_configure_trigger(struct iio_trigger *trig, bool state)
|
||||||
{
|
{
|
||||||
struct iio_dev *idev = iio_trigger_get_drvdata(trig);
|
struct iio_dev *idev = iio_trigger_get_drvdata(trig);
|
||||||
struct at91_adc_state *st = iio_priv(idev);
|
struct at91_adc_state *st = iio_priv(idev);
|
||||||
struct iio_buffer *buffer = idev->buffer;
|
|
||||||
struct at91_adc_reg_desc *reg = st->registers;
|
struct at91_adc_reg_desc *reg = st->registers;
|
||||||
u32 status = at91_adc_readl(st, reg->trigger_register);
|
u32 status = at91_adc_readl(st, reg->trigger_register);
|
||||||
int value;
|
int value;
|
||||||
|
@ -564,7 +563,7 @@ static int at91_adc_configure_trigger(struct iio_trigger *trig, bool state)
|
||||||
at91_adc_writel(st, reg->trigger_register,
|
at91_adc_writel(st, reg->trigger_register,
|
||||||
status | value);
|
status | value);
|
||||||
|
|
||||||
for_each_set_bit(bit, buffer->scan_mask,
|
for_each_set_bit(bit, idev->active_scan_mask,
|
||||||
st->num_channels) {
|
st->num_channels) {
|
||||||
struct iio_chan_spec const *chan = idev->channels + bit;
|
struct iio_chan_spec const *chan = idev->channels + bit;
|
||||||
at91_adc_writel(st, AT91_ADC_CHER,
|
at91_adc_writel(st, AT91_ADC_CHER,
|
||||||
|
@ -579,7 +578,7 @@ static int at91_adc_configure_trigger(struct iio_trigger *trig, bool state)
|
||||||
at91_adc_writel(st, reg->trigger_register,
|
at91_adc_writel(st, reg->trigger_register,
|
||||||
status & ~value);
|
status & ~value);
|
||||||
|
|
||||||
for_each_set_bit(bit, buffer->scan_mask,
|
for_each_set_bit(bit, idev->active_scan_mask,
|
||||||
st->num_channels) {
|
st->num_channels) {
|
||||||
struct iio_chan_spec const *chan = idev->channels + bit;
|
struct iio_chan_spec const *chan = idev->channels + bit;
|
||||||
at91_adc_writel(st, AT91_ADC_CHDR,
|
at91_adc_writel(st, AT91_ADC_CHDR,
|
||||||
|
|
|
@ -188,12 +188,11 @@ static int tiadc_buffer_preenable(struct iio_dev *indio_dev)
|
||||||
static int tiadc_buffer_postenable(struct iio_dev *indio_dev)
|
static int tiadc_buffer_postenable(struct iio_dev *indio_dev)
|
||||||
{
|
{
|
||||||
struct tiadc_device *adc_dev = iio_priv(indio_dev);
|
struct tiadc_device *adc_dev = iio_priv(indio_dev);
|
||||||
struct iio_buffer *buffer = indio_dev->buffer;
|
|
||||||
unsigned int enb = 0;
|
unsigned int enb = 0;
|
||||||
u8 bit;
|
u8 bit;
|
||||||
|
|
||||||
tiadc_step_config(indio_dev);
|
tiadc_step_config(indio_dev);
|
||||||
for_each_set_bit(bit, buffer->scan_mask, adc_dev->channels)
|
for_each_set_bit(bit, indio_dev->active_scan_mask, adc_dev->channels)
|
||||||
enb |= (get_adc_step_bit(adc_dev, bit) << 1);
|
enb |= (get_adc_step_bit(adc_dev, bit) << 1);
|
||||||
adc_dev->buffer_en_ch_steps = enb;
|
adc_dev->buffer_en_ch_steps = enb;
|
||||||
|
|
||||||
|
|
|
@ -141,9 +141,13 @@ struct vf610_adc {
|
||||||
struct regulator *vref;
|
struct regulator *vref;
|
||||||
struct vf610_adc_feature adc_feature;
|
struct vf610_adc_feature adc_feature;
|
||||||
|
|
||||||
|
u32 sample_freq_avail[5];
|
||||||
|
|
||||||
struct completion completion;
|
struct completion completion;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
static const u32 vf610_hw_avgs[] = { 1, 4, 8, 16, 32 };
|
||||||
|
|
||||||
#define VF610_ADC_CHAN(_idx, _chan_type) { \
|
#define VF610_ADC_CHAN(_idx, _chan_type) { \
|
||||||
.type = (_chan_type), \
|
.type = (_chan_type), \
|
||||||
.indexed = 1, \
|
.indexed = 1, \
|
||||||
|
@ -180,35 +184,47 @@ static const struct iio_chan_spec vf610_adc_iio_channels[] = {
|
||||||
/* sentinel */
|
/* sentinel */
|
||||||
};
|
};
|
||||||
|
|
||||||
/*
|
static inline void vf610_adc_calculate_rates(struct vf610_adc *info)
|
||||||
* ADC sample frequency, unit is ADCK cycles.
|
{
|
||||||
* ADC clk source is ipg clock, which is the same as bus clock.
|
unsigned long adck_rate, ipg_rate = clk_get_rate(info->clk);
|
||||||
*
|
int i;
|
||||||
* ADC conversion time = SFCAdder + AverageNum x (BCT + LSTAdder)
|
|
||||||
* SFCAdder: fixed to 6 ADCK cycles
|
/*
|
||||||
* AverageNum: 1, 4, 8, 16, 32 samples for hardware average.
|
* Calculate ADC sample frequencies
|
||||||
* BCT (Base Conversion Time): fixed to 25 ADCK cycles for 12 bit mode
|
* Sample time unit is ADCK cycles. ADCK clk source is ipg clock,
|
||||||
* LSTAdder(Long Sample Time): fixed to 3 ADCK cycles
|
* which is the same as bus clock.
|
||||||
*
|
*
|
||||||
* By default, enable 12 bit resolution mode, clock source
|
* ADC conversion time = SFCAdder + AverageNum x (BCT + LSTAdder)
|
||||||
* set to ipg clock, So get below frequency group:
|
* SFCAdder: fixed to 6 ADCK cycles
|
||||||
*/
|
* AverageNum: 1, 4, 8, 16, 32 samples for hardware average.
|
||||||
static const u32 vf610_sample_freq_avail[5] =
|
* BCT (Base Conversion Time): fixed to 25 ADCK cycles for 12 bit mode
|
||||||
{1941176, 559332, 286957, 145374, 73171};
|
* LSTAdder(Long Sample Time): fixed to 3 ADCK cycles
|
||||||
|
*/
|
||||||
|
adck_rate = ipg_rate / info->adc_feature.clk_div;
|
||||||
|
for (i = 0; i < ARRAY_SIZE(vf610_hw_avgs); i++)
|
||||||
|
info->sample_freq_avail[i] =
|
||||||
|
adck_rate / (6 + vf610_hw_avgs[i] * (25 + 3));
|
||||||
|
}
|
||||||
|
|
||||||
static inline void vf610_adc_cfg_init(struct vf610_adc *info)
|
static inline void vf610_adc_cfg_init(struct vf610_adc *info)
|
||||||
{
|
{
|
||||||
|
struct vf610_adc_feature *adc_feature = &info->adc_feature;
|
||||||
|
|
||||||
/* set default Configuration for ADC controller */
|
/* set default Configuration for ADC controller */
|
||||||
info->adc_feature.clk_sel = VF610_ADCIOC_BUSCLK_SET;
|
adc_feature->clk_sel = VF610_ADCIOC_BUSCLK_SET;
|
||||||
info->adc_feature.vol_ref = VF610_ADCIOC_VR_VREF_SET;
|
adc_feature->vol_ref = VF610_ADCIOC_VR_VREF_SET;
|
||||||
|
|
||||||
info->adc_feature.calibration = true;
|
adc_feature->calibration = true;
|
||||||
info->adc_feature.ovwren = true;
|
adc_feature->ovwren = true;
|
||||||
|
|
||||||
info->adc_feature.clk_div = 1;
|
adc_feature->res_mode = 12;
|
||||||
info->adc_feature.res_mode = 12;
|
adc_feature->sample_rate = 1;
|
||||||
info->adc_feature.sample_rate = 1;
|
adc_feature->lpm = true;
|
||||||
info->adc_feature.lpm = true;
|
|
||||||
|
/* Use a save ADCK which is below 20MHz on all devices */
|
||||||
|
adc_feature->clk_div = 8;
|
||||||
|
|
||||||
|
vf610_adc_calculate_rates(info);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void vf610_adc_cfg_post_set(struct vf610_adc *info)
|
static void vf610_adc_cfg_post_set(struct vf610_adc *info)
|
||||||
|
@ -290,12 +306,10 @@ static void vf610_adc_cfg_set(struct vf610_adc *info)
|
||||||
|
|
||||||
cfg_data = readl(info->regs + VF610_REG_ADC_CFG);
|
cfg_data = readl(info->regs + VF610_REG_ADC_CFG);
|
||||||
|
|
||||||
/* low power configuration */
|
|
||||||
cfg_data &= ~VF610_ADC_ADLPC_EN;
|
cfg_data &= ~VF610_ADC_ADLPC_EN;
|
||||||
if (adc_feature->lpm)
|
if (adc_feature->lpm)
|
||||||
cfg_data |= VF610_ADC_ADLPC_EN;
|
cfg_data |= VF610_ADC_ADLPC_EN;
|
||||||
|
|
||||||
/* disable high speed */
|
|
||||||
cfg_data &= ~VF610_ADC_ADHSC_EN;
|
cfg_data &= ~VF610_ADC_ADHSC_EN;
|
||||||
|
|
||||||
writel(cfg_data, info->regs + VF610_REG_ADC_CFG);
|
writel(cfg_data, info->regs + VF610_REG_ADC_CFG);
|
||||||
|
@ -435,10 +449,27 @@ static irqreturn_t vf610_adc_isr(int irq, void *dev_id)
|
||||||
return IRQ_HANDLED;
|
return IRQ_HANDLED;
|
||||||
}
|
}
|
||||||
|
|
||||||
static IIO_CONST_ATTR_SAMP_FREQ_AVAIL("1941176, 559332, 286957, 145374, 73171");
|
static ssize_t vf610_show_samp_freq_avail(struct device *dev,
|
||||||
|
struct device_attribute *attr, char *buf)
|
||||||
|
{
|
||||||
|
struct vf610_adc *info = iio_priv(dev_to_iio_dev(dev));
|
||||||
|
size_t len = 0;
|
||||||
|
int i;
|
||||||
|
|
||||||
|
for (i = 0; i < ARRAY_SIZE(info->sample_freq_avail); i++)
|
||||||
|
len += scnprintf(buf + len, PAGE_SIZE - len,
|
||||||
|
"%u ", info->sample_freq_avail[i]);
|
||||||
|
|
||||||
|
/* replace trailing space by newline */
|
||||||
|
buf[len - 1] = '\n';
|
||||||
|
|
||||||
|
return len;
|
||||||
|
}
|
||||||
|
|
||||||
|
static IIO_DEV_ATTR_SAMP_FREQ_AVAIL(vf610_show_samp_freq_avail);
|
||||||
|
|
||||||
static struct attribute *vf610_attributes[] = {
|
static struct attribute *vf610_attributes[] = {
|
||||||
&iio_const_attr_sampling_frequency_available.dev_attr.attr,
|
&iio_dev_attr_sampling_frequency_available.dev_attr.attr,
|
||||||
NULL
|
NULL
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -502,7 +533,7 @@ static int vf610_read_raw(struct iio_dev *indio_dev,
|
||||||
return IIO_VAL_FRACTIONAL_LOG2;
|
return IIO_VAL_FRACTIONAL_LOG2;
|
||||||
|
|
||||||
case IIO_CHAN_INFO_SAMP_FREQ:
|
case IIO_CHAN_INFO_SAMP_FREQ:
|
||||||
*val = vf610_sample_freq_avail[info->adc_feature.sample_rate];
|
*val = info->sample_freq_avail[info->adc_feature.sample_rate];
|
||||||
*val2 = 0;
|
*val2 = 0;
|
||||||
return IIO_VAL_INT;
|
return IIO_VAL_INT;
|
||||||
|
|
||||||
|
@ -525,9 +556,9 @@ static int vf610_write_raw(struct iio_dev *indio_dev,
|
||||||
switch (mask) {
|
switch (mask) {
|
||||||
case IIO_CHAN_INFO_SAMP_FREQ:
|
case IIO_CHAN_INFO_SAMP_FREQ:
|
||||||
for (i = 0;
|
for (i = 0;
|
||||||
i < ARRAY_SIZE(vf610_sample_freq_avail);
|
i < ARRAY_SIZE(info->sample_freq_avail);
|
||||||
i++)
|
i++)
|
||||||
if (val == vf610_sample_freq_avail[i]) {
|
if (val == info->sample_freq_avail[i]) {
|
||||||
info->adc_feature.sample_rate = i;
|
info->adc_feature.sample_rate = i;
|
||||||
vf610_adc_sample_set(info);
|
vf610_adc_sample_set(info);
|
||||||
return 0;
|
return 0;
|
||||||
|
|
|
@ -822,7 +822,7 @@ static irqreturn_t bmg160_trigger_handler(int irq, void *p)
|
||||||
int bit, ret, i = 0;
|
int bit, ret, i = 0;
|
||||||
|
|
||||||
mutex_lock(&data->mutex);
|
mutex_lock(&data->mutex);
|
||||||
for_each_set_bit(bit, indio_dev->buffer->scan_mask,
|
for_each_set_bit(bit, indio_dev->active_scan_mask,
|
||||||
indio_dev->masklength) {
|
indio_dev->masklength) {
|
||||||
ret = i2c_smbus_read_word_data(data->client,
|
ret = i2c_smbus_read_word_data(data->client,
|
||||||
BMG160_AXIS_TO_REG(bit));
|
BMG160_AXIS_TO_REG(bit));
|
||||||
|
|
|
@ -60,7 +60,7 @@ int adis_probe_trigger(struct adis *adis, struct iio_dev *indio_dev)
|
||||||
iio_trigger_set_drvdata(adis->trig, adis);
|
iio_trigger_set_drvdata(adis->trig, adis);
|
||||||
ret = iio_trigger_register(adis->trig);
|
ret = iio_trigger_register(adis->trig);
|
||||||
|
|
||||||
indio_dev->trig = adis->trig;
|
indio_dev->trig = iio_trigger_get(adis->trig);
|
||||||
if (ret)
|
if (ret)
|
||||||
goto error_free_irq;
|
goto error_free_irq;
|
||||||
|
|
||||||
|
|
|
@ -410,42 +410,46 @@ error_read_raw:
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static int inv_mpu6050_write_fsr(struct inv_mpu6050_state *st, int fsr)
|
static int inv_mpu6050_write_gyro_scale(struct inv_mpu6050_state *st, int val)
|
||||||
{
|
{
|
||||||
int result;
|
int result, i;
|
||||||
u8 d;
|
u8 d;
|
||||||
|
|
||||||
if (fsr < 0 || fsr > INV_MPU6050_MAX_GYRO_FS_PARAM)
|
for (i = 0; i < ARRAY_SIZE(gyro_scale_6050); ++i) {
|
||||||
return -EINVAL;
|
if (gyro_scale_6050[i] == val) {
|
||||||
if (fsr == st->chip_config.fsr)
|
d = (i << INV_MPU6050_GYRO_CONFIG_FSR_SHIFT);
|
||||||
return 0;
|
result = inv_mpu6050_write_reg(st,
|
||||||
|
st->reg->gyro_config, d);
|
||||||
|
if (result)
|
||||||
|
return result;
|
||||||
|
|
||||||
d = (fsr << INV_MPU6050_GYRO_CONFIG_FSR_SHIFT);
|
st->chip_config.fsr = i;
|
||||||
result = inv_mpu6050_write_reg(st, st->reg->gyro_config, d);
|
return 0;
|
||||||
if (result)
|
}
|
||||||
return result;
|
}
|
||||||
st->chip_config.fsr = fsr;
|
|
||||||
|
|
||||||
return 0;
|
return -EINVAL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int inv_mpu6050_write_accel_fs(struct inv_mpu6050_state *st, int fs)
|
static int inv_mpu6050_write_accel_scale(struct inv_mpu6050_state *st, int val)
|
||||||
{
|
{
|
||||||
int result;
|
int result, i;
|
||||||
u8 d;
|
u8 d;
|
||||||
|
|
||||||
if (fs < 0 || fs > INV_MPU6050_MAX_ACCL_FS_PARAM)
|
for (i = 0; i < ARRAY_SIZE(accel_scale); ++i) {
|
||||||
return -EINVAL;
|
if (accel_scale[i] == val) {
|
||||||
if (fs == st->chip_config.accl_fs)
|
d = (i << INV_MPU6050_ACCL_CONFIG_FSR_SHIFT);
|
||||||
return 0;
|
result = inv_mpu6050_write_reg(st,
|
||||||
|
st->reg->accl_config, d);
|
||||||
|
if (result)
|
||||||
|
return result;
|
||||||
|
|
||||||
d = (fs << INV_MPU6050_ACCL_CONFIG_FSR_SHIFT);
|
st->chip_config.accl_fs = i;
|
||||||
result = inv_mpu6050_write_reg(st, st->reg->accl_config, d);
|
return 0;
|
||||||
if (result)
|
}
|
||||||
return result;
|
}
|
||||||
st->chip_config.accl_fs = fs;
|
|
||||||
|
|
||||||
return 0;
|
return -EINVAL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int inv_mpu6050_write_raw(struct iio_dev *indio_dev,
|
static int inv_mpu6050_write_raw(struct iio_dev *indio_dev,
|
||||||
|
@ -471,10 +475,10 @@ static int inv_mpu6050_write_raw(struct iio_dev *indio_dev,
|
||||||
case IIO_CHAN_INFO_SCALE:
|
case IIO_CHAN_INFO_SCALE:
|
||||||
switch (chan->type) {
|
switch (chan->type) {
|
||||||
case IIO_ANGL_VEL:
|
case IIO_ANGL_VEL:
|
||||||
result = inv_mpu6050_write_fsr(st, val);
|
result = inv_mpu6050_write_gyro_scale(st, val2);
|
||||||
break;
|
break;
|
||||||
case IIO_ACCEL:
|
case IIO_ACCEL:
|
||||||
result = inv_mpu6050_write_accel_fs(st, val);
|
result = inv_mpu6050_write_accel_scale(st, val2);
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
result = -EINVAL;
|
result = -EINVAL;
|
||||||
|
|
|
@ -24,6 +24,16 @@
|
||||||
#include <linux/poll.h>
|
#include <linux/poll.h>
|
||||||
#include "inv_mpu_iio.h"
|
#include "inv_mpu_iio.h"
|
||||||
|
|
||||||
|
static void inv_clear_kfifo(struct inv_mpu6050_state *st)
|
||||||
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
|
/* take the spin lock sem to avoid interrupt kick in */
|
||||||
|
spin_lock_irqsave(&st->time_stamp_lock, flags);
|
||||||
|
kfifo_reset(&st->timestamps);
|
||||||
|
spin_unlock_irqrestore(&st->time_stamp_lock, flags);
|
||||||
|
}
|
||||||
|
|
||||||
int inv_reset_fifo(struct iio_dev *indio_dev)
|
int inv_reset_fifo(struct iio_dev *indio_dev)
|
||||||
{
|
{
|
||||||
int result;
|
int result;
|
||||||
|
@ -50,6 +60,10 @@ int inv_reset_fifo(struct iio_dev *indio_dev)
|
||||||
INV_MPU6050_BIT_FIFO_RST);
|
INV_MPU6050_BIT_FIFO_RST);
|
||||||
if (result)
|
if (result)
|
||||||
goto reset_fifo_fail;
|
goto reset_fifo_fail;
|
||||||
|
|
||||||
|
/* clear timestamps fifo */
|
||||||
|
inv_clear_kfifo(st);
|
||||||
|
|
||||||
/* enable interrupt */
|
/* enable interrupt */
|
||||||
if (st->chip_config.accl_fifo_enable ||
|
if (st->chip_config.accl_fifo_enable ||
|
||||||
st->chip_config.gyro_fifo_enable) {
|
st->chip_config.gyro_fifo_enable) {
|
||||||
|
@ -83,16 +97,6 @@ reset_fifo_fail:
|
||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void inv_clear_kfifo(struct inv_mpu6050_state *st)
|
|
||||||
{
|
|
||||||
unsigned long flags;
|
|
||||||
|
|
||||||
/* take the spin lock sem to avoid interrupt kick in */
|
|
||||||
spin_lock_irqsave(&st->time_stamp_lock, flags);
|
|
||||||
kfifo_reset(&st->timestamps);
|
|
||||||
spin_unlock_irqrestore(&st->time_stamp_lock, flags);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* inv_mpu6050_irq_handler() - Cache a timestamp at each data ready interrupt.
|
* inv_mpu6050_irq_handler() - Cache a timestamp at each data ready interrupt.
|
||||||
*/
|
*/
|
||||||
|
@ -184,7 +188,6 @@ end_session:
|
||||||
flush_fifo:
|
flush_fifo:
|
||||||
/* Flush HW and SW FIFOs. */
|
/* Flush HW and SW FIFOs. */
|
||||||
inv_reset_fifo(indio_dev);
|
inv_reset_fifo(indio_dev);
|
||||||
inv_clear_kfifo(st);
|
|
||||||
mutex_unlock(&indio_dev->mlock);
|
mutex_unlock(&indio_dev->mlock);
|
||||||
iio_trigger_notify_done(indio_dev->trig);
|
iio_trigger_notify_done(indio_dev->trig);
|
||||||
|
|
||||||
|
|
|
@ -1227,7 +1227,7 @@ static irqreturn_t kmx61_trigger_handler(int irq, void *p)
|
||||||
base = KMX61_MAG_XOUT_L;
|
base = KMX61_MAG_XOUT_L;
|
||||||
|
|
||||||
mutex_lock(&data->lock);
|
mutex_lock(&data->lock);
|
||||||
for_each_set_bit(bit, indio_dev->buffer->scan_mask,
|
for_each_set_bit(bit, indio_dev->active_scan_mask,
|
||||||
indio_dev->masklength) {
|
indio_dev->masklength) {
|
||||||
ret = kmx61_read_measurement(data, base, bit);
|
ret = kmx61_read_measurement(data, base, bit);
|
||||||
if (ret < 0) {
|
if (ret < 0) {
|
||||||
|
|
|
@ -847,8 +847,7 @@ static int iio_device_add_channel_sysfs(struct iio_dev *indio_dev,
|
||||||
* @attr_list: List of IIO device attributes
|
* @attr_list: List of IIO device attributes
|
||||||
*
|
*
|
||||||
* This function frees the memory allocated for each of the IIO device
|
* This function frees the memory allocated for each of the IIO device
|
||||||
* attributes in the list. Note: if you want to reuse the list after calling
|
* attributes in the list.
|
||||||
* this function you have to reinitialize it using INIT_LIST_HEAD().
|
|
||||||
*/
|
*/
|
||||||
void iio_free_chan_devattr_list(struct list_head *attr_list)
|
void iio_free_chan_devattr_list(struct list_head *attr_list)
|
||||||
{
|
{
|
||||||
|
@ -856,6 +855,7 @@ void iio_free_chan_devattr_list(struct list_head *attr_list)
|
||||||
|
|
||||||
list_for_each_entry_safe(p, n, attr_list, l) {
|
list_for_each_entry_safe(p, n, attr_list, l) {
|
||||||
kfree(p->dev_attr.attr.name);
|
kfree(p->dev_attr.attr.name);
|
||||||
|
list_del(&p->l);
|
||||||
kfree(p);
|
kfree(p);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -936,6 +936,7 @@ static void iio_device_unregister_sysfs(struct iio_dev *indio_dev)
|
||||||
|
|
||||||
iio_free_chan_devattr_list(&indio_dev->channel_attr_list);
|
iio_free_chan_devattr_list(&indio_dev->channel_attr_list);
|
||||||
kfree(indio_dev->chan_attr_group.attrs);
|
kfree(indio_dev->chan_attr_group.attrs);
|
||||||
|
indio_dev->chan_attr_group.attrs = NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void iio_dev_release(struct device *device)
|
static void iio_dev_release(struct device *device)
|
||||||
|
|
|
@ -500,6 +500,7 @@ int iio_device_register_eventset(struct iio_dev *indio_dev)
|
||||||
error_free_setup_event_lines:
|
error_free_setup_event_lines:
|
||||||
iio_free_chan_devattr_list(&indio_dev->event_interface->dev_attr_list);
|
iio_free_chan_devattr_list(&indio_dev->event_interface->dev_attr_list);
|
||||||
kfree(indio_dev->event_interface);
|
kfree(indio_dev->event_interface);
|
||||||
|
indio_dev->event_interface = NULL;
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -494,7 +494,7 @@ static irqreturn_t sx9500_trigger_handler(int irq, void *private)
|
||||||
|
|
||||||
mutex_lock(&data->mutex);
|
mutex_lock(&data->mutex);
|
||||||
|
|
||||||
for_each_set_bit(bit, indio_dev->buffer->scan_mask,
|
for_each_set_bit(bit, indio_dev->active_scan_mask,
|
||||||
indio_dev->masklength) {
|
indio_dev->masklength) {
|
||||||
ret = sx9500_read_proximity(data, &indio_dev->channels[bit],
|
ret = sx9500_read_proximity(data, &indio_dev->channels[bit],
|
||||||
&val);
|
&val);
|
||||||
|
|
|
@ -99,6 +99,14 @@ struct ib_umem *ib_umem_get(struct ib_ucontext *context, unsigned long addr,
|
||||||
if (dmasync)
|
if (dmasync)
|
||||||
dma_set_attr(DMA_ATTR_WRITE_BARRIER, &attrs);
|
dma_set_attr(DMA_ATTR_WRITE_BARRIER, &attrs);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* If the combination of the addr and size requested for this memory
|
||||||
|
* region causes an integer overflow, return error.
|
||||||
|
*/
|
||||||
|
if ((PAGE_ALIGN(addr + size) <= size) ||
|
||||||
|
(PAGE_ALIGN(addr + size) <= addr))
|
||||||
|
return ERR_PTR(-EINVAL);
|
||||||
|
|
||||||
if (!can_do_mlock())
|
if (!can_do_mlock())
|
||||||
return ERR_PTR(-EPERM);
|
return ERR_PTR(-EPERM);
|
||||||
|
|
||||||
|
|
|
@ -1154,10 +1154,28 @@ out:
|
||||||
mutex_unlock(&alps_mutex);
|
mutex_unlock(&alps_mutex);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void alps_report_bare_ps2_packet(struct input_dev *dev,
|
static void alps_report_bare_ps2_packet(struct psmouse *psmouse,
|
||||||
unsigned char packet[],
|
unsigned char packet[],
|
||||||
bool report_buttons)
|
bool report_buttons)
|
||||||
{
|
{
|
||||||
|
struct alps_data *priv = psmouse->private;
|
||||||
|
struct input_dev *dev;
|
||||||
|
|
||||||
|
/* Figure out which device to use to report the bare packet */
|
||||||
|
if (priv->proto_version == ALPS_PROTO_V2 &&
|
||||||
|
(priv->flags & ALPS_DUALPOINT)) {
|
||||||
|
/* On V2 devices the DualPoint Stick reports bare packets */
|
||||||
|
dev = priv->dev2;
|
||||||
|
} else if (unlikely(IS_ERR_OR_NULL(priv->dev3))) {
|
||||||
|
/* Register dev3 mouse if we received PS/2 packet first time */
|
||||||
|
if (!IS_ERR(priv->dev3))
|
||||||
|
psmouse_queue_work(psmouse, &priv->dev3_register_work,
|
||||||
|
0);
|
||||||
|
return;
|
||||||
|
} else {
|
||||||
|
dev = priv->dev3;
|
||||||
|
}
|
||||||
|
|
||||||
if (report_buttons)
|
if (report_buttons)
|
||||||
alps_report_buttons(dev, NULL,
|
alps_report_buttons(dev, NULL,
|
||||||
packet[0] & 1, packet[0] & 2, packet[0] & 4);
|
packet[0] & 1, packet[0] & 2, packet[0] & 4);
|
||||||
|
@ -1232,8 +1250,8 @@ static psmouse_ret_t alps_handle_interleaved_ps2(struct psmouse *psmouse)
|
||||||
* de-synchronization.
|
* de-synchronization.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
alps_report_bare_ps2_packet(priv->dev2,
|
alps_report_bare_ps2_packet(psmouse, &psmouse->packet[3],
|
||||||
&psmouse->packet[3], false);
|
false);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Continue with the standard ALPS protocol handling,
|
* Continue with the standard ALPS protocol handling,
|
||||||
|
@ -1289,18 +1307,9 @@ static psmouse_ret_t alps_process_byte(struct psmouse *psmouse)
|
||||||
* properly we only do this if the device is fully synchronized.
|
* properly we only do this if the device is fully synchronized.
|
||||||
*/
|
*/
|
||||||
if (!psmouse->out_of_sync_cnt && (psmouse->packet[0] & 0xc8) == 0x08) {
|
if (!psmouse->out_of_sync_cnt && (psmouse->packet[0] & 0xc8) == 0x08) {
|
||||||
|
|
||||||
/* Register dev3 mouse if we received PS/2 packet first time */
|
|
||||||
if (unlikely(!priv->dev3))
|
|
||||||
psmouse_queue_work(psmouse,
|
|
||||||
&priv->dev3_register_work, 0);
|
|
||||||
|
|
||||||
if (psmouse->pktcnt == 3) {
|
if (psmouse->pktcnt == 3) {
|
||||||
/* Once dev3 mouse device is registered report data */
|
alps_report_bare_ps2_packet(psmouse, psmouse->packet,
|
||||||
if (likely(!IS_ERR_OR_NULL(priv->dev3)))
|
true);
|
||||||
alps_report_bare_ps2_packet(priv->dev3,
|
|
||||||
psmouse->packet,
|
|
||||||
true);
|
|
||||||
return PSMOUSE_FULL_PACKET;
|
return PSMOUSE_FULL_PACKET;
|
||||||
}
|
}
|
||||||
return PSMOUSE_GOOD_DATA;
|
return PSMOUSE_GOOD_DATA;
|
||||||
|
@ -2281,10 +2290,12 @@ static int alps_set_protocol(struct psmouse *psmouse,
|
||||||
priv->set_abs_params = alps_set_abs_params_mt;
|
priv->set_abs_params = alps_set_abs_params_mt;
|
||||||
priv->nibble_commands = alps_v3_nibble_commands;
|
priv->nibble_commands = alps_v3_nibble_commands;
|
||||||
priv->addr_command = PSMOUSE_CMD_RESET_WRAP;
|
priv->addr_command = PSMOUSE_CMD_RESET_WRAP;
|
||||||
priv->x_max = 1360;
|
|
||||||
priv->y_max = 660;
|
|
||||||
priv->x_bits = 23;
|
priv->x_bits = 23;
|
||||||
priv->y_bits = 12;
|
priv->y_bits = 12;
|
||||||
|
|
||||||
|
if (alps_dolphin_get_device_area(psmouse, priv))
|
||||||
|
return -EIO;
|
||||||
|
|
||||||
break;
|
break;
|
||||||
|
|
||||||
case ALPS_PROTO_V6:
|
case ALPS_PROTO_V6:
|
||||||
|
@ -2303,9 +2314,8 @@ static int alps_set_protocol(struct psmouse *psmouse,
|
||||||
priv->set_abs_params = alps_set_abs_params_mt;
|
priv->set_abs_params = alps_set_abs_params_mt;
|
||||||
priv->nibble_commands = alps_v3_nibble_commands;
|
priv->nibble_commands = alps_v3_nibble_commands;
|
||||||
priv->addr_command = PSMOUSE_CMD_RESET_WRAP;
|
priv->addr_command = PSMOUSE_CMD_RESET_WRAP;
|
||||||
|
priv->x_max = 0xfff;
|
||||||
if (alps_dolphin_get_device_area(psmouse, priv))
|
priv->y_max = 0x7ff;
|
||||||
return -EIO;
|
|
||||||
|
|
||||||
if (priv->fw_ver[1] != 0xba)
|
if (priv->fw_ver[1] != 0xba)
|
||||||
priv->flags |= ALPS_BUTTONPAD;
|
priv->flags |= ALPS_BUTTONPAD;
|
||||||
|
|
|
@ -152,6 +152,11 @@ static const struct min_max_quirk min_max_pnpid_table[] = {
|
||||||
{ANY_BOARD_ID, ANY_BOARD_ID},
|
{ANY_BOARD_ID, ANY_BOARD_ID},
|
||||||
1024, 5022, 2508, 4832
|
1024, 5022, 2508, 4832
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
(const char * const []){"LEN2006", NULL},
|
||||||
|
{2691, 2691},
|
||||||
|
1024, 5045, 2457, 4832
|
||||||
|
},
|
||||||
{
|
{
|
||||||
(const char * const []){"LEN2006", NULL},
|
(const char * const []){"LEN2006", NULL},
|
||||||
{ANY_BOARD_ID, ANY_BOARD_ID},
|
{ANY_BOARD_ID, ANY_BOARD_ID},
|
||||||
|
@ -189,7 +194,7 @@ static const char * const topbuttonpad_pnp_ids[] = {
|
||||||
"LEN2003",
|
"LEN2003",
|
||||||
"LEN2004", /* L440 */
|
"LEN2004", /* L440 */
|
||||||
"LEN2005",
|
"LEN2005",
|
||||||
"LEN2006",
|
"LEN2006", /* Edge E440/E540 */
|
||||||
"LEN2007",
|
"LEN2007",
|
||||||
"LEN2008",
|
"LEN2008",
|
||||||
"LEN2009",
|
"LEN2009",
|
||||||
|
|
|
@ -1288,10 +1288,13 @@ static phys_addr_t arm_smmu_iova_to_phys(struct iommu_domain *domain,
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
spin_lock_irqsave(&smmu_domain->pgtbl_lock, flags);
|
spin_lock_irqsave(&smmu_domain->pgtbl_lock, flags);
|
||||||
if (smmu_domain->smmu->features & ARM_SMMU_FEAT_TRANS_OPS)
|
if (smmu_domain->smmu->features & ARM_SMMU_FEAT_TRANS_OPS &&
|
||||||
|
smmu_domain->stage == ARM_SMMU_DOMAIN_S1) {
|
||||||
ret = arm_smmu_iova_to_phys_hard(domain, iova);
|
ret = arm_smmu_iova_to_phys_hard(domain, iova);
|
||||||
else
|
} else {
|
||||||
ret = ops->iova_to_phys(ops, iova);
|
ret = ops->iova_to_phys(ops, iova);
|
||||||
|
}
|
||||||
|
|
||||||
spin_unlock_irqrestore(&smmu_domain->pgtbl_lock, flags);
|
spin_unlock_irqrestore(&smmu_domain->pgtbl_lock, flags);
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
|
@ -1556,7 +1559,7 @@ static int arm_smmu_device_cfg_probe(struct arm_smmu_device *smmu)
|
||||||
return -ENODEV;
|
return -ENODEV;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (smmu->version == 1 || (!(id & ID0_ATOSNS) && (id & ID0_S1TS))) {
|
if ((id & ID0_S1TS) && ((smmu->version == 1) || (id & ID0_ATOSNS))) {
|
||||||
smmu->features |= ARM_SMMU_FEAT_TRANS_OPS;
|
smmu->features |= ARM_SMMU_FEAT_TRANS_OPS;
|
||||||
dev_notice(smmu->dev, "\taddress translation ops\n");
|
dev_notice(smmu->dev, "\taddress translation ops\n");
|
||||||
}
|
}
|
||||||
|
|
|
@ -1742,9 +1742,8 @@ static int domain_init(struct dmar_domain *domain, int guest_width)
|
||||||
|
|
||||||
static void domain_exit(struct dmar_domain *domain)
|
static void domain_exit(struct dmar_domain *domain)
|
||||||
{
|
{
|
||||||
struct dmar_drhd_unit *drhd;
|
|
||||||
struct intel_iommu *iommu;
|
|
||||||
struct page *freelist = NULL;
|
struct page *freelist = NULL;
|
||||||
|
int i;
|
||||||
|
|
||||||
/* Domain 0 is reserved, so dont process it */
|
/* Domain 0 is reserved, so dont process it */
|
||||||
if (!domain)
|
if (!domain)
|
||||||
|
@ -1764,8 +1763,8 @@ static void domain_exit(struct dmar_domain *domain)
|
||||||
|
|
||||||
/* clear attached or cached domains */
|
/* clear attached or cached domains */
|
||||||
rcu_read_lock();
|
rcu_read_lock();
|
||||||
for_each_active_iommu(iommu, drhd)
|
for_each_set_bit(i, domain->iommu_bmp, g_num_of_iommus)
|
||||||
iommu_detach_domain(domain, iommu);
|
iommu_detach_domain(domain, g_iommus[i]);
|
||||||
rcu_read_unlock();
|
rcu_read_unlock();
|
||||||
|
|
||||||
dma_free_pagelist(freelist);
|
dma_free_pagelist(freelist);
|
||||||
|
|
|
@ -851,6 +851,7 @@ static int ipmmu_remove(struct platform_device *pdev)
|
||||||
|
|
||||||
static const struct of_device_id ipmmu_of_ids[] = {
|
static const struct of_device_id ipmmu_of_ids[] = {
|
||||||
{ .compatible = "renesas,ipmmu-vmsa", },
|
{ .compatible = "renesas,ipmmu-vmsa", },
|
||||||
|
{ }
|
||||||
};
|
};
|
||||||
|
|
||||||
static struct platform_driver ipmmu_driver = {
|
static struct platform_driver ipmmu_driver = {
|
||||||
|
|
|
@ -169,7 +169,7 @@ static void its_encode_cmd(struct its_cmd_block *cmd, u8 cmd_nr)
|
||||||
|
|
||||||
static void its_encode_devid(struct its_cmd_block *cmd, u32 devid)
|
static void its_encode_devid(struct its_cmd_block *cmd, u32 devid)
|
||||||
{
|
{
|
||||||
cmd->raw_cmd[0] &= ~(0xffffUL << 32);
|
cmd->raw_cmd[0] &= BIT_ULL(32) - 1;
|
||||||
cmd->raw_cmd[0] |= ((u64)devid) << 32;
|
cmd->raw_cmd[0] |= ((u64)devid) << 32;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -802,6 +802,7 @@ static int its_alloc_tables(struct its_node *its)
|
||||||
int i;
|
int i;
|
||||||
int psz = SZ_64K;
|
int psz = SZ_64K;
|
||||||
u64 shr = GITS_BASER_InnerShareable;
|
u64 shr = GITS_BASER_InnerShareable;
|
||||||
|
u64 cache = GITS_BASER_WaWb;
|
||||||
|
|
||||||
for (i = 0; i < GITS_BASER_NR_REGS; i++) {
|
for (i = 0; i < GITS_BASER_NR_REGS; i++) {
|
||||||
u64 val = readq_relaxed(its->base + GITS_BASER + i * 8);
|
u64 val = readq_relaxed(its->base + GITS_BASER + i * 8);
|
||||||
|
@ -848,7 +849,7 @@ retry_baser:
|
||||||
val = (virt_to_phys(base) |
|
val = (virt_to_phys(base) |
|
||||||
(type << GITS_BASER_TYPE_SHIFT) |
|
(type << GITS_BASER_TYPE_SHIFT) |
|
||||||
((entry_size - 1) << GITS_BASER_ENTRY_SIZE_SHIFT) |
|
((entry_size - 1) << GITS_BASER_ENTRY_SIZE_SHIFT) |
|
||||||
GITS_BASER_WaWb |
|
cache |
|
||||||
shr |
|
shr |
|
||||||
GITS_BASER_VALID);
|
GITS_BASER_VALID);
|
||||||
|
|
||||||
|
@ -874,9 +875,12 @@ retry_baser:
|
||||||
* Shareability didn't stick. Just use
|
* Shareability didn't stick. Just use
|
||||||
* whatever the read reported, which is likely
|
* whatever the read reported, which is likely
|
||||||
* to be the only thing this redistributor
|
* to be the only thing this redistributor
|
||||||
* supports.
|
* supports. If that's zero, make it
|
||||||
|
* non-cacheable as well.
|
||||||
*/
|
*/
|
||||||
shr = tmp & GITS_BASER_SHAREABILITY_MASK;
|
shr = tmp & GITS_BASER_SHAREABILITY_MASK;
|
||||||
|
if (!shr)
|
||||||
|
cache = GITS_BASER_nC;
|
||||||
goto retry_baser;
|
goto retry_baser;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -980,16 +984,39 @@ static void its_cpu_init_lpis(void)
|
||||||
tmp = readq_relaxed(rbase + GICR_PROPBASER);
|
tmp = readq_relaxed(rbase + GICR_PROPBASER);
|
||||||
|
|
||||||
if ((tmp ^ val) & GICR_PROPBASER_SHAREABILITY_MASK) {
|
if ((tmp ^ val) & GICR_PROPBASER_SHAREABILITY_MASK) {
|
||||||
|
if (!(tmp & GICR_PROPBASER_SHAREABILITY_MASK)) {
|
||||||
|
/*
|
||||||
|
* The HW reports non-shareable, we must
|
||||||
|
* remove the cacheability attributes as
|
||||||
|
* well.
|
||||||
|
*/
|
||||||
|
val &= ~(GICR_PROPBASER_SHAREABILITY_MASK |
|
||||||
|
GICR_PROPBASER_CACHEABILITY_MASK);
|
||||||
|
val |= GICR_PROPBASER_nC;
|
||||||
|
writeq_relaxed(val, rbase + GICR_PROPBASER);
|
||||||
|
}
|
||||||
pr_info_once("GIC: using cache flushing for LPI property table\n");
|
pr_info_once("GIC: using cache flushing for LPI property table\n");
|
||||||
gic_rdists->flags |= RDIST_FLAGS_PROPBASE_NEEDS_FLUSHING;
|
gic_rdists->flags |= RDIST_FLAGS_PROPBASE_NEEDS_FLUSHING;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* set PENDBASE */
|
/* set PENDBASE */
|
||||||
val = (page_to_phys(pend_page) |
|
val = (page_to_phys(pend_page) |
|
||||||
GICR_PROPBASER_InnerShareable |
|
GICR_PENDBASER_InnerShareable |
|
||||||
GICR_PROPBASER_WaWb);
|
GICR_PENDBASER_WaWb);
|
||||||
|
|
||||||
writeq_relaxed(val, rbase + GICR_PENDBASER);
|
writeq_relaxed(val, rbase + GICR_PENDBASER);
|
||||||
|
tmp = readq_relaxed(rbase + GICR_PENDBASER);
|
||||||
|
|
||||||
|
if (!(tmp & GICR_PENDBASER_SHAREABILITY_MASK)) {
|
||||||
|
/*
|
||||||
|
* The HW reports non-shareable, we must remove the
|
||||||
|
* cacheability attributes as well.
|
||||||
|
*/
|
||||||
|
val &= ~(GICR_PENDBASER_SHAREABILITY_MASK |
|
||||||
|
GICR_PENDBASER_CACHEABILITY_MASK);
|
||||||
|
val |= GICR_PENDBASER_nC;
|
||||||
|
writeq_relaxed(val, rbase + GICR_PENDBASER);
|
||||||
|
}
|
||||||
|
|
||||||
/* Enable LPIs */
|
/* Enable LPIs */
|
||||||
val = readl_relaxed(rbase + GICR_CTLR);
|
val = readl_relaxed(rbase + GICR_CTLR);
|
||||||
|
@ -1026,7 +1053,7 @@ static void its_cpu_init_collection(void)
|
||||||
* This ITS wants a linear CPU number.
|
* This ITS wants a linear CPU number.
|
||||||
*/
|
*/
|
||||||
target = readq_relaxed(gic_data_rdist_rd_base() + GICR_TYPER);
|
target = readq_relaxed(gic_data_rdist_rd_base() + GICR_TYPER);
|
||||||
target = GICR_TYPER_CPU_NUMBER(target);
|
target = GICR_TYPER_CPU_NUMBER(target) << 16;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Perform collection mapping */
|
/* Perform collection mapping */
|
||||||
|
@ -1422,14 +1449,26 @@ static int its_probe(struct device_node *node, struct irq_domain *parent)
|
||||||
|
|
||||||
writeq_relaxed(baser, its->base + GITS_CBASER);
|
writeq_relaxed(baser, its->base + GITS_CBASER);
|
||||||
tmp = readq_relaxed(its->base + GITS_CBASER);
|
tmp = readq_relaxed(its->base + GITS_CBASER);
|
||||||
writeq_relaxed(0, its->base + GITS_CWRITER);
|
|
||||||
writel_relaxed(GITS_CTLR_ENABLE, its->base + GITS_CTLR);
|
|
||||||
|
|
||||||
if ((tmp ^ baser) & GITS_BASER_SHAREABILITY_MASK) {
|
if ((tmp ^ baser) & GITS_CBASER_SHAREABILITY_MASK) {
|
||||||
|
if (!(tmp & GITS_CBASER_SHAREABILITY_MASK)) {
|
||||||
|
/*
|
||||||
|
* The HW reports non-shareable, we must
|
||||||
|
* remove the cacheability attributes as
|
||||||
|
* well.
|
||||||
|
*/
|
||||||
|
baser &= ~(GITS_CBASER_SHAREABILITY_MASK |
|
||||||
|
GITS_CBASER_CACHEABILITY_MASK);
|
||||||
|
baser |= GITS_CBASER_nC;
|
||||||
|
writeq_relaxed(baser, its->base + GITS_CBASER);
|
||||||
|
}
|
||||||
pr_info("ITS: using cache flushing for cmd queue\n");
|
pr_info("ITS: using cache flushing for cmd queue\n");
|
||||||
its->flags |= ITS_FLAGS_CMDQ_NEEDS_FLUSHING;
|
its->flags |= ITS_FLAGS_CMDQ_NEEDS_FLUSHING;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
writeq_relaxed(0, its->base + GITS_CWRITER);
|
||||||
|
writel_relaxed(GITS_CTLR_ENABLE, its->base + GITS_CTLR);
|
||||||
|
|
||||||
if (of_property_read_bool(its->msi_chip.of_node, "msi-controller")) {
|
if (of_property_read_bool(its->msi_chip.of_node, "msi-controller")) {
|
||||||
its->domain = irq_domain_add_tree(NULL, &its_domain_ops, its);
|
its->domain = irq_domain_add_tree(NULL, &its_domain_ops, its);
|
||||||
if (!its->domain) {
|
if (!its->domain) {
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
config LGUEST
|
config LGUEST
|
||||||
tristate "Linux hypervisor example code"
|
tristate "Linux hypervisor example code"
|
||||||
depends on X86_32 && EVENTFD && TTY
|
depends on X86_32 && EVENTFD && TTY && PCI_DIRECT
|
||||||
select HVC_DRIVER
|
select HVC_DRIVER
|
||||||
---help---
|
---help---
|
||||||
This is a very simple module which allows you to run
|
This is a very simple module which allows you to run
|
||||||
|
|
|
@ -3850,7 +3850,8 @@ static inline int bond_slave_override(struct bonding *bond,
|
||||||
/* Find out if any slaves have the same mapping as this skb. */
|
/* Find out if any slaves have the same mapping as this skb. */
|
||||||
bond_for_each_slave_rcu(bond, slave, iter) {
|
bond_for_each_slave_rcu(bond, slave, iter) {
|
||||||
if (slave->queue_id == skb->queue_mapping) {
|
if (slave->queue_id == skb->queue_mapping) {
|
||||||
if (bond_slave_can_tx(slave)) {
|
if (bond_slave_is_up(slave) &&
|
||||||
|
slave->link == BOND_LINK_UP) {
|
||||||
bond_dev_queue_xmit(bond, skb, slave->dev);
|
bond_dev_queue_xmit(bond, skb, slave->dev);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
|
@ -592,13 +592,12 @@ static int flexcan_poll_state(struct net_device *dev, u32 reg_esr)
|
||||||
rx_state = unlikely(reg_esr & FLEXCAN_ESR_RX_WRN) ?
|
rx_state = unlikely(reg_esr & FLEXCAN_ESR_RX_WRN) ?
|
||||||
CAN_STATE_ERROR_WARNING : CAN_STATE_ERROR_ACTIVE;
|
CAN_STATE_ERROR_WARNING : CAN_STATE_ERROR_ACTIVE;
|
||||||
new_state = max(tx_state, rx_state);
|
new_state = max(tx_state, rx_state);
|
||||||
} else if (unlikely(flt == FLEXCAN_ESR_FLT_CONF_PASSIVE)) {
|
} else {
|
||||||
__flexcan_get_berr_counter(dev, &bec);
|
__flexcan_get_berr_counter(dev, &bec);
|
||||||
new_state = CAN_STATE_ERROR_PASSIVE;
|
new_state = flt == FLEXCAN_ESR_FLT_CONF_PASSIVE ?
|
||||||
|
CAN_STATE_ERROR_PASSIVE : CAN_STATE_BUS_OFF;
|
||||||
rx_state = bec.rxerr >= bec.txerr ? new_state : 0;
|
rx_state = bec.rxerr >= bec.txerr ? new_state : 0;
|
||||||
tx_state = bec.rxerr <= bec.txerr ? new_state : 0;
|
tx_state = bec.rxerr <= bec.txerr ? new_state : 0;
|
||||||
} else {
|
|
||||||
new_state = CAN_STATE_BUS_OFF;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* state hasn't changed */
|
/* state hasn't changed */
|
||||||
|
@ -1158,12 +1157,19 @@ static int flexcan_probe(struct platform_device *pdev)
|
||||||
const struct flexcan_devtype_data *devtype_data;
|
const struct flexcan_devtype_data *devtype_data;
|
||||||
struct net_device *dev;
|
struct net_device *dev;
|
||||||
struct flexcan_priv *priv;
|
struct flexcan_priv *priv;
|
||||||
|
struct regulator *reg_xceiver;
|
||||||
struct resource *mem;
|
struct resource *mem;
|
||||||
struct clk *clk_ipg = NULL, *clk_per = NULL;
|
struct clk *clk_ipg = NULL, *clk_per = NULL;
|
||||||
void __iomem *base;
|
void __iomem *base;
|
||||||
int err, irq;
|
int err, irq;
|
||||||
u32 clock_freq = 0;
|
u32 clock_freq = 0;
|
||||||
|
|
||||||
|
reg_xceiver = devm_regulator_get(&pdev->dev, "xceiver");
|
||||||
|
if (PTR_ERR(reg_xceiver) == -EPROBE_DEFER)
|
||||||
|
return -EPROBE_DEFER;
|
||||||
|
else if (IS_ERR(reg_xceiver))
|
||||||
|
reg_xceiver = NULL;
|
||||||
|
|
||||||
if (pdev->dev.of_node)
|
if (pdev->dev.of_node)
|
||||||
of_property_read_u32(pdev->dev.of_node,
|
of_property_read_u32(pdev->dev.of_node,
|
||||||
"clock-frequency", &clock_freq);
|
"clock-frequency", &clock_freq);
|
||||||
|
@ -1224,9 +1230,7 @@ static int flexcan_probe(struct platform_device *pdev)
|
||||||
priv->pdata = dev_get_platdata(&pdev->dev);
|
priv->pdata = dev_get_platdata(&pdev->dev);
|
||||||
priv->devtype_data = devtype_data;
|
priv->devtype_data = devtype_data;
|
||||||
|
|
||||||
priv->reg_xceiver = devm_regulator_get(&pdev->dev, "xceiver");
|
priv->reg_xceiver = reg_xceiver;
|
||||||
if (IS_ERR(priv->reg_xceiver))
|
|
||||||
priv->reg_xceiver = NULL;
|
|
||||||
|
|
||||||
netif_napi_add(dev, &priv->napi, flexcan_poll, FLEXCAN_NAPI_WEIGHT);
|
netif_napi_add(dev, &priv->napi, flexcan_poll, FLEXCAN_NAPI_WEIGHT);
|
||||||
|
|
||||||
|
|
|
@ -901,6 +901,8 @@ static int gs_usb_probe(struct usb_interface *intf, const struct usb_device_id *
|
||||||
}
|
}
|
||||||
|
|
||||||
dev = kzalloc(sizeof(*dev), GFP_KERNEL);
|
dev = kzalloc(sizeof(*dev), GFP_KERNEL);
|
||||||
|
if (!dev)
|
||||||
|
return -ENOMEM;
|
||||||
init_usb_anchor(&dev->rx_submitted);
|
init_usb_anchor(&dev->rx_submitted);
|
||||||
|
|
||||||
atomic_set(&dev->active_channels, 0);
|
atomic_set(&dev->active_channels, 0);
|
||||||
|
|
|
@ -25,7 +25,6 @@
|
||||||
#include <linux/can/dev.h>
|
#include <linux/can/dev.h>
|
||||||
#include <linux/can/error.h>
|
#include <linux/can/error.h>
|
||||||
|
|
||||||
#define MAX_TX_URBS 16
|
|
||||||
#define MAX_RX_URBS 4
|
#define MAX_RX_URBS 4
|
||||||
#define START_TIMEOUT 1000 /* msecs */
|
#define START_TIMEOUT 1000 /* msecs */
|
||||||
#define STOP_TIMEOUT 1000 /* msecs */
|
#define STOP_TIMEOUT 1000 /* msecs */
|
||||||
|
@ -443,6 +442,7 @@ struct kvaser_usb_error_summary {
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
|
/* Context for an outstanding, not yet ACKed, transmission */
|
||||||
struct kvaser_usb_tx_urb_context {
|
struct kvaser_usb_tx_urb_context {
|
||||||
struct kvaser_usb_net_priv *priv;
|
struct kvaser_usb_net_priv *priv;
|
||||||
u32 echo_index;
|
u32 echo_index;
|
||||||
|
@ -456,8 +456,13 @@ struct kvaser_usb {
|
||||||
struct usb_endpoint_descriptor *bulk_in, *bulk_out;
|
struct usb_endpoint_descriptor *bulk_in, *bulk_out;
|
||||||
struct usb_anchor rx_submitted;
|
struct usb_anchor rx_submitted;
|
||||||
|
|
||||||
|
/* @max_tx_urbs: Firmware-reported maximum number of oustanding,
|
||||||
|
* not yet ACKed, transmissions on this device. This value is
|
||||||
|
* also used as a sentinel for marking free tx contexts.
|
||||||
|
*/
|
||||||
u32 fw_version;
|
u32 fw_version;
|
||||||
unsigned int nchannels;
|
unsigned int nchannels;
|
||||||
|
unsigned int max_tx_urbs;
|
||||||
enum kvaser_usb_family family;
|
enum kvaser_usb_family family;
|
||||||
|
|
||||||
bool rxinitdone;
|
bool rxinitdone;
|
||||||
|
@ -467,19 +472,18 @@ struct kvaser_usb {
|
||||||
|
|
||||||
struct kvaser_usb_net_priv {
|
struct kvaser_usb_net_priv {
|
||||||
struct can_priv can;
|
struct can_priv can;
|
||||||
|
struct can_berr_counter bec;
|
||||||
spinlock_t tx_contexts_lock;
|
|
||||||
int active_tx_contexts;
|
|
||||||
struct kvaser_usb_tx_urb_context tx_contexts[MAX_TX_URBS];
|
|
||||||
|
|
||||||
struct usb_anchor tx_submitted;
|
|
||||||
struct completion start_comp, stop_comp;
|
|
||||||
|
|
||||||
struct kvaser_usb *dev;
|
struct kvaser_usb *dev;
|
||||||
struct net_device *netdev;
|
struct net_device *netdev;
|
||||||
int channel;
|
int channel;
|
||||||
|
|
||||||
struct can_berr_counter bec;
|
struct completion start_comp, stop_comp;
|
||||||
|
struct usb_anchor tx_submitted;
|
||||||
|
|
||||||
|
spinlock_t tx_contexts_lock;
|
||||||
|
int active_tx_contexts;
|
||||||
|
struct kvaser_usb_tx_urb_context tx_contexts[];
|
||||||
};
|
};
|
||||||
|
|
||||||
static const struct usb_device_id kvaser_usb_table[] = {
|
static const struct usb_device_id kvaser_usb_table[] = {
|
||||||
|
@ -592,8 +596,8 @@ static int kvaser_usb_wait_msg(const struct kvaser_usb *dev, u8 id,
|
||||||
* for further details.
|
* for further details.
|
||||||
*/
|
*/
|
||||||
if (tmp->len == 0) {
|
if (tmp->len == 0) {
|
||||||
pos = round_up(pos,
|
pos = round_up(pos, le16_to_cpu(dev->bulk_in->
|
||||||
dev->bulk_in->wMaxPacketSize);
|
wMaxPacketSize));
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -657,9 +661,13 @@ static int kvaser_usb_get_software_info(struct kvaser_usb *dev)
|
||||||
switch (dev->family) {
|
switch (dev->family) {
|
||||||
case KVASER_LEAF:
|
case KVASER_LEAF:
|
||||||
dev->fw_version = le32_to_cpu(msg.u.leaf.softinfo.fw_version);
|
dev->fw_version = le32_to_cpu(msg.u.leaf.softinfo.fw_version);
|
||||||
|
dev->max_tx_urbs =
|
||||||
|
le16_to_cpu(msg.u.leaf.softinfo.max_outstanding_tx);
|
||||||
break;
|
break;
|
||||||
case KVASER_USBCAN:
|
case KVASER_USBCAN:
|
||||||
dev->fw_version = le32_to_cpu(msg.u.usbcan.softinfo.fw_version);
|
dev->fw_version = le32_to_cpu(msg.u.usbcan.softinfo.fw_version);
|
||||||
|
dev->max_tx_urbs =
|
||||||
|
le16_to_cpu(msg.u.usbcan.softinfo.max_outstanding_tx);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -715,7 +723,7 @@ static void kvaser_usb_tx_acknowledge(const struct kvaser_usb *dev,
|
||||||
|
|
||||||
stats = &priv->netdev->stats;
|
stats = &priv->netdev->stats;
|
||||||
|
|
||||||
context = &priv->tx_contexts[tid % MAX_TX_URBS];
|
context = &priv->tx_contexts[tid % dev->max_tx_urbs];
|
||||||
|
|
||||||
/* Sometimes the state change doesn't come after a bus-off event */
|
/* Sometimes the state change doesn't come after a bus-off event */
|
||||||
if (priv->can.restart_ms &&
|
if (priv->can.restart_ms &&
|
||||||
|
@ -744,7 +752,7 @@ static void kvaser_usb_tx_acknowledge(const struct kvaser_usb *dev,
|
||||||
spin_lock_irqsave(&priv->tx_contexts_lock, flags);
|
spin_lock_irqsave(&priv->tx_contexts_lock, flags);
|
||||||
|
|
||||||
can_get_echo_skb(priv->netdev, context->echo_index);
|
can_get_echo_skb(priv->netdev, context->echo_index);
|
||||||
context->echo_index = MAX_TX_URBS;
|
context->echo_index = dev->max_tx_urbs;
|
||||||
--priv->active_tx_contexts;
|
--priv->active_tx_contexts;
|
||||||
netif_wake_queue(priv->netdev);
|
netif_wake_queue(priv->netdev);
|
||||||
|
|
||||||
|
@ -1329,7 +1337,8 @@ static void kvaser_usb_read_bulk_callback(struct urb *urb)
|
||||||
* number of events in case of a heavy rx load on the bus.
|
* number of events in case of a heavy rx load on the bus.
|
||||||
*/
|
*/
|
||||||
if (msg->len == 0) {
|
if (msg->len == 0) {
|
||||||
pos = round_up(pos, dev->bulk_in->wMaxPacketSize);
|
pos = round_up(pos, le16_to_cpu(dev->bulk_in->
|
||||||
|
wMaxPacketSize));
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1512,11 +1521,13 @@ error:
|
||||||
|
|
||||||
static void kvaser_usb_reset_tx_urb_contexts(struct kvaser_usb_net_priv *priv)
|
static void kvaser_usb_reset_tx_urb_contexts(struct kvaser_usb_net_priv *priv)
|
||||||
{
|
{
|
||||||
int i;
|
int i, max_tx_urbs;
|
||||||
|
|
||||||
|
max_tx_urbs = priv->dev->max_tx_urbs;
|
||||||
|
|
||||||
priv->active_tx_contexts = 0;
|
priv->active_tx_contexts = 0;
|
||||||
for (i = 0; i < MAX_TX_URBS; i++)
|
for (i = 0; i < max_tx_urbs; i++)
|
||||||
priv->tx_contexts[i].echo_index = MAX_TX_URBS;
|
priv->tx_contexts[i].echo_index = max_tx_urbs;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* This method might sleep. Do not call it in the atomic context
|
/* This method might sleep. Do not call it in the atomic context
|
||||||
|
@ -1702,14 +1713,14 @@ static netdev_tx_t kvaser_usb_start_xmit(struct sk_buff *skb,
|
||||||
*msg_tx_can_flags |= MSG_FLAG_REMOTE_FRAME;
|
*msg_tx_can_flags |= MSG_FLAG_REMOTE_FRAME;
|
||||||
|
|
||||||
spin_lock_irqsave(&priv->tx_contexts_lock, flags);
|
spin_lock_irqsave(&priv->tx_contexts_lock, flags);
|
||||||
for (i = 0; i < ARRAY_SIZE(priv->tx_contexts); i++) {
|
for (i = 0; i < dev->max_tx_urbs; i++) {
|
||||||
if (priv->tx_contexts[i].echo_index == MAX_TX_URBS) {
|
if (priv->tx_contexts[i].echo_index == dev->max_tx_urbs) {
|
||||||
context = &priv->tx_contexts[i];
|
context = &priv->tx_contexts[i];
|
||||||
|
|
||||||
context->echo_index = i;
|
context->echo_index = i;
|
||||||
can_put_echo_skb(skb, netdev, context->echo_index);
|
can_put_echo_skb(skb, netdev, context->echo_index);
|
||||||
++priv->active_tx_contexts;
|
++priv->active_tx_contexts;
|
||||||
if (priv->active_tx_contexts >= MAX_TX_URBS)
|
if (priv->active_tx_contexts >= dev->max_tx_urbs)
|
||||||
netif_stop_queue(netdev);
|
netif_stop_queue(netdev);
|
||||||
|
|
||||||
break;
|
break;
|
||||||
|
@ -1743,7 +1754,7 @@ static netdev_tx_t kvaser_usb_start_xmit(struct sk_buff *skb,
|
||||||
spin_lock_irqsave(&priv->tx_contexts_lock, flags);
|
spin_lock_irqsave(&priv->tx_contexts_lock, flags);
|
||||||
|
|
||||||
can_free_echo_skb(netdev, context->echo_index);
|
can_free_echo_skb(netdev, context->echo_index);
|
||||||
context->echo_index = MAX_TX_URBS;
|
context->echo_index = dev->max_tx_urbs;
|
||||||
--priv->active_tx_contexts;
|
--priv->active_tx_contexts;
|
||||||
netif_wake_queue(netdev);
|
netif_wake_queue(netdev);
|
||||||
|
|
||||||
|
@ -1881,7 +1892,9 @@ static int kvaser_usb_init_one(struct usb_interface *intf,
|
||||||
if (err)
|
if (err)
|
||||||
return err;
|
return err;
|
||||||
|
|
||||||
netdev = alloc_candev(sizeof(*priv), MAX_TX_URBS);
|
netdev = alloc_candev(sizeof(*priv) +
|
||||||
|
dev->max_tx_urbs * sizeof(*priv->tx_contexts),
|
||||||
|
dev->max_tx_urbs);
|
||||||
if (!netdev) {
|
if (!netdev) {
|
||||||
dev_err(&intf->dev, "Cannot alloc candev\n");
|
dev_err(&intf->dev, "Cannot alloc candev\n");
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
@ -2009,6 +2022,13 @@ static int kvaser_usb_probe(struct usb_interface *intf,
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
dev_dbg(&intf->dev, "Firmware version: %d.%d.%d\n",
|
||||||
|
((dev->fw_version >> 24) & 0xff),
|
||||||
|
((dev->fw_version >> 16) & 0xff),
|
||||||
|
(dev->fw_version & 0xffff));
|
||||||
|
|
||||||
|
dev_dbg(&intf->dev, "Max oustanding tx = %d URBs\n", dev->max_tx_urbs);
|
||||||
|
|
||||||
err = kvaser_usb_get_card_info(dev);
|
err = kvaser_usb_get_card_info(dev);
|
||||||
if (err) {
|
if (err) {
|
||||||
dev_err(&intf->dev,
|
dev_err(&intf->dev,
|
||||||
|
@ -2016,11 +2036,6 @@ static int kvaser_usb_probe(struct usb_interface *intf,
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
dev_dbg(&intf->dev, "Firmware version: %d.%d.%d\n",
|
|
||||||
((dev->fw_version >> 24) & 0xff),
|
|
||||||
((dev->fw_version >> 16) & 0xff),
|
|
||||||
(dev->fw_version & 0xffff));
|
|
||||||
|
|
||||||
for (i = 0; i < dev->nchannels; i++) {
|
for (i = 0; i < dev->nchannels; i++) {
|
||||||
err = kvaser_usb_init_one(intf, id, i);
|
err = kvaser_usb_init_one(intf, id, i);
|
||||||
if (err) {
|
if (err) {
|
||||||
|
|
|
@ -26,8 +26,8 @@
|
||||||
#define PUCAN_CMD_FILTER_STD 0x008
|
#define PUCAN_CMD_FILTER_STD 0x008
|
||||||
#define PUCAN_CMD_TX_ABORT 0x009
|
#define PUCAN_CMD_TX_ABORT 0x009
|
||||||
#define PUCAN_CMD_WR_ERR_CNT 0x00a
|
#define PUCAN_CMD_WR_ERR_CNT 0x00a
|
||||||
#define PUCAN_CMD_RX_FRAME_ENABLE 0x00b
|
#define PUCAN_CMD_SET_EN_OPTION 0x00b
|
||||||
#define PUCAN_CMD_RX_FRAME_DISABLE 0x00c
|
#define PUCAN_CMD_CLR_DIS_OPTION 0x00c
|
||||||
#define PUCAN_CMD_END_OF_COLLECTION 0x3ff
|
#define PUCAN_CMD_END_OF_COLLECTION 0x3ff
|
||||||
|
|
||||||
/* uCAN received messages list */
|
/* uCAN received messages list */
|
||||||
|
@ -101,14 +101,15 @@ struct __packed pucan_wr_err_cnt {
|
||||||
u16 unused;
|
u16 unused;
|
||||||
};
|
};
|
||||||
|
|
||||||
/* uCAN RX_FRAME_ENABLE command fields */
|
/* uCAN SET_EN/CLR_DIS _OPTION command fields */
|
||||||
#define PUCAN_FLTEXT_ERROR 0x0001
|
#define PUCAN_OPTION_ERROR 0x0001
|
||||||
#define PUCAN_FLTEXT_BUSLOAD 0x0002
|
#define PUCAN_OPTION_BUSLOAD 0x0002
|
||||||
|
#define PUCAN_OPTION_CANDFDISO 0x0004
|
||||||
|
|
||||||
struct __packed pucan_filter_ext {
|
struct __packed pucan_options {
|
||||||
__le16 opcode_channel;
|
__le16 opcode_channel;
|
||||||
|
|
||||||
__le16 ext_mask;
|
__le16 options;
|
||||||
u32 unused;
|
u32 unused;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -110,13 +110,13 @@ struct __packed pcan_ufd_led {
|
||||||
u8 unused[5];
|
u8 unused[5];
|
||||||
};
|
};
|
||||||
|
|
||||||
/* Extended usage of uCAN commands CMD_RX_FRAME_xxxABLE for PCAN-USB Pro FD */
|
/* Extended usage of uCAN commands CMD_xxx_xx_OPTION for PCAN-USB Pro FD */
|
||||||
#define PCAN_UFD_FLTEXT_CALIBRATION 0x8000
|
#define PCAN_UFD_FLTEXT_CALIBRATION 0x8000
|
||||||
|
|
||||||
struct __packed pcan_ufd_filter_ext {
|
struct __packed pcan_ufd_options {
|
||||||
__le16 opcode_channel;
|
__le16 opcode_channel;
|
||||||
|
|
||||||
__le16 ext_mask;
|
__le16 ucan_mask;
|
||||||
u16 unused;
|
u16 unused;
|
||||||
__le16 usb_mask;
|
__le16 usb_mask;
|
||||||
};
|
};
|
||||||
|
@ -251,6 +251,27 @@ static int pcan_usb_fd_build_restart_cmd(struct peak_usb_device *dev, u8 *buf)
|
||||||
/* moves the pointer forward */
|
/* moves the pointer forward */
|
||||||
pc += sizeof(struct pucan_wr_err_cnt);
|
pc += sizeof(struct pucan_wr_err_cnt);
|
||||||
|
|
||||||
|
/* add command to switch from ISO to non-ISO mode, if fw allows it */
|
||||||
|
if (dev->can.ctrlmode_supported & CAN_CTRLMODE_FD_NON_ISO) {
|
||||||
|
struct pucan_options *puo = (struct pucan_options *)pc;
|
||||||
|
|
||||||
|
puo->opcode_channel =
|
||||||
|
(dev->can.ctrlmode & CAN_CTRLMODE_FD_NON_ISO) ?
|
||||||
|
pucan_cmd_opcode_channel(dev,
|
||||||
|
PUCAN_CMD_CLR_DIS_OPTION) :
|
||||||
|
pucan_cmd_opcode_channel(dev, PUCAN_CMD_SET_EN_OPTION);
|
||||||
|
|
||||||
|
puo->options = cpu_to_le16(PUCAN_OPTION_CANDFDISO);
|
||||||
|
|
||||||
|
/* to be sure that no other extended bits will be taken into
|
||||||
|
* account
|
||||||
|
*/
|
||||||
|
puo->unused = 0;
|
||||||
|
|
||||||
|
/* moves the pointer forward */
|
||||||
|
pc += sizeof(struct pucan_options);
|
||||||
|
}
|
||||||
|
|
||||||
/* next, go back to operational mode */
|
/* next, go back to operational mode */
|
||||||
cmd = (struct pucan_command *)pc;
|
cmd = (struct pucan_command *)pc;
|
||||||
cmd->opcode_channel = pucan_cmd_opcode_channel(dev,
|
cmd->opcode_channel = pucan_cmd_opcode_channel(dev,
|
||||||
|
@ -321,21 +342,21 @@ static int pcan_usb_fd_set_filter_std(struct peak_usb_device *dev, int idx,
|
||||||
return pcan_usb_fd_send_cmd(dev, cmd);
|
return pcan_usb_fd_send_cmd(dev, cmd);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* set/unset notifications filter:
|
/* set/unset options
|
||||||
*
|
*
|
||||||
* onoff sets(1)/unset(0) notifications
|
* onoff set(1)/unset(0) options
|
||||||
* mask each bit defines a kind of notification to set/unset
|
* mask each bit defines a kind of options to set/unset
|
||||||
*/
|
*/
|
||||||
static int pcan_usb_fd_set_filter_ext(struct peak_usb_device *dev,
|
static int pcan_usb_fd_set_options(struct peak_usb_device *dev,
|
||||||
bool onoff, u16 ext_mask, u16 usb_mask)
|
bool onoff, u16 ucan_mask, u16 usb_mask)
|
||||||
{
|
{
|
||||||
struct pcan_ufd_filter_ext *cmd = pcan_usb_fd_cmd_buffer(dev);
|
struct pcan_ufd_options *cmd = pcan_usb_fd_cmd_buffer(dev);
|
||||||
|
|
||||||
cmd->opcode_channel = pucan_cmd_opcode_channel(dev,
|
cmd->opcode_channel = pucan_cmd_opcode_channel(dev,
|
||||||
(onoff) ? PUCAN_CMD_RX_FRAME_ENABLE :
|
(onoff) ? PUCAN_CMD_SET_EN_OPTION :
|
||||||
PUCAN_CMD_RX_FRAME_DISABLE);
|
PUCAN_CMD_CLR_DIS_OPTION);
|
||||||
|
|
||||||
cmd->ext_mask = cpu_to_le16(ext_mask);
|
cmd->ucan_mask = cpu_to_le16(ucan_mask);
|
||||||
cmd->usb_mask = cpu_to_le16(usb_mask);
|
cmd->usb_mask = cpu_to_le16(usb_mask);
|
||||||
|
|
||||||
/* send the command */
|
/* send the command */
|
||||||
|
@ -770,9 +791,9 @@ static int pcan_usb_fd_start(struct peak_usb_device *dev)
|
||||||
&pcan_usb_pro_fd);
|
&pcan_usb_pro_fd);
|
||||||
|
|
||||||
/* enable USB calibration messages */
|
/* enable USB calibration messages */
|
||||||
err = pcan_usb_fd_set_filter_ext(dev, 1,
|
err = pcan_usb_fd_set_options(dev, 1,
|
||||||
PUCAN_FLTEXT_ERROR,
|
PUCAN_OPTION_ERROR,
|
||||||
PCAN_UFD_FLTEXT_CALIBRATION);
|
PCAN_UFD_FLTEXT_CALIBRATION);
|
||||||
}
|
}
|
||||||
|
|
||||||
pdev->usb_if->dev_opened_count++;
|
pdev->usb_if->dev_opened_count++;
|
||||||
|
@ -806,9 +827,9 @@ static int pcan_usb_fd_stop(struct peak_usb_device *dev)
|
||||||
|
|
||||||
/* turn off special msgs for that interface if no other dev opened */
|
/* turn off special msgs for that interface if no other dev opened */
|
||||||
if (pdev->usb_if->dev_opened_count == 1)
|
if (pdev->usb_if->dev_opened_count == 1)
|
||||||
pcan_usb_fd_set_filter_ext(dev, 0,
|
pcan_usb_fd_set_options(dev, 0,
|
||||||
PUCAN_FLTEXT_ERROR,
|
PUCAN_OPTION_ERROR,
|
||||||
PCAN_UFD_FLTEXT_CALIBRATION);
|
PCAN_UFD_FLTEXT_CALIBRATION);
|
||||||
pdev->usb_if->dev_opened_count--;
|
pdev->usb_if->dev_opened_count--;
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -860,8 +881,14 @@ static int pcan_usb_fd_init(struct peak_usb_device *dev)
|
||||||
pdev->usb_if->fw_info.fw_version[2],
|
pdev->usb_if->fw_info.fw_version[2],
|
||||||
dev->adapter->ctrl_count);
|
dev->adapter->ctrl_count);
|
||||||
|
|
||||||
/* the currently supported hw is non-ISO */
|
/* check for ability to switch between ISO/non-ISO modes */
|
||||||
dev->can.ctrlmode = CAN_CTRLMODE_FD_NON_ISO;
|
if (pdev->usb_if->fw_info.fw_version[0] >= 2) {
|
||||||
|
/* firmware >= 2.x supports ISO/non-ISO switching */
|
||||||
|
dev->can.ctrlmode_supported |= CAN_CTRLMODE_FD_NON_ISO;
|
||||||
|
} else {
|
||||||
|
/* firmware < 2.x only supports fixed(!) non-ISO */
|
||||||
|
dev->can.ctrlmode |= CAN_CTRLMODE_FD_NON_ISO;
|
||||||
|
}
|
||||||
|
|
||||||
/* tell the hardware the can driver is running */
|
/* tell the hardware the can driver is running */
|
||||||
err = pcan_usb_fd_drv_loaded(dev, 1);
|
err = pcan_usb_fd_drv_loaded(dev, 1);
|
||||||
|
@ -937,9 +964,9 @@ static void pcan_usb_fd_exit(struct peak_usb_device *dev)
|
||||||
if (dev->ctrl_idx == 0) {
|
if (dev->ctrl_idx == 0) {
|
||||||
/* turn off calibration message if any device were opened */
|
/* turn off calibration message if any device were opened */
|
||||||
if (pdev->usb_if->dev_opened_count > 0)
|
if (pdev->usb_if->dev_opened_count > 0)
|
||||||
pcan_usb_fd_set_filter_ext(dev, 0,
|
pcan_usb_fd_set_options(dev, 0,
|
||||||
PUCAN_FLTEXT_ERROR,
|
PUCAN_OPTION_ERROR,
|
||||||
PCAN_UFD_FLTEXT_CALIBRATION);
|
PCAN_UFD_FLTEXT_CALIBRATION);
|
||||||
|
|
||||||
/* tell USB adapter that the driver is being unloaded */
|
/* tell USB adapter that the driver is being unloaded */
|
||||||
pcan_usb_fd_drv_loaded(dev, 0);
|
pcan_usb_fd_drv_loaded(dev, 0);
|
||||||
|
|
|
@ -1811,7 +1811,7 @@ struct bnx2x {
|
||||||
int stats_state;
|
int stats_state;
|
||||||
|
|
||||||
/* used for synchronization of concurrent threads statistics handling */
|
/* used for synchronization of concurrent threads statistics handling */
|
||||||
spinlock_t stats_lock;
|
struct mutex stats_lock;
|
||||||
|
|
||||||
/* used by dmae command loader */
|
/* used by dmae command loader */
|
||||||
struct dmae_command stats_dmae;
|
struct dmae_command stats_dmae;
|
||||||
|
@ -1935,8 +1935,6 @@ struct bnx2x {
|
||||||
|
|
||||||
int fp_array_size;
|
int fp_array_size;
|
||||||
u32 dump_preset_idx;
|
u32 dump_preset_idx;
|
||||||
bool stats_started;
|
|
||||||
struct semaphore stats_sema;
|
|
||||||
|
|
||||||
u8 phys_port_id[ETH_ALEN];
|
u8 phys_port_id[ETH_ALEN];
|
||||||
|
|
||||||
|
|
|
@ -129,8 +129,8 @@ struct bnx2x_mac_vals {
|
||||||
u32 xmac_val;
|
u32 xmac_val;
|
||||||
u32 emac_addr;
|
u32 emac_addr;
|
||||||
u32 emac_val;
|
u32 emac_val;
|
||||||
u32 umac_addr;
|
u32 umac_addr[2];
|
||||||
u32 umac_val;
|
u32 umac_val[2];
|
||||||
u32 bmac_addr;
|
u32 bmac_addr;
|
||||||
u32 bmac_val[2];
|
u32 bmac_val[2];
|
||||||
};
|
};
|
||||||
|
@ -7866,6 +7866,20 @@ int bnx2x_init_hw_func_cnic(struct bnx2x *bp)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* previous driver DMAE transaction may have occurred when pre-boot stage ended
|
||||||
|
* and boot began, or when kdump kernel was loaded. Either case would invalidate
|
||||||
|
* the addresses of the transaction, resulting in was-error bit set in the pci
|
||||||
|
* causing all hw-to-host pcie transactions to timeout. If this happened we want
|
||||||
|
* to clear the interrupt which detected this from the pglueb and the was done
|
||||||
|
* bit
|
||||||
|
*/
|
||||||
|
static void bnx2x_clean_pglue_errors(struct bnx2x *bp)
|
||||||
|
{
|
||||||
|
if (!CHIP_IS_E1x(bp))
|
||||||
|
REG_WR(bp, PGLUE_B_REG_WAS_ERROR_PF_7_0_CLR,
|
||||||
|
1 << BP_ABS_FUNC(bp));
|
||||||
|
}
|
||||||
|
|
||||||
static int bnx2x_init_hw_func(struct bnx2x *bp)
|
static int bnx2x_init_hw_func(struct bnx2x *bp)
|
||||||
{
|
{
|
||||||
int port = BP_PORT(bp);
|
int port = BP_PORT(bp);
|
||||||
|
@ -7958,8 +7972,7 @@ static int bnx2x_init_hw_func(struct bnx2x *bp)
|
||||||
|
|
||||||
bnx2x_init_block(bp, BLOCK_PGLUE_B, init_phase);
|
bnx2x_init_block(bp, BLOCK_PGLUE_B, init_phase);
|
||||||
|
|
||||||
if (!CHIP_IS_E1x(bp))
|
bnx2x_clean_pglue_errors(bp);
|
||||||
REG_WR(bp, PGLUE_B_REG_WAS_ERROR_PF_7_0_CLR, func);
|
|
||||||
|
|
||||||
bnx2x_init_block(bp, BLOCK_ATC, init_phase);
|
bnx2x_init_block(bp, BLOCK_ATC, init_phase);
|
||||||
bnx2x_init_block(bp, BLOCK_DMAE, init_phase);
|
bnx2x_init_block(bp, BLOCK_DMAE, init_phase);
|
||||||
|
@ -10141,6 +10154,25 @@ static u32 bnx2x_get_pretend_reg(struct bnx2x *bp)
|
||||||
return base + (BP_ABS_FUNC(bp)) * stride;
|
return base + (BP_ABS_FUNC(bp)) * stride;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static bool bnx2x_prev_unload_close_umac(struct bnx2x *bp,
|
||||||
|
u8 port, u32 reset_reg,
|
||||||
|
struct bnx2x_mac_vals *vals)
|
||||||
|
{
|
||||||
|
u32 mask = MISC_REGISTERS_RESET_REG_2_UMAC0 << port;
|
||||||
|
u32 base_addr;
|
||||||
|
|
||||||
|
if (!(mask & reset_reg))
|
||||||
|
return false;
|
||||||
|
|
||||||
|
BNX2X_DEV_INFO("Disable umac Rx %02x\n", port);
|
||||||
|
base_addr = port ? GRCBASE_UMAC1 : GRCBASE_UMAC0;
|
||||||
|
vals->umac_addr[port] = base_addr + UMAC_REG_COMMAND_CONFIG;
|
||||||
|
vals->umac_val[port] = REG_RD(bp, vals->umac_addr[port]);
|
||||||
|
REG_WR(bp, vals->umac_addr[port], 0);
|
||||||
|
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
static void bnx2x_prev_unload_close_mac(struct bnx2x *bp,
|
static void bnx2x_prev_unload_close_mac(struct bnx2x *bp,
|
||||||
struct bnx2x_mac_vals *vals)
|
struct bnx2x_mac_vals *vals)
|
||||||
{
|
{
|
||||||
|
@ -10149,10 +10181,7 @@ static void bnx2x_prev_unload_close_mac(struct bnx2x *bp,
|
||||||
u8 port = BP_PORT(bp);
|
u8 port = BP_PORT(bp);
|
||||||
|
|
||||||
/* reset addresses as they also mark which values were changed */
|
/* reset addresses as they also mark which values were changed */
|
||||||
vals->bmac_addr = 0;
|
memset(vals, 0, sizeof(*vals));
|
||||||
vals->umac_addr = 0;
|
|
||||||
vals->xmac_addr = 0;
|
|
||||||
vals->emac_addr = 0;
|
|
||||||
|
|
||||||
reset_reg = REG_RD(bp, MISC_REG_RESET_REG_2);
|
reset_reg = REG_RD(bp, MISC_REG_RESET_REG_2);
|
||||||
|
|
||||||
|
@ -10201,15 +10230,11 @@ static void bnx2x_prev_unload_close_mac(struct bnx2x *bp,
|
||||||
REG_WR(bp, vals->xmac_addr, 0);
|
REG_WR(bp, vals->xmac_addr, 0);
|
||||||
mac_stopped = true;
|
mac_stopped = true;
|
||||||
}
|
}
|
||||||
mask = MISC_REGISTERS_RESET_REG_2_UMAC0 << port;
|
|
||||||
if (mask & reset_reg) {
|
mac_stopped |= bnx2x_prev_unload_close_umac(bp, 0,
|
||||||
BNX2X_DEV_INFO("Disable umac Rx\n");
|
reset_reg, vals);
|
||||||
base_addr = BP_PORT(bp) ? GRCBASE_UMAC1 : GRCBASE_UMAC0;
|
mac_stopped |= bnx2x_prev_unload_close_umac(bp, 1,
|
||||||
vals->umac_addr = base_addr + UMAC_REG_COMMAND_CONFIG;
|
reset_reg, vals);
|
||||||
vals->umac_val = REG_RD(bp, vals->umac_addr);
|
|
||||||
REG_WR(bp, vals->umac_addr, 0);
|
|
||||||
mac_stopped = true;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (mac_stopped)
|
if (mac_stopped)
|
||||||
|
@ -10505,8 +10530,11 @@ static int bnx2x_prev_unload_common(struct bnx2x *bp)
|
||||||
/* Close the MAC Rx to prevent BRB from filling up */
|
/* Close the MAC Rx to prevent BRB from filling up */
|
||||||
bnx2x_prev_unload_close_mac(bp, &mac_vals);
|
bnx2x_prev_unload_close_mac(bp, &mac_vals);
|
||||||
|
|
||||||
/* close LLH filters towards the BRB */
|
/* close LLH filters for both ports towards the BRB */
|
||||||
bnx2x_set_rx_filter(&bp->link_params, 0);
|
bnx2x_set_rx_filter(&bp->link_params, 0);
|
||||||
|
bp->link_params.port ^= 1;
|
||||||
|
bnx2x_set_rx_filter(&bp->link_params, 0);
|
||||||
|
bp->link_params.port ^= 1;
|
||||||
|
|
||||||
/* Check if the UNDI driver was previously loaded */
|
/* Check if the UNDI driver was previously loaded */
|
||||||
if (bnx2x_prev_is_after_undi(bp)) {
|
if (bnx2x_prev_is_after_undi(bp)) {
|
||||||
|
@ -10553,8 +10581,10 @@ static int bnx2x_prev_unload_common(struct bnx2x *bp)
|
||||||
|
|
||||||
if (mac_vals.xmac_addr)
|
if (mac_vals.xmac_addr)
|
||||||
REG_WR(bp, mac_vals.xmac_addr, mac_vals.xmac_val);
|
REG_WR(bp, mac_vals.xmac_addr, mac_vals.xmac_val);
|
||||||
if (mac_vals.umac_addr)
|
if (mac_vals.umac_addr[0])
|
||||||
REG_WR(bp, mac_vals.umac_addr, mac_vals.umac_val);
|
REG_WR(bp, mac_vals.umac_addr[0], mac_vals.umac_val[0]);
|
||||||
|
if (mac_vals.umac_addr[1])
|
||||||
|
REG_WR(bp, mac_vals.umac_addr[1], mac_vals.umac_val[1]);
|
||||||
if (mac_vals.emac_addr)
|
if (mac_vals.emac_addr)
|
||||||
REG_WR(bp, mac_vals.emac_addr, mac_vals.emac_val);
|
REG_WR(bp, mac_vals.emac_addr, mac_vals.emac_val);
|
||||||
if (mac_vals.bmac_addr) {
|
if (mac_vals.bmac_addr) {
|
||||||
|
@ -10571,26 +10601,6 @@ static int bnx2x_prev_unload_common(struct bnx2x *bp)
|
||||||
return bnx2x_prev_mcp_done(bp);
|
return bnx2x_prev_mcp_done(bp);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* previous driver DMAE transaction may have occurred when pre-boot stage ended
|
|
||||||
* and boot began, or when kdump kernel was loaded. Either case would invalidate
|
|
||||||
* the addresses of the transaction, resulting in was-error bit set in the pci
|
|
||||||
* causing all hw-to-host pcie transactions to timeout. If this happened we want
|
|
||||||
* to clear the interrupt which detected this from the pglueb and the was done
|
|
||||||
* bit
|
|
||||||
*/
|
|
||||||
static void bnx2x_prev_interrupted_dmae(struct bnx2x *bp)
|
|
||||||
{
|
|
||||||
if (!CHIP_IS_E1x(bp)) {
|
|
||||||
u32 val = REG_RD(bp, PGLUE_B_REG_PGLUE_B_INT_STS);
|
|
||||||
if (val & PGLUE_B_PGLUE_B_INT_STS_REG_WAS_ERROR_ATTN) {
|
|
||||||
DP(BNX2X_MSG_SP,
|
|
||||||
"'was error' bit was found to be set in pglueb upon startup. Clearing\n");
|
|
||||||
REG_WR(bp, PGLUE_B_REG_WAS_ERROR_PF_7_0_CLR,
|
|
||||||
1 << BP_FUNC(bp));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
static int bnx2x_prev_unload(struct bnx2x *bp)
|
static int bnx2x_prev_unload(struct bnx2x *bp)
|
||||||
{
|
{
|
||||||
int time_counter = 10;
|
int time_counter = 10;
|
||||||
|
@ -10600,7 +10610,7 @@ static int bnx2x_prev_unload(struct bnx2x *bp)
|
||||||
/* clear hw from errors which may have resulted from an interrupted
|
/* clear hw from errors which may have resulted from an interrupted
|
||||||
* dmae transaction.
|
* dmae transaction.
|
||||||
*/
|
*/
|
||||||
bnx2x_prev_interrupted_dmae(bp);
|
bnx2x_clean_pglue_errors(bp);
|
||||||
|
|
||||||
/* Release previously held locks */
|
/* Release previously held locks */
|
||||||
hw_lock_reg = (BP_FUNC(bp) <= 5) ?
|
hw_lock_reg = (BP_FUNC(bp) <= 5) ?
|
||||||
|
@ -12037,9 +12047,8 @@ static int bnx2x_init_bp(struct bnx2x *bp)
|
||||||
mutex_init(&bp->port.phy_mutex);
|
mutex_init(&bp->port.phy_mutex);
|
||||||
mutex_init(&bp->fw_mb_mutex);
|
mutex_init(&bp->fw_mb_mutex);
|
||||||
mutex_init(&bp->drv_info_mutex);
|
mutex_init(&bp->drv_info_mutex);
|
||||||
|
mutex_init(&bp->stats_lock);
|
||||||
bp->drv_info_mng_owner = false;
|
bp->drv_info_mng_owner = false;
|
||||||
spin_lock_init(&bp->stats_lock);
|
|
||||||
sema_init(&bp->stats_sema, 1);
|
|
||||||
|
|
||||||
INIT_DELAYED_WORK(&bp->sp_task, bnx2x_sp_task);
|
INIT_DELAYED_WORK(&bp->sp_task, bnx2x_sp_task);
|
||||||
INIT_DELAYED_WORK(&bp->sp_rtnl_task, bnx2x_sp_rtnl_task);
|
INIT_DELAYED_WORK(&bp->sp_rtnl_task, bnx2x_sp_rtnl_task);
|
||||||
|
@ -13668,9 +13677,9 @@ static int bnx2x_eeh_nic_unload(struct bnx2x *bp)
|
||||||
cancel_delayed_work_sync(&bp->sp_task);
|
cancel_delayed_work_sync(&bp->sp_task);
|
||||||
cancel_delayed_work_sync(&bp->period_task);
|
cancel_delayed_work_sync(&bp->period_task);
|
||||||
|
|
||||||
spin_lock_bh(&bp->stats_lock);
|
mutex_lock(&bp->stats_lock);
|
||||||
bp->stats_state = STATS_STATE_DISABLED;
|
bp->stats_state = STATS_STATE_DISABLED;
|
||||||
spin_unlock_bh(&bp->stats_lock);
|
mutex_unlock(&bp->stats_lock);
|
||||||
|
|
||||||
bnx2x_save_statistics(bp);
|
bnx2x_save_statistics(bp);
|
||||||
|
|
||||||
|
|
|
@ -2238,7 +2238,9 @@ int bnx2x_vf_close(struct bnx2x *bp, struct bnx2x_virtf *vf)
|
||||||
|
|
||||||
cookie.vf = vf;
|
cookie.vf = vf;
|
||||||
cookie.state = VF_ACQUIRED;
|
cookie.state = VF_ACQUIRED;
|
||||||
bnx2x_stats_safe_exec(bp, bnx2x_set_vf_state, &cookie);
|
rc = bnx2x_stats_safe_exec(bp, bnx2x_set_vf_state, &cookie);
|
||||||
|
if (rc)
|
||||||
|
goto op_err;
|
||||||
}
|
}
|
||||||
|
|
||||||
DP(BNX2X_MSG_IOV, "set state to acquired\n");
|
DP(BNX2X_MSG_IOV, "set state to acquired\n");
|
||||||
|
|
|
@ -123,36 +123,28 @@ static void bnx2x_dp_stats(struct bnx2x *bp)
|
||||||
*/
|
*/
|
||||||
static void bnx2x_storm_stats_post(struct bnx2x *bp)
|
static void bnx2x_storm_stats_post(struct bnx2x *bp)
|
||||||
{
|
{
|
||||||
if (!bp->stats_pending) {
|
int rc;
|
||||||
int rc;
|
|
||||||
|
|
||||||
spin_lock_bh(&bp->stats_lock);
|
if (bp->stats_pending)
|
||||||
|
return;
|
||||||
|
|
||||||
if (bp->stats_pending) {
|
bp->fw_stats_req->hdr.drv_stats_counter =
|
||||||
spin_unlock_bh(&bp->stats_lock);
|
cpu_to_le16(bp->stats_counter++);
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
bp->fw_stats_req->hdr.drv_stats_counter =
|
DP(BNX2X_MSG_STATS, "Sending statistics ramrod %d\n",
|
||||||
cpu_to_le16(bp->stats_counter++);
|
le16_to_cpu(bp->fw_stats_req->hdr.drv_stats_counter));
|
||||||
|
|
||||||
DP(BNX2X_MSG_STATS, "Sending statistics ramrod %d\n",
|
/* adjust the ramrod to include VF queues statistics */
|
||||||
le16_to_cpu(bp->fw_stats_req->hdr.drv_stats_counter));
|
bnx2x_iov_adjust_stats_req(bp);
|
||||||
|
bnx2x_dp_stats(bp);
|
||||||
|
|
||||||
/* adjust the ramrod to include VF queues statistics */
|
/* send FW stats ramrod */
|
||||||
bnx2x_iov_adjust_stats_req(bp);
|
rc = bnx2x_sp_post(bp, RAMROD_CMD_ID_COMMON_STAT_QUERY, 0,
|
||||||
bnx2x_dp_stats(bp);
|
U64_HI(bp->fw_stats_req_mapping),
|
||||||
|
U64_LO(bp->fw_stats_req_mapping),
|
||||||
/* send FW stats ramrod */
|
NONE_CONNECTION_TYPE);
|
||||||
rc = bnx2x_sp_post(bp, RAMROD_CMD_ID_COMMON_STAT_QUERY, 0,
|
if (rc == 0)
|
||||||
U64_HI(bp->fw_stats_req_mapping),
|
bp->stats_pending = 1;
|
||||||
U64_LO(bp->fw_stats_req_mapping),
|
|
||||||
NONE_CONNECTION_TYPE);
|
|
||||||
if (rc == 0)
|
|
||||||
bp->stats_pending = 1;
|
|
||||||
|
|
||||||
spin_unlock_bh(&bp->stats_lock);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void bnx2x_hw_stats_post(struct bnx2x *bp)
|
static void bnx2x_hw_stats_post(struct bnx2x *bp)
|
||||||
|
@ -221,7 +213,7 @@ static void bnx2x_stats_comp(struct bnx2x *bp)
|
||||||
*/
|
*/
|
||||||
|
|
||||||
/* should be called under stats_sema */
|
/* should be called under stats_sema */
|
||||||
static void __bnx2x_stats_pmf_update(struct bnx2x *bp)
|
static void bnx2x_stats_pmf_update(struct bnx2x *bp)
|
||||||
{
|
{
|
||||||
struct dmae_command *dmae;
|
struct dmae_command *dmae;
|
||||||
u32 opcode;
|
u32 opcode;
|
||||||
|
@ -519,7 +511,7 @@ static void bnx2x_func_stats_init(struct bnx2x *bp)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* should be called under stats_sema */
|
/* should be called under stats_sema */
|
||||||
static void __bnx2x_stats_start(struct bnx2x *bp)
|
static void bnx2x_stats_start(struct bnx2x *bp)
|
||||||
{
|
{
|
||||||
if (IS_PF(bp)) {
|
if (IS_PF(bp)) {
|
||||||
if (bp->port.pmf)
|
if (bp->port.pmf)
|
||||||
|
@ -531,34 +523,13 @@ static void __bnx2x_stats_start(struct bnx2x *bp)
|
||||||
bnx2x_hw_stats_post(bp);
|
bnx2x_hw_stats_post(bp);
|
||||||
bnx2x_storm_stats_post(bp);
|
bnx2x_storm_stats_post(bp);
|
||||||
}
|
}
|
||||||
|
|
||||||
bp->stats_started = true;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void bnx2x_stats_start(struct bnx2x *bp)
|
|
||||||
{
|
|
||||||
if (down_timeout(&bp->stats_sema, HZ/10))
|
|
||||||
BNX2X_ERR("Unable to acquire stats lock\n");
|
|
||||||
__bnx2x_stats_start(bp);
|
|
||||||
up(&bp->stats_sema);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void bnx2x_stats_pmf_start(struct bnx2x *bp)
|
static void bnx2x_stats_pmf_start(struct bnx2x *bp)
|
||||||
{
|
{
|
||||||
if (down_timeout(&bp->stats_sema, HZ/10))
|
|
||||||
BNX2X_ERR("Unable to acquire stats lock\n");
|
|
||||||
bnx2x_stats_comp(bp);
|
bnx2x_stats_comp(bp);
|
||||||
__bnx2x_stats_pmf_update(bp);
|
bnx2x_stats_pmf_update(bp);
|
||||||
__bnx2x_stats_start(bp);
|
bnx2x_stats_start(bp);
|
||||||
up(&bp->stats_sema);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void bnx2x_stats_pmf_update(struct bnx2x *bp)
|
|
||||||
{
|
|
||||||
if (down_timeout(&bp->stats_sema, HZ/10))
|
|
||||||
BNX2X_ERR("Unable to acquire stats lock\n");
|
|
||||||
__bnx2x_stats_pmf_update(bp);
|
|
||||||
up(&bp->stats_sema);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void bnx2x_stats_restart(struct bnx2x *bp)
|
static void bnx2x_stats_restart(struct bnx2x *bp)
|
||||||
|
@ -568,11 +539,9 @@ static void bnx2x_stats_restart(struct bnx2x *bp)
|
||||||
*/
|
*/
|
||||||
if (IS_VF(bp))
|
if (IS_VF(bp))
|
||||||
return;
|
return;
|
||||||
if (down_timeout(&bp->stats_sema, HZ/10))
|
|
||||||
BNX2X_ERR("Unable to acquire stats lock\n");
|
|
||||||
bnx2x_stats_comp(bp);
|
bnx2x_stats_comp(bp);
|
||||||
__bnx2x_stats_start(bp);
|
bnx2x_stats_start(bp);
|
||||||
up(&bp->stats_sema);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void bnx2x_bmac_stats_update(struct bnx2x *bp)
|
static void bnx2x_bmac_stats_update(struct bnx2x *bp)
|
||||||
|
@ -1246,18 +1215,12 @@ static void bnx2x_stats_update(struct bnx2x *bp)
|
||||||
{
|
{
|
||||||
u32 *stats_comp = bnx2x_sp(bp, stats_comp);
|
u32 *stats_comp = bnx2x_sp(bp, stats_comp);
|
||||||
|
|
||||||
/* we run update from timer context, so give up
|
if (bnx2x_edebug_stats_stopped(bp))
|
||||||
* if somebody is in the middle of transition
|
|
||||||
*/
|
|
||||||
if (down_trylock(&bp->stats_sema))
|
|
||||||
return;
|
return;
|
||||||
|
|
||||||
if (bnx2x_edebug_stats_stopped(bp) || !bp->stats_started)
|
|
||||||
goto out;
|
|
||||||
|
|
||||||
if (IS_PF(bp)) {
|
if (IS_PF(bp)) {
|
||||||
if (*stats_comp != DMAE_COMP_VAL)
|
if (*stats_comp != DMAE_COMP_VAL)
|
||||||
goto out;
|
return;
|
||||||
|
|
||||||
if (bp->port.pmf)
|
if (bp->port.pmf)
|
||||||
bnx2x_hw_stats_update(bp);
|
bnx2x_hw_stats_update(bp);
|
||||||
|
@ -1267,7 +1230,7 @@ static void bnx2x_stats_update(struct bnx2x *bp)
|
||||||
BNX2X_ERR("storm stats were not updated for 3 times\n");
|
BNX2X_ERR("storm stats were not updated for 3 times\n");
|
||||||
bnx2x_panic();
|
bnx2x_panic();
|
||||||
}
|
}
|
||||||
goto out;
|
return;
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
/* vf doesn't collect HW statistics, and doesn't get completions
|
/* vf doesn't collect HW statistics, and doesn't get completions
|
||||||
|
@ -1281,7 +1244,7 @@ static void bnx2x_stats_update(struct bnx2x *bp)
|
||||||
|
|
||||||
/* vf is done */
|
/* vf is done */
|
||||||
if (IS_VF(bp))
|
if (IS_VF(bp))
|
||||||
goto out;
|
return;
|
||||||
|
|
||||||
if (netif_msg_timer(bp)) {
|
if (netif_msg_timer(bp)) {
|
||||||
struct bnx2x_eth_stats *estats = &bp->eth_stats;
|
struct bnx2x_eth_stats *estats = &bp->eth_stats;
|
||||||
|
@ -1292,9 +1255,6 @@ static void bnx2x_stats_update(struct bnx2x *bp)
|
||||||
|
|
||||||
bnx2x_hw_stats_post(bp);
|
bnx2x_hw_stats_post(bp);
|
||||||
bnx2x_storm_stats_post(bp);
|
bnx2x_storm_stats_post(bp);
|
||||||
|
|
||||||
out:
|
|
||||||
up(&bp->stats_sema);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void bnx2x_port_stats_stop(struct bnx2x *bp)
|
static void bnx2x_port_stats_stop(struct bnx2x *bp)
|
||||||
|
@ -1358,12 +1318,7 @@ static void bnx2x_port_stats_stop(struct bnx2x *bp)
|
||||||
|
|
||||||
static void bnx2x_stats_stop(struct bnx2x *bp)
|
static void bnx2x_stats_stop(struct bnx2x *bp)
|
||||||
{
|
{
|
||||||
int update = 0;
|
bool update = false;
|
||||||
|
|
||||||
if (down_timeout(&bp->stats_sema, HZ/10))
|
|
||||||
BNX2X_ERR("Unable to acquire stats lock\n");
|
|
||||||
|
|
||||||
bp->stats_started = false;
|
|
||||||
|
|
||||||
bnx2x_stats_comp(bp);
|
bnx2x_stats_comp(bp);
|
||||||
|
|
||||||
|
@ -1381,8 +1336,6 @@ static void bnx2x_stats_stop(struct bnx2x *bp)
|
||||||
bnx2x_hw_stats_post(bp);
|
bnx2x_hw_stats_post(bp);
|
||||||
bnx2x_stats_comp(bp);
|
bnx2x_stats_comp(bp);
|
||||||
}
|
}
|
||||||
|
|
||||||
up(&bp->stats_sema);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void bnx2x_stats_do_nothing(struct bnx2x *bp)
|
static void bnx2x_stats_do_nothing(struct bnx2x *bp)
|
||||||
|
@ -1410,18 +1363,28 @@ static const struct {
|
||||||
|
|
||||||
void bnx2x_stats_handle(struct bnx2x *bp, enum bnx2x_stats_event event)
|
void bnx2x_stats_handle(struct bnx2x *bp, enum bnx2x_stats_event event)
|
||||||
{
|
{
|
||||||
enum bnx2x_stats_state state;
|
enum bnx2x_stats_state state = bp->stats_state;
|
||||||
void (*action)(struct bnx2x *bp);
|
|
||||||
if (unlikely(bp->panic))
|
if (unlikely(bp->panic))
|
||||||
return;
|
return;
|
||||||
|
|
||||||
spin_lock_bh(&bp->stats_lock);
|
/* Statistics update run from timer context, and we don't want to stop
|
||||||
state = bp->stats_state;
|
* that context in case someone is in the middle of a transition.
|
||||||
bp->stats_state = bnx2x_stats_stm[state][event].next_state;
|
* For other events, wait a bit until lock is taken.
|
||||||
action = bnx2x_stats_stm[state][event].action;
|
*/
|
||||||
spin_unlock_bh(&bp->stats_lock);
|
if (!mutex_trylock(&bp->stats_lock)) {
|
||||||
|
if (event == STATS_EVENT_UPDATE)
|
||||||
|
return;
|
||||||
|
|
||||||
action(bp);
|
DP(BNX2X_MSG_STATS,
|
||||||
|
"Unlikely stats' lock contention [event %d]\n", event);
|
||||||
|
mutex_lock(&bp->stats_lock);
|
||||||
|
}
|
||||||
|
|
||||||
|
bnx2x_stats_stm[state][event].action(bp);
|
||||||
|
bp->stats_state = bnx2x_stats_stm[state][event].next_state;
|
||||||
|
|
||||||
|
mutex_unlock(&bp->stats_lock);
|
||||||
|
|
||||||
if ((event != STATS_EVENT_UPDATE) || netif_msg_timer(bp))
|
if ((event != STATS_EVENT_UPDATE) || netif_msg_timer(bp))
|
||||||
DP(BNX2X_MSG_STATS, "state %d -> event %d -> state %d\n",
|
DP(BNX2X_MSG_STATS, "state %d -> event %d -> state %d\n",
|
||||||
|
@ -1998,13 +1961,34 @@ void bnx2x_afex_collect_stats(struct bnx2x *bp, void *void_afex_stats,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void bnx2x_stats_safe_exec(struct bnx2x *bp,
|
int bnx2x_stats_safe_exec(struct bnx2x *bp,
|
||||||
void (func_to_exec)(void *cookie),
|
void (func_to_exec)(void *cookie),
|
||||||
void *cookie){
|
void *cookie)
|
||||||
if (down_timeout(&bp->stats_sema, HZ/10))
|
{
|
||||||
BNX2X_ERR("Unable to acquire stats lock\n");
|
int cnt = 10, rc = 0;
|
||||||
|
|
||||||
|
/* Wait for statistics to end [while blocking further requests],
|
||||||
|
* then run supplied function 'safely'.
|
||||||
|
*/
|
||||||
|
mutex_lock(&bp->stats_lock);
|
||||||
|
|
||||||
bnx2x_stats_comp(bp);
|
bnx2x_stats_comp(bp);
|
||||||
|
while (bp->stats_pending && cnt--)
|
||||||
|
if (bnx2x_storm_stats_update(bp))
|
||||||
|
usleep_range(1000, 2000);
|
||||||
|
if (bp->stats_pending) {
|
||||||
|
BNX2X_ERR("Failed to wait for stats pending to clear [possibly FW is stuck]\n");
|
||||||
|
rc = -EBUSY;
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
func_to_exec(cookie);
|
func_to_exec(cookie);
|
||||||
__bnx2x_stats_start(bp);
|
|
||||||
up(&bp->stats_sema);
|
out:
|
||||||
|
/* No need to restart statistics - if they're enabled, the timer
|
||||||
|
* will restart the statistics.
|
||||||
|
*/
|
||||||
|
mutex_unlock(&bp->stats_lock);
|
||||||
|
|
||||||
|
return rc;
|
||||||
}
|
}
|
||||||
|
|
|
@ -539,9 +539,9 @@ struct bnx2x;
|
||||||
void bnx2x_memset_stats(struct bnx2x *bp);
|
void bnx2x_memset_stats(struct bnx2x *bp);
|
||||||
void bnx2x_stats_init(struct bnx2x *bp);
|
void bnx2x_stats_init(struct bnx2x *bp);
|
||||||
void bnx2x_stats_handle(struct bnx2x *bp, enum bnx2x_stats_event event);
|
void bnx2x_stats_handle(struct bnx2x *bp, enum bnx2x_stats_event event);
|
||||||
void bnx2x_stats_safe_exec(struct bnx2x *bp,
|
int bnx2x_stats_safe_exec(struct bnx2x *bp,
|
||||||
void (func_to_exec)(void *cookie),
|
void (func_to_exec)(void *cookie),
|
||||||
void *cookie);
|
void *cookie);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* bnx2x_save_statistics - save statistics when unloading.
|
* bnx2x_save_statistics - save statistics when unloading.
|
||||||
|
|
|
@ -376,8 +376,6 @@ enum {
|
||||||
enum {
|
enum {
|
||||||
INGQ_EXTRAS = 2, /* firmware event queue and */
|
INGQ_EXTRAS = 2, /* firmware event queue and */
|
||||||
/* forwarded interrupts */
|
/* forwarded interrupts */
|
||||||
MAX_EGRQ = MAX_ETH_QSETS*2 + MAX_OFLD_QSETS*2
|
|
||||||
+ MAX_CTRL_QUEUES + MAX_RDMA_QUEUES + MAX_ISCSI_QUEUES,
|
|
||||||
MAX_INGQ = MAX_ETH_QSETS + MAX_OFLD_QSETS + MAX_RDMA_QUEUES
|
MAX_INGQ = MAX_ETH_QSETS + MAX_OFLD_QSETS + MAX_RDMA_QUEUES
|
||||||
+ MAX_RDMA_CIQS + MAX_ISCSI_QUEUES + INGQ_EXTRAS,
|
+ MAX_RDMA_CIQS + MAX_ISCSI_QUEUES + INGQ_EXTRAS,
|
||||||
};
|
};
|
||||||
|
@ -616,11 +614,13 @@ struct sge {
|
||||||
unsigned int idma_qid[2]; /* SGE IDMA Hung Ingress Queue ID */
|
unsigned int idma_qid[2]; /* SGE IDMA Hung Ingress Queue ID */
|
||||||
|
|
||||||
unsigned int egr_start;
|
unsigned int egr_start;
|
||||||
|
unsigned int egr_sz;
|
||||||
unsigned int ingr_start;
|
unsigned int ingr_start;
|
||||||
void *egr_map[MAX_EGRQ]; /* qid->queue egress queue map */
|
unsigned int ingr_sz;
|
||||||
struct sge_rspq *ingr_map[MAX_INGQ]; /* qid->queue ingress queue map */
|
void **egr_map; /* qid->queue egress queue map */
|
||||||
DECLARE_BITMAP(starving_fl, MAX_EGRQ);
|
struct sge_rspq **ingr_map; /* qid->queue ingress queue map */
|
||||||
DECLARE_BITMAP(txq_maperr, MAX_EGRQ);
|
unsigned long *starving_fl;
|
||||||
|
unsigned long *txq_maperr;
|
||||||
struct timer_list rx_timer; /* refills starving FLs */
|
struct timer_list rx_timer; /* refills starving FLs */
|
||||||
struct timer_list tx_timer; /* checks Tx queues */
|
struct timer_list tx_timer; /* checks Tx queues */
|
||||||
};
|
};
|
||||||
|
@ -1136,6 +1136,8 @@ int cxgb4_t4_bar2_sge_qregs(struct adapter *adapter,
|
||||||
|
|
||||||
unsigned int qtimer_val(const struct adapter *adap,
|
unsigned int qtimer_val(const struct adapter *adap,
|
||||||
const struct sge_rspq *q);
|
const struct sge_rspq *q);
|
||||||
|
|
||||||
|
int t4_init_devlog_params(struct adapter *adapter);
|
||||||
int t4_init_sge_params(struct adapter *adapter);
|
int t4_init_sge_params(struct adapter *adapter);
|
||||||
int t4_init_tp_params(struct adapter *adap);
|
int t4_init_tp_params(struct adapter *adap);
|
||||||
int t4_filter_field_shift(const struct adapter *adap, int filter_sel);
|
int t4_filter_field_shift(const struct adapter *adap, int filter_sel);
|
||||||
|
|
|
@ -670,9 +670,13 @@ static int cctrl_tbl_show(struct seq_file *seq, void *v)
|
||||||
"0.9375" };
|
"0.9375" };
|
||||||
|
|
||||||
int i;
|
int i;
|
||||||
u16 incr[NMTUS][NCCTRL_WIN];
|
u16 (*incr)[NCCTRL_WIN];
|
||||||
struct adapter *adap = seq->private;
|
struct adapter *adap = seq->private;
|
||||||
|
|
||||||
|
incr = kmalloc(sizeof(*incr) * NMTUS, GFP_KERNEL);
|
||||||
|
if (!incr)
|
||||||
|
return -ENOMEM;
|
||||||
|
|
||||||
t4_read_cong_tbl(adap, incr);
|
t4_read_cong_tbl(adap, incr);
|
||||||
|
|
||||||
for (i = 0; i < NCCTRL_WIN; ++i) {
|
for (i = 0; i < NCCTRL_WIN; ++i) {
|
||||||
|
@ -685,6 +689,8 @@ static int cctrl_tbl_show(struct seq_file *seq, void *v)
|
||||||
adap->params.a_wnd[i],
|
adap->params.a_wnd[i],
|
||||||
dec_fac[adap->params.b_wnd[i]]);
|
dec_fac[adap->params.b_wnd[i]]);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
kfree(incr);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -920,7 +920,7 @@ static void quiesce_rx(struct adapter *adap)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
for (i = 0; i < ARRAY_SIZE(adap->sge.ingr_map); i++) {
|
for (i = 0; i < adap->sge.ingr_sz; i++) {
|
||||||
struct sge_rspq *q = adap->sge.ingr_map[i];
|
struct sge_rspq *q = adap->sge.ingr_map[i];
|
||||||
|
|
||||||
if (q && q->handler) {
|
if (q && q->handler) {
|
||||||
|
@ -934,6 +934,21 @@ static void quiesce_rx(struct adapter *adap)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* Disable interrupt and napi handler */
|
||||||
|
static void disable_interrupts(struct adapter *adap)
|
||||||
|
{
|
||||||
|
if (adap->flags & FULL_INIT_DONE) {
|
||||||
|
t4_intr_disable(adap);
|
||||||
|
if (adap->flags & USING_MSIX) {
|
||||||
|
free_msix_queue_irqs(adap);
|
||||||
|
free_irq(adap->msix_info[0].vec, adap);
|
||||||
|
} else {
|
||||||
|
free_irq(adap->pdev->irq, adap);
|
||||||
|
}
|
||||||
|
quiesce_rx(adap);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Enable NAPI scheduling and interrupt generation for all Rx queues.
|
* Enable NAPI scheduling and interrupt generation for all Rx queues.
|
||||||
*/
|
*/
|
||||||
|
@ -941,7 +956,7 @@ static void enable_rx(struct adapter *adap)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
for (i = 0; i < ARRAY_SIZE(adap->sge.ingr_map); i++) {
|
for (i = 0; i < adap->sge.ingr_sz; i++) {
|
||||||
struct sge_rspq *q = adap->sge.ingr_map[i];
|
struct sge_rspq *q = adap->sge.ingr_map[i];
|
||||||
|
|
||||||
if (!q)
|
if (!q)
|
||||||
|
@ -970,8 +985,8 @@ static int setup_sge_queues(struct adapter *adap)
|
||||||
int err, msi_idx, i, j;
|
int err, msi_idx, i, j;
|
||||||
struct sge *s = &adap->sge;
|
struct sge *s = &adap->sge;
|
||||||
|
|
||||||
bitmap_zero(s->starving_fl, MAX_EGRQ);
|
bitmap_zero(s->starving_fl, s->egr_sz);
|
||||||
bitmap_zero(s->txq_maperr, MAX_EGRQ);
|
bitmap_zero(s->txq_maperr, s->egr_sz);
|
||||||
|
|
||||||
if (adap->flags & USING_MSIX)
|
if (adap->flags & USING_MSIX)
|
||||||
msi_idx = 1; /* vector 0 is for non-queue interrupts */
|
msi_idx = 1; /* vector 0 is for non-queue interrupts */
|
||||||
|
@ -983,6 +998,19 @@ static int setup_sge_queues(struct adapter *adap)
|
||||||
msi_idx = -((int)s->intrq.abs_id + 1);
|
msi_idx = -((int)s->intrq.abs_id + 1);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* NOTE: If you add/delete any Ingress/Egress Queue allocations in here,
|
||||||
|
* don't forget to update the following which need to be
|
||||||
|
* synchronized to and changes here.
|
||||||
|
*
|
||||||
|
* 1. The calculations of MAX_INGQ in cxgb4.h.
|
||||||
|
*
|
||||||
|
* 2. Update enable_msix/name_msix_vecs/request_msix_queue_irqs
|
||||||
|
* to accommodate any new/deleted Ingress Queues
|
||||||
|
* which need MSI-X Vectors.
|
||||||
|
*
|
||||||
|
* 3. Update sge_qinfo_show() to include information on the
|
||||||
|
* new/deleted queues.
|
||||||
|
*/
|
||||||
err = t4_sge_alloc_rxq(adap, &s->fw_evtq, true, adap->port[0],
|
err = t4_sge_alloc_rxq(adap, &s->fw_evtq, true, adap->port[0],
|
||||||
msi_idx, NULL, fwevtq_handler);
|
msi_idx, NULL, fwevtq_handler);
|
||||||
if (err) {
|
if (err) {
|
||||||
|
@ -4244,19 +4272,12 @@ static int cxgb_up(struct adapter *adap)
|
||||||
|
|
||||||
static void cxgb_down(struct adapter *adapter)
|
static void cxgb_down(struct adapter *adapter)
|
||||||
{
|
{
|
||||||
t4_intr_disable(adapter);
|
|
||||||
cancel_work_sync(&adapter->tid_release_task);
|
cancel_work_sync(&adapter->tid_release_task);
|
||||||
cancel_work_sync(&adapter->db_full_task);
|
cancel_work_sync(&adapter->db_full_task);
|
||||||
cancel_work_sync(&adapter->db_drop_task);
|
cancel_work_sync(&adapter->db_drop_task);
|
||||||
adapter->tid_release_task_busy = false;
|
adapter->tid_release_task_busy = false;
|
||||||
adapter->tid_release_head = NULL;
|
adapter->tid_release_head = NULL;
|
||||||
|
|
||||||
if (adapter->flags & USING_MSIX) {
|
|
||||||
free_msix_queue_irqs(adapter);
|
|
||||||
free_irq(adapter->msix_info[0].vec, adapter);
|
|
||||||
} else
|
|
||||||
free_irq(adapter->pdev->irq, adapter);
|
|
||||||
quiesce_rx(adapter);
|
|
||||||
t4_sge_stop(adapter);
|
t4_sge_stop(adapter);
|
||||||
t4_free_sge_resources(adapter);
|
t4_free_sge_resources(adapter);
|
||||||
adapter->flags &= ~FULL_INIT_DONE;
|
adapter->flags &= ~FULL_INIT_DONE;
|
||||||
|
@ -4733,8 +4754,9 @@ static int adap_init1(struct adapter *adap, struct fw_caps_config_cmd *c)
|
||||||
if (ret < 0)
|
if (ret < 0)
|
||||||
return ret;
|
return ret;
|
||||||
|
|
||||||
ret = t4_cfg_pfvf(adap, adap->fn, adap->fn, 0, MAX_EGRQ, 64, MAX_INGQ,
|
ret = t4_cfg_pfvf(adap, adap->fn, adap->fn, 0, adap->sge.egr_sz, 64,
|
||||||
0, 0, 4, 0xf, 0xf, 16, FW_CMD_CAP_PF, FW_CMD_CAP_PF);
|
MAX_INGQ, 0, 0, 4, 0xf, 0xf, 16, FW_CMD_CAP_PF,
|
||||||
|
FW_CMD_CAP_PF);
|
||||||
if (ret < 0)
|
if (ret < 0)
|
||||||
return ret;
|
return ret;
|
||||||
|
|
||||||
|
@ -5088,10 +5110,15 @@ static int adap_init0(struct adapter *adap)
|
||||||
enum dev_state state;
|
enum dev_state state;
|
||||||
u32 params[7], val[7];
|
u32 params[7], val[7];
|
||||||
struct fw_caps_config_cmd caps_cmd;
|
struct fw_caps_config_cmd caps_cmd;
|
||||||
struct fw_devlog_cmd devlog_cmd;
|
|
||||||
u32 devlog_meminfo;
|
|
||||||
int reset = 1;
|
int reset = 1;
|
||||||
|
|
||||||
|
/* Grab Firmware Device Log parameters as early as possible so we have
|
||||||
|
* access to it for debugging, etc.
|
||||||
|
*/
|
||||||
|
ret = t4_init_devlog_params(adap);
|
||||||
|
if (ret < 0)
|
||||||
|
return ret;
|
||||||
|
|
||||||
/* Contact FW, advertising Master capability */
|
/* Contact FW, advertising Master capability */
|
||||||
ret = t4_fw_hello(adap, adap->mbox, adap->mbox, MASTER_MAY, &state);
|
ret = t4_fw_hello(adap, adap->mbox, adap->mbox, MASTER_MAY, &state);
|
||||||
if (ret < 0) {
|
if (ret < 0) {
|
||||||
|
@ -5169,30 +5196,6 @@ static int adap_init0(struct adapter *adap)
|
||||||
if (ret < 0)
|
if (ret < 0)
|
||||||
goto bye;
|
goto bye;
|
||||||
|
|
||||||
/* Read firmware device log parameters. We really need to find a way
|
|
||||||
* to get these parameters initialized with some default values (which
|
|
||||||
* are likely to be correct) for the case where we either don't
|
|
||||||
* attache to the firmware or it's crashed when we probe the adapter.
|
|
||||||
* That way we'll still be able to perform early firmware startup
|
|
||||||
* debugging ... If the request to get the Firmware's Device Log
|
|
||||||
* parameters fails, we'll live so we don't make that a fatal error.
|
|
||||||
*/
|
|
||||||
memset(&devlog_cmd, 0, sizeof(devlog_cmd));
|
|
||||||
devlog_cmd.op_to_write = htonl(FW_CMD_OP_V(FW_DEVLOG_CMD) |
|
|
||||||
FW_CMD_REQUEST_F | FW_CMD_READ_F);
|
|
||||||
devlog_cmd.retval_len16 = htonl(FW_LEN16(devlog_cmd));
|
|
||||||
ret = t4_wr_mbox(adap, adap->mbox, &devlog_cmd, sizeof(devlog_cmd),
|
|
||||||
&devlog_cmd);
|
|
||||||
if (ret == 0) {
|
|
||||||
devlog_meminfo =
|
|
||||||
ntohl(devlog_cmd.memtype_devlog_memaddr16_devlog);
|
|
||||||
adap->params.devlog.memtype =
|
|
||||||
FW_DEVLOG_CMD_MEMTYPE_DEVLOG_G(devlog_meminfo);
|
|
||||||
adap->params.devlog.start =
|
|
||||||
FW_DEVLOG_CMD_MEMADDR16_DEVLOG_G(devlog_meminfo) << 4;
|
|
||||||
adap->params.devlog.size = ntohl(devlog_cmd.memsize_devlog);
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Find out what ports are available to us. Note that we need to do
|
* Find out what ports are available to us. Note that we need to do
|
||||||
* this before calling adap_init0_no_config() since it needs nports
|
* this before calling adap_init0_no_config() since it needs nports
|
||||||
|
@ -5293,6 +5296,51 @@ static int adap_init0(struct adapter *adap)
|
||||||
adap->tids.nftids = val[4] - val[3] + 1;
|
adap->tids.nftids = val[4] - val[3] + 1;
|
||||||
adap->sge.ingr_start = val[5];
|
adap->sge.ingr_start = val[5];
|
||||||
|
|
||||||
|
/* qids (ingress/egress) returned from firmware can be anywhere
|
||||||
|
* in the range from EQ(IQFLINT)_START to EQ(IQFLINT)_END.
|
||||||
|
* Hence driver needs to allocate memory for this range to
|
||||||
|
* store the queue info. Get the highest IQFLINT/EQ index returned
|
||||||
|
* in FW_EQ_*_CMD.alloc command.
|
||||||
|
*/
|
||||||
|
params[0] = FW_PARAM_PFVF(EQ_END);
|
||||||
|
params[1] = FW_PARAM_PFVF(IQFLINT_END);
|
||||||
|
ret = t4_query_params(adap, adap->mbox, adap->fn, 0, 2, params, val);
|
||||||
|
if (ret < 0)
|
||||||
|
goto bye;
|
||||||
|
adap->sge.egr_sz = val[0] - adap->sge.egr_start + 1;
|
||||||
|
adap->sge.ingr_sz = val[1] - adap->sge.ingr_start + 1;
|
||||||
|
|
||||||
|
adap->sge.egr_map = kcalloc(adap->sge.egr_sz,
|
||||||
|
sizeof(*adap->sge.egr_map), GFP_KERNEL);
|
||||||
|
if (!adap->sge.egr_map) {
|
||||||
|
ret = -ENOMEM;
|
||||||
|
goto bye;
|
||||||
|
}
|
||||||
|
|
||||||
|
adap->sge.ingr_map = kcalloc(adap->sge.ingr_sz,
|
||||||
|
sizeof(*adap->sge.ingr_map), GFP_KERNEL);
|
||||||
|
if (!adap->sge.ingr_map) {
|
||||||
|
ret = -ENOMEM;
|
||||||
|
goto bye;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Allocate the memory for the vaious egress queue bitmaps
|
||||||
|
* ie starving_fl and txq_maperr.
|
||||||
|
*/
|
||||||
|
adap->sge.starving_fl = kcalloc(BITS_TO_LONGS(adap->sge.egr_sz),
|
||||||
|
sizeof(long), GFP_KERNEL);
|
||||||
|
if (!adap->sge.starving_fl) {
|
||||||
|
ret = -ENOMEM;
|
||||||
|
goto bye;
|
||||||
|
}
|
||||||
|
|
||||||
|
adap->sge.txq_maperr = kcalloc(BITS_TO_LONGS(adap->sge.egr_sz),
|
||||||
|
sizeof(long), GFP_KERNEL);
|
||||||
|
if (!adap->sge.txq_maperr) {
|
||||||
|
ret = -ENOMEM;
|
||||||
|
goto bye;
|
||||||
|
}
|
||||||
|
|
||||||
params[0] = FW_PARAM_PFVF(CLIP_START);
|
params[0] = FW_PARAM_PFVF(CLIP_START);
|
||||||
params[1] = FW_PARAM_PFVF(CLIP_END);
|
params[1] = FW_PARAM_PFVF(CLIP_END);
|
||||||
ret = t4_query_params(adap, adap->mbox, adap->fn, 0, 2, params, val);
|
ret = t4_query_params(adap, adap->mbox, adap->fn, 0, 2, params, val);
|
||||||
|
@ -5501,6 +5549,10 @@ static int adap_init0(struct adapter *adap)
|
||||||
* happened to HW/FW, stop issuing commands.
|
* happened to HW/FW, stop issuing commands.
|
||||||
*/
|
*/
|
||||||
bye:
|
bye:
|
||||||
|
kfree(adap->sge.egr_map);
|
||||||
|
kfree(adap->sge.ingr_map);
|
||||||
|
kfree(adap->sge.starving_fl);
|
||||||
|
kfree(adap->sge.txq_maperr);
|
||||||
if (ret != -ETIMEDOUT && ret != -EIO)
|
if (ret != -ETIMEDOUT && ret != -EIO)
|
||||||
t4_fw_bye(adap, adap->mbox);
|
t4_fw_bye(adap, adap->mbox);
|
||||||
return ret;
|
return ret;
|
||||||
|
@ -5528,6 +5580,7 @@ static pci_ers_result_t eeh_err_detected(struct pci_dev *pdev,
|
||||||
netif_carrier_off(dev);
|
netif_carrier_off(dev);
|
||||||
}
|
}
|
||||||
spin_unlock(&adap->stats_lock);
|
spin_unlock(&adap->stats_lock);
|
||||||
|
disable_interrupts(adap);
|
||||||
if (adap->flags & FULL_INIT_DONE)
|
if (adap->flags & FULL_INIT_DONE)
|
||||||
cxgb_down(adap);
|
cxgb_down(adap);
|
||||||
rtnl_unlock();
|
rtnl_unlock();
|
||||||
|
@ -5912,6 +5965,10 @@ static void free_some_resources(struct adapter *adapter)
|
||||||
|
|
||||||
t4_free_mem(adapter->l2t);
|
t4_free_mem(adapter->l2t);
|
||||||
t4_free_mem(adapter->tids.tid_tab);
|
t4_free_mem(adapter->tids.tid_tab);
|
||||||
|
kfree(adapter->sge.egr_map);
|
||||||
|
kfree(adapter->sge.ingr_map);
|
||||||
|
kfree(adapter->sge.starving_fl);
|
||||||
|
kfree(adapter->sge.txq_maperr);
|
||||||
disable_msi(adapter);
|
disable_msi(adapter);
|
||||||
|
|
||||||
for_each_port(adapter, i)
|
for_each_port(adapter, i)
|
||||||
|
@ -6237,6 +6294,8 @@ static void remove_one(struct pci_dev *pdev)
|
||||||
if (is_offload(adapter))
|
if (is_offload(adapter))
|
||||||
detach_ulds(adapter);
|
detach_ulds(adapter);
|
||||||
|
|
||||||
|
disable_interrupts(adapter);
|
||||||
|
|
||||||
for_each_port(adapter, i)
|
for_each_port(adapter, i)
|
||||||
if (adapter->port[i]->reg_state == NETREG_REGISTERED)
|
if (adapter->port[i]->reg_state == NETREG_REGISTERED)
|
||||||
unregister_netdev(adapter->port[i]);
|
unregister_netdev(adapter->port[i]);
|
||||||
|
|
|
@ -2171,7 +2171,7 @@ static void sge_rx_timer_cb(unsigned long data)
|
||||||
struct adapter *adap = (struct adapter *)data;
|
struct adapter *adap = (struct adapter *)data;
|
||||||
struct sge *s = &adap->sge;
|
struct sge *s = &adap->sge;
|
||||||
|
|
||||||
for (i = 0; i < ARRAY_SIZE(s->starving_fl); i++)
|
for (i = 0; i < BITS_TO_LONGS(s->egr_sz); i++)
|
||||||
for (m = s->starving_fl[i]; m; m &= m - 1) {
|
for (m = s->starving_fl[i]; m; m &= m - 1) {
|
||||||
struct sge_eth_rxq *rxq;
|
struct sge_eth_rxq *rxq;
|
||||||
unsigned int id = __ffs(m) + i * BITS_PER_LONG;
|
unsigned int id = __ffs(m) + i * BITS_PER_LONG;
|
||||||
|
@ -2259,7 +2259,7 @@ static void sge_tx_timer_cb(unsigned long data)
|
||||||
struct adapter *adap = (struct adapter *)data;
|
struct adapter *adap = (struct adapter *)data;
|
||||||
struct sge *s = &adap->sge;
|
struct sge *s = &adap->sge;
|
||||||
|
|
||||||
for (i = 0; i < ARRAY_SIZE(s->txq_maperr); i++)
|
for (i = 0; i < BITS_TO_LONGS(s->egr_sz); i++)
|
||||||
for (m = s->txq_maperr[i]; m; m &= m - 1) {
|
for (m = s->txq_maperr[i]; m; m &= m - 1) {
|
||||||
unsigned long id = __ffs(m) + i * BITS_PER_LONG;
|
unsigned long id = __ffs(m) + i * BITS_PER_LONG;
|
||||||
struct sge_ofld_txq *txq = s->egr_map[id];
|
struct sge_ofld_txq *txq = s->egr_map[id];
|
||||||
|
@ -2741,7 +2741,8 @@ void t4_free_sge_resources(struct adapter *adap)
|
||||||
free_rspq_fl(adap, &adap->sge.intrq, NULL);
|
free_rspq_fl(adap, &adap->sge.intrq, NULL);
|
||||||
|
|
||||||
/* clear the reverse egress queue map */
|
/* clear the reverse egress queue map */
|
||||||
memset(adap->sge.egr_map, 0, sizeof(adap->sge.egr_map));
|
memset(adap->sge.egr_map, 0,
|
||||||
|
adap->sge.egr_sz * sizeof(*adap->sge.egr_map));
|
||||||
}
|
}
|
||||||
|
|
||||||
void t4_sge_start(struct adapter *adap)
|
void t4_sge_start(struct adapter *adap)
|
||||||
|
|
|
@ -4458,6 +4458,59 @@ int cxgb4_t4_bar2_sge_qregs(struct adapter *adapter,
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* t4_init_devlog_params - initialize adapter->params.devlog
|
||||||
|
* @adap: the adapter
|
||||||
|
*
|
||||||
|
* Initialize various fields of the adapter's Firmware Device Log
|
||||||
|
* Parameters structure.
|
||||||
|
*/
|
||||||
|
int t4_init_devlog_params(struct adapter *adap)
|
||||||
|
{
|
||||||
|
struct devlog_params *dparams = &adap->params.devlog;
|
||||||
|
u32 pf_dparams;
|
||||||
|
unsigned int devlog_meminfo;
|
||||||
|
struct fw_devlog_cmd devlog_cmd;
|
||||||
|
int ret;
|
||||||
|
|
||||||
|
/* If we're dealing with newer firmware, the Device Log Paramerters
|
||||||
|
* are stored in a designated register which allows us to access the
|
||||||
|
* Device Log even if we can't talk to the firmware.
|
||||||
|
*/
|
||||||
|
pf_dparams =
|
||||||
|
t4_read_reg(adap, PCIE_FW_REG(PCIE_FW_PF_A, PCIE_FW_PF_DEVLOG));
|
||||||
|
if (pf_dparams) {
|
||||||
|
unsigned int nentries, nentries128;
|
||||||
|
|
||||||
|
dparams->memtype = PCIE_FW_PF_DEVLOG_MEMTYPE_G(pf_dparams);
|
||||||
|
dparams->start = PCIE_FW_PF_DEVLOG_ADDR16_G(pf_dparams) << 4;
|
||||||
|
|
||||||
|
nentries128 = PCIE_FW_PF_DEVLOG_NENTRIES128_G(pf_dparams);
|
||||||
|
nentries = (nentries128 + 1) * 128;
|
||||||
|
dparams->size = nentries * sizeof(struct fw_devlog_e);
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Otherwise, ask the firmware for it's Device Log Parameters.
|
||||||
|
*/
|
||||||
|
memset(&devlog_cmd, 0, sizeof(devlog_cmd));
|
||||||
|
devlog_cmd.op_to_write = htonl(FW_CMD_OP_V(FW_DEVLOG_CMD) |
|
||||||
|
FW_CMD_REQUEST_F | FW_CMD_READ_F);
|
||||||
|
devlog_cmd.retval_len16 = htonl(FW_LEN16(devlog_cmd));
|
||||||
|
ret = t4_wr_mbox(adap, adap->mbox, &devlog_cmd, sizeof(devlog_cmd),
|
||||||
|
&devlog_cmd);
|
||||||
|
if (ret)
|
||||||
|
return ret;
|
||||||
|
|
||||||
|
devlog_meminfo = ntohl(devlog_cmd.memtype_devlog_memaddr16_devlog);
|
||||||
|
dparams->memtype = FW_DEVLOG_CMD_MEMTYPE_DEVLOG_G(devlog_meminfo);
|
||||||
|
dparams->start = FW_DEVLOG_CMD_MEMADDR16_DEVLOG_G(devlog_meminfo) << 4;
|
||||||
|
dparams->size = ntohl(devlog_cmd.memsize_devlog);
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* t4_init_sge_params - initialize adap->params.sge
|
* t4_init_sge_params - initialize adap->params.sge
|
||||||
* @adapter: the adapter
|
* @adapter: the adapter
|
||||||
|
|
|
@ -63,6 +63,8 @@
|
||||||
#define MC_BIST_STATUS_REG(reg_addr, idx) ((reg_addr) + (idx) * 4)
|
#define MC_BIST_STATUS_REG(reg_addr, idx) ((reg_addr) + (idx) * 4)
|
||||||
#define EDC_BIST_STATUS_REG(reg_addr, idx) ((reg_addr) + (idx) * 4)
|
#define EDC_BIST_STATUS_REG(reg_addr, idx) ((reg_addr) + (idx) * 4)
|
||||||
|
|
||||||
|
#define PCIE_FW_REG(reg_addr, idx) ((reg_addr) + (idx) * 4)
|
||||||
|
|
||||||
#define SGE_PF_KDOORBELL_A 0x0
|
#define SGE_PF_KDOORBELL_A 0x0
|
||||||
|
|
||||||
#define QID_S 15
|
#define QID_S 15
|
||||||
|
@ -707,6 +709,7 @@
|
||||||
#define PFNUM_V(x) ((x) << PFNUM_S)
|
#define PFNUM_V(x) ((x) << PFNUM_S)
|
||||||
|
|
||||||
#define PCIE_FW_A 0x30b8
|
#define PCIE_FW_A 0x30b8
|
||||||
|
#define PCIE_FW_PF_A 0x30bc
|
||||||
|
|
||||||
#define PCIE_CORE_UTL_SYSTEM_BUS_AGENT_STATUS_A 0x5908
|
#define PCIE_CORE_UTL_SYSTEM_BUS_AGENT_STATUS_A 0x5908
|
||||||
|
|
||||||
|
|
|
@ -101,7 +101,7 @@ enum fw_wr_opcodes {
|
||||||
FW_RI_BIND_MW_WR = 0x18,
|
FW_RI_BIND_MW_WR = 0x18,
|
||||||
FW_RI_FR_NSMR_WR = 0x19,
|
FW_RI_FR_NSMR_WR = 0x19,
|
||||||
FW_RI_INV_LSTAG_WR = 0x1a,
|
FW_RI_INV_LSTAG_WR = 0x1a,
|
||||||
FW_LASTC2E_WR = 0x40
|
FW_LASTC2E_WR = 0x70
|
||||||
};
|
};
|
||||||
|
|
||||||
struct fw_wr_hdr {
|
struct fw_wr_hdr {
|
||||||
|
@ -993,6 +993,7 @@ enum fw_memtype_cf {
|
||||||
FW_MEMTYPE_CF_EXTMEM = 0x2,
|
FW_MEMTYPE_CF_EXTMEM = 0x2,
|
||||||
FW_MEMTYPE_CF_FLASH = 0x4,
|
FW_MEMTYPE_CF_FLASH = 0x4,
|
||||||
FW_MEMTYPE_CF_INTERNAL = 0x5,
|
FW_MEMTYPE_CF_INTERNAL = 0x5,
|
||||||
|
FW_MEMTYPE_CF_EXTMEM1 = 0x6,
|
||||||
};
|
};
|
||||||
|
|
||||||
struct fw_caps_config_cmd {
|
struct fw_caps_config_cmd {
|
||||||
|
@ -1035,6 +1036,7 @@ enum fw_params_mnem {
|
||||||
FW_PARAMS_MNEM_PFVF = 2, /* function params */
|
FW_PARAMS_MNEM_PFVF = 2, /* function params */
|
||||||
FW_PARAMS_MNEM_REG = 3, /* limited register access */
|
FW_PARAMS_MNEM_REG = 3, /* limited register access */
|
||||||
FW_PARAMS_MNEM_DMAQ = 4, /* dma queue params */
|
FW_PARAMS_MNEM_DMAQ = 4, /* dma queue params */
|
||||||
|
FW_PARAMS_MNEM_CHNET = 5, /* chnet params */
|
||||||
FW_PARAMS_MNEM_LAST
|
FW_PARAMS_MNEM_LAST
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -3102,7 +3104,8 @@ enum fw_devlog_facility {
|
||||||
FW_DEVLOG_FACILITY_FCOE = 0x2E,
|
FW_DEVLOG_FACILITY_FCOE = 0x2E,
|
||||||
FW_DEVLOG_FACILITY_FOISCSI = 0x30,
|
FW_DEVLOG_FACILITY_FOISCSI = 0x30,
|
||||||
FW_DEVLOG_FACILITY_FOFCOE = 0x32,
|
FW_DEVLOG_FACILITY_FOFCOE = 0x32,
|
||||||
FW_DEVLOG_FACILITY_MAX = 0x32,
|
FW_DEVLOG_FACILITY_CHNET = 0x34,
|
||||||
|
FW_DEVLOG_FACILITY_MAX = 0x34,
|
||||||
};
|
};
|
||||||
|
|
||||||
/* log message format */
|
/* log message format */
|
||||||
|
@ -3139,4 +3142,36 @@ struct fw_devlog_cmd {
|
||||||
(((x) >> FW_DEVLOG_CMD_MEMADDR16_DEVLOG_S) & \
|
(((x) >> FW_DEVLOG_CMD_MEMADDR16_DEVLOG_S) & \
|
||||||
FW_DEVLOG_CMD_MEMADDR16_DEVLOG_M)
|
FW_DEVLOG_CMD_MEMADDR16_DEVLOG_M)
|
||||||
|
|
||||||
|
/* P C I E F W P F 7 R E G I S T E R */
|
||||||
|
|
||||||
|
/* PF7 stores the Firmware Device Log parameters which allows Host Drivers to
|
||||||
|
* access the "devlog" which needing to contact firmware. The encoding is
|
||||||
|
* mostly the same as that returned by the DEVLOG command except for the size
|
||||||
|
* which is encoded as the number of entries in multiples-1 of 128 here rather
|
||||||
|
* than the memory size as is done in the DEVLOG command. Thus, 0 means 128
|
||||||
|
* and 15 means 2048. This of course in turn constrains the allowed values
|
||||||
|
* for the devlog size ...
|
||||||
|
*/
|
||||||
|
#define PCIE_FW_PF_DEVLOG 7
|
||||||
|
|
||||||
|
#define PCIE_FW_PF_DEVLOG_NENTRIES128_S 28
|
||||||
|
#define PCIE_FW_PF_DEVLOG_NENTRIES128_M 0xf
|
||||||
|
#define PCIE_FW_PF_DEVLOG_NENTRIES128_V(x) \
|
||||||
|
((x) << PCIE_FW_PF_DEVLOG_NENTRIES128_S)
|
||||||
|
#define PCIE_FW_PF_DEVLOG_NENTRIES128_G(x) \
|
||||||
|
(((x) >> PCIE_FW_PF_DEVLOG_NENTRIES128_S) & \
|
||||||
|
PCIE_FW_PF_DEVLOG_NENTRIES128_M)
|
||||||
|
|
||||||
|
#define PCIE_FW_PF_DEVLOG_ADDR16_S 4
|
||||||
|
#define PCIE_FW_PF_DEVLOG_ADDR16_M 0xffffff
|
||||||
|
#define PCIE_FW_PF_DEVLOG_ADDR16_V(x) ((x) << PCIE_FW_PF_DEVLOG_ADDR16_S)
|
||||||
|
#define PCIE_FW_PF_DEVLOG_ADDR16_G(x) \
|
||||||
|
(((x) >> PCIE_FW_PF_DEVLOG_ADDR16_S) & PCIE_FW_PF_DEVLOG_ADDR16_M)
|
||||||
|
|
||||||
|
#define PCIE_FW_PF_DEVLOG_MEMTYPE_S 0
|
||||||
|
#define PCIE_FW_PF_DEVLOG_MEMTYPE_M 0xf
|
||||||
|
#define PCIE_FW_PF_DEVLOG_MEMTYPE_V(x) ((x) << PCIE_FW_PF_DEVLOG_MEMTYPE_S)
|
||||||
|
#define PCIE_FW_PF_DEVLOG_MEMTYPE_G(x) \
|
||||||
|
(((x) >> PCIE_FW_PF_DEVLOG_MEMTYPE_S) & PCIE_FW_PF_DEVLOG_MEMTYPE_M)
|
||||||
|
|
||||||
#endif /* _T4FW_INTERFACE_H_ */
|
#endif /* _T4FW_INTERFACE_H_ */
|
||||||
|
|
|
@ -36,13 +36,13 @@
|
||||||
#define __T4FW_VERSION_H__
|
#define __T4FW_VERSION_H__
|
||||||
|
|
||||||
#define T4FW_VERSION_MAJOR 0x01
|
#define T4FW_VERSION_MAJOR 0x01
|
||||||
#define T4FW_VERSION_MINOR 0x0C
|
#define T4FW_VERSION_MINOR 0x0D
|
||||||
#define T4FW_VERSION_MICRO 0x19
|
#define T4FW_VERSION_MICRO 0x20
|
||||||
#define T4FW_VERSION_BUILD 0x00
|
#define T4FW_VERSION_BUILD 0x00
|
||||||
|
|
||||||
#define T5FW_VERSION_MAJOR 0x01
|
#define T5FW_VERSION_MAJOR 0x01
|
||||||
#define T5FW_VERSION_MINOR 0x0C
|
#define T5FW_VERSION_MINOR 0x0D
|
||||||
#define T5FW_VERSION_MICRO 0x19
|
#define T5FW_VERSION_MICRO 0x20
|
||||||
#define T5FW_VERSION_BUILD 0x00
|
#define T5FW_VERSION_BUILD 0x00
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -1004,7 +1004,7 @@ static inline void ring_tx_db(struct adapter *adapter, struct sge_txq *tq,
|
||||||
? (tq->pidx - 1)
|
? (tq->pidx - 1)
|
||||||
: (tq->size - 1));
|
: (tq->size - 1));
|
||||||
__be64 *src = (__be64 *)&tq->desc[index];
|
__be64 *src = (__be64 *)&tq->desc[index];
|
||||||
__be64 __iomem *dst = (__be64 *)(tq->bar2_addr +
|
__be64 __iomem *dst = (__be64 __iomem *)(tq->bar2_addr +
|
||||||
SGE_UDB_WCDOORBELL);
|
SGE_UDB_WCDOORBELL);
|
||||||
unsigned int count = EQ_UNIT / sizeof(__be64);
|
unsigned int count = EQ_UNIT / sizeof(__be64);
|
||||||
|
|
||||||
|
@ -1018,7 +1018,11 @@ static inline void ring_tx_db(struct adapter *adapter, struct sge_txq *tq,
|
||||||
* DMA.
|
* DMA.
|
||||||
*/
|
*/
|
||||||
while (count) {
|
while (count) {
|
||||||
writeq(*src, dst);
|
/* the (__force u64) is because the compiler
|
||||||
|
* doesn't understand the endian swizzling
|
||||||
|
* going on
|
||||||
|
*/
|
||||||
|
writeq((__force u64)*src, dst);
|
||||||
src++;
|
src++;
|
||||||
dst++;
|
dst++;
|
||||||
count--;
|
count--;
|
||||||
|
@ -1252,8 +1256,8 @@ int t4vf_eth_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
BUG_ON(DIV_ROUND_UP(ETHTXQ_MAX_HDR, TXD_PER_EQ_UNIT) > 1);
|
BUG_ON(DIV_ROUND_UP(ETHTXQ_MAX_HDR, TXD_PER_EQ_UNIT) > 1);
|
||||||
wr = (void *)&txq->q.desc[txq->q.pidx];
|
wr = (void *)&txq->q.desc[txq->q.pidx];
|
||||||
wr->equiq_to_len16 = cpu_to_be32(wr_mid);
|
wr->equiq_to_len16 = cpu_to_be32(wr_mid);
|
||||||
wr->r3[0] = cpu_to_be64(0);
|
wr->r3[0] = cpu_to_be32(0);
|
||||||
wr->r3[1] = cpu_to_be64(0);
|
wr->r3[1] = cpu_to_be32(0);
|
||||||
skb_copy_from_linear_data(skb, (void *)wr->ethmacdst, fw_hdr_copy_len);
|
skb_copy_from_linear_data(skb, (void *)wr->ethmacdst, fw_hdr_copy_len);
|
||||||
end = (u64 *)wr + flits;
|
end = (u64 *)wr + flits;
|
||||||
|
|
||||||
|
|
|
@ -210,10 +210,10 @@ int t4vf_wr_mbox_core(struct adapter *adapter, const void *cmd, int size,
|
||||||
|
|
||||||
if (rpl) {
|
if (rpl) {
|
||||||
/* request bit in high-order BE word */
|
/* request bit in high-order BE word */
|
||||||
WARN_ON((be32_to_cpu(*(const u32 *)cmd)
|
WARN_ON((be32_to_cpu(*(const __be32 *)cmd)
|
||||||
& FW_CMD_REQUEST_F) == 0);
|
& FW_CMD_REQUEST_F) == 0);
|
||||||
get_mbox_rpl(adapter, rpl, size, mbox_data);
|
get_mbox_rpl(adapter, rpl, size, mbox_data);
|
||||||
WARN_ON((be32_to_cpu(*(u32 *)rpl)
|
WARN_ON((be32_to_cpu(*(__be32 *)rpl)
|
||||||
& FW_CMD_REQUEST_F) != 0);
|
& FW_CMD_REQUEST_F) != 0);
|
||||||
}
|
}
|
||||||
t4_write_reg(adapter, mbox_ctl,
|
t4_write_reg(adapter, mbox_ctl,
|
||||||
|
@ -484,7 +484,7 @@ int t4_bar2_sge_qregs(struct adapter *adapter,
|
||||||
* o The BAR2 Queue ID.
|
* o The BAR2 Queue ID.
|
||||||
* o The BAR2 Queue ID Offset into the BAR2 page.
|
* o The BAR2 Queue ID Offset into the BAR2 page.
|
||||||
*/
|
*/
|
||||||
bar2_page_offset = ((qid >> qpp_shift) << page_shift);
|
bar2_page_offset = ((u64)(qid >> qpp_shift) << page_shift);
|
||||||
bar2_qid = qid & qpp_mask;
|
bar2_qid = qid & qpp_mask;
|
||||||
bar2_qid_offset = bar2_qid * SGE_UDB_SIZE;
|
bar2_qid_offset = bar2_qid * SGE_UDB_SIZE;
|
||||||
|
|
||||||
|
|
|
@ -1954,6 +1954,7 @@ static int fec_enet_mii_init(struct platform_device *pdev)
|
||||||
struct fec_enet_private *fep = netdev_priv(ndev);
|
struct fec_enet_private *fep = netdev_priv(ndev);
|
||||||
struct device_node *node;
|
struct device_node *node;
|
||||||
int err = -ENXIO, i;
|
int err = -ENXIO, i;
|
||||||
|
u32 mii_speed, holdtime;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* The i.MX28 dual fec interfaces are not equal.
|
* The i.MX28 dual fec interfaces are not equal.
|
||||||
|
@ -1991,10 +1992,33 @@ static int fec_enet_mii_init(struct platform_device *pdev)
|
||||||
* Reference Manual has an error on this, and gets fixed on i.MX6Q
|
* Reference Manual has an error on this, and gets fixed on i.MX6Q
|
||||||
* document.
|
* document.
|
||||||
*/
|
*/
|
||||||
fep->phy_speed = DIV_ROUND_UP(clk_get_rate(fep->clk_ipg), 5000000);
|
mii_speed = DIV_ROUND_UP(clk_get_rate(fep->clk_ipg), 5000000);
|
||||||
if (fep->quirks & FEC_QUIRK_ENET_MAC)
|
if (fep->quirks & FEC_QUIRK_ENET_MAC)
|
||||||
fep->phy_speed--;
|
mii_speed--;
|
||||||
fep->phy_speed <<= 1;
|
if (mii_speed > 63) {
|
||||||
|
dev_err(&pdev->dev,
|
||||||
|
"fec clock (%lu) to fast to get right mii speed\n",
|
||||||
|
clk_get_rate(fep->clk_ipg));
|
||||||
|
err = -EINVAL;
|
||||||
|
goto err_out;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* The i.MX28 and i.MX6 types have another filed in the MSCR (aka
|
||||||
|
* MII_SPEED) register that defines the MDIO output hold time. Earlier
|
||||||
|
* versions are RAZ there, so just ignore the difference and write the
|
||||||
|
* register always.
|
||||||
|
* The minimal hold time according to IEE802.3 (clause 22) is 10 ns.
|
||||||
|
* HOLDTIME + 1 is the number of clk cycles the fec is holding the
|
||||||
|
* output.
|
||||||
|
* The HOLDTIME bitfield takes values between 0 and 7 (inclusive).
|
||||||
|
* Given that ceil(clkrate / 5000000) <= 64, the calculation for
|
||||||
|
* holdtime cannot result in a value greater than 3.
|
||||||
|
*/
|
||||||
|
holdtime = DIV_ROUND_UP(clk_get_rate(fep->clk_ipg), 100000000) - 1;
|
||||||
|
|
||||||
|
fep->phy_speed = mii_speed << 1 | holdtime << 8;
|
||||||
|
|
||||||
writel(fep->phy_speed, fep->hwp + FEC_MII_SPEED);
|
writel(fep->phy_speed, fep->hwp + FEC_MII_SPEED);
|
||||||
|
|
||||||
fep->mii_bus = mdiobus_alloc();
|
fep->mii_bus = mdiobus_alloc();
|
||||||
|
|
|
@ -3893,6 +3893,9 @@ static int ucc_geth_probe(struct platform_device* ofdev)
|
||||||
ugeth->phy_interface = phy_interface;
|
ugeth->phy_interface = phy_interface;
|
||||||
ugeth->max_speed = max_speed;
|
ugeth->max_speed = max_speed;
|
||||||
|
|
||||||
|
/* Carrier starts down, phylib will bring it up */
|
||||||
|
netif_carrier_off(dev);
|
||||||
|
|
||||||
err = register_netdev(dev);
|
err = register_netdev(dev);
|
||||||
if (err) {
|
if (err) {
|
||||||
if (netif_msg_probe(ugeth))
|
if (netif_msg_probe(ugeth))
|
||||||
|
|
|
@ -2658,16 +2658,11 @@ static int mvneta_stop(struct net_device *dev)
|
||||||
static int mvneta_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
|
static int mvneta_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
|
||||||
{
|
{
|
||||||
struct mvneta_port *pp = netdev_priv(dev);
|
struct mvneta_port *pp = netdev_priv(dev);
|
||||||
int ret;
|
|
||||||
|
|
||||||
if (!pp->phy_dev)
|
if (!pp->phy_dev)
|
||||||
return -ENOTSUPP;
|
return -ENOTSUPP;
|
||||||
|
|
||||||
ret = phy_mii_ioctl(pp->phy_dev, ifr, cmd);
|
return phy_mii_ioctl(pp->phy_dev, ifr, cmd);
|
||||||
if (!ret)
|
|
||||||
mvneta_adjust_link(dev);
|
|
||||||
|
|
||||||
return ret;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Ethtool methods */
|
/* Ethtool methods */
|
||||||
|
|
|
@ -724,7 +724,8 @@ static int mlx4_cmd_wait(struct mlx4_dev *dev, u64 in_param, u64 *out_param,
|
||||||
* on the host, we deprecate the error message for this
|
* on the host, we deprecate the error message for this
|
||||||
* specific command/input_mod/opcode_mod/fw-status to be debug.
|
* specific command/input_mod/opcode_mod/fw-status to be debug.
|
||||||
*/
|
*/
|
||||||
if (op == MLX4_CMD_SET_PORT && in_modifier == 1 &&
|
if (op == MLX4_CMD_SET_PORT &&
|
||||||
|
(in_modifier == 1 || in_modifier == 2) &&
|
||||||
op_modifier == 0 && context->fw_status == CMD_STAT_BAD_SIZE)
|
op_modifier == 0 && context->fw_status == CMD_STAT_BAD_SIZE)
|
||||||
mlx4_dbg(dev, "command 0x%x failed: fw status = 0x%x\n",
|
mlx4_dbg(dev, "command 0x%x failed: fw status = 0x%x\n",
|
||||||
op, context->fw_status);
|
op, context->fw_status);
|
||||||
|
@ -1993,7 +1994,6 @@ static void mlx4_master_do_cmd(struct mlx4_dev *dev, int slave, u8 cmd,
|
||||||
goto reset_slave;
|
goto reset_slave;
|
||||||
slave_state[slave].vhcr_dma = ((u64) param) << 48;
|
slave_state[slave].vhcr_dma = ((u64) param) << 48;
|
||||||
priv->mfunc.master.slave_state[slave].cookie = 0;
|
priv->mfunc.master.slave_state[slave].cookie = 0;
|
||||||
mutex_init(&priv->mfunc.master.gen_eqe_mutex[slave]);
|
|
||||||
break;
|
break;
|
||||||
case MLX4_COMM_CMD_VHCR1:
|
case MLX4_COMM_CMD_VHCR1:
|
||||||
if (slave_state[slave].last_cmd != MLX4_COMM_CMD_VHCR0)
|
if (slave_state[slave].last_cmd != MLX4_COMM_CMD_VHCR0)
|
||||||
|
@ -2225,6 +2225,7 @@ int mlx4_multi_func_init(struct mlx4_dev *dev)
|
||||||
for (i = 0; i < dev->num_slaves; ++i) {
|
for (i = 0; i < dev->num_slaves; ++i) {
|
||||||
s_state = &priv->mfunc.master.slave_state[i];
|
s_state = &priv->mfunc.master.slave_state[i];
|
||||||
s_state->last_cmd = MLX4_COMM_CMD_RESET;
|
s_state->last_cmd = MLX4_COMM_CMD_RESET;
|
||||||
|
mutex_init(&priv->mfunc.master.gen_eqe_mutex[i]);
|
||||||
for (j = 0; j < MLX4_EVENT_TYPES_NUM; ++j)
|
for (j = 0; j < MLX4_EVENT_TYPES_NUM; ++j)
|
||||||
s_state->event_eq[j].eqn = -1;
|
s_state->event_eq[j].eqn = -1;
|
||||||
__raw_writel((__force u32) 0,
|
__raw_writel((__force u32) 0,
|
||||||
|
|
|
@ -2805,13 +2805,6 @@ int mlx4_en_init_netdev(struct mlx4_en_dev *mdev, int port,
|
||||||
netif_carrier_off(dev);
|
netif_carrier_off(dev);
|
||||||
mlx4_en_set_default_moderation(priv);
|
mlx4_en_set_default_moderation(priv);
|
||||||
|
|
||||||
err = register_netdev(dev);
|
|
||||||
if (err) {
|
|
||||||
en_err(priv, "Netdev registration failed for port %d\n", port);
|
|
||||||
goto out;
|
|
||||||
}
|
|
||||||
priv->registered = 1;
|
|
||||||
|
|
||||||
en_warn(priv, "Using %d TX rings\n", prof->tx_ring_num);
|
en_warn(priv, "Using %d TX rings\n", prof->tx_ring_num);
|
||||||
en_warn(priv, "Using %d RX rings\n", prof->rx_ring_num);
|
en_warn(priv, "Using %d RX rings\n", prof->rx_ring_num);
|
||||||
|
|
||||||
|
@ -2853,6 +2846,14 @@ int mlx4_en_init_netdev(struct mlx4_en_dev *mdev, int port,
|
||||||
|
|
||||||
mlx4_set_stats_bitmap(mdev->dev, &priv->stats_bitmap);
|
mlx4_set_stats_bitmap(mdev->dev, &priv->stats_bitmap);
|
||||||
|
|
||||||
|
err = register_netdev(dev);
|
||||||
|
if (err) {
|
||||||
|
en_err(priv, "Netdev registration failed for port %d\n", port);
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
|
priv->registered = 1;
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
out:
|
out:
|
||||||
|
|
|
@ -153,12 +153,10 @@ void mlx4_gen_slave_eqe(struct work_struct *work)
|
||||||
|
|
||||||
/* All active slaves need to receive the event */
|
/* All active slaves need to receive the event */
|
||||||
if (slave == ALL_SLAVES) {
|
if (slave == ALL_SLAVES) {
|
||||||
for (i = 0; i < dev->num_slaves; i++) {
|
for (i = 0; i <= dev->persist->num_vfs; i++) {
|
||||||
if (i != dev->caps.function &&
|
if (mlx4_GEN_EQE(dev, i, eqe))
|
||||||
master->slave_state[i].active)
|
mlx4_warn(dev, "Failed to generate event for slave %d\n",
|
||||||
if (mlx4_GEN_EQE(dev, i, eqe))
|
i);
|
||||||
mlx4_warn(dev, "Failed to generate event for slave %d\n",
|
|
||||||
i);
|
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
if (mlx4_GEN_EQE(dev, slave, eqe))
|
if (mlx4_GEN_EQE(dev, slave, eqe))
|
||||||
|
@ -203,13 +201,11 @@ static void mlx4_slave_event(struct mlx4_dev *dev, int slave,
|
||||||
struct mlx4_eqe *eqe)
|
struct mlx4_eqe *eqe)
|
||||||
{
|
{
|
||||||
struct mlx4_priv *priv = mlx4_priv(dev);
|
struct mlx4_priv *priv = mlx4_priv(dev);
|
||||||
struct mlx4_slave_state *s_slave =
|
|
||||||
&priv->mfunc.master.slave_state[slave];
|
|
||||||
|
|
||||||
if (!s_slave->active) {
|
if (slave < 0 || slave > dev->persist->num_vfs ||
|
||||||
/*mlx4_warn(dev, "Trying to pass event to inactive slave\n");*/
|
slave == dev->caps.function ||
|
||||||
|
!priv->mfunc.master.slave_state[slave].active)
|
||||||
return;
|
return;
|
||||||
}
|
|
||||||
|
|
||||||
slave_event(dev, slave, eqe);
|
slave_event(dev, slave, eqe);
|
||||||
}
|
}
|
||||||
|
|
|
@ -3095,6 +3095,12 @@ int mlx4_GEN_EQE(struct mlx4_dev *dev, int slave, struct mlx4_eqe *eqe)
|
||||||
if (!priv->mfunc.master.slave_state)
|
if (!priv->mfunc.master.slave_state)
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
|
||||||
|
/* check for slave valid, slave not PF, and slave active */
|
||||||
|
if (slave < 0 || slave > dev->persist->num_vfs ||
|
||||||
|
slave == dev->caps.function ||
|
||||||
|
!priv->mfunc.master.slave_state[slave].active)
|
||||||
|
return 0;
|
||||||
|
|
||||||
event_eq = &priv->mfunc.master.slave_state[slave].event_eq[eqe->type];
|
event_eq = &priv->mfunc.master.slave_state[slave].event_eq[eqe->type];
|
||||||
|
|
||||||
/* Create the event only if the slave is registered */
|
/* Create the event only if the slave is registered */
|
||||||
|
|
|
@ -4468,10 +4468,16 @@ static int rocker_port_master_changed(struct net_device *dev)
|
||||||
struct net_device *master = netdev_master_upper_dev_get(dev);
|
struct net_device *master = netdev_master_upper_dev_get(dev);
|
||||||
int err = 0;
|
int err = 0;
|
||||||
|
|
||||||
|
/* There are currently three cases handled here:
|
||||||
|
* 1. Joining a bridge
|
||||||
|
* 2. Leaving a previously joined bridge
|
||||||
|
* 3. Other, e.g. being added to or removed from a bond or openvswitch,
|
||||||
|
* in which case nothing is done
|
||||||
|
*/
|
||||||
if (master && master->rtnl_link_ops &&
|
if (master && master->rtnl_link_ops &&
|
||||||
!strcmp(master->rtnl_link_ops->kind, "bridge"))
|
!strcmp(master->rtnl_link_ops->kind, "bridge"))
|
||||||
err = rocker_port_bridge_join(rocker_port, master);
|
err = rocker_port_bridge_join(rocker_port, master);
|
||||||
else
|
else if (rocker_port_is_bridged(rocker_port))
|
||||||
err = rocker_port_bridge_leave(rocker_port);
|
err = rocker_port_bridge_leave(rocker_port);
|
||||||
|
|
||||||
return err;
|
return err;
|
||||||
|
|
|
@ -114,7 +114,9 @@ unsigned int ipvlan_mac_hash(const unsigned char *addr);
|
||||||
rx_handler_result_t ipvlan_handle_frame(struct sk_buff **pskb);
|
rx_handler_result_t ipvlan_handle_frame(struct sk_buff **pskb);
|
||||||
int ipvlan_queue_xmit(struct sk_buff *skb, struct net_device *dev);
|
int ipvlan_queue_xmit(struct sk_buff *skb, struct net_device *dev);
|
||||||
void ipvlan_ht_addr_add(struct ipvl_dev *ipvlan, struct ipvl_addr *addr);
|
void ipvlan_ht_addr_add(struct ipvl_dev *ipvlan, struct ipvl_addr *addr);
|
||||||
bool ipvlan_addr_busy(struct ipvl_dev *ipvlan, void *iaddr, bool is_v6);
|
struct ipvl_addr *ipvlan_find_addr(const struct ipvl_dev *ipvlan,
|
||||||
|
const void *iaddr, bool is_v6);
|
||||||
|
bool ipvlan_addr_busy(struct ipvl_port *port, void *iaddr, bool is_v6);
|
||||||
struct ipvl_addr *ipvlan_ht_addr_lookup(const struct ipvl_port *port,
|
struct ipvl_addr *ipvlan_ht_addr_lookup(const struct ipvl_port *port,
|
||||||
const void *iaddr, bool is_v6);
|
const void *iaddr, bool is_v6);
|
||||||
void ipvlan_ht_addr_del(struct ipvl_addr *addr, bool sync);
|
void ipvlan_ht_addr_del(struct ipvl_addr *addr, bool sync);
|
||||||
|
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue