7264 lines
222 KiB
C
7264 lines
222 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/* Copyright (c) 2013, Intel Corporation. */
|
|
|
|
#ifndef _KCOMPAT_H_
|
|
#define _KCOMPAT_H_
|
|
|
|
#ifndef LINUX_VERSION_CODE
|
|
#include <linux/version.h>
|
|
#else
|
|
#define KERNEL_VERSION(a,b,c) (((a) << 16) + ((b) << 8) + (c))
|
|
#endif
|
|
#include <linux/io.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/etherdevice.h>
|
|
#include <linux/ethtool.h>
|
|
#include <linux/if_vlan.h>
|
|
#include <linux/in.h>
|
|
#include <linux/if_link.h>
|
|
#include <linux/init.h>
|
|
#include <linux/ioport.h>
|
|
#include <linux/ip.h>
|
|
#include <linux/ipv6.h>
|
|
#include <linux/list.h>
|
|
#include <linux/mii.h>
|
|
#include <linux/module.h>
|
|
#include <linux/netdevice.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/skbuff.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/string.h>
|
|
#include <linux/tcp.h>
|
|
#include <linux/types.h>
|
|
#include <linux/udp.h>
|
|
#include <linux/vmalloc.h>
|
|
|
|
#ifndef GCC_VERSION
|
|
#define GCC_VERSION (__GNUC__ * 10000 \
|
|
+ __GNUC_MINOR__ * 100 \
|
|
+ __GNUC_PATCHLEVEL__)
|
|
#endif /* GCC_VERSION */
|
|
|
|
/* Backport macros for controlling GCC diagnostics */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(4,18,0) )
|
|
|
|
/* Compilers before gcc-4.6 do not understand "#pragma GCC diagnostic push" */
|
|
#if GCC_VERSION >= 40600
|
|
#define __diag_str1(s) #s
|
|
#define __diag_str(s) __diag_str1(s)
|
|
#define __diag(s) _Pragma(__diag_str(GCC diagnostic s))
|
|
#else
|
|
#define __diag(s)
|
|
#endif /* GCC_VERSION >= 4.6 */
|
|
#define __diag_push() __diag(push)
|
|
#define __diag_pop() __diag(pop)
|
|
#endif /* LINUX_VERSION < 4.18.0 */
|
|
|
|
#ifndef NSEC_PER_MSEC
|
|
#define NSEC_PER_MSEC 1000000L
|
|
#endif
|
|
#include <net/ipv6.h>
|
|
/* UTS_RELEASE is in a different header starting in kernel 2.6.18 */
|
|
#ifndef UTS_RELEASE
|
|
/* utsrelease.h changed locations in 2.6.33 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,33) )
|
|
#include <linux/utsrelease.h>
|
|
#else
|
|
#include <generated/utsrelease.h>
|
|
#endif
|
|
#endif
|
|
|
|
/* NAPI enable/disable flags here */
|
|
|
|
/* and finally set defines so that the code sees the changes */
|
|
#ifdef NAPI
|
|
#else
|
|
#endif /* NAPI */
|
|
|
|
/* Dynamic LTR and deeper C-State support disable/enable */
|
|
|
|
/* packet split disable/enable */
|
|
#ifdef DISABLE_PACKET_SPLIT
|
|
#endif /* DISABLE_PACKET_SPLIT */
|
|
|
|
/* MSI compatibility code for all kernels and drivers */
|
|
#ifdef DISABLE_PCI_MSI
|
|
#undef CONFIG_PCI_MSI
|
|
#endif
|
|
#ifndef CONFIG_PCI_MSI
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,8) )
|
|
struct msix_entry {
|
|
u16 vector; /* kernel uses to write allocated vector */
|
|
u16 entry; /* driver uses to specify entry, OS writes */
|
|
};
|
|
#endif
|
|
#undef pci_enable_msi
|
|
#define pci_enable_msi(a) -ENOTSUPP
|
|
#undef pci_disable_msi
|
|
#define pci_disable_msi(a) do {} while (0)
|
|
#undef pci_enable_msix
|
|
#define pci_enable_msix(a, b, c) -ENOTSUPP
|
|
#undef pci_disable_msix
|
|
#define pci_disable_msix(a) do {} while (0)
|
|
#define msi_remove_pci_irq_vectors(a) do {} while (0)
|
|
#endif /* CONFIG_PCI_MSI */
|
|
#ifdef DISABLE_PM
|
|
#undef CONFIG_PM
|
|
#endif
|
|
|
|
#ifdef DISABLE_NET_POLL_CONTROLLER
|
|
#undef CONFIG_NET_POLL_CONTROLLER
|
|
#endif
|
|
|
|
#ifndef PMSG_SUSPEND
|
|
#define PMSG_SUSPEND 3
|
|
#endif
|
|
|
|
/* generic boolean compatibility */
|
|
#undef TRUE
|
|
#undef FALSE
|
|
#define TRUE true
|
|
#define FALSE false
|
|
#ifdef GCC_VERSION
|
|
#if ( GCC_VERSION < 3000 )
|
|
#define _Bool char
|
|
#endif
|
|
#else
|
|
#define _Bool char
|
|
#endif
|
|
|
|
#ifndef BIT
|
|
#define BIT(nr) (1UL << (nr))
|
|
#endif
|
|
|
|
#undef __always_unused
|
|
#define __always_unused __attribute__((__unused__))
|
|
|
|
#undef __maybe_unused
|
|
#define __maybe_unused __attribute__((__unused__))
|
|
|
|
/* kernels less than 2.4.14 don't have this */
|
|
#ifndef ETH_P_8021Q
|
|
#define ETH_P_8021Q 0x8100
|
|
#endif
|
|
|
|
#ifndef module_param
|
|
#define module_param(v,t,p) MODULE_PARM(v, "i");
|
|
#endif
|
|
|
|
#ifndef DMA_64BIT_MASK
|
|
#define DMA_64BIT_MASK 0xffffffffffffffffULL
|
|
#endif
|
|
|
|
#ifndef DMA_32BIT_MASK
|
|
#define DMA_32BIT_MASK 0x00000000ffffffffULL
|
|
#endif
|
|
|
|
#ifndef PCI_CAP_ID_EXP
|
|
#define PCI_CAP_ID_EXP 0x10
|
|
#endif
|
|
|
|
#ifndef uninitialized_var
|
|
#define uninitialized_var(x) x = x
|
|
#endif
|
|
|
|
#ifndef PCIE_LINK_STATE_L0S
|
|
#define PCIE_LINK_STATE_L0S 1
|
|
#endif
|
|
#ifndef PCIE_LINK_STATE_L1
|
|
#define PCIE_LINK_STATE_L1 2
|
|
#endif
|
|
|
|
#ifndef SET_NETDEV_DEV
|
|
#define SET_NETDEV_DEV(net, pdev)
|
|
#endif
|
|
|
|
#if !defined(HAVE_FREE_NETDEV) && ( LINUX_VERSION_CODE < KERNEL_VERSION(3,1,0) )
|
|
#define free_netdev(x) kfree(x)
|
|
#endif
|
|
|
|
#ifdef HAVE_POLL_CONTROLLER
|
|
#define CONFIG_NET_POLL_CONTROLLER
|
|
#endif
|
|
|
|
#ifndef SKB_DATAREF_SHIFT
|
|
/* if we do not have the infrastructure to detect if skb_header is cloned
|
|
just return false in all cases */
|
|
#define skb_header_cloned(x) 0
|
|
#endif
|
|
|
|
#ifndef NETIF_F_GSO
|
|
#define gso_size tso_size
|
|
#define gso_segs tso_segs
|
|
#endif
|
|
|
|
#ifndef NETIF_F_GRO
|
|
#define vlan_gro_receive(_napi, _vlgrp, _vlan, _skb) \
|
|
vlan_hwaccel_receive_skb(_skb, _vlgrp, _vlan)
|
|
#define napi_gro_receive(_napi, _skb) netif_receive_skb(_skb)
|
|
#endif
|
|
|
|
#ifndef NETIF_F_SCTP_CSUM
|
|
#define NETIF_F_SCTP_CSUM 0
|
|
#endif
|
|
|
|
#ifndef NETIF_F_LRO
|
|
#define NETIF_F_LRO BIT(15)
|
|
#endif
|
|
|
|
#ifndef NETIF_F_NTUPLE
|
|
#define NETIF_F_NTUPLE BIT(27)
|
|
#endif
|
|
|
|
#ifndef NETIF_F_ALL_FCOE
|
|
#define NETIF_F_ALL_FCOE (NETIF_F_FCOE_CRC | NETIF_F_FCOE_MTU | \
|
|
NETIF_F_FSO)
|
|
#endif
|
|
|
|
#ifndef IPPROTO_SCTP
|
|
#define IPPROTO_SCTP 132
|
|
#endif
|
|
|
|
#ifndef IPPROTO_UDPLITE
|
|
#define IPPROTO_UDPLITE 136
|
|
#endif
|
|
|
|
#ifndef CHECKSUM_PARTIAL
|
|
#define CHECKSUM_PARTIAL CHECKSUM_HW
|
|
#define CHECKSUM_COMPLETE CHECKSUM_HW
|
|
#endif
|
|
|
|
#ifndef __read_mostly
|
|
#define __read_mostly
|
|
#endif
|
|
|
|
#ifndef MII_RESV1
|
|
#define MII_RESV1 0x17 /* Reserved... */
|
|
#endif
|
|
|
|
#ifndef unlikely
|
|
#define unlikely(_x) _x
|
|
#define likely(_x) _x
|
|
#endif
|
|
|
|
#ifndef WARN_ON
|
|
#define WARN_ON(x) ({0;})
|
|
#endif
|
|
|
|
#ifndef PCI_DEVICE
|
|
#define PCI_DEVICE(vend,dev) \
|
|
.vendor = (vend), .device = (dev), \
|
|
.subvendor = PCI_ANY_ID, .subdevice = PCI_ANY_ID
|
|
#endif
|
|
|
|
#ifndef node_online
|
|
#define node_online(node) ((node) == 0)
|
|
#endif
|
|
|
|
#ifndef _LINUX_RANDOM_H
|
|
#include <linux/random.h>
|
|
#endif
|
|
|
|
#ifndef BITS_PER_TYPE
|
|
#define BITS_PER_TYPE(type) (sizeof(type) * BITS_PER_BYTE)
|
|
#endif
|
|
|
|
#ifndef BITS_TO_LONGS
|
|
#define BITS_TO_LONGS(bits) (((bits)+BITS_PER_LONG-1)/BITS_PER_LONG)
|
|
#endif
|
|
|
|
#ifndef DECLARE_BITMAP
|
|
#define DECLARE_BITMAP(name,bits) long name[BITS_TO_LONGS(bits)]
|
|
#endif
|
|
|
|
#ifndef VLAN_HLEN
|
|
#define VLAN_HLEN 4
|
|
#endif
|
|
|
|
#ifndef VLAN_ETH_HLEN
|
|
#define VLAN_ETH_HLEN 18
|
|
#endif
|
|
|
|
#ifndef VLAN_ETH_FRAME_LEN
|
|
#define VLAN_ETH_FRAME_LEN 1518
|
|
#endif
|
|
|
|
#ifndef DCA_GET_TAG_TWO_ARGS
|
|
#define dca3_get_tag(a,b) dca_get_tag(b)
|
|
#endif
|
|
|
|
#ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
|
|
#if defined(__i386__) || defined(__x86_64__)
|
|
#define CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
|
|
#endif
|
|
#endif
|
|
|
|
/* taken from 2.6.24 definition in linux/kernel.h */
|
|
#ifndef IS_ALIGNED
|
|
#define IS_ALIGNED(x,a) (((x) % ((typeof(x))(a))) == 0)
|
|
#endif
|
|
|
|
#ifdef IS_ENABLED
|
|
#undef IS_ENABLED
|
|
#undef __ARG_PLACEHOLDER_1
|
|
#undef config_enabled
|
|
#undef _config_enabled
|
|
#undef __config_enabled
|
|
#undef ___config_enabled
|
|
#endif
|
|
|
|
#define __ARG_PLACEHOLDER_1 0,
|
|
#define config_enabled(cfg) _config_enabled(cfg)
|
|
#ifdef __CHECKER__
|
|
/* cppcheck-suppress preprocessorErrorDirective */
|
|
#endif /* __CHECKER__ */
|
|
#define _config_enabled(value) __config_enabled(__ARG_PLACEHOLDER_##value)
|
|
#define __config_enabled(arg1_or_junk) ___config_enabled(arg1_or_junk 1, 0)
|
|
#define ___config_enabled(__ignored, val, ...) val
|
|
|
|
#define IS_ENABLED(option) \
|
|
(config_enabled(option) || config_enabled(option##_MODULE))
|
|
|
|
#if !defined(NETIF_F_HW_VLAN_TX) && !defined(NETIF_F_HW_VLAN_CTAG_TX)
|
|
struct _kc_vlan_ethhdr {
|
|
unsigned char h_dest[ETH_ALEN];
|
|
unsigned char h_source[ETH_ALEN];
|
|
__be16 h_vlan_proto;
|
|
__be16 h_vlan_TCI;
|
|
__be16 h_vlan_encapsulated_proto;
|
|
};
|
|
#define vlan_ethhdr _kc_vlan_ethhdr
|
|
struct _kc_vlan_hdr {
|
|
__be16 h_vlan_TCI;
|
|
__be16 h_vlan_encapsulated_proto;
|
|
};
|
|
#define vlan_hdr _kc_vlan_hdr
|
|
#define vlan_tx_tag_present(_skb) 0
|
|
#define vlan_tx_tag_get(_skb) 0
|
|
#endif /* NETIF_F_HW_VLAN_TX && NETIF_F_HW_VLAN_CTAG_TX */
|
|
|
|
#ifndef VLAN_PRIO_SHIFT
|
|
#define VLAN_PRIO_SHIFT 13
|
|
#endif
|
|
|
|
#ifndef PCI_EXP_LNKSTA_CLS_2_5GB
|
|
#define PCI_EXP_LNKSTA_CLS_2_5GB 0x0001
|
|
#endif
|
|
|
|
#ifndef PCI_EXP_LNKSTA_CLS_5_0GB
|
|
#define PCI_EXP_LNKSTA_CLS_5_0GB 0x0002
|
|
#endif
|
|
|
|
#ifndef PCI_EXP_LNKSTA_CLS_8_0GB
|
|
#define PCI_EXP_LNKSTA_CLS_8_0GB 0x0003
|
|
#endif
|
|
|
|
#ifndef PCI_EXP_LNKSTA_NLW_X1
|
|
#define PCI_EXP_LNKSTA_NLW_X1 0x0010
|
|
#endif
|
|
|
|
#ifndef PCI_EXP_LNKSTA_NLW_X2
|
|
#define PCI_EXP_LNKSTA_NLW_X2 0x0020
|
|
#endif
|
|
|
|
#ifndef PCI_EXP_LNKSTA_NLW_X4
|
|
#define PCI_EXP_LNKSTA_NLW_X4 0x0040
|
|
#endif
|
|
|
|
#ifndef PCI_EXP_LNKSTA_NLW_X8
|
|
#define PCI_EXP_LNKSTA_NLW_X8 0x0080
|
|
#endif
|
|
|
|
#ifndef __GFP_COLD
|
|
#define __GFP_COLD 0
|
|
#endif
|
|
|
|
#ifndef __GFP_COMP
|
|
#define __GFP_COMP 0
|
|
#endif
|
|
|
|
#ifndef IP_OFFSET
|
|
#define IP_OFFSET 0x1FFF /* "Fragment Offset" part */
|
|
#endif
|
|
|
|
/*****************************************************************************/
|
|
/* Installations with ethtool version without eeprom, adapter id, or statistics
|
|
* support */
|
|
|
|
#ifndef ETH_GSTRING_LEN
|
|
#define ETH_GSTRING_LEN 32
|
|
#endif
|
|
|
|
#ifndef ETHTOOL_GSTATS
|
|
#define ETHTOOL_GSTATS 0x1d
|
|
#undef ethtool_drvinfo
|
|
#define ethtool_drvinfo k_ethtool_drvinfo
|
|
struct k_ethtool_drvinfo {
|
|
u32 cmd;
|
|
char driver[32];
|
|
char version[32];
|
|
char fw_version[32];
|
|
char bus_info[32];
|
|
char reserved1[32];
|
|
char reserved2[16];
|
|
u32 n_stats;
|
|
u32 testinfo_len;
|
|
u32 eedump_len;
|
|
u32 regdump_len;
|
|
};
|
|
|
|
struct ethtool_stats {
|
|
u32 cmd;
|
|
u32 n_stats;
|
|
u64 data[0];
|
|
};
|
|
#endif /* ETHTOOL_GSTATS */
|
|
|
|
#ifndef ETHTOOL_PHYS_ID
|
|
#define ETHTOOL_PHYS_ID 0x1c
|
|
#endif /* ETHTOOL_PHYS_ID */
|
|
|
|
#ifndef ETHTOOL_GSTRINGS
|
|
#define ETHTOOL_GSTRINGS 0x1b
|
|
enum ethtool_stringset {
|
|
ETH_SS_TEST = 0,
|
|
ETH_SS_STATS,
|
|
};
|
|
struct ethtool_gstrings {
|
|
u32 cmd; /* ETHTOOL_GSTRINGS */
|
|
u32 string_set; /* string set id e.c. ETH_SS_TEST, etc*/
|
|
u32 len; /* number of strings in the string set */
|
|
u8 data[0];
|
|
};
|
|
#endif /* ETHTOOL_GSTRINGS */
|
|
|
|
#ifndef ETHTOOL_TEST
|
|
#define ETHTOOL_TEST 0x1a
|
|
enum ethtool_test_flags {
|
|
ETH_TEST_FL_OFFLINE = BIT(0),
|
|
ETH_TEST_FL_FAILED = BIT(1),
|
|
};
|
|
struct ethtool_test {
|
|
u32 cmd;
|
|
u32 flags;
|
|
u32 reserved;
|
|
u32 len;
|
|
u64 data[0];
|
|
};
|
|
#endif /* ETHTOOL_TEST */
|
|
|
|
#ifndef ETHTOOL_GEEPROM
|
|
#define ETHTOOL_GEEPROM 0xb
|
|
#undef ETHTOOL_GREGS
|
|
struct ethtool_eeprom {
|
|
u32 cmd;
|
|
u32 magic;
|
|
u32 offset;
|
|
u32 len;
|
|
u8 data[0];
|
|
};
|
|
|
|
struct ethtool_value {
|
|
u32 cmd;
|
|
u32 data;
|
|
};
|
|
#endif /* ETHTOOL_GEEPROM */
|
|
|
|
#ifndef ETHTOOL_GLINK
|
|
#define ETHTOOL_GLINK 0xa
|
|
#endif /* ETHTOOL_GLINK */
|
|
|
|
#ifndef ETHTOOL_GWOL
|
|
#define ETHTOOL_GWOL 0x5
|
|
#define ETHTOOL_SWOL 0x6
|
|
#define SOPASS_MAX 6
|
|
struct ethtool_wolinfo {
|
|
u32 cmd;
|
|
u32 supported;
|
|
u32 wolopts;
|
|
u8 sopass[SOPASS_MAX]; /* SecureOn(tm) password */
|
|
};
|
|
#endif /* ETHTOOL_GWOL */
|
|
|
|
#ifndef ETHTOOL_GREGS
|
|
#define ETHTOOL_GREGS 0x00000004 /* Get NIC registers */
|
|
#define ethtool_regs _kc_ethtool_regs
|
|
/* for passing big chunks of data */
|
|
struct _kc_ethtool_regs {
|
|
u32 cmd;
|
|
u32 version; /* driver-specific, indicates different chips/revs */
|
|
u32 len; /* bytes */
|
|
u8 data[0];
|
|
};
|
|
#endif /* ETHTOOL_GREGS */
|
|
|
|
#ifndef ETHTOOL_GMSGLVL
|
|
#define ETHTOOL_GMSGLVL 0x00000007 /* Get driver message level */
|
|
#endif
|
|
#ifndef ETHTOOL_SMSGLVL
|
|
#define ETHTOOL_SMSGLVL 0x00000008 /* Set driver msg level, priv. */
|
|
#endif
|
|
#ifndef ETHTOOL_NWAY_RST
|
|
#define ETHTOOL_NWAY_RST 0x00000009 /* Restart autonegotiation, priv */
|
|
#endif
|
|
#ifndef ETHTOOL_GLINK
|
|
#define ETHTOOL_GLINK 0x0000000a /* Get link status */
|
|
#endif
|
|
#ifndef ETHTOOL_GEEPROM
|
|
#define ETHTOOL_GEEPROM 0x0000000b /* Get EEPROM data */
|
|
#endif
|
|
#ifndef ETHTOOL_SEEPROM
|
|
#define ETHTOOL_SEEPROM 0x0000000c /* Set EEPROM data */
|
|
#endif
|
|
#ifndef ETHTOOL_GCOALESCE
|
|
#define ETHTOOL_GCOALESCE 0x0000000e /* Get coalesce config */
|
|
/* for configuring coalescing parameters of chip */
|
|
#define ethtool_coalesce _kc_ethtool_coalesce
|
|
struct _kc_ethtool_coalesce {
|
|
u32 cmd; /* ETHTOOL_{G,S}COALESCE */
|
|
|
|
/* How many usecs to delay an RX interrupt after
|
|
* a packet arrives. If 0, only rx_max_coalesced_frames
|
|
* is used.
|
|
*/
|
|
u32 rx_coalesce_usecs;
|
|
|
|
/* How many packets to delay an RX interrupt after
|
|
* a packet arrives. If 0, only rx_coalesce_usecs is
|
|
* used. It is illegal to set both usecs and max frames
|
|
* to zero as this would cause RX interrupts to never be
|
|
* generated.
|
|
*/
|
|
u32 rx_max_coalesced_frames;
|
|
|
|
/* Same as above two parameters, except that these values
|
|
* apply while an IRQ is being serviced by the host. Not
|
|
* all cards support this feature and the values are ignored
|
|
* in that case.
|
|
*/
|
|
u32 rx_coalesce_usecs_irq;
|
|
u32 rx_max_coalesced_frames_irq;
|
|
|
|
/* How many usecs to delay a TX interrupt after
|
|
* a packet is sent. If 0, only tx_max_coalesced_frames
|
|
* is used.
|
|
*/
|
|
u32 tx_coalesce_usecs;
|
|
|
|
/* How many packets to delay a TX interrupt after
|
|
* a packet is sent. If 0, only tx_coalesce_usecs is
|
|
* used. It is illegal to set both usecs and max frames
|
|
* to zero as this would cause TX interrupts to never be
|
|
* generated.
|
|
*/
|
|
u32 tx_max_coalesced_frames;
|
|
|
|
/* Same as above two parameters, except that these values
|
|
* apply while an IRQ is being serviced by the host. Not
|
|
* all cards support this feature and the values are ignored
|
|
* in that case.
|
|
*/
|
|
u32 tx_coalesce_usecs_irq;
|
|
u32 tx_max_coalesced_frames_irq;
|
|
|
|
/* How many usecs to delay in-memory statistics
|
|
* block updates. Some drivers do not have an in-memory
|
|
* statistic block, and in such cases this value is ignored.
|
|
* This value must not be zero.
|
|
*/
|
|
u32 stats_block_coalesce_usecs;
|
|
|
|
/* Adaptive RX/TX coalescing is an algorithm implemented by
|
|
* some drivers to improve latency under low packet rates and
|
|
* improve throughput under high packet rates. Some drivers
|
|
* only implement one of RX or TX adaptive coalescing. Anything
|
|
* not implemented by the driver causes these values to be
|
|
* silently ignored.
|
|
*/
|
|
u32 use_adaptive_rx_coalesce;
|
|
u32 use_adaptive_tx_coalesce;
|
|
|
|
/* When the packet rate (measured in packets per second)
|
|
* is below pkt_rate_low, the {rx,tx}_*_low parameters are
|
|
* used.
|
|
*/
|
|
u32 pkt_rate_low;
|
|
u32 rx_coalesce_usecs_low;
|
|
u32 rx_max_coalesced_frames_low;
|
|
u32 tx_coalesce_usecs_low;
|
|
u32 tx_max_coalesced_frames_low;
|
|
|
|
/* When the packet rate is below pkt_rate_high but above
|
|
* pkt_rate_low (both measured in packets per second) the
|
|
* normal {rx,tx}_* coalescing parameters are used.
|
|
*/
|
|
|
|
/* When the packet rate is (measured in packets per second)
|
|
* is above pkt_rate_high, the {rx,tx}_*_high parameters are
|
|
* used.
|
|
*/
|
|
u32 pkt_rate_high;
|
|
u32 rx_coalesce_usecs_high;
|
|
u32 rx_max_coalesced_frames_high;
|
|
u32 tx_coalesce_usecs_high;
|
|
u32 tx_max_coalesced_frames_high;
|
|
|
|
/* How often to do adaptive coalescing packet rate sampling,
|
|
* measured in seconds. Must not be zero.
|
|
*/
|
|
u32 rate_sample_interval;
|
|
};
|
|
#endif /* ETHTOOL_GCOALESCE */
|
|
|
|
#ifndef ETHTOOL_SCOALESCE
|
|
#define ETHTOOL_SCOALESCE 0x0000000f /* Set coalesce config. */
|
|
#endif
|
|
#ifndef ETHTOOL_GRINGPARAM
|
|
#define ETHTOOL_GRINGPARAM 0x00000010 /* Get ring parameters */
|
|
/* for configuring RX/TX ring parameters */
|
|
#define ethtool_ringparam _kc_ethtool_ringparam
|
|
struct _kc_ethtool_ringparam {
|
|
u32 cmd; /* ETHTOOL_{G,S}RINGPARAM */
|
|
|
|
/* Read only attributes. These indicate the maximum number
|
|
* of pending RX/TX ring entries the driver will allow the
|
|
* user to set.
|
|
*/
|
|
u32 rx_max_pending;
|
|
u32 rx_mini_max_pending;
|
|
u32 rx_jumbo_max_pending;
|
|
u32 tx_max_pending;
|
|
|
|
/* Values changeable by the user. The valid values are
|
|
* in the range 1 to the "*_max_pending" counterpart above.
|
|
*/
|
|
u32 rx_pending;
|
|
u32 rx_mini_pending;
|
|
u32 rx_jumbo_pending;
|
|
u32 tx_pending;
|
|
};
|
|
#endif /* ETHTOOL_GRINGPARAM */
|
|
|
|
#ifndef ETHTOOL_SRINGPARAM
|
|
#define ETHTOOL_SRINGPARAM 0x00000011 /* Set ring parameters, priv. */
|
|
#endif
|
|
#ifndef ETHTOOL_GPAUSEPARAM
|
|
#define ETHTOOL_GPAUSEPARAM 0x00000012 /* Get pause parameters */
|
|
/* for configuring link flow control parameters */
|
|
#define ethtool_pauseparam _kc_ethtool_pauseparam
|
|
struct _kc_ethtool_pauseparam {
|
|
u32 cmd; /* ETHTOOL_{G,S}PAUSEPARAM */
|
|
|
|
/* If the link is being auto-negotiated (via ethtool_cmd.autoneg
|
|
* being true) the user may set 'autoneg' here non-zero to have the
|
|
* pause parameters be auto-negotiated too. In such a case, the
|
|
* {rx,tx}_pause values below determine what capabilities are
|
|
* advertised.
|
|
*
|
|
* If 'autoneg' is zero or the link is not being auto-negotiated,
|
|
* then {rx,tx}_pause force the driver to use/not-use pause
|
|
* flow control.
|
|
*/
|
|
u32 autoneg;
|
|
u32 rx_pause;
|
|
u32 tx_pause;
|
|
};
|
|
#endif /* ETHTOOL_GPAUSEPARAM */
|
|
|
|
#ifndef ETHTOOL_SPAUSEPARAM
|
|
#define ETHTOOL_SPAUSEPARAM 0x00000013 /* Set pause parameters. */
|
|
#endif
|
|
#ifndef ETHTOOL_GRXCSUM
|
|
#define ETHTOOL_GRXCSUM 0x00000014 /* Get RX hw csum enable (ethtool_value) */
|
|
#endif
|
|
#ifndef ETHTOOL_SRXCSUM
|
|
#define ETHTOOL_SRXCSUM 0x00000015 /* Set RX hw csum enable (ethtool_value) */
|
|
#endif
|
|
#ifndef ETHTOOL_GTXCSUM
|
|
#define ETHTOOL_GTXCSUM 0x00000016 /* Get TX hw csum enable (ethtool_value) */
|
|
#endif
|
|
#ifndef ETHTOOL_STXCSUM
|
|
#define ETHTOOL_STXCSUM 0x00000017 /* Set TX hw csum enable (ethtool_value) */
|
|
#endif
|
|
#ifndef ETHTOOL_GSG
|
|
#define ETHTOOL_GSG 0x00000018 /* Get scatter-gather enable
|
|
* (ethtool_value) */
|
|
#endif
|
|
#ifndef ETHTOOL_SSG
|
|
#define ETHTOOL_SSG 0x00000019 /* Set scatter-gather enable
|
|
* (ethtool_value). */
|
|
#endif
|
|
#ifndef ETHTOOL_TEST
|
|
#define ETHTOOL_TEST 0x0000001a /* execute NIC self-test, priv. */
|
|
#endif
|
|
#ifndef ETHTOOL_GSTRINGS
|
|
#define ETHTOOL_GSTRINGS 0x0000001b /* get specified string set */
|
|
#endif
|
|
#ifndef ETHTOOL_PHYS_ID
|
|
#define ETHTOOL_PHYS_ID 0x0000001c /* identify the NIC */
|
|
#endif
|
|
#ifndef ETHTOOL_GSTATS
|
|
#define ETHTOOL_GSTATS 0x0000001d /* get NIC-specific statistics */
|
|
#endif
|
|
#ifndef ETHTOOL_GTSO
|
|
#define ETHTOOL_GTSO 0x0000001e /* Get TSO enable (ethtool_value) */
|
|
#endif
|
|
#ifndef ETHTOOL_STSO
|
|
#define ETHTOOL_STSO 0x0000001f /* Set TSO enable (ethtool_value) */
|
|
#endif
|
|
|
|
#ifndef ETHTOOL_BUSINFO_LEN
|
|
#define ETHTOOL_BUSINFO_LEN 32
|
|
#endif
|
|
|
|
#ifndef WAKE_FILTER
|
|
#define WAKE_FILTER BIT(7)
|
|
#endif
|
|
|
|
#ifndef SPEED_2500
|
|
#define SPEED_2500 2500
|
|
#endif
|
|
#ifndef SPEED_5000
|
|
#define SPEED_5000 5000
|
|
#endif
|
|
#ifndef SPEED_14000
|
|
#define SPEED_14000 14000
|
|
#endif
|
|
#ifndef SPEED_25000
|
|
#define SPEED_25000 25000
|
|
#endif
|
|
#ifndef SPEED_50000
|
|
#define SPEED_50000 50000
|
|
#endif
|
|
#ifndef SPEED_56000
|
|
#define SPEED_56000 56000
|
|
#endif
|
|
#ifndef SPEED_100000
|
|
#define SPEED_100000 100000
|
|
#endif
|
|
|
|
#ifndef RHEL_RELEASE_VERSION
|
|
#define RHEL_RELEASE_VERSION(a,b) (((a) << 8) + (b))
|
|
#endif
|
|
#ifndef AX_RELEASE_VERSION
|
|
#define AX_RELEASE_VERSION(a,b) (((a) << 8) + (b))
|
|
#endif
|
|
|
|
#ifndef AX_RELEASE_CODE
|
|
#define AX_RELEASE_CODE 0
|
|
#endif
|
|
|
|
#if (AX_RELEASE_CODE && AX_RELEASE_CODE == AX_RELEASE_VERSION(3,0))
|
|
#define RHEL_RELEASE_CODE RHEL_RELEASE_VERSION(5,0)
|
|
#elif (AX_RELEASE_CODE && AX_RELEASE_CODE == AX_RELEASE_VERSION(3,1))
|
|
#define RHEL_RELEASE_CODE RHEL_RELEASE_VERSION(5,1)
|
|
#elif (AX_RELEASE_CODE && AX_RELEASE_CODE == AX_RELEASE_VERSION(3,2))
|
|
#define RHEL_RELEASE_CODE RHEL_RELEASE_VERSION(5,3)
|
|
#endif
|
|
|
|
#ifndef RHEL_RELEASE_CODE
|
|
/* NOTE: RHEL_RELEASE_* introduced in RHEL4.5 */
|
|
#define RHEL_RELEASE_CODE 0
|
|
#endif
|
|
|
|
/* RHEL 7 didn't backport the parameter change in
|
|
* create_singlethread_workqueue.
|
|
* If/when RH corrects this we will want to tighten up the version check.
|
|
*/
|
|
#if (RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,0))
|
|
#undef create_singlethread_workqueue
|
|
#define create_singlethread_workqueue(name) \
|
|
alloc_ordered_workqueue("%s", WQ_MEM_RECLAIM, name)
|
|
#endif
|
|
|
|
/* Ubuntu Release ABI is the 4th digit of their kernel version. You can find
|
|
* it in /usr/src/linux/$(uname -r)/include/generated/utsrelease.h for new
|
|
* enough versions of Ubuntu. Otherwise you can simply see it in the output of
|
|
* uname as the 4th digit of the kernel. The UTS_UBUNTU_RELEASE_ABI is not in
|
|
* the linux-source package, but in the linux-headers package. It begins to
|
|
* appear in later releases of 14.04 and 14.10.
|
|
*
|
|
* Ex:
|
|
* <Ubuntu 14.04.1>
|
|
* $uname -r
|
|
* 3.13.0-45-generic
|
|
* ABI is 45
|
|
*
|
|
* <Ubuntu 14.10>
|
|
* $uname -r
|
|
* 3.16.0-23-generic
|
|
* ABI is 23
|
|
*/
|
|
#ifndef UTS_UBUNTU_RELEASE_ABI
|
|
#define UTS_UBUNTU_RELEASE_ABI 0
|
|
#define UBUNTU_VERSION_CODE 0
|
|
#else
|
|
/* Ubuntu does not provide actual release version macro, so we use the kernel
|
|
* version plus the ABI to generate a unique version code specific to Ubuntu.
|
|
* In addition, we mask the lower 8 bits of LINUX_VERSION_CODE in order to
|
|
* ignore differences in sublevel which are not important since we have the
|
|
* ABI value. Otherwise, it becomes impossible to correlate ABI to version for
|
|
* ordering checks.
|
|
*
|
|
* This also lets us store an ABI value up to 65535, since it can take the
|
|
* space that would use the lower byte of the Linux version code.
|
|
*/
|
|
#define UBUNTU_VERSION_CODE (((~0xFF & LINUX_VERSION_CODE) << 8) + \
|
|
UTS_UBUNTU_RELEASE_ABI)
|
|
|
|
#if UTS_UBUNTU_RELEASE_ABI > 65535
|
|
#error UTS_UBUNTU_RELEASE_ABI is larger than 65535...
|
|
#endif /* UTS_UBUNTU_RELEASE_ABI > 65535 */
|
|
|
|
#if ( LINUX_VERSION_CODE <= KERNEL_VERSION(3,0,0) )
|
|
/* Our version code scheme does not make sense for non 3.x or newer kernels,
|
|
* and we have no support in kcompat for this scenario. Thus, treat this as a
|
|
* non-Ubuntu kernel. Possibly might be better to error here.
|
|
*/
|
|
#define UTS_UBUNTU_RELEASE_ABI 0
|
|
#define UBUNTU_VERSION_CODE 0
|
|
#endif /* <= 3.0.0 */
|
|
#endif /* !UTS_UBUNTU_RELEASE_ABI */
|
|
|
|
/* We ignore the 3rd digit since we want to give precedence to the additional
|
|
* ABI value provided by Ubuntu.
|
|
*/
|
|
#define UBUNTU_VERSION(a,b,c,d) (((a) << 24) + ((b) << 16) + (d))
|
|
|
|
/* SLE_VERSION is used to generate a 3-digit encoding that can order SLE
|
|
* kernels based on their major release, service pack, and a possible
|
|
* maintenance release.
|
|
*/
|
|
#define SLE_VERSION(a,b,c) (((a) << 16) + ((b) << 8) + (c))
|
|
|
|
/* The SLE_LOCALVERSION_CODE comes from a 3-digit code added as part of the
|
|
* Linux kernel version. It is extracted by the driver Makefile. This macro is
|
|
* used to generate codes for making comparisons below.
|
|
*/
|
|
#define SLE_LOCALVERSION(a,b,c) (((a) << 16) + ((b) << 8) + (c))
|
|
|
|
#ifdef CONFIG_SUSE_KERNEL
|
|
/* Starting since at least SLE 12sp4 and SLE 15, the SUSE kernels have
|
|
* provided CONFIG_SUSE_VERSION, CONFIG_SUSE_PATCHLEVEL and
|
|
* CONFIG_SUSE_AUXRELEASE. Use these to generate SLE_VERSION if available.
|
|
* Only fall back to the manual table otherwise. We expect all future versions
|
|
* of SLE kernels to include these values, so the table will remain only for
|
|
* the older releases.
|
|
*/
|
|
#ifdef CONFIG_SUSE_VERSION
|
|
#ifndef CONFIG_SUSE_PATCHLEVEL
|
|
#error "CONFIG_SUSE_VERSION exists but CONFIG_SUSE_PATCHLEVEL is missing"
|
|
#endif
|
|
#ifndef CONFIG_SUSE_AUXRELEASE
|
|
#error "CONFIG_SUSE_VERSION exists but CONFIG_SUSE_AUXRELEASE is missing"
|
|
#endif
|
|
#define SLE_VERSION_CODE SLE_VERSION(CONFIG_SUSE_VERSION, CONFIG_SUSE_PATCHLEVEL, CONFIG_SUSE_AUXRELEASE)
|
|
#else
|
|
/* If we do not have the CONFIG_SUSE_VERSION configuration values, fall back
|
|
* to the following table for older releases.
|
|
*/
|
|
#if ( LINUX_VERSION_CODE == KERNEL_VERSION(2,6,27) )
|
|
/* SLES11 GA is 2.6.27 based */
|
|
#define SLE_VERSION_CODE SLE_VERSION(11,0,0)
|
|
#elif ( LINUX_VERSION_CODE == KERNEL_VERSION(2,6,32) )
|
|
/* SLES11 SP1 is 2.6.32 based */
|
|
#define SLE_VERSION_CODE SLE_VERSION(11,1,0)
|
|
#elif ( LINUX_VERSION_CODE == KERNEL_VERSION(3,0,13) )
|
|
/* SLES11 SP2 GA is 3.0.13-0.27 */
|
|
#define SLE_VERSION_CODE SLE_VERSION(11,2,0)
|
|
#elif ((LINUX_VERSION_CODE == KERNEL_VERSION(3,0,76)))
|
|
/* SLES11 SP3 GA is 3.0.76-0.11 */
|
|
#define SLE_VERSION_CODE SLE_VERSION(11,3,0)
|
|
#elif (LINUX_VERSION_CODE == KERNEL_VERSION(3,0,101))
|
|
#if (SLE_LOCALVERSION_CODE < SLE_LOCALVERSION(0,8,0))
|
|
/* some SLES11sp2 update kernels up to 3.0.101-0.7.x */
|
|
#define SLE_VERSION_CODE SLE_VERSION(11,2,0)
|
|
#elif (SLE_LOCALVERSION_CODE < SLE_LOCALVERSION(63,0,0))
|
|
/* most SLES11sp3 update kernels */
|
|
#define SLE_VERSION_CODE SLE_VERSION(11,3,0)
|
|
#else
|
|
/* SLES11 SP4 GA (3.0.101-63) and update kernels 3.0.101-63+ */
|
|
#define SLE_VERSION_CODE SLE_VERSION(11,4,0)
|
|
#endif
|
|
#elif (LINUX_VERSION_CODE == KERNEL_VERSION(3,12,28))
|
|
/* SLES12 GA is 3.12.28-4
|
|
* kernel updates 3.12.xx-<33 through 52>[.yy] */
|
|
#define SLE_VERSION_CODE SLE_VERSION(12,0,0)
|
|
#elif (LINUX_VERSION_CODE == KERNEL_VERSION(3,12,49))
|
|
/* SLES12 SP1 GA is 3.12.49-11
|
|
* updates 3.12.xx-60.yy where xx={51..} */
|
|
#define SLE_VERSION_CODE SLE_VERSION(12,1,0)
|
|
#elif ((LINUX_VERSION_CODE >= KERNEL_VERSION(4,4,21) && \
|
|
(LINUX_VERSION_CODE <= KERNEL_VERSION(4,4,59))) || \
|
|
(LINUX_VERSION_CODE >= KERNEL_VERSION(4,4,74) && \
|
|
LINUX_VERSION_CODE < KERNEL_VERSION(4,5,0) && \
|
|
SLE_LOCALVERSION_CODE >= KERNEL_VERSION(92,0,0) && \
|
|
SLE_LOCALVERSION_CODE < KERNEL_VERSION(93,0,0)))
|
|
/* SLES12 SP2 GA is 4.4.21-69.
|
|
* SLES12 SP2 updates before SLES12 SP3 are: 4.4.{21,38,49,59}
|
|
* SLES12 SP2 updates after SLES12 SP3 are: 4.4.{74,90,103,114,120}
|
|
* but they all use a SLE_LOCALVERSION_CODE matching 92.nn.y */
|
|
#define SLE_VERSION_CODE SLE_VERSION(12,2,0)
|
|
#elif ((LINUX_VERSION_CODE == KERNEL_VERSION(4,4,73) || \
|
|
LINUX_VERSION_CODE == KERNEL_VERSION(4,4,82) || \
|
|
LINUX_VERSION_CODE == KERNEL_VERSION(4,4,92)) || \
|
|
(LINUX_VERSION_CODE == KERNEL_VERSION(4,4,103) && \
|
|
(SLE_LOCALVERSION_CODE == KERNEL_VERSION(6,33,0) || \
|
|
SLE_LOCALVERSION_CODE == KERNEL_VERSION(6,38,0))) || \
|
|
(LINUX_VERSION_CODE >= KERNEL_VERSION(4,4,114) && \
|
|
LINUX_VERSION_CODE < KERNEL_VERSION(4,5,0) && \
|
|
SLE_LOCALVERSION_CODE >= KERNEL_VERSION(94,0,0) && \
|
|
SLE_LOCALVERSION_CODE < KERNEL_VERSION(95,0,0)) )
|
|
/* SLES12 SP3 GM is 4.4.73-5 and update kernels are 4.4.82-6.3.
|
|
* SLES12 SP3 updates not conflicting with SP2 are: 4.4.{82,92}
|
|
* SLES12 SP3 updates conflicting with SP2 are:
|
|
* - 4.4.103-6.33.1, 4.4.103-6.38.1
|
|
* - 4.4.{114,120}-94.nn.y */
|
|
#define SLE_VERSION_CODE SLE_VERSION(12,3,0)
|
|
#else
|
|
#error "This looks like a SUSE kernel, but it has an unrecognized local version code."
|
|
#endif /* LINUX_VERSION_CODE == KERNEL_VERSION(x,y,z) */
|
|
#endif /* !CONFIG_SUSE_VERSION */
|
|
#endif /* CONFIG_SUSE_KERNEL */
|
|
#ifndef SLE_VERSION_CODE
|
|
#define SLE_VERSION_CODE 0
|
|
#endif /* SLE_VERSION_CODE */
|
|
#ifndef SLE_LOCALVERSION_CODE
|
|
#define SLE_LOCALVERSION_CODE 0
|
|
#endif /* SLE_LOCALVERSION_CODE */
|
|
|
|
/*
|
|
* Include the definitions file for HAVE/NEED flags for the standard upstream
|
|
* kernels.
|
|
*
|
|
* Then, based on the distribution we detect, load the distribution specific
|
|
* definitions file that customizes the definitions for the target
|
|
* distribution.
|
|
*/
|
|
#include "kcompat_std_defs.h"
|
|
|
|
#ifdef CONFIG_SUSE_KERNEL
|
|
#include "kcompat_sles_defs.h"
|
|
#elif UBUNTU_VERSION_CODE
|
|
#include "kcompat_ubuntu_defs.h"
|
|
#elif RHEL_RELEASE_CODE
|
|
#include "kcompat_rhel_defs.h"
|
|
#endif
|
|
|
|
/*
|
|
* ADQ depends on __TC_MQPRIO_MODE_MAX and related kernel code
|
|
* added around 4.15. Some distributions (e.g. Oracle Linux 7.7)
|
|
* have done a partial back-port of that to their kernels based
|
|
* on older mainline kernels that did not include all the necessary
|
|
* kernel enablement to support ADQ.
|
|
* Undefine __TC_MQPRIO_MODE_MAX for all OSV distributions with
|
|
* kernels based on mainline kernels older than 4.15 except for
|
|
* RHEL, SLES and Ubuntu which are known to have good back-ports.
|
|
*/
|
|
#if (!RHEL_RELEASE_CODE && !SLE_VERSION_CODE && !UBUNTU_VERSION_CODE)
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,15,0))
|
|
#undef __TC_MQPRIO_MODE_MAX
|
|
#endif /* LINUX_VERSION_CODE == KERNEL_VERSION(4,15,0) */
|
|
#endif /* if (NOT RHEL && NOT SLES && NOT UBUNTU) */
|
|
|
|
#ifdef __KLOCWORK__
|
|
/* The following are not compiled into the binary driver; they are here
|
|
* only to tune Klocwork scans to workaround false-positive issues.
|
|
*/
|
|
#ifdef ARRAY_SIZE
|
|
#undef ARRAY_SIZE
|
|
#define ARRAY_SIZE(x) (sizeof(x) / sizeof((x)[0]))
|
|
#endif
|
|
|
|
#define memcpy(dest, src, len) memcpy_s(dest, len, src, len)
|
|
#define memset(dest, ch, len) memset_s(dest, len, ch, len)
|
|
|
|
static inline int _kc_test_and_clear_bit(int nr, volatile unsigned long *addr)
|
|
{
|
|
unsigned long mask = BIT_MASK(nr);
|
|
unsigned long *p = ((unsigned long *)addr) + BIT_WORD(nr);
|
|
unsigned long old;
|
|
unsigned long flags = 0;
|
|
|
|
_atomic_spin_lock_irqsave(p, flags);
|
|
old = *p;
|
|
*p = old & ~mask;
|
|
_atomic_spin_unlock_irqrestore(p, flags);
|
|
|
|
return (old & mask) != 0;
|
|
}
|
|
#define test_and_clear_bit(nr, addr) _kc_test_and_clear_bit(nr, addr)
|
|
|
|
static inline int _kc_test_and_set_bit(int nr, volatile unsigned long *addr)
|
|
{
|
|
unsigned long mask = BIT_MASK(nr);
|
|
unsigned long *p = ((unsigned long *)addr) + BIT_WORD(nr);
|
|
unsigned long old;
|
|
unsigned long flags = 0;
|
|
|
|
_atomic_spin_lock_irqsave(p, flags);
|
|
old = *p;
|
|
*p = old | mask;
|
|
_atomic_spin_unlock_irqrestore(p, flags);
|
|
|
|
return (old & mask) != 0;
|
|
}
|
|
#define test_and_set_bit(nr, addr) _kc_test_and_set_bit(nr, addr)
|
|
|
|
#ifdef CONFIG_DYNAMIC_DEBUG
|
|
#undef dev_dbg
|
|
#define dev_dbg(dev, format, arg...) dev_printk(KERN_DEBUG, dev, format, ##arg)
|
|
#undef pr_debug
|
|
#define pr_debug(format, arg...) printk(KERN_DEBUG format, ##arg)
|
|
#endif /* CONFIG_DYNAMIC_DEBUG */
|
|
|
|
#undef hlist_for_each_entry_safe
|
|
#define hlist_for_each_entry_safe(pos, n, head, member) \
|
|
for (n = NULL, pos = hlist_entry_safe((head)->first, typeof(*(pos)), \
|
|
member); \
|
|
pos; \
|
|
pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
|
|
|
|
#ifdef uninitialized_var
|
|
#undef uninitialized_var
|
|
#define uninitialized_var(x) x = *(&(x))
|
|
#endif
|
|
|
|
#ifdef WRITE_ONCE
|
|
#undef WRITE_ONCE
|
|
#define WRITE_ONCE(x, val) ((x) = (val))
|
|
#endif /* WRITE_ONCE */
|
|
|
|
#ifdef wait_event_interruptible_timeout
|
|
#undef wait_event_interruptible_timeout
|
|
#define wait_event_interruptible_timeout(wq_head, condition, timeout) ({ \
|
|
long ret; \
|
|
if ((condition)) \
|
|
ret = timeout; \
|
|
else \
|
|
ret = 0; \
|
|
ret; \
|
|
})
|
|
#endif /* wait_event_interruptible_timeout */
|
|
|
|
#ifdef max_t
|
|
#undef max_t
|
|
#define max_t(type, x, y) ({ \
|
|
type __x = (x); \
|
|
type __y = (y); \
|
|
__x > __y ? __x : __y; \
|
|
})
|
|
#endif /* max_t */
|
|
|
|
#ifdef min_t
|
|
#undef min_t
|
|
#define min_t(type, x, y) ({ \
|
|
type __x = (x); \
|
|
type __y = (y); \
|
|
__x < __y ? __x : __y; \
|
|
})
|
|
#endif /* min_t */
|
|
#endif /* __KLOCWORK__ */
|
|
|
|
/* Older versions of GCC will trigger -Wformat-nonliteral warnings for const
|
|
* char * strings. Unfortunately, the implementation of do_trace_printk does
|
|
* this, in order to add a storage attribute to the memory. This was fixed in
|
|
* GCC 5.1, but we still use older distributions built with GCC 4.x.
|
|
*
|
|
* The string pointer is only passed as a const char * to the __trace_bprintk
|
|
* function. Since that function has the __printf attribute, it will trigger
|
|
* the warnings. We can't remove the attribute, so instead we'll use the
|
|
* __diag macro to disable -Wformat-nonliteral around the call to
|
|
* __trace_bprintk.
|
|
*/
|
|
#if GCC_VERSION < 50100
|
|
#define __trace_bprintk(ip, fmt, args...) ({ \
|
|
int err; \
|
|
__diag_push(); \
|
|
__diag(ignored "-Wformat-nonliteral"); \
|
|
err = __trace_bprintk(ip, fmt, ##args); \
|
|
__diag_pop(); \
|
|
err; \
|
|
})
|
|
#endif /* GCC_VERSION < 5.1.0 */
|
|
|
|
/* Newer kernels removed <linux/pci-aspm.h> */
|
|
#if ((LINUX_VERSION_CODE < KERNEL_VERSION(5,4,0)) && \
|
|
(!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,3)) && \
|
|
!(SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(15,3,0)))))
|
|
#define HAVE_PCI_ASPM_H
|
|
#endif
|
|
|
|
/*****************************************************************************/
|
|
/* 2.4.3 => 2.4.0 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,3) )
|
|
|
|
/**************************************/
|
|
/* PCI DRIVER API */
|
|
|
|
#ifndef pci_set_dma_mask
|
|
#define pci_set_dma_mask _kc_pci_set_dma_mask
|
|
int _kc_pci_set_dma_mask(struct pci_dev *dev, dma_addr_t mask);
|
|
#endif
|
|
|
|
#ifndef pci_request_regions
|
|
#define pci_request_regions _kc_pci_request_regions
|
|
int _kc_pci_request_regions(struct pci_dev *pdev, char *res_name);
|
|
#endif
|
|
|
|
#ifndef pci_release_regions
|
|
#define pci_release_regions _kc_pci_release_regions
|
|
void _kc_pci_release_regions(struct pci_dev *pdev);
|
|
#endif
|
|
|
|
/**************************************/
|
|
/* NETWORK DRIVER API */
|
|
|
|
#ifndef alloc_etherdev
|
|
#define alloc_etherdev _kc_alloc_etherdev
|
|
struct net_device * _kc_alloc_etherdev(int sizeof_priv);
|
|
#endif
|
|
|
|
#ifndef is_valid_ether_addr
|
|
#define is_valid_ether_addr _kc_is_valid_ether_addr
|
|
int _kc_is_valid_ether_addr(u8 *addr);
|
|
#endif
|
|
|
|
/**************************************/
|
|
/* MISCELLANEOUS */
|
|
|
|
#ifndef INIT_TQUEUE
|
|
#define INIT_TQUEUE(_tq, _routine, _data) \
|
|
do { \
|
|
INIT_LIST_HEAD(&(_tq)->list); \
|
|
(_tq)->sync = 0; \
|
|
(_tq)->routine = _routine; \
|
|
(_tq)->data = _data; \
|
|
} while (0)
|
|
#endif
|
|
|
|
#endif /* 2.4.3 => 2.4.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,5) )
|
|
/* Generic MII registers. */
|
|
#define MII_BMCR 0x00 /* Basic mode control register */
|
|
#define MII_BMSR 0x01 /* Basic mode status register */
|
|
#define MII_PHYSID1 0x02 /* PHYS ID 1 */
|
|
#define MII_PHYSID2 0x03 /* PHYS ID 2 */
|
|
#define MII_ADVERTISE 0x04 /* Advertisement control reg */
|
|
#define MII_LPA 0x05 /* Link partner ability reg */
|
|
#define MII_EXPANSION 0x06 /* Expansion register */
|
|
/* Basic mode control register. */
|
|
#define BMCR_FULLDPLX 0x0100 /* Full duplex */
|
|
#define BMCR_ANENABLE 0x1000 /* Enable auto negotiation */
|
|
/* Basic mode status register. */
|
|
#define BMSR_ERCAP 0x0001 /* Ext-reg capability */
|
|
#define BMSR_ANEGCAPABLE 0x0008 /* Able to do auto-negotiation */
|
|
#define BMSR_10HALF 0x0800 /* Can do 10mbps, half-duplex */
|
|
#define BMSR_10FULL 0x1000 /* Can do 10mbps, full-duplex */
|
|
#define BMSR_100HALF 0x2000 /* Can do 100mbps, half-duplex */
|
|
#define BMSR_100FULL 0x4000 /* Can do 100mbps, full-duplex */
|
|
/* Advertisement control register. */
|
|
#define ADVERTISE_CSMA 0x0001 /* Only selector supported */
|
|
#define ADVERTISE_10HALF 0x0020 /* Try for 10mbps half-duplex */
|
|
#define ADVERTISE_10FULL 0x0040 /* Try for 10mbps full-duplex */
|
|
#define ADVERTISE_100HALF 0x0080 /* Try for 100mbps half-duplex */
|
|
#define ADVERTISE_100FULL 0x0100 /* Try for 100mbps full-duplex */
|
|
#define ADVERTISE_ALL (ADVERTISE_10HALF | ADVERTISE_10FULL | \
|
|
ADVERTISE_100HALF | ADVERTISE_100FULL)
|
|
/* Expansion register for auto-negotiation. */
|
|
#define EXPANSION_ENABLENPAGE 0x0004 /* This enables npage words */
|
|
#endif
|
|
|
|
/*****************************************************************************/
|
|
/* 2.4.6 => 2.4.3 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,6) )
|
|
|
|
#ifndef pci_set_power_state
|
|
#define pci_set_power_state _kc_pci_set_power_state
|
|
int _kc_pci_set_power_state(struct pci_dev *dev, int state);
|
|
#endif
|
|
|
|
#ifndef pci_enable_wake
|
|
#define pci_enable_wake _kc_pci_enable_wake
|
|
int _kc_pci_enable_wake(struct pci_dev *pdev, u32 state, int enable);
|
|
#endif
|
|
|
|
#ifndef pci_disable_device
|
|
#define pci_disable_device _kc_pci_disable_device
|
|
void _kc_pci_disable_device(struct pci_dev *pdev);
|
|
#endif
|
|
|
|
/* PCI PM entry point syntax changed, so don't support suspend/resume */
|
|
#undef CONFIG_PM
|
|
|
|
#endif /* 2.4.6 => 2.4.3 */
|
|
|
|
#ifndef HAVE_PCI_SET_MWI
|
|
#define pci_set_mwi(X) pci_write_config_word(X, \
|
|
PCI_COMMAND, adapter->hw.bus.pci_cmd_word | \
|
|
PCI_COMMAND_INVALIDATE);
|
|
#define pci_clear_mwi(X) pci_write_config_word(X, \
|
|
PCI_COMMAND, adapter->hw.bus.pci_cmd_word & \
|
|
~PCI_COMMAND_INVALIDATE);
|
|
#endif
|
|
|
|
/*****************************************************************************/
|
|
/* 2.4.10 => 2.4.9 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,10) )
|
|
|
|
/**************************************/
|
|
/* MODULE API */
|
|
|
|
#ifndef MODULE_LICENSE
|
|
#define MODULE_LICENSE(X)
|
|
#endif
|
|
|
|
/**************************************/
|
|
/* OTHER */
|
|
|
|
#undef min
|
|
#define min(x,y) ({ \
|
|
const typeof(x) _x = (x); \
|
|
const typeof(y) _y = (y); \
|
|
(void) (&_x == &_y); \
|
|
_x < _y ? _x : _y; })
|
|
|
|
#undef max
|
|
#define max(x,y) ({ \
|
|
const typeof(x) _x = (x); \
|
|
const typeof(y) _y = (y); \
|
|
(void) (&_x == &_y); \
|
|
_x > _y ? _x : _y; })
|
|
|
|
#define min_t(type,x,y) ({ \
|
|
type _x = (x); \
|
|
type _y = (y); \
|
|
_x < _y ? _x : _y; })
|
|
|
|
#define max_t(type,x,y) ({ \
|
|
type _x = (x); \
|
|
type _y = (y); \
|
|
_x > _y ? _x : _y; })
|
|
|
|
#ifndef list_for_each_safe
|
|
#define list_for_each_safe(pos, n, head) \
|
|
for (pos = (head)->next, n = pos->next; pos != (head); \
|
|
pos = n, n = pos->next)
|
|
#endif
|
|
|
|
#ifndef ____cacheline_aligned_in_smp
|
|
#ifdef CONFIG_SMP
|
|
#define ____cacheline_aligned_in_smp ____cacheline_aligned
|
|
#else
|
|
#define ____cacheline_aligned_in_smp
|
|
#endif /* CONFIG_SMP */
|
|
#endif
|
|
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,8) )
|
|
int _kc_snprintf(char * buf, size_t size, const char *fmt, ...);
|
|
#define snprintf(buf, size, fmt, args...) _kc_snprintf(buf, size, fmt, ##args)
|
|
int _kc_vsnprintf(char *buf, size_t size, const char *fmt, va_list args);
|
|
#define vsnprintf(buf, size, fmt, args) _kc_vsnprintf(buf, size, fmt, args)
|
|
#else /* 2.4.8 => 2.4.9 */
|
|
int snprintf(char * buf, size_t size, const char *fmt, ...);
|
|
int vsnprintf(char *buf, size_t size, const char *fmt, va_list args);
|
|
#endif
|
|
#endif /* 2.4.10 -> 2.4.6 */
|
|
|
|
|
|
/*****************************************************************************/
|
|
/* 2.4.12 => 2.4.10 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,12) )
|
|
#ifndef HAVE_NETIF_MSG
|
|
#define HAVE_NETIF_MSG 1
|
|
enum {
|
|
NETIF_MSG_DRV = 0x0001,
|
|
NETIF_MSG_PROBE = 0x0002,
|
|
NETIF_MSG_LINK = 0x0004,
|
|
NETIF_MSG_TIMER = 0x0008,
|
|
NETIF_MSG_IFDOWN = 0x0010,
|
|
NETIF_MSG_IFUP = 0x0020,
|
|
NETIF_MSG_RX_ERR = 0x0040,
|
|
NETIF_MSG_TX_ERR = 0x0080,
|
|
NETIF_MSG_TX_QUEUED = 0x0100,
|
|
NETIF_MSG_INTR = 0x0200,
|
|
NETIF_MSG_TX_DONE = 0x0400,
|
|
NETIF_MSG_RX_STATUS = 0x0800,
|
|
NETIF_MSG_PKTDATA = 0x1000,
|
|
NETIF_MSG_HW = 0x2000,
|
|
NETIF_MSG_WOL = 0x4000,
|
|
};
|
|
|
|
#define netif_msg_drv(p) ((p)->msg_enable & NETIF_MSG_DRV)
|
|
#define netif_msg_probe(p) ((p)->msg_enable & NETIF_MSG_PROBE)
|
|
#define netif_msg_link(p) ((p)->msg_enable & NETIF_MSG_LINK)
|
|
#define netif_msg_timer(p) ((p)->msg_enable & NETIF_MSG_TIMER)
|
|
#define netif_msg_ifdown(p) ((p)->msg_enable & NETIF_MSG_IFDOWN)
|
|
#define netif_msg_ifup(p) ((p)->msg_enable & NETIF_MSG_IFUP)
|
|
#define netif_msg_rx_err(p) ((p)->msg_enable & NETIF_MSG_RX_ERR)
|
|
#define netif_msg_tx_err(p) ((p)->msg_enable & NETIF_MSG_TX_ERR)
|
|
#define netif_msg_tx_queued(p) ((p)->msg_enable & NETIF_MSG_TX_QUEUED)
|
|
#define netif_msg_intr(p) ((p)->msg_enable & NETIF_MSG_INTR)
|
|
#define netif_msg_tx_done(p) ((p)->msg_enable & NETIF_MSG_TX_DONE)
|
|
#define netif_msg_rx_status(p) ((p)->msg_enable & NETIF_MSG_RX_STATUS)
|
|
#define netif_msg_pktdata(p) ((p)->msg_enable & NETIF_MSG_PKTDATA)
|
|
#endif /* !HAVE_NETIF_MSG */
|
|
#endif /* 2.4.12 => 2.4.10 */
|
|
|
|
/*****************************************************************************/
|
|
/* 2.4.13 => 2.4.12 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,13) )
|
|
|
|
/**************************************/
|
|
/* PCI DMA MAPPING */
|
|
|
|
#ifndef virt_to_page
|
|
#define virt_to_page(v) (mem_map + (virt_to_phys(v) >> PAGE_SHIFT))
|
|
#endif
|
|
|
|
#ifndef pci_map_page
|
|
#define pci_map_page _kc_pci_map_page
|
|
u64 _kc_pci_map_page(struct pci_dev *dev, struct page *page, unsigned long offset, size_t size, int direction);
|
|
#endif
|
|
|
|
#ifndef pci_unmap_page
|
|
#define pci_unmap_page _kc_pci_unmap_page
|
|
void _kc_pci_unmap_page(struct pci_dev *dev, u64 dma_addr, size_t size, int direction);
|
|
#endif
|
|
|
|
/* pci_set_dma_mask takes dma_addr_t, which is only 32-bits prior to 2.4.13 */
|
|
|
|
#undef DMA_32BIT_MASK
|
|
#define DMA_32BIT_MASK 0xffffffff
|
|
#undef DMA_64BIT_MASK
|
|
#define DMA_64BIT_MASK 0xffffffff
|
|
|
|
/**************************************/
|
|
/* OTHER */
|
|
|
|
#ifndef cpu_relax
|
|
#define cpu_relax() rep_nop()
|
|
#endif
|
|
|
|
struct vlan_ethhdr {
|
|
unsigned char h_dest[ETH_ALEN];
|
|
unsigned char h_source[ETH_ALEN];
|
|
unsigned short h_vlan_proto;
|
|
unsigned short h_vlan_TCI;
|
|
unsigned short h_vlan_encapsulated_proto;
|
|
};
|
|
#endif /* 2.4.13 => 2.4.12 */
|
|
|
|
/*****************************************************************************/
|
|
/* 2.4.17 => 2.4.12 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,17) )
|
|
|
|
#ifndef __devexit_p
|
|
#define __devexit_p(x) &(x)
|
|
#endif
|
|
|
|
#endif /* 2.4.17 => 2.4.13 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,18) )
|
|
#define NETIF_MSG_HW 0x2000
|
|
#define NETIF_MSG_WOL 0x4000
|
|
|
|
#ifndef netif_msg_hw
|
|
#define netif_msg_hw(p) ((p)->msg_enable & NETIF_MSG_HW)
|
|
#endif
|
|
#ifndef netif_msg_wol
|
|
#define netif_msg_wol(p) ((p)->msg_enable & NETIF_MSG_WOL)
|
|
#endif
|
|
#endif /* 2.4.18 */
|
|
|
|
/*****************************************************************************/
|
|
|
|
/*****************************************************************************/
|
|
/* 2.4.20 => 2.4.19 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,20) )
|
|
|
|
/* we won't support NAPI on less than 2.4.20 */
|
|
#ifdef NAPI
|
|
#undef NAPI
|
|
#endif
|
|
|
|
#endif /* 2.4.20 => 2.4.19 */
|
|
|
|
/*****************************************************************************/
|
|
/* 2.4.22 => 2.4.17 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,22) )
|
|
#define pci_name(x) ((x)->slot_name)
|
|
|
|
#ifndef SUPPORTED_10000baseT_Full
|
|
#define SUPPORTED_10000baseT_Full BIT(12)
|
|
#endif
|
|
#ifndef ADVERTISED_10000baseT_Full
|
|
#define ADVERTISED_10000baseT_Full BIT(12)
|
|
#endif
|
|
#endif
|
|
|
|
/*****************************************************************************/
|
|
/*****************************************************************************/
|
|
/* 2.4.23 => 2.4.22 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,23) )
|
|
/*****************************************************************************/
|
|
#ifdef NAPI
|
|
#ifndef netif_poll_disable
|
|
#define netif_poll_disable(x) _kc_netif_poll_disable(x)
|
|
static inline void _kc_netif_poll_disable(struct net_device *netdev)
|
|
{
|
|
while (test_and_set_bit(__LINK_STATE_RX_SCHED, &netdev->state)) {
|
|
/* No hurry */
|
|
current->state = TASK_INTERRUPTIBLE;
|
|
schedule_timeout(1);
|
|
}
|
|
}
|
|
#endif
|
|
#ifndef netif_poll_enable
|
|
#define netif_poll_enable(x) _kc_netif_poll_enable(x)
|
|
static inline void _kc_netif_poll_enable(struct net_device *netdev)
|
|
{
|
|
clear_bit(__LINK_STATE_RX_SCHED, &netdev->state);
|
|
}
|
|
#endif
|
|
#endif /* NAPI */
|
|
#ifndef netif_tx_disable
|
|
#define netif_tx_disable(x) _kc_netif_tx_disable(x)
|
|
static inline void _kc_netif_tx_disable(struct net_device *dev)
|
|
{
|
|
spin_lock_bh(&dev->xmit_lock);
|
|
netif_stop_queue(dev);
|
|
spin_unlock_bh(&dev->xmit_lock);
|
|
}
|
|
#endif
|
|
#else /* 2.4.23 => 2.4.22 */
|
|
#define HAVE_SCTP
|
|
#endif /* 2.4.23 => 2.4.22 */
|
|
|
|
/*****************************************************************************/
|
|
/* 2.6.4 => 2.6.0 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,25) || \
|
|
( LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,0) && \
|
|
LINUX_VERSION_CODE < KERNEL_VERSION(2,6,4) ) )
|
|
#define ETHTOOL_OPS_COMPAT
|
|
#endif /* 2.6.4 => 2.6.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,27) )
|
|
#define __user
|
|
#endif /* < 2.4.27 */
|
|
|
|
/*****************************************************************************/
|
|
/* 2.5.71 => 2.4.x */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,5,71) )
|
|
#define sk_protocol protocol
|
|
#define pci_get_device pci_find_device
|
|
#endif /* 2.5.70 => 2.4.x */
|
|
|
|
/*****************************************************************************/
|
|
/* < 2.4.27 or 2.6.0 <= 2.6.5 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,27) || \
|
|
( LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,0) && \
|
|
LINUX_VERSION_CODE < KERNEL_VERSION(2,6,5) ) )
|
|
|
|
#ifndef netif_msg_init
|
|
#define netif_msg_init _kc_netif_msg_init
|
|
static inline u32 _kc_netif_msg_init(int debug_value, int default_msg_enable_bits)
|
|
{
|
|
/* use default */
|
|
if (debug_value < 0 || debug_value >= (sizeof(u32) * 8))
|
|
return default_msg_enable_bits;
|
|
if (debug_value == 0) /* no output */
|
|
return 0;
|
|
/* set low N bits */
|
|
return (1 << debug_value) -1;
|
|
}
|
|
#endif
|
|
|
|
#endif /* < 2.4.27 or 2.6.0 <= 2.6.5 */
|
|
/*****************************************************************************/
|
|
#if (( LINUX_VERSION_CODE < KERNEL_VERSION(2,4,27) ) || \
|
|
(( LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,0) ) && \
|
|
( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,3) )))
|
|
#define netdev_priv(x) x->priv
|
|
#endif
|
|
|
|
/*****************************************************************************/
|
|
/* <= 2.5.0 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,5,0) )
|
|
#include <linux/rtnetlink.h>
|
|
#undef pci_register_driver
|
|
#define pci_register_driver pci_module_init
|
|
|
|
/*
|
|
* Most of the dma compat code is copied/modifed from the 2.4.37
|
|
* /include/linux/libata-compat.h header file
|
|
*/
|
|
/* These definitions mirror those in pci.h, so they can be used
|
|
* interchangeably with their PCI_ counterparts */
|
|
enum dma_data_direction {
|
|
DMA_BIDIRECTIONAL = 0,
|
|
DMA_TO_DEVICE = 1,
|
|
DMA_FROM_DEVICE = 2,
|
|
DMA_NONE = 3,
|
|
};
|
|
|
|
struct device {
|
|
struct pci_dev pdev;
|
|
};
|
|
|
|
static inline struct pci_dev *to_pci_dev (struct device *dev)
|
|
{
|
|
return (struct pci_dev *) dev;
|
|
}
|
|
static inline struct device *pci_dev_to_dev(struct pci_dev *pdev)
|
|
{
|
|
return (struct device *) pdev;
|
|
}
|
|
#define pdev_printk(lvl, pdev, fmt, args...) \
|
|
printk("%s %s: " fmt, lvl, pci_name(pdev), ## args)
|
|
#define dev_err(dev, fmt, args...) \
|
|
pdev_printk(KERN_ERR, to_pci_dev(dev), fmt, ## args)
|
|
#define dev_info(dev, fmt, args...) \
|
|
pdev_printk(KERN_INFO, to_pci_dev(dev), fmt, ## args)
|
|
#define dev_warn(dev, fmt, args...) \
|
|
pdev_printk(KERN_WARNING, to_pci_dev(dev), fmt, ## args)
|
|
#define dev_notice(dev, fmt, args...) \
|
|
pdev_printk(KERN_NOTICE, to_pci_dev(dev), fmt, ## args)
|
|
#define dev_dbg(dev, fmt, args...) \
|
|
pdev_printk(KERN_DEBUG, to_pci_dev(dev), fmt, ## args)
|
|
|
|
/* NOTE: dangerous! we ignore the 'gfp' argument */
|
|
#define dma_alloc_coherent(dev,sz,dma,gfp) \
|
|
pci_alloc_consistent(to_pci_dev(dev),(sz),(dma))
|
|
#define dma_free_coherent(dev,sz,addr,dma_addr) \
|
|
pci_free_consistent(to_pci_dev(dev),(sz),(addr),(dma_addr))
|
|
|
|
#define dma_map_page(dev,a,b,c,d) \
|
|
pci_map_page(to_pci_dev(dev),(a),(b),(c),(d))
|
|
#define dma_unmap_page(dev,a,b,c) \
|
|
pci_unmap_page(to_pci_dev(dev),(a),(b),(c))
|
|
|
|
#define dma_map_single(dev,a,b,c) \
|
|
pci_map_single(to_pci_dev(dev),(a),(b),(c))
|
|
#define dma_unmap_single(dev,a,b,c) \
|
|
pci_unmap_single(to_pci_dev(dev),(a),(b),(c))
|
|
|
|
#define dma_map_sg(dev, sg, nents, dir) \
|
|
pci_map_sg(to_pci_dev(dev), (sg), (nents), (dir)
|
|
#define dma_unmap_sg(dev, sg, nents, dir) \
|
|
pci_unmap_sg(to_pci_dev(dev), (sg), (nents), (dir)
|
|
|
|
#define dma_sync_single(dev,a,b,c) \
|
|
pci_dma_sync_single(to_pci_dev(dev),(a),(b),(c))
|
|
|
|
/* for range just sync everything, that's all the pci API can do */
|
|
#define dma_sync_single_range(dev,addr,off,sz,dir) \
|
|
pci_dma_sync_single(to_pci_dev(dev),(addr),(off)+(sz),(dir))
|
|
|
|
#define dma_set_mask(dev,mask) \
|
|
pci_set_dma_mask(to_pci_dev(dev),(mask))
|
|
|
|
/* hlist_* code - double linked lists */
|
|
struct hlist_head {
|
|
struct hlist_node *first;
|
|
};
|
|
|
|
struct hlist_node {
|
|
struct hlist_node *next, **pprev;
|
|
};
|
|
|
|
static inline void __hlist_del(struct hlist_node *n)
|
|
{
|
|
struct hlist_node *next = n->next;
|
|
struct hlist_node **pprev = n->pprev;
|
|
*pprev = next;
|
|
if (next)
|
|
next->pprev = pprev;
|
|
}
|
|
|
|
static inline void hlist_del(struct hlist_node *n)
|
|
{
|
|
__hlist_del(n);
|
|
n->next = NULL;
|
|
n->pprev = NULL;
|
|
}
|
|
|
|
static inline void hlist_add_head(struct hlist_node *n, struct hlist_head *h)
|
|
{
|
|
struct hlist_node *first = h->first;
|
|
n->next = first;
|
|
if (first)
|
|
first->pprev = &n->next;
|
|
h->first = n;
|
|
n->pprev = &h->first;
|
|
}
|
|
|
|
static inline int hlist_empty(const struct hlist_head *h)
|
|
{
|
|
return !h->first;
|
|
}
|
|
#define HLIST_HEAD_INIT { .first = NULL }
|
|
#define HLIST_HEAD(name) struct hlist_head name = { .first = NULL }
|
|
#define INIT_HLIST_HEAD(ptr) ((ptr)->first = NULL)
|
|
static inline void INIT_HLIST_NODE(struct hlist_node *h)
|
|
{
|
|
h->next = NULL;
|
|
h->pprev = NULL;
|
|
}
|
|
|
|
#ifndef might_sleep
|
|
#define might_sleep()
|
|
#endif
|
|
#else
|
|
static inline struct device *pci_dev_to_dev(struct pci_dev *pdev)
|
|
{
|
|
return &pdev->dev;
|
|
}
|
|
#endif /* <= 2.5.0 */
|
|
|
|
/*****************************************************************************/
|
|
/* 2.5.28 => 2.4.23 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,5,28) )
|
|
|
|
#include <linux/tqueue.h>
|
|
#define work_struct tq_struct
|
|
#undef INIT_WORK
|
|
#define INIT_WORK(a,b) INIT_TQUEUE(a,(void (*)(void *))b,a)
|
|
#undef container_of
|
|
#define container_of list_entry
|
|
#define schedule_work schedule_task
|
|
#define flush_scheduled_work flush_scheduled_tasks
|
|
#define cancel_work_sync(x) flush_scheduled_work()
|
|
|
|
#endif /* 2.5.28 => 2.4.17 */
|
|
|
|
/*****************************************************************************/
|
|
/* 2.6.0 => 2.5.28 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0) )
|
|
#ifndef read_barrier_depends
|
|
#define read_barrier_depends() rmb()
|
|
#endif
|
|
|
|
#ifndef rcu_head
|
|
struct __kc_callback_head {
|
|
struct __kc_callback_head *next;
|
|
void (*func)(struct callback_head *head);
|
|
};
|
|
#define rcu_head __kc_callback_head
|
|
#endif
|
|
|
|
#undef get_cpu
|
|
#define get_cpu() smp_processor_id()
|
|
#undef put_cpu
|
|
#define put_cpu() do { } while(0)
|
|
#define MODULE_INFO(version, _version)
|
|
#ifndef CONFIG_E1000_DISABLE_PACKET_SPLIT
|
|
#define CONFIG_E1000_DISABLE_PACKET_SPLIT 1
|
|
#endif
|
|
#ifndef CONFIG_IGB_DISABLE_PACKET_SPLIT
|
|
#define CONFIG_IGB_DISABLE_PACKET_SPLIT 1
|
|
#endif
|
|
#ifndef CONFIG_IGC_DISABLE_PACKET_SPLIT
|
|
#define CONFIG_IGC_DISABLE_PACKET_SPLIT 1
|
|
#endif
|
|
|
|
#define dma_set_coherent_mask(dev,mask) 1
|
|
|
|
#undef dev_put
|
|
#define dev_put(dev) __dev_put(dev)
|
|
|
|
#ifndef skb_fill_page_desc
|
|
#define skb_fill_page_desc _kc_skb_fill_page_desc
|
|
void _kc_skb_fill_page_desc(struct sk_buff *skb, int i, struct page *page, int off, int size);
|
|
#endif
|
|
|
|
#undef ALIGN
|
|
#define ALIGN(x,a) (((x)+(a)-1)&~((a)-1))
|
|
|
|
#ifndef page_count
|
|
#define page_count(p) atomic_read(&(p)->count)
|
|
#endif
|
|
|
|
#ifdef MAX_NUMNODES
|
|
#undef MAX_NUMNODES
|
|
#endif
|
|
#define MAX_NUMNODES 1
|
|
|
|
/* find_first_bit and find_next bit are not defined for most
|
|
* 2.4 kernels (except for the redhat 2.4.21 kernels
|
|
*/
|
|
#include <linux/bitops.h>
|
|
#define BITOP_WORD(nr) ((nr) / BITS_PER_LONG)
|
|
#undef find_next_bit
|
|
#define find_next_bit _kc_find_next_bit
|
|
unsigned long _kc_find_next_bit(const unsigned long *addr, unsigned long size,
|
|
unsigned long offset);
|
|
#define find_first_bit(addr, size) find_next_bit((addr), (size), 0)
|
|
|
|
#ifndef netdev_name
|
|
static inline const char *_kc_netdev_name(const struct net_device *dev)
|
|
{
|
|
if (strchr(dev->name, '%'))
|
|
return "(unregistered net_device)";
|
|
return dev->name;
|
|
}
|
|
#define netdev_name(netdev) _kc_netdev_name(netdev)
|
|
#endif /* netdev_name */
|
|
|
|
#ifndef strlcpy
|
|
#define strlcpy _kc_strlcpy
|
|
size_t _kc_strlcpy(char *dest, const char *src, size_t size);
|
|
#endif /* strlcpy */
|
|
|
|
#ifndef do_div
|
|
#if BITS_PER_LONG == 64
|
|
# define do_div(n,base) ({ \
|
|
uint32_t __base = (base); \
|
|
uint32_t __rem; \
|
|
__rem = ((uint64_t)(n)) % __base; \
|
|
(n) = ((uint64_t)(n)) / __base; \
|
|
__rem; \
|
|
})
|
|
#elif BITS_PER_LONG == 32
|
|
uint32_t _kc__div64_32(uint64_t *dividend, uint32_t divisor);
|
|
# define do_div(n,base) ({ \
|
|
uint32_t __base = (base); \
|
|
uint32_t __rem; \
|
|
if (likely(((n) >> 32) == 0)) { \
|
|
__rem = (uint32_t)(n) % __base; \
|
|
(n) = (uint32_t)(n) / __base; \
|
|
} else \
|
|
__rem = _kc__div64_32(&(n), __base); \
|
|
__rem; \
|
|
})
|
|
#else /* BITS_PER_LONG == ?? */
|
|
# error do_div() does not yet support the C64
|
|
#endif /* BITS_PER_LONG */
|
|
#endif /* do_div */
|
|
|
|
#ifndef NSEC_PER_SEC
|
|
#define NSEC_PER_SEC 1000000000L
|
|
#endif
|
|
|
|
#undef HAVE_I2C_SUPPORT
|
|
#else /* 2.6.0 */
|
|
|
|
#endif /* 2.6.0 => 2.5.28 */
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,3) )
|
|
#define dma_pool pci_pool
|
|
#define dma_pool_destroy pci_pool_destroy
|
|
#define dma_pool_alloc pci_pool_alloc
|
|
#define dma_pool_free pci_pool_free
|
|
|
|
#define dma_pool_create(name,dev,size,align,allocation) \
|
|
pci_pool_create((name),to_pci_dev(dev),(size),(align),(allocation))
|
|
#endif /* < 2.6.3 */
|
|
|
|
/*****************************************************************************/
|
|
/* 2.6.4 => 2.6.0 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,4) )
|
|
#define MODULE_VERSION(_version) MODULE_INFO(version, _version)
|
|
#endif /* 2.6.4 => 2.6.0 */
|
|
|
|
/*****************************************************************************/
|
|
/* 2.6.5 => 2.6.0 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,5) )
|
|
#define dma_sync_single_for_cpu dma_sync_single
|
|
#define dma_sync_single_for_device dma_sync_single
|
|
#define dma_sync_single_range_for_cpu dma_sync_single_range
|
|
#define dma_sync_single_range_for_device dma_sync_single_range
|
|
#ifndef pci_dma_mapping_error
|
|
#define pci_dma_mapping_error _kc_pci_dma_mapping_error
|
|
static inline int _kc_pci_dma_mapping_error(dma_addr_t dma_addr)
|
|
{
|
|
return dma_addr == 0;
|
|
}
|
|
#endif
|
|
#endif /* 2.6.5 => 2.6.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,4) )
|
|
int _kc_scnprintf(char * buf, size_t size, const char *fmt, ...);
|
|
#define scnprintf(buf, size, fmt, args...) _kc_scnprintf(buf, size, fmt, ##args)
|
|
#endif /* < 2.6.4 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,6) )
|
|
/* taken from 2.6 include/linux/bitmap.h */
|
|
#undef bitmap_zero
|
|
#define bitmap_zero _kc_bitmap_zero
|
|
static inline void _kc_bitmap_zero(unsigned long *dst, int nbits)
|
|
{
|
|
if (nbits <= BITS_PER_LONG)
|
|
*dst = 0UL;
|
|
else {
|
|
int len = BITS_TO_LONGS(nbits) * sizeof(unsigned long);
|
|
memset(dst, 0, len);
|
|
}
|
|
}
|
|
#define page_to_nid(x) 0
|
|
|
|
#endif /* < 2.6.6 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,7) )
|
|
#undef if_mii
|
|
#define if_mii _kc_if_mii
|
|
static inline struct mii_ioctl_data *_kc_if_mii(struct ifreq *rq)
|
|
{
|
|
return (struct mii_ioctl_data *) &rq->ifr_ifru;
|
|
}
|
|
|
|
#ifndef __force
|
|
#define __force
|
|
#endif
|
|
#endif /* < 2.6.7 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,8) )
|
|
#ifndef PCI_EXP_DEVCTL
|
|
#define PCI_EXP_DEVCTL 8
|
|
#endif
|
|
#ifndef PCI_EXP_DEVCTL_CERE
|
|
#define PCI_EXP_DEVCTL_CERE 0x0001
|
|
#endif
|
|
#define PCI_EXP_FLAGS 2 /* Capabilities register */
|
|
#define PCI_EXP_FLAGS_VERS 0x000f /* Capability version */
|
|
#define PCI_EXP_FLAGS_TYPE 0x00f0 /* Device/Port type */
|
|
#define PCI_EXP_TYPE_ENDPOINT 0x0 /* Express Endpoint */
|
|
#define PCI_EXP_TYPE_LEG_END 0x1 /* Legacy Endpoint */
|
|
#define PCI_EXP_TYPE_ROOT_PORT 0x4 /* Root Port */
|
|
#define PCI_EXP_TYPE_DOWNSTREAM 0x6 /* Downstream Port */
|
|
#define PCI_EXP_FLAGS_SLOT 0x0100 /* Slot implemented */
|
|
#define PCI_EXP_DEVCAP 4 /* Device capabilities */
|
|
#define PCI_EXP_DEVSTA 10 /* Device Status */
|
|
#define msleep(x) do { set_current_state(TASK_UNINTERRUPTIBLE); \
|
|
schedule_timeout((x * HZ)/1000 + 2); \
|
|
} while (0)
|
|
|
|
#endif /* < 2.6.8 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,9))
|
|
#include <net/dsfield.h>
|
|
#define __iomem
|
|
|
|
#ifndef kcalloc
|
|
#define kcalloc(n, size, flags) _kc_kzalloc(((n) * (size)), flags)
|
|
void *_kc_kzalloc(size_t size, int flags);
|
|
#endif
|
|
#define MSEC_PER_SEC 1000L
|
|
static inline unsigned int _kc_jiffies_to_msecs(const unsigned long j)
|
|
{
|
|
#if HZ <= MSEC_PER_SEC && !(MSEC_PER_SEC % HZ)
|
|
return (MSEC_PER_SEC / HZ) * j;
|
|
#elif HZ > MSEC_PER_SEC && !(HZ % MSEC_PER_SEC)
|
|
return (j + (HZ / MSEC_PER_SEC) - 1)/(HZ / MSEC_PER_SEC);
|
|
#else
|
|
return (j * MSEC_PER_SEC) / HZ;
|
|
#endif
|
|
}
|
|
static inline unsigned long _kc_msecs_to_jiffies(const unsigned int m)
|
|
{
|
|
if (m > _kc_jiffies_to_msecs(MAX_JIFFY_OFFSET))
|
|
return MAX_JIFFY_OFFSET;
|
|
#if HZ <= MSEC_PER_SEC && !(MSEC_PER_SEC % HZ)
|
|
return (m + (MSEC_PER_SEC / HZ) - 1) / (MSEC_PER_SEC / HZ);
|
|
#elif HZ > MSEC_PER_SEC && !(HZ % MSEC_PER_SEC)
|
|
return m * (HZ / MSEC_PER_SEC);
|
|
#else
|
|
return (m * HZ + MSEC_PER_SEC - 1) / MSEC_PER_SEC;
|
|
#endif
|
|
}
|
|
|
|
#define msleep_interruptible _kc_msleep_interruptible
|
|
static inline unsigned long _kc_msleep_interruptible(unsigned int msecs)
|
|
{
|
|
unsigned long timeout = _kc_msecs_to_jiffies(msecs) + 1;
|
|
|
|
while (timeout && !signal_pending(current)) {
|
|
__set_current_state(TASK_INTERRUPTIBLE);
|
|
timeout = schedule_timeout(timeout);
|
|
}
|
|
return _kc_jiffies_to_msecs(timeout);
|
|
}
|
|
|
|
/* Basic mode control register. */
|
|
#define BMCR_SPEED1000 0x0040 /* MSB of Speed (1000) */
|
|
|
|
#ifndef __le16
|
|
#define __le16 u16
|
|
#endif
|
|
#ifndef __le32
|
|
#define __le32 u32
|
|
#endif
|
|
#ifndef __le64
|
|
#define __le64 u64
|
|
#endif
|
|
#ifndef __be16
|
|
#define __be16 u16
|
|
#endif
|
|
#ifndef __be32
|
|
#define __be32 u32
|
|
#endif
|
|
#ifndef __be64
|
|
#define __be64 u64
|
|
#endif
|
|
|
|
static inline struct vlan_ethhdr *vlan_eth_hdr(const struct sk_buff *skb)
|
|
{
|
|
return (struct vlan_ethhdr *)skb->mac.raw;
|
|
}
|
|
|
|
/* Wake-On-Lan options. */
|
|
#define WAKE_PHY BIT(0)
|
|
#define WAKE_UCAST BIT(1)
|
|
#define WAKE_MCAST BIT(2)
|
|
#define WAKE_BCAST BIT(3)
|
|
#define WAKE_ARP BIT(4)
|
|
#define WAKE_MAGIC BIT(5)
|
|
#define WAKE_MAGICSECURE BIT(6) /* only meaningful if WAKE_MAGIC */
|
|
|
|
#define skb_header_pointer _kc_skb_header_pointer
|
|
static inline void *_kc_skb_header_pointer(const struct sk_buff *skb,
|
|
int offset, int len, void *buffer)
|
|
{
|
|
int hlen = skb_headlen(skb);
|
|
|
|
if (hlen - offset >= len)
|
|
return skb->data + offset;
|
|
|
|
#ifdef MAX_SKB_FRAGS
|
|
if (skb_copy_bits(skb, offset, buffer, len) < 0)
|
|
return NULL;
|
|
|
|
return buffer;
|
|
#else
|
|
return NULL;
|
|
#endif
|
|
|
|
#ifndef NETDEV_TX_OK
|
|
#define NETDEV_TX_OK 0
|
|
#endif
|
|
#ifndef NETDEV_TX_BUSY
|
|
#define NETDEV_TX_BUSY 1
|
|
#endif
|
|
#ifndef NETDEV_TX_LOCKED
|
|
#define NETDEV_TX_LOCKED -1
|
|
#endif
|
|
}
|
|
|
|
#ifndef __bitwise
|
|
#define __bitwise
|
|
#endif
|
|
#endif /* < 2.6.9 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,10) )
|
|
#ifdef module_param_array_named
|
|
#undef module_param_array_named
|
|
#define module_param_array_named(name, array, type, nump, perm) \
|
|
static struct kparam_array __param_arr_##name \
|
|
= { ARRAY_SIZE(array), nump, param_set_##type, param_get_##type, \
|
|
sizeof(array[0]), array }; \
|
|
module_param_call(name, param_array_set, param_array_get, \
|
|
&__param_arr_##name, perm)
|
|
#endif /* module_param_array_named */
|
|
/*
|
|
* num_online is broken for all < 2.6.10 kernels. This is needed to support
|
|
* Node module parameter of ixgbe.
|
|
*/
|
|
#undef num_online_nodes
|
|
#define num_online_nodes(n) 1
|
|
extern DECLARE_BITMAP(_kcompat_node_online_map, MAX_NUMNODES);
|
|
#undef node_online_map
|
|
#define node_online_map _kcompat_node_online_map
|
|
#define pci_get_class pci_find_class
|
|
#endif /* < 2.6.10 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,11) )
|
|
#define PCI_D0 0
|
|
#define PCI_D1 1
|
|
#define PCI_D2 2
|
|
#define PCI_D3hot 3
|
|
#define PCI_D3cold 4
|
|
typedef int pci_power_t;
|
|
#define pci_choose_state(pdev,state) state
|
|
#define PMSG_SUSPEND 3
|
|
#define PCI_EXP_LNKCTL 16
|
|
|
|
#undef NETIF_F_LLTX
|
|
|
|
#ifndef ARCH_HAS_PREFETCH
|
|
#define prefetch(X)
|
|
#endif
|
|
|
|
#ifndef NET_IP_ALIGN
|
|
#define NET_IP_ALIGN 2
|
|
#endif
|
|
|
|
#define KC_USEC_PER_SEC 1000000L
|
|
#define usecs_to_jiffies _kc_usecs_to_jiffies
|
|
static inline unsigned int _kc_jiffies_to_usecs(const unsigned long j)
|
|
{
|
|
#if HZ <= KC_USEC_PER_SEC && !(KC_USEC_PER_SEC % HZ)
|
|
return (KC_USEC_PER_SEC / HZ) * j;
|
|
#elif HZ > KC_USEC_PER_SEC && !(HZ % KC_USEC_PER_SEC)
|
|
return (j + (HZ / KC_USEC_PER_SEC) - 1)/(HZ / KC_USEC_PER_SEC);
|
|
#else
|
|
return (j * KC_USEC_PER_SEC) / HZ;
|
|
#endif
|
|
}
|
|
static inline unsigned long _kc_usecs_to_jiffies(const unsigned int m)
|
|
{
|
|
if (m > _kc_jiffies_to_usecs(MAX_JIFFY_OFFSET))
|
|
return MAX_JIFFY_OFFSET;
|
|
#if HZ <= KC_USEC_PER_SEC && !(KC_USEC_PER_SEC % HZ)
|
|
return (m + (KC_USEC_PER_SEC / HZ) - 1) / (KC_USEC_PER_SEC / HZ);
|
|
#elif HZ > KC_USEC_PER_SEC && !(HZ % KC_USEC_PER_SEC)
|
|
return m * (HZ / KC_USEC_PER_SEC);
|
|
#else
|
|
return (m * HZ + KC_USEC_PER_SEC - 1) / KC_USEC_PER_SEC;
|
|
#endif
|
|
}
|
|
|
|
#define PCI_EXP_LNKCAP 12 /* Link Capabilities */
|
|
#define PCI_EXP_LNKSTA 18 /* Link Status */
|
|
#define PCI_EXP_SLTCAP 20 /* Slot Capabilities */
|
|
#define PCI_EXP_SLTCTL 24 /* Slot Control */
|
|
#define PCI_EXP_SLTSTA 26 /* Slot Status */
|
|
#define PCI_EXP_RTCTL 28 /* Root Control */
|
|
#define PCI_EXP_RTCAP 30 /* Root Capabilities */
|
|
#define PCI_EXP_RTSTA 32 /* Root Status */
|
|
#endif /* < 2.6.11 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,12) )
|
|
#include <linux/reboot.h>
|
|
#define USE_REBOOT_NOTIFIER
|
|
|
|
/* Generic MII registers. */
|
|
#define MII_CTRL1000 0x09 /* 1000BASE-T control */
|
|
#define MII_STAT1000 0x0a /* 1000BASE-T status */
|
|
/* Advertisement control register. */
|
|
#define ADVERTISE_PAUSE_CAP 0x0400 /* Try for pause */
|
|
#define ADVERTISE_PAUSE_ASYM 0x0800 /* Try for asymmetric pause */
|
|
/* Link partner ability register. */
|
|
#define LPA_PAUSE_CAP 0x0400 /* Can pause */
|
|
#define LPA_PAUSE_ASYM 0x0800 /* Can pause asymetrically */
|
|
/* 1000BASE-T Control register */
|
|
#define ADVERTISE_1000FULL 0x0200 /* Advertise 1000BASE-T full duplex */
|
|
#define ADVERTISE_1000HALF 0x0100 /* Advertise 1000BASE-T half duplex */
|
|
/* 1000BASE-T Status register */
|
|
#define LPA_1000LOCALRXOK 0x2000 /* Link partner local receiver status */
|
|
#define LPA_1000REMRXOK 0x1000 /* Link partner remote receiver status */
|
|
|
|
#ifndef is_zero_ether_addr
|
|
#define is_zero_ether_addr _kc_is_zero_ether_addr
|
|
static inline int _kc_is_zero_ether_addr(const u8 *addr)
|
|
{
|
|
return !(addr[0] | addr[1] | addr[2] | addr[3] | addr[4] | addr[5]);
|
|
}
|
|
#endif /* is_zero_ether_addr */
|
|
#ifndef is_multicast_ether_addr
|
|
#define is_multicast_ether_addr _kc_is_multicast_ether_addr
|
|
static inline int _kc_is_multicast_ether_addr(const u8 *addr)
|
|
{
|
|
return addr[0] & 0x01;
|
|
}
|
|
#endif /* is_multicast_ether_addr */
|
|
#endif /* < 2.6.12 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,13) )
|
|
#ifndef kstrdup
|
|
#define kstrdup _kc_kstrdup
|
|
char *_kc_kstrdup(const char *s, unsigned int gfp);
|
|
#endif
|
|
#endif /* < 2.6.13 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,14) )
|
|
#define pm_message_t u32
|
|
#ifndef kzalloc
|
|
#define kzalloc _kc_kzalloc
|
|
void *_kc_kzalloc(size_t size, int flags);
|
|
#endif
|
|
|
|
/* Generic MII registers. */
|
|
#define MII_ESTATUS 0x0f /* Extended Status */
|
|
/* Basic mode status register. */
|
|
#define BMSR_ESTATEN 0x0100 /* Extended Status in R15 */
|
|
/* Extended status register. */
|
|
#define ESTATUS_1000_TFULL 0x2000 /* Can do 1000BT Full */
|
|
#define ESTATUS_1000_THALF 0x1000 /* Can do 1000BT Half */
|
|
|
|
#define SUPPORTED_Pause BIT(13)
|
|
#define SUPPORTED_Asym_Pause BIT(14)
|
|
#define ADVERTISED_Pause BIT(13)
|
|
#define ADVERTISED_Asym_Pause BIT(14)
|
|
|
|
#if (!(RHEL_RELEASE_CODE && \
|
|
(RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(4,3)) && \
|
|
(RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(5,0))))
|
|
#if ((LINUX_VERSION_CODE == KERNEL_VERSION(2,6,9)) && !defined(gfp_t))
|
|
#define gfp_t unsigned
|
|
#else
|
|
typedef unsigned gfp_t;
|
|
#endif
|
|
#endif /* !RHEL4.3->RHEL5.0 */
|
|
|
|
#if ( LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,9) )
|
|
#ifdef CONFIG_X86_64
|
|
#define dma_sync_single_range_for_cpu(dev, addr, off, sz, dir) \
|
|
dma_sync_single_for_cpu((dev), (addr), (off) + (sz), (dir))
|
|
#define dma_sync_single_range_for_device(dev, addr, off, sz, dir) \
|
|
dma_sync_single_for_device((dev), (addr), (off) + (sz), (dir))
|
|
#endif
|
|
#endif
|
|
#endif /* < 2.6.14 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,15) )
|
|
#ifndef kfree_rcu
|
|
/* this is placed here due to a lack of rcu_barrier in previous kernels */
|
|
#define kfree_rcu(_ptr, _offset) kfree(_ptr)
|
|
#endif /* kfree_rcu */
|
|
#ifndef vmalloc_node
|
|
#define vmalloc_node(a,b) vmalloc(a)
|
|
#endif /* vmalloc_node*/
|
|
|
|
#define setup_timer(_timer, _function, _data) \
|
|
do { \
|
|
(_timer)->function = _function; \
|
|
(_timer)->data = _data; \
|
|
init_timer(_timer); \
|
|
} while (0)
|
|
#ifndef device_can_wakeup
|
|
#define device_can_wakeup(dev) (1)
|
|
#endif
|
|
#ifndef device_set_wakeup_enable
|
|
#define device_set_wakeup_enable(dev, val) do{}while(0)
|
|
#endif
|
|
#ifndef device_init_wakeup
|
|
#define device_init_wakeup(dev,val) do {} while (0)
|
|
#endif
|
|
static inline unsigned _kc_compare_ether_addr(const u8 *addr1, const u8 *addr2)
|
|
{
|
|
const u16 *a = (const u16 *) addr1;
|
|
const u16 *b = (const u16 *) addr2;
|
|
|
|
return ((a[0] ^ b[0]) | (a[1] ^ b[1]) | (a[2] ^ b[2])) != 0;
|
|
}
|
|
#undef compare_ether_addr
|
|
#define compare_ether_addr(addr1, addr2) _kc_compare_ether_addr(addr1, addr2)
|
|
#endif /* < 2.6.15 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,16) )
|
|
#undef DEFINE_MUTEX
|
|
#define DEFINE_MUTEX(x) DECLARE_MUTEX(x)
|
|
#define mutex_lock(x) down_interruptible(x)
|
|
#define mutex_unlock(x) up(x)
|
|
|
|
#ifndef ____cacheline_internodealigned_in_smp
|
|
#ifdef CONFIG_SMP
|
|
#define ____cacheline_internodealigned_in_smp ____cacheline_aligned_in_smp
|
|
#else
|
|
#define ____cacheline_internodealigned_in_smp
|
|
#endif /* CONFIG_SMP */
|
|
#endif /* ____cacheline_internodealigned_in_smp */
|
|
#undef HAVE_PCI_ERS
|
|
#else /* 2.6.16 and above */
|
|
#undef HAVE_PCI_ERS
|
|
#define HAVE_PCI_ERS
|
|
#if ( SLE_VERSION_CODE && SLE_VERSION_CODE == SLE_VERSION(10,4,0) )
|
|
#ifdef device_can_wakeup
|
|
#undef device_can_wakeup
|
|
#endif /* device_can_wakeup */
|
|
#define device_can_wakeup(dev) 1
|
|
#endif /* SLE_VERSION(10,4,0) */
|
|
#endif /* < 2.6.16 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,17) )
|
|
#ifndef dev_notice
|
|
#define dev_notice(dev, fmt, args...) \
|
|
dev_printk(KERN_NOTICE, dev, fmt, ## args)
|
|
#endif
|
|
|
|
#ifndef first_online_node
|
|
#define first_online_node 0
|
|
#endif
|
|
#ifndef NET_SKB_PAD
|
|
#define NET_SKB_PAD 16
|
|
#endif
|
|
#endif /* < 2.6.17 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,18) )
|
|
|
|
#ifndef IRQ_HANDLED
|
|
#define irqreturn_t void
|
|
#define IRQ_HANDLED
|
|
#define IRQ_NONE
|
|
#endif
|
|
|
|
#ifndef IRQF_PROBE_SHARED
|
|
#ifdef SA_PROBEIRQ
|
|
#define IRQF_PROBE_SHARED SA_PROBEIRQ
|
|
#else
|
|
#define IRQF_PROBE_SHARED 0
|
|
#endif
|
|
#endif
|
|
|
|
#ifndef IRQF_SHARED
|
|
#define IRQF_SHARED SA_SHIRQ
|
|
#endif
|
|
|
|
#ifndef ARRAY_SIZE
|
|
#define ARRAY_SIZE(x) (sizeof(x) / sizeof((x)[0]))
|
|
#endif
|
|
|
|
#ifndef skb_is_gso
|
|
#ifdef NETIF_F_TSO
|
|
#define skb_is_gso _kc_skb_is_gso
|
|
static inline int _kc_skb_is_gso(const struct sk_buff *skb)
|
|
{
|
|
return skb_shinfo(skb)->gso_size;
|
|
}
|
|
#else
|
|
#define skb_is_gso(a) 0
|
|
#endif
|
|
#endif
|
|
|
|
#ifndef resource_size_t
|
|
#define resource_size_t unsigned long
|
|
#endif
|
|
|
|
#ifdef skb_pad
|
|
#undef skb_pad
|
|
#endif
|
|
#define skb_pad(x,y) _kc_skb_pad(x, y)
|
|
int _kc_skb_pad(struct sk_buff *skb, int pad);
|
|
#ifdef skb_padto
|
|
#undef skb_padto
|
|
#endif
|
|
#define skb_padto(x,y) _kc_skb_padto(x, y)
|
|
static inline int _kc_skb_padto(struct sk_buff *skb, unsigned int len)
|
|
{
|
|
unsigned int size = skb->len;
|
|
if(likely(size >= len))
|
|
return 0;
|
|
return _kc_skb_pad(skb, len - size);
|
|
}
|
|
|
|
#ifndef DECLARE_PCI_UNMAP_ADDR
|
|
#define DECLARE_PCI_UNMAP_ADDR(ADDR_NAME) \
|
|
dma_addr_t ADDR_NAME
|
|
#define DECLARE_PCI_UNMAP_LEN(LEN_NAME) \
|
|
u32 LEN_NAME
|
|
#define pci_unmap_addr(PTR, ADDR_NAME) \
|
|
((PTR)->ADDR_NAME)
|
|
#define pci_unmap_addr_set(PTR, ADDR_NAME, VAL) \
|
|
(((PTR)->ADDR_NAME) = (VAL))
|
|
#define pci_unmap_len(PTR, LEN_NAME) \
|
|
((PTR)->LEN_NAME)
|
|
#define pci_unmap_len_set(PTR, LEN_NAME, VAL) \
|
|
(((PTR)->LEN_NAME) = (VAL))
|
|
#endif /* DECLARE_PCI_UNMAP_ADDR */
|
|
#endif /* < 2.6.18 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,19) )
|
|
enum pcie_link_width {
|
|
PCIE_LNK_WIDTH_RESRV = 0x00,
|
|
PCIE_LNK_X1 = 0x01,
|
|
PCIE_LNK_X2 = 0x02,
|
|
PCIE_LNK_X4 = 0x04,
|
|
PCIE_LNK_X8 = 0x08,
|
|
PCIE_LNK_X12 = 0x0C,
|
|
PCIE_LNK_X16 = 0x10,
|
|
PCIE_LNK_X32 = 0x20,
|
|
PCIE_LNK_WIDTH_UNKNOWN = 0xFF,
|
|
};
|
|
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(5,0)))
|
|
#define i_private u.generic_ip
|
|
#endif /* >= RHEL 5.0 */
|
|
|
|
#ifndef DIV_ROUND_UP
|
|
#define DIV_ROUND_UP(n,d) (((n) + (d) - 1) / (d))
|
|
#endif
|
|
#ifndef __ALIGN_MASK
|
|
#define __ALIGN_MASK(x, mask) (((x) + (mask)) & ~(mask))
|
|
#endif
|
|
#if ( LINUX_VERSION_CODE >= KERNEL_VERSION(2,5,0) )
|
|
#if (!((RHEL_RELEASE_CODE && \
|
|
((RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(4,4) && \
|
|
RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(5,0)) || \
|
|
(RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(5,0))))))
|
|
typedef irqreturn_t (*irq_handler_t)(int, void*, struct pt_regs *);
|
|
#endif
|
|
#if (RHEL_RELEASE_CODE && RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(6,0))
|
|
#undef CONFIG_INET_LRO
|
|
#undef CONFIG_INET_LRO_MODULE
|
|
#endif
|
|
typedef irqreturn_t (*new_handler_t)(int, void*);
|
|
static inline irqreturn_t _kc_request_irq(unsigned int irq, new_handler_t handler, unsigned long flags, const char *devname, void *dev_id)
|
|
#else /* 2.4.x */
|
|
typedef void (*irq_handler_t)(int, void*, struct pt_regs *);
|
|
typedef void (*new_handler_t)(int, void*);
|
|
static inline int _kc_request_irq(unsigned int irq, new_handler_t handler, unsigned long flags, const char *devname, void *dev_id)
|
|
#endif /* >= 2.5.x */
|
|
{
|
|
irq_handler_t new_handler = (irq_handler_t) handler;
|
|
return request_irq(irq, new_handler, flags, devname, dev_id);
|
|
}
|
|
|
|
#undef request_irq
|
|
#define request_irq(irq, handler, flags, devname, dev_id) _kc_request_irq((irq), (handler), (flags), (devname), (dev_id))
|
|
|
|
#define irq_handler_t new_handler_t
|
|
|
|
#if ( LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,11) )
|
|
#ifndef skb_checksum_help
|
|
static inline int __kc_skb_checksum_help(struct sk_buff *skb)
|
|
{
|
|
return skb_checksum_help(skb, 0);
|
|
}
|
|
#define skb_checksum_help(skb) __kc_skb_checksum_help((skb))
|
|
#endif
|
|
#endif /* < 2.6.19 && >= 2.6.11 */
|
|
|
|
/* pci_restore_state and pci_save_state handles MSI/PCIE from 2.6.19 */
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(5,4)))
|
|
#define PCIE_CONFIG_SPACE_LEN 256
|
|
#define PCI_CONFIG_SPACE_LEN 64
|
|
#define PCIE_LINK_STATUS 0x12
|
|
#define pci_config_space_ich8lan() do {} while(0)
|
|
#undef pci_save_state
|
|
int _kc_pci_save_state(struct pci_dev *);
|
|
#define pci_save_state(pdev) _kc_pci_save_state(pdev)
|
|
#undef pci_restore_state
|
|
void _kc_pci_restore_state(struct pci_dev *);
|
|
#define pci_restore_state(pdev) _kc_pci_restore_state(pdev)
|
|
#endif /* !(RHEL_RELEASE_CODE >= RHEL 5.4) */
|
|
|
|
#ifdef HAVE_PCI_ERS
|
|
#undef free_netdev
|
|
void _kc_free_netdev(struct net_device *);
|
|
#define free_netdev(netdev) _kc_free_netdev(netdev)
|
|
#endif
|
|
static inline int pci_enable_pcie_error_reporting(struct pci_dev __always_unused *dev)
|
|
{
|
|
return 0;
|
|
}
|
|
#define pci_disable_pcie_error_reporting(dev) do {} while (0)
|
|
#define pci_cleanup_aer_uncorrect_error_status(dev) do {} while (0)
|
|
|
|
void *_kc_kmemdup(const void *src, size_t len, unsigned gfp);
|
|
#define kmemdup(src, len, gfp) _kc_kmemdup(src, len, gfp)
|
|
#ifndef bool
|
|
#define bool _Bool
|
|
#define true 1
|
|
#define false 0
|
|
#endif
|
|
#else /* 2.6.19 */
|
|
#include <linux/aer.h>
|
|
#include <linux/pci_hotplug.h>
|
|
|
|
#define NEW_SKB_CSUM_HELP
|
|
#endif /* < 2.6.19 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20) )
|
|
#if ( LINUX_VERSION_CODE >= KERNEL_VERSION(2,5,28) )
|
|
#undef INIT_WORK
|
|
#define INIT_WORK(_work, _func) \
|
|
do { \
|
|
INIT_LIST_HEAD(&(_work)->entry); \
|
|
(_work)->pending = 0; \
|
|
(_work)->func = (void (*)(void *))_func; \
|
|
(_work)->data = _work; \
|
|
init_timer(&(_work)->timer); \
|
|
} while (0)
|
|
#endif
|
|
|
|
#ifndef PCI_VDEVICE
|
|
#define PCI_VDEVICE(ven, dev) \
|
|
PCI_VENDOR_ID_##ven, (dev), \
|
|
PCI_ANY_ID, PCI_ANY_ID, 0, 0
|
|
#endif
|
|
|
|
#ifndef PCI_VENDOR_ID_INTEL
|
|
#define PCI_VENDOR_ID_INTEL 0x8086
|
|
#endif
|
|
|
|
#ifndef round_jiffies
|
|
#define round_jiffies(x) x
|
|
#endif
|
|
|
|
#define csum_offset csum
|
|
|
|
#define HAVE_EARLY_VMALLOC_NODE
|
|
#define dev_to_node(dev) -1
|
|
#undef set_dev_node
|
|
/* remove compiler warning with b=b, for unused variable */
|
|
#define set_dev_node(a, b) do { (b) = (b); } while(0)
|
|
|
|
#if (!(RHEL_RELEASE_CODE && \
|
|
(((RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(4,7)) && \
|
|
(RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(5,0))) || \
|
|
(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(5,6)))) && \
|
|
!(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(10,2,0)))
|
|
typedef __u16 __bitwise __sum16;
|
|
typedef __u32 __bitwise __wsum;
|
|
#endif
|
|
|
|
#if (!(RHEL_RELEASE_CODE && \
|
|
(((RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(4,7)) && \
|
|
(RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(5,0))) || \
|
|
(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(5,4)))) && \
|
|
!(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(10,2,0)))
|
|
static inline __wsum csum_unfold(__sum16 n)
|
|
{
|
|
return (__force __wsum)n;
|
|
}
|
|
#endif
|
|
|
|
#else /* < 2.6.20 */
|
|
#define HAVE_DEVICE_NUMA_NODE
|
|
#endif /* < 2.6.20 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,21) )
|
|
#define to_net_dev(class) container_of(class, struct net_device, class_dev)
|
|
#define NETDEV_CLASS_DEV
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(5,5)))
|
|
#define vlan_group_get_device(vg, id) (vg->vlan_devices[id])
|
|
#define vlan_group_set_device(vg, id, dev) \
|
|
do { \
|
|
if (vg) vg->vlan_devices[id] = dev; \
|
|
} while (0)
|
|
#endif /* !(RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(5,5)) */
|
|
#define pci_channel_offline(pdev) (pdev->error_state && \
|
|
pdev->error_state != pci_channel_io_normal)
|
|
#define pci_request_selected_regions(pdev, bars, name) \
|
|
pci_request_regions(pdev, name)
|
|
#define pci_release_selected_regions(pdev, bars) pci_release_regions(pdev);
|
|
|
|
#ifndef __aligned
|
|
#define __aligned(x) __attribute__((aligned(x)))
|
|
#endif
|
|
|
|
struct pci_dev *_kc_netdev_to_pdev(struct net_device *netdev);
|
|
#define netdev_to_dev(netdev) \
|
|
pci_dev_to_dev(_kc_netdev_to_pdev(netdev))
|
|
#define devm_kzalloc(dev, size, flags) kzalloc(size, flags)
|
|
#define devm_kfree(dev, p) kfree(p)
|
|
#else /* 2.6.21 */
|
|
static inline struct device *netdev_to_dev(struct net_device *netdev)
|
|
{
|
|
return &netdev->dev;
|
|
}
|
|
|
|
#endif /* < 2.6.21 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,22) )
|
|
#define tcp_hdr(skb) (skb->h.th)
|
|
#define tcp_hdrlen(skb) (skb->h.th->doff << 2)
|
|
#define skb_transport_offset(skb) (skb->h.raw - skb->data)
|
|
#define skb_transport_header(skb) (skb->h.raw)
|
|
#define ipv6_hdr(skb) (skb->nh.ipv6h)
|
|
#define ip_hdr(skb) (skb->nh.iph)
|
|
#define skb_network_offset(skb) (skb->nh.raw - skb->data)
|
|
#define skb_network_header(skb) (skb->nh.raw)
|
|
#define skb_tail_pointer(skb) skb->tail
|
|
#define skb_reset_tail_pointer(skb) \
|
|
do { \
|
|
skb->tail = skb->data; \
|
|
} while (0)
|
|
#define skb_set_tail_pointer(skb, offset) \
|
|
do { \
|
|
skb->tail = skb->data + offset; \
|
|
} while (0)
|
|
#define skb_copy_to_linear_data(skb, from, len) \
|
|
memcpy(skb->data, from, len)
|
|
#define skb_copy_to_linear_data_offset(skb, offset, from, len) \
|
|
memcpy(skb->data + offset, from, len)
|
|
#define skb_network_header_len(skb) (skb->h.raw - skb->nh.raw)
|
|
#define pci_register_driver pci_module_init
|
|
#define skb_mac_header(skb) skb->mac.raw
|
|
|
|
#ifdef NETIF_F_MULTI_QUEUE
|
|
#ifndef alloc_etherdev_mq
|
|
#define alloc_etherdev_mq(_a, _b) alloc_etherdev(_a)
|
|
#endif
|
|
#endif /* NETIF_F_MULTI_QUEUE */
|
|
|
|
#ifndef ETH_FCS_LEN
|
|
#define ETH_FCS_LEN 4
|
|
#endif
|
|
#define cancel_work_sync(x) flush_scheduled_work()
|
|
#ifndef udp_hdr
|
|
#define udp_hdr _udp_hdr
|
|
static inline struct udphdr *_udp_hdr(const struct sk_buff *skb)
|
|
{
|
|
return (struct udphdr *)skb_transport_header(skb);
|
|
}
|
|
#endif
|
|
|
|
#ifdef cpu_to_be16
|
|
#undef cpu_to_be16
|
|
#endif
|
|
#define cpu_to_be16(x) __constant_htons(x)
|
|
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(5,1)))
|
|
enum {
|
|
DUMP_PREFIX_NONE,
|
|
DUMP_PREFIX_ADDRESS,
|
|
DUMP_PREFIX_OFFSET
|
|
};
|
|
#endif /* !(RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(5,1)) */
|
|
#ifndef hex_asc
|
|
#define hex_asc(x) "0123456789abcdef"[x]
|
|
#endif
|
|
#include <linux/ctype.h>
|
|
void _kc_print_hex_dump(const char *level, const char *prefix_str,
|
|
int prefix_type, int rowsize, int groupsize,
|
|
const void *buf, size_t len, bool ascii);
|
|
#define print_hex_dump(lvl, s, t, r, g, b, l, a) \
|
|
_kc_print_hex_dump(lvl, s, t, r, g, b, l, a)
|
|
#ifndef ADVERTISED_2500baseX_Full
|
|
#define ADVERTISED_2500baseX_Full BIT(15)
|
|
#endif
|
|
#ifndef SUPPORTED_2500baseX_Full
|
|
#define SUPPORTED_2500baseX_Full BIT(15)
|
|
#endif
|
|
|
|
#ifndef ETH_P_PAUSE
|
|
#define ETH_P_PAUSE 0x8808
|
|
#endif
|
|
|
|
static inline int compound_order(struct page *page)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
#define __must_be_array(a) 0
|
|
|
|
#ifndef SKB_WITH_OVERHEAD
|
|
#define SKB_WITH_OVERHEAD(X) \
|
|
((X) - SKB_DATA_ALIGN(sizeof(struct skb_shared_info)))
|
|
#endif
|
|
#else /* 2.6.22 */
|
|
#define ETH_TYPE_TRANS_SETS_DEV
|
|
#define HAVE_NETDEV_STATS_IN_NETDEV
|
|
#endif /* < 2.6.22 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE > KERNEL_VERSION(2,6,22) )
|
|
#endif /* > 2.6.22 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,23) )
|
|
#define netif_subqueue_stopped(_a, _b) 0
|
|
#ifndef PTR_ALIGN
|
|
#define PTR_ALIGN(p, a) ((typeof(p))ALIGN((unsigned long)(p), (a)))
|
|
#endif
|
|
|
|
#ifndef CONFIG_PM_SLEEP
|
|
#define CONFIG_PM_SLEEP CONFIG_PM
|
|
#endif
|
|
|
|
#if ( LINUX_VERSION_CODE > KERNEL_VERSION(2,6,13) )
|
|
#define HAVE_ETHTOOL_GET_PERM_ADDR
|
|
#endif /* 2.6.14 through 2.6.22 */
|
|
|
|
static inline int __kc_skb_cow_head(struct sk_buff *skb, unsigned int headroom)
|
|
{
|
|
int delta = 0;
|
|
|
|
if (headroom > (skb->data - skb->head))
|
|
delta = headroom - (skb->data - skb->head);
|
|
|
|
if (delta || skb_header_cloned(skb))
|
|
return pskb_expand_head(skb, ALIGN(delta, NET_SKB_PAD), 0,
|
|
GFP_ATOMIC);
|
|
return 0;
|
|
}
|
|
#define skb_cow_head(s, h) __kc_skb_cow_head((s), (h))
|
|
#endif /* < 2.6.23 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,24) )
|
|
#ifndef ETH_FLAG_LRO
|
|
#define ETH_FLAG_LRO NETIF_F_LRO
|
|
#endif
|
|
|
|
#ifndef ACCESS_ONCE
|
|
#define ACCESS_ONCE(x) (*(volatile typeof(x) *)&(x))
|
|
#endif
|
|
|
|
/* if GRO is supported then the napi struct must already exist */
|
|
#ifndef NETIF_F_GRO
|
|
/* NAPI API changes in 2.6.24 break everything */
|
|
struct napi_struct {
|
|
/* used to look up the real NAPI polling routine */
|
|
int (*poll)(struct napi_struct *, int);
|
|
struct net_device *dev;
|
|
int weight;
|
|
};
|
|
#endif
|
|
|
|
#ifdef NAPI
|
|
int __kc_adapter_clean(struct net_device *, int *);
|
|
/* The following defines only provide limited support for NAPI calls and
|
|
* should only be used by drivers which are not multi-queue enabled.
|
|
*/
|
|
|
|
#define napi_to_poll_dev(_napi) (_napi)->dev
|
|
|
|
static inline void __kc_netif_napi_add(struct net_device *dev, struct napi_struct *napi,
|
|
int (*poll)(struct napi_struct *, int), int weight)
|
|
{
|
|
dev->poll = __kc_adapter_clean;
|
|
dev->weight = weight;
|
|
napi->poll = poll;
|
|
napi->dev = dev;
|
|
}
|
|
#define netif_napi_add __kc_netif_napi_add
|
|
|
|
#define netif_napi_del(_a) do {} while (0)
|
|
#define napi_schedule_prep(_napi) netif_rx_schedule_prep((_napi)->dev)
|
|
#define napi_schedule(_napi) netif_rx_schedule((_napi)->dev)
|
|
#define napi_enable(_napi) netif_poll_enable(napi_to_poll_dev(_napi))
|
|
#define napi_disable(_napi) netif_poll_disable(napi_to_poll_dev(_napi))
|
|
#ifdef CONFIG_SMP
|
|
static inline void napi_synchronize(const struct napi_struct *n)
|
|
{
|
|
struct net_device *dev = napi_to_poll_dev(n);
|
|
|
|
while (test_bit(__LINK_STATE_RX_SCHED, &dev->state)) {
|
|
/* No hurry. */
|
|
msleep(1);
|
|
}
|
|
}
|
|
#else
|
|
#define napi_synchronize(n) barrier()
|
|
#endif /* CONFIG_SMP */
|
|
#define __napi_schedule(_napi) __netif_rx_schedule(napi_to_poll_dev(_napi))
|
|
static inline void _kc_napi_complete(struct napi_struct *napi)
|
|
{
|
|
#ifdef NETIF_F_GRO
|
|
napi_gro_flush(napi);
|
|
#endif
|
|
netif_rx_complete(napi_to_poll_dev(napi));
|
|
}
|
|
#define napi_complete _kc_napi_complete
|
|
#else /* NAPI */
|
|
|
|
/* The following definitions are only used if we don't support NAPI at all. */
|
|
|
|
static inline __kc_netif_napi_add(struct net_device *dev, struct napi_struct *napi,
|
|
int (*poll)(struct napi_struct *, int), int weight)
|
|
{
|
|
dev->poll = poll;
|
|
dev->weight = weight;
|
|
napi->poll = poll;
|
|
napi->weight = weight;
|
|
napi->dev = dev;
|
|
}
|
|
#define netif_napi_del(_a) do {} while (0)
|
|
#endif /* NAPI */
|
|
|
|
#undef dev_get_by_name
|
|
#define dev_get_by_name(_a, _b) dev_get_by_name(_b)
|
|
#define __netif_subqueue_stopped(_a, _b) netif_subqueue_stopped(_a, _b)
|
|
#ifndef DMA_BIT_MASK
|
|
#define DMA_BIT_MASK(n) (((n) == 64) ? DMA_64BIT_MASK : ((1ULL<<(n))-1))
|
|
#endif
|
|
|
|
#ifdef NETIF_F_TSO6
|
|
#define skb_is_gso_v6 _kc_skb_is_gso_v6
|
|
static inline int _kc_skb_is_gso_v6(const struct sk_buff *skb)
|
|
{
|
|
return skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6;
|
|
}
|
|
#endif /* NETIF_F_TSO6 */
|
|
|
|
#ifndef KERN_CONT
|
|
#define KERN_CONT ""
|
|
#endif
|
|
#ifndef pr_err
|
|
#define pr_err(fmt, arg...) \
|
|
printk(KERN_ERR fmt, ##arg)
|
|
#endif
|
|
|
|
#ifndef rounddown_pow_of_two
|
|
#define rounddown_pow_of_two(n) \
|
|
__builtin_constant_p(n) ? ( \
|
|
(n == 1) ? 0 : \
|
|
(1UL << ilog2(n))) : \
|
|
(1UL << (fls_long(n) - 1))
|
|
#endif
|
|
|
|
#else /* < 2.6.24 */
|
|
#define HAVE_ETHTOOL_GET_SSET_COUNT
|
|
#define HAVE_NETDEV_NAPI_LIST
|
|
#endif /* < 2.6.24 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE > KERNEL_VERSION(2,6,24) )
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,2,0) )
|
|
#define INCLUDE_PM_QOS_PARAMS_H
|
|
#include <linux/pm_qos_params.h>
|
|
#else /* >= 3.2.0 */
|
|
#include <linux/pm_qos.h>
|
|
#endif /* else >= 3.2.0 */
|
|
#endif /* > 2.6.24 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,25) )
|
|
#define PM_QOS_CPU_DMA_LATENCY 1
|
|
|
|
#if ( LINUX_VERSION_CODE > KERNEL_VERSION(2,6,18) )
|
|
#include <linux/latency.h>
|
|
#define PM_QOS_DEFAULT_VALUE INFINITE_LATENCY
|
|
#define pm_qos_add_requirement(pm_qos_class, name, value) \
|
|
set_acceptable_latency(name, value)
|
|
#define pm_qos_remove_requirement(pm_qos_class, name) \
|
|
remove_acceptable_latency(name)
|
|
#define pm_qos_update_requirement(pm_qos_class, name, value) \
|
|
modify_acceptable_latency(name, value)
|
|
#else
|
|
#define PM_QOS_DEFAULT_VALUE -1
|
|
#define pm_qos_add_requirement(pm_qos_class, name, value)
|
|
#define pm_qos_remove_requirement(pm_qos_class, name)
|
|
#define pm_qos_update_requirement(pm_qos_class, name, value) { \
|
|
if (value != PM_QOS_DEFAULT_VALUE) { \
|
|
printk(KERN_WARNING "%s: unable to set PM QoS requirement\n", \
|
|
pci_name(adapter->pdev)); \
|
|
} \
|
|
}
|
|
|
|
#endif /* > 2.6.18 */
|
|
|
|
#define pci_enable_device_mem(pdev) pci_enable_device(pdev)
|
|
|
|
#ifndef DEFINE_PCI_DEVICE_TABLE
|
|
#define DEFINE_PCI_DEVICE_TABLE(_table) struct pci_device_id _table[]
|
|
#endif /* DEFINE_PCI_DEVICE_TABLE */
|
|
|
|
#ifndef strict_strtol
|
|
#define strict_strtol(s, b, r) _kc_strict_strtol(s, b, r)
|
|
static inline int _kc_strict_strtol(const char *buf, unsigned int base, long *res)
|
|
{
|
|
/* adapted from strict_strtoul() in 2.6.25 */
|
|
char *tail;
|
|
long val;
|
|
size_t len;
|
|
|
|
*res = 0;
|
|
len = strlen(buf);
|
|
if (!len)
|
|
return -EINVAL;
|
|
val = simple_strtol(buf, &tail, base);
|
|
if (tail == buf)
|
|
return -EINVAL;
|
|
if ((*tail == '\0') ||
|
|
((len == (size_t)(tail - buf) + 1) && (*tail == '\n'))) {
|
|
*res = val;
|
|
return 0;
|
|
}
|
|
|
|
return -EINVAL;
|
|
}
|
|
#endif
|
|
|
|
#else /* < 2.6.25 */
|
|
|
|
#endif /* < 2.6.25 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,26) )
|
|
#ifndef clamp_t
|
|
#define clamp_t(type, val, min, max) ({ \
|
|
type __val = (val); \
|
|
type __min = (min); \
|
|
type __max = (max); \
|
|
__val = __val < __min ? __min : __val; \
|
|
__val > __max ? __max : __val; })
|
|
#endif /* clamp_t */
|
|
#undef kzalloc_node
|
|
#define kzalloc_node(_size, _flags, _node) kzalloc(_size, _flags)
|
|
|
|
void _kc_pci_disable_link_state(struct pci_dev *dev, int state);
|
|
#define pci_disable_link_state(p, s) _kc_pci_disable_link_state(p, s)
|
|
#else /* < 2.6.26 */
|
|
#define NETDEV_CAN_SET_GSO_MAX_SIZE
|
|
#ifdef HAVE_PCI_ASPM_H
|
|
#include <linux/pci-aspm.h>
|
|
#endif
|
|
#define HAVE_NETDEV_VLAN_FEATURES
|
|
#ifndef PCI_EXP_LNKCAP_ASPMS
|
|
#define PCI_EXP_LNKCAP_ASPMS 0x00000c00 /* ASPM Support */
|
|
#endif /* PCI_EXP_LNKCAP_ASPMS */
|
|
#endif /* < 2.6.26 */
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,27) )
|
|
static inline void _kc_ethtool_cmd_speed_set(struct ethtool_cmd *ep,
|
|
__u32 speed)
|
|
{
|
|
ep->speed = (__u16)speed;
|
|
/* ep->speed_hi = (__u16)(speed >> 16); */
|
|
}
|
|
#define ethtool_cmd_speed_set _kc_ethtool_cmd_speed_set
|
|
|
|
static inline __u32 _kc_ethtool_cmd_speed(struct ethtool_cmd *ep)
|
|
{
|
|
/* no speed_hi before 2.6.27, and probably no need for it yet */
|
|
return (__u32)ep->speed;
|
|
}
|
|
#define ethtool_cmd_speed _kc_ethtool_cmd_speed
|
|
|
|
#if ( LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,15) )
|
|
#if ((LINUX_VERSION_CODE < KERNEL_VERSION(2,6,23)) && defined(CONFIG_PM))
|
|
#define ANCIENT_PM 1
|
|
#elif ((LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,23)) && \
|
|
(LINUX_VERSION_CODE < KERNEL_VERSION(2,6,26)) && \
|
|
defined(CONFIG_PM_SLEEP))
|
|
#define NEWER_PM 1
|
|
#endif
|
|
#if defined(ANCIENT_PM) || defined(NEWER_PM)
|
|
#undef device_set_wakeup_enable
|
|
#define device_set_wakeup_enable(dev, val) \
|
|
do { \
|
|
u16 pmc = 0; \
|
|
int pm = pci_find_capability(adapter->pdev, PCI_CAP_ID_PM); \
|
|
if (pm) { \
|
|
pci_read_config_word(adapter->pdev, pm + PCI_PM_PMC, \
|
|
&pmc); \
|
|
} \
|
|
(dev)->power.can_wakeup = !!(pmc >> 11); \
|
|
(dev)->power.should_wakeup = (val && (pmc >> 11)); \
|
|
} while (0)
|
|
#endif /* 2.6.15-2.6.22 and CONFIG_PM or 2.6.23-2.6.25 and CONFIG_PM_SLEEP */
|
|
#endif /* 2.6.15 through 2.6.27 */
|
|
#ifndef netif_napi_del
|
|
#define netif_napi_del(_a) do {} while (0)
|
|
#ifdef NAPI
|
|
#ifdef CONFIG_NETPOLL
|
|
#undef netif_napi_del
|
|
#define netif_napi_del(_a) list_del(&(_a)->dev_list);
|
|
#endif
|
|
#endif
|
|
#endif /* netif_napi_del */
|
|
#ifdef dma_mapping_error
|
|
#undef dma_mapping_error
|
|
#endif
|
|
#define dma_mapping_error(dev, dma_addr) pci_dma_mapping_error(dma_addr)
|
|
|
|
#ifndef DMA_ATTR_WEAK_ORDERING
|
|
#define DMA_ATTR_WEAK_ORDERING 0
|
|
#endif
|
|
|
|
#ifdef HAVE_TX_MQ
|
|
void _kc_netif_tx_stop_all_queues(struct net_device *);
|
|
void _kc_netif_tx_wake_all_queues(struct net_device *);
|
|
void _kc_netif_tx_start_all_queues(struct net_device *);
|
|
#define netif_tx_stop_all_queues(a) _kc_netif_tx_stop_all_queues(a)
|
|
#define netif_tx_wake_all_queues(a) _kc_netif_tx_wake_all_queues(a)
|
|
#define netif_tx_start_all_queues(a) _kc_netif_tx_start_all_queues(a)
|
|
#undef netif_stop_subqueue
|
|
#define netif_stop_subqueue(_ndev,_qi) do { \
|
|
if (netif_is_multiqueue((_ndev))) \
|
|
netif_stop_subqueue((_ndev), (_qi)); \
|
|
else \
|
|
netif_stop_queue((_ndev)); \
|
|
} while (0)
|
|
#undef netif_start_subqueue
|
|
#define netif_start_subqueue(_ndev,_qi) do { \
|
|
if (netif_is_multiqueue((_ndev))) \
|
|
netif_start_subqueue((_ndev), (_qi)); \
|
|
else \
|
|
netif_start_queue((_ndev)); \
|
|
} while (0)
|
|
#else /* HAVE_TX_MQ */
|
|
#define netif_tx_stop_all_queues(a) netif_stop_queue(a)
|
|
#define netif_tx_wake_all_queues(a) netif_wake_queue(a)
|
|
#if ( LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,12) )
|
|
#define netif_tx_start_all_queues(a) netif_start_queue(a)
|
|
#else
|
|
#define netif_tx_start_all_queues(a) do {} while (0)
|
|
#endif
|
|
#define netif_stop_subqueue(_ndev,_qi) netif_stop_queue((_ndev))
|
|
#define netif_start_subqueue(_ndev,_qi) netif_start_queue((_ndev))
|
|
#endif /* HAVE_TX_MQ */
|
|
#ifndef NETIF_F_MULTI_QUEUE
|
|
#define NETIF_F_MULTI_QUEUE 0
|
|
#define netif_is_multiqueue(a) 0
|
|
#define netif_wake_subqueue(a, b)
|
|
#endif /* NETIF_F_MULTI_QUEUE */
|
|
|
|
#ifndef __WARN_printf
|
|
void __kc_warn_slowpath(const char *file, const int line,
|
|
const char *fmt, ...) __attribute__((format(printf, 3, 4)));
|
|
#define __WARN_printf(arg...) __kc_warn_slowpath(__FILE__, __LINE__, arg)
|
|
#endif /* __WARN_printf */
|
|
|
|
#ifndef WARN
|
|
#define WARN(condition, format...) ({ \
|
|
int __ret_warn_on = !!(condition); \
|
|
if (unlikely(__ret_warn_on)) \
|
|
__WARN_printf(format); \
|
|
unlikely(__ret_warn_on); \
|
|
})
|
|
#endif /* WARN */
|
|
#undef HAVE_IXGBE_DEBUG_FS
|
|
#undef HAVE_IGB_DEBUG_FS
|
|
#else /* < 2.6.27 */
|
|
#define ethtool_cmd_speed_set _kc_ethtool_cmd_speed_set
|
|
static inline void _kc_ethtool_cmd_speed_set(struct ethtool_cmd *ep,
|
|
__u32 speed)
|
|
{
|
|
ep->speed = (__u16)(speed & 0xFFFF);
|
|
ep->speed_hi = (__u16)(speed >> 16);
|
|
}
|
|
#define HAVE_TX_MQ
|
|
#define HAVE_NETDEV_SELECT_QUEUE
|
|
#ifdef CONFIG_DEBUG_FS
|
|
#define HAVE_IXGBE_DEBUG_FS
|
|
#define HAVE_IGB_DEBUG_FS
|
|
#endif /* CONFIG_DEBUG_FS */
|
|
#endif /* < 2.6.27 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,28) )
|
|
#define pci_ioremap_bar(pdev, bar) ioremap(pci_resource_start(pdev, bar), \
|
|
pci_resource_len(pdev, bar))
|
|
#define pci_wake_from_d3 _kc_pci_wake_from_d3
|
|
#define pci_prepare_to_sleep _kc_pci_prepare_to_sleep
|
|
int _kc_pci_wake_from_d3(struct pci_dev *dev, bool enable);
|
|
int _kc_pci_prepare_to_sleep(struct pci_dev *dev);
|
|
#define netdev_alloc_page(a) alloc_page(GFP_ATOMIC)
|
|
#ifndef __skb_queue_head_init
|
|
static inline void __kc_skb_queue_head_init(struct sk_buff_head *list)
|
|
{
|
|
list->prev = list->next = (struct sk_buff *)list;
|
|
list->qlen = 0;
|
|
}
|
|
#define __skb_queue_head_init(_q) __kc_skb_queue_head_init(_q)
|
|
#endif
|
|
|
|
#define PCI_EXP_DEVCAP2 36 /* Device Capabilities 2 */
|
|
#define PCI_EXP_DEVCTL2 40 /* Device Control 2 */
|
|
|
|
#define PCI_EXP_DEVCAP_FLR 0x10000000 /* Function Level Reset */
|
|
#define PCI_EXP_DEVCTL_BCR_FLR 0x8000 /* Bridge Configuration Retry / FLR */
|
|
|
|
#endif /* < 2.6.28 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,29) )
|
|
#ifndef swap
|
|
#define swap(a, b) \
|
|
do { typeof(a) __tmp = (a); (a) = (b); (b) = __tmp; } while (0)
|
|
#endif
|
|
#define pci_request_selected_regions_exclusive(pdev, bars, name) \
|
|
pci_request_selected_regions(pdev, bars, name)
|
|
#ifndef CONFIG_NR_CPUS
|
|
#define CONFIG_NR_CPUS 1
|
|
#endif /* CONFIG_NR_CPUS */
|
|
#ifndef pcie_aspm_enabled
|
|
#define pcie_aspm_enabled() (1)
|
|
#endif /* pcie_aspm_enabled */
|
|
|
|
#define PCI_EXP_SLTSTA_PDS 0x0040 /* Presence Detect State */
|
|
|
|
#ifndef PCI_EXP_LNKSTA_CLS
|
|
#define PCI_EXP_LNKSTA_CLS 0x000f /* Current Link Speed */
|
|
#endif
|
|
#ifndef PCI_EXP_LNKSTA_NLW
|
|
#define PCI_EXP_LNKSTA_NLW 0x03f0 /* Negotiated Link Width */
|
|
#endif
|
|
|
|
#ifndef pci_clear_master
|
|
void _kc_pci_clear_master(struct pci_dev *dev);
|
|
#define pci_clear_master(dev) _kc_pci_clear_master(dev)
|
|
#endif
|
|
|
|
#ifndef PCI_EXP_LNKCTL_ASPMC
|
|
#define PCI_EXP_LNKCTL_ASPMC 0x0003 /* ASPM Control */
|
|
#endif
|
|
|
|
#ifndef PCI_EXP_LNKCAP_MLW
|
|
#define PCI_EXP_LNKCAP_MLW 0x000003f0 /* Maximum Link Width */
|
|
#endif
|
|
|
|
#else /* < 2.6.29 */
|
|
#ifndef HAVE_NET_DEVICE_OPS
|
|
#define HAVE_NET_DEVICE_OPS
|
|
#endif
|
|
#ifdef CONFIG_DCB
|
|
#define HAVE_PFC_MODE_ENABLE
|
|
#endif /* CONFIG_DCB */
|
|
#endif /* < 2.6.29 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,30) )
|
|
#define NO_PTP_SUPPORT
|
|
#define skb_rx_queue_recorded(a) false
|
|
#define skb_get_rx_queue(a) 0
|
|
#define skb_record_rx_queue(a, b) do {} while (0)
|
|
#define skb_tx_hash(n, s) ___kc_skb_tx_hash((n), (s), (n)->real_num_tx_queues)
|
|
#ifndef CONFIG_PCI_IOV
|
|
#undef pci_enable_sriov
|
|
#define pci_enable_sriov(a, b) -ENOTSUPP
|
|
#undef pci_disable_sriov
|
|
#define pci_disable_sriov(a) do {} while (0)
|
|
#endif /* CONFIG_PCI_IOV */
|
|
#ifndef pr_cont
|
|
#define pr_cont(fmt, ...) \
|
|
printk(KERN_CONT fmt, ##__VA_ARGS__)
|
|
#endif /* pr_cont */
|
|
static inline void _kc_synchronize_irq(unsigned int a)
|
|
{
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,5,28) )
|
|
synchronize_irq();
|
|
#else /* < 2.5.28 */
|
|
synchronize_irq(a);
|
|
#endif /* < 2.5.28 */
|
|
}
|
|
#undef synchronize_irq
|
|
#define synchronize_irq(a) _kc_synchronize_irq(a)
|
|
|
|
#define PCI_EXP_LNKCTL2 48 /* Link Control 2 */
|
|
|
|
#ifdef nr_cpus_node
|
|
#undef nr_cpus_node
|
|
#define nr_cpus_node(node) cpumask_weight(cpumask_of_node(node))
|
|
#endif
|
|
|
|
#if (RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(5,5))
|
|
#define HAVE_PCI_DEV_IS_VIRTFN_BIT
|
|
#endif /* RHEL >= 5.5 */
|
|
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(5,5)))
|
|
static inline bool pci_is_root_bus(struct pci_bus *pbus)
|
|
{
|
|
return !(pbus->parent);
|
|
}
|
|
#endif
|
|
|
|
#else /* < 2.6.30 */
|
|
#define HAVE_ASPM_QUIRKS
|
|
#define HAVE_PCI_DEV_IS_VIRTFN_BIT
|
|
#endif /* < 2.6.30 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,31) )
|
|
#define ETH_P_1588 0x88F7
|
|
#define ETH_P_FIP 0x8914
|
|
#ifndef netdev_uc_count
|
|
#define netdev_uc_count(dev) ((dev)->uc_count)
|
|
#endif
|
|
#ifndef netdev_for_each_uc_addr
|
|
#define netdev_for_each_uc_addr(uclist, dev) \
|
|
for (uclist = dev->uc_list; uclist; uclist = uclist->next)
|
|
#endif
|
|
#ifndef PORT_OTHER
|
|
#define PORT_OTHER 0xff
|
|
#endif
|
|
#ifndef MDIO_PHY_ID_PRTAD
|
|
#define MDIO_PHY_ID_PRTAD 0x03e0
|
|
#endif
|
|
#ifndef MDIO_PHY_ID_DEVAD
|
|
#define MDIO_PHY_ID_DEVAD 0x001f
|
|
#endif
|
|
#ifndef skb_dst
|
|
#define skb_dst(s) ((s)->dst)
|
|
#endif
|
|
|
|
#ifndef SUPPORTED_1000baseKX_Full
|
|
#define SUPPORTED_1000baseKX_Full BIT(17)
|
|
#endif
|
|
#ifndef SUPPORTED_10000baseKX4_Full
|
|
#define SUPPORTED_10000baseKX4_Full BIT(18)
|
|
#endif
|
|
#ifndef SUPPORTED_10000baseKR_Full
|
|
#define SUPPORTED_10000baseKR_Full BIT(19)
|
|
#endif
|
|
|
|
#ifndef ADVERTISED_1000baseKX_Full
|
|
#define ADVERTISED_1000baseKX_Full BIT(17)
|
|
#endif
|
|
#ifndef ADVERTISED_10000baseKX4_Full
|
|
#define ADVERTISED_10000baseKX4_Full BIT(18)
|
|
#endif
|
|
#ifndef ADVERTISED_10000baseKR_Full
|
|
#define ADVERTISED_10000baseKR_Full BIT(19)
|
|
#endif
|
|
|
|
static inline unsigned long dev_trans_start(struct net_device *dev)
|
|
{
|
|
return dev->trans_start;
|
|
}
|
|
#else /* < 2.6.31 */
|
|
#ifndef HAVE_NETDEV_STORAGE_ADDRESS
|
|
#define HAVE_NETDEV_STORAGE_ADDRESS
|
|
#endif
|
|
#ifndef HAVE_NETDEV_HW_ADDR
|
|
#define HAVE_NETDEV_HW_ADDR
|
|
#endif
|
|
#ifndef HAVE_TRANS_START_IN_QUEUE
|
|
#define HAVE_TRANS_START_IN_QUEUE
|
|
#endif
|
|
#ifndef HAVE_INCLUDE_LINUX_MDIO_H
|
|
#define HAVE_INCLUDE_LINUX_MDIO_H
|
|
#endif
|
|
#include <linux/mdio.h>
|
|
#endif /* < 2.6.31 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,32) )
|
|
#undef netdev_tx_t
|
|
#define netdev_tx_t int
|
|
#if defined(CONFIG_FCOE) || defined(CONFIG_FCOE_MODULE)
|
|
#ifndef NETIF_F_FCOE_MTU
|
|
#define NETIF_F_FCOE_MTU BIT(26)
|
|
#endif
|
|
#endif /* CONFIG_FCOE || CONFIG_FCOE_MODULE */
|
|
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0) )
|
|
static inline int _kc_pm_runtime_get_sync()
|
|
{
|
|
return 1;
|
|
}
|
|
#define pm_runtime_get_sync(dev) _kc_pm_runtime_get_sync()
|
|
#else /* 2.6.0 => 2.6.32 */
|
|
static inline int _kc_pm_runtime_get_sync(struct device __always_unused *dev)
|
|
{
|
|
return 1;
|
|
}
|
|
#ifndef pm_runtime_get_sync
|
|
#define pm_runtime_get_sync(dev) _kc_pm_runtime_get_sync(dev)
|
|
#endif
|
|
#endif /* 2.6.0 => 2.6.32 */
|
|
#ifndef pm_runtime_put
|
|
#define pm_runtime_put(dev) do {} while (0)
|
|
#endif
|
|
#ifndef pm_runtime_put_sync
|
|
#define pm_runtime_put_sync(dev) do {} while (0)
|
|
#endif
|
|
#ifndef pm_runtime_resume
|
|
#define pm_runtime_resume(dev) do {} while (0)
|
|
#endif
|
|
#ifndef pm_schedule_suspend
|
|
#define pm_schedule_suspend(dev, t) do {} while (0)
|
|
#endif
|
|
#ifndef pm_runtime_set_suspended
|
|
#define pm_runtime_set_suspended(dev) do {} while (0)
|
|
#endif
|
|
#ifndef pm_runtime_disable
|
|
#define pm_runtime_disable(dev) do {} while (0)
|
|
#endif
|
|
#ifndef pm_runtime_put_noidle
|
|
#define pm_runtime_put_noidle(dev) do {} while (0)
|
|
#endif
|
|
#ifndef pm_runtime_set_active
|
|
#define pm_runtime_set_active(dev) do {} while (0)
|
|
#endif
|
|
#ifndef pm_runtime_enable
|
|
#define pm_runtime_enable(dev) do {} while (0)
|
|
#endif
|
|
#ifndef pm_runtime_get_noresume
|
|
#define pm_runtime_get_noresume(dev) do {} while (0)
|
|
#endif
|
|
#else /* < 2.6.32 */
|
|
#if (RHEL_RELEASE_CODE && \
|
|
(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,2)) && \
|
|
(RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,0)))
|
|
#define HAVE_RHEL6_NET_DEVICE_EXTENDED
|
|
#endif /* RHEL >= 6.2 && RHEL < 7.0 */
|
|
#if (RHEL_RELEASE_CODE && \
|
|
(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,6)) && \
|
|
(RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,0)))
|
|
#define HAVE_RHEL6_NET_DEVICE_OPS_EXT
|
|
#define HAVE_NDO_SET_FEATURES
|
|
#endif /* RHEL >= 6.6 && RHEL < 7.0 */
|
|
#if defined(CONFIG_FCOE) || defined(CONFIG_FCOE_MODULE)
|
|
#ifndef HAVE_NETDEV_OPS_FCOE_ENABLE
|
|
#define HAVE_NETDEV_OPS_FCOE_ENABLE
|
|
#endif
|
|
#endif /* CONFIG_FCOE || CONFIG_FCOE_MODULE */
|
|
#ifdef CONFIG_DCB
|
|
#ifndef HAVE_DCBNL_OPS_GETAPP
|
|
#define HAVE_DCBNL_OPS_GETAPP
|
|
#endif
|
|
#endif /* CONFIG_DCB */
|
|
#include <linux/pm_runtime.h>
|
|
/* IOV bad DMA target work arounds require at least this kernel rev support */
|
|
#define HAVE_PCIE_TYPE
|
|
#endif /* < 2.6.32 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,33) )
|
|
#ifndef pci_pcie_cap
|
|
#define pci_pcie_cap(pdev) pci_find_capability(pdev, PCI_CAP_ID_EXP)
|
|
#endif
|
|
#ifndef IPV4_FLOW
|
|
#define IPV4_FLOW 0x10
|
|
#endif /* IPV4_FLOW */
|
|
#ifndef IPV6_FLOW
|
|
#define IPV6_FLOW 0x11
|
|
#endif /* IPV6_FLOW */
|
|
/* Features back-ported to RHEL6 or SLES11 SP1 after 2.6.32 */
|
|
#if ( (RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,0)) || \
|
|
(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(11,1,0)) )
|
|
#if defined(CONFIG_FCOE) || defined(CONFIG_FCOE_MODULE)
|
|
#ifndef HAVE_NETDEV_OPS_FCOE_GETWWN
|
|
#define HAVE_NETDEV_OPS_FCOE_GETWWN
|
|
#endif
|
|
#endif /* CONFIG_FCOE || CONFIG_FCOE_MODULE */
|
|
#endif /* RHEL6 or SLES11 SP1 */
|
|
#ifndef __percpu
|
|
#define __percpu
|
|
#endif /* __percpu */
|
|
|
|
#ifndef PORT_DA
|
|
#define PORT_DA PORT_OTHER
|
|
#endif /* PORT_DA */
|
|
#ifndef PORT_NONE
|
|
#define PORT_NONE PORT_OTHER
|
|
#endif
|
|
|
|
#if ((RHEL_RELEASE_CODE && \
|
|
(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,3)) && \
|
|
(RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,0))))
|
|
#if !defined(CONFIG_X86_32) && !defined(CONFIG_NEED_DMA_MAP_STATE)
|
|
#undef DEFINE_DMA_UNMAP_ADDR
|
|
#define DEFINE_DMA_UNMAP_ADDR(ADDR_NAME) dma_addr_t ADDR_NAME
|
|
#undef DEFINE_DMA_UNMAP_LEN
|
|
#define DEFINE_DMA_UNMAP_LEN(LEN_NAME) __u32 LEN_NAME
|
|
#undef dma_unmap_addr
|
|
#define dma_unmap_addr(PTR, ADDR_NAME) ((PTR)->ADDR_NAME)
|
|
#undef dma_unmap_addr_set
|
|
#define dma_unmap_addr_set(PTR, ADDR_NAME, VAL) (((PTR)->ADDR_NAME) = (VAL))
|
|
#undef dma_unmap_len
|
|
#define dma_unmap_len(PTR, LEN_NAME) ((PTR)->LEN_NAME)
|
|
#undef dma_unmap_len_set
|
|
#define dma_unmap_len_set(PTR, LEN_NAME, VAL) (((PTR)->LEN_NAME) = (VAL))
|
|
#endif /* CONFIG_X86_64 && !CONFIG_NEED_DMA_MAP_STATE */
|
|
#endif /* RHEL_RELEASE_CODE */
|
|
|
|
#if (!(RHEL_RELEASE_CODE && \
|
|
(((RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(5,8)) && \
|
|
(RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(6,0))) || \
|
|
((RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,1)) && \
|
|
(RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,0))))))
|
|
static inline bool pci_is_pcie(struct pci_dev *dev)
|
|
{
|
|
return !!pci_pcie_cap(dev);
|
|
}
|
|
#endif /* RHEL_RELEASE_CODE */
|
|
|
|
#if (!(RHEL_RELEASE_CODE && \
|
|
(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,2))))
|
|
#define sk_tx_queue_get(_sk) (-1)
|
|
#define sk_tx_queue_set(_sk, _tx_queue) do {} while(0)
|
|
#endif /* !(RHEL >= 6.2) */
|
|
|
|
#if (RHEL_RELEASE_CODE && \
|
|
(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,4)) && \
|
|
(RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,0)))
|
|
#define HAVE_RHEL6_ETHTOOL_OPS_EXT_STRUCT
|
|
#define HAVE_ETHTOOL_GRXFHINDIR_SIZE
|
|
#define HAVE_ETHTOOL_SET_PHYS_ID
|
|
#define HAVE_ETHTOOL_GET_TS_INFO
|
|
#if (RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(6,5))
|
|
#define HAVE_ETHTOOL_GSRSSH
|
|
#define HAVE_RHEL6_SRIOV_CONFIGURE
|
|
#define HAVE_RXFH_NONCONST
|
|
#endif /* RHEL > 6.5 */
|
|
#endif /* RHEL >= 6.4 && RHEL < 7.0 */
|
|
|
|
#else /* < 2.6.33 */
|
|
#if defined(CONFIG_FCOE) || defined(CONFIG_FCOE_MODULE)
|
|
#ifndef HAVE_NETDEV_OPS_FCOE_GETWWN
|
|
#define HAVE_NETDEV_OPS_FCOE_GETWWN
|
|
#endif
|
|
#endif /* CONFIG_FCOE || CONFIG_FCOE_MODULE */
|
|
#endif /* < 2.6.33 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,34) )
|
|
#if (RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(6,0))
|
|
#ifndef pci_num_vf
|
|
#define pci_num_vf(pdev) _kc_pci_num_vf(pdev)
|
|
int _kc_pci_num_vf(struct pci_dev *dev);
|
|
#endif
|
|
#endif /* RHEL_RELEASE_CODE */
|
|
|
|
#ifndef dev_is_pci
|
|
#define dev_is_pci(d) ((d)->bus == &pci_bus_type)
|
|
#endif
|
|
|
|
#ifndef ETH_FLAG_NTUPLE
|
|
#define ETH_FLAG_NTUPLE NETIF_F_NTUPLE
|
|
#endif
|
|
|
|
#ifndef netdev_mc_count
|
|
#define netdev_mc_count(dev) ((dev)->mc_count)
|
|
#endif
|
|
#ifndef netdev_mc_empty
|
|
#define netdev_mc_empty(dev) (netdev_mc_count(dev) == 0)
|
|
#endif
|
|
#ifndef netdev_for_each_mc_addr
|
|
#define netdev_for_each_mc_addr(mclist, dev) \
|
|
for (mclist = dev->mc_list; mclist; mclist = mclist->next)
|
|
#endif
|
|
#ifndef netdev_uc_count
|
|
#define netdev_uc_count(dev) ((dev)->uc.count)
|
|
#endif
|
|
#ifndef netdev_uc_empty
|
|
#define netdev_uc_empty(dev) (netdev_uc_count(dev) == 0)
|
|
#endif
|
|
#ifndef netdev_for_each_uc_addr
|
|
#define netdev_for_each_uc_addr(ha, dev) \
|
|
list_for_each_entry(ha, &dev->uc.list, list)
|
|
#endif
|
|
#ifndef dma_set_coherent_mask
|
|
#define dma_set_coherent_mask(dev,mask) \
|
|
pci_set_consistent_dma_mask(to_pci_dev(dev),(mask))
|
|
#endif
|
|
#ifndef pci_dev_run_wake
|
|
#define pci_dev_run_wake(pdev) (0)
|
|
#endif
|
|
|
|
/* netdev logging taken from include/linux/netdevice.h */
|
|
#ifndef netdev_name
|
|
static inline const char *_kc_netdev_name(const struct net_device *dev)
|
|
{
|
|
if (dev->reg_state != NETREG_REGISTERED)
|
|
return "(unregistered net_device)";
|
|
return dev->name;
|
|
}
|
|
#define netdev_name(netdev) _kc_netdev_name(netdev)
|
|
#endif /* netdev_name */
|
|
|
|
#undef netdev_printk
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0) )
|
|
#define netdev_printk(level, netdev, format, args...) \
|
|
do { \
|
|
struct pci_dev *pdev = _kc_netdev_to_pdev(netdev); \
|
|
printk(level "%s: " format, pci_name(pdev), ##args); \
|
|
} while(0)
|
|
#elif ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,21) )
|
|
#define netdev_printk(level, netdev, format, args...) \
|
|
do { \
|
|
struct pci_dev *pdev = _kc_netdev_to_pdev(netdev); \
|
|
struct device *dev = pci_dev_to_dev(pdev); \
|
|
dev_printk(level, dev, "%s: " format, \
|
|
netdev_name(netdev), ##args); \
|
|
} while(0)
|
|
#else /* 2.6.21 => 2.6.34 */
|
|
#define netdev_printk(level, netdev, format, args...) \
|
|
dev_printk(level, (netdev)->dev.parent, \
|
|
"%s: " format, \
|
|
netdev_name(netdev), ##args)
|
|
#endif /* <2.6.0 <2.6.21 <2.6.34 */
|
|
#undef netdev_emerg
|
|
#define netdev_emerg(dev, format, args...) \
|
|
netdev_printk(KERN_EMERG, dev, format, ##args)
|
|
#undef netdev_alert
|
|
#define netdev_alert(dev, format, args...) \
|
|
netdev_printk(KERN_ALERT, dev, format, ##args)
|
|
#undef netdev_crit
|
|
#define netdev_crit(dev, format, args...) \
|
|
netdev_printk(KERN_CRIT, dev, format, ##args)
|
|
#undef netdev_err
|
|
#define netdev_err(dev, format, args...) \
|
|
netdev_printk(KERN_ERR, dev, format, ##args)
|
|
#undef netdev_warn
|
|
#define netdev_warn(dev, format, args...) \
|
|
netdev_printk(KERN_WARNING, dev, format, ##args)
|
|
#undef netdev_notice
|
|
#define netdev_notice(dev, format, args...) \
|
|
netdev_printk(KERN_NOTICE, dev, format, ##args)
|
|
#undef netdev_info
|
|
#define netdev_info(dev, format, args...) \
|
|
netdev_printk(KERN_INFO, dev, format, ##args)
|
|
#undef netdev_dbg
|
|
#if defined(DEBUG)
|
|
#define netdev_dbg(__dev, format, args...) \
|
|
netdev_printk(KERN_DEBUG, __dev, format, ##args)
|
|
#elif defined(CONFIG_DYNAMIC_DEBUG)
|
|
#define netdev_dbg(__dev, format, args...) \
|
|
do { \
|
|
dynamic_dev_dbg((__dev)->dev.parent, "%s: " format, \
|
|
netdev_name(__dev), ##args); \
|
|
} while (0)
|
|
#else /* DEBUG */
|
|
#define netdev_dbg(__dev, format, args...) \
|
|
({ \
|
|
if (0) \
|
|
netdev_printk(KERN_DEBUG, __dev, format, ##args); \
|
|
0; \
|
|
})
|
|
#endif /* DEBUG */
|
|
|
|
#undef netif_printk
|
|
#define netif_printk(priv, type, level, dev, fmt, args...) \
|
|
do { \
|
|
if (netif_msg_##type(priv)) \
|
|
netdev_printk(level, (dev), fmt, ##args); \
|
|
} while (0)
|
|
|
|
#undef netif_emerg
|
|
#define netif_emerg(priv, type, dev, fmt, args...) \
|
|
netif_level(emerg, priv, type, dev, fmt, ##args)
|
|
#undef netif_alert
|
|
#define netif_alert(priv, type, dev, fmt, args...) \
|
|
netif_level(alert, priv, type, dev, fmt, ##args)
|
|
#undef netif_crit
|
|
#define netif_crit(priv, type, dev, fmt, args...) \
|
|
netif_level(crit, priv, type, dev, fmt, ##args)
|
|
#undef netif_err
|
|
#define netif_err(priv, type, dev, fmt, args...) \
|
|
netif_level(err, priv, type, dev, fmt, ##args)
|
|
#undef netif_warn
|
|
#define netif_warn(priv, type, dev, fmt, args...) \
|
|
netif_level(warn, priv, type, dev, fmt, ##args)
|
|
#undef netif_notice
|
|
#define netif_notice(priv, type, dev, fmt, args...) \
|
|
netif_level(notice, priv, type, dev, fmt, ##args)
|
|
#undef netif_info
|
|
#define netif_info(priv, type, dev, fmt, args...) \
|
|
netif_level(info, priv, type, dev, fmt, ##args)
|
|
#undef netif_dbg
|
|
#define netif_dbg(priv, type, dev, fmt, args...) \
|
|
netif_level(dbg, priv, type, dev, fmt, ##args)
|
|
|
|
#ifdef SET_SYSTEM_SLEEP_PM_OPS
|
|
#define HAVE_SYSTEM_SLEEP_PM_OPS
|
|
#endif
|
|
|
|
#ifndef for_each_set_bit
|
|
#define for_each_set_bit(bit, addr, size) \
|
|
for ((bit) = find_first_bit((addr), (size)); \
|
|
(bit) < (size); \
|
|
(bit) = find_next_bit((addr), (size), (bit) + 1))
|
|
#endif /* for_each_set_bit */
|
|
|
|
#ifndef DEFINE_DMA_UNMAP_ADDR
|
|
#define DEFINE_DMA_UNMAP_ADDR DECLARE_PCI_UNMAP_ADDR
|
|
#define DEFINE_DMA_UNMAP_LEN DECLARE_PCI_UNMAP_LEN
|
|
#define dma_unmap_addr pci_unmap_addr
|
|
#define dma_unmap_addr_set pci_unmap_addr_set
|
|
#define dma_unmap_len pci_unmap_len
|
|
#define dma_unmap_len_set pci_unmap_len_set
|
|
#endif /* DEFINE_DMA_UNMAP_ADDR */
|
|
|
|
#if (RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(6,3))
|
|
#ifdef IGB_HWMON
|
|
#ifdef CONFIG_DEBUG_LOCK_ALLOC
|
|
#define sysfs_attr_init(attr) \
|
|
do { \
|
|
static struct lock_class_key __key; \
|
|
(attr)->key = &__key; \
|
|
} while (0)
|
|
#else
|
|
#define sysfs_attr_init(attr) do {} while (0)
|
|
#endif /* CONFIG_DEBUG_LOCK_ALLOC */
|
|
#endif /* IGB_HWMON */
|
|
#endif /* RHEL_RELEASE_CODE */
|
|
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0) )
|
|
static inline bool _kc_pm_runtime_suspended()
|
|
{
|
|
return false;
|
|
}
|
|
#define pm_runtime_suspended(dev) _kc_pm_runtime_suspended()
|
|
#else /* 2.6.0 => 2.6.34 */
|
|
static inline bool _kc_pm_runtime_suspended(struct device __always_unused *dev)
|
|
{
|
|
return false;
|
|
}
|
|
#ifndef pm_runtime_suspended
|
|
#define pm_runtime_suspended(dev) _kc_pm_runtime_suspended(dev)
|
|
#endif
|
|
#endif /* 2.6.0 => 2.6.34 */
|
|
|
|
#ifndef pci_bus_speed
|
|
/* override pci_bus_speed introduced in 2.6.19 with an expanded enum type */
|
|
enum _kc_pci_bus_speed {
|
|
_KC_PCIE_SPEED_2_5GT = 0x14,
|
|
_KC_PCIE_SPEED_5_0GT = 0x15,
|
|
_KC_PCIE_SPEED_8_0GT = 0x16,
|
|
_KC_PCI_SPEED_UNKNOWN = 0xff,
|
|
};
|
|
#define pci_bus_speed _kc_pci_bus_speed
|
|
#define PCIE_SPEED_2_5GT _KC_PCIE_SPEED_2_5GT
|
|
#define PCIE_SPEED_5_0GT _KC_PCIE_SPEED_5_0GT
|
|
#define PCIE_SPEED_8_0GT _KC_PCIE_SPEED_8_0GT
|
|
#define PCI_SPEED_UNKNOWN _KC_PCI_SPEED_UNKNOWN
|
|
#endif /* pci_bus_speed */
|
|
|
|
#else /* < 2.6.34 */
|
|
#define HAVE_SYSTEM_SLEEP_PM_OPS
|
|
#ifndef HAVE_SET_RX_MODE
|
|
#define HAVE_SET_RX_MODE
|
|
#endif
|
|
|
|
#endif /* < 2.6.34 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,35) )
|
|
ssize_t _kc_simple_write_to_buffer(void *to, size_t available, loff_t *ppos,
|
|
const void __user *from, size_t count);
|
|
#define simple_write_to_buffer _kc_simple_write_to_buffer
|
|
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,4)))
|
|
static inline struct pci_dev *pci_physfn(struct pci_dev *dev)
|
|
{
|
|
#ifdef HAVE_PCI_DEV_IS_VIRTFN_BIT
|
|
#ifdef CONFIG_PCI_IOV
|
|
if (dev->is_virtfn)
|
|
dev = dev->physfn;
|
|
#endif /* CONFIG_PCI_IOV */
|
|
#endif /* HAVE_PCI_DEV_IS_VIRTFN_BIT */
|
|
return dev;
|
|
}
|
|
#endif /* ! RHEL >= 6.4 */
|
|
|
|
#ifndef PCI_EXP_LNKSTA_NLW_SHIFT
|
|
#define PCI_EXP_LNKSTA_NLW_SHIFT 4
|
|
#endif
|
|
|
|
#ifndef numa_node_id
|
|
#define numa_node_id() 0
|
|
#endif
|
|
#ifndef numa_mem_id
|
|
#define numa_mem_id numa_node_id
|
|
#endif
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,0)))
|
|
#ifdef HAVE_TX_MQ
|
|
#include <net/sch_generic.h>
|
|
#ifndef CONFIG_NETDEVICES_MULTIQUEUE
|
|
int _kc_netif_set_real_num_tx_queues(struct net_device *, unsigned int);
|
|
#else /* CONFIG_NETDEVICES_MULTI_QUEUE */
|
|
static inline int _kc_netif_set_real_num_tx_queues(struct net_device *dev,
|
|
unsigned int txq)
|
|
{
|
|
dev->egress_subqueue_count = txq;
|
|
return 0;
|
|
}
|
|
#endif /* CONFIG_NETDEVICES_MULTI_QUEUE */
|
|
#else /* HAVE_TX_MQ */
|
|
static inline int _kc_netif_set_real_num_tx_queues(struct net_device __always_unused *dev,
|
|
unsigned int __always_unused txq)
|
|
{
|
|
return 0;
|
|
}
|
|
#endif /* HAVE_TX_MQ */
|
|
#define netif_set_real_num_tx_queues(dev, txq) \
|
|
_kc_netif_set_real_num_tx_queues(dev, txq)
|
|
#endif /* !(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,0)) */
|
|
#ifndef ETH_FLAG_RXHASH
|
|
#define ETH_FLAG_RXHASH (1<<28)
|
|
#endif /* ETH_FLAG_RXHASH */
|
|
#if (RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,0))
|
|
#define HAVE_IRQ_AFFINITY_HINT
|
|
#endif
|
|
struct device_node;
|
|
#else /* < 2.6.35 */
|
|
#define HAVE_STRUCT_DEVICE_OF_NODE
|
|
#define HAVE_PM_QOS_REQUEST_LIST
|
|
#define HAVE_IRQ_AFFINITY_HINT
|
|
#include <linux/of.h>
|
|
#endif /* < 2.6.35 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,36) )
|
|
int _kc_ethtool_op_set_flags(struct net_device *, u32, u32);
|
|
#define ethtool_op_set_flags _kc_ethtool_op_set_flags
|
|
u32 _kc_ethtool_op_get_flags(struct net_device *);
|
|
#define ethtool_op_get_flags _kc_ethtool_op_get_flags
|
|
|
|
enum {
|
|
WQ_UNBOUND = 0,
|
|
WQ_RESCUER = 0,
|
|
};
|
|
|
|
#ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
|
|
#ifdef NET_IP_ALIGN
|
|
#undef NET_IP_ALIGN
|
|
#endif
|
|
#define NET_IP_ALIGN 0
|
|
#endif /* CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS */
|
|
|
|
#ifdef NET_SKB_PAD
|
|
#undef NET_SKB_PAD
|
|
#endif
|
|
|
|
#if (L1_CACHE_BYTES > 32)
|
|
#define NET_SKB_PAD L1_CACHE_BYTES
|
|
#else
|
|
#define NET_SKB_PAD 32
|
|
#endif
|
|
|
|
static inline struct sk_buff *_kc_netdev_alloc_skb_ip_align(struct net_device *dev,
|
|
unsigned int length)
|
|
{
|
|
struct sk_buff *skb;
|
|
|
|
skb = alloc_skb(length + NET_SKB_PAD + NET_IP_ALIGN, GFP_ATOMIC);
|
|
if (skb) {
|
|
#if (NET_IP_ALIGN + NET_SKB_PAD)
|
|
skb_reserve(skb, NET_IP_ALIGN + NET_SKB_PAD);
|
|
#endif
|
|
skb->dev = dev;
|
|
}
|
|
return skb;
|
|
}
|
|
|
|
#ifdef netdev_alloc_skb_ip_align
|
|
#undef netdev_alloc_skb_ip_align
|
|
#endif
|
|
#define netdev_alloc_skb_ip_align(n, l) _kc_netdev_alloc_skb_ip_align(n, l)
|
|
|
|
#undef netif_level
|
|
#define netif_level(level, priv, type, dev, fmt, args...) \
|
|
do { \
|
|
if (netif_msg_##type(priv)) \
|
|
netdev_##level(dev, fmt, ##args); \
|
|
} while (0)
|
|
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,3)))
|
|
#undef usleep_range
|
|
#define usleep_range(min, max) msleep(DIV_ROUND_UP(min, 1000))
|
|
#endif
|
|
|
|
#define u64_stats_update_begin(a) do { } while(0)
|
|
#define u64_stats_update_end(a) do { } while(0)
|
|
#define u64_stats_fetch_begin(a) do { } while(0)
|
|
#define u64_stats_fetch_retry_bh(a,b) (0)
|
|
#define u64_stats_fetch_begin_bh(a) (0)
|
|
|
|
#if (RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,1))
|
|
#define HAVE_8021P_SUPPORT
|
|
#endif
|
|
|
|
/* RHEL6.4 and SLES11sp2 backported skb_tx_timestamp */
|
|
#if (!(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,4)) && \
|
|
!(SLE_VERSION_CODE >= SLE_VERSION(11,2,0)))
|
|
static inline void skb_tx_timestamp(struct sk_buff __always_unused *skb)
|
|
{
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
#else /* < 2.6.36 */
|
|
|
|
#define HAVE_PM_QOS_REQUEST_ACTIVE
|
|
#define HAVE_8021P_SUPPORT
|
|
#define HAVE_NDO_GET_STATS64
|
|
#endif /* < 2.6.36 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,37) )
|
|
#define HAVE_NON_CONST_PCI_DRIVER_NAME
|
|
#ifndef netif_set_real_num_tx_queues
|
|
static inline int _kc_netif_set_real_num_tx_queues(struct net_device *dev,
|
|
unsigned int txq)
|
|
{
|
|
netif_set_real_num_tx_queues(dev, txq);
|
|
return 0;
|
|
}
|
|
#define netif_set_real_num_tx_queues(dev, txq) \
|
|
_kc_netif_set_real_num_tx_queues(dev, txq)
|
|
#endif
|
|
#ifndef netif_set_real_num_rx_queues
|
|
static inline int __kc_netif_set_real_num_rx_queues(struct net_device __always_unused *dev,
|
|
unsigned int __always_unused rxq)
|
|
{
|
|
return 0;
|
|
}
|
|
#define netif_set_real_num_rx_queues(dev, rxq) \
|
|
__kc_netif_set_real_num_rx_queues((dev), (rxq))
|
|
#endif
|
|
#ifndef ETHTOOL_RXNTUPLE_ACTION_CLEAR
|
|
#define ETHTOOL_RXNTUPLE_ACTION_CLEAR (-2)
|
|
#endif
|
|
#ifndef VLAN_N_VID
|
|
#define VLAN_N_VID VLAN_GROUP_ARRAY_LEN
|
|
#endif /* VLAN_N_VID */
|
|
#ifndef ETH_FLAG_TXVLAN
|
|
#define ETH_FLAG_TXVLAN BIT(7)
|
|
#endif /* ETH_FLAG_TXVLAN */
|
|
#ifndef ETH_FLAG_RXVLAN
|
|
#define ETH_FLAG_RXVLAN BIT(8)
|
|
#endif /* ETH_FLAG_RXVLAN */
|
|
|
|
#define WQ_MEM_RECLAIM WQ_RESCUER
|
|
|
|
static inline void _kc_skb_checksum_none_assert(struct sk_buff *skb)
|
|
{
|
|
WARN_ON(skb->ip_summed != CHECKSUM_NONE);
|
|
}
|
|
#define skb_checksum_none_assert(skb) _kc_skb_checksum_none_assert(skb)
|
|
|
|
static inline void *_kc_vzalloc_node(unsigned long size, int node)
|
|
{
|
|
void *addr = vmalloc_node(size, node);
|
|
if (addr)
|
|
memset(addr, 0, size);
|
|
return addr;
|
|
}
|
|
#define vzalloc_node(_size, _node) _kc_vzalloc_node(_size, _node)
|
|
|
|
static inline void *_kc_vzalloc(unsigned long size)
|
|
{
|
|
void *addr = vmalloc(size);
|
|
if (addr)
|
|
memset(addr, 0, size);
|
|
return addr;
|
|
}
|
|
#define vzalloc(_size) _kc_vzalloc(_size)
|
|
|
|
#if (!(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(5,7)) || \
|
|
(RHEL_RELEASE_CODE == RHEL_RELEASE_VERSION(6,0)))
|
|
static inline __be16 vlan_get_protocol(const struct sk_buff *skb)
|
|
{
|
|
if (vlan_tx_tag_present(skb) ||
|
|
skb->protocol != cpu_to_be16(ETH_P_8021Q))
|
|
return skb->protocol;
|
|
|
|
if (skb_headlen(skb) < sizeof(struct vlan_ethhdr))
|
|
return 0;
|
|
|
|
return ((struct vlan_ethhdr*)skb->data)->h_vlan_encapsulated_proto;
|
|
}
|
|
#endif /* !RHEL5.7+ || RHEL6.0 */
|
|
|
|
#ifdef HAVE_HW_TIME_STAMP
|
|
#define SKBTX_HW_TSTAMP BIT(0)
|
|
#define SKBTX_IN_PROGRESS BIT(2)
|
|
#define SKB_SHARED_TX_IS_UNION
|
|
#endif
|
|
|
|
#ifndef device_wakeup_enable
|
|
#define device_wakeup_enable(dev) device_set_wakeup_enable(dev, true)
|
|
#endif
|
|
|
|
#if ( LINUX_VERSION_CODE > KERNEL_VERSION(2,4,18) )
|
|
#ifndef HAVE_VLAN_RX_REGISTER
|
|
#define HAVE_VLAN_RX_REGISTER
|
|
#endif
|
|
#endif /* > 2.4.18 */
|
|
#endif /* < 2.6.37 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,38) )
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,22) )
|
|
#define skb_checksum_start_offset(skb) skb_transport_offset(skb)
|
|
#else /* 2.6.22 -> 2.6.37 */
|
|
static inline int _kc_skb_checksum_start_offset(const struct sk_buff *skb)
|
|
{
|
|
return skb->csum_start - skb_headroom(skb);
|
|
}
|
|
#define skb_checksum_start_offset(skb) _kc_skb_checksum_start_offset(skb)
|
|
#endif /* 2.6.22 -> 2.6.37 */
|
|
#if IS_ENABLED(CONFIG_DCB)
|
|
#ifndef IEEE_8021QAZ_MAX_TCS
|
|
#define IEEE_8021QAZ_MAX_TCS 8
|
|
#endif
|
|
#ifndef DCB_CAP_DCBX_HOST
|
|
#define DCB_CAP_DCBX_HOST 0x01
|
|
#endif
|
|
#ifndef DCB_CAP_DCBX_LLD_MANAGED
|
|
#define DCB_CAP_DCBX_LLD_MANAGED 0x02
|
|
#endif
|
|
#ifndef DCB_CAP_DCBX_VER_CEE
|
|
#define DCB_CAP_DCBX_VER_CEE 0x04
|
|
#endif
|
|
#ifndef DCB_CAP_DCBX_VER_IEEE
|
|
#define DCB_CAP_DCBX_VER_IEEE 0x08
|
|
#endif
|
|
#ifndef DCB_CAP_DCBX_STATIC
|
|
#define DCB_CAP_DCBX_STATIC 0x10
|
|
#endif
|
|
#endif /* CONFIG_DCB */
|
|
#if (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,2))
|
|
#define CONFIG_XPS
|
|
#endif /* RHEL_RELEASE_VERSION(6,2) */
|
|
#endif /* < 2.6.38 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,39) )
|
|
#ifndef TC_BITMASK
|
|
#define TC_BITMASK 15
|
|
#endif
|
|
#ifndef NETIF_F_RXCSUM
|
|
#define NETIF_F_RXCSUM BIT(29)
|
|
#endif
|
|
#ifndef skb_queue_reverse_walk_safe
|
|
#define skb_queue_reverse_walk_safe(queue, skb, tmp) \
|
|
for (skb = (queue)->prev, tmp = skb->prev; \
|
|
skb != (struct sk_buff *)(queue); \
|
|
skb = tmp, tmp = skb->prev)
|
|
#endif
|
|
|
|
#ifndef udp_csum
|
|
#define udp_csum __kc_udp_csum
|
|
static inline __wsum __kc_udp_csum(struct sk_buff *skb)
|
|
{
|
|
__wsum csum = csum_partial(skb_transport_header(skb),
|
|
sizeof(struct udphdr), skb->csum);
|
|
|
|
for (skb = skb_shinfo(skb)->frag_list; skb; skb = skb->next) {
|
|
csum = csum_add(csum, skb->csum);
|
|
}
|
|
return csum;
|
|
}
|
|
#endif /* udp_csum */
|
|
#else /* < 2.6.39 */
|
|
#if defined(CONFIG_FCOE) || defined(CONFIG_FCOE_MODULE)
|
|
#ifndef HAVE_NETDEV_OPS_FCOE_DDP_TARGET
|
|
#define HAVE_NETDEV_OPS_FCOE_DDP_TARGET
|
|
#endif
|
|
#endif /* CONFIG_FCOE || CONFIG_FCOE_MODULE */
|
|
#ifndef HAVE_MQPRIO
|
|
#define HAVE_MQPRIO
|
|
#endif
|
|
#ifndef HAVE_SETUP_TC
|
|
#define HAVE_SETUP_TC
|
|
#endif
|
|
#ifdef CONFIG_DCB
|
|
#ifndef HAVE_DCBNL_IEEE
|
|
#define HAVE_DCBNL_IEEE
|
|
#endif
|
|
#endif /* CONFIG_DCB */
|
|
#ifndef HAVE_NDO_SET_FEATURES
|
|
#define HAVE_NDO_SET_FEATURES
|
|
#endif
|
|
#define HAVE_IRQ_AFFINITY_NOTIFY
|
|
#endif /* < 2.6.39 */
|
|
|
|
/*****************************************************************************/
|
|
/* use < 2.6.40 because of a Fedora 15 kernel update where they
|
|
* updated the kernel version to 2.6.40.x and they back-ported 3.0 features
|
|
* like set_phys_id for ethtool.
|
|
*/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,40) )
|
|
#ifdef ETHTOOL_GRXRINGS
|
|
#ifndef FLOW_EXT
|
|
#define FLOW_EXT 0x80000000
|
|
union _kc_ethtool_flow_union {
|
|
struct ethtool_tcpip4_spec tcp_ip4_spec;
|
|
struct ethtool_usrip4_spec usr_ip4_spec;
|
|
__u8 hdata[60];
|
|
};
|
|
struct _kc_ethtool_flow_ext {
|
|
__be16 vlan_etype;
|
|
__be16 vlan_tci;
|
|
__be32 data[2];
|
|
};
|
|
struct _kc_ethtool_rx_flow_spec {
|
|
__u32 flow_type;
|
|
union _kc_ethtool_flow_union h_u;
|
|
struct _kc_ethtool_flow_ext h_ext;
|
|
union _kc_ethtool_flow_union m_u;
|
|
struct _kc_ethtool_flow_ext m_ext;
|
|
__u64 ring_cookie;
|
|
__u32 location;
|
|
};
|
|
#define ethtool_rx_flow_spec _kc_ethtool_rx_flow_spec
|
|
#endif /* FLOW_EXT */
|
|
#endif
|
|
|
|
#define pci_disable_link_state_locked pci_disable_link_state
|
|
|
|
#ifndef PCI_LTR_VALUE_MASK
|
|
#define PCI_LTR_VALUE_MASK 0x000003ff
|
|
#endif
|
|
#ifndef PCI_LTR_SCALE_MASK
|
|
#define PCI_LTR_SCALE_MASK 0x00001c00
|
|
#endif
|
|
#ifndef PCI_LTR_SCALE_SHIFT
|
|
#define PCI_LTR_SCALE_SHIFT 10
|
|
#endif
|
|
|
|
#else /* < 2.6.40 */
|
|
#define HAVE_ETHTOOL_SET_PHYS_ID
|
|
#endif /* < 2.6.40 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,0,0) )
|
|
#define USE_LEGACY_PM_SUPPORT
|
|
#ifndef kfree_rcu
|
|
#define kfree_rcu(_ptr, _rcu_head) do { \
|
|
void __kc_kfree_rcu(struct rcu_head *rcu_head) \
|
|
{ \
|
|
void *ptr = container_of(rcu_head, \
|
|
typeof(*_ptr), \
|
|
_rcu_head); \
|
|
kfree(ptr); \
|
|
} \
|
|
call_rcu(&(_ptr)->_rcu_head, __kc_kfree_rcu); \
|
|
} while (0)
|
|
#define HAVE_KFREE_RCU_BARRIER
|
|
#endif /* kfree_rcu */
|
|
#ifndef kstrtol_from_user
|
|
#define kstrtol_from_user(s, c, b, r) _kc_kstrtol_from_user(s, c, b, r)
|
|
static inline int _kc_kstrtol_from_user(const char __user *s, size_t count,
|
|
unsigned int base, long *res)
|
|
{
|
|
/* sign, base 2 representation, newline, terminator */
|
|
char buf[1 + sizeof(long) * 8 + 1 + 1];
|
|
|
|
count = min(count, sizeof(buf) - 1);
|
|
if (copy_from_user(buf, s, count))
|
|
return -EFAULT;
|
|
buf[count] = '\0';
|
|
return strict_strtol(buf, base, res);
|
|
}
|
|
#endif
|
|
|
|
#if (RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,0) || \
|
|
RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(5,7)))
|
|
/* 20000base_blah_full Supported and Advertised Registers */
|
|
#define SUPPORTED_20000baseMLD2_Full BIT(21)
|
|
#define SUPPORTED_20000baseKR2_Full BIT(22)
|
|
#define ADVERTISED_20000baseMLD2_Full BIT(21)
|
|
#define ADVERTISED_20000baseKR2_Full BIT(22)
|
|
#endif /* RHEL_RELEASE_CODE */
|
|
#endif /* < 3.0.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,1,0) )
|
|
#ifndef __netdev_alloc_skb_ip_align
|
|
#define __netdev_alloc_skb_ip_align(d,l,_g) netdev_alloc_skb_ip_align(d,l)
|
|
#endif /* __netdev_alloc_skb_ip_align */
|
|
#define dcb_ieee_setapp(dev, app) dcb_setapp(dev, app)
|
|
#define dcb_ieee_delapp(dev, app) 0
|
|
#define dcb_ieee_getapp_mask(dev, app) (1 << app->priority)
|
|
|
|
/* 1000BASE-T Control register */
|
|
#define CTL1000_AS_MASTER 0x0800
|
|
#define CTL1000_ENABLE_MASTER 0x1000
|
|
|
|
/* kernels less than 3.0.0 don't have this */
|
|
#ifndef ETH_P_8021AD
|
|
#define ETH_P_8021AD 0x88A8
|
|
#endif
|
|
|
|
/* Stub definition for !CONFIG_OF is introduced later */
|
|
#ifdef CONFIG_OF
|
|
static inline struct device_node *
|
|
pci_device_to_OF_node(struct pci_dev __maybe_unused *pdev)
|
|
{
|
|
#ifdef HAVE_STRUCT_DEVICE_OF_NODE
|
|
return pdev ? pdev->dev.of_node : NULL;
|
|
#else
|
|
return NULL;
|
|
#endif /* !HAVE_STRUCT_DEVICE_OF_NODE */
|
|
}
|
|
#endif /* CONFIG_OF */
|
|
#else /* < 3.1.0 */
|
|
#ifndef HAVE_DCBNL_IEEE_DELAPP
|
|
#define HAVE_DCBNL_IEEE_DELAPP
|
|
#endif
|
|
#endif /* < 3.1.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,2,0) )
|
|
#ifndef dma_zalloc_coherent
|
|
#define dma_zalloc_coherent(d, s, h, f) _kc_dma_zalloc_coherent(d, s, h, f)
|
|
static inline void *_kc_dma_zalloc_coherent(struct device *dev, size_t size,
|
|
dma_addr_t *dma_handle, gfp_t flag)
|
|
{
|
|
void *ret = dma_alloc_coherent(dev, size, dma_handle, flag);
|
|
if (ret)
|
|
memset(ret, 0, size);
|
|
return ret;
|
|
}
|
|
#endif
|
|
#ifdef ETHTOOL_GRXRINGS
|
|
#define HAVE_ETHTOOL_GET_RXNFC_VOID_RULE_LOCS
|
|
#endif /* ETHTOOL_GRXRINGS */
|
|
|
|
#ifndef skb_frag_size
|
|
#define skb_frag_size(frag) _kc_skb_frag_size(frag)
|
|
static inline unsigned int _kc_skb_frag_size(const skb_frag_t *frag)
|
|
{
|
|
return frag->size;
|
|
}
|
|
#endif /* skb_frag_size */
|
|
|
|
#ifndef skb_frag_size_sub
|
|
#define skb_frag_size_sub(frag, delta) _kc_skb_frag_size_sub(frag, delta)
|
|
static inline void _kc_skb_frag_size_sub(skb_frag_t *frag, int delta)
|
|
{
|
|
frag->size -= delta;
|
|
}
|
|
#endif /* skb_frag_size_sub */
|
|
|
|
#ifndef skb_frag_page
|
|
#define skb_frag_page(frag) _kc_skb_frag_page(frag)
|
|
static inline struct page *_kc_skb_frag_page(const skb_frag_t *frag)
|
|
{
|
|
return frag->page;
|
|
}
|
|
#endif /* skb_frag_page */
|
|
|
|
#ifndef skb_frag_address
|
|
#define skb_frag_address(frag) _kc_skb_frag_address(frag)
|
|
static inline void *_kc_skb_frag_address(const skb_frag_t *frag)
|
|
{
|
|
return page_address(skb_frag_page(frag)) + frag->page_offset;
|
|
}
|
|
#endif /* skb_frag_address */
|
|
|
|
#ifndef skb_frag_dma_map
|
|
#if ( LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,0) )
|
|
#include <linux/dma-mapping.h>
|
|
#endif
|
|
#define skb_frag_dma_map(dev,frag,offset,size,dir) \
|
|
_kc_skb_frag_dma_map(dev,frag,offset,size,dir)
|
|
static inline dma_addr_t _kc_skb_frag_dma_map(struct device *dev,
|
|
const skb_frag_t *frag,
|
|
size_t offset, size_t size,
|
|
enum dma_data_direction dir)
|
|
{
|
|
return dma_map_page(dev, skb_frag_page(frag),
|
|
frag->page_offset + offset, size, dir);
|
|
}
|
|
#endif /* skb_frag_dma_map */
|
|
|
|
#ifndef __skb_frag_unref
|
|
#define __skb_frag_unref(frag) __kc_skb_frag_unref(frag)
|
|
static inline void __kc_skb_frag_unref(skb_frag_t *frag)
|
|
{
|
|
put_page(skb_frag_page(frag));
|
|
}
|
|
#endif /* __skb_frag_unref */
|
|
|
|
#ifndef SPEED_UNKNOWN
|
|
#define SPEED_UNKNOWN -1
|
|
#endif
|
|
#ifndef DUPLEX_UNKNOWN
|
|
#define DUPLEX_UNKNOWN 0xff
|
|
#endif
|
|
#if ((RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,3)) ||\
|
|
(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(11,3,0)))
|
|
#ifndef HAVE_PCI_DEV_FLAGS_ASSIGNED
|
|
#define HAVE_PCI_DEV_FLAGS_ASSIGNED
|
|
#endif
|
|
#endif
|
|
#else /* < 3.2.0 */
|
|
#ifndef HAVE_PCI_DEV_FLAGS_ASSIGNED
|
|
#define HAVE_PCI_DEV_FLAGS_ASSIGNED
|
|
#define HAVE_VF_SPOOFCHK_CONFIGURE
|
|
#endif
|
|
#ifndef HAVE_SKB_L4_RXHASH
|
|
#define HAVE_SKB_L4_RXHASH
|
|
#endif
|
|
#define HAVE_IOMMU_PRESENT
|
|
#define HAVE_PM_QOS_REQUEST_LIST_NEW
|
|
#endif /* < 3.2.0 */
|
|
|
|
#if (RHEL_RELEASE_CODE && RHEL_RELEASE_CODE == RHEL_RELEASE_VERSION(6,2))
|
|
#undef ixgbe_get_netdev_tc_txq
|
|
#define ixgbe_get_netdev_tc_txq(dev, tc) (&netdev_extended(dev)->qos_data.tc_to_txq[tc])
|
|
#endif
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,3,0) )
|
|
/* NOTE: the order of parameters to _kc_alloc_workqueue() is different than
|
|
* alloc_workqueue() to avoid compiler warning from -Wvarargs
|
|
*/
|
|
static inline struct workqueue_struct * __attribute__ ((format(printf, 3, 4)))
|
|
_kc_alloc_workqueue(__maybe_unused int flags, __maybe_unused int max_active,
|
|
const char *fmt, ...)
|
|
{
|
|
struct workqueue_struct *wq;
|
|
va_list args, temp;
|
|
unsigned int len;
|
|
char *p;
|
|
|
|
va_start(args, fmt);
|
|
va_copy(temp, args);
|
|
len = vsnprintf(NULL, 0, fmt, temp);
|
|
va_end(temp);
|
|
|
|
p = kmalloc(len + 1, GFP_KERNEL);
|
|
if (!p) {
|
|
va_end(args);
|
|
return NULL;
|
|
}
|
|
|
|
vsnprintf(p, len + 1, fmt, args);
|
|
va_end(args);
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,36) )
|
|
wq = create_workqueue(p);
|
|
#else
|
|
wq = alloc_workqueue(p, flags, max_active);
|
|
#endif
|
|
kfree(p);
|
|
|
|
return wq;
|
|
}
|
|
#ifdef alloc_workqueue
|
|
#undef alloc_workqueue
|
|
#endif
|
|
#define alloc_workqueue(fmt, flags, max_active, args...) \
|
|
_kc_alloc_workqueue(flags, max_active, fmt, ##args)
|
|
|
|
#if !(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,5))
|
|
typedef u32 netdev_features_t;
|
|
#endif
|
|
#undef PCI_EXP_TYPE_RC_EC
|
|
#define PCI_EXP_TYPE_RC_EC 0xa /* Root Complex Event Collector */
|
|
#ifndef CONFIG_BQL
|
|
#define netdev_tx_completed_queue(_q, _p, _b) do {} while (0)
|
|
#define netdev_completed_queue(_n, _p, _b) do {} while (0)
|
|
#define netdev_tx_sent_queue(_q, _b) do {} while (0)
|
|
#define netdev_sent_queue(_n, _b) do {} while (0)
|
|
#define netdev_tx_reset_queue(_q) do {} while (0)
|
|
#define netdev_reset_queue(_n) do {} while (0)
|
|
#endif
|
|
#if (SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(11,3,0))
|
|
#define HAVE_ETHTOOL_GRXFHINDIR_SIZE
|
|
#endif /* SLE_VERSION(11,3,0) */
|
|
#define netif_xmit_stopped(_q) netif_tx_queue_stopped(_q)
|
|
#if !(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(11,4,0))
|
|
static inline int __kc_ipv6_skip_exthdr(const struct sk_buff *skb, int start,
|
|
u8 *nexthdrp,
|
|
__be16 __always_unused *frag_offp)
|
|
{
|
|
return ipv6_skip_exthdr(skb, start, nexthdrp);
|
|
}
|
|
#undef ipv6_skip_exthdr
|
|
#define ipv6_skip_exthdr(a,b,c,d) __kc_ipv6_skip_exthdr((a), (b), (c), (d))
|
|
#endif /* !SLES11sp4 or greater */
|
|
|
|
#if (!(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,4)) && \
|
|
!(SLE_VERSION_CODE >= SLE_VERSION(11,3,0)))
|
|
static inline u32 ethtool_rxfh_indir_default(u32 index, u32 n_rx_rings)
|
|
{
|
|
return index % n_rx_rings;
|
|
}
|
|
#endif
|
|
|
|
#else /* ! < 3.3.0 */
|
|
#define HAVE_ETHTOOL_GRXFHINDIR_SIZE
|
|
#define HAVE_INT_NDO_VLAN_RX_ADD_VID
|
|
#ifdef ETHTOOL_SRXNTUPLE
|
|
#undef ETHTOOL_SRXNTUPLE
|
|
#endif
|
|
#endif /* < 3.3.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,4,0) )
|
|
#ifndef NETIF_F_RXFCS
|
|
#define NETIF_F_RXFCS 0
|
|
#endif /* NETIF_F_RXFCS */
|
|
#ifndef NETIF_F_RXALL
|
|
#define NETIF_F_RXALL 0
|
|
#endif /* NETIF_F_RXALL */
|
|
|
|
#if !(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(11,3,0))
|
|
#define NUMTCS_RETURNS_U8
|
|
|
|
int _kc_simple_open(struct inode *inode, struct file *file);
|
|
#define simple_open _kc_simple_open
|
|
#endif /* !(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(11,3,0)) */
|
|
|
|
#ifndef skb_add_rx_frag
|
|
#define skb_add_rx_frag _kc_skb_add_rx_frag
|
|
void _kc_skb_add_rx_frag(struct sk_buff * skb, int i, struct page *page,
|
|
int off, int size, unsigned int truesize);
|
|
#endif
|
|
#ifdef NET_ADDR_RANDOM
|
|
#define eth_hw_addr_random(N) do { \
|
|
eth_random_addr(N->dev_addr); \
|
|
N->addr_assign_type |= NET_ADDR_RANDOM; \
|
|
} while (0)
|
|
#else /* NET_ADDR_RANDOM */
|
|
#define eth_hw_addr_random(N) eth_random_addr(N->dev_addr)
|
|
#endif /* NET_ADDR_RANDOM */
|
|
|
|
#ifndef for_each_set_bit_from
|
|
#define for_each_set_bit_from(bit, addr, size) \
|
|
for ((bit) = find_next_bit((addr), (size), (bit)); \
|
|
(bit) < (size); \
|
|
(bit) = find_next_bit((addr), (size), (bit) + 1))
|
|
#endif /* for_each_set_bit_from */
|
|
|
|
#else /* < 3.4.0 */
|
|
#include <linux/kconfig.h>
|
|
#endif /* >= 3.4.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE >= KERNEL_VERSION(3,0,0) ) || \
|
|
( RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,4) )
|
|
#if !defined(NO_PTP_SUPPORT) && IS_ENABLED(CONFIG_PTP_1588_CLOCK)
|
|
#define HAVE_PTP_1588_CLOCK
|
|
#endif /* !NO_PTP_SUPPORT && IS_ENABLED(CONFIG_PTP_1588_CLOCK) */
|
|
#endif /* >= 3.0.0 || RHEL_RELEASE > 6.4 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,5,0) )
|
|
|
|
#ifndef SIZE_MAX
|
|
#define SIZE_MAX (~(size_t)0)
|
|
#endif
|
|
|
|
#ifndef BITS_PER_LONG_LONG
|
|
#define BITS_PER_LONG_LONG 64
|
|
#endif
|
|
|
|
#ifndef ether_addr_equal
|
|
static inline bool __kc_ether_addr_equal(const u8 *addr1, const u8 *addr2)
|
|
{
|
|
return !compare_ether_addr(addr1, addr2);
|
|
}
|
|
#define ether_addr_equal(_addr1, _addr2) __kc_ether_addr_equal((_addr1),(_addr2))
|
|
#endif
|
|
|
|
/* Definitions for !CONFIG_OF_NET are introduced in 3.10 */
|
|
#ifdef CONFIG_OF_NET
|
|
static inline int of_get_phy_mode(struct device_node __always_unused *np)
|
|
{
|
|
return -ENODEV;
|
|
}
|
|
|
|
static inline const void *
|
|
of_get_mac_address(struct device_node __always_unused *np)
|
|
{
|
|
return NULL;
|
|
}
|
|
#endif
|
|
#else
|
|
#include <linux/of_net.h>
|
|
#define HAVE_FDB_OPS
|
|
#define HAVE_ETHTOOL_GET_TS_INFO
|
|
#endif /* < 3.5.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,6,0) )
|
|
#define PCI_EXP_LNKCAP2 44 /* Link Capability 2 */
|
|
|
|
#ifndef MDIO_EEE_100TX
|
|
#define MDIO_EEE_100TX 0x0002 /* 100TX EEE cap */
|
|
#endif
|
|
#ifndef MDIO_EEE_1000T
|
|
#define MDIO_EEE_1000T 0x0004 /* 1000T EEE cap */
|
|
#endif
|
|
#ifndef MDIO_EEE_10GT
|
|
#define MDIO_EEE_10GT 0x0008 /* 10GT EEE cap */
|
|
#endif
|
|
#ifndef MDIO_EEE_1000KX
|
|
#define MDIO_EEE_1000KX 0x0010 /* 1000KX EEE cap */
|
|
#endif
|
|
#ifndef MDIO_EEE_10GKX4
|
|
#define MDIO_EEE_10GKX4 0x0020 /* 10G KX4 EEE cap */
|
|
#endif
|
|
#ifndef MDIO_EEE_10GKR
|
|
#define MDIO_EEE_10GKR 0x0040 /* 10G KR EEE cap */
|
|
#endif
|
|
|
|
#ifndef __GFP_MEMALLOC
|
|
#define __GFP_MEMALLOC 0
|
|
#endif
|
|
|
|
#ifndef eth_broadcast_addr
|
|
#define eth_broadcast_addr _kc_eth_broadcast_addr
|
|
static inline void _kc_eth_broadcast_addr(u8 *addr)
|
|
{
|
|
memset(addr, 0xff, ETH_ALEN);
|
|
}
|
|
#endif
|
|
|
|
#ifndef eth_random_addr
|
|
#define eth_random_addr _kc_eth_random_addr
|
|
static inline void _kc_eth_random_addr(u8 *addr)
|
|
{
|
|
get_random_bytes(addr, ETH_ALEN);
|
|
addr[0] &= 0xfe; /* clear multicast */
|
|
addr[0] |= 0x02; /* set local assignment */
|
|
}
|
|
#endif /* eth_random_addr */
|
|
|
|
#ifndef DMA_ATTR_SKIP_CPU_SYNC
|
|
#define DMA_ATTR_SKIP_CPU_SYNC 0
|
|
#endif
|
|
#else /* < 3.6.0 */
|
|
#define HAVE_STRUCT_PAGE_PFMEMALLOC
|
|
#endif /* < 3.6.0 */
|
|
|
|
/******************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,7,0) )
|
|
#include <linux/workqueue.h>
|
|
#ifndef ADVERTISED_40000baseKR4_Full
|
|
/* these defines were all added in one commit, so should be safe
|
|
* to trigger activiation on one define
|
|
*/
|
|
#define SUPPORTED_40000baseKR4_Full BIT(23)
|
|
#define SUPPORTED_40000baseCR4_Full BIT(24)
|
|
#define SUPPORTED_40000baseSR4_Full BIT(25)
|
|
#define SUPPORTED_40000baseLR4_Full BIT(26)
|
|
#define ADVERTISED_40000baseKR4_Full BIT(23)
|
|
#define ADVERTISED_40000baseCR4_Full BIT(24)
|
|
#define ADVERTISED_40000baseSR4_Full BIT(25)
|
|
#define ADVERTISED_40000baseLR4_Full BIT(26)
|
|
#endif
|
|
|
|
#ifndef mmd_eee_cap_to_ethtool_sup_t
|
|
/**
|
|
* mmd_eee_cap_to_ethtool_sup_t
|
|
* @eee_cap: value of the MMD EEE Capability register
|
|
*
|
|
* A small helper function that translates MMD EEE Capability (3.20) bits
|
|
* to ethtool supported settings.
|
|
*/
|
|
static inline u32 __kc_mmd_eee_cap_to_ethtool_sup_t(u16 eee_cap)
|
|
{
|
|
u32 supported = 0;
|
|
|
|
if (eee_cap & MDIO_EEE_100TX)
|
|
supported |= SUPPORTED_100baseT_Full;
|
|
if (eee_cap & MDIO_EEE_1000T)
|
|
supported |= SUPPORTED_1000baseT_Full;
|
|
if (eee_cap & MDIO_EEE_10GT)
|
|
supported |= SUPPORTED_10000baseT_Full;
|
|
if (eee_cap & MDIO_EEE_1000KX)
|
|
supported |= SUPPORTED_1000baseKX_Full;
|
|
if (eee_cap & MDIO_EEE_10GKX4)
|
|
supported |= SUPPORTED_10000baseKX4_Full;
|
|
if (eee_cap & MDIO_EEE_10GKR)
|
|
supported |= SUPPORTED_10000baseKR_Full;
|
|
|
|
return supported;
|
|
}
|
|
#define mmd_eee_cap_to_ethtool_sup_t(eee_cap) \
|
|
__kc_mmd_eee_cap_to_ethtool_sup_t(eee_cap)
|
|
#endif /* mmd_eee_cap_to_ethtool_sup_t */
|
|
|
|
#ifndef mmd_eee_adv_to_ethtool_adv_t
|
|
/**
|
|
* mmd_eee_adv_to_ethtool_adv_t
|
|
* @eee_adv: value of the MMD EEE Advertisement/Link Partner Ability registers
|
|
*
|
|
* A small helper function that translates the MMD EEE Advertisement (7.60)
|
|
* and MMD EEE Link Partner Ability (7.61) bits to ethtool advertisement
|
|
* settings.
|
|
*/
|
|
static inline u32 __kc_mmd_eee_adv_to_ethtool_adv_t(u16 eee_adv)
|
|
{
|
|
u32 adv = 0;
|
|
|
|
if (eee_adv & MDIO_EEE_100TX)
|
|
adv |= ADVERTISED_100baseT_Full;
|
|
if (eee_adv & MDIO_EEE_1000T)
|
|
adv |= ADVERTISED_1000baseT_Full;
|
|
if (eee_adv & MDIO_EEE_10GT)
|
|
adv |= ADVERTISED_10000baseT_Full;
|
|
if (eee_adv & MDIO_EEE_1000KX)
|
|
adv |= ADVERTISED_1000baseKX_Full;
|
|
if (eee_adv & MDIO_EEE_10GKX4)
|
|
adv |= ADVERTISED_10000baseKX4_Full;
|
|
if (eee_adv & MDIO_EEE_10GKR)
|
|
adv |= ADVERTISED_10000baseKR_Full;
|
|
|
|
return adv;
|
|
}
|
|
|
|
#define mmd_eee_adv_to_ethtool_adv_t(eee_adv) \
|
|
__kc_mmd_eee_adv_to_ethtool_adv_t(eee_adv)
|
|
#endif /* mmd_eee_adv_to_ethtool_adv_t */
|
|
|
|
#ifndef ethtool_adv_to_mmd_eee_adv_t
|
|
/**
|
|
* ethtool_adv_to_mmd_eee_adv_t
|
|
* @adv: the ethtool advertisement settings
|
|
*
|
|
* A small helper function that translates ethtool advertisement settings
|
|
* to EEE advertisements for the MMD EEE Advertisement (7.60) and
|
|
* MMD EEE Link Partner Ability (7.61) registers.
|
|
*/
|
|
static inline u16 __kc_ethtool_adv_to_mmd_eee_adv_t(u32 adv)
|
|
{
|
|
u16 reg = 0;
|
|
|
|
if (adv & ADVERTISED_100baseT_Full)
|
|
reg |= MDIO_EEE_100TX;
|
|
if (adv & ADVERTISED_1000baseT_Full)
|
|
reg |= MDIO_EEE_1000T;
|
|
if (adv & ADVERTISED_10000baseT_Full)
|
|
reg |= MDIO_EEE_10GT;
|
|
if (adv & ADVERTISED_1000baseKX_Full)
|
|
reg |= MDIO_EEE_1000KX;
|
|
if (adv & ADVERTISED_10000baseKX4_Full)
|
|
reg |= MDIO_EEE_10GKX4;
|
|
if (adv & ADVERTISED_10000baseKR_Full)
|
|
reg |= MDIO_EEE_10GKR;
|
|
|
|
return reg;
|
|
}
|
|
#define ethtool_adv_to_mmd_eee_adv_t(adv) __kc_ethtool_adv_to_mmd_eee_adv_t(adv)
|
|
#endif /* ethtool_adv_to_mmd_eee_adv_t */
|
|
|
|
#ifndef pci_pcie_type
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,24) )
|
|
static inline u8 pci_pcie_type(struct pci_dev *pdev)
|
|
{
|
|
int pos;
|
|
u16 reg16;
|
|
|
|
pos = pci_find_capability(pdev, PCI_CAP_ID_EXP);
|
|
BUG_ON(!pos);
|
|
pci_read_config_word(pdev, pos + PCI_EXP_FLAGS, ®16);
|
|
return (reg16 & PCI_EXP_FLAGS_TYPE) >> 4;
|
|
}
|
|
#else /* < 2.6.24 */
|
|
#define pci_pcie_type(x) (x)->pcie_type
|
|
#endif /* < 2.6.24 */
|
|
#endif /* pci_pcie_type */
|
|
|
|
#if ( ! ( RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,4) ) ) && \
|
|
( ! ( SLE_VERSION_CODE >= SLE_VERSION(11,3,0) ) ) && \
|
|
( LINUX_VERSION_CODE >= KERNEL_VERSION(3,0,0) )
|
|
#define ptp_clock_register(caps, args...) ptp_clock_register(caps)
|
|
#endif
|
|
|
|
#ifndef pcie_capability_read_word
|
|
int __kc_pcie_capability_read_word(struct pci_dev *dev, int pos, u16 *val);
|
|
#define pcie_capability_read_word(d,p,v) __kc_pcie_capability_read_word(d,p,v)
|
|
#endif /* pcie_capability_read_word */
|
|
|
|
#ifndef pcie_capability_read_dword
|
|
int __kc_pcie_capability_read_dword(struct pci_dev *dev, int pos, u32 *val);
|
|
#define pcie_capability_read_dword(d,p,v) __kc_pcie_capability_read_dword(d,p,v)
|
|
#endif
|
|
|
|
#ifndef pcie_capability_write_word
|
|
int __kc_pcie_capability_write_word(struct pci_dev *dev, int pos, u16 val);
|
|
#define pcie_capability_write_word(d,p,v) __kc_pcie_capability_write_word(d,p,v)
|
|
#endif /* pcie_capability_write_word */
|
|
|
|
#ifndef pcie_capability_clear_and_set_word
|
|
int __kc_pcie_capability_clear_and_set_word(struct pci_dev *dev, int pos,
|
|
u16 clear, u16 set);
|
|
#define pcie_capability_clear_and_set_word(d,p,c,s) \
|
|
__kc_pcie_capability_clear_and_set_word(d,p,c,s)
|
|
#endif /* pcie_capability_clear_and_set_word */
|
|
|
|
#ifndef pcie_capability_clear_word
|
|
int __kc_pcie_capability_clear_word(struct pci_dev *dev, int pos,
|
|
u16 clear);
|
|
#define pcie_capability_clear_word(d, p, c) \
|
|
__kc_pcie_capability_clear_word(d, p, c)
|
|
#endif /* pcie_capability_clear_word */
|
|
|
|
#ifndef PCI_EXP_LNKSTA2
|
|
#define PCI_EXP_LNKSTA2 50 /* Link Status 2 */
|
|
#endif
|
|
|
|
#if (SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(11,3,0))
|
|
#define USE_CONST_DEV_UC_CHAR
|
|
#define HAVE_NDO_FDB_ADD_NLATTR
|
|
#endif
|
|
|
|
#if !(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,8))
|
|
#define napi_gro_flush(_napi, _flush_old) napi_gro_flush(_napi)
|
|
#endif /* !RHEL6.8+ */
|
|
|
|
#if (RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,6))
|
|
#include <linux/hashtable.h>
|
|
#else
|
|
|
|
#define DEFINE_HASHTABLE(name, bits) \
|
|
struct hlist_head name[1 << (bits)] = \
|
|
{ [0 ... ((1 << (bits)) - 1)] = HLIST_HEAD_INIT }
|
|
|
|
#define DEFINE_READ_MOSTLY_HASHTABLE(name, bits) \
|
|
struct hlist_head name[1 << (bits)] __read_mostly = \
|
|
{ [0 ... ((1 << (bits)) - 1)] = HLIST_HEAD_INIT }
|
|
|
|
#define DECLARE_HASHTABLE(name, bits) \
|
|
struct hlist_head name[1 << (bits)]
|
|
|
|
#define HASH_SIZE(name) (ARRAY_SIZE(name))
|
|
#define HASH_BITS(name) ilog2(HASH_SIZE(name))
|
|
|
|
/* Use hash_32 when possible to allow for fast 32bit hashing in 64bit kernels. */
|
|
#define hash_min(val, bits) \
|
|
(sizeof(val) <= 4 ? hash_32(val, bits) : hash_long(val, bits))
|
|
|
|
static inline void __hash_init(struct hlist_head *ht, unsigned int sz)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < sz; i++)
|
|
INIT_HLIST_HEAD(&ht[i]);
|
|
}
|
|
|
|
#define hash_init(hashtable) __hash_init(hashtable, HASH_SIZE(hashtable))
|
|
|
|
#define hash_add(hashtable, node, key) \
|
|
hlist_add_head(node, &hashtable[hash_min(key, HASH_BITS(hashtable))])
|
|
|
|
static inline bool hash_hashed(struct hlist_node *node)
|
|
{
|
|
return !hlist_unhashed(node);
|
|
}
|
|
|
|
static inline bool __hash_empty(struct hlist_head *ht, unsigned int sz)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < sz; i++)
|
|
if (!hlist_empty(&ht[i]))
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
#define hash_empty(hashtable) __hash_empty(hashtable, HASH_SIZE(hashtable))
|
|
|
|
static inline void hash_del(struct hlist_node *node)
|
|
{
|
|
hlist_del_init(node);
|
|
}
|
|
#endif /* RHEL >= 6.6 */
|
|
|
|
/* We don't have @flags support prior to 3.7, so we'll simply ignore the flags
|
|
* parameter on these older kernels.
|
|
*/
|
|
#define __setup_timer(_timer, _fn, _data, _flags) \
|
|
setup_timer((_timer), (_fn), (_data)) \
|
|
|
|
#if ( ! ( RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,7) ) ) && \
|
|
( ! ( SLE_VERSION_CODE >= SLE_VERSION(12,0,0) ) )
|
|
|
|
#ifndef mod_delayed_work
|
|
/**
|
|
* __mod_delayed_work - modify delay or queue delayed work
|
|
* @wq: workqueue to use
|
|
* @dwork: delayed work to queue
|
|
* @delay: number of jiffies to wait before queueing
|
|
*
|
|
* Return: %true if @dwork was pending and was rescheduled;
|
|
* %false if it wasn't pending
|
|
*
|
|
* Note: the dwork parameter was declared as a void*
|
|
* to avoid comptibility problems with early 2.6 kernels
|
|
* where struct delayed_work is not declared. Unlike the original
|
|
* implementation flags are not preserved and it shouldn't be
|
|
* used in the interrupt context.
|
|
*/
|
|
static inline bool __mod_delayed_work(struct workqueue_struct *wq,
|
|
void *dwork,
|
|
unsigned long delay)
|
|
{
|
|
bool ret = cancel_delayed_work(dwork);
|
|
queue_delayed_work(wq, dwork, delay);
|
|
return ret;
|
|
}
|
|
#define mod_delayed_work(wq, dwork, delay) __mod_delayed_work(wq, dwork, delay)
|
|
#endif /* mod_delayed_work */
|
|
|
|
#endif /* !(RHEL >= 6.7) && !(SLE >= 12.0) */
|
|
#else /* >= 3.7.0 */
|
|
#include <linux/hashtable.h>
|
|
#define HAVE_CONST_STRUCT_PCI_ERROR_HANDLERS
|
|
#define USE_CONST_DEV_UC_CHAR
|
|
#define HAVE_NDO_FDB_ADD_NLATTR
|
|
#endif /* >= 3.7.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,8,0) )
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,5)) && \
|
|
!(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(11,4,0)))
|
|
#ifndef pci_sriov_set_totalvfs
|
|
static inline int __kc_pci_sriov_set_totalvfs(struct pci_dev __always_unused *dev, u16 __always_unused numvfs)
|
|
{
|
|
return 0;
|
|
}
|
|
#define pci_sriov_set_totalvfs(a, b) __kc_pci_sriov_set_totalvfs((a), (b))
|
|
#endif
|
|
#endif /* !(RHEL_RELEASE_CODE >= 6.5 && SLE_VERSION_CODE >= 11.4) */
|
|
#ifndef PCI_EXP_LNKCTL_ASPM_L0S
|
|
#define PCI_EXP_LNKCTL_ASPM_L0S 0x01 /* L0s Enable */
|
|
#endif
|
|
#ifndef PCI_EXP_LNKCTL_ASPM_L1
|
|
#define PCI_EXP_LNKCTL_ASPM_L1 0x02 /* L1 Enable */
|
|
#endif
|
|
#define HAVE_CONFIG_HOTPLUG
|
|
/* Reserved Ethernet Addresses per IEEE 802.1Q */
|
|
static const u8 eth_reserved_addr_base[ETH_ALEN] __aligned(2) = {
|
|
0x01, 0x80, 0xc2, 0x00, 0x00, 0x00 };
|
|
|
|
#ifndef is_link_local_ether_addr
|
|
static inline bool __kc_is_link_local_ether_addr(const u8 *addr)
|
|
{
|
|
__be16 *a = (__be16 *)addr;
|
|
static const __be16 *b = (const __be16 *)eth_reserved_addr_base;
|
|
static const __be16 m = cpu_to_be16(0xfff0);
|
|
|
|
return ((a[0] ^ b[0]) | (a[1] ^ b[1]) | ((a[2] ^ b[2]) & m)) == 0;
|
|
}
|
|
#define is_link_local_ether_addr(addr) __kc_is_link_local_ether_addr(addr)
|
|
#endif /* is_link_local_ether_addr */
|
|
|
|
#ifndef FLOW_MAC_EXT
|
|
#define FLOW_MAC_EXT 0x40000000
|
|
#endif /* FLOW_MAC_EXT */
|
|
|
|
#if (SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(11,4,0))
|
|
#define HAVE_SRIOV_CONFIGURE
|
|
#endif
|
|
|
|
#ifndef PCI_EXP_LNKCAP_SLS_2_5GB
|
|
#define PCI_EXP_LNKCAP_SLS_2_5GB 0x00000001 /* LNKCAP2 SLS Vector bit 0 */
|
|
#endif
|
|
|
|
#ifndef PCI_EXP_LNKCAP_SLS_5_0GB
|
|
#define PCI_EXP_LNKCAP_SLS_5_0GB 0x00000002 /* LNKCAP2 SLS Vector bit 1 */
|
|
#endif
|
|
|
|
#undef PCI_EXP_LNKCAP2_SLS_2_5GB
|
|
#define PCI_EXP_LNKCAP2_SLS_2_5GB 0x00000002 /* Supported Speed 2.5GT/s */
|
|
|
|
#undef PCI_EXP_LNKCAP2_SLS_5_0GB
|
|
#define PCI_EXP_LNKCAP2_SLS_5_0GB 0x00000004 /* Supported Speed 5GT/s */
|
|
|
|
#undef PCI_EXP_LNKCAP2_SLS_8_0GB
|
|
#define PCI_EXP_LNKCAP2_SLS_8_0GB 0x00000008 /* Supported Speed 8GT/s */
|
|
|
|
#else /* >= 3.8.0 */
|
|
#ifndef __devinit
|
|
#define __devinit
|
|
#endif
|
|
|
|
#ifndef __devinitdata
|
|
#define __devinitdata
|
|
#endif
|
|
|
|
#ifndef __devinitconst
|
|
#define __devinitconst
|
|
#endif
|
|
|
|
#ifndef __devexit
|
|
#define __devexit
|
|
#endif
|
|
|
|
#ifndef __devexit_p
|
|
#define __devexit_p
|
|
#endif
|
|
|
|
#ifndef HAVE_ENCAP_CSUM_OFFLOAD
|
|
#define HAVE_ENCAP_CSUM_OFFLOAD
|
|
#endif
|
|
|
|
#ifndef HAVE_GRE_ENCAP_OFFLOAD
|
|
#define HAVE_GRE_ENCAP_OFFLOAD
|
|
#endif
|
|
|
|
#ifndef HAVE_SRIOV_CONFIGURE
|
|
#define HAVE_SRIOV_CONFIGURE
|
|
#endif
|
|
|
|
#define HAVE_BRIDGE_ATTRIBS
|
|
#ifndef BRIDGE_MODE_VEB
|
|
#define BRIDGE_MODE_VEB 0 /* Default loopback mode */
|
|
#endif /* BRIDGE_MODE_VEB */
|
|
#ifndef BRIDGE_MODE_VEPA
|
|
#define BRIDGE_MODE_VEPA 1 /* 802.1Qbg defined VEPA mode */
|
|
#endif /* BRIDGE_MODE_VEPA */
|
|
#endif /* >= 3.8.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,9,0) )
|
|
|
|
#undef BUILD_BUG_ON
|
|
#ifdef __CHECKER__
|
|
#define BUILD_BUG_ON(condition) (0)
|
|
#else /* __CHECKER__ */
|
|
#ifndef __compiletime_warning
|
|
#if defined(__GNUC__) && ((__GNUC__ * 10000 + __GNUC_MINOR__ * 100) >= 40400)
|
|
#define __compiletime_warning(message) __attribute__((warning(message)))
|
|
#else /* __GNUC__ */
|
|
#define __compiletime_warning(message)
|
|
#endif /* __GNUC__ */
|
|
#endif /* __compiletime_warning */
|
|
#ifndef __compiletime_error
|
|
#if defined(__GNUC__) && ((__GNUC__ * 10000 + __GNUC_MINOR__ * 100) >= 40400)
|
|
#define __compiletime_error(message) __attribute__((error(message)))
|
|
#define __compiletime_error_fallback(condition) do { } while (0)
|
|
#else /* __GNUC__ */
|
|
#define __compiletime_error(message)
|
|
#define __compiletime_error_fallback(condition) \
|
|
do { ((void)sizeof(char[1 - 2 * condition])); } while (0)
|
|
#endif /* __GNUC__ */
|
|
#else /* __compiletime_error */
|
|
#define __compiletime_error_fallback(condition) do { } while (0)
|
|
#endif /* __compiletime_error */
|
|
#define __compiletime_assert(condition, msg, prefix, suffix) \
|
|
do { \
|
|
bool __cond = !(condition); \
|
|
extern void prefix ## suffix(void) __compiletime_error(msg); \
|
|
if (__cond) \
|
|
prefix ## suffix(); \
|
|
__compiletime_error_fallback(__cond); \
|
|
} while (0)
|
|
|
|
#define _compiletime_assert(condition, msg, prefix, suffix) \
|
|
__compiletime_assert(condition, msg, prefix, suffix)
|
|
#define compiletime_assert(condition, msg) \
|
|
_compiletime_assert(condition, msg, __compiletime_assert_, __LINE__)
|
|
#define BUILD_BUG_ON_MSG(cond, msg) compiletime_assert(!(cond), msg)
|
|
#ifndef __OPTIMIZE__
|
|
#define BUILD_BUG_ON(condition) ((void)sizeof(char[1 - 2*!!(condition)]))
|
|
#else /* __OPTIMIZE__ */
|
|
#define BUILD_BUG_ON(condition) \
|
|
BUILD_BUG_ON_MSG(condition, "BUILD_BUG_ON failed: " #condition)
|
|
#endif /* __OPTIMIZE__ */
|
|
#endif /* __CHECKER__ */
|
|
|
|
#undef hlist_entry
|
|
#define hlist_entry(ptr, type, member) container_of(ptr,type,member)
|
|
|
|
#undef hlist_entry_safe
|
|
#define hlist_entry_safe(ptr, type, member) \
|
|
({ typeof(ptr) ____ptr = (ptr); \
|
|
____ptr ? hlist_entry(____ptr, type, member) : NULL; \
|
|
})
|
|
|
|
#undef hlist_for_each_entry
|
|
#define hlist_for_each_entry(pos, head, member) \
|
|
for (pos = hlist_entry_safe((head)->first, typeof(*(pos)), member); \
|
|
pos; \
|
|
pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
|
|
|
|
#undef hlist_for_each_entry_safe
|
|
#define hlist_for_each_entry_safe(pos, n, head, member) \
|
|
for (pos = hlist_entry_safe((head)->first, typeof(*pos), member); \
|
|
pos && ({ n = pos->member.next; 1; }); \
|
|
pos = hlist_entry_safe(n, typeof(*pos), member))
|
|
|
|
#undef hlist_for_each_entry_continue
|
|
#define hlist_for_each_entry_continue(pos, member) \
|
|
for (pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member);\
|
|
pos; \
|
|
pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
|
|
|
|
#undef hlist_for_each_entry_from
|
|
#define hlist_for_each_entry_from(pos, member) \
|
|
for (; pos; \
|
|
pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
|
|
|
|
#undef hash_for_each
|
|
#define hash_for_each(name, bkt, obj, member) \
|
|
for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
|
|
(bkt)++)\
|
|
hlist_for_each_entry(obj, &name[bkt], member)
|
|
|
|
#undef hash_for_each_safe
|
|
#define hash_for_each_safe(name, bkt, tmp, obj, member) \
|
|
for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
|
|
(bkt)++)\
|
|
hlist_for_each_entry_safe(obj, tmp, &name[bkt], member)
|
|
|
|
#undef hash_for_each_possible
|
|
#define hash_for_each_possible(name, obj, member, key) \
|
|
hlist_for_each_entry(obj, &name[hash_min(key, HASH_BITS(name))], member)
|
|
|
|
#undef hash_for_each_possible_safe
|
|
#define hash_for_each_possible_safe(name, obj, tmp, member, key) \
|
|
hlist_for_each_entry_safe(obj, tmp,\
|
|
&name[hash_min(key, HASH_BITS(name))], member)
|
|
|
|
#ifdef CONFIG_XPS
|
|
int __kc_netif_set_xps_queue(struct net_device *, const struct cpumask *, u16);
|
|
#define netif_set_xps_queue(_dev, _mask, _idx) __kc_netif_set_xps_queue((_dev), (_mask), (_idx))
|
|
#else /* CONFIG_XPS */
|
|
#define netif_set_xps_queue(_dev, _mask, _idx) do {} while (0)
|
|
#endif /* CONFIG_XPS */
|
|
|
|
#ifdef HAVE_NETDEV_SELECT_QUEUE
|
|
#define _kc_hashrnd 0xd631614b /* not so random hash salt */
|
|
u16 __kc_netdev_pick_tx(struct net_device *dev, struct sk_buff *skb);
|
|
#define __netdev_pick_tx __kc_netdev_pick_tx
|
|
#endif /* HAVE_NETDEV_SELECT_QUEUE */
|
|
#else
|
|
#define HAVE_BRIDGE_FILTER
|
|
#define HAVE_FDB_DEL_NLATTR
|
|
#endif /* < 3.9.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,10,0) )
|
|
#ifndef NAPI_POLL_WEIGHT
|
|
#define NAPI_POLL_WEIGHT 64
|
|
#endif
|
|
#ifdef CONFIG_PCI_IOV
|
|
int __kc_pci_vfs_assigned(struct pci_dev *dev);
|
|
#else
|
|
static inline int __kc_pci_vfs_assigned(struct pci_dev __always_unused *dev)
|
|
{
|
|
return 0;
|
|
}
|
|
#endif
|
|
#define pci_vfs_assigned(dev) __kc_pci_vfs_assigned(dev)
|
|
|
|
#ifndef list_first_entry_or_null
|
|
#define list_first_entry_or_null(ptr, type, member) \
|
|
(!list_empty(ptr) ? list_first_entry(ptr, type, member) : NULL)
|
|
#endif
|
|
|
|
#ifndef VLAN_TX_COOKIE_MAGIC
|
|
static inline struct sk_buff *__kc__vlan_hwaccel_put_tag(struct sk_buff *skb,
|
|
u16 vlan_tci)
|
|
{
|
|
#ifdef VLAN_TAG_PRESENT
|
|
vlan_tci |= VLAN_TAG_PRESENT;
|
|
#endif
|
|
skb->vlan_tci = vlan_tci;
|
|
return skb;
|
|
}
|
|
#define __vlan_hwaccel_put_tag(skb, vlan_proto, vlan_tci) \
|
|
__kc__vlan_hwaccel_put_tag(skb, vlan_tci)
|
|
#endif
|
|
|
|
#ifdef HAVE_FDB_OPS
|
|
#if defined(HAVE_NDO_FDB_ADD_NLATTR)
|
|
int __kc_ndo_dflt_fdb_add(struct ndmsg *ndm, struct nlattr *tb[],
|
|
struct net_device *dev,
|
|
const unsigned char *addr, u16 flags);
|
|
#elif defined(USE_CONST_DEV_UC_CHAR)
|
|
int __kc_ndo_dflt_fdb_add(struct ndmsg *ndm, struct net_device *dev,
|
|
const unsigned char *addr, u16 flags);
|
|
#else
|
|
int __kc_ndo_dflt_fdb_add(struct ndmsg *ndm, struct net_device *dev,
|
|
unsigned char *addr, u16 flags);
|
|
#endif /* HAVE_NDO_FDB_ADD_NLATTR */
|
|
#if defined(HAVE_FDB_DEL_NLATTR)
|
|
int __kc_ndo_dflt_fdb_del(struct ndmsg *ndm, struct nlattr *tb[],
|
|
struct net_device *dev,
|
|
const unsigned char *addr);
|
|
#elif defined(USE_CONST_DEV_UC_CHAR)
|
|
int __kc_ndo_dflt_fdb_del(struct ndmsg *ndm, struct net_device *dev,
|
|
const unsigned char *addr);
|
|
#else
|
|
int __kc_ndo_dflt_fdb_del(struct ndmsg *ndm, struct net_device *dev,
|
|
unsigned char *addr);
|
|
#endif /* HAVE_FDB_DEL_NLATTR */
|
|
#define ndo_dflt_fdb_add __kc_ndo_dflt_fdb_add
|
|
#define ndo_dflt_fdb_del __kc_ndo_dflt_fdb_del
|
|
#endif /* HAVE_FDB_OPS */
|
|
|
|
#ifndef PCI_DEVID
|
|
#define PCI_DEVID(bus, devfn) ((((u16)(bus)) << 8) | (devfn))
|
|
#endif
|
|
|
|
/* The definitions for these functions when CONFIG_OF_NET is defined are
|
|
* pulled in from <linux/of_net.h>. For kernels older than 3.5 we already have
|
|
* backports for when CONFIG_OF_NET is true. These are separated and
|
|
* duplicated in order to cover all cases so that all kernels get either the
|
|
* real definitions (when CONFIG_OF_NET is defined) or the stub definitions
|
|
* (when CONFIG_OF_NET is not defined, or the kernel is too old to have real
|
|
* definitions).
|
|
*/
|
|
#ifndef CONFIG_OF_NET
|
|
static inline int of_get_phy_mode(struct device_node __always_unused *np)
|
|
{
|
|
return -ENODEV;
|
|
}
|
|
|
|
static inline const void *
|
|
of_get_mac_address(struct device_node __always_unused *np)
|
|
{
|
|
return NULL;
|
|
}
|
|
#endif
|
|
|
|
#else /* >= 3.10.0 */
|
|
#define HAVE_ENCAP_TSO_OFFLOAD
|
|
#define USE_DEFAULT_FDB_DEL_DUMP
|
|
#define HAVE_SKB_INNER_NETWORK_HEADER
|
|
|
|
#if (RHEL_RELEASE_CODE && \
|
|
(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,0)))
|
|
#if (RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(8,0))
|
|
#define HAVE_RHEL7_PCI_DRIVER_RH
|
|
#if (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,2))
|
|
#define HAVE_RHEL7_PCI_RESET_NOTIFY
|
|
#endif /* RHEL >= 7.2 */
|
|
#if (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,3))
|
|
#if (RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,5))
|
|
#define HAVE_GENEVE_RX_OFFLOAD
|
|
#endif /* RHEL < 7.5 */
|
|
#define HAVE_ETHTOOL_FLOW_UNION_IP6_SPEC
|
|
#define HAVE_RHEL7_NET_DEVICE_OPS_EXT
|
|
#if !defined(HAVE_UDP_ENC_TUNNEL) && IS_ENABLED(CONFIG_GENEVE)
|
|
#define HAVE_UDP_ENC_TUNNEL
|
|
#endif /* !HAVE_UDP_ENC_TUNNEL && CONFIG_GENEVE */
|
|
#endif /* RHEL >= 7.3 */
|
|
|
|
/* new hooks added to net_device_ops_extended in RHEL7.4 */
|
|
#if (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,4))
|
|
#define HAVE_RHEL7_NETDEV_OPS_EXT_NDO_SET_VF_VLAN
|
|
#define HAVE_RHEL7_NETDEV_OPS_EXT_NDO_UDP_TUNNEL
|
|
#define HAVE_UDP_ENC_RX_OFFLOAD
|
|
#endif /* RHEL >= 7.4 */
|
|
#else /* RHEL >= 8.0 */
|
|
#define HAVE_TCF_BLOCK_CB_REGISTER_EXTACK
|
|
#define NO_NETDEV_BPF_PROG_ATTACHED
|
|
#define HAVE_NDO_SELECT_QUEUE_SB_DEV
|
|
#endif /* RHEL >= 8.0 */
|
|
#endif /* RHEL >= 7.0 */
|
|
#endif /* >= 3.10.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,11,0) )
|
|
#define netdev_notifier_info_to_dev(ptr) ptr
|
|
#ifndef time_in_range64
|
|
#define time_in_range64(a, b, c) \
|
|
(time_after_eq64(a, b) && \
|
|
time_before_eq64(a, c))
|
|
#endif /* time_in_range64 */
|
|
#if ((RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,6)) ||\
|
|
(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(11,4,0)))
|
|
#define HAVE_NDO_SET_VF_LINK_STATE
|
|
#endif
|
|
#if RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,2))
|
|
#define HAVE_NDO_SELECT_QUEUE_ACCEL_FALLBACK
|
|
#endif
|
|
#else /* >= 3.11.0 */
|
|
#define HAVE_NDO_SET_VF_LINK_STATE
|
|
#define HAVE_SKB_INNER_PROTOCOL
|
|
#define HAVE_MPLS_FEATURES
|
|
#endif /* >= 3.11.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,12,0) )
|
|
int __kc_pcie_get_minimum_link(struct pci_dev *dev, enum pci_bus_speed *speed,
|
|
enum pcie_link_width *width);
|
|
#ifndef pcie_get_minimum_link
|
|
#define pcie_get_minimum_link(_p, _s, _w) __kc_pcie_get_minimum_link(_p, _s, _w)
|
|
#endif
|
|
|
|
#if (RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(6,7))
|
|
int _kc_pci_wait_for_pending_transaction(struct pci_dev *dev);
|
|
#define pci_wait_for_pending_transaction _kc_pci_wait_for_pending_transaction
|
|
#endif /* <RHEL6.7 */
|
|
|
|
#else /* >= 3.12.0 */
|
|
#if ( SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(12,0,0))
|
|
#define HAVE_NDO_SELECT_QUEUE_ACCEL_FALLBACK
|
|
#endif
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(4,8,0) )
|
|
#define HAVE_VXLAN_RX_OFFLOAD
|
|
#if !defined(HAVE_UDP_ENC_TUNNEL) && IS_ENABLED(CONFIG_VXLAN)
|
|
#define HAVE_UDP_ENC_TUNNEL
|
|
#endif
|
|
#endif /* < 4.8.0 */
|
|
#define HAVE_NDO_GET_PHYS_PORT_ID
|
|
#define HAVE_NETIF_SET_XPS_QUEUE_CONST_MASK
|
|
#endif /* >= 3.12.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,13,0) )
|
|
#define dma_set_mask_and_coherent(_p, _m) __kc_dma_set_mask_and_coherent(_p, _m)
|
|
int __kc_dma_set_mask_and_coherent(struct device *dev, u64 mask);
|
|
#ifndef u64_stats_init
|
|
#define u64_stats_init(a) do { } while(0)
|
|
#endif
|
|
#undef BIT_ULL
|
|
#define BIT_ULL(n) (1ULL << (n))
|
|
|
|
#if (!(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(12,0,0)) && \
|
|
!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,0)))
|
|
static inline struct pci_dev *pci_upstream_bridge(struct pci_dev *dev)
|
|
{
|
|
dev = pci_physfn(dev);
|
|
if (pci_is_root_bus(dev->bus))
|
|
return NULL;
|
|
|
|
return dev->bus->self;
|
|
}
|
|
#endif
|
|
|
|
#if (SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(12,1,0))
|
|
#undef HAVE_STRUCT_PAGE_PFMEMALLOC
|
|
#define HAVE_DCBNL_OPS_SETAPP_RETURN_INT
|
|
#endif
|
|
#ifndef list_next_entry
|
|
#define list_next_entry(pos, member) \
|
|
list_entry((pos)->member.next, typeof(*(pos)), member)
|
|
#endif
|
|
#ifndef list_prev_entry
|
|
#define list_prev_entry(pos, member) \
|
|
list_entry((pos)->member.prev, typeof(*(pos)), member)
|
|
#endif
|
|
|
|
#if ( LINUX_VERSION_CODE > KERNEL_VERSION(2,6,20) )
|
|
#define devm_kcalloc(dev, cnt, size, flags) \
|
|
devm_kzalloc(dev, (cnt) * (size), flags)
|
|
#endif /* > 2.6.20 */
|
|
|
|
#if (!(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,2)))
|
|
#define list_last_entry(ptr, type, member) list_entry((ptr)->prev, type, member)
|
|
#endif
|
|
|
|
#if (RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,0))
|
|
bool _kc_pci_device_is_present(struct pci_dev *pdev);
|
|
#define pci_device_is_present _kc_pci_device_is_present
|
|
#endif /* <RHEL7.0 */
|
|
#else /* >= 3.13.0 */
|
|
#define HAVE_VXLAN_CHECKS
|
|
#if (UBUNTU_VERSION_CODE && UBUNTU_VERSION_CODE >= UBUNTU_VERSION(3,13,0,24))
|
|
#define HAVE_NDO_SELECT_QUEUE_ACCEL_FALLBACK
|
|
#else
|
|
#define HAVE_NDO_SELECT_QUEUE_ACCEL
|
|
#endif
|
|
#define HAVE_HWMON_DEVICE_REGISTER_WITH_GROUPS
|
|
#endif
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,14,0) )
|
|
|
|
#ifndef U16_MAX
|
|
#define U16_MAX ((u16)~0U)
|
|
#endif
|
|
|
|
#ifndef U32_MAX
|
|
#define U32_MAX ((u32)~0U)
|
|
#endif
|
|
|
|
#ifndef U64_MAX
|
|
#define U64_MAX ((u64)~0ULL)
|
|
#endif
|
|
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,2)))
|
|
#define dev_consume_skb_any(x) dev_kfree_skb_any(x)
|
|
#define dev_consume_skb_irq(x) dev_kfree_skb_irq(x)
|
|
#endif
|
|
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,0)) && \
|
|
!(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(12,0,0)))
|
|
|
|
/* it isn't expected that this would be a #define unless we made it so */
|
|
#ifndef skb_set_hash
|
|
|
|
#define PKT_HASH_TYPE_NONE 0
|
|
#define PKT_HASH_TYPE_L2 1
|
|
#define PKT_HASH_TYPE_L3 2
|
|
#define PKT_HASH_TYPE_L4 3
|
|
|
|
enum _kc_pkt_hash_types {
|
|
_KC_PKT_HASH_TYPE_NONE = PKT_HASH_TYPE_NONE,
|
|
_KC_PKT_HASH_TYPE_L2 = PKT_HASH_TYPE_L2,
|
|
_KC_PKT_HASH_TYPE_L3 = PKT_HASH_TYPE_L3,
|
|
_KC_PKT_HASH_TYPE_L4 = PKT_HASH_TYPE_L4,
|
|
};
|
|
#define pkt_hash_types _kc_pkt_hash_types
|
|
|
|
#define skb_set_hash __kc_skb_set_hash
|
|
static inline void __kc_skb_set_hash(struct sk_buff __maybe_unused *skb,
|
|
u32 __maybe_unused hash,
|
|
int __maybe_unused type)
|
|
{
|
|
#ifdef HAVE_SKB_L4_RXHASH
|
|
skb->l4_rxhash = (type == PKT_HASH_TYPE_L4);
|
|
#endif
|
|
#ifdef NETIF_F_RXHASH
|
|
skb->rxhash = hash;
|
|
#endif
|
|
}
|
|
#endif /* !skb_set_hash */
|
|
|
|
#else /* RHEL_RELEASE_CODE >= 7.0 || SLE_VERSION_CODE >= 12.0 */
|
|
|
|
#if ((RHEL_RELEASE_CODE && RHEL_RELEASE_CODE <= RHEL_RELEASE_VERSION(7,0)) ||\
|
|
(SLE_VERSION_CODE && SLE_VERSION_CODE <= SLE_VERSION(12,1,0)))
|
|
/* GPLv2 code taken from 5.10-rc2 kernel source include/linux/pci.h, Copyright
|
|
* original authors.
|
|
*/
|
|
static inline int pci_enable_msix_exact(struct pci_dev *dev,
|
|
struct msix_entry *entries, int nvec)
|
|
{
|
|
int rc = pci_enable_msix_range(dev, entries, nvec, nvec);
|
|
if (rc < 0)
|
|
return rc;
|
|
return 0;
|
|
}
|
|
#endif /* <=EL7.0 || <=SLES 12.1 */
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,5)))
|
|
#ifndef HAVE_VXLAN_RX_OFFLOAD
|
|
#define HAVE_VXLAN_RX_OFFLOAD
|
|
#endif /* HAVE_VXLAN_RX_OFFLOAD */
|
|
#endif
|
|
|
|
#if !defined(HAVE_UDP_ENC_TUNNEL) && IS_ENABLED(CONFIG_VXLAN)
|
|
#define HAVE_UDP_ENC_TUNNEL
|
|
#endif
|
|
|
|
#ifndef HAVE_VXLAN_CHECKS
|
|
#define HAVE_VXLAN_CHECKS
|
|
#endif /* HAVE_VXLAN_CHECKS */
|
|
#endif /* !(RHEL_RELEASE_CODE >= 7.0 && SLE_VERSION_CODE >= 12.0) */
|
|
|
|
#if ((RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,3)) ||\
|
|
(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(12,0,0)))
|
|
#define HAVE_NDO_DFWD_OPS
|
|
#endif
|
|
|
|
#ifndef pci_enable_msix_range
|
|
int __kc_pci_enable_msix_range(struct pci_dev *dev, struct msix_entry *entries,
|
|
int minvec, int maxvec);
|
|
#define pci_enable_msix_range __kc_pci_enable_msix_range
|
|
#endif
|
|
|
|
#ifndef ether_addr_copy
|
|
#define ether_addr_copy __kc_ether_addr_copy
|
|
static inline void __kc_ether_addr_copy(u8 *dst, const u8 *src)
|
|
{
|
|
#if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)
|
|
*(u32 *)dst = *(const u32 *)src;
|
|
*(u16 *)(dst + 4) = *(const u16 *)(src + 4);
|
|
#else
|
|
u16 *a = (u16 *)dst;
|
|
const u16 *b = (const u16 *)src;
|
|
|
|
a[0] = b[0];
|
|
a[1] = b[1];
|
|
a[2] = b[2];
|
|
#endif
|
|
}
|
|
#endif /* ether_addr_copy */
|
|
int __kc_ipv6_find_hdr(const struct sk_buff *skb, unsigned int *offset,
|
|
int target, unsigned short *fragoff, int *flags);
|
|
#define ipv6_find_hdr(a, b, c, d, e) __kc_ipv6_find_hdr((a), (b), (c), (d), (e))
|
|
|
|
#ifndef OPTIMIZE_HIDE_VAR
|
|
#ifdef __GNUC__
|
|
#define OPTIMIZER_HIDE_VAR(var) __asm__ ("" : "=r" (var) : "0" (var))
|
|
#else
|
|
#include <linux/barrier.h>
|
|
#define OPTIMIZE_HIDE_VAR(var) barrier()
|
|
#endif
|
|
#endif
|
|
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,0)) && \
|
|
!(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(10,4,0)))
|
|
static inline __u32 skb_get_hash_raw(const struct sk_buff *skb)
|
|
{
|
|
#ifdef NETIF_F_RXHASH
|
|
return skb->rxhash;
|
|
#else
|
|
return 0;
|
|
#endif /* NETIF_F_RXHASH */
|
|
}
|
|
#endif /* !RHEL > 5.9 && !SLES >= 10.4 */
|
|
|
|
#if (RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,5))
|
|
#define request_firmware_direct request_firmware
|
|
#endif /* !RHEL || RHEL < 7.5 */
|
|
|
|
#else /* >= 3.14.0 */
|
|
|
|
/* for ndo_dfwd_ ops add_station, del_station and _start_xmit */
|
|
#ifndef HAVE_NDO_DFWD_OPS
|
|
#define HAVE_NDO_DFWD_OPS
|
|
#endif
|
|
#define HAVE_NDO_SELECT_QUEUE_ACCEL_FALLBACK
|
|
#endif /* 3.14.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,15,0) )
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(2,6,35) )
|
|
#define HAVE_SKBUFF_RXHASH
|
|
#endif /* >= 2.6.35 */
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,1)) && \
|
|
!(UBUNTU_VERSION_CODE && UBUNTU_VERSION_CODE >= UBUNTU_VERSION(3,13,0,30)))
|
|
#define u64_stats_fetch_begin_irq u64_stats_fetch_begin_bh
|
|
#define u64_stats_fetch_retry_irq u64_stats_fetch_retry_bh
|
|
#endif
|
|
|
|
char *_kc_devm_kstrdup(struct device *dev, const char *s, gfp_t gfp);
|
|
#define devm_kstrdup(dev, s, gfp) _kc_devm_kstrdup(dev, s, gfp)
|
|
|
|
#else /* >= 3.15.0 */
|
|
#define HAVE_NET_GET_RANDOM_ONCE
|
|
#define HAVE_PTP_1588_CLOCK_PINS
|
|
#define HAVE_NETDEV_PORT
|
|
#endif /* 3.15.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,16,0) )
|
|
#ifndef smp_mb__before_atomic
|
|
#define smp_mb__before_atomic() smp_mb()
|
|
#define smp_mb__after_atomic() smp_mb()
|
|
#endif
|
|
#ifndef __dev_uc_sync
|
|
#ifdef HAVE_SET_RX_MODE
|
|
#ifdef NETDEV_HW_ADDR_T_UNICAST
|
|
int __kc_hw_addr_sync_dev(struct netdev_hw_addr_list *list,
|
|
struct net_device *dev,
|
|
int (*sync)(struct net_device *, const unsigned char *),
|
|
int (*unsync)(struct net_device *, const unsigned char *));
|
|
void __kc_hw_addr_unsync_dev(struct netdev_hw_addr_list *list,
|
|
struct net_device *dev,
|
|
int (*unsync)(struct net_device *, const unsigned char *));
|
|
#endif
|
|
#ifndef NETDEV_HW_ADDR_T_MULTICAST
|
|
int __kc_dev_addr_sync_dev(struct dev_addr_list **list, int *count,
|
|
struct net_device *dev,
|
|
int (*sync)(struct net_device *, const unsigned char *),
|
|
int (*unsync)(struct net_device *, const unsigned char *));
|
|
void __kc_dev_addr_unsync_dev(struct dev_addr_list **list, int *count,
|
|
struct net_device *dev,
|
|
int (*unsync)(struct net_device *, const unsigned char *));
|
|
#endif
|
|
#endif /* HAVE_SET_RX_MODE */
|
|
|
|
static inline int __kc_dev_uc_sync(struct net_device __maybe_unused *dev,
|
|
int __maybe_unused (*sync)(struct net_device *, const unsigned char *),
|
|
int __maybe_unused (*unsync)(struct net_device *, const unsigned char *))
|
|
{
|
|
#ifdef NETDEV_HW_ADDR_T_UNICAST
|
|
return __kc_hw_addr_sync_dev(&dev->uc, dev, sync, unsync);
|
|
#elif defined(HAVE_SET_RX_MODE)
|
|
return __kc_dev_addr_sync_dev(&dev->uc_list, &dev->uc_count,
|
|
dev, sync, unsync);
|
|
#else
|
|
return 0;
|
|
#endif
|
|
}
|
|
#define __dev_uc_sync __kc_dev_uc_sync
|
|
|
|
static inline void __kc_dev_uc_unsync(struct net_device __maybe_unused *dev,
|
|
int __maybe_unused (*unsync)(struct net_device *, const unsigned char *))
|
|
{
|
|
#ifdef HAVE_SET_RX_MODE
|
|
#ifdef NETDEV_HW_ADDR_T_UNICAST
|
|
__kc_hw_addr_unsync_dev(&dev->uc, dev, unsync);
|
|
#else /* NETDEV_HW_ADDR_T_MULTICAST */
|
|
__kc_dev_addr_unsync_dev(&dev->uc_list, &dev->uc_count, dev, unsync);
|
|
#endif /* NETDEV_HW_ADDR_T_UNICAST */
|
|
#endif /* HAVE_SET_RX_MODE */
|
|
}
|
|
#define __dev_uc_unsync __kc_dev_uc_unsync
|
|
|
|
static inline int __kc_dev_mc_sync(struct net_device __maybe_unused *dev,
|
|
int __maybe_unused (*sync)(struct net_device *, const unsigned char *),
|
|
int __maybe_unused (*unsync)(struct net_device *, const unsigned char *))
|
|
{
|
|
#ifdef NETDEV_HW_ADDR_T_MULTICAST
|
|
return __kc_hw_addr_sync_dev(&dev->mc, dev, sync, unsync);
|
|
#elif defined(HAVE_SET_RX_MODE)
|
|
return __kc_dev_addr_sync_dev(&dev->mc_list, &dev->mc_count,
|
|
dev, sync, unsync);
|
|
#else
|
|
return 0;
|
|
#endif
|
|
|
|
}
|
|
#define __dev_mc_sync __kc_dev_mc_sync
|
|
|
|
static inline void __kc_dev_mc_unsync(struct net_device __maybe_unused *dev,
|
|
int __maybe_unused (*unsync)(struct net_device *, const unsigned char *))
|
|
{
|
|
#ifdef HAVE_SET_RX_MODE
|
|
#ifdef NETDEV_HW_ADDR_T_MULTICAST
|
|
__kc_hw_addr_unsync_dev(&dev->mc, dev, unsync);
|
|
#else /* NETDEV_HW_ADDR_T_MULTICAST */
|
|
__kc_dev_addr_unsync_dev(&dev->mc_list, &dev->mc_count, dev, unsync);
|
|
#endif /* NETDEV_HW_ADDR_T_MULTICAST */
|
|
#endif /* HAVE_SET_RX_MODE */
|
|
}
|
|
#define __dev_mc_unsync __kc_dev_mc_unsync
|
|
#endif /* __dev_uc_sync */
|
|
|
|
#if RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,1))
|
|
#define HAVE_NDO_SET_VF_MIN_MAX_TX_RATE
|
|
#endif
|
|
|
|
#ifndef NETIF_F_GSO_UDP_TUNNEL_CSUM
|
|
/* if someone backports this, hopefully they backport as a #define.
|
|
* declare it as zero on older kernels so that if it get's or'd in
|
|
* it won't effect anything, therefore preventing core driver changes
|
|
*/
|
|
#define NETIF_F_GSO_UDP_TUNNEL_CSUM 0
|
|
#define SKB_GSO_UDP_TUNNEL_CSUM 0
|
|
#endif
|
|
void *__kc_devm_kmemdup(struct device *dev, const void *src, size_t len,
|
|
gfp_t gfp);
|
|
#define devm_kmemdup __kc_devm_kmemdup
|
|
|
|
#else
|
|
#if ( ( LINUX_VERSION_CODE < KERNEL_VERSION(4,13,0) ) && \
|
|
! ( SLE_VERSION_CODE && ( SLE_VERSION_CODE >= SLE_VERSION(12,4,0)) ) )
|
|
#define HAVE_PCI_ERROR_HANDLER_RESET_NOTIFY
|
|
#endif /* >= 3.16.0 && < 4.13.0 && !(SLES >= 12sp4) */
|
|
#define HAVE_NDO_SET_VF_MIN_MAX_TX_RATE
|
|
#endif /* 3.16.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,17,0) )
|
|
#if !(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,8) && \
|
|
RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,0)) && \
|
|
!(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,2))
|
|
#ifndef timespec64
|
|
#define timespec64 timespec
|
|
static inline struct timespec64 timespec_to_timespec64(const struct timespec ts)
|
|
{
|
|
return ts;
|
|
}
|
|
static inline struct timespec timespec64_to_timespec(const struct timespec64 ts64)
|
|
{
|
|
return ts64;
|
|
}
|
|
#define timespec64_equal timespec_equal
|
|
#define timespec64_compare timespec_compare
|
|
#define set_normalized_timespec64 set_normalized_timespec
|
|
#define timespec64_add_safe timespec_add_safe
|
|
#define timespec64_add timespec_add
|
|
#define timespec64_sub timespec_sub
|
|
#define timespec64_valid timespec_valid
|
|
#define timespec64_valid_strict timespec_valid_strict
|
|
#define timespec64_to_ns timespec_to_ns
|
|
#define ns_to_timespec64 ns_to_timespec
|
|
#define ktime_to_timespec64 ktime_to_timespec
|
|
#define ktime_get_ts64 ktime_get_ts
|
|
#define ktime_get_real_ts64 ktime_get_real_ts
|
|
#define timespec64_add_ns timespec_add_ns
|
|
#endif /* timespec64 */
|
|
#endif /* !(RHEL6.8<RHEL7.0) && !RHEL7.2+ */
|
|
|
|
#if (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,8) && \
|
|
RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,0))
|
|
static inline void ktime_get_real_ts64(struct timespec64 *ts)
|
|
{
|
|
*ts = ktime_to_timespec64(ktime_get_real());
|
|
}
|
|
|
|
static inline void ktime_get_ts64(struct timespec64 *ts)
|
|
{
|
|
*ts = ktime_to_timespec64(ktime_get());
|
|
}
|
|
#endif
|
|
|
|
#if !(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,4))
|
|
#define hlist_add_behind(_a, _b) hlist_add_after(_b, _a)
|
|
#endif
|
|
|
|
#if (RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,5))
|
|
#endif /* RHEL_RELEASE_CODE < RHEL7.5 */
|
|
|
|
#if RHEL_RELEASE_CODE && \
|
|
RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(6,3) && \
|
|
RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,3)
|
|
static inline u64 ktime_get_ns(void)
|
|
{
|
|
return ktime_to_ns(ktime_get());
|
|
}
|
|
|
|
static inline u64 ktime_get_real_ns(void)
|
|
{
|
|
return ktime_to_ns(ktime_get_real());
|
|
}
|
|
|
|
static inline u64 ktime_get_boot_ns(void)
|
|
{
|
|
return ktime_to_ns(ktime_get_boottime());
|
|
}
|
|
#endif /* RHEL < 7.3 */
|
|
|
|
#else
|
|
#define HAVE_DCBNL_OPS_SETAPP_RETURN_INT
|
|
#include <linux/time64.h>
|
|
#define HAVE_RHASHTABLE
|
|
#endif /* 3.17.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,18,0) )
|
|
#ifndef NO_PTP_SUPPORT
|
|
#include <linux/errqueue.h>
|
|
struct sk_buff *__kc_skb_clone_sk(struct sk_buff *skb);
|
|
void __kc_skb_complete_tx_timestamp(struct sk_buff *skb,
|
|
struct skb_shared_hwtstamps *hwtstamps);
|
|
#define skb_clone_sk __kc_skb_clone_sk
|
|
#define skb_complete_tx_timestamp __kc_skb_complete_tx_timestamp
|
|
#endif
|
|
#if (!(RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,2))))
|
|
u32 __kc_eth_get_headlen(const struct net_device *dev, unsigned char *data,
|
|
unsigned int max_len);
|
|
#else
|
|
unsigned int __kc_eth_get_headlen(unsigned char *data, unsigned int max_len);
|
|
#endif /* !RHEL >= 8.2 */
|
|
|
|
#define eth_get_headlen __kc_eth_get_headlen
|
|
#ifndef ETH_P_XDSA
|
|
#define ETH_P_XDSA 0x00F8
|
|
#endif
|
|
/* RHEL 7.1 backported csum_level, but SLES 12 and 12-SP1 did not */
|
|
#if RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,1))
|
|
#define HAVE_SKBUFF_CSUM_LEVEL
|
|
#endif /* >= RH 7.1 */
|
|
|
|
/* RHEL 7.3 backported xmit_more */
|
|
#if (RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,3))
|
|
#define HAVE_SKB_XMIT_MORE
|
|
#endif /* >= RH 7.3 */
|
|
|
|
#undef GENMASK
|
|
#define GENMASK(h, l) \
|
|
(((~0UL) << (l)) & (~0UL >> (BITS_PER_LONG - 1 - (h))))
|
|
#undef GENMASK_ULL
|
|
#define GENMASK_ULL(h, l) \
|
|
(((~0ULL) << (l)) & (~0ULL >> (BITS_PER_LONG_LONG - 1 - (h))))
|
|
|
|
#else /* 3.18.0 */
|
|
#define HAVE_SKBUFF_CSUM_LEVEL
|
|
#define HAVE_SKB_XMIT_MORE
|
|
#define HAVE_SKB_INNER_PROTOCOL_TYPE
|
|
#endif /* 3.18.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,18,4) )
|
|
#else
|
|
#define HAVE_NDO_FEATURES_CHECK
|
|
#endif /* 3.18.4 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,18,13) )
|
|
#ifndef WRITE_ONCE
|
|
#define WRITE_ONCE(x, val) ({ ACCESS_ONCE(x) = (val); })
|
|
#endif
|
|
#endif /* 3.18.13 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,19,0) )
|
|
/* netdev_phys_port_id renamed to netdev_phys_item_id */
|
|
#define netdev_phys_item_id netdev_phys_port_id
|
|
|
|
static inline void _kc_napi_complete_done(struct napi_struct *napi,
|
|
int __always_unused work_done) {
|
|
napi_complete(napi);
|
|
}
|
|
/* don't use our backport if the distro kernels already have it */
|
|
#if (SLE_VERSION_CODE && (SLE_VERSION_CODE < SLE_VERSION(12,3,0))) || \
|
|
(RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,5)))
|
|
#define napi_complete_done _kc_napi_complete_done
|
|
#endif
|
|
|
|
int _kc_bitmap_print_to_pagebuf(bool list, char *buf,
|
|
const unsigned long *maskp, int nmaskbits);
|
|
#define bitmap_print_to_pagebuf _kc_bitmap_print_to_pagebuf
|
|
|
|
#ifndef NETDEV_RSS_KEY_LEN
|
|
#define NETDEV_RSS_KEY_LEN (13 * 4)
|
|
#endif
|
|
#if (!(RHEL_RELEASE_CODE && \
|
|
((RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,7) && RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,0)) || \
|
|
(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,2)))))
|
|
#define netdev_rss_key_fill(buffer, len) __kc_netdev_rss_key_fill(buffer, len)
|
|
#endif /* RHEL_RELEASE_CODE */
|
|
void __kc_netdev_rss_key_fill(void *buffer, size_t len);
|
|
#define SPEED_20000 20000
|
|
#define SPEED_40000 40000
|
|
#ifndef dma_rmb
|
|
#define dma_rmb() rmb()
|
|
#endif
|
|
#ifndef dev_alloc_pages
|
|
#ifndef NUMA_NO_NODE
|
|
#define NUMA_NO_NODE -1
|
|
#endif
|
|
#define dev_alloc_pages(_order) alloc_pages_node(NUMA_NO_NODE, (GFP_ATOMIC | __GFP_COLD | __GFP_COMP | __GFP_MEMALLOC), (_order))
|
|
#endif
|
|
#ifndef dev_alloc_page
|
|
#define dev_alloc_page() dev_alloc_pages(0)
|
|
#endif
|
|
#if !defined(eth_skb_pad) && !defined(skb_put_padto)
|
|
/**
|
|
* __kc_skb_put_padto - increase size and pad an skbuff up to a minimal size
|
|
* @skb: buffer to pad
|
|
* @len: minimal length
|
|
*
|
|
* Pads up a buffer to ensure the trailing bytes exist and are
|
|
* blanked. If the buffer already contains sufficient data it
|
|
* is untouched. Otherwise it is extended. Returns zero on
|
|
* success. The skb is freed on error.
|
|
*/
|
|
static inline int __kc_skb_put_padto(struct sk_buff *skb, unsigned int len)
|
|
{
|
|
unsigned int size = skb->len;
|
|
|
|
if (unlikely(size < len)) {
|
|
len -= size;
|
|
if (skb_pad(skb, len))
|
|
return -ENOMEM;
|
|
__skb_put(skb, len);
|
|
}
|
|
return 0;
|
|
}
|
|
#define skb_put_padto(skb, len) __kc_skb_put_padto(skb, len)
|
|
|
|
static inline int __kc_eth_skb_pad(struct sk_buff *skb)
|
|
{
|
|
return __kc_skb_put_padto(skb, ETH_ZLEN);
|
|
}
|
|
#define eth_skb_pad(skb) __kc_eth_skb_pad(skb)
|
|
#endif /* eth_skb_pad && skb_put_padto */
|
|
|
|
#ifndef SKB_ALLOC_NAPI
|
|
/* RHEL 7.2 backported napi_alloc_skb and friends */
|
|
static inline struct sk_buff *__kc_napi_alloc_skb(struct napi_struct *napi, unsigned int length)
|
|
{
|
|
return netdev_alloc_skb_ip_align(napi->dev, length);
|
|
}
|
|
#define napi_alloc_skb(napi,len) __kc_napi_alloc_skb(napi,len)
|
|
#define __napi_alloc_skb(napi,len,mask) __kc_napi_alloc_skb(napi,len)
|
|
#endif /* SKB_ALLOC_NAPI */
|
|
#define HAVE_CONFIG_PM_RUNTIME
|
|
#if (RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(6,7)) && \
|
|
(RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,0)))
|
|
#define HAVE_RXFH_HASHFUNC
|
|
#endif /* 6.7 < RHEL < 7.0 */
|
|
#if RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,1))
|
|
#define HAVE_RXFH_HASHFUNC
|
|
#define NDO_DFLT_BRIDGE_GETLINK_HAS_BRFLAGS
|
|
#endif /* RHEL > 7.1 */
|
|
#ifndef napi_schedule_irqoff
|
|
#define napi_schedule_irqoff napi_schedule
|
|
#endif
|
|
#ifndef READ_ONCE
|
|
#define READ_ONCE(_x) ACCESS_ONCE(_x)
|
|
#endif
|
|
#if RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,2))
|
|
#define HAVE_NDO_FDB_ADD_VID
|
|
#endif
|
|
#ifndef ETH_MODULE_SFF_8636
|
|
#define ETH_MODULE_SFF_8636 0x3
|
|
#endif
|
|
#ifndef ETH_MODULE_SFF_8636_LEN
|
|
#define ETH_MODULE_SFF_8636_LEN 256
|
|
#endif
|
|
#ifndef ETH_MODULE_SFF_8436
|
|
#define ETH_MODULE_SFF_8436 0x4
|
|
#endif
|
|
#ifndef ETH_MODULE_SFF_8436_LEN
|
|
#define ETH_MODULE_SFF_8436_LEN 256
|
|
#endif
|
|
#ifndef writel_relaxed
|
|
#define writel_relaxed writel
|
|
#endif
|
|
#else /* 3.19.0 */
|
|
#define HAVE_NDO_FDB_ADD_VID
|
|
#define HAVE_RXFH_HASHFUNC
|
|
#define NDO_DFLT_BRIDGE_GETLINK_HAS_BRFLAGS
|
|
#endif /* 3.19.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,20,0) )
|
|
/* vlan_tx_xx functions got renamed to skb_vlan */
|
|
#ifndef skb_vlan_tag_get
|
|
#define skb_vlan_tag_get vlan_tx_tag_get
|
|
#endif
|
|
#ifndef skb_vlan_tag_present
|
|
#define skb_vlan_tag_present vlan_tx_tag_present
|
|
#endif
|
|
#if RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,1))
|
|
#define HAVE_INCLUDE_LINUX_TIMECOUNTER_H
|
|
#endif
|
|
#if RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,2))
|
|
#define HAVE_NDO_BRIDGE_SET_DEL_LINK_FLAGS
|
|
#endif
|
|
#else
|
|
#define HAVE_INCLUDE_LINUX_TIMECOUNTER_H
|
|
#define HAVE_NDO_BRIDGE_SET_DEL_LINK_FLAGS
|
|
#endif /* 3.20.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(4,0,0) )
|
|
/* Definition for CONFIG_OF was introduced earlier */
|
|
#if !defined(CONFIG_OF) && \
|
|
!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,2))
|
|
static inline struct device_node *
|
|
pci_device_to_OF_node(const struct pci_dev __always_unused *pdev) { return NULL; }
|
|
#else /* !CONFIG_OF && RHEL < 7.3 */
|
|
#define HAVE_DDP_PROFILE_UPLOAD_SUPPORT
|
|
#endif /* !CONFIG_OF && RHEL < 7.3 */
|
|
#else /* < 4.0 */
|
|
#define HAVE_DDP_PROFILE_UPLOAD_SUPPORT
|
|
#endif /* < 4.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(4,1,0) )
|
|
#ifndef NO_PTP_SUPPORT
|
|
#ifdef HAVE_INCLUDE_LINUX_TIMECOUNTER_H
|
|
#include <linux/timecounter.h>
|
|
#else
|
|
#include <linux/clocksource.h>
|
|
#endif
|
|
static inline void __kc_timecounter_adjtime(struct timecounter *tc, s64 delta)
|
|
{
|
|
tc->nsec += delta;
|
|
}
|
|
|
|
static inline struct net_device *
|
|
of_find_net_device_by_node(struct device_node __always_unused *np)
|
|
{
|
|
return NULL;
|
|
}
|
|
|
|
#define timecounter_adjtime __kc_timecounter_adjtime
|
|
#endif
|
|
#if ((RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,2))) || \
|
|
(SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(12,2,0))))
|
|
#define HAVE_NDO_SET_VF_RSS_QUERY_EN
|
|
#endif
|
|
#if RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,2))
|
|
#define HAVE_NDO_BRIDGE_GETLINK_NLFLAGS
|
|
#define HAVE_RHEL7_EXTENDED_NDO_SET_TX_MAXRATE
|
|
#define HAVE_NDO_SET_TX_MAXRATE
|
|
#endif
|
|
#if !((RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(6,8) && RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,0)) && \
|
|
(RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,2)) && \
|
|
(SLE_VERSION_CODE > SLE_VERSION(12,1,0)))
|
|
unsigned int _kc_cpumask_local_spread(unsigned int i, int node);
|
|
#define cpumask_local_spread _kc_cpumask_local_spread
|
|
#endif
|
|
#ifdef HAVE_RHASHTABLE
|
|
#define rhashtable_loopup_fast(ht, key, params) \
|
|
do { \
|
|
(void)params; \
|
|
rhashtable_lookup((ht), (key)); \
|
|
} while (0)
|
|
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(3,19,0) )
|
|
#define rhashtable_insert_fast(ht, obj, params) \
|
|
do { \
|
|
(void)params; \
|
|
rhashtable_insert((ht), (obj), GFP_KERNEL); \
|
|
} while (0)
|
|
|
|
#define rhashtable_remove_fast(ht, obj, params) \
|
|
do { \
|
|
(void)params; \
|
|
rhashtable_remove((ht), (obj), GFP_KERNEL); \
|
|
} while (0)
|
|
|
|
#else /* >= 3,19,0 */
|
|
#define rhashtable_insert_fast(ht, obj, params) \
|
|
do { \
|
|
(void)params; \
|
|
rhashtable_insert((ht), (obj)); \
|
|
} while (0)
|
|
|
|
#define rhashtable_remove_fast(ht, obj, params) \
|
|
do { \
|
|
(void)params; \
|
|
rhashtable_remove((ht), (obj)); \
|
|
} while (0)
|
|
|
|
#endif /* 3,19,0 */
|
|
#endif /* HAVE_RHASHTABLE */
|
|
#else /* >= 4,1,0 */
|
|
#define HAVE_NDO_GET_PHYS_PORT_NAME
|
|
#define HAVE_PTP_CLOCK_INFO_GETTIME64
|
|
#define HAVE_NDO_BRIDGE_GETLINK_NLFLAGS
|
|
#define HAVE_PASSTHRU_FEATURES_CHECK
|
|
#define HAVE_NDO_SET_VF_RSS_QUERY_EN
|
|
#define HAVE_NDO_SET_TX_MAXRATE
|
|
#endif /* 4,1,0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,1,9))
|
|
#if (!(RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,2)) && \
|
|
!((SLE_VERSION_CODE == SLE_VERSION(11,3,0)) && \
|
|
(SLE_LOCALVERSION_CODE >= SLE_LOCALVERSION(0,47,71))) && \
|
|
!((SLE_VERSION_CODE == SLE_VERSION(11,4,0)) && \
|
|
(SLE_LOCALVERSION_CODE >= SLE_LOCALVERSION(65,0,0))) && \
|
|
!(SLE_VERSION_CODE >= SLE_VERSION(12,1,0)))
|
|
static inline bool page_is_pfmemalloc(struct page __maybe_unused *page)
|
|
{
|
|
#ifdef HAVE_STRUCT_PAGE_PFMEMALLOC
|
|
return page->pfmemalloc;
|
|
#else
|
|
return false;
|
|
#endif
|
|
}
|
|
#endif /* !RHEL7.2+ && !SLES11sp3(3.0.101-0.47.71+ update) && !SLES11sp4(3.0.101-65+ update) & !SLES12sp1+ */
|
|
#else
|
|
#undef HAVE_STRUCT_PAGE_PFMEMALLOC
|
|
#endif /* 4.1.9 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,2,0))
|
|
#if (!(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,2)) && \
|
|
!(SLE_VERSION_CODE >= SLE_VERSION(12,1,0)))
|
|
#define ETHTOOL_RX_FLOW_SPEC_RING 0x00000000FFFFFFFFULL
|
|
#define ETHTOOL_RX_FLOW_SPEC_RING_VF 0x000000FF00000000ULL
|
|
#define ETHTOOL_RX_FLOW_SPEC_RING_VF_OFF 32
|
|
static inline __u64 ethtool_get_flow_spec_ring(__u64 ring_cookie)
|
|
{
|
|
return ETHTOOL_RX_FLOW_SPEC_RING & ring_cookie;
|
|
};
|
|
|
|
static inline __u64 ethtool_get_flow_spec_ring_vf(__u64 ring_cookie)
|
|
{
|
|
return (ETHTOOL_RX_FLOW_SPEC_RING_VF & ring_cookie) >>
|
|
ETHTOOL_RX_FLOW_SPEC_RING_VF_OFF;
|
|
};
|
|
#endif /* ! RHEL >= 7.2 && ! SLES >= 12.1 */
|
|
#if (RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,4))
|
|
#define HAVE_NDO_DFLT_BRIDGE_GETLINK_VLAN_SUPPORT
|
|
#endif
|
|
|
|
#if (LINUX_VERSION_CODE > KERNEL_VERSION(2,6,27))
|
|
#if (!((RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,8) && \
|
|
RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,0)) || \
|
|
RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,2)))
|
|
static inline bool pci_ari_enabled(struct pci_bus *bus)
|
|
{
|
|
return bus->self && bus->self->ari_enabled;
|
|
}
|
|
#if (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,2))
|
|
#define HAVE_VF_STATS
|
|
#endif /* (RHEL7.2+) */
|
|
#endif /* !(RHEL6.8+ || RHEL7.2+) */
|
|
#else
|
|
static inline bool pci_ari_enabled(struct pci_bus *bus)
|
|
{
|
|
return false;
|
|
}
|
|
#endif /* 2.6.27 */
|
|
#else
|
|
#define HAVE_NDO_DFLT_BRIDGE_GETLINK_VLAN_SUPPORT
|
|
#define HAVE_VF_STATS
|
|
#endif /* 4.2.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,3,0))
|
|
#if (!(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,4)) && \
|
|
!(SLE_VERSION_CODE >= SLE_VERSION(12,2,0)))
|
|
/**
|
|
* _kc_flow_dissector_key_ipv4_addrs:
|
|
* @src: source ip address
|
|
* @dst: destination ip address
|
|
*/
|
|
struct _kc_flow_dissector_key_ipv4_addrs {
|
|
__be32 src;
|
|
__be32 dst;
|
|
};
|
|
|
|
/**
|
|
* _kc_flow_dissector_key_ipv6_addrs:
|
|
* @src: source ip address
|
|
* @dst: destination ip address
|
|
*/
|
|
struct _kc_flow_dissector_key_ipv6_addrs {
|
|
struct in6_addr src;
|
|
struct in6_addr dst;
|
|
};
|
|
|
|
/**
|
|
* _kc_flow_dissector_key_addrs:
|
|
* @v4addrs: IPv4 addresses
|
|
* @v6addrs: IPv6 addresses
|
|
*/
|
|
struct _kc_flow_dissector_key_addrs {
|
|
union {
|
|
struct _kc_flow_dissector_key_ipv4_addrs v4addrs;
|
|
struct _kc_flow_dissector_key_ipv6_addrs v6addrs;
|
|
};
|
|
};
|
|
|
|
/**
|
|
* _kc_flow_dissector_key_tp_ports:
|
|
* @ports: port numbers of Transport header
|
|
* src: source port number
|
|
* dst: destination port number
|
|
*/
|
|
struct _kc_flow_dissector_key_ports {
|
|
union {
|
|
__be32 ports;
|
|
struct {
|
|
__be16 src;
|
|
__be16 dst;
|
|
};
|
|
};
|
|
};
|
|
|
|
/**
|
|
* _kc_flow_dissector_key_basic:
|
|
* @n_proto: Network header protocol (eg. IPv4/IPv6)
|
|
* @ip_proto: Transport header protocol (eg. TCP/UDP)
|
|
* @padding: padding for alignment
|
|
*/
|
|
struct _kc_flow_dissector_key_basic {
|
|
__be16 n_proto;
|
|
u8 ip_proto;
|
|
u8 padding;
|
|
};
|
|
|
|
struct _kc_flow_keys {
|
|
struct _kc_flow_dissector_key_basic basic;
|
|
struct _kc_flow_dissector_key_ports ports;
|
|
struct _kc_flow_dissector_key_addrs addrs;
|
|
};
|
|
|
|
/* These are all the include files for kernels inside this #ifdef block that
|
|
* have any reference to the in kernel definition of struct flow_keys. The
|
|
* reason for putting them here is to make 100% sure that these files do not get
|
|
* included after re-defining flow_keys to _kc_flow_keys. This is done to
|
|
* prevent any possible ABI issues that this structure re-definition could case.
|
|
*/
|
|
#if ((LINUX_VERSION_CODE >= KERNEL_VERSION(3,3,0) && \
|
|
LINUX_VERSION_CODE < KERNEL_VERSION(4,2,0)) || \
|
|
RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(6,7) || \
|
|
SLE_VERSION_CODE >= SLE_VERSION(11,4,0))
|
|
#include <net/flow_keys.h>
|
|
#endif /* (>= 3.3.0 && < 4.2.0) || >= RHEL 6.7 || >= SLE 11.4 */
|
|
#if (LINUX_VERSION_CODE == KERNEL_VERSION(4,2,0))
|
|
#include <net/flow_dissector.h>
|
|
#endif /* 4.2.0 */
|
|
#include <linux/skbuff.h>
|
|
#include <net/ip.h>
|
|
#include <net/ipv6.h>
|
|
#include <net/flow.h>
|
|
|
|
#define flow_keys _kc_flow_keys
|
|
bool
|
|
_kc_skb_flow_dissect_flow_keys(const struct sk_buff *skb,
|
|
struct flow_keys *flow,
|
|
unsigned int __always_unused flags);
|
|
#define skb_flow_dissect_flow_keys _kc_skb_flow_dissect_flow_keys
|
|
#endif /* ! >= RHEL 7.4 && ! >= SLES 12.2 */
|
|
|
|
#if ((RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,3)) || \
|
|
(SLE_VERSION_CODE >= SLE_VERSION(12,2,0)))
|
|
#include <net/dst_metadata.h>
|
|
#endif /* >= RHEL7.3 || >= SLE12sp2 */
|
|
#else /* >= 4.3.0 */
|
|
#include <net/dst_metadata.h>
|
|
#endif /* 4.3.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,4,0))
|
|
#if (RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,3))
|
|
#define HAVE_NDO_SET_VF_TRUST
|
|
#endif /* (RHEL_RELEASE >= 7.3) */
|
|
#ifndef CONFIG_64BIT
|
|
#if (LINUX_VERSION_CODE >= KERNEL_VERSION(3,3,0))
|
|
#include <asm-generic/io-64-nonatomic-lo-hi.h> /* 32-bit readq/writeq */
|
|
#else /* 3.3.0 => 4.3.x */
|
|
#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,26))
|
|
#include <asm-generic/int-ll64.h>
|
|
#endif /* 2.6.26 => 3.3.0 */
|
|
#ifndef readq
|
|
static inline __u64 readq(const volatile void __iomem *addr)
|
|
{
|
|
const volatile u32 __iomem *p = addr;
|
|
u32 low, high;
|
|
|
|
low = readl(p);
|
|
high = readl(p + 1);
|
|
|
|
return low + ((u64)high << 32);
|
|
}
|
|
#define readq readq
|
|
#endif
|
|
|
|
#ifndef writeq
|
|
static inline void writeq(__u64 val, volatile void __iomem *addr)
|
|
{
|
|
writel(val, addr);
|
|
writel(val >> 32, (u8 *)addr + 4);
|
|
}
|
|
#define writeq writeq
|
|
#endif
|
|
#endif /* < 3.3.0 */
|
|
#endif /* !CONFIG_64BIT */
|
|
#else /* < 4.4.0 */
|
|
#define HAVE_NDO_SET_VF_TRUST
|
|
|
|
#ifndef CONFIG_64BIT
|
|
#include <linux/io-64-nonatomic-lo-hi.h> /* 32-bit readq/writeq */
|
|
#endif /* !CONFIG_64BIT */
|
|
#endif /* 4.4.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,5,0))
|
|
/* protect against a likely backport */
|
|
#ifndef NETIF_F_CSUM_MASK
|
|
#define NETIF_F_CSUM_MASK NETIF_F_ALL_CSUM
|
|
#endif /* NETIF_F_CSUM_MASK */
|
|
#ifndef NETIF_F_SCTP_CRC
|
|
#define NETIF_F_SCTP_CRC NETIF_F_SCTP_CSUM
|
|
#endif /* NETIF_F_SCTP_CRC */
|
|
#if (!(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,3)))
|
|
#define eth_platform_get_mac_address _kc_eth_platform_get_mac_address
|
|
int _kc_eth_platform_get_mac_address(struct device *dev __maybe_unused,
|
|
u8 *mac_addr __maybe_unused);
|
|
#endif /* !(RHEL_RELEASE >= 7.3) */
|
|
#else /* 4.5.0 */
|
|
#if ( LINUX_VERSION_CODE < KERNEL_VERSION(4,8,0) )
|
|
#define HAVE_GENEVE_RX_OFFLOAD
|
|
#if !defined(HAVE_UDP_ENC_TUNNEL) && IS_ENABLED(CONFIG_GENEVE)
|
|
#define HAVE_UDP_ENC_TUNNEL
|
|
#endif
|
|
#endif /* < 4.8.0 */
|
|
#define HAVE_NETIF_NAPI_ADD_CALLS_NAPI_HASH_ADD
|
|
#define HAVE_NETDEV_UPPER_INFO
|
|
#endif /* 4.5.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,6,0))
|
|
#if !(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,3))
|
|
static inline unsigned char *skb_checksum_start(const struct sk_buff *skb)
|
|
{
|
|
#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,22))
|
|
return skb->head + skb->csum_start;
|
|
#else /* < 2.6.22 */
|
|
return skb_transport_header(skb);
|
|
#endif
|
|
}
|
|
#endif
|
|
|
|
#if !(UBUNTU_VERSION_CODE && \
|
|
UBUNTU_VERSION_CODE >= UBUNTU_VERSION(4,4,0,21)) && \
|
|
!(RHEL_RELEASE_CODE && \
|
|
(RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,2))) && \
|
|
!(SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(12,3,0)))
|
|
static inline void napi_consume_skb(struct sk_buff *skb,
|
|
int __always_unused budget)
|
|
{
|
|
dev_consume_skb_any(skb);
|
|
}
|
|
|
|
#endif /* UBUNTU 4,4,0,21, RHEL 7.2, SLES12 SP3 */
|
|
#if !(SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(12,3,0))) && \
|
|
!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,4))
|
|
static inline void csum_replace_by_diff(__sum16 *sum, __wsum diff)
|
|
{
|
|
* sum = csum_fold(csum_add(diff, ~csum_unfold(*sum)));
|
|
}
|
|
#endif
|
|
#if !(RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(7,2))) && \
|
|
!(SLE_VERSION_CODE && (SLE_VERSION_CODE > SLE_VERSION(12,3,0)))
|
|
static inline void page_ref_inc(struct page *page)
|
|
{
|
|
get_page(page);
|
|
}
|
|
#else
|
|
#define HAVE_PAGE_COUNT_BULK_UPDATE
|
|
#endif
|
|
#ifndef IPV4_USER_FLOW
|
|
#define IPV4_USER_FLOW 0x0d /* spec only (usr_ip4_spec) */
|
|
#endif
|
|
|
|
#if (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,4))
|
|
#define HAVE_TC_SETUP_CLSFLOWER
|
|
#define HAVE_TC_FLOWER_ENC
|
|
#endif
|
|
|
|
#if ((RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,7)) || \
|
|
(SLE_VERSION_CODE >= SLE_VERSION(12,2,0)))
|
|
#define HAVE_TC_SETUP_CLSU32
|
|
#endif
|
|
|
|
#if (SLE_VERSION_CODE >= SLE_VERSION(12,2,0))
|
|
#define HAVE_TC_SETUP_CLSFLOWER
|
|
#endif
|
|
|
|
#else /* >= 4.6.0 */
|
|
#define HAVE_PAGE_COUNT_BULK_UPDATE
|
|
#define HAVE_ETHTOOL_FLOW_UNION_IP6_SPEC
|
|
#define HAVE_PTP_CROSSTIMESTAMP
|
|
#define HAVE_TC_SETUP_CLSFLOWER
|
|
#define HAVE_TC_SETUP_CLSU32
|
|
#endif /* 4.6.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,7,0))
|
|
#if ((SLE_VERSION_CODE >= SLE_VERSION(12,3,0)) ||\
|
|
(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,4)))
|
|
#define HAVE_NETIF_TRANS_UPDATE
|
|
#endif /* SLES12sp3+ || RHEL7.4+ */
|
|
#if ((RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,3)) ||\
|
|
(SLE_VERSION_CODE >= SLE_VERSION(12,3,0)))
|
|
#define HAVE_ETHTOOL_25G_BITS
|
|
#define HAVE_ETHTOOL_50G_BITS
|
|
#define HAVE_ETHTOOL_100G_BITS
|
|
#endif /* RHEL7.3+ || SLES12sp3+ */
|
|
#else /* 4.7.0 */
|
|
#define HAVE_NETIF_TRANS_UPDATE
|
|
#define HAVE_ETHTOOL_CONVERT_U32_AND_LINK_MODE
|
|
#define HAVE_ETHTOOL_25G_BITS
|
|
#define HAVE_ETHTOOL_50G_BITS
|
|
#define HAVE_ETHTOOL_100G_BITS
|
|
#define HAVE_TCF_MIRRED_REDIRECT
|
|
#endif /* 4.7.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,8,0))
|
|
#if !(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,4))
|
|
enum udp_parsable_tunnel_type {
|
|
UDP_TUNNEL_TYPE_VXLAN,
|
|
UDP_TUNNEL_TYPE_GENEVE,
|
|
};
|
|
struct udp_tunnel_info {
|
|
unsigned short type;
|
|
sa_family_t sa_family;
|
|
__be16 port;
|
|
};
|
|
#endif
|
|
|
|
#if (UBUNTU_VERSION_CODE && UBUNTU_VERSION_CODE < UBUNTU_VERSION(4,8,0,0))
|
|
#define tc_no_actions(_exts) true
|
|
#define tc_for_each_action(_a, _exts) while (0)
|
|
#endif
|
|
#if !(SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(12,3,0))) &&\
|
|
!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,4))
|
|
static inline int
|
|
#ifdef HAVE_NON_CONST_PCI_DRIVER_NAME
|
|
pci_request_io_regions(struct pci_dev *pdev, char *name)
|
|
#else
|
|
pci_request_io_regions(struct pci_dev *pdev, const char *name)
|
|
#endif
|
|
{
|
|
return pci_request_selected_regions(pdev,
|
|
pci_select_bars(pdev, IORESOURCE_IO), name);
|
|
}
|
|
|
|
static inline void
|
|
pci_release_io_regions(struct pci_dev *pdev)
|
|
{
|
|
return pci_release_selected_regions(pdev,
|
|
pci_select_bars(pdev, IORESOURCE_IO));
|
|
}
|
|
|
|
static inline int
|
|
#ifdef HAVE_NON_CONST_PCI_DRIVER_NAME
|
|
pci_request_mem_regions(struct pci_dev *pdev, char *name)
|
|
#else
|
|
pci_request_mem_regions(struct pci_dev *pdev, const char *name)
|
|
#endif
|
|
{
|
|
return pci_request_selected_regions(pdev,
|
|
pci_select_bars(pdev, IORESOURCE_MEM), name);
|
|
}
|
|
|
|
static inline void
|
|
pci_release_mem_regions(struct pci_dev *pdev)
|
|
{
|
|
return pci_release_selected_regions(pdev,
|
|
pci_select_bars(pdev, IORESOURCE_MEM));
|
|
}
|
|
#endif /* !SLE_VERSION(12,3,0) */
|
|
#if ((RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,4)) ||\
|
|
(SLE_VERSION_CODE >= SLE_VERSION(12,3,0)))
|
|
#define HAVE_ETHTOOL_NEW_50G_BITS
|
|
#endif /* RHEL7.4+ || SLES12sp3+ */
|
|
#else
|
|
#define HAVE_UDP_ENC_RX_OFFLOAD
|
|
#define HAVE_ETHTOOL_NEW_50G_BITS
|
|
#endif /* 4.8.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,9,0))
|
|
#ifdef HAVE_TC_SETUP_CLSFLOWER
|
|
#if (!(RHEL_RELEASE_CODE) && !(SLE_VERSION_CODE) || \
|
|
(SLE_VERSION_CODE && (SLE_VERSION_CODE < SLE_VERSION(12,3,0))))
|
|
#define HAVE_TC_FLOWER_VLAN_IN_TAGS
|
|
#endif /* !RHEL_RELEASE_CODE && !SLE_VERSION_CODE || <SLE_VERSION(12,3,0) */
|
|
#endif /* HAVE_TC_SETUP_CLSFLOWER */
|
|
#if (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,4))
|
|
#define HAVE_ETHTOOL_NEW_1G_BITS
|
|
#define HAVE_ETHTOOL_NEW_10G_BITS
|
|
#endif /* RHEL7.4+ */
|
|
#if (!(SLE_VERSION_CODE) && !(RHEL_RELEASE_CODE)) || \
|
|
SLE_VERSION_CODE && (SLE_VERSION_CODE <= SLE_VERSION(12,3,0)) || \
|
|
RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE <= RHEL_RELEASE_VERSION(7,5))
|
|
#define time_is_before_jiffies64(a) time_after64(get_jiffies_64(), a)
|
|
#endif /* !SLE_VERSION_CODE && !RHEL_RELEASE_CODE || (SLES <= 12.3.0) || (RHEL <= 7.5) */
|
|
#if (RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,4))
|
|
static inline void bitmap_from_u64(unsigned long *dst, u64 mask)
|
|
{
|
|
dst[0] = mask & ULONG_MAX;
|
|
|
|
if (sizeof(mask) > sizeof(unsigned long))
|
|
dst[1] = mask >> 32;
|
|
}
|
|
#endif /* <RHEL7.4 */
|
|
#if (!(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,4)) && \
|
|
!(SLE_VERSION_CODE >= SLE_VERSION(12,3,0)) && \
|
|
!(UBUNTU_VERSION_CODE >= UBUNTU_VERSION(4,13,0,16)))
|
|
static inline bool eth_type_vlan(__be16 ethertype)
|
|
{
|
|
switch (ethertype) {
|
|
case htons(ETH_P_8021Q):
|
|
#ifdef ETH_P_8021AD
|
|
case htons(ETH_P_8021AD):
|
|
#endif
|
|
return true;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
#endif /* Linux < 4.9 || RHEL < 7.4 || SLES < 12.3 || Ubuntu < 4.3.0-16 */
|
|
#else /* >=4.9 */
|
|
#define HAVE_FLOW_DISSECTOR_KEY_VLAN_PRIO
|
|
#define HAVE_ETHTOOL_NEW_1G_BITS
|
|
#define HAVE_ETHTOOL_NEW_10G_BITS
|
|
#endif /* KERNEL_VERSION(4.9.0) */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,10,0))
|
|
/* SLES 12.3 and RHEL 7.5 backported this interface */
|
|
#if (!SLE_VERSION_CODE && !RHEL_RELEASE_CODE) || \
|
|
(SLE_VERSION_CODE && (SLE_VERSION_CODE < SLE_VERSION(12,3,0))) || \
|
|
(RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(7,5)))
|
|
static inline bool _kc_napi_complete_done2(struct napi_struct *napi,
|
|
int __always_unused work_done)
|
|
{
|
|
/* it was really hard to get napi_complete_done to be safe to call
|
|
* recursively without running into our own kcompat, so just use
|
|
* napi_complete
|
|
*/
|
|
napi_complete(napi);
|
|
|
|
/* true means that the stack is telling the driver to go-ahead and
|
|
* re-enable interrupts
|
|
*/
|
|
return true;
|
|
}
|
|
|
|
#ifdef napi_complete_done
|
|
#undef napi_complete_done
|
|
#endif
|
|
#define napi_complete_done _kc_napi_complete_done2
|
|
#endif /* sles and rhel exclusion for < 4.10 */
|
|
#if (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,4))
|
|
#define HAVE_DEV_WALK_API
|
|
#define HAVE_ETHTOOL_NEW_2500MB_BITS
|
|
#define HAVE_ETHTOOL_5G_BITS
|
|
#endif /* RHEL7.4+ */
|
|
#if (SLE_VERSION_CODE && (SLE_VERSION_CODE == SLE_VERSION(12,3,0)))
|
|
#define HAVE_STRUCT_DMA_ATTRS
|
|
#endif /* (SLES == 12.3.0) */
|
|
#if (SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(12,3,0)))
|
|
#define HAVE_NETDEVICE_MIN_MAX_MTU
|
|
#endif /* (SLES >= 12.3.0) */
|
|
#if (RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,5)))
|
|
#define HAVE_STRUCT_DMA_ATTRS
|
|
#define HAVE_RHEL7_EXTENDED_MIN_MAX_MTU
|
|
#define HAVE_NETDEVICE_MIN_MAX_MTU
|
|
#endif
|
|
#if (!(SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(12,3,0))) && \
|
|
!(RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,5))))
|
|
#ifndef dma_map_page_attrs
|
|
#define dma_map_page_attrs __kc_dma_map_page_attrs
|
|
static inline dma_addr_t __kc_dma_map_page_attrs(struct device *dev,
|
|
struct page *page,
|
|
size_t offset, size_t size,
|
|
enum dma_data_direction dir,
|
|
unsigned long __always_unused attrs)
|
|
{
|
|
return dma_map_page(dev, page, offset, size, dir);
|
|
}
|
|
#endif
|
|
|
|
#ifndef dma_unmap_page_attrs
|
|
#define dma_unmap_page_attrs __kc_dma_unmap_page_attrs
|
|
static inline void __kc_dma_unmap_page_attrs(struct device *dev,
|
|
dma_addr_t addr, size_t size,
|
|
enum dma_data_direction dir,
|
|
unsigned long __always_unused attrs)
|
|
{
|
|
dma_unmap_page(dev, addr, size, dir);
|
|
}
|
|
#endif
|
|
|
|
static inline void __page_frag_cache_drain(struct page *page,
|
|
unsigned int count)
|
|
{
|
|
#ifdef HAVE_PAGE_COUNT_BULK_UPDATE
|
|
if (!page_ref_sub_and_test(page, count))
|
|
return;
|
|
|
|
init_page_count(page);
|
|
#else
|
|
BUG_ON(count > 1);
|
|
if (!count)
|
|
return;
|
|
#endif
|
|
__free_pages(page, compound_order(page));
|
|
}
|
|
#endif /* !SLE_VERSION(12,3,0) && !RHEL_VERSION(7,5) */
|
|
#if ((SLE_VERSION_CODE && (SLE_VERSION_CODE > SLE_VERSION(12,3,0))) ||\
|
|
(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,5)))
|
|
#define HAVE_SWIOTLB_SKIP_CPU_SYNC
|
|
#endif
|
|
|
|
#if ((SLE_VERSION_CODE && (SLE_VERSION_CODE < SLE_VERSION(15,0,0))) ||\
|
|
(RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE <= RHEL_RELEASE_VERSION(7,4))))
|
|
#define page_frag_free __free_page_frag
|
|
#endif
|
|
#ifndef ETH_MIN_MTU
|
|
#define ETH_MIN_MTU 68
|
|
#endif /* ETH_MIN_MTU */
|
|
|
|
/* If kernel is older than 4.10 but distro is RHEL >= 7.5 || SLES > 12SP4,
|
|
* it does have support for NAPI_STATE
|
|
*/
|
|
#if ((RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,5))) ||\
|
|
(SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(12,4,0))))
|
|
#define HAVE_NAPI_STATE_IN_BUSY_POLL
|
|
#endif /* RHEL >= 7.5 || SLES >=12.4 */
|
|
#else /* >= 4.10 */
|
|
#define HAVE_TC_FLOWER_ENC
|
|
#define HAVE_NETDEVICE_MIN_MAX_MTU
|
|
#define HAVE_SWIOTLB_SKIP_CPU_SYNC
|
|
#define HAVE_NETDEV_TC_RESETS_XPS
|
|
#define HAVE_XPS_QOS_SUPPORT
|
|
#define HAVE_DEV_WALK_API
|
|
#define HAVE_ETHTOOL_NEW_2500MB_BITS
|
|
#define HAVE_ETHTOOL_5G_BITS
|
|
/* kernel 4.10 onwards, as part of busy_poll rewrite, new state were added
|
|
* which is part of NAPI:state. If NAPI:state=NAPI_STATE_IN_BUSY_POLL,
|
|
* it means napi_poll is invoked in busy_poll context
|
|
*/
|
|
#define HAVE_NAPI_STATE_IN_BUSY_POLL
|
|
#define HAVE_TCF_MIRRED_EGRESS_REDIRECT
|
|
#define HAVE_PTP_CLOCK_INFO_ADJFINE
|
|
#endif /* 4.10.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,11,0))
|
|
#ifdef CONFIG_NET_RX_BUSY_POLL
|
|
#define HAVE_NDO_BUSY_POLL
|
|
#endif /* CONFIG_NET_RX_BUSY_POLL */
|
|
#if ((SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(12,3,0))) || \
|
|
(RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,5))))
|
|
#define HAVE_VOID_NDO_GET_STATS64
|
|
#endif /* (SLES >= 12.3.0) && (RHEL >= 7.5) */
|
|
|
|
static inline void _kc_dev_kfree_skb_irq(struct sk_buff *skb)
|
|
{
|
|
if (!skb)
|
|
return;
|
|
dev_kfree_skb_irq(skb);
|
|
}
|
|
|
|
#undef dev_kfree_skb_irq
|
|
#define dev_kfree_skb_irq _kc_dev_kfree_skb_irq
|
|
|
|
static inline void _kc_dev_consume_skb_irq(struct sk_buff *skb)
|
|
{
|
|
if (!skb)
|
|
return;
|
|
dev_consume_skb_irq(skb);
|
|
}
|
|
|
|
#undef dev_consume_skb_irq
|
|
#define dev_consume_skb_irq _kc_dev_consume_skb_irq
|
|
|
|
static inline void _kc_dev_kfree_skb_any(struct sk_buff *skb)
|
|
{
|
|
if (!skb)
|
|
return;
|
|
dev_kfree_skb_any(skb);
|
|
}
|
|
|
|
#undef dev_kfree_skb_any
|
|
#define dev_kfree_skb_any _kc_dev_kfree_skb_any
|
|
|
|
static inline void _kc_dev_consume_skb_any(struct sk_buff *skb)
|
|
{
|
|
if (!skb)
|
|
return;
|
|
dev_consume_skb_any(skb);
|
|
}
|
|
|
|
#undef dev_consume_skb_any
|
|
#define dev_consume_skb_any _kc_dev_consume_skb_any
|
|
|
|
#else /* > 4.11 */
|
|
#define HAVE_VOID_NDO_GET_STATS64
|
|
#define HAVE_VM_OPS_FAULT_NO_VMA
|
|
#endif /* 4.11.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,12,0))
|
|
#if (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,7) && \
|
|
RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(8,0))
|
|
/* The RHEL 7.7+ NL_SET_ERR_MSG_MOD triggers unused parameter warnings */
|
|
#undef NL_SET_ERR_MSG_MOD
|
|
#endif
|
|
/* If kernel is older than 4.12 but distro is RHEL >= 7.5 || SLES > 12SP4,
|
|
* it does have support for MIN_NAPI_ID
|
|
*/
|
|
#if ((RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,5))) || \
|
|
(SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(12,4,0))))
|
|
#define HAVE_MIN_NAPI_ID
|
|
#endif /* RHEL >= 7.5 || SLES >= 12.4 */
|
|
#ifndef NL_SET_ERR_MSG_MOD
|
|
#define NL_SET_ERR_MSG_MOD(extack, msg) \
|
|
do { \
|
|
uninitialized_var(extack); \
|
|
pr_err(KBUILD_MODNAME ": " msg); \
|
|
} while (0)
|
|
#endif /* !NL_SET_ERR_MSG_MOD */
|
|
#else /* >= 4.12 */
|
|
#define HAVE_NAPI_BUSY_LOOP
|
|
#define HAVE_MIN_NAPI_ID
|
|
#endif /* 4.12 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,13,0))
|
|
#if ((SLE_VERSION_CODE && (SLE_VERSION_CODE > SLE_VERSION(12,3,0))) || \
|
|
(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,5)))
|
|
#define HAVE_TCF_EXTS_HAS_ACTION
|
|
#endif
|
|
#define PCI_EXP_LNKCAP_SLS_8_0GB 0x00000003 /* LNKCAP2 SLS Vector bit 2 */
|
|
#if (SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(12,4,0)))
|
|
#define HAVE_PCI_ERROR_HANDLER_RESET_PREPARE
|
|
#endif /* SLES >= 12sp4 */
|
|
#if (!(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,5)) && \
|
|
!(SLE_VERSION_CODE >= SLE_VERSION(12,4,0)))
|
|
#define UUID_SIZE 16
|
|
typedef struct {
|
|
__u8 b[UUID_SIZE];
|
|
} uuid_t;
|
|
#define UUID_INIT(a, b, c, d0, d1, d2, d3, d4, d5, d6, d7) \
|
|
((uuid_t) \
|
|
{{ ((a) >> 24) & 0xff, ((a) >> 16) & 0xff, ((a) >> 8) & 0xff, (a) & 0xff, \
|
|
((b) >> 8) & 0xff, (b) & 0xff, \
|
|
((c) >> 8) & 0xff, (c) & 0xff, \
|
|
(d0), (d1), (d2), (d3), (d4), (d5), (d6), (d7) }})
|
|
|
|
static inline bool uuid_equal(const uuid_t *u1, const uuid_t *u2)
|
|
{
|
|
return memcmp(u1, u2, sizeof(uuid_t)) == 0;
|
|
}
|
|
#else
|
|
#define HAVE_METADATA_PORT_INFO
|
|
#endif /* !(RHEL >= 7.5) && !(SLES >= 12.4) */
|
|
#else /* > 4.13 */
|
|
#define HAVE_METADATA_PORT_INFO
|
|
#define HAVE_HWTSTAMP_FILTER_NTP_ALL
|
|
#define HAVE_NDO_SETUP_TC_CHAIN_INDEX
|
|
#define HAVE_PCI_ERROR_HANDLER_RESET_PREPARE
|
|
#define HAVE_PTP_CLOCK_DO_AUX_WORK
|
|
#endif /* 4.13.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,14,0))
|
|
#ifdef ETHTOOL_GLINKSETTINGS
|
|
#ifndef ethtool_link_ksettings_del_link_mode
|
|
#define ethtool_link_ksettings_del_link_mode(ptr, name, mode) \
|
|
__clear_bit(ETHTOOL_LINK_MODE_ ## mode ## _BIT, (ptr)->link_modes.name)
|
|
#endif
|
|
#endif /* ETHTOOL_GLINKSETTINGS */
|
|
#if (SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(12,4,0)))
|
|
#define HAVE_NDO_SETUP_TC_REMOVE_TC_TO_NETDEV
|
|
#endif
|
|
|
|
#if (RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,5)))
|
|
#define HAVE_NDO_SETUP_TC_REMOVE_TC_TO_NETDEV
|
|
#define HAVE_RHEL7_NETDEV_OPS_EXT_NDO_SETUP_TC
|
|
#endif
|
|
|
|
#define TIMER_DATA_TYPE unsigned long
|
|
#define TIMER_FUNC_TYPE void (*)(TIMER_DATA_TYPE)
|
|
|
|
#define timer_setup(timer, callback, flags) \
|
|
__setup_timer((timer), (TIMER_FUNC_TYPE)(callback), \
|
|
(TIMER_DATA_TYPE)(timer), (flags))
|
|
|
|
#define from_timer(var, callback_timer, timer_fieldname) \
|
|
container_of(callback_timer, typeof(*var), timer_fieldname)
|
|
|
|
#ifndef xdp_do_flush_map
|
|
#define xdp_do_flush_map() do {} while (0)
|
|
#endif
|
|
struct _kc_xdp_buff {
|
|
void *data;
|
|
void *data_end;
|
|
void *data_hard_start;
|
|
};
|
|
#define xdp_buff _kc_xdp_buff
|
|
struct _kc_bpf_prog {
|
|
};
|
|
#define bpf_prog _kc_bpf_prog
|
|
#ifndef DIV_ROUND_DOWN_ULL
|
|
#define DIV_ROUND_DOWN_ULL(ll, d) \
|
|
({ unsigned long long _tmp = (ll); do_div(_tmp, d); _tmp; })
|
|
#endif /* DIV_ROUND_DOWN_ULL */
|
|
#else /* > 4.14 */
|
|
#define HAVE_XDP_SUPPORT
|
|
#define HAVE_NDO_SETUP_TC_REMOVE_TC_TO_NETDEV
|
|
#define HAVE_TCF_EXTS_HAS_ACTION
|
|
#endif /* 4.14.0 */
|
|
|
|
/*****************************************************************************/
|
|
#ifndef ETHTOOL_GLINKSETTINGS
|
|
|
|
#define __ETHTOOL_LINK_MODE_MASK_NBITS 32
|
|
#define ETHTOOL_LINK_MASK_SIZE BITS_TO_LONGS(__ETHTOOL_LINK_MODE_MASK_NBITS)
|
|
|
|
/**
|
|
* struct ethtool_link_ksettings
|
|
* @link_modes: supported and advertising, single item arrays
|
|
* @link_modes.supported: bitmask of supported link speeds
|
|
* @link_modes.advertising: bitmask of currently advertised speeds
|
|
* @base: base link details
|
|
* @base.speed: current link speed
|
|
* @base.port: current port type
|
|
* @base.duplex: current duplex mode
|
|
* @base.autoneg: current autonegotiation settings
|
|
*
|
|
* This struct and the following macros provide a way to support the old
|
|
* ethtool get/set_settings API on older kernels, but in the style of the new
|
|
* GLINKSETTINGS API. In this way, the same code can be used to support both
|
|
* APIs as seemlessly as possible.
|
|
*
|
|
* It should be noted the old API only has support up to the first 32 bits.
|
|
*/
|
|
struct ethtool_link_ksettings {
|
|
struct {
|
|
u32 speed;
|
|
u8 port;
|
|
u8 duplex;
|
|
u8 autoneg;
|
|
} base;
|
|
struct {
|
|
unsigned long supported[ETHTOOL_LINK_MASK_SIZE];
|
|
unsigned long advertising[ETHTOOL_LINK_MASK_SIZE];
|
|
} link_modes;
|
|
};
|
|
|
|
#define ETHTOOL_LINK_NAME_advertising(mode) ADVERTISED_ ## mode
|
|
#define ETHTOOL_LINK_NAME_supported(mode) SUPPORTED_ ## mode
|
|
#define ETHTOOL_LINK_NAME(name) ETHTOOL_LINK_NAME_ ## name
|
|
#define ETHTOOL_LINK_CONVERT(name, mode) ETHTOOL_LINK_NAME(name)(mode)
|
|
|
|
/**
|
|
* ethtool_link_ksettings_zero_link_mode
|
|
* @ptr: ptr to ksettings struct
|
|
* @name: supported or advertising
|
|
*/
|
|
#define ethtool_link_ksettings_zero_link_mode(ptr, name)\
|
|
(*((ptr)->link_modes.name) = 0x0)
|
|
|
|
/**
|
|
* ethtool_link_ksettings_add_link_mode
|
|
* @ptr: ptr to ksettings struct
|
|
* @name: supported or advertising
|
|
* @mode: link mode to add
|
|
*/
|
|
#define ethtool_link_ksettings_add_link_mode(ptr, name, mode)\
|
|
(*((ptr)->link_modes.name) |= (typeof(*((ptr)->link_modes.name)))ETHTOOL_LINK_CONVERT(name, mode))
|
|
|
|
/**
|
|
* ethtool_link_ksettings_del_link_mode
|
|
* @ptr: ptr to ksettings struct
|
|
* @name: supported or advertising
|
|
* @mode: link mode to delete
|
|
*/
|
|
#define ethtool_link_ksettings_del_link_mode(ptr, name, mode)\
|
|
(*((ptr)->link_modes.name) &= ~(typeof(*((ptr)->link_modes.name)))ETHTOOL_LINK_CONVERT(name, mode))
|
|
|
|
/**
|
|
* ethtool_link_ksettings_test_link_mode
|
|
* @ptr: ptr to ksettings struct
|
|
* @name: supported or advertising
|
|
* @mode: link mode to add
|
|
*/
|
|
#define ethtool_link_ksettings_test_link_mode(ptr, name, mode)\
|
|
(!!(*((ptr)->link_modes.name) & ETHTOOL_LINK_CONVERT(name, mode)))
|
|
|
|
/**
|
|
* _kc_ethtool_ksettings_to_cmd - Convert ethtool_link_ksettings to ethtool_cmd
|
|
* @ks: ethtool_link_ksettings struct
|
|
* @cmd: ethtool_cmd struct
|
|
*
|
|
* Convert an ethtool_link_ksettings structure into the older ethtool_cmd
|
|
* structure. We provide this in kcompat.h so that drivers can easily
|
|
* implement the older .{get|set}_settings as wrappers around the new api.
|
|
* Hence, we keep it prefixed with _kc_ to make it clear this isn't actually
|
|
* a real function in the kernel.
|
|
*/
|
|
static inline void
|
|
_kc_ethtool_ksettings_to_cmd(struct ethtool_link_ksettings *ks,
|
|
struct ethtool_cmd *cmd)
|
|
{
|
|
cmd->supported = (u32)ks->link_modes.supported[0];
|
|
cmd->advertising = (u32)ks->link_modes.advertising[0];
|
|
ethtool_cmd_speed_set(cmd, ks->base.speed);
|
|
cmd->duplex = ks->base.duplex;
|
|
cmd->autoneg = ks->base.autoneg;
|
|
cmd->port = ks->base.port;
|
|
}
|
|
|
|
#endif /* !ETHTOOL_GLINKSETTINGS */
|
|
|
|
/*****************************************************************************/
|
|
#if ((LINUX_VERSION_CODE < KERNEL_VERSION(4,14,0)) || \
|
|
(SLE_VERSION_CODE && (SLE_VERSION_CODE <= SLE_VERSION(12,3,0))) || \
|
|
(RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE <= RHEL_RELEASE_VERSION(7,5))))
|
|
#define phy_speed_to_str _kc_phy_speed_to_str
|
|
const char *_kc_phy_speed_to_str(int speed);
|
|
#else /* (LINUX >= 4.14.0) || (SLES > 12.3.0) || (RHEL > 7.5) */
|
|
#include <linux/phy.h>
|
|
#endif /* (LINUX < 4.14.0) || (SLES <= 12.3.0) || (RHEL <= 7.5) */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,15,0))
|
|
#if ((RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,6))) || \
|
|
(SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(15,1,0))))
|
|
#define HAVE_TC_CB_AND_SETUP_QDISC_MQPRIO
|
|
#define HAVE_TCF_BLOCK
|
|
#else /* RHEL >= 7.6 || SLES >= 15.1 */
|
|
#endif /* !(RHEL >= 7.6) && !(SLES >= 15.1) */
|
|
void _kc_ethtool_intersect_link_masks(struct ethtool_link_ksettings *dst,
|
|
struct ethtool_link_ksettings *src);
|
|
#define ethtool_intersect_link_masks _kc_ethtool_intersect_link_masks
|
|
#else /* >= 4.15 */
|
|
#define HAVE_NDO_BPF
|
|
#define HAVE_XDP_BUFF_DATA_META
|
|
#define HAVE_TC_CB_AND_SETUP_QDISC_MQPRIO
|
|
#define HAVE_TCF_BLOCK
|
|
#endif /* 4.15.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,16,0))
|
|
#if (!(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,7)) && \
|
|
!(SLE_VERSION_CODE >= SLE_VERSION(12,4,0) && \
|
|
SLE_VERSION_CODE < SLE_VERSION(15,0,0)) && \
|
|
!(SLE_VERSION_CODE >= SLE_VERSION(15,1,0)))
|
|
/* The return value of the strscpy() and strlcpy() functions is different.
|
|
* This could be potentially hazard for the future.
|
|
* To avoid this the void result is forced.
|
|
* So it is not possible use this function with the return value.
|
|
* Return value is required in kernel 4.3 through 4.15
|
|
*/
|
|
#define strscpy(...) (void)(strlcpy(__VA_ARGS__))
|
|
#endif /* !RHEL >= 7.7 && !SLES12sp4+ && !SLES15sp1+ */
|
|
|
|
#define pci_printk(level, pdev, fmt, arg...) \
|
|
dev_printk(level, &(pdev)->dev, fmt, ##arg)
|
|
#define pci_emerg(pdev, fmt, arg...) dev_emerg(&(pdev)->dev, fmt, ##arg)
|
|
#define pci_alert(pdev, fmt, arg...) dev_alert(&(pdev)->dev, fmt, ##arg)
|
|
#define pci_crit(pdev, fmt, arg...) dev_crit(&(pdev)->dev, fmt, ##arg)
|
|
#define pci_err(pdev, fmt, arg...) dev_err(&(pdev)->dev, fmt, ##arg)
|
|
#define pci_warn(pdev, fmt, arg...) dev_warn(&(pdev)->dev, fmt, ##arg)
|
|
#define pci_notice(pdev, fmt, arg...) dev_notice(&(pdev)->dev, fmt, ##arg)
|
|
#define pci_info(pdev, fmt, arg...) dev_info(&(pdev)->dev, fmt, ##arg)
|
|
#define pci_dbg(pdev, fmt, arg...) dev_dbg(&(pdev)->dev, fmt, ##arg)
|
|
|
|
#ifndef array_index_nospec
|
|
static inline unsigned long _kc_array_index_mask_nospec(unsigned long index,
|
|
unsigned long size)
|
|
{
|
|
/*
|
|
* Always calculate and emit the mask even if the compiler
|
|
* thinks the mask is not needed. The compiler does not take
|
|
* into account the value of @index under speculation.
|
|
*/
|
|
OPTIMIZER_HIDE_VAR(index);
|
|
return ~(long)(index | (size - 1UL - index)) >> (BITS_PER_LONG - 1);
|
|
}
|
|
|
|
#define array_index_nospec(index, size) \
|
|
({ \
|
|
typeof(index) _i = (index); \
|
|
typeof(size) _s = (size); \
|
|
unsigned long _mask = _kc_array_index_mask_nospec(_i, _s); \
|
|
\
|
|
BUILD_BUG_ON(sizeof(_i) > sizeof(long)); \
|
|
BUILD_BUG_ON(sizeof(_s) > sizeof(long)); \
|
|
\
|
|
(typeof(_i)) (_i & _mask); \
|
|
})
|
|
#endif /* array_index_nospec */
|
|
#ifndef sizeof_field
|
|
#define sizeof_field(TYPE, MEMBER) (sizeof((((TYPE *)0)->MEMBER)))
|
|
#endif /* sizeof_field */
|
|
#if !(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,0)) && \
|
|
!(SLE_VERSION_CODE >= SLE_VERSION(12,5,0) && \
|
|
SLE_VERSION_CODE < SLE_VERSION(15,0,0) || \
|
|
SLE_VERSION_CODE >= SLE_VERSION(15,1,0))
|
|
/*
|
|
* Copy bitmap and clear tail bits in last word.
|
|
*/
|
|
static inline void
|
|
bitmap_copy_clear_tail(unsigned long *dst, const unsigned long *src, unsigned int nbits)
|
|
{
|
|
bitmap_copy(dst, src, nbits);
|
|
if (nbits % BITS_PER_LONG)
|
|
dst[nbits / BITS_PER_LONG] &= BITMAP_LAST_WORD_MASK(nbits);
|
|
}
|
|
|
|
/*
|
|
* On 32-bit systems bitmaps are represented as u32 arrays internally, and
|
|
* therefore conversion is not needed when copying data from/to arrays of u32.
|
|
*/
|
|
#if BITS_PER_LONG == 64
|
|
void bitmap_from_arr32(unsigned long *bitmap, const u32 *buf, unsigned int nbits);
|
|
#else
|
|
#define bitmap_from_arr32(bitmap, buf, nbits) \
|
|
bitmap_copy_clear_tail((unsigned long *) (bitmap), \
|
|
(const unsigned long *) (buf), (nbits))
|
|
#endif /* BITS_PER_LONG == 64 */
|
|
#endif /* !(RHEL >= 8.0) && !(SLES >= 12.5 && SLES < 15.0 || SLES >= 15.1) */
|
|
#else /* >= 4.16 */
|
|
#include <linux/nospec.h>
|
|
#define HAVE_XDP_BUFF_RXQ
|
|
#define HAVE_TC_FLOWER_OFFLOAD_COMMON_EXTACK
|
|
#define HAVE_TCF_MIRRED_DEV
|
|
#define HAVE_VF_STATS_DROPPED
|
|
#endif /* 4.16.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,17,0))
|
|
#include <linux/pci_regs.h>
|
|
#include <linux/pci.h>
|
|
#define PCIE_SPEED_16_0GT 0x17
|
|
#define PCI_EXP_LNKCAP_SLS_16_0GB 0x00000004 /* LNKCAP2 SLS Vector bit 3 */
|
|
#define PCI_EXP_LNKSTA_CLS_16_0GB 0x0004 /* Current Link Speed 16.0GT/s */
|
|
#define PCI_EXP_LNKCAP2_SLS_16_0GB 0x00000010 /* Supported Speed 16GT/s */
|
|
void _kc_pcie_print_link_status(struct pci_dev *dev);
|
|
#define pcie_print_link_status _kc_pcie_print_link_status
|
|
#else /* >= 4.17.0 */
|
|
#define HAVE_XDP_BUFF_IN_XDP_H
|
|
#endif /* 4.17.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,18,0))
|
|
#include "kcompat_overflow.h"
|
|
|
|
#if (SLE_VERSION_CODE < SLE_VERSION(15,1,0))
|
|
#define firmware_request_nowarn request_firmware_direct
|
|
#endif /* SLES < 15.1 */
|
|
|
|
#else
|
|
#include <linux/overflow.h>
|
|
#include <net/xdp_sock.h>
|
|
#define HAVE_XDP_FRAME_STRUCT
|
|
#define HAVE_XDP_SOCK
|
|
#define HAVE_NDO_XDP_XMIT_BULK_AND_FLAGS
|
|
#define NO_NDO_XDP_FLUSH
|
|
#define HAVE_AF_XDP_SUPPORT
|
|
#endif /* 4.18.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,19,0))
|
|
#if (RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,0)) && \
|
|
(RHEL_RELEASE_CODE <= RHEL_RELEASE_VERSION(8,2)))
|
|
#define HAVE_DEVLINK_REGIONS
|
|
#endif /* RHEL >= 8.0 && RHEL <= 8.2 */
|
|
#define bitmap_alloc(nbits, flags) \
|
|
kmalloc_array(BITS_TO_LONGS(nbits), sizeof(unsigned long), flags)
|
|
#define bitmap_zalloc(nbits, flags) bitmap_alloc(nbits, ((flags) | __GFP_ZERO))
|
|
#define bitmap_free(bitmap) kfree(bitmap)
|
|
#ifdef ETHTOOL_GLINKSETTINGS
|
|
#define ethtool_ks_clear(ptr, name) \
|
|
ethtool_link_ksettings_zero_link_mode(ptr, name)
|
|
#define ethtool_ks_add_mode(ptr, name, mode) \
|
|
ethtool_link_ksettings_add_link_mode(ptr, name, mode)
|
|
#define ethtool_ks_del_mode(ptr, name, mode) \
|
|
ethtool_link_ksettings_del_link_mode(ptr, name, mode)
|
|
#define ethtool_ks_test(ptr, name, mode) \
|
|
ethtool_link_ksettings_test_link_mode(ptr, name, mode)
|
|
#endif /* ETHTOOL_GLINKSETTINGS */
|
|
#define HAVE_NETPOLL_CONTROLLER
|
|
#define REQUIRE_PCI_CLEANUP_AER_ERROR_STATUS
|
|
#if (SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(15,1,0)))
|
|
#define HAVE_TCF_MIRRED_DEV
|
|
#define HAVE_NDO_SELECT_QUEUE_SB_DEV
|
|
#define HAVE_TCF_BLOCK_CB_REGISTER_EXTACK
|
|
#endif
|
|
|
|
static inline void __kc_metadata_dst_free(void *md_dst)
|
|
{
|
|
kfree(md_dst);
|
|
}
|
|
|
|
#define metadata_dst_free(md_dst) __kc_metadata_dst_free(md_dst)
|
|
#else /* >= 4.19.0 */
|
|
#define HAVE_TCF_BLOCK_CB_REGISTER_EXTACK
|
|
#define NO_NETDEV_BPF_PROG_ATTACHED
|
|
#define HAVE_NDO_SELECT_QUEUE_SB_DEV
|
|
#define HAVE_NETDEV_SB_DEV
|
|
#define HAVE_TCF_VLAN_TPID
|
|
#define HAVE_RHASHTABLE_TYPES
|
|
#define HAVE_DEVLINK_REGIONS
|
|
#define HAVE_DEVLINK_PARAMS
|
|
#endif /* 4.19.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,20,0))
|
|
#define HAVE_XDP_UMEM_PROPS
|
|
#ifdef HAVE_AF_XDP_SUPPORT
|
|
#ifndef napi_if_scheduled_mark_missed
|
|
static inline bool __kc_napi_if_scheduled_mark_missed(struct napi_struct *n)
|
|
{
|
|
unsigned long val, new;
|
|
|
|
do {
|
|
val = READ_ONCE(n->state);
|
|
if (val & NAPIF_STATE_DISABLE)
|
|
return true;
|
|
|
|
if (!(val & NAPIF_STATE_SCHED))
|
|
return false;
|
|
|
|
new = val | NAPIF_STATE_MISSED;
|
|
} while (cmpxchg(&n->state, val, new) != val);
|
|
|
|
return true;
|
|
}
|
|
|
|
#define napi_if_scheduled_mark_missed __kc_napi_if_scheduled_mark_missed
|
|
#endif /* !napi_if_scheduled_mark_missed */
|
|
#endif /* HAVE_AF_XDP_SUPPORT */
|
|
#if (RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,0)))
|
|
#define HAVE_DEVLINK_ESWITCH_OPS_EXTACK
|
|
#endif /* RHEL >= 8.0 */
|
|
#if ((SLE_VERSION_CODE >= SLE_VERSION(12,5,0) && \
|
|
SLE_VERSION_CODE < SLE_VERSION(15,0,0)) || \
|
|
(SLE_VERSION_CODE >= SLE_VERSION(15,1,0)))
|
|
#define HAVE_DEVLINK_ESWITCH_OPS_EXTACK
|
|
#endif /* SLE == 12sp5 || SLE >= 15sp1 */
|
|
#else /* >= 4.20.0 */
|
|
#define HAVE_DEVLINK_ESWITCH_OPS_EXTACK
|
|
#define HAVE_AF_XDP_ZC_SUPPORT
|
|
#define HAVE_VXLAN_TYPE
|
|
#define HAVE_ETF_SUPPORT /* Earliest TxTime First */
|
|
#endif /* 4.20.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(5,0,0))
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(8,0)))
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(4,12,0))
|
|
#define NETLINK_MAX_COOKIE_LEN 20
|
|
struct netlink_ext_ack {
|
|
const char *_msg;
|
|
const struct nlattr *bad_attr;
|
|
u8 cookie[NETLINK_MAX_COOKIE_LEN];
|
|
u8 cookie_len;
|
|
};
|
|
|
|
#endif /* < 4.12 */
|
|
static inline int _kc_dev_open(struct net_device *netdev,
|
|
struct netlink_ext_ack __always_unused *extack)
|
|
{
|
|
return dev_open(netdev);
|
|
}
|
|
|
|
#define dev_open _kc_dev_open
|
|
|
|
static inline int
|
|
_kc_dev_change_flags(struct net_device *netdev, unsigned int flags,
|
|
struct netlink_ext_ack __always_unused *extack)
|
|
{
|
|
return dev_change_flags(netdev, flags);
|
|
}
|
|
|
|
#define dev_change_flags _kc_dev_change_flags
|
|
#endif /* !(RHEL_RELEASE_CODE && RHEL > RHEL(8,0)) */
|
|
#if (RHEL_RELEASE_CODE && \
|
|
(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(7,7) && \
|
|
RHEL_RELEASE_CODE < RHEL_RELEASE_VERSION(8,0)) || \
|
|
(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,1)))
|
|
#define HAVE_PTP_SYS_OFFSET_EXTENDED_IOCTL
|
|
#define HAVE_PTP_CLOCK_INFO_GETTIMEX64
|
|
#else /* RHEL >= 7.7 && RHEL < 8.0 || RHEL >= 8.1 */
|
|
struct ptp_system_timestamp {
|
|
struct timespec64 pre_ts;
|
|
struct timespec64 post_ts;
|
|
};
|
|
|
|
static inline void
|
|
ptp_read_system_prets(struct ptp_system_timestamp __always_unused *sts)
|
|
{
|
|
;
|
|
}
|
|
|
|
static inline void
|
|
ptp_read_system_postts(struct ptp_system_timestamp __always_unused *sts)
|
|
{
|
|
;
|
|
}
|
|
#endif /* !(RHEL >= 7.7 && RHEL != 8.0) */
|
|
#if (RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,1)))
|
|
#define HAVE_NDO_BRIDGE_SETLINK_EXTACK
|
|
#endif /* RHEL 8.1 */
|
|
#if (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,2))
|
|
#define HAVE_TC_INDIR_BLOCK
|
|
#endif /* RHEL 8.2 */
|
|
#define INDIRECT_CALLABLE_DECLARE(x) x
|
|
#else /* >= 5.0.0 */
|
|
#define HAVE_PTP_SYS_OFFSET_EXTENDED_IOCTL
|
|
#define HAVE_PTP_CLOCK_INFO_GETTIMEX64
|
|
#define HAVE_NDO_BRIDGE_SETLINK_EXTACK
|
|
#define HAVE_DMA_ALLOC_COHERENT_ZEROES_MEM
|
|
#define HAVE_GENEVE_TYPE
|
|
#define HAVE_TC_INDIR_BLOCK
|
|
#define HAVE_INDIRECT_CALL_WRAPPER_HEADER
|
|
#endif /* 5.0.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(5,1,0))
|
|
#if (RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,1)))
|
|
#define HAVE_TC_FLOW_RULE_INFRASTRUCTURE
|
|
#define HAVE_NDO_FDB_ADD_EXTACK
|
|
#define HAVE_DEVLINK_INFO_GET
|
|
#define HAVE_DEVLINK_FLASH_UPDATE
|
|
#else /* RHEL < 8.1 */
|
|
#ifdef HAVE_TC_SETUP_CLSFLOWER
|
|
#include <net/pkt_cls.h>
|
|
|
|
struct flow_match {
|
|
struct flow_dissector *dissector;
|
|
void *mask;
|
|
void *key;
|
|
};
|
|
|
|
struct flow_match_basic {
|
|
struct flow_dissector_key_basic *key, *mask;
|
|
};
|
|
|
|
struct flow_match_control {
|
|
struct flow_dissector_key_control *key, *mask;
|
|
};
|
|
|
|
struct flow_match_eth_addrs {
|
|
struct flow_dissector_key_eth_addrs *key, *mask;
|
|
};
|
|
|
|
#ifdef HAVE_TC_FLOWER_ENC
|
|
struct flow_match_enc_keyid {
|
|
struct flow_dissector_key_keyid *key, *mask;
|
|
};
|
|
#endif
|
|
|
|
#ifndef HAVE_TC_FLOWER_VLAN_IN_TAGS
|
|
struct flow_match_vlan {
|
|
struct flow_dissector_key_vlan *key, *mask;
|
|
};
|
|
#endif
|
|
|
|
struct flow_match_ipv4_addrs {
|
|
struct flow_dissector_key_ipv4_addrs *key, *mask;
|
|
};
|
|
|
|
struct flow_match_ipv6_addrs {
|
|
struct flow_dissector_key_ipv6_addrs *key, *mask;
|
|
};
|
|
|
|
struct flow_match_ports {
|
|
struct flow_dissector_key_ports *key, *mask;
|
|
};
|
|
|
|
struct flow_rule {
|
|
struct flow_match match;
|
|
};
|
|
|
|
void flow_rule_match_basic(const struct flow_rule *rule,
|
|
struct flow_match_basic *out);
|
|
void flow_rule_match_control(const struct flow_rule *rule,
|
|
struct flow_match_control *out);
|
|
void flow_rule_match_eth_addrs(const struct flow_rule *rule,
|
|
struct flow_match_eth_addrs *out);
|
|
#ifndef HAVE_TC_FLOWER_VLAN_IN_TAGS
|
|
void flow_rule_match_vlan(const struct flow_rule *rule,
|
|
struct flow_match_vlan *out);
|
|
#endif
|
|
void flow_rule_match_ipv4_addrs(const struct flow_rule *rule,
|
|
struct flow_match_ipv4_addrs *out);
|
|
void flow_rule_match_ipv6_addrs(const struct flow_rule *rule,
|
|
struct flow_match_ipv6_addrs *out);
|
|
void flow_rule_match_ports(const struct flow_rule *rule,
|
|
struct flow_match_ports *out);
|
|
#ifdef HAVE_TC_FLOWER_ENC
|
|
void flow_rule_match_enc_ports(const struct flow_rule *rule,
|
|
struct flow_match_ports *out);
|
|
void flow_rule_match_enc_control(const struct flow_rule *rule,
|
|
struct flow_match_control *out);
|
|
void flow_rule_match_enc_ipv4_addrs(const struct flow_rule *rule,
|
|
struct flow_match_ipv4_addrs *out);
|
|
void flow_rule_match_enc_ipv6_addrs(const struct flow_rule *rule,
|
|
struct flow_match_ipv6_addrs *out);
|
|
void flow_rule_match_enc_keyid(const struct flow_rule *rule,
|
|
struct flow_match_enc_keyid *out);
|
|
#endif
|
|
|
|
static inline struct flow_rule *
|
|
tc_cls_flower_offload_flow_rule(struct tc_cls_flower_offload *tc_flow_cmd)
|
|
{
|
|
return (struct flow_rule *)&tc_flow_cmd->dissector;
|
|
}
|
|
|
|
static inline bool flow_rule_match_key(const struct flow_rule *rule,
|
|
enum flow_dissector_key_id key)
|
|
{
|
|
return dissector_uses_key(rule->match.dissector, key);
|
|
}
|
|
#endif /* HAVE_TC_SETUP_CLSFLOWER */
|
|
|
|
#endif /* RHEL < 8.1 */
|
|
|
|
#if (!(RHEL_RELEASE_CODE && RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,1)))
|
|
#define devlink_params_publish(devlink) do { } while (0)
|
|
#define devlink_params_unpublish(devlink) do { } while (0)
|
|
#endif
|
|
|
|
#else /* >= 5.1.0 */
|
|
#define HAVE_NDO_FDB_ADD_EXTACK
|
|
#define NO_XDP_QUERY_XSK_UMEM
|
|
#define HAVE_AF_XDP_NETDEV_UMEM
|
|
#define HAVE_TC_FLOW_RULE_INFRASTRUCTURE
|
|
#define HAVE_TC_FLOWER_ENC_IP
|
|
#define HAVE_DEVLINK_INFO_GET
|
|
#define HAVE_DEVLINK_FLASH_UPDATE
|
|
#define HAVE_DEVLINK_PORT_PARAMS
|
|
#endif /* 5.1.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(5,2,0))
|
|
#if (defined HAVE_SKB_XMIT_MORE) && \
|
|
(!(RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,2))))
|
|
#define netdev_xmit_more() (skb->xmit_more)
|
|
#else
|
|
#define netdev_xmit_more() (0)
|
|
#endif
|
|
|
|
#if (!(RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,2))))
|
|
#ifndef eth_get_headlen
|
|
static inline u32
|
|
__kc_eth_get_headlen(const struct net_device __always_unused *dev, void *data,
|
|
unsigned int len)
|
|
{
|
|
return eth_get_headlen(data, len);
|
|
}
|
|
|
|
#define eth_get_headlen(dev, data, len) __kc_eth_get_headlen(dev, data, len)
|
|
#endif /* !eth_get_headlen */
|
|
#endif /* !RHEL >= 8.2 */
|
|
|
|
#ifndef mmiowb
|
|
#ifdef CONFIG_IA64
|
|
#define mmiowb() asm volatile ("mf.a" ::: "memory")
|
|
#else
|
|
#define mmiowb()
|
|
#endif
|
|
#endif /* mmiowb */
|
|
|
|
#if (RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(8,1))
|
|
#define HAVE_NDO_GET_DEVLINK_PORT
|
|
#endif /* RHEL > 8.1 */
|
|
|
|
#else /* >= 5.2.0 */
|
|
#define HAVE_NDO_SELECT_QUEUE_FALLBACK_REMOVED
|
|
#define SPIN_UNLOCK_IMPLIES_MMIOWB
|
|
#define HAVE_NDO_GET_DEVLINK_PORT
|
|
#endif /* 5.2.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(5,3,0))
|
|
#if (!(RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,2)))
|
|
#define flow_block_offload tc_block_offload
|
|
#define flow_block_command tc_block_command
|
|
#define flow_cls_offload tc_cls_flower_offload
|
|
#define flow_block_binder_type tcf_block_binder_type
|
|
#define flow_cls_common_offload tc_cls_common_offload
|
|
#define flow_cls_offload_flow_rule tc_cls_flower_offload_flow_rule
|
|
#define FLOW_CLS_REPLACE TC_CLSFLOWER_REPLACE
|
|
#define FLOW_CLS_DESTROY TC_CLSFLOWER_DESTROY
|
|
#define FLOW_CLS_STATS TC_CLSFLOWER_STATS
|
|
#define FLOW_CLS_TMPLT_CREATE TC_CLSFLOWER_TMPLT_CREATE
|
|
#define FLOW_CLS_TMPLT_DESTROY TC_CLSFLOWER_TMPLT_DESTROY
|
|
#define FLOW_BLOCK_BINDER_TYPE_CLSACT_INGRESS \
|
|
TCF_BLOCK_BINDER_TYPE_CLSACT_INGRESS
|
|
#define FLOW_BLOCK_BIND TC_BLOCK_BIND
|
|
#define FLOW_BLOCK_UNBIND TC_BLOCK_UNBIND
|
|
|
|
#ifdef HAVE_TC_CB_AND_SETUP_QDISC_MQPRIO
|
|
#include <net/pkt_cls.h>
|
|
|
|
int _kc_flow_block_cb_setup_simple(struct flow_block_offload *f,
|
|
struct list_head *driver_list,
|
|
tc_setup_cb_t *cb,
|
|
void *cb_ident, void *cb_priv,
|
|
bool ingress_only);
|
|
|
|
#define flow_block_cb_setup_simple(f, driver_list, cb, cb_ident, cb_priv, \
|
|
ingress_only) \
|
|
_kc_flow_block_cb_setup_simple(f, driver_list, cb, cb_ident, cb_priv, \
|
|
ingress_only)
|
|
#endif /* HAVE_TC_CB_AND_SETUP_QDISC_MQPRIO */
|
|
#else /* RHEL >= 8.2 */
|
|
#define HAVE_FLOW_BLOCK_API
|
|
#define HAVE_DEVLINK_PORT_ATTR_PCI_VF
|
|
#endif /* RHEL >= 8.2 */
|
|
|
|
#ifndef ETH_P_LLDP
|
|
#define ETH_P_LLDP 0x88CC
|
|
#endif /* !ETH_P_LLDP */
|
|
|
|
#else /* >= 5.3.0 */
|
|
#define XSK_UMEM_RETURNS_XDP_DESC
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(5,8,0))
|
|
#if !(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(15,3,0))
|
|
#define HAVE_XSK_UMEM_HAS_ADDRS
|
|
#endif /* SLE < 15.3 */
|
|
#endif /* < 5.8.0*/
|
|
#define HAVE_FLOW_BLOCK_API
|
|
#define HAVE_DEVLINK_PORT_ATTR_PCI_VF
|
|
#if IS_ENABLED(CONFIG_DIMLIB)
|
|
#define HAVE_CONFIG_DIMLIB
|
|
#endif
|
|
#endif /* 5.3.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(5,4,0))
|
|
#if (SLE_VERSION_CODE >= SLE_VERSION(15,2,0))
|
|
#define HAVE_NDO_XSK_WAKEUP
|
|
#endif /* SLES15sp2 */
|
|
#else /* >= 5.4.0 */
|
|
#define HAVE_NDO_XSK_WAKEUP
|
|
#endif /* 5.4.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(5,5,0))
|
|
static inline unsigned long _kc_bitmap_get_value8(const unsigned long *map,
|
|
unsigned long start)
|
|
{
|
|
const size_t index = BIT_WORD(start);
|
|
const unsigned long offset = start % BITS_PER_LONG;
|
|
|
|
return (map[index] >> offset) & 0xFF;
|
|
}
|
|
#define bitmap_get_value8 _kc_bitmap_get_value8
|
|
|
|
static inline void _kc_bitmap_set_value8(unsigned long *map,
|
|
unsigned long value,
|
|
unsigned long start)
|
|
{
|
|
const size_t index = BIT_WORD(start);
|
|
const unsigned long offset = start % BITS_PER_LONG;
|
|
|
|
map[index] &= ~(0xFFUL << offset);
|
|
map[index] |= value << offset;
|
|
}
|
|
#define bitmap_set_value8 _kc_bitmap_set_value8
|
|
|
|
#endif /* 5.5.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(5,6,0))
|
|
#ifdef HAVE_AF_XDP_SUPPORT
|
|
#define xsk_umem_release_addr xsk_umem_discard_addr
|
|
#define xsk_umem_release_addr_rq xsk_umem_discard_addr_rq
|
|
#endif /* HAVE_AF_XDP_SUPPORT */
|
|
#if (RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,3)) || \
|
|
(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(15,3,0)))
|
|
#define HAVE_TX_TIMEOUT_TXQUEUE
|
|
#endif
|
|
#else /* >= 5.6.0 */
|
|
#define HAVE_TX_TIMEOUT_TXQUEUE
|
|
#endif /* 5.6.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(5,7,0))
|
|
u64 _kc_pci_get_dsn(struct pci_dev *dev);
|
|
#define pci_get_dsn(dev) _kc_pci_get_dsn(dev)
|
|
#if !(SLE_VERSION_CODE > SLE_VERSION(15,2,0)) && \
|
|
!((LINUX_VERSION_CODE == KERNEL_VERSION(5,3,18)) && \
|
|
(SLE_LOCALVERSION_CODE >= KERNEL_VERSION(14,0,0))) && \
|
|
!(RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,3)))
|
|
#define pci_aer_clear_nonfatal_status pci_cleanup_aer_uncorrect_error_status
|
|
#endif
|
|
|
|
#define cpu_latency_qos_update_request pm_qos_update_request
|
|
#define cpu_latency_qos_add_request(arg1, arg2) pm_qos_add_request(arg1, PM_QOS_CPU_DMA_LATENCY, arg2)
|
|
#define cpu_latency_qos_remove_request pm_qos_remove_request
|
|
|
|
#ifndef DEVLINK_INFO_VERSION_GENERIC_FW_BUNDLE_ID
|
|
#define DEVLINK_INFO_VERSION_GENERIC_FW_BUNDLE_ID "fw.bundle_id"
|
|
#endif
|
|
#else /* >= 5.7.0 */
|
|
#define HAVE_DEVLINK_REGION_OPS_SNAPSHOT
|
|
#define HAVE_ETHTOOL_COALESCE_PARAMS_SUPPORT
|
|
#endif /* 5.7.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(5,8,0))
|
|
#if !(RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,4))) && \
|
|
!(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(15,3,0))
|
|
#define xdp_convert_buff_to_frame convert_to_xdp_frame
|
|
#endif /* (RHEL < 8.4) || (SLE < 15.3) */
|
|
#define flex_array_size(p, member, count) \
|
|
array_size(count, sizeof(*(p)->member) + __must_be_array((p)->member))
|
|
#if (!(SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(15,3,0)))
|
|
#ifdef HAVE_AF_XDP_ZC_SUPPORT
|
|
#ifndef xsk_umem_get_rx_frame_size
|
|
static inline u32 _xsk_umem_get_rx_frame_size(struct xdp_umem *umem)
|
|
{
|
|
return umem->chunk_size_nohr - XDP_PACKET_HEADROOM;
|
|
}
|
|
|
|
#define xsk_umem_get_rx_frame_size _xsk_umem_get_rx_frame_size
|
|
#endif /* xsk_umem_get_rx_frame_size */
|
|
#endif /* HAVE_AF_XDP_ZC_SUPPORT */
|
|
#else /* SLE >= 15.3 */
|
|
#define HAVE_XDP_BUFF_FRAME_SZ
|
|
#define HAVE_MEM_TYPE_XSK_BUFF_POOL
|
|
#endif /* SLE >= 15.3 */
|
|
#else /* >= 5.8.0 */
|
|
#define HAVE_TC_FLOW_INDIR_DEV
|
|
#define HAVE_TC_FLOW_INDIR_BLOCK_CLEANUP
|
|
#define HAVE_XDP_BUFF_FRAME_SZ
|
|
#define HAVE_MEM_TYPE_XSK_BUFF_POOL
|
|
#endif /* 5.8.0 */
|
|
#if (RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,3)))
|
|
#define HAVE_TC_FLOW_INDIR_DEV
|
|
#endif
|
|
#if (SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(15,3,0)))
|
|
#define HAVE_TC_FLOW_INDIR_DEV
|
|
#endif /* SLE_VERSION_CODE && SLE_VERSION_CODE >= SLES15SP3 */
|
|
|
|
/*****************************************************************************/
|
|
#if (RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE >= RHEL_RELEASE_VERSION(8,4)))
|
|
#define HAVE_TC_FLOW_INDIR_BLOCK_CLEANUP
|
|
#endif /* (RHEL >= 8.4) */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(5,9,0))
|
|
#else /* >= 5.9.0 */
|
|
#define HAVE_FLOW_INDIR_BLOCK_QDISC
|
|
#define HAVE_UDP_TUNNEL_NIC_INFO
|
|
#endif /* 5.9.0 */
|
|
#if (RHEL_RELEASE_CODE && (RHEL_RELEASE_CODE > RHEL_RELEASE_VERSION(8,3)))
|
|
#define HAVE_FLOW_INDIR_BLOCK_QDISC
|
|
#endif
|
|
#if (SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(15,3,0)))
|
|
#define HAVE_FLOW_INDIR_BLOCK_QDISC
|
|
#endif /* SLE_VERSION_CODE && SLE_VERSION_CODE >= SLES15SP3 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(5,10,0))
|
|
#if (SLE_VERSION_CODE && SLE_VERSION_CODE >= SLE_VERSION(15,3,0))
|
|
#define HAVE_DEVLINK_REGION_OPS_SNAPSHOT_OPS
|
|
#define HAVE_DEVLINK_FLASH_UPDATE_PARAMS
|
|
#else /* SLE >= 15.3 */
|
|
struct devlink_flash_update_params {
|
|
const char *file_name;
|
|
const char *component;
|
|
u32 overwrite_mask;
|
|
};
|
|
|
|
#ifndef DEVLINK_FLASH_OVERWRITE_SETTINGS
|
|
#define DEVLINK_FLASH_OVERWRITE_SETTINGS BIT(0)
|
|
#endif
|
|
|
|
#ifndef DEVLINK_FLASH_OVERWRITE_IDENTIFIERS
|
|
#define DEVLINK_FLASH_OVERWRITE_IDENTIFIERS BIT(1)
|
|
#endif
|
|
#endif /* !(SLE >= 15.3) */
|
|
|
|
#if (!(SLE_VERSION_CODE && (SLE_VERSION_CODE >= SLE_VERSION(15,3,0))))
|
|
#define XDP_SETUP_XSK_POOL XDP_SETUP_XSK_UMEM
|
|
#define xsk_get_pool_from_qid xdp_get_umem_from_qid
|
|
#define xsk_pool_get_rx_frame_size xsk_umem_get_rx_frame_size
|
|
#define xsk_pool_set_rxq_info xsk_buff_set_rxq_info
|
|
#define xsk_pool_dma_unmap xsk_buff_dma_unmap
|
|
#define xsk_pool_dma_map xsk_buff_dma_map
|
|
#define xsk_tx_peek_desc xsk_umem_consume_tx
|
|
#define xsk_tx_release xsk_umem_consume_tx_done
|
|
#define xsk_tx_completed xsk_umem_complete_tx
|
|
#define xsk_uses_need_wakeup xsk_umem_uses_need_wakeup
|
|
#ifdef HAVE_MEM_TYPE_XSK_BUFF_POOL
|
|
#include <net/xdp_sock_drv.h>
|
|
static inline void
|
|
_kc_xsk_buff_dma_sync_for_cpu(struct xdp_buff *xdp,
|
|
void __always_unused *pool)
|
|
{
|
|
xsk_buff_dma_sync_for_cpu(xdp);
|
|
}
|
|
|
|
#define xsk_buff_dma_sync_for_cpu(xdp, pool) \
|
|
_kc_xsk_buff_dma_sync_for_cpu(xdp, pool)
|
|
#endif /* HAVE_MEM_TYPE_XSK_BUFF_POOL */
|
|
#else /* SLE >= 15.3 */
|
|
#define HAVE_NETDEV_BPF_XSK_POOL
|
|
#endif /* SLE >= 15.3 */
|
|
#else /* >= 5.10.0 */
|
|
#define HAVE_DEVLINK_REGION_OPS_SNAPSHOT_OPS
|
|
#define HAVE_DEVLINK_FLASH_UPDATE_PARAMS
|
|
#define HAVE_NETDEV_BPF_XSK_POOL
|
|
#endif /* 5.10.0 */
|
|
|
|
/*****************************************************************************/
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(5,11,0))
|
|
#ifdef HAVE_XDP_BUFF_RXQ
|
|
#include <net/xdp.h>
|
|
static inline int
|
|
_kc_xdp_rxq_info_reg(struct xdp_rxq_info *xdp_rxq, struct net_device *dev,
|
|
u32 queue_index, unsigned int __always_unused napi_id)
|
|
{
|
|
return xdp_rxq_info_reg(xdp_rxq, dev, queue_index);
|
|
}
|
|
|
|
#define xdp_rxq_info_reg(xdp_rxq, dev, queue_index, napi_id) \
|
|
_kc_xdp_rxq_info_reg(xdp_rxq, dev, queue_index, napi_id)
|
|
#endif /* HAVE_XDP_BUFF_RXQ */
|
|
#ifdef HAVE_NAPI_BUSY_LOOP
|
|
#ifdef CONFIG_NET_RX_BUSY_POLL
|
|
#include <net/busy_poll.h>
|
|
static inline void
|
|
_kc_napi_busy_loop(unsigned int napi_id,
|
|
bool (*loop_end)(void *, unsigned long), void *loop_end_arg,
|
|
bool __always_unused prefer_busy_poll,
|
|
u16 __always_unused budget)
|
|
{
|
|
napi_busy_loop(napi_id, loop_end, loop_end_arg);
|
|
}
|
|
|
|
#define napi_busy_loop(napi_id, loop_end, loop_end_arg, prefer_busy_poll, budget) \
|
|
_kc_napi_busy_loop(napi_id, loop_end, loop_end_arg, prefer_busy_poll, budget)
|
|
#endif /* CONFIG_NET_RX_BUSY_POLL */
|
|
#endif /* HAVE_NAPI_BUSY_LOOP */
|
|
#define HAVE_DEVLINK_FLASH_UPDATE_BEGIN_END_NOTIFY
|
|
#else /* >= 5.11.0 */
|
|
#define HAVE_DEVLINK_FLASH_UPDATE_PARAMS_FW
|
|
#endif /* 5.11.0 */
|
|
|
|
/*
|
|
* Load the implementations file which actually defines kcompat backports.
|
|
* Legacy backports still exist in this file, but all new backports must be
|
|
* implemented using kcompat_*defs.h and kcompat_impl.h
|
|
*/
|
|
#include "kcompat_impl.h"
|
|
|
|
#endif /* _KCOMPAT_H_ */
|