skip to main content

Flex System CN4058S and CN4052 10Gb Virtual Fabric Adapters

Product Guide (withdrawn product)

Home
Top
Author
Updated
15 Jan 2019
Form Number
TIPS1164
PDF size
19 pages, 605 KB

Abstract

The Flex System CN4058S 8-port 10Gb Virtual Fabric Adapter and CN4052 2-port 10Gb Virtual Fabric Adapter are part of the VFA5 family of System x and Flex System adapters. These adapter supports up to four virtual NIC (vNIC) devices per port (for a total of 32 for the CN4058S and 8 for the CN4052), where each physical 10 GbE port can be divided into four virtual ports with flexible bandwidth allocation. These adapters also feature RDMA over Converged Ethernet (RoCE) capability, and support iSCSI, and FCoE protocols with the addition of a Features on Demand (FoD) license upgrade.

Watch the video walk-through of the adapters with David Watts and Tom Boucher

Introduction

The Flex System CN4058S 8-port 10Gb Virtual Fabric Adapter and CN4052 2-port 10Gb Virtual Fabric Adapter are part of the VFA5 family of System x and Flex System adapters. These adapter supports up to four virtual NIC (vNIC) devices per port (for a total of 32 for the CN4058S and 8 for the CN4052), where each physical 10 GbE port can be divided into four virtual ports with flexible bandwidth allocation. These adapters also feature RDMA over Converged Ethernet (RoCE) capability, and support iSCSI, and FCoE protocols with the addition of a Features on Demand (FoD) license upgrade.

With hardware protocol offloads for TCP/IP and FCoE, the CN4058S and CN4052 provides maximum bandwidth with minimum use of CPU resources and enables more VMs per server, which provides greater cost saving to optimize return on investment. With up to eight ports, the CN4058S in particular makes full use of the capabilities of all supported Ethernet switches in the Flex System portfolio.

Figure 1 shows the CN4058S 8-port 10Gb Virtual Fabric Adapter.

0.53E.jpg
Figure 1. Flex System CN4058S 8-port 10Gb Virtual Fabric Adapter

Did you know?

This CN4058S is based on industry-standard PCIe 3.0 architecture and offers the flexibility to operate as a Virtual NIC Fabric Adapter or as a 8-port 10 Gb or 1 Gb Ethernet device. Because this adapter supports up to 32 virtual NICs on a single adapter, you see virtualization-related benefits, such as cost, power and cooling, and data center footprint by deploying less hardware. Servers, such as the new x240 M5 Compute Node, support up to two of these adapters for a total of 64 virtual NICs per system.

Part number information

Withdrawn: The adapters are withdrawn from marketing.

Table 1. Ordering part numbers and feature codes

Option Feature Code Description
Adapters (withdrawn from marketing)
94Y5160* A4R6 Flex System CN4058S 8-port 10Gb Virtual Fabric Adapter
00JY800* A5RP Flex System CN4052 2-port 10Gb Virtual Fabric Adapter
License upgrades
94Y5164 A4R9 Flex System CN4058S Virtual Fabric Adapter SW Upgrade (FoD)
00JY804 A5RV Flex System CN4052 Virtual Fabric Adapter SW Upgrade (FoD)

* Withdrawn from marketing

Features

The CN4058S 8-port 10Gb Virtual Fabric Adapter and CN4052 2-port 10Gb Virtual Fabric Adapter, which are part of the VFA5 family of System x and Flex System adapters, reduce cost by enabling a converged infrastructure and improves performance with powerful offload engines. The adapters have the following features and benefits:

  • Multi-protocol support for 10 GbE

    The adapters offers two (CN4052) or eight (CN4058S) 10 GbE connections and are cost- and performance-optimized for integrated and converged infrastructures. They offer a “triple play” of converged data, storage, and low latency RDMA networking on a common Ethernet fabric. The adapters provides customers with a flexible storage protocol option for running heterogeneous workloads on their increasingly converged infrastructures.

  • Virtual NIC emulation

    The Emulex VFA5 family supports three NIC virtualization modes, right out of the box: Virtual Fabric mode (vNIC1), switch independent mode (vNIC2), and Unified Fabric Port (UFP). With NIC virtualization, each of the two physical ports on the adapter can be logically configured to emulate up to four virtual NIC (vNIC) functions with user-definable bandwidth settings. Additionally, each physical port can simultaneously support a storage protocol (FCoE or iSCSI).

  • Full hardware storage offloads

    These adapters support an optional FCoE hardware offload engine, which is enabled via Lenovo Features on Demand (FoD). This offload engine accelerates storage protocol processing and delivers up to 1.5 million I/O operations per second (IOPS), which enables the server’s processing resources to focus on applications and improves the server’s performance.

  • Power savings

    When compared with previous generation Emulex VFA3 adapters, the Emulex VFA5 adapters can save up to 50 watts per server, reducing energy and cooling OPEX through improved storage offloads.

  • Lenovo Features on Demand (FoD)

    The adapters use FoD software activation technology. FoD enables the adapters to be initially deployed as a low-cost Ethernet NICs, and then later upgraded in the field to support FCoE or iSCSI hardware offload.

  • Emulex Virtual Network Acceleration (VNeX) technology support

    Emulex VNeX supports Microsoft supported network virtualization by using generic routing encapsulation (NVGRE) and VMware-supported virtual extensible LAN (VXLAN). These technologies create more logical LANs for traffic isolation that are needed by cloud architectures. Because these protocols increase the processing burden on the server’s processor, the adapters have an offload engine that is designed specifically for processing these tags. The resulting benefit is that cloud providers can use the benefits of VxLAN/NVGRE while not being penalized with a reduction in the server’s performance.

  • Advanced RDMA capabilities

    RDMA over Converged Ethernet (RoCE) delivers application and storage acceleration through faster I/O operations. It is best suited for environments, such as Microsoft SMB Direct in a Windows Server 2012 R2 environment. It can also support Linux NFS protocols.

    With RoCE, the VFA5 adapter helps accelerate workloads in the following ways:

    • Capability to deliver a 4x boost in small packet network performance vs. previous generation adapters, which is critical for transaction-heavy workloads
    • Desktop-like experiences for VDI with up to 1.5 million FCoE or iSCSI IOPS
    • Ability to scale Microsoft SQL Server, Exchange, and SharePoint using SMB Direct optimized with RoCE
    • More VDI desktops/server due to up to 18% higher CPU effectiveness (the percentage of server CPU utilization for every 1 Mbps I/O throughput)
    • Superior application performance for VMware hybrid cloud workloads due to up to 129% higher I/O throughput compared to adapters without offload
  • Deploy faster and manage less when Virtual Fabric adapters (VFAs) and Host Bus Adapters (HBAs) are combined that are developed by Emulex

    VFAs and HBAs that are developed by Emulex use the same installation and configuration process, which streamlines the effort to get your server running and saves you valuable time. They also use the same Fibre Channel drivers, which reduces time to qualify and manage storage connectivity. With Emulex's OneCommand Manager, you can manage VFAs and HBAs that are developed by Emulex through the data center from a single console.

Specifications

The Flex System CN4058S 8-port 10Gb Virtual Fabric Adapter features the following specifications:

  • Eight-port 10 Gb Ethernet adapter
  • Dual-ASIC controller using the Emulex XE104 design
  • Two PCIe Express 3.0 x8 host interfaces (8 GT/s)
  • MSI-X support
  • Fabric Manager support
  • Power consumption: 25 W maximum

The CN4052 2-port 10Gb Virtual Fabric Adapter has these specifications:

  • Two-port 10 Gb Ethernet adapter
  • Single-ASIC controller using the Emulex XE104 design
  • One PCIe Express 3.0 x8 host interface (8 GT/s)
  • MSI-X support
  • Fabric Manager support
  • Power consumption: 25 W maximum

Ethernet features

The following Ethernet features are included in both adapters:

  • IPv4/IPv6 TCP and UDP checksum offload, Large Send Offload (LSO), Large Receive Offload, Receive Side Scaling (RSS), TCP Segmentation Offload (TSO)
  • VLAN insertion and extraction
  • Jumbo frames up to 9216 Bytes
  • Priority Flow Control (PFC) for Ethernet traffic
  • Interrupt coalescing
  • Load balancing and failover support, including adapter fault tolerance (AFT), switch fault tolerance (SFT), adaptive load balancing (ALB), link aggregation, and IEEE 802.1AX
  • 4 PCIe Physical Functions (PFs) per port
  • 31 PCIe Virtual Functions (VFs) for NIC per port configurable
  • Advance Error Reporting (AER)
  • Supports D0, D3 (hot & cold) power management modes
  • Completion Timeout (CTO)
  • Function Level Reset (FLR)
  • Network boot support:
    • PXE Boot
    • iSCSI Boot (with FoD upgrade)
    • FCoE Boot (with FoD upgrade)
  • Virtual Fabric support:
    • Virtual Fabric mode (vNIC1)
    • Switch Independent mode (vNIC2)
    • Unified Fabric Port (UFP) support
  • Emulex VNeX Technology full hardware offload support:
    • Increases VM Density and decreases CPU power draw
    • Microsoft Network Virtualization using Generic Routing Encapsulation (NVGRE)
    • VMware Virtual Extensible LAN (VXLAN)
    • VM mobility acceleration, which offloads header encapsulation from processor to network controller
  • Remote Direct Memory Access over Converged Ethernet (RoCE) support:
    • Accelerates application and storage I/O operations
    • Supports Windows Server SMB Direct
    • Is required to deliver low latency performance
  • Single Root I/O Virtualization (SRIOV) support:
    • Maximum of 128 Virtual Functions
    • Minimizes VM I/O overhead and improves adapter performance

FCoE features (with FoD upgrade)

The following FCoE features (with FoD upgrade) are included:

  • Hardware offloads of Ethernet TCP/IP
  • ANSI T11 FC-BB-5 Support
  • Programmable World Wide Name (WWN)
  • Support for FIP and FCoE Ether Types
  • Support for up to 255 NPIV Interfaces per port
  • FCoE Initiator
  • Common driver for Emulex Universal CNA and Fibre Channel HBAs
  • Fabric Provided MAC Addressing (FPMA) support
  • 1024 concurrent port logins (RPIs) per port
  • 1024 active exchanges (XRIs) per port

iSCSI features (with FoD upgrade)

The following iSCSI features (with FoD upgrade) are included:

  • Full iSCSI Protocol Offload
  • Header, Data Digest (CRC), and PDU
  • Direct data placement of SCSI data
  • 2048 Offloaded iSCSI connections
  • iSCSI initiator and concurrent initiator or target modes
  • Multipath I/O
  • OS-independent INT13 based iSCSIboot and iSCSI crash dump support
  • RFC 4171 Internet Storage Name Service (iSNS)

Standards

The adapter supports the following IEEE standards:

  • PCI Express base spec 2.0, PCI Bus Power Management Interface, rev. 1.2, Advanced Error Reporting (AER)
  • 802.3-2008 10Gbase Ethernet port
  • 802.1Q vLAN
  • 802.3x Flow Control with pause Frames
  • 802.1 Qbg Edge Virtual Bridging
  • 802.1Qaz Enhanced transmission Selection (ETS) Data Center Bridging Capability (DCBX)
  • 802.1Qbb Priority Flow Control
  • 802.3ad link Aggregation/LACP
  • 802.1ab Link Layer Discovery Protocol
  • 802.3ae (SR Optics)
  • 802.1AX (Link Aggregation)
  • 802.3p (Priority of Service)
  • 802.1Qau (Congestion Notification)
  • IPV4 (RFQ 791)
  • IPV6 (RFC 2460)

Modes of operation

The CN4058S and CN4052 support the following vNIC modes of operation and pNIC mode:

  • Virtual Fabric Mode (also known as vNIC1 mode). This mode works only with any of the following switches that are installed in the chassis:
    • Flex System Fabric CN4093 10Gb Converged Scalable Switch
    • Flex System Fabric EN4093R 10Gb Scalable Switch
    • Flex System Fabric EN4093 10Gb Scalable Switch

    In this mode, the adapter communicates using DCBX with the switch module to obtain vNIC parameters. Also, a special tag within each data packet is added and later removed by the NIC and switch for each vNIC group to maintain separation of the virtual channels.

    In vNIC mode, each physical port is divided into four virtual ports, which provides a total of 32 (CN4058S) or 8 (CN4052) virtual NICs per adapter. The default bandwidth for each vNIC is 2.5 Gbps. Bandwidth for each vNIC can be configured at the switch 100 Mbps - 10 Gbps, up to a total of 10 Gbps per physical port. The vNICs can also be configured to have 0 bandwidth if you must allocate the available bandwidth to fewer than eight vNICs. In Virtual Fabric Mode, you can change the bandwidth allocations through the switch user interfaces without requiring a reboot of the server.

    When storage protocols are enabled on the adapter (enabled with the appropriate FoD upgrade as listed in Table 1), 28 ports are Ethernet, and eight ports are iSCSI or FCoE.

  • Switch Independent Mode (also known as vNIC2 mode), where the adapter works with the following switches:
    • Cisco Nexus B22 Fabric Extender for Flex System
    • Flex System EN4023 10Gb Scalable Switch
    • Flex System Fabric CN4093 10Gb Converged Scalable Switch
    • Flex System Fabric EN4093R 10Gb Scalable Switch
    • Flex System Fabric EN4093 10Gb Scalable Switch
    • Flex System Fabric SI4091 System Interconnect Module
    • Flex System Fabric SI4093 System Interconnect Module
    • Flex System EN4091 10Gb Ethernet Pass-thru and a top-of-rack (TOR) switch

    Switch Independent Mode offers the same capabilities as Virtual Fabric Mode in terms of the number of vNICs and the bandwidth that each can be configured to include. However, Switch Independent Mode extends the existing customer VLANs to the virtual NIC interfaces. The IEEE 802.1Q VLAN tag is essential to the separation of the vNIC groups by the NIC adapter or driver and the switch. The VLAN tags are added to the packet by the applications or drivers at each end station rather than by the switch.
  • Unified Fabric Port (UFP) provides a feature-rich solution compared to the original vNIC Virtual Fabric mode. As with Virtual Fabric mode vNIC, UFP allows carving up a single 10 Gb port into four virtual NICs (called vPorts in UFP). UFP also has the following modes that are associated with it:
    • Tunnel mode: Provides Q-in-Q mode, where the vPort is customer VLAN-independent (very similar to vNIC Virtual Fabric Dedicated Uplink Mode).
    • Trunk mode: Provides a traditional 802.1Q trunk mode (multi-VLAN trunk link) to the virtual NIC (vPort) interface; that is. permits host side tagging.
    • Access mode: Provides a traditional access mode (single untagged VLAN) to the virtual NIC (vPort) interface, which is similar to a physical port in access mode.
    • FCoE mode: Provides FCoE functionality to the vPort.
    • Auto-VLAN mode: Auto VLAN creation for Qbg and VMready environments.

    Only one vPort (vPort 2) per physical port can be bound to FCoE. If FCoE is not wanted, vPort 2 can be configured for one of the other modes. UFP works with the following switches:
    • Flex System Fabric CN4093 10Gb Converged Scalable Switch
    • Flex System Fabric EN4093R 10Gb Scalable Switch
    • Flex System Fabric SI4093 System Interconnect Module
  • In pNIC mode, the adapter can operate as a standard 10 Gbps or 1 Gbps 4-port Ethernet expansion card. When in pNIC mode, the adapter functions with all supported I/O modules.

    In pNIC mode, the adapter with the FoD upgrade applied operates in traditional Converged Network Adapter (CNA) mode with eight ports (CN4058S) or two ports (CN4052) of Ethernet and eight ports (CN4058S) or two ports (CN4052) of iSCSI or FCoE available to the operating system.

Embedded 10Gb Virtual Fabric Adapter

Some models of the X6 compute nodes include one Embedded 4-port 10Gb Virtual Fabric controller (VFA), which is also known as LAN on Motherboard or LOM. This controller is built into the system board. The controller is Emulex XE104-based like the CN4058S 8-port 10Gb Virtual Fabric Adapter and has the same features.

Using the Fabric Connector, two 10Gb ports routed to I/O module bay 1 and two 10 Gb ports are routed to I/O module bay 2. A compatible I/O module must be installed in those I/O module bays.

The Embedded 10Gb VFA also supports iSCSI and FCoE using the Lenovo Features on Demand (FoD) license upgrade listed in the following table. Only one upgrade is needed per compute node.

Withdrawn: This FoD license upgrade is now withdrawn from marketing.

Table 2. Feature on Demand upgrade for FCoE and iSCSI support

Part number Feature code Description
00FM258* A5TM Flex System X6 Embedded 10Gb Virtual Fabric Upgrade

* Withdrawn from marketing

Supported servers

The following table lists the Flex System compute nodes that support the adapters.

Table 3. Supported servers

Description Part number x220 (7906) x222 (7916) x240 (8737, E5-2600) x240 (8737, E5-2600 v2) x240 (7162) x240 M5 (9532) x440 (7917) 3.1DB6.gif x880/x480/x280 X6 (7903) x280 / x480 / x880 X6 (7196) PCIe Expansion Node
CN4058S 8-port 10Gb Virtual Fabric Adapter 94Y5160 N N N N Y Y N N N Y* Y
CN4058S Virtual Fabric Adapter Software Upgrade (FoD) 94Y5164 N N N N Y Y N N N Y Y
Flex System CN4052 2-port 10Gb Virtual Fabric Adapter 00JY800 N N N N Y Y N N N Y Y
Flex System CN4052 Virtual Fabric Adapter SW Upgrade (FoD) 00JY804 N N N N Y Y N N N N Y

* The CN4058S 8-port 10Gb Virtual Fabric Adapter has two ASICs and is not supported in slots 3 and 4

For more information about the expansion cards that are supported by each blade server type, visit
http://www.lenovo.com/us/en/serverproven/

I/O adapter cards are installed in the slot in supported servers, such as the x240 M5, as highlighted in the following figure.

4.1E.jpg
Figure 2. Location of the I/O adapter slots in the Flex System x240 M5 Compute Node

Supported I/O modules

These adapters can be installed in any I/O adapter slot of a supported Flex System compute node. One or two compatible 1 Gb or 10 Gb I/O modules must be installed in the corresponding I/O bays in the chassis. The following table lists the switches that are supported. When the adapter is connected to the 1 Gb switch, the adapter operates at 1 Gb speeds.

To maximize the number of adapter ports that are usable, you might also need to order switch upgrades to enable more ports as listed in the table. Alternatively, for CN4093, EN4093R, and SI4093 switches, you can use Flexible Port Mapping, which is a new feature of Networking OS 7.8 with which you can minimize the number of upgrades that are needed. For more information, see the Product Guides for the switches, which are available at this website:
http://lenovopress.com/flexsystem/iomodules

Note: Adapter ports 7 and 8 are reserved for future use. The chassis supports all eight ports, but there are no switches that are available that connect to these ports.

Table 4. I/O modules and upgrades for use with the CN4058S and CN4052 adapter

Description Part number Port count
(per pair
of switches)*
Lenovo Flex System Fabric EN4093R 10Gb Scalable Switch
+ EN4093 10Gb Scalable Switch (Upgrade 1)
00FM514
49Y4798
4
Lenovo Flex System Fabric CN4093 10Gb Converged Scalable Switch
+ CN4093 10Gb Converged Scalable Switch (Upgrade 1)
00FM510
49Y4798
4
Lenovo Flex System SI4091 10Gb System Interconnect Module 00FE327 2
Lenovo Flex System Fabric SI4093 System Interconnect Module
+ SI4093 System Interconnect Module (Upgrade 1)
00FM518
95Y3318
4
CN4093 10Gb Converged Scalable Switch:
+ CN4093 10Gb Converged Scalable Switch (Upgrade 1)
+ CN4093 10Gb Converged Scalable Switch (Upgrade 2)
00D5823
00D5845
00D5847
6
EN4093R 10Gb Scalable Switch:
+ EN4093 10Gb Scalable Switch (Upgrade 1)
+ EN4093 10Gb Scalable Switch (Upgrade 2)
95Y3309
49Y4798
88Y6037
6
EN4093 10Gb Scalable Switch:
+ EN4093 10Gb Scalable Switch (Upgrade 1)
+ EN4093 10Gb Scalable Switch (Upgrade 2)
49Y4270**
49Y4798
88Y6037
6
Flex System EN4091 10Gb Ethernet Pass-thru 88Y6043 2
SI4093 System Interconnect Module:
+ SI4093 System Interconnect Module (Upgrade 1)
+ SI4093 System Interconnect Module (Upgrade 2)
95Y3313
95Y3318
95Y3320
6
EN2092 1Gb Ethernet Scalable Switch
+ EN2092 1Gb Ethernet Scalable Switch (Upgrade 1)
49Y4294
90Y3562
4
Cisco Nexus B22 Fabric Extender for Flex System 94Y5350 2
EN4023 10Gb Scalable Switch
(Base switch has 24 port licenses; Upgrades 1 & 2 may be needed)
94Y5212 6

* This column indicates the number of adapter ports that are active if indicated upgrades are installed.
** Withdrawn from marketing.

The following table shows the connections between adapters that are installed in the compute nodes to the switch bays in the chassis.

Note: The CN4058S is not supported in slots 3 and 4 of the x880 X6, x480 X6 and x280 X6.

Table 5. Adapter to I/O bay correspondence

I/O adapter slot
in the server
Port on the adapter*
Corresponding I/O module bay
in the chassis
Bay 1 Bay 2 Bay 3 Bay 4
Slot 1 Port 1 Yes No No No
Port 2 No Yes No No
Port 3 Yes No No No
Port 4 No Yes No No
Port 5 Yes No No No
Port 6 No Yes No No
Port 7** Yes No No No
Port 8** No Yes No No
Slot 2 Port 1 No No Yes No
Port 2 No No No Yes
Port 3 No No Yes No
Port 4 No No No Yes
Port 5 No No Yes No
Port 6 No No No Yes
Port 7** No No Yes** No
Port 8** No No No Yes**

* The use of adapter ports 3, 4, 5, and 6 require upgrades to the switches as described in Table 4. The EN4091 Pass-thru only supports ports 1 and 2 (and only when two Pass-thru modules are installed).
** Adapter ports 7 and 8 are reserved for future use. The chassis supports all eight ports, but there are no switches that are available that connect to these ports.

The following figure shows the internal layout of the CN4058S showing how the adapter ports are routed to the I/O module ports. Red lines indicate connections from ASIC 1 on the adapter and blue lines are the connections from ASIC 2. The dotted blue lines are reserved for future use when switches are available that support all 8 ports of the adapter.

8.1E5C.jpg
Figure 3. Internal layout of the CN4058S adapter ports

The following figure shows the internal layout of the CN4052 showing how the adapter ports are routed to the I/O module ports.

16.4DF8.jpg
Figure 4. Internal layout of the CN4052 adapter ports

The connections between the CN4058S 8-port adapters that are installed in the compute nodes and the switch bays in the chassis are shown in the following figure. See Table 3 for specific server compatibility. See Table 4 for specific switch compatibility.

Logical layout of the interconnects between I/O adapters and I/O modules
Figure 5. Logical layout of the interconnects between I/O adapters and I/O modules

Operating system support

The CN4058S and CN4052 support the following operating systems:

  • Microsoft Windows Server 2012
  • Microsoft Windows Server 2012 R2
  • Microsoft Windows Server 2016
  • Microsoft Windows Server 2019
  • Microsoft Windows Server version 1709
  • Red Hat Enterprise Linux 6 Server x64 Edition
  • SUSE Linux Enterprise Server 11 for AMD64/EM64T
  • SUSE Linux Enterprise Server 11 with Xen for AMD64/EM64T
  • SUSE Linux Enterprise Server 12
  • SUSE Linux Enterprise Server 12 with Xen
  • SUSE Linux Enterprise Server 15
  • SUSE Linux Enterprise Server 15 with Xen
  • VMware vSphere 5.1 (ESXi)
  • VMware vSphere Hypervisor (ESXi) 5.5
  • VMware vSphere Hypervisor (ESXi) 6.5
  • VMware vSphere Hypervisor (ESXi) 6.7

Operating system support may vary by server. For more information, see ServerProven at http://www.lenovo.com/us/en/serverproven/. Select the server in question, scroll down to the adapter, and click the + icon in that row to show the supported operating systems.

Warranty

There is a 1-year, customer-replaceable unit (CRU) limited warranty. When installed in a server, these adapters assume your system’s base warranty and any Lenovo warranty service upgrade purchased for the system.

Physical specifications

The adapter features the following dimensions and weight:

  • Width: 100 mm (3.9 in.)
  • Depth: 80 mm (3.1 in.)
  • Weight: 13 g (0.3 lb)

The adapter features the following shipping dimensions and weight (approximate):

  • Height: 58 mm (2.3 in.)
  • Width: 229 mm (9.0 in.)
  • Depth: 208 mm (8.2 in.)
  • Weight: 0.4 kg (0.89 lb)

Regulatory compliance

The adapter conforms to the following regulatory standards:

  • United States FCC 47 CFR Part 15, Subpart B, ANSI C63.4 (2003), Class A
  • United States UL 60950-1, Second Edition
  • IEC/EN 60950-1, Second Edition
  • FCC - Verified to comply with Part 15 of the FCC Rules, Class A
  • Canada ICES-003, issue 4, Class A
  • UL/IEC 60950-1
  • CSA C22.2 No. 60950-1-03
  • Japan VCCI, Class A
  • Australia/New Zealand AS/NZS CISPR 22:2006, Class A
  • IEC 60950-1(CB Certificate and CB Test Report)
  • Taiwan BSMI CNS13438, Class A
  • Korea KN22, Class A; KN24
  • Russia/GOST ME01, IEC-60950-1, GOST R 51318.22-99, GOST R 51318.24-99, GOST R 51317.3.2-2006, GOST R 51317.3.3-99
  • IEC 60950-1 (CB Certificate and CB Test Report)
  • CE Mark (EN55022 Class A, EN60950-1, EN55024, EN61000-3-2, EN61000-3-3)
  • CISPR 22, Class A

Related product families

Product families related to this document are the following:

Trademarks

Lenovo and the Lenovo logo are trademarks or registered trademarks of Lenovo in the United States, other countries, or both. A current list of Lenovo trademarks is available on the Web at https://www.lenovo.com/us/en/legal/copytrade/.

The following terms are trademarks of Lenovo in the United States, other countries, or both:
Lenovo®
ServerProven®
System x®

The following terms are trademarks of other companies:

Linux® is the trademark of Linus Torvalds in the U.S. and other countries.

Microsoft®, SQL Server®, SharePoint®, Windows Server®, and Windows® are trademarks of Microsoft Corporation in the United States, other countries, or both.

Other company, product, or service names may be trademarks or service marks of others.