Quantcast
Channel: Mellanox Interconnect Community: Message List
Viewing all 6227 articles
Browse latest View live

part_man.exe not working on Windows 10 build 10240 with WinOF 5.10 or 5.22 using ConnectX-3 dual-port and IS5022 Infiniscale IB Switches

$
0
0

I am pretty new to Infiniband and I am having one heck of a time trying to get the right secret combination to allow for me to successfully run part_man.exe to add virtual interfaces based on partitions.conf with opensm.  Am I missing something? 

 

DEVICE INFO:

 

C:\Program Files\Mellanox\MLNX_VPI\IB\Tools>ibv_devinfo.exe

hca_id: ibv_device0

        fw_ver:                         2.36.5000

        node_guid:                      f452:1403:007e:fae0

        sys_image_guid:                 f452:1403:007e:fae3

        vendor_id:                      0x02c9

        vendor_part_id:                 4099

        hw_ver:                         0x0

        phys_port_cnt:                  2

                port:   1

                        state:                  PORT_ACTIVE (4)

                        max_mtu:                4096 (5)

                        active_mtu:             4096 (5)

                        sm_lid:                 8

                        port_lid:               8

                        port_lmc:               0x00

                        transport:              IB

 

 

                port:   2

                        state:                  PORT_ACTIVE (4)

                        max_mtu:                4096 (5)

                        active_mtu:             4096 (5)

                        sm_lid:                 8

                        port_lid:               9

                        port_lmc:               0x00

                        transport:              IB

 

Below is the error that I receive when trying to configure according to several other posts and documentation on using part_man.exe:

 

C:\Program Files\Mellanox\MLNX_VPI\IB\Tools>part_man -v add "Ethernet" ipoib_vlan0111 806f

==> GetAdapterKey: searchByKey 'Name', KeyValue 'Ethernet', DataKey 'PnpInstanceID'

<== GetAdapterKey: result - adapterData 'IBA\ConnectX-3_IPoIB\e2fa7e00031452f4-ffff-IPoIB', status 0

 

 

ERROR:  Illegal adapter name 'Ethernet'.  Please select valid physical Mellanox IPoIB adapter.

 

C:\Program Files\Mellanox\MLNX_VPI\IB\Tools>part_man -v add "Ethernet 1" ipoib_vlan0111 806f

==> GetAdapterKey: searchByKey 'Name', KeyValue 'Ethernet 1', DataKey 'PnpInstanceID'

<== GetAdapterKey: result - adapterData 'IBA\ConnectX-3_IPoIB\e1fa7e00031452f4-ffff-IPoIB', status 0

 

 

ERROR:  Illegal adapter name 'Ethernet 1'.  Please select valid physical Mellanox IPoIB adapter.

 

Below is the output when executing 'mlxtool dbg pkeys':

 

IPoIB NIC: Ethernet

              ---------------- ----------------

             |   PKEY index   |      PKEY      |

              ---------------- ----------------

             |           0    |        ffff    |

             |           1    |        806f    |

              ---------------- ----------------

IPoIB NIC: Ethernet 1

              ---------------- ----------------

             |   PKEY index   |      PKEY      |

              ---------------- ----------------

             |           0    |        ffff    |

             |           1    |        806f    |

              ---------------- ----------------

 

Below is the output from opensm partitions.conf file:

 

Default=0xffff, rate=7, mtu=5, scope=2, defmember=full:

          ALL, ALL_SWITCHES=full;

Default=0xffff, ipoib, rate=7, mtu=5, scope=2:

          ALL=full, ALL_SWITCHES=full;

 

VLAN0111=0x006f, rate=7, mtu=5, scope=2, defmember=full:

          ALL, ALL_SWITCHES=full;

VLAN0111=0x006f, ipoib, rate=7, mtu=5, scope=2:

          ALL=full, ALL_SWITCHES=full;


Connecting Intel FVL 40 GbE QSFP+ nic to Mellanox 40 GbE nic QSFP+

$
0
0

Hi,

Will connecting Intel FVL 40 GbE QSFP+  NIC to Mellanox QSFP+ 40 GbE nic with Intel cable work ? will it work with Mellanox cables (and if so, with which Mellanox cables)?

More details:

I am talking about QSFP+  nics like Intel QSFP+ Fortville and Mellanox QSFP+ MCX314A-BCCT, and about cables like https://www.amazon.com/Intel-Ethernet-Twinaxial-Cable-meters/dp/B00KB3MA40 ,

See also for exaemple such  FVL NIC here:

Intel XL710QDA2 Ethernet Converged Network Adapters XL710 10/40 GbE PCIe 3.0, x8 Dual port - Newegg.com

  See also the mellanox NIC here for example:

http://www.mellanox.com/related-docs/user_manuals/ConnectX-3_Pro_Ethernet_Single_and_Dual_QSFP _Port_Adapter_Card_User_Manual.pdf

Thanks,

Kevin

Connectx-4 MCX4131A-BCAT : is it a dual port or a single port ?

$
0
0

Hello,

According  to this doc:

http://www.mellanox.com/related-docs/prod_adapter_cards/PB_ConnectX-4_Lx-EN_Card.pdf

the MCX4131A-BCAT is a single port.

However, in all websites which sale this card (Amazon, eBay, newegg and others) there is a picture of a dual port NIC.

Does anyone know about this nic ? could it be that there is an error in the product brief  and it is a dual port nic indeed?

 

Second question: in

http://www.mellanox.com/page/products_dyn?product_family=219&mtag=connectx_4_lx_en_car

I see (on the left side): "ConnectX-4 EN" and "ConnectX-4 Lx EN".

what does the "Lx" stands for ?

 

Regards,

Kevin

Re: SR-IOV on ESXi

$
0
0

Hi!

[Q] Is there by any chance a script to clean up an ESXi host of all previous Mellanox and OFED drivers completely… ? Indeed, each installation requires a manual removal of each module, which is very time consuming. This is especially true when updating the ESXi hosts and the drivers having to be uninstalled before the update and reinstalled after the update due to the strict acceptance level of the ESXi update process.

 

[A] Did you try using VMware Power CLI that based on Windows PowerShell?

I made a custom image for my R610 ESXi hosts.

This image removed original Mellanox inbox drivers then I can install vSphere OFED after ESXi host installed...:)

Re: IPoIB Interop VMware ESXi 6 to Windows 10/2012 R2 nodes with IS5022 InfiniScale "unmanaged" switch partitions.conf pkeys configured nodes not able to reach each other

$
0
0

Thanks for your reply.  However, all hosts both Windows and VMware ESXi are using IPoIB connectivity to the IS5022 switch.  It turns out that the issue wasn't at all with the infrastructure components once they were configured correctly. I reformatted the Windows 10 host as Windows 2012 R2 leveraging the 5.22 OFED driver package and everything worked like it should.  I was able to ping the ESXi 6 hosts and the VM guests that were leveraging port-groups configured with the IPoIB uplinks.  This is probably due to issues seen with other networking products like Intel PROSet and the creation of VLANs and Teams with Windows 10.  Microsoft has drastically changed the driver architecture for the kernel drivers used for networking.  Intel is been working with Microsoft since November of last year trying to resolve issues seen with the build 1511 for Windows 10.  Looks like Mellanox should join the list of vendors who also have issues with Microsoft's flagship OS.

Re: SR-IOV on ESXi

$
0
0

I don't know why but I had not seen mlnx_ofed_esxi v2.40 until now.

Was unsuccessfully trying with v2.3.3.1 on ESXi 6.0.

 

So, just to confirm if I understood correctly:

- According to the restrictions you mentioned, it is not possible to assign an Infiniband VF function to a Linux guest in ESXi 6 (hence no IPoIB) ?

- SR-IOV is not available for Connect-IB in ESXi ?

- There is no storage protocol (iSER / SRP) available in OFED 2.4 ?

 

Many thanks for the detailed information.

Re: SR-IOV on ESXi

$
0
0

Hi Jae-Hoon,

 

I use ESXi-Customizer-PS to inject the drivers but it does not help when the ESXi is already installed on the host.

 

We have variety of hosts and adapter types (Connect-X2, ConnectX-3, Connect-IB and probably soon ConnectX-4 or 5), and the manual cleaning up of drivers and VIBs are painful and time consuming because the list of items to remove is different from host to host, which makes a simple list impossible to apply to "esxcli software vib remove..."

 

Any ideas are welcome.

 

Thanks for your contribution.

Can anybody point me to location where i can get network direct ddk package for windows 2012 R2

$
0
0

I want to verify ndping.exe, ndpingpong.exe etc.Can anybody please point me to location where i can get network direct ddk package


Re: IPoIB Interop VMware ESXi 6 to Windows 10/2012 R2 nodes with IS5022 InfiniScale "unmanaged" switch partitions.conf pkeys configured nodes not able to reach each other

$
0
0

If you were using an inbox InfiniBand driver for Win 10 (or any O.S.), this is a very limited driver and the Mellanox OFED driver should be installed as best practice.

The WinOF 5.22 technically supports Windows 10. Support for Windows 10 client (64 bit only) was added in WinOF 5.10.

Glad to hear you got it working on Windows 2012R2.

Re: Connecting Intel FVL 40 GbE QSFP+ nic to Mellanox 40 GbE nic QSFP+

$
0
0

Hi Kevin,

Thank you for your patience.

 

Regarding your question , Mellanox only guarantees a valid connection when using Mellanox products from end to end .
Therefore using other products than Mellanox products in the environment cannot be guaranteed be functional .

 

Please note that we have never tested tested this setup.
The tested and supported devices are listed in release notes of the firmware of the card.

For your convenience we please refer to the URL below (Section 1.2.3 Validated and Supported 4GbE Cables)

http://www.mellanox.com/pdf/firmware/ConnectX3Pro-FW-2_36_5000-release_notes.pdf

 

Thanks and regards,

~Martijn

Change the speed of Infiniband adaptor card

$
0
0

Hi, We have Mellanox SX6536 648-Port InfiniBand Director Switch (56 Gbps) and Mellanox Technologies MT27500 Family [ConnectX-3] adapter cards connected to the Intel servers running CentOS 6.2 (64 bit).

We have OFED version : 3.3.4-1

We want to degrade the spped of IB cards from 56 Gbps to 40 Gbps. How to do it?

Re: Compiling with mellanox headers

Re: mckey program does not work

$
0
0

'-m'  option should be used with multicast address. In addtion, you can bind it to specific device using '-b' parameter. Example

Server:

$mckey -m 234.5.6.7 -b 10.11.22.1

 

Client:

$mckey -m 234.5.6.7 -b 10.11.22.2 -s

Re: OFED 3.3, RHEL 6.8, MVAPICH2 not present

$
0
0

Hello Sergey,

 

MVAPICH2 is not included MOFED 3.3, so you need to manually compile. Sorry for the inconvenience.

I will mention this internally.

 

Thanks,

~Rage

Re: Change the speed of Infiniband adaptor card

$
0
0

Hi

You can just change the speed on the switch port to 40 Gbps and the adapter card should auto-negotiate to that speed.

Example commands:

1. enable

2 configure terminal

3. interface ib 1/1/1 speed fdr10 force


what does IBV_WC_SUCCESS mean?

$
0
0

I am a newbie in RDMA.

I try to write data from local memory to a remote machine with IBV_WR_RDMA_WRITE (I use unreliable connection), and then I try to poll a completion event and its status is IBV_WC_SUCCESS, I am curious does this status mean the data is already successfully written to the remote machine?

 

Many thanks for you time and patience.

Re: Trouble with ConnectX-3 VPI VFs with SR-IOV

$
0
0

Hello Lasse,

 

I'm not familiar with this error. As I was looking into your error internally, i noticed you have opened a support case with us for the same issue already.

We will continue to assist you on that case.

 

Thanks.

.R

Re: SR-IOV on ESXi

$
0
0

Hi!
I pray these scrips to help you.

01. Uninstall Mellanox vSphere OFED 1.8.2.4 & 1.8.3 from ESXi host

 

1st you must check your installed driver list with esxcli software vib list | grep Mellanox

 

esxcli software vib remove -n scsi-ib-srp

esxcli software vib remove -n net-ib-ipoib

esxcli software vib remove -n net-mlx4-ib

esxcli software vib remove -n scsi-ib-iser

esxcli software vib remove -n net-rdma-cm

esxcli software vib remove -n net-ib-addr

esxcli software vib remove -n net-ib-cm

esxcli software vib remove -n net-ib-umad

esxcli software vib remove -n net-ib-sa

esxcli software vib remove -n net-ib-mad

esxcli software vib remove -n net-ib-core

esxcli software vib remove -n net-mlx4-en

esxcli software vib remove -n net-mlx4-core

esxcli software vib remove -n net-mlx-compat

esxcli software vib remove -n net-memtrack

 

then reboot your ESXi host

 

 

 

02. Uninstall Mellanox vSphere OFED  2.3.3.1 & 2.4.0 from ESXi host

 

1st you must check your installed driver list with esxcli software vib list | grep MEL

 

esxcli software vib remove -n net-ib-ipoib

esxcli software vib remove -n net-mlx4-ib

esxcli software vib remove -n net-ib-sa

esxcli software vib remove -n net-ib-mad

esxcli software vib remove -n net-ib-core

esxcli software vib remove -n net-mlx4-en

esxcli software vib remove -n net-mlx4-core

esxcli software vib remove -n net-mlx-compat

 

reboot your ESXi host

 

You're correct.

This is a time consuming work, but there is no choice...

 

AND If you want to upgrade ESXi host to new version, you must remove inbox driver from original ESXi image.

 

If you want build a custom image for your ESXi host, you can build with ESXi offline bundle and vSphere PowerCLI.

 

I have a 10 of Dell PE R610 ESXi host.

 

I'm always remove Mellanox inbox driver from Dell Customized offline bundle then create custom ESXi ISO image for my hosts.

 

This custom ESXi ISO image - removed inbox driver - can upgrade ESXi host to new version directly.

 

I'm also build a custom ESXi ISO image that include Mellanox vSphere OFED driver 1.8.2.4.

- also remove inbox driver then inject driver 1.8.2.4

 

All was successfully work for my ESXi hosts.

 

If you have standard your host configuration and driver version, you can build a custom images for your host.

 

I think build a custom image will help you.

Re: permanent setting log_num_mtt and log_mtts_per_seg

$
0
0

Hello,

 

If the file /etc/modprobe.d/mlx4_core.conf does not exist, create file then input the desired parameter followed by restart of the driver(# /etc/init.d/openibd restart)

 

Cheers,

~R

Odd, unsymmetric ib_send_lat results?

$
0
0

I have two small IB clusters set up for testing:

  • Both have SB7700 IB switch
  • Two servers, each with a MCX455A-ECAT ConnectX-4 VPI adapter, are connected to each switch.

 

Essential system and software info:

[root@fs00 ~]# uname -a

Linux fs00 3.10.0-327.22.2.el7.x86_64 #1 SMP Thu Jun 23 17:05:11 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux

[root@fs00 ~]# rpm -qa |grep ofed

ofed-scripts-3.3-OFED.3.3.1.0.0.x86_64

mlnx-ofed-all-3.3-1.0.0.0.noarch

 

I have been testing the two clusters using ib_send_lat and observed the following that I don't understand:

Cluster A:

 

Server:

I got what the latency numbers that I anticipated.  Reverse the role of client and server, results more or less the same. Again, that's what I anticipated.

 

[root@fs01 ~]# ib_send_lat -a -c UD

 

 

************************************

* Waiting for client to connect... *

************************************

Max msg size in UD is MTU 4096

Changing to this MTU

---------------------------------------------------------------------------------------

                    Send Latency Test

Dual-port       : OFF Device         : mlx5_0

Number of qps   : 1 Transport type : IB

Connection type : UD Using SRQ      : OFF

RX depth        : 1000

Mtu             : 4096[B]

Link type       : IB

Max inline data : 188[B]

rdma_cm QPs : OFF

Data ex. method : Ethernet

---------------------------------------------------------------------------------------

local address: LID 0x02 QPN 0x0028 PSN 0xd0b6fe

remote address: LID 0x03 QPN 0x0028 PSN 0x91642e

---------------------------------------------------------------------------------------

#bytes #iterations    t_min[usec]    t_max[usec]  t_typical[usec]

2       1000          0.81           4.47         0.83  

4       1000          0.82           3.88         0.83  

8       1000          0.81           2.95         0.83  

16      1000          0.82           3.31         0.84  

32      1000          0.88           3.40         0.90  

64      1000          0.88           3.27         0.90  

128     1000          0.91           3.54         0.93  

256     1000          1.23           3.55         1.25  

512     1000          1.29           4.17         1.32  

1024    1000          1.49           3.15         1.51  

2048    1000          1.72           4.32         1.74  

4096    1000          2.15           4.32         2.20  

---------------------------------------------------------------------------------------

Client:

[root@fs00 ~]#  ib_send_lat -a -c UD  192.168.11.151

Max msg size in UD is MTU 4096

Changing to this MTU

---------------------------------------------------------------------------------------

                    Send Latency Test

Dual-port       : OFF Device         : mlx5_0

Number of qps   : 1 Transport type : IB

Connection type : UD Using SRQ      : OFF

TX depth        : 1

Mtu             : 4096[B]

Link type       : IB

Max inline data : 188[B]

rdma_cm QPs : OFF

Data ex. method : Ethernet

---------------------------------------------------------------------------------------

local address: LID 0x03 QPN 0x0028 PSN 0x91642e

remote address: LID 0x02 QPN 0x0028 PSN 0xd0b6fe

---------------------------------------------------------------------------------------

#bytes #iterations    t_min[usec]    t_max[usec]  t_typical[usec]

2       1000          0.81           8.37         0.83  

4       1000          0.82           3.87         0.83  

8       1000          0.81           2.97         0.83  

16      1000          0.82           3.31         0.84  

32      1000          0.88           3.41         0.89  

64      1000          0.88           3.27         0.90  

128     1000          0.91           3.55         0.93  

256     1000          1.23           3.56         1.25  

512     1000          1.30           4.15         1.32  

1024    1000          1.48           3.17         1.51  

2048    1000          1.72           4.32         1.74  

4096    1000          2.16           4.32         2.20  

---------------------------------------------------------------------------------------

Cluster B:

As shown below, in Direction I, the client side max latency is about 10X larger.  What's odd is that once I reversed the role of client and server, both showed the latency numbers that I anticipated.

 

Server:

Direct I

[root@fs11 ~]# ib_send_lat -a -c UD

 

 

************************************

* Waiting for client to connect... *

************************************

Max msg size in UD is MTU 4096

Changing to this MTU

---------------------------------------------------------------------------------------

                    Send Latency Test

Dual-port       : OFF Device         : mlx5_0

Number of qps   : 1 Transport type : IB

Connection type : UD Using SRQ      : OFF

RX depth        : 1000

Mtu             : 4096[B]

Link type       : IB

Max inline data : 188[B]

rdma_cm QPs : OFF

Data ex. method : Ethernet

---------------------------------------------------------------------------------------

local address: LID 0x02 QPN 0x002b PSN 0x79fb69

remote address: LID 0x03 QPN 0x002b PSN 0xfbae7e

---------------------------------------------------------------------------------------

#bytes #iterations    t_min[usec]    t_max[usec]  t_typical[usec]

2       1000          0.76           4.93         0.78  

4       1000          0.77           3.60         0.79  

8       1000          0.76           4.16         0.78  

16      1000          0.77           3.54         0.79  

32      1000          0.83           3.60         0.85  

64      1000          0.83           3.74         0.85  

128     1000          0.86           3.52         0.88  

256     1000          1.18           4.68         1.20  

512     1000          1.25           3.88         1.27  

1024    1000          1.44           4.71         1.46  

2048    1000          1.68           4.20         1.70  

4096    1000          2.13           3.91         2.16  

---------------------------------------------------------------------------------------

 

Client:

[root@fs10 ~]# ib_send_lat -a -c UD  192.168.12.151

Max msg size in UD is MTU 4096

Changing to this MTU

---------------------------------------------------------------------------------------

                    Send Latency Test

Dual-port       : OFF Device         : mlx5_0

Number of qps   : 1 Transport type : IB

Connection type : UD Using SRQ      : OFF

TX depth        : 1

Mtu             : 4096[B]

Link type       : IB

Max inline data : 188[B]

rdma_cm QPs : OFF

Data ex. method : Ethernet

---------------------------------------------------------------------------------------

local address: LID 0x03 QPN 0x002a PSN 0x544e64

remote address: LID 0x02 QPN 0x002a PSN 0x7babed

---------------------------------------------------------------------------------------

#bytes #iterations    t_min[usec]    t_max[usec]  t_typical[usec]

2       1000          0.76           45.78       0.78  

4       1000          0.77           30.98        0.79  

8       1000          0.76           37.99        0.78  

16      1000          0.77           43.70        0.79  

32      1000          0.83           47.34        0.85  

64      1000          0.84           39.94        0.86  

128     1000          0.86           41.16        0.88  

256     1000          1.18           37.54        1.20  

512     1000          1.24           42.94        1.26  

1024    1000          1.43           39.50        1.45  

2048    1000          1.66           42.06        1.69  

4096    1000          2.11           40.37       2.15  

---------------------------------------------------------------------------------------

 

Direct II

 

Server:

[root@fs10 ~]# ib_send_lat -a -c UD

 

 

************************************

* Waiting for client to connect... *

************************************

Max msg size in UD is MTU 4096

Changing to this MTU

---------------------------------------------------------------------------------------

                    Send Latency Test

Dual-port       : OFF Device         : mlx5_0

Number of qps   : 1 Transport type : IB

Connection type : UD Using SRQ      : OFF

RX depth        : 1000

Mtu             : 4096[B]

Link type       : IB

Max inline data : 188[B]

rdma_cm QPs : OFF

Data ex. method : Ethernet

---------------------------------------------------------------------------------------

local address: LID 0x03 QPN 0x002c PSN 0x8f46d

remote address: LID 0x02 QPN 0x002c PSN 0x9c2fe5

---------------------------------------------------------------------------------------

#bytes #iterations    t_min[usec]    t_max[usec]  t_typical[usec]

2       1000          0.76           5.30         0.78  

4       1000          0.78           4.56         0.79  

8       1000          0.76           3.80         0.78  

16      1000          0.77           3.39         0.79  

32      1000          0.83           3.07         0.84  

64      1000          0.84           5.82         0.86  

128     1000          0.86           3.95         0.88  

256     1000          1.17           4.01         1.19  

512     1000          1.25           4.64         1.27  

1024    1000          1.45           3.70         1.46  

2048    1000          1.67           5.21         1.70  

4096    1000          2.13           4.72         2.16  

---------------------------------------------------------------------------------------

Client:

[root@fs11 ~]# ib_send_lat -a -c UD 192.168.12.150

Max msg size in UD is MTU 4096

Changing to this MTU

---------------------------------------------------------------------------------------

                    Send Latency Test

Dual-port       : OFF Device         : mlx5_0

Number of qps   : 1 Transport type : IB

Connection type : UD Using SRQ      : OFF

TX depth        : 1

Mtu             : 4096[B]

Link type       : IB

Max inline data : 188[B]

rdma_cm QPs : OFF

Data ex. method : Ethernet

---------------------------------------------------------------------------------------

local address: LID 0x02 QPN 0x002c PSN 0x9c2fe5

remote address: LID 0x03 QPN 0x002c PSN 0x8f46d

---------------------------------------------------------------------------------------

#bytes #iterations    t_min[usec]    t_max[usec]  t_typical[usec]

2       1000          0.76           5.29         0.78  

4       1000          0.77           4.57         0.79  

8       1000          0.77           3.80         0.78  

16      1000          0.77           3.38         0.79  

32      1000          0.83           3.06         0.84  

64      1000          0.84           5.77         0.86  

128     1000          0.86           3.95         0.88  

256     1000          1.17           3.97         1.19  

512     1000          1.25           4.65         1.27  

1024    1000          1.44           3.69         1.47  

2048    1000          1.67           5.18         1.70  

4096    1000          2.13           4.68         2.16  

---------------------------------------------------------------------------------------

I am very puzzled by the above outcome.  Would appreciate any hints as to what I can do to figure out what's causing the large latency.

 

Viewing all 6227 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>