Forum Discussion

kishore_chennup's avatar
kishore_chennup
Icon for Nimbostratus rankNimbostratus
Nov 28, 2012

Cannot ping default GW VIPRION

HI experts,

 

 

I have a situation where I am not able to ping the default GW from my LTM VIPRION. I have subnet 10.20.20.0/24.

 

My self ip : 10.20.20.10

 

my default GW : 10.20.20.1

 

 

The default GW sits on the checkpoint FW as shown below. The vlans are all included properly. Now, the strange thing is that I can ping the self -ip from the FW but not the other way around i.e ping the GW from the F5. Also, I see packet loss while pinging the self-ip from the FW. The trunks are all fine up and running as well. No errors as well.

 

Network connectivity.

 

F5--------cisco 6500 VSS------checkpoint FW

 

Any help would be appreciated. If you need any information , I am happy to provide

 

 

Regards,

 

 

9 Replies

  • Ignoring the packet loss for a moment, can you confirm the CP FW has a rule allowing ICMP ECHO REQUESTS from the F5 to itself?
  • Hi yes it does. The F5 cannot resolve arp as well. Weird as it was all working fine couple of days ago and now suddenly it doesn't..

     

     

  • OK thanks. So are we talking about single devices or redundant pairs? I take it you are trunking using LACP and not using VPC. Are you tagging too? How is the tagging setup on the F5 and the 6500? Does the 6500 have an IP in the same VLAN, can you test from that each way?
  • Thanks mate. We are talking about VSS on cisco side and the F5 is a standalone. We are trunking using LACP. We are tagging too. On the F5 the tagging is setup by the hypervisor.. We using vCMP

     

     

    I couldnt set up an SVI to rule out issues between VSS and the F5 as we are in a change freeze.

     

     

    The weird thing is that i can ping from FW to the F5 but not the other way..Doh??

     

  • Thanks for the info. So on the F5 side all VLANs are tagged over the trunk. What about the Cisco side? What's set as the native VLAN etc.
  •  

    It's not surprising that, despite not being able to resolve the ARP entry, the LTM can reply to pings. LTM uses auto-lasthop by default, which will cause all reply packets to go back to the same source MAC from which the previous packet in a connection was received, whether or not that device would normally be part of the L3 route. In effect, LTM sees the echo request come from L2 address abc, so it sends the response back to L2 address abc.

     

     

    Do you have other vCMP guests on that chassis to test against? (You won't be able to ping the hypervisor addresses on those vlans, even if you have assigned them.) If you do, are they having any issues reaching the FW?

     

     

    When trying to ping from the LTM, do you see the ARP requests being received on the FW? If you can't capture there, can you see them if you tcpdump on the Hypervisor (not the guest)?

     

     

    I have seen some devices (the L2/3 switch in my lab) use the same link in the LAG for all packets that it cannot hash using it's default mechanism. I don't know if F5 does this or not, but if so, is it possible that the link it's using to send the ARPs through the LAG is dropping them? For example, if the 6500 hashes the echo request to LAG member 3, and ltm hashes the echo response to member 4 all would be well. But if the LTM used member 2 for the ARP requests, and member 2 is dropping short packets (or whatever), the effect would be what you describe.

     

     

    Another potential issue would be the vCMP guest blade assignment. By that I mean, if the guest is provisioned for "all-slots", but the Hypervisor has slot 3 disabled, the guest could be trying to send the ARP request to a link on blade 3, not realizing that the blade is disabled at the hypervisor level. (Yes, you'd expect the guest to detect the status of the physical blade and disable the virtual blade to match, but such is not the case in my experience.) This scenario could also benefit from the auto-lasthop feature because the Hypervisor blade that receives and forwards the echo request to the guest would not use any disabled blades in the chassis, so the guest would only receive the request on a cluster member residing on a non-disabled hypervisor blade.

     

     

     

    I realize that my description of the second scenario may be a bit difficult to follow, especially if you're not entirely familiar with how the guest sees itself. The short version is that, if a hypervisor blade is disabled, you'll want to disable the corresponding blade within the guest to be sure the LTM doesn't try to use it (and the virtual network links associated with it).

     

     

     

    Hope this helps,

     

    --jesse

     

     

     

  • Thanks Jesse. We are trying to isolate the issue. When i ping from the LTM to the FW, the FW doesnt see any thing come through at all. I havent done a tcpdump on the Hypervisor as its able to ping the FW etc. Its this specific guest tha tis being the bad boy. We have 3 other guests and they seem to be fine. I am suspecting the issue to be the port channel or the interfaces. WIll let you know after we have done some additional physical layer integrity testing

     

    • Brian_Aguirre_1's avatar
      Brian_Aguirre_1
      Icon for Nimbostratus rankNimbostratus
      Kishore, did you ever resolve your issue? we are having a similar issue where we deployed 2 new F5 5000's and after about 36 hours we lost connectivity and noticed that we are losing arp to the directly connected nexus switches.