Home » Mailing lists » Users » Using multicast in virtual servers
Using multicast in virtual servers [message #20707] |
Tue, 25 September 2007 12:01 |
Peter Hinse
Messages: 7 Registered: September 2007
|
Junior Member |
|
|
Hi all,
we have several OpenVZ instances (CentOS 4.5) running on several
physical servers (CentOS 5.0) as a QA-/testing environment for Java
applications running in JBoss application server. Since we do have to
test clustering, multicast has to work for all virtual servers, no
matter which physical host they run on.
Right now, we use veth interfaces with local IPs (192.168.*.*) and
tested multicast with ssmping (http://www.venaas.no/multicast/ssmping/)
without success.
Any idea how to get this running?
Regards,
Peter
|
|
|
|
|
Re: Using multicast in virtual servers [message #20716 is a reply to message #20710] |
Tue, 25 September 2007 14:03 |
dev
Messages: 1693 Registered: September 2005 Location: Moscow
|
Senior Member |
|
|
Peter,
Is the same setup working without openvz?
Have you used multicast before? Multicast is a bit complex to set up,
requires support from routers/switches etc., so this might well be not openvz-related.
But we setting up this test case right now to check ourselfes.
Can you please also provide a bit more information about your configuration
like whether you use bridge for veth-eth0 traffic bridging or routed networking,
any configuration options (including sysctl) you used/changed etc.?
Thanks,
Kirill
Peter Hinse wrote:
> Kirill Korotaev wrote:
>
>>what OVZ kernel do you use?
>>we'll try to check with this tool as well.
>
>
> we use the rhel5 kernel series:
>
> Linux bladeG4 2.6.18-8.1.8.el5.028stab039.1 #1 SMP Mon Jul 23 18:02:32
> MSD 2007 x86_64 x86_64 x86_64 GNU/Linux
>
>
>>Regards,
>>Kirill
>>
>>
>>>Hi all,
>>>
>>>we have several OpenVZ instances (CentOS 4.5) running on several
>>>physical servers (CentOS 5.0) as a QA-/testing environment for Java
>>>applications running in JBoss application server. Since we do have to
>>>test clustering, multicast has to work for all virtual servers, no
>>>matter which physical host they run on.
>>>
>>>Right now, we use veth interfaces with local IPs (192.168.*.*) and
>>>tested multicast with ssmping (http://www.venaas.no/multicast/ssmping/)
>>>without success.
>>>
>>>Any idea how to get this running?
>>>
>>>Regards,
>>>
>>> Peter
|
|
|
Re: Using multicast in virtual servers [message #20717 is a reply to message #20707] |
Tue, 25 September 2007 14:18 |
Vitaliy Gusev
Messages: 7 Registered: September 2007
|
Junior Member |
|
|
On the Tuesday 25 September 2007 16:01 Peter Hinse, wrote:
> Hi all,
>
> we have several OpenVZ instances (CentOS 4.5) running on several
> physical servers (CentOS 5.0) as a QA-/testing environment for Java
> applications running in JBoss application server. Since we do have to
> test clustering, multicast has to work for all virtual servers, no
> matter which physical host they run on.
>
> Right now, we use veth interfaces with local IPs (192.168.*.*) and
> tested multicast with ssmping (http://www.venaas.no/multicast/ssmping/)
> without success.
>
> Any idea how to get this running?
Please, print ssmping command, output for ifconfig, route -n, etc.
>
> Regards,
>
> Peter
--
Thanks,
Vitaliy Gusev
|
|
|
Re: Using multicast in virtual servers [message #20760 is a reply to message #20716] |
Wed, 26 September 2007 03:18 |
Daniel Pittman
Messages: 26 Registered: January 2007
|
Junior Member |
|
|
Kirill Korotaev <dev@sw.ru> writes:
> Is the same setup working without openvz? Have you used multicast
> before? Multicast is a bit complex to set up, requires support from
> routers/switches etc., so this might well be not openvz-related. But
> we setting up this test case right now to check ourselfes.
>
> Can you please also provide a bit more information about your
> configuration like whether you use bridge for veth-eth0 traffic
> bridging or routed networking, any configuration options (including
> sysctl) you used/changed etc.?
One thing that is worth noting: I found a bug in the ... veth code, I
think, where it wouldn't pass a multicast packet through. The code
checked with the 'is_broadcast' flag, for a matching mac, and assumed
that anything else was not for this host.
Perhaps this is a similar issue? I can try to dig out the fault report
if it helps, but at the time it was simply changing is_broadcast to
include an is_multicast test on the Ethernet MAC.
Regards,
Daniel
I hope that actually helps. :)
--
Daniel Pittman <daniel@cybersource.com.au> Phone: 03 9621 2377
Level 4, 10 Queen St, Melbourne Web: http://www.cyber.com.au
Cybersource: Australia's Leading Linux and Open Source Solutions Company
|
|
|
|
Re: Using multicast in virtual servers [message #20776 is a reply to message #20760] |
Wed, 26 September 2007 07:13 |
dev
Messages: 1693 Registered: September 2005 Location: Moscow
|
Senior Member |
|
|
Daniel Pittman wrote:
> Kirill Korotaev <dev@sw.ru> writes:
>
>
>>Is the same setup working without openvz? Have you used multicast
>>before? Multicast is a bit complex to set up, requires support from
>>routers/switches etc., so this might well be not openvz-related. But
>>we setting up this test case right now to check ourselfes.
>>
>>Can you please also provide a bit more information about your
>>configuration like whether you use bridge for veth-eth0 traffic
>>bridging or routed networking, any configuration options (including
>>sysctl) you used/changed etc.?
>
>
> One thing that is worth noting: I found a bug in the ... veth code, I
> think, where it wouldn't pass a multicast packet through. The code
> checked with the 'is_broadcast' flag, for a matching mac, and assumed
> that anything else was not for this host.
plz make sure you really use and look at the sources of 028stab039 kernel.
this check in veth_xmit() was fixed in 028stab034 with this commit:
http://git.openvz.org/?p=linux-2.6.18-openvz;a=commitdiff;h=993241dcdfc8ae22d339e08ed78db6e9760b1d89
>
> Perhaps this is a similar issue? I can try to dig out the fault report
> if it helps, but at the time it was simply changing is_broadcast to
> include an is_multicast test on the Ethernet MAC.
and it started to work?
Thanks,
Kirill
|
|
|
Re: Using multicast in virtual servers [message #20777 is a reply to message #20716] |
Wed, 26 September 2007 07:28 |
Peter Hinse
Messages: 7 Registered: September 2007
|
Junior Member |
|
|
Kirill Korotaev wrote:
> Peter,
>
> Is the same setup working without openvz?
> Have you used multicast before? Multicast is a bit complex to set up,
> requires support from routers/switches etc., so this might well be not openvz-related.
> But we setting up this test case right now to check ourselfes.
We use several JBoss clusters with multicast in our datacenters.
Multicast itself works between the physical hosts:
ssmping joined (S,G) = (192.168.198.54,232.43.211.234)
pinging S from 192.168.198.53
unicast from 192.168.198.54, seq=1 dist=0 time=1.513 ms
multicast from 192.168.198.54, seq=1 dist=0 time=1.528 ms
unicast from 192.168.198.54, seq=2 dist=0 time=0.107 ms
multicast from 192.168.198.54, seq=2 dist=0 time=0.115 ms
unicast from 192.168.198.54, seq=3 dist=0 time=0.097 ms
multicast from 192.168.198.54, seq=3 dist=0 time=0.104 ms
unicast from 192.168.198.54, seq=4 dist=0 time=0.113 ms
multicast from 192.168.198.54, seq=4 dist=0 time=0.123 ms
unicast from 192.168.198.54, seq=5 dist=0 time=0.102 ms
multicast from 192.168.198.54, seq=5 dist=0 time=0.113 ms
unicast from 192.168.198.54, seq=6 dist=0 time=0.119 ms
multicast from 192.168.198.54, seq=6 dist=0 time=0.130 ms
--- 192.168.198.54 statistics ---
6 packets transmitted, time 5276 ms
unicast:
6 packets received, 0% packet loss
rtt min/avg/max/std-dev = 0.097/0.341/1.513/0.524 ms
multicast:
6 packets received, 0% packet loss since first mc packet (seq 1) recvd
rtt min/avg/max/std-dev = 0.104/0.352/1.528/0.526 ms
If I try the ssmpingd/ssmping between two virtual instances on two
different hosts (or from one virtual instance to a physical host an vice
versa):
pinging S from 192.168.198.142
unicast from 192.168.198.132, seq=1 dist=1 time=2627.063 ms
unicast from 192.168.198.132, seq=2 dist=1 time=1626.828 ms
unicast from 192.168.198.132, seq=3 dist=1 time=626.718 ms
unicast from 192.168.198.132, seq=4 dist=1 time=0.100 ms
unicast from 192.168.198.132, seq=5 dist=1 time=0.101 ms
unicast from 192.168.198.132, seq=6 dist=1 time=0.150 ms
--- 192.168.198.132 statistics ---
6 packets transmitted, time 5372 ms
unicast:
6 packets received, 0% packet loss
rtt min/avg/max/std-dev = 0.100/813.493/2627.063/997.511 ms
multicast:
0 packets received, 100% packet loss
> Can you please also provide a bit more information about your configuration
> like whether you use bridge for veth-eth0 traffic bridging or routed networking,
> any configuration options (including sysctl) you used/changed etc.?
sysctl settings for the virtual interface on one of the host systems:
net.ipv4.conf.veth1981320.promote_secondaries = 0
net.ipv4.conf.veth1981320.force_igmp_version = 0
net.ipv4.conf.veth1981320.disable_policy = 0
net.ipv4.conf.veth1981320.disable_xfrm = 0
net.ipv4.conf.veth1981320.arp_accept = 0
net.ipv4.conf.veth1981320.arp_ignore = 0
net.ipv4.conf.veth1981320.arp_announce = 0
net.ipv4.conf.veth1981320.arp_filter = 0
net.ipv4.conf.veth1981320.tag = 0
net.ipv4.conf.veth1981320.log_martians = 0
net.ipv4.conf.veth1981320.bootp_relay = 0
net.ipv4.conf.veth1981320.medium_id = 0
net.ipv4.conf.veth1981320.proxy_arp = 1
net.ipv4.conf.veth1981320.accept_source_route = 1
net.ipv4.conf.veth1981320.send_redirects = 1
net.ipv4.conf.veth1981320.rp_filter = 0
net.ipv4.conf.veth1981320.shared_media = 1
net.ipv4.conf.veth1981320.secure_redirects = 1
net.ipv4.conf.veth1981320.accept_redirects = 1
net.ipv4.conf.veth1981320.mc_forwarding = 0
net.ipv4.conf.veth1981320.forwarding = 1
Network config on the VPS:
eth0 Link encap:Ethernet HWaddr 00:0C:29:0A:3D:C6
inet addr:192.168.198.132 Bcast:192.168.198.255
Mask:255.255.255.0
inet6 addr: fe80::20c:29ff:fe0a:3dc6/64 Scope:Link
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:2855871 errors:0 dropped:0 overruns:0 frame:0
TX packets:5495903 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:0
RX bytes:1453942778 (1.3 GiB) TX bytes:3721100335 (3.4 GiB)
Kernel IP routing table
Destination Gateway Genmask Flags Metric Ref Use
Iface
192.168.198.0 0.0.0.0 255.255.255.0 U 0 0 0 eth0
169.254.0.0 0.0.0.0 255.255.0.0 U 0 0 0 eth0
224.0.0.0 0.0.0.0 240.0.0.0 U 0 0 0 eth0
Any more information you need?
Regards,
Peter
|
|
|
Re: Using multicast in virtual servers [message #20779 is a reply to message #20777] |
Wed, 26 September 2007 07:35 |
dev
Messages: 1693 Registered: September 2005 Location: Moscow
|
Senior Member |
|
|
# uname -a
(from both VEs if you VE <-> VE multicasting)
Thanks,
Kirill
P.S. plz check my another today email about veth code.
Peter Hinse wrote:
> Kirill Korotaev wrote:
>
>>Peter,
>>
>>Is the same setup working without openvz?
>>Have you used multicast before? Multicast is a bit complex to set up,
>>requires support from routers/switches etc., so this might well be not openvz-related.
>>But we setting up this test case right now to check ourselfes.
>
>
> We use several JBoss clusters with multicast in our datacenters.
> Multicast itself works between the physical hosts:
>
> ssmping joined (S,G) = (192.168.198.54,232.43.211.234)
> pinging S from 192.168.198.53
> unicast from 192.168.198.54, seq=1 dist=0 time=1.513 ms
> multicast from 192.168.198.54, seq=1 dist=0 time=1.528 ms
> unicast from 192.168.198.54, seq=2 dist=0 time=0.107 ms
> multicast from 192.168.198.54, seq=2 dist=0 time=0.115 ms
> unicast from 192.168.198.54, seq=3 dist=0 time=0.097 ms
> multicast from 192.168.198.54, seq=3 dist=0 time=0.104 ms
> unicast from 192.168.198.54, seq=4 dist=0 time=0.113 ms
> multicast from 192.168.198.54, seq=4 dist=0 time=0.123 ms
> unicast from 192.168.198.54, seq=5 dist=0 time=0.102 ms
> multicast from 192.168.198.54, seq=5 dist=0 time=0.113 ms
> unicast from 192.168.198.54, seq=6 dist=0 time=0.119 ms
> multicast from 192.168.198.54, seq=6 dist=0 time=0.130 ms
>
> --- 192.168.198.54 statistics ---
> 6 packets transmitted, time 5276 ms
> unicast:
> 6 packets received, 0% packet loss
> rtt min/avg/max/std-dev = 0.097/0.341/1.513/0.524 ms
> multicast:
> 6 packets received, 0% packet loss since first mc packet (seq 1) recvd
> rtt min/avg/max/std-dev = 0.104/0.352/1.528/0.526 ms
>
> If I try the ssmpingd/ssmping between two virtual instances on two
> different hosts (or from one virtual instance to a physical host an vice
> versa):
>
> pinging S from 192.168.198.142
> unicast from 192.168.198.132, seq=1 dist=1 time=2627.063 ms
> unicast from 192.168.198.132, seq=2 dist=1 time=1626.828 ms
> unicast from 192.168.198.132, seq=3 dist=1 time=626.718 ms
> unicast from 192.168.198.132, seq=4 dist=1 time=0.100 ms
> unicast from 192.168.198.132, seq=5 dist=1 time=0.101 ms
> unicast from 192.168.198.132, seq=6 dist=1 time=0.150 ms
>
> --- 192.168.198.132 statistics ---
> 6 packets transmitted, time 5372 ms
> unicast:
> 6 packets received, 0% packet loss
> rtt min/avg/max/std-dev = 0.100/813.493/2627.063/997.511 ms
> multicast:
> 0 packets received, 100% packet loss
>
>
>
>>Can you please also provide a bit more information about your configuration
>>like whether you use bridge for veth-eth0 traffic bridging or routed networking,
>>any configuration options (including sysctl) you used/changed etc.?
>
>
> sysctl settings for the virtual interface on one of the host systems:
>
> net.ipv4.conf.veth1981320.promote_secondaries = 0
> net.ipv4.conf.veth1981320.force_igmp_version = 0
> net.ipv4.conf.veth1981320.disable_policy = 0
> net.ipv4.conf.veth1981320.disable_xfrm = 0
> net.ipv4.conf.veth1981320.arp_accept = 0
> net.ipv4.conf.veth1981320.arp_ignore = 0
> net.ipv4.conf.veth1981320.arp_announce = 0
> net.ipv4.conf.veth1981320.arp_filter = 0
> net.ipv4.conf.veth1981320.tag = 0
> net.ipv4.conf.veth1981320.log_martians = 0
> net.ipv4.conf.veth1981320.bootp_relay = 0
> net.ipv4.conf.veth1981320.medium_id = 0
> net.ipv4.conf.veth1981320.proxy_arp = 1
> net.ipv4.conf.veth1981320.accept_source_route = 1
> net.ipv4.conf.veth1981320.send_redirects = 1
> net.ipv4.conf.veth1981320.rp_filter = 0
> net.ipv4.conf.veth1981320.shared_media = 1
> net.ipv4.conf.veth1981320.secure_redirects = 1
> net.ipv4.conf.veth1981320.accept_redirects = 1
> net.ipv4.conf.veth1981320.mc_forwarding = 0
> net.ipv4.conf.veth1981320.forwarding = 1
>
>
> Network config on the VPS:
>
> eth0 Link encap:Ethernet HWaddr 00:0C:29:0A:3D:C6
> inet addr:192.168.198.132 Bcast:192.168.198.255
> Mask:255.255.255.0
> inet6 addr: fe80::20c:29ff:fe0a:3dc6/64 Scope:Link
> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
> RX packets:2855871 errors:0 dropped:0 overruns:0 frame:0
> TX packets:5495903 errors:0 dropped:0 overruns:0 carrier:0
> collisions:0 txqueuelen:0
> RX bytes:1453942778 (1.3 GiB) TX bytes:3721100335 (3.4 GiB)
>
> Kernel IP routing table
> Destination Gateway Genmask Flags Metric Ref Use
> Iface
> 192.168.198.0 0.0.0.0 255.255.255.0 U 0 0 0 eth0
> 169.254.0.0 0.0.0.0 255.255.0.0 U 0 0 0 eth0
> 224.0.0.0 0.0.0.0 240.0.0.0 U 0 0 0 eth0
>
> Any more information you need?
>
> Regards,
>
> Peter
>
|
|
|
|
|
Re: Using multicast in virtual servers [message #20787 is a reply to message #20776] |
Wed, 26 September 2007 08:31 |
Daniel Pittman
Messages: 26 Registered: January 2007
|
Junior Member |
|
|
Kirill Korotaev <dev@sw.ru> writes:
> Daniel Pittman wrote:
>> Kirill Korotaev <dev@sw.ru> writes:
>>
>>>Is the same setup working without openvz? Have you used multicast
>>>before? Multicast is a bit complex to set up, requires support from
>>>routers/switches etc., so this might well be not openvz-related. But
>>>we setting up this test case right now to check ourselfes.
>>>
>>>Can you please also provide a bit more information about your
>>>configuration like whether you use bridge for veth-eth0 traffic
>>>bridging or routed networking, any configuration options (including
>>>sysctl) you used/changed etc.?
>>
>> One thing that is worth noting: I found a bug in the ... veth code, I
>> think, where it wouldn't pass a multicast packet through. The code
>> checked with the 'is_broadcast' flag, for a matching mac, and assumed
>> that anything else was not for this host.
>
> plz make sure you really use and look at the sources of 028stab039 kernel.
> this check in veth_xmit() was fixed in 028stab034 with this commit:
> http://git.openvz.org/?p=linux-2.6.18-openvz;a=commitdiff;h=993241dcdfc8ae22d339e08ed78db6e9760b1d89
I suspected that you would remember. :)
>> Perhaps this is a similar issue? I can try to dig out the fault report
>> if it helps, but at the time it was simply changing is_broadcast to
>> include an is_multicast test on the Ethernet MAC.
>
> and it started to work?
I never got to test it; that particular job (enable CUPS server browse
announcements in a VE) is still outstanding on my list because it was
low priority. (sorry)
Daniel
--
Daniel Pittman <daniel@cybersource.com.au> Phone: 03 9621 2377
Level 4, 10 Queen St, Melbourne Web: http://www.cyber.com.au
Cybersource: Australia's Leading Linux and Open Source Solutions Company
|
|
|
|
Re: Using multicast in virtual servers [message #20789 is a reply to message #20785] |
Wed, 26 September 2007 09:00 |
Peter Hinse
Messages: 7 Registered: September 2007
|
Junior Member |
|
|
Vitaliy Gusev wrote:
> Please, print output for ssmping in VE (which doesn't work), route for VE0,
> brctl show for VE0, ifconfig for VE0.
ssmping output in VE:
pinging S from 192.168.198.142
unicast from 192.168.198.132, seq=1 dist=1 time=2627.063 ms
unicast from 192.168.198.132, seq=2 dist=1 time=1626.828 ms
unicast from 192.168.198.132, seq=3 dist=1 time=626.718 ms
unicast from 192.168.198.132, seq=4 dist=1 time=0.100 ms
unicast from 192.168.198.132, seq=5 dist=1 time=0.101 ms
unicast from 192.168.198.132, seq=6 dist=1 time=0.150 ms
--- 192.168.198.132 statistics ---
6 packets transmitted, time 5372 ms
unicast:
6 packets received, 0% packet loss
rtt min/avg/max/std-dev = 0.100/813.493/2627.063/997.511 ms
multicast:
0 packets received, 100% packet loss
route for VE0:
Kernel IP routing table
Destination Gateway Genmask Flags Metric Ref Use
Iface
192.168.198.141 0.0.0.0 255.255.255.255 UH 0 0 0
veth1981410
192.168.198.142 0.0.0.0 255.255.255.255 UH 0 0 0
veth1981420
195.x.x.x 0.0.0.0 255.255.255.255 UH 0 0 0
veth1981411
195.x.x.x 0.0.0.0 255.255.255.224 U 0 0 0 eth0
192.168.198.0 0.0.0.0 255.255.255.0 U 0 0 0 eth0
169.254.0.0 0.0.0.0 255.255.0.0 U 0 0 0 eth0
224.0.0.0 0.0.0.0 240.0.0.0 U 0 0 0 eth0
0.0.0.0 195.x.x.x 0.0.0.0 UG 0 0 0 eth0
ifconfig for VE0:
eth0 Link encap:Ethernet HWaddr 00:1A:64:32:0A:F8
inet addr:192.168.198.54 Bcast:192.168.198.255
Mask:255.255.255.0
inet6 addr: fe80::21a:64ff:fe32:af8/64 Scope:Link
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:19330162 errors:0 dropped:0 overruns:0 frame:0
TX packets:3542937 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:1000
RX bytes:4931551683 (4.5 GiB) TX bytes:1232779800 (1.1 GiB)
Interrupt:98 Memory:da000000-da011100
eth0:1 Link encap:Ethernet HWaddr 00:1A:64:32:0A:F8
inet addr:195.x.x.x Bcast:195.x.x.x Mask:255.255.255.224
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
Interrupt:98 Memory:da000000-da011100
lo Link encap:Local Loopback
inet addr:127.0.0.1 Mask:255.0.0.0
inet6 addr: ::1/128 Scope:Host
UP LOOPBACK RUNNING MTU:16436 Metric:1
RX packets:788 errors:0 dropped:0 overruns:0 frame:0
TX packets:788 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:0
RX bytes:246284 (240.5 KiB) TX bytes:246284 (240.5 KiB)
venet0 Link encap:UNSPEC HWaddr
00-00-00-00-00-00-00-00-00-00-00-00-00-00-00-00
UP BROADCAST POINTOPOINT RUNNING NOARP MTU:1500 Metric:1
RX packets:0 errors:0 dropped:0 overruns:0 frame:0
TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:0
RX bytes:0 (0.0 b) TX bytes:0 (0.0 b)
veth1981410 Link encap:Ethernet HWaddr 00:0C:29:91:B1:81
inet6 addr: fe80::20c:29ff:fe91:b181/64 Scope:Link
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:345225 errors:0 dropped:0 overruns:0 frame:0
TX packets:379368 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:0
RX bytes:221623267 (211.3 MiB) TX bytes:349531019 (333.3 MiB)
veth1981411 Link encap:Ethernet HWaddr 00:0C:29:91:B1:83
inet6 addr: fe80::20c:29ff:fe91:b183/64 Scope:Link
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:558932 errors:0 dropped:0 overruns:0 frame:0
TX packets:696100 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:0
RX bytes:75514675 (72.0 MiB) TX bytes:432185018 (412.1 MiB)
veth1981420 Link encap:Ethernet HWaddr 00:0C:29:F7:A0:88
inet6 addr: fe80::20c:29ff:fef7:a088/64 Scope:Link
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:4101690 errors:0 dropped:0 overruns:0 frame:0
TX packets:3166888 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:0
RX bytes:1071935233 (1022.2 MiB) TX bytes:3124980969 (2.9 GiB)
No bridging used right now.
|
|
|
|
|
Re: Using multicast in virtual servers [message #21090 is a reply to message #20803] |
Mon, 01 October 2007 17:00 |
Peter Hinse
Messages: 7 Registered: September 2007
|
Junior Member |
|
|
Vitaliy Gusev schrieb:
>> On the Wednesday 26 September 2007 13:00 Peter Hinse, wrote:
>>> veth1981420 Link encap:Ethernet HWaddr 00:0C:29:F7:A0:88
>>> inet6 addr: fe80::20c:29ff:fef7:a088/64 Scope:Link
>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
>>> RX packets:4101690 errors:0 dropped:0 overruns:0 frame:0
>>> TX packets:3166888 errors:0 dropped:0 overruns:0 carrier:0
>>> collisions:0 txqueuelen:0
>>> RX bytes:1071935233 (1022.2 MiB) TX bytes:3124980969 (2.9 GiB)
>>>
>>> No bridging used right now.
> You must use a bridge. Multicast packets is not forwarded.
>
OK, multicast works with bridging enabled. Big thx for help!
Regards,
Peter
|
|
|
Goto Forum:
Current Time: Tue Nov 05 18:52:16 GMT 2024
Total time taken to generate the page: 0.03638 seconds
|