To all, I am running a Windows based high performance computing application that uses "reliable" multicast (29West) on a gigabit LAN. All systems are logically on the same VLAN and even on the same physical switch The application is set to use an 8k buffer and therefore results in IP fragmentation when datagrams are transmitted. The application is sensitive to any latency or data loss (of course) and uses a proprietary mechanism to create TCP-like retransmissions in case there is any actual data loss. Unfortunately, becasue of the fragmentation during the retransmission window all ip fragments must be resent even though only one may have been lost. If the buffer size is tweeked to the ~1460 this may fix the fragmentation but will the side effects be less throughput and possibly more latency. Is there a sweet spot for UDP on an ethernet segment? Philip