Not to belabor this, but so interesting... I need a FEC-for-Dummies or FEC-for-IP/Ethernet-Engineers...
Shown below, my 400g interface with NO config at all... Interface has no traffic at all, no packets at all.... BUT, lots of FEC hits. Interesting this FEC-thing. I'd love to have a fiber splitter and see if wireshark could read it and show me what FEC looks like...but something tells me i would need a 400g sniffer to read it, lol
It's like FEC (fec119 in this case) is this automatic thing running between interfaces (hardware i guess), with no protocols and nothing needed at all in order to function.
-Aaron
{master}
me@mx960> show configuration interfaces et-7/1/4 | display set
{master}
me@mx960>
{master}
me@mx960> clear interfaces statistics et-7/1/4
{master}
me@mx960> show interfaces et-7/1/4 | grep packet
Input packets : 0
Output packets: 0
{master}
me@mx960> show interfaces et-7/1/4 | grep "put rate"
Input rate : 0 bps (0 pps)
Output rate : 0 bps (0 pps)
{master}
me@mx960> show interfaces et-7/1/4 | grep rror
Link-level type: Ethernet, MTU: 1514, MRU: 1522, Speed: 400Gbps, BPDU Error: None, Loop Detect PDU Error: None, Loopback: Disabled, Source filtering: Disabled,
Bit errors 0
Errored blocks 0
Ethernet FEC statistics Errors
FEC Corrected Errors 28209
FEC Uncorrected Errors 0
FEC Corrected Errors Rate 2347
FEC Uncorrected Errors Rate 0
{master}
me@mx960> show interfaces et-7/1/4 | grep packet
Input packets : 0
Output packets: 0
{master}
me@mx960> show interfaces et-7/1/4 | grep "put rate"
Input rate : 0 bps (0 pps)
Output rate : 0 bps (0 pps)
{master}
me@mx960> show interfaces et-7/1/4 | grep rror
Link-level type: Ethernet, MTU: 1514, MRU: 1522, Speed: 400Gbps, BPDU Error: None, Loop Detect PDU Error: None, Loopback: Disabled, Source filtering: Disabled,
Bit errors 0
Errored blocks 0
Ethernet FEC statistics Errors
FEC Corrected Errors 45153
FEC Uncorrected Errors 0
FEC Corrected Errors Rate 29
FEC Uncorrected Errors Rate 0
{master}
me@mx960> show interfaces et-7/1/4 | grep packet
Input packets : 0
Output packets: 0
{master}
me@mx960> show interfaces et-7/1/4 | grep "put rate"
Input rate : 0 bps (0 pps)
Output rate : 0 bps (0 pps)
{master}
me@mx960> show interfaces et-7/1/4 | grep rror
Link-level type: Ethernet, MTU: 1514, MRU: 1522, Speed: 400Gbps, BPDU Error: None, Loop Detect PDU Error: None, Loopback: Disabled, Source filtering: Disabled,
Bit errors 0
Errored blocks 0
Ethernet FEC statistics Errors
FEC Corrected Errors 57339
FEC Uncorrected Errors 0
FEC Corrected Errors Rate 2378
FEC Uncorrected Errors Rate 0
{master}
me@mx960>
On 4/18/2024 7:13 AM, Mark Tinka wrote:
On 4/17/24 23:24, Aaron Gould wrote:
Well JTAC just said that it seems ok, and
that 400g is going to show 4x more than 100g "This is due to
having to synchronize much more to support higher data."
We've seen the same between Juniper and Arista boxes in the same
rack running at 100G, despite cleaning fibres, swapping optics,
moving ports, moving line cards, e.t.c. TAC said it's a non-issue,
and to be expected, and shared the same KB's.
It's a bit disconcerting when you plot the data on your NMS, but
it's not material.
Mark.
--
-Aaron