Just bare logics (a lost art in the modern datacom world, i guess).
logics/insanity, l2/l3, no real difference, right...
Make this gedunkenexperiment: take each mcast packet replicating point and replace it with a cache with very small retention time. The thing will replicate data in exactly the same way; and the packets will flow along the same paths. Therefore it is _at least_ as efficient
So you're saying that an application-specific cache would be more efficient than non-application-specific multicast? OK, let's consider this. How would distribution/replication be done? What was the destination address of the original packet .. let's say, 225.X.X.X? Wow!
(Note that i do not consider possibility of building mcast trees dependent on traffic or bandwidth reservation - the algorithmic complexity involved makes that an intractable problem (it is believed to be NP-complete in general case); even the best heuristic algorithms for such planning place it beyond realm of computable for the networks fraction of size of present Internet).
But we're not discussing that. we're discussing multicast v/s unicast (or was it QoS?), but you're somehow tying the two together when they're completely different things. and are you suggesting that dense-mode (blindly distributing), to coin the term, is more efficient than say, sparse-mode (multicast?) .. not even in your bare logics world.
Caching is not employing any routing information exchange. Therefore it is a) oblivious to the state of other caches or to the changes in network topology and b) is invulnerable to the bogus routing information and flap-like DoS attacks.
a) Yes, it indeed does have different challenges than multicast. It still relies on the underlying layers which b) are not oblivious to such DoS attacks .. in todays Internet (maybe not yours, but todays) where 2 million line prefix filters are about as real as millions of multicast groups.
99% of Web content is write-once. It does not need any fancy management. The remaining 1% can be delivered end-to-end.
I don't recall scoping this to only "Web", but it does seem to best augment your argument, better stick to it :-) I'm of the opinion that an application-agnostic approach is a better idea.
(BTW, i do consider intelligent cache-synchronization development efforts seriously misguided; there's a much simpler and much more scalable solution to the cache performance problem. If someone wants to invest, i'd like to talk about it :)
Ahh, I see. Everything sux, but give me cash (pun intended :-) and I'll give you cold fusion.
It is not that caching is a silver bullet, it is rather that multicating is unuseable at a large scale.
many to many, perhaps it does have it's challenges. one to many, it works today.
Well, philosophical note: science is _all_ about generalizing. For an inventor of perpetuum mobile the flat refusal of a modern physicist to look into details to assert that it will not work sure looks as an oversimplifying. After all, the details of actual construction sure are a lot more complex than the second law of thermodynamics.
Second law of thermodynamics: OK, coke loses it's fizz, but putting a cap on the bottle wasn't that hard.
In this case, i just do not care to go into details of implementations. The L2/L3 mcasting is not scalable and _cannot be made_ scalable for reasons having nothing to do with deficiencies of protocols.
Summary: It sux, I can't can't tell you why, but it just sux!
Caching algorithms do not have similar limitations, solely because they do not rely on distributed computations. So they have a chance of working. Of course, nothing "just works".
PS To those who point that provider ABC already sells mcast service: there's an old saying at NASA that with enough thrust even pigs can fly. However, no reactively propulsed hog is likely to make it to an orbit all on its own.
However, if he sends enough mail to NANOG, he'll likely find someone to fund even shoving a rocket up a dogs ass :-) -danny (who won't waste any additional time in this thread)