Patrick,

> Matt: Are you arguing the CDNs are at fault because the game companies tell everyone to download simultaneously, and
> the ISPs sold the users connectivity to do that download?

While a gross oversimplification, yes, that's basically what I'm saying; I know it may not be a popular opinion, but I stand by it. There aren't any villains here though, just lots of good suggestions in this thread to make the internet work better for everyone, without spending large swaths of money to cover the demand of an infrequent, large, update for a single game.

CNDs do, however, have a responsibility to be good netizens and get this data out in a manner that doesn't cause disruption. They know the technical challenges of distributing that much data to the masses, the game company does not, that's why they outsourced it to a CDN. If the CDN knows what the gaming company is asking for is pushing the limits of our current infrastructure, they have a responsibility to relay those limitations that are outside of their control to their customer, as any responsible vendor should. Instead, there may be an element of "oh yeah sure, we can do that" or "the customer is always right" going on here and modern limitations are being disregarded.

The idea behind the internet is not that every user can always have their entire capacity available for a single destination regardless of what everyone else is doing (and especially if they're all going to the same place too), the user has purchased that capacity into their provider's network as a whole, gaining access to all of their connections to all of the various endpoints on the internet at a backbone and peering capacity that is economically viable given normal peak demand with some cushion built-in for redundancy. If that's the desire to have full capacity available to Akamai available at all times, then everyone needs dedicated P2P circuits direct to Akamai, but that's not practical.

If you own an ISP and you're not oversubscribing, you're not making money, period. To use your analogy, if you've ever been to a gym in January, you've seen a similar phenomenon first-hand. There aren't enough machines for everyone, and the gym isn't going to add them because this is a once-a-year thing and it goes away after a few weeks when many people get tired of fulfilling their new year's resolutions. Why should the gym limit its sales to exactly the capacity it has available (or add a lot more machines), when it knows that for the overwhelming majority of the year, there will always be dozens of empty machines across the floor?

If Akamai was doing these updates more frequently (weekly for example) then sure, it is on the ISP to augment, because this has become the "new normal". But these updates for a single game that happen once per quarter are hardly able to be considered normal. Sure, some day 50GB updates will be the norm, but that's not today, and when it is, somebody else will be pushing out 250GB updates quarterly. This problem isn't going away soon, and it can't be fixed permanently by just adding more capacity, it's a complex technical challenge that CDN's ought to give some more thought, and game publishers should start considering too.

-Matt

On Thu, Apr 1, 2021 at 3:30 PM Patrick W. Gilmore <patrick@ianai.net> wrote:
I am sorry, maybe I misunderstand.

Matt: Are you arguing the CDNs are at fault because the game companies tell everyone to download simultaneously, and the ISPs sold the users connectivity to do that download?

If so, are you really arguing “I sold my users XXX Mbps, but if they try to use it, I want *YOU* to tell them no”? Because that is what it sounds like to me.

Imagine a gym sold 10,000 memberships with 10 machines because they figured everyone would sit on their ass. They would be right most of the time - and rake in that sweet, sweet monthly cash for zero effort after the initial sale. But if Oprah or Cher or Biden or some other person famous enough to go by one name tweets “get your ass to the gym!!", does the gym really think getting mad at Oprah is the solution? Or do they expect Oprah to pay for the extra machines they have to buy now?

Selling a service you know will not work if everyone uses it simultaneously can be profitable, but there is risk. Do not blame third parties when you lose that bet.

-- 
TTFN,
patrick

On Apr 1, 2021, at 5:04 PM, Tom Beecher <beecher@beecher.cc> wrote:

No disrespect taken, or intended back in your direction, but again, I disagree. 

If thousands of users are downloading 50G files at the same time, it really doesn't matter if they are pulling from a CDN or the origin directly. The volume of traffic still has to be handled. Yes, it's a burden on the ISP, but it's a burden created by the usage created by their subscribers. 


On Thu, Apr 1, 2021 at 4:57 PM Matt Erculiani <merculiani@gmail.com> wrote:
Tom,

All due respect, but there is a massive difference between one user downloading 50G and thousands of users each downloading 50G when they all go to play their videogame of choice at around the same time.

-Matt



On Thu, Apr 1, 2021 at 2:46 PM Tom Beecher <beecher@beecher.cc> wrote:
A user sends a few megabytes of request and receives 50 gigs of reply. They aren't DDoSing the network, but they're amplifying a single 50 gig copy they receive from the mothership and turning it into likely tens of terabytes of traffic.
Yes, that's a CDN's job, but that volume of legitimate traffic and the very tiny window with which it is transmitted is likely to be a burden for even the largest residential ISPs.

I'm sitting at home, and I could send a 50k request for a 50G file right now from a source not fronted by a CDN. What do? My ISP is still has to deliver it to me. The fact that the 50G file does or does not come from a CDN is irrelevant. The CDN just happens to be a point source that a lot of users happen to connect to. 

CDNs want to have the best performance to users because that's what brings them business. A poorly performing CDN will lose customers to a better performing one. The trend for years has been instead of ISPs investing in infrastructure to effectively handle the traffic that their users request, they turf that to CDNs. In many cases, a CDN will put a cache box in or extend a circuit at a loss to them, because they know if the performance metrics get bad, business will be taken elsewhere, even if the CAUSE of the poor performance is actually at the edge of, or inside , the ISPs network. 

ISPs in the US can get away with this because their users are captive and rarely have an alternative choice of provider.  


On Thu, Apr 1, 2021 at 4:33 PM Matt Erculiani <merculiani@gmail.com> wrote:
Patrick,

> First, to be blunt, if you really think Akamai nodes are “sitting idle for weeks” before CoD comes out with a new game,
> you are clearly confused.

"Idle" in the sense that when you look at a graph of traffic before and after a large push such as this makes the rest of the week's traffic look like a horizontal line at the bottom, admittedly poor word choice, yes, but far from "confused" as to what CDNs do under relatively normal circumstances. Otherwise very valid points you've raised.

Tom,

> Akamai, and other CDNs, do not **generate** traffic ; they serve the requests generated by users.  

A user sends a few megabytes of request and receives 50 gigs of reply. They aren't DDoSing the network, but they're amplifying a single 50 gig copy they receive from the mothership and turning it into likely tens of terabytes of traffic.
Yes, that's a CDN's job, but that volume of legitimate traffic and the very tiny window with which it is transmitted is likely to be a burden for even the largest residential ISPs.

-Matt

On Thu, Apr 1, 2021 at 2:09 PM Patrick W. Gilmore <patrick@ianai.net> wrote:
Matt:

I am going to disagree with your characterization of how Akamai - and many other CDNs - manage things. First, to be blunt, if you really think Akamai nodes are “sitting idle for weeks” before CoD comes out with a new game, you are clearly confused.

More importantly, I know for a fact Akamai has spent ungodly amounts of money & resources putting content precisely where the ISPs ask them to put it, deliver it over the pipes the ISPs ask them to deliver it, at precisely the capacity the ISPs tell them.

On the other hand, I agree with your characterization of residential broadband. It is ridiculous to expect a neighborhood with 1,000 homes each with 1 Gbps links to have a terabit of uplink capacity. But it also should have a lot more than 10 Gbps, IMHO. Unfortunately, most neighborhoods I have seen are closer to the latter than the former.

Finally, this could quickly devolve into finger pointing. You say the CDNs bear some responsibility? They may well respond that the large broadband providers ask for cash to interconnect - but still require the CDNs to do all the work. The CDNs did not create the content, or tell the users which content to pull. When I pay $NATIONAL_PROVIDER, I expect them to provide me with access to the Internet. Not just to the content that pays that provider.

Personally, I have zero problems with the ISPs saying “give me a cache to put here with this sized uplink” or “please deliver to these users over this xconn / IX / whatever”. I have a huge problem with the ISPs blaming the ISPs for delivering what the ISP’s users request.

Of course, this could all be solved if there were more competition in broadband in the US (and many other countries). But that is a totally different 10,000 post thread (that we have had many dozens of times).

-- 
TTFN,
patrick

On Apr 1, 2021, at 3:53 PM, Matt Erculiani <merculiani@gmail.com> wrote:

Niels,

I think to clarify Jean's point, when you buy a 300mbps circuit, you're paying for 300mbps of internet access. 

That does not mean that a network should (and in this case small-medium ones simply can't) build all of their capacity to service a large number of customer circuits at line rate at the same time for an extended period, ESPECIALLY to the exact same endpoint. It's just not economically reasonable to expect that. Remember we're talking about residential service here, not enterprise circuits.

Therefore, how do you prevent this spike of [insert large number here] gigabits traversing the network at the same time from causing issues? Build more network? That sounds easy, but there are plenty of legitimate reasons why ISPs can't or don't want to do that, particularly for an event that only occurs once per quarter or so.

Does Akamai bear some burden here to make these rollouts less troublesome for the ISPs they traverse through the last mile(s)? IMO yes, yes they do. When you're doing something new and unprecedented, as Akamai frequently brags about on Twitter, like having rapid, bursty growth of traffic, you need to consider that just because you can generate it, doesn't mean it can be delivered.  They've gotta be more sophisticated than a bunch of servers with SSD arrays, ramdisks, and 100 gig interfaces, so there's no excuse for them here to just blindly fill every link they have after sitting idle for weeks/months at a time and expect everything to come out alright and nobody to complain about it.

On Thu, Apr 1, 2021 at 1:21 PM Niels Bakker <niels=nanog@bakker.net> wrote:
* nanog@nanog.org (Jean St-Laurent via NANOG) [Thu 01 Apr 2021, 21:03 CEST]:
>An artificial roll out penalty somehow? Probably not at the ISP
>level, but more at the game level. Well, ISP could also have some
>mechanisms to reduce the impact or even Akamai could force a
>progressive roll out.

It's an online game. You can't play the game with outdated assets.
You'd not see walls where other players would, for example.

What you're suggesting is the ability of ISPs to market Internet access
at a certain speed but not have to deliver it based on conditions they
create.


        -- Niels.


--
Matt Erculiani
ERCUL-ARIN



--
Matt Erculiani
ERCUL-ARIN


--
Matt Erculiani
ERCUL-ARIN



--
Matt Erculiani
ERCUL-ARIN