Jump to content

Proxy server failure?


Lucia Nightfire
 Share

You are about to reply to a thread that has been inactive for 262 days.

Please take a moment to consider if this thread is worth bumping.

Recommended Posts

Not that precisely but about 15 minutes ago I was getting random errors, mostly script errors without consequences on Teeglepet horses when freshly rezzed.  It may not be connected but it IS unusual.

Some appeared to be the consequence of the item not being able to communicate with a central server, so maybe those ARE connected.

Nothing on the GSP yet...hardly a surprise though.

ETA: this was on the LeTigre RC channel with the new build number.

Edited by Aishagain
Link to comment
Share on other sites

On aws portal  - amazon backend having issues

Network Connectivity Issues [05:21 PM PDT] We continue to work toward resolving the increased networking latencies and errors affecting Availability Zones (usw2-az1 and usw2-az2) in the US-WEST-2 Region. We have successfully applied an update to the subsystem responsible for network mapping propagation to address resource contention. We have seen network mapping propagation times stabilize but they have not yet begun to trend towards normal levels. We expect that to begin over the next 30 minutes, at which time we expect latencies and error rates to improve. We will continue to keep you updated on our progress towards full recovery.

[04:20 PM PDT] We continue to progress toward resolving the increased networking latencies and errors affecting Availability Zones (usw2-az1 and usw2-az2) in the US-WEST-2 Region. At this time, we are approximately 50% completed with the update to address resource contention within the subsystem responsible for network mappings propagation in the usw2-az2 Availability Zone. Once we complete the update in usw2-az2, we will then move on to usw2-az1. Our current expectation is to have both Availability Zones fully resolved within the next 60 to 90 minutes, and we will continue to provide updates as recovery progresses.

[03:33 PM PDT] We continue to make progress towards resolving the increased networking latencies and errors affecting Availability Zones (usw2-az1 and usw2-az2) in the US-WEST-2 Region. In the last 30 minutes, we’ve continued applying an update to address resource contention within the subsystem responsible for network mappings propagation and are seeing early signs of improvement. We will continue to monitor before deploying this change more broadly and will continue to provide updates.

[03:02 PM PDT] We continue to make progress towards resolving the increased networking latencies and errors affecting Availability Zones (usw2-az1 and usw2-az2) in the US-WEST-2 Region. In the last hour, we applied an update to address resource contention within the subsystem responsible for network mappings propagation and are seeing early signs of improvement. We will continue to monitor before deploying this change more broadly and will continue to provide updates.

[02:22 PM PDT] We continue to make progress towards resolving the increased networking latencies and errors affecting Availability Zones (usw2-az1 and usw2-az2) in the US-WEST-2 Region. While we continue to make progress in addressing the issue, we wanted to provide some more details on the issue. Within Amazon Virtual Private Cloud (VPC) any changes to the network configuration - including launching an EC2 instance, attaching an Elastic IP address or Elastic Network Interface - needs to be propagated to the underlying hardware to ensure that network packets can flow between source and destination. We call this network configuration “network mappings”, as it contains information about network paths or mappings. Starting at 10:00 AM PDT this morning, we have been experiencing a delay in the propagation of these mappings within a single cell (part of the Availability Zone) in usw2-az1 and usw2-az2 Availability Zones. The root cause appears to be increased load to the subsystem responsible for the handling of these network mappings. We have been working to reduce the load on this service to improve propagation times, but while we have made some progress, mapping propagation latencies have not returned to normal levels. We continue to work to identify all forms of resource contention that could be leading to load, and have a few additional updates that we are currently working on.

Link to comment
Share on other sites

[05:56 PM PDT] We continue to work toward resolving the increased networking latencies and errors affecting Availability Zones (usw2-az1 and usw2-az2) in the US-WEST-2 Region. While network mapping propagation times have remained stable, we have not yet seen the improvement in propagation latencies that we had hoped for. In parallel, we are working on several other updates to address the resource contention within the subsystem responsible for network mapping propagation. We will continue to keep you updated on our progress towards full recovery.

Link to comment
Share on other sites

Reading some of the above posts, it seems likely that the issue is another AWS-related glitch.  As such we are unlikely to see any comment from LL via the GSP unless the problems increase markedly.

The current spate of AWS issues is concerning, to a wider userbase than just SL, so socks will need to be pulled up quickly at AWS unless they want some heavyweight complaints.

ETA:  I should know better!  It is now on the GSP.

Edited by Aishagain
Link to comment
Share on other sites

I miss April Linden. She always gave us REALLY good reports to keep us informed and with compassion. We need that now that AWS says their stuff is resolved. 

"Engineers worked to identify the root cause and resolve the resource contention affecting the specific subsystem. By 9:15 PM PDT, the propagation time for network mappings had returned to normal levels. The issue has been resolved, and the service is operating normally."

 

Do SL regions just need the Tuesday restart to be back on their feet? We're seeing inconsistency with world linden balance and web linden balances, too; we have no idea what is happening. 

some information and some clarification, perhaps just a touch of compassion, would be lovely from the lab :D

  • Thanks 1
Link to comment
Share on other sites

  • Lindens

Information is still coming in (*really* looking forward to AWS' explanation).  HTTP-Out was running at elevated levels (including higher error rates) from 21:30slt yesterday until 2:45slt today.  That's now running as expected.  Teleports remained unreliable (~80% successful) until around 6:30slt today.  They've now recovered.  Lingering issues are likely and we do want to hear about them.  Please contact support.

  • Like 2
  • Thanks 1
Link to comment
Share on other sites

You are about to reply to a thread that has been inactive for 262 days.

Please take a moment to consider if this thread is worth bumping.

Please sign in to comment

You will be able to leave a comment after signing in



Sign In Now
 Share

×
×
  • Create New...