| Commit message (Collapse) | Author | Age | Files | Lines |
| |
|
|
|
|
| |
or else they'll see an echo of their own messages after teleporting.
|
| |
|
| |
|
|
|
|
|
|
| |
join/drop appropriately, invitechatboxes.
The major departure from flotsam is to send only one message per destination region, as opposed to one message per group member. This reduces messaging considerably in large groups that have clusters of members in certain regions.
|
| |
|
|
|
|
| |
1sec). Group chat is going over the EQ... Hopefully this won't increase CPU when there's nothing going on, but we need to watch for that.
|
| |
|
|
|
|
| |
aren't processing it in any meaningful way, and it seems to get invoked everytime someone types a message in group chat.
|
|
|
|
| |
copy of the LocalSimulationConnector but never initializes it (hence ServiceVersion was never set)
|
|
|
|
|
| |
This is a test setup failure since code paths when adding a duplicate root scene presence now require the EntityTransferModule to be present.
Test fixed by adding this module to test setup
|
|\ |
|
| | |
|
| |\ |
|
| | |
| | |
| | |
| | | |
though we specified jpg extention it was actually a png and thus confusing the viewer silently.
|
| |/
|/|
| |
| | |
week's SIMULATOR/0.1 protocol for now.
|
|/
|
|
|
| |
These were genuine failures caused by ScenePresence.CompleteMovement() waiting for an UpdateAgent from NPC introduction that would never come.
Instead, we do not wait if the agent is an NPC.
|
| |
|
|
|
|
| |
because DisableSimulator is now being sent via UDP
|
|
|
|
|
|
|
|
|
| |
as needed. (Although sometimes Justin's state machine kicks in and doesn't let you) The EventQueues are a hairy mess, and it's very easy to mess things up. But it looks like this commit makes them work right. Here's what's going on:
- Child and root agents are only closed after 15 sec, maybe
- If the user comes back, they aren't closed, and everything is reused
- On the receiving side, clients and scene presences are reused if they already exist
- Caps are always recreated (this is where I spent most of my time!). It turns out that, because the agents carry the seeds around, the seed gets the same URL, except for the root agent coming back to a far away region, which gets a new seed (because we don't know what was its seed in the departing region, and we can't send it back to the client when the agent returns there).
|
|
|
|
|
|
|
|
|
| |
seconds before closing root agent after 15.
This is because a returning viewer by teleport before 15 seconds are up will be disrupted by the close.
The 2 second delay is within the scope where a normal viewer would not allow a teleport back anyway.
Simulator/0.2 (V2) protocol will continue with the longer delay since this is actually the behaviour viewers get from the ll grid
and an early close causes other issues (avatar being sent to infinite locations temporarily, etc.)
|
|
|
|
| |
DoNotClose to DoNotCloseAfterTeleport
|
|
|
|
|
|
|
|
| |
was a previous root agent, do not close that child agent at the end of the 15 sec teleport timer.
This prevents an issue if the user teleports back to the neighbour simulator of a source before 15 seconds have elapsed.
This more closely emulates observed linden behaviour, though the timeout there is 50 secs and applies to all the pre-teleport agents.
Currently sticks a DoNotClose flag on ScenePresence though this may be temporary as possibly it could be incorporated into the ETM state machine
|
|
|
|
| |
was always present, this seems to have fixed it.
|
| |
|
|
|
|
| |
on TP_V1 and adding a big red warning on top of those lines.
|
| |
|
|
|
|
| |
that eliminates the temporary placement at infinity upon TPs
|
| |
|
| |
|
|
|
|
| |
with high priority and hopefully gets to the client before AgentMovementComplete
|
| |
|
|
|
|
| |
besides ViaLogin.
|
| |
|
| |
|
|
|
|
|
|
|
|
| |
the destination is being checked)
In this new protocol, and as committed before, the viewer is not sent EnableSimulator/EstablishChildCommunication for the destination. Instead, it is sent TeleportFinish directly. TeleportFinish, in turn, makes the viewer send a UserCircuitCode packet followed by CompleteMovementIntoRegion packet. These 2 packets tend to occur one after the other almost immediately to the point that when CMIR arrives the client is not even connected yet and that packet is ignored (there might have been some race conditions here before); then the viewer sends CMIR again within 5-8 secs. But the delay between them may be higher in busier regions, which may lead to race conditions.
This commit improves the process so there are are no race conditions at the destination. CompleteMovement (triggered by the viewer) waits until Update has been sent from the origin. Update, in turn, waits until there is a *root* scene presence -- so making sure CompleteMovement has run MakeRoot. In other words, there are two threadlets at the destination, one from the viewer and one from the origin region, waiting for each other to do the right thing. That makes it safe to close the agent at the origin upon return of the Update call without having to wait for callback, because we are absolutely sure that the viewer knows it is in th new region.
Note also that in the V1 protocol, the destination was getting UseCircuitCode from the viewer twice -- once on EstablishAgentCommunication and then again on TeleportFinish. The second UCC was being ignored, but it shows how we were not following the expected steps...
|
|
|
|
|
|
| |
15sec before closing the agent. This seems to be working fairly well. The viewer seems to have an 8 sec delay between UseCircuitCode and CompleteMovement.
Also added back the position on UpdateAgent, because it's needed for TPing between neighboring regions.
|
| |
|
| |
|
|\ |
|
| |
| |
| |
| | |
what looks like a very significant reducing in GUN occurrances
|
| |\ |
|
| | |
| | |
| | |
| | | |
than it should have been (though internal use was correct)
|
| | |
| | |
| | |
| | | |
(Note: different from the defunct see_into_neighboring_sim, which used to control the process from the other end). This enables child agents in neighbors for which the root agent doesn't have permission to be in.
|
| |/
|/|
| |
| | |
nerves that it takes so long to let go of old info.
|
| | |
|
| | |
|
| |
| |
| |
| | |
folder. The viewer doesn't like that.
|
| | |
|
|/ |
|