diff options
author | Tiger Wang <ziwei.tiger@outlook.com> | 2021-01-06 01:35:42 +0100 |
---|---|---|
committer | Tiger Wang <ziwei.tiger@outlook.com> | 2021-01-12 13:34:34 +0100 |
commit | 054a89dd9e5d6819adede9d7ba781b69f98ff2f4 (patch) | |
tree | 820a104cae2d9e4ec912b0a1bd7debac52cf4cc9 /src/ChunkSender.cpp | |
parent | Convert most calls to blocking GetHeight/GetBiomeAt to direct chunk accesses (diff) | |
download | cuberite-054a89dd9e5d6819adede9d7ba781b69f98ff2f4.tar cuberite-054a89dd9e5d6819adede9d7ba781b69f98ff2f4.tar.gz cuberite-054a89dd9e5d6819adede9d7ba781b69f98ff2f4.tar.bz2 cuberite-054a89dd9e5d6819adede9d7ba781b69f98ff2f4.tar.lz cuberite-054a89dd9e5d6819adede9d7ba781b69f98ff2f4.tar.xz cuberite-054a89dd9e5d6819adede9d7ba781b69f98ff2f4.tar.zst cuberite-054a89dd9e5d6819adede9d7ba781b69f98ff2f4.zip |
Diffstat (limited to '')
-rw-r--r-- | src/ChunkSender.cpp | 75 |
1 files changed, 38 insertions, 37 deletions
diff --git a/src/ChunkSender.cpp b/src/ChunkSender.cpp index c93a764b2..0a7f58bc7 100644 --- a/src/ChunkSender.cpp +++ b/src/ChunkSender.cpp @@ -104,13 +104,13 @@ void cChunkSender::QueueSendChunkTo(int a_ChunkX, int a_ChunkZ, Priority a_Prior m_SendChunks.push(sChunkQueue{a_Priority, Chunk}); info.m_Priority = a_Priority; } - info.m_Clients.insert(a_Client); + info.m_Clients.insert(a_Client->shared_from_this()); } else { m_SendChunks.push(sChunkQueue{a_Priority, Chunk}); auto info = sSendChunk{Chunk, a_Priority}; - info.m_Clients.insert(a_Client); + info.m_Clients.insert(a_Client->shared_from_this()); m_ChunkInfo.emplace(Chunk, info); } } @@ -135,13 +135,19 @@ void cChunkSender::QueueSendChunkTo(int a_ChunkX, int a_ChunkZ, Priority a_Prior m_SendChunks.push(sChunkQueue{a_Priority, Chunk}); info.m_Priority = a_Priority; } - info.m_Clients.insert(a_Clients.begin(), a_Clients.end()); + for (const auto & Client : a_Clients) + { + info.m_Clients.insert(Client->shared_from_this()); + } } else { m_SendChunks.push(sChunkQueue{a_Priority, Chunk}); auto info = sSendChunk{Chunk, a_Priority}; - info.m_Clients.insert(a_Clients.begin(), a_Clients.end()); + for (const auto & Client : a_Clients) + { + info.m_Clients.insert(Client->shared_from_this()); + } m_ChunkInfo.emplace(Chunk, info); } } @@ -152,24 +158,6 @@ void cChunkSender::QueueSendChunkTo(int a_ChunkX, int a_ChunkZ, Priority a_Prior -void cChunkSender::RemoveClient(cClientHandle * a_Client) -{ - { - cCSLock Lock(m_CS); - for (auto && pair : m_ChunkInfo) - { - auto && clients = pair.second.m_Clients; - clients.erase(a_Client); // nop for sets that do not contain a_Client - } - } - m_evtQueue.Set(); - m_evtRemoved.Wait(); // Wait for all remaining instances of a_Client to be processed (Execute() makes a copy of m_ChunkInfo) -} - - - - - void cChunkSender::Execute(void) { while (!m_ShouldTerminate) @@ -189,16 +177,13 @@ void cChunkSender::Execute(void) continue; } - std::unordered_set<cClientHandle *> clients; - std::swap(itr->second.m_Clients, clients); + auto clients = std::move(itr->second.m_Clients); m_ChunkInfo.erase(itr); cCSUnlock Unlock(Lock); SendChunk(Chunk.m_ChunkX, Chunk.m_ChunkZ, clients); } } - - m_evtRemoved.SetAll(); // Notify all waiting threads that all clients are processed and thus safe to destroy } // while (!m_ShouldTerminate) } @@ -206,21 +191,27 @@ void cChunkSender::Execute(void) -void cChunkSender::SendChunk(int a_ChunkX, int a_ChunkZ, std::unordered_set<cClientHandle *> a_Clients) +void cChunkSender::SendChunk(int a_ChunkX, int a_ChunkZ, const WeakClients & a_Clients) { + // Contains strong pointers to clienthandles. + std::vector<std::shared_ptr<cClientHandle>> Clients; + // Ask the client if it still wants the chunk: - for (auto itr = a_Clients.begin(); itr != a_Clients.end();) + for (const auto & WeakClient : a_Clients) { - if (!(*itr)->WantsSendChunk(a_ChunkX, a_ChunkZ)) + auto Client = WeakClient.lock(); + if ((Client != nullptr) && Client->WantsSendChunk(a_ChunkX, a_ChunkZ)) { - itr = a_Clients.erase(itr); - } - else - { - itr++; + Clients.push_back(std::move(Client)); } } + // Bail early if every requester disconnected: + if (Clients.empty()) + { + return; + } + // If the chunk has no clients, no need to packetize it: if (!m_World.HasChunkAnyClients(a_ChunkX, a_ChunkZ)) { @@ -247,9 +238,9 @@ void cChunkSender::SendChunk(int a_ChunkX, int a_ChunkZ, std::unordered_set<cCli } // Send: - m_Serializer.SendToClients(a_ChunkX, a_ChunkZ, m_Data, m_BiomeMap, a_Clients); + m_Serializer.SendToClients(a_ChunkX, a_ChunkZ, m_Data, m_BiomeMap, Clients); - for (const auto Client : a_Clients) + for (const auto & Client : Clients) { // Send block-entity packets: for (const auto & Pos : m_BlockEntities) @@ -270,7 +261,17 @@ void cChunkSender::SendChunk(int a_ChunkX, int a_ChunkZ, std::unordered_set<cCli Client->GetUsername().c_str() ); */ - a_Entity.SpawnOn(*Client); + + /* This check looks highly suspect. + Its purpose is to check the client still has a valid player object associated, + since the player destroys itself when the client is destroyed. + It's done within the world lock to ensure correctness. + A better way involves fixing chunk sending (GH #3696) to obviate calling SpawnOn from this thread in the first place. */ + if (!Client->IsDestroyed()) + { + a_Entity.SpawnOn(*Client); + } + return true; }); } |