Skip to content

Commit

Permalink
Fix up syntax, remove unused comments
Browse files Browse the repository at this point in the history
  • Loading branch information
krishung5 committed Sep 21, 2023
1 parent 858a271 commit ebef5c9
Show file tree
Hide file tree
Showing 4 changed files with 1 addition and 9 deletions.
3 changes: 0 additions & 3 deletions src/memory_manager.h
Original file line number Diff line number Diff line change
Expand Up @@ -72,8 +72,6 @@ class BackendMemoryRecord : public MemoryRecord {
/// message queue asking the memory manager to deallocate the GPU tensor.
class MemoryManager {
public:
// MemoryManager(std::unique_ptr<MessageQueue<intptr_t>>&&
// memory_message_queue);
MemoryManager(
std::unique_ptr<SharedMemoryManager>& shm_pool,
std::unique_ptr<MessageQueue<bi::managed_external_buffer::handle_t>>&&
Expand All @@ -85,7 +83,6 @@ class MemoryManager {
private:
std::thread thread_;
std::unordered_map<intptr_t, std::unique_ptr<MemoryRecord>> records_;
// std::unique_ptr<MessageQueue<intptr_t>> message_queue_;
std::unique_ptr<MessageQueue<bi::managed_external_buffer::handle_t>>
message_queue_;
void QueueMonitorThread();
Expand Down
1 change: 0 additions & 1 deletion src/pb_stub.h
Original file line number Diff line number Diff line change
Expand Up @@ -365,7 +365,6 @@ class Stub {
stub_to_parent_mq_;
std::unique_ptr<MessageQueue<bi::managed_external_buffer::handle_t>>
parent_to_stub_mq_;
// std::unique_ptr<MessageQueue<uint64_t>> memory_manager_message_queue_;
std::unique_ptr<MessageQueue<bi::managed_external_buffer::handle_t>>
memory_manager_message_queue_;
bool initialized_;
Expand Down
2 changes: 1 addition & 1 deletion src/python_be.cc
Original file line number Diff line number Diff line change
Expand Up @@ -1578,7 +1578,7 @@ ModelInstanceState::ProcessRequests(

#ifdef TRITON_ENABLE_GPU
for (auto& output_tensor : infer_response->OutputTensors()) {
if ((output_tensor->MemoryType() == TRITONSERVER_MEMORY_GPU)) {
if (output_tensor->MemoryType() == TRITONSERVER_MEMORY_GPU) {
// Attempt to use the cuda shared memory pool for GPU tensor.
ShareCUDAMemoryPool(output_tensor->MemoryTypeId());
}
Expand Down
4 changes: 0 additions & 4 deletions src/stub_launcher.cc
Original file line number Diff line number Diff line change
Expand Up @@ -167,10 +167,6 @@ StubLauncher::Setup()

std::unique_ptr<MessageQueue<bi::managed_external_buffer::handle_t>>
memory_manager_message_queue;
// RETURN_IF_EXCEPTION(
// memory_manager_message_queue =
// MessageQueue<intptr_t>::Create(shm_pool_,
// shm_message_queue_size_));
RETURN_IF_EXCEPTION(
memory_manager_message_queue =
MessageQueue<bi::managed_external_buffer::handle_t>::Create(
Expand Down

0 comments on commit ebef5c9

Please sign in to comment.