!4928 revert gpu buffer

Merge pull request !4928 from qianlong21st/revert_gpu_queue
This commit is contained in:
mindspore-ci-bot 2020-08-21 20:06:59 +08:00 committed by Gitee
commit b7d9d5e0cf
4 changed files with 2 additions and 51 deletions

View File

@ -52,7 +52,6 @@ BlockQueueStatus_T GpuQueue::Push(const std::vector<DataItemGpu> &data) {
CHECK_CUDA_RET_WITH_ERROR(cudaEventCreate(&(*(node_info_[tail_].event_))), "Cuda Create Event Failed");
node_info_[tail_].data_ = data;
tail_ = (tail_ + 1) % (capacity_);
++size_;
return SUCCESS;
}
@ -70,7 +69,6 @@ BlockQueueStatus_T GpuQueue::Front(void **addr, size_t *len) const {
BlockQueueStatus_T GpuQueue::Pop() {
head_ = (head_ + 1) % (capacity_);
--size_;
return SUCCESS;
}

View File

@ -44,15 +44,13 @@ class GpuQueue {
void RegisterRelease(const std::function<void(void *)> &func) { host_release_ = func; }
inline bool IsEmpty() const { return size_ == 0; }
inline bool IsFull() const { return size_ == capacity_; }
inline bool IsEmpty() const { return head_ == tail_; }
inline bool IsFull() const { return head_ == ((tail_ + 1) % (capacity_)); }
BlockQueueStatus_T Push(const std::vector<DataItemGpu> &data);
BlockQueueStatus_T Front(void **ptr, size_t *len) const;
BlockQueueStatus_T Pop();
bool Destroy();
size_t Size() { return size_; }
size_t Capacity() { return capacity_; }
private:
struct NodeInfo {
@ -65,7 +63,6 @@ class GpuQueue {
size_t tail_;
std::vector<size_t> shape_;
size_t len_;
size_t size_;
size_t capacity_;
cudaStream_t stream_;
std::unique_ptr<NodeInfo[]> node_info_;
@ -86,8 +83,6 @@ class BlockingQueue {
BlockQueueStatus_T Front(void **ptr, size_t *len);
BlockQueueStatus_T Pop();
bool Destroy();
size_t Size() { return queue_->Size(); }
size_t Capacity() { return queue_->Capacity(); }
private:
std::mutex mutex_;

View File

@ -187,39 +187,5 @@ bool GpuBufferMgr::CloseNotify() {
}
void GpuBufferMgr::CloseConfirm() { sema.Signal(); }
size_t GpuBufferMgr::Size(unsigned int handle) {
if (handle == HandleMgr::INVALID_HANDLE) {
MS_LOG(ERROR) << "handle is invalid";
return 0;
}
return handle_queue_map_.at(handle)->Size();
}
size_t GpuBufferMgr::Size(unsigned int device_id, const std::string &channel_name) {
std::string name = std::to_string(device_id) + std::string("_") + channel_name;
if (!name_queue_map_.count(name)) {
MS_LOG(ERROR) << "Queue not exist " << name;
return 0;
}
return name_queue_map_.at(name)->Size();
}
size_t GpuBufferMgr::Capacity(unsigned int handle) {
if (handle == HandleMgr::INVALID_HANDLE) {
MS_LOG(ERROR) << "handle is invalid";
return 0;
}
return handle_queue_map_.at(handle)->Capacity();
}
size_t GpuBufferMgr::Capacity(unsigned int device_id, const std::string &channel_name) {
std::string name = std::to_string(device_id) + std::string("_") + channel_name;
if (!name_queue_map_.count(name)) {
MS_LOG(ERROR) << "Queue not exist " << name;
return 0;
}
return name_queue_map_.at(name)->Capacity();
}
} // namespace device
} // namespace mindspore

View File

@ -111,14 +111,6 @@ class GpuBufferMgr {
// call for dataset send thread
EXPORT void CloseConfirm();
EXPORT size_t Size(unsigned int handle);
EXPORT size_t Size(unsigned int device_id, const std::string &channel_name);
EXPORT size_t Capacity(unsigned int handle);
EXPORT size_t Capacity(unsigned int device_id, const std::string &channel_name);
private:
void set_device() const;