Files
clang-p2996/openmp/libomptarget/plugins/amdgpu/impl/machine.cpp
Jon Chesterfield d0b312955f [libomptarget] Implement host plugin for amdgpu
[libomptarget] Implement host plugin for amdgpu

Replacement for D71384. Primary difference is inlining the dependency on atmi
followed by extensive simplification and bugfixes. This is the latest version
from https://github.com/ROCm-Developer-Tools/amd-llvm-project/tree/aomp12 with
minor patches and a rename from hsa to amdgpu, on the basis that this can't be
used by other implementations of hsa without additional work.

This will not build unless the ROCM_DIR variable is passed so won't break other
builds. That variable is used to locate two amdgpu specific libraries that ship
as part of rocm:
libhsakmt at https://github.com/RadeonOpenCompute/ROCT-Thunk-Interface
libhsa-runtime64 at https://github.com/RadeonOpenCompute/ROCR-Runtime
These libraries build from source. The build scripts in those repos are for
shared libraries, but can be adapted to statically link both into this plugin.

There are caveats.
- This works well enough to run various tests and benchmarks, and will be used
  to support the current clang bring up
- It is adequately thread safe for the above but there will be races remaining
- It is not stylistically correct for llvm, though has had clang-format run
- It has suboptimal memory management and locking strategies
- The debug printing / error handling is inconsistent

I would like to contribute this pretty much as-is and then improve it in-tree.
This would be advantagous because the aomp12 branch that was in use for fixing
this codebase has just been joined with the amd internal rocm dev process.

Reviewed By: jdoerfert

Differential Revision: https://reviews.llvm.org/D85742
2020-08-15 23:58:28 +01:00

129 lines
3.7 KiB
C++

/*===--------------------------------------------------------------------------
* ATMI (Asynchronous Task and Memory Interface)
*
* This file is distributed under the MIT License. See LICENSE.txt for details.
*===------------------------------------------------------------------------*/
#include "machine.h"
#include "atmi_runtime.h"
#include "internal.h"
#include <cassert>
#include <hsa.h>
#include <hsa_ext_amd.h>
#include <stdio.h>
#include <stdlib.h>
#include <vector>
extern ATLMachine g_atl_machine;
extern hsa_region_t atl_cpu_kernarg_region;
void *ATLMemory::alloc(size_t sz) {
void *ret;
hsa_status_t err = hsa_amd_memory_pool_allocate(memory_pool_, sz, 0, &ret);
ErrorCheck(Allocate from memory pool, err);
return ret;
}
void ATLMemory::free(void *ptr) {
hsa_status_t err = hsa_amd_memory_pool_free(ptr);
ErrorCheck(Allocate from memory pool, err);
}
void ATLProcessor::addMemory(const ATLMemory &mem) {
for (auto &mem_obj : memories_) {
// if the memory already exists, then just return
if (mem.memory().handle == mem_obj.memory().handle)
return;
}
memories_.push_back(mem);
}
const std::vector<ATLMemory> &ATLProcessor::memories() const {
return memories_;
}
template <> std::vector<ATLCPUProcessor> &ATLMachine::processors() {
return cpu_processors_;
}
template <> std::vector<ATLGPUProcessor> &ATLMachine::processors() {
return gpu_processors_;
}
hsa_amd_memory_pool_t get_memory_pool(const ATLProcessor &proc,
const int mem_id) {
hsa_amd_memory_pool_t pool;
const std::vector<ATLMemory> &mems = proc.memories();
assert(mems.size() && mem_id >= 0 && mem_id < mems.size() &&
"Invalid memory pools for this processor");
pool = mems[mem_id].memory();
return pool;
}
template <> void ATLMachine::addProcessor(const ATLCPUProcessor &p) {
cpu_processors_.push_back(p);
}
template <> void ATLMachine::addProcessor(const ATLGPUProcessor &p) {
gpu_processors_.push_back(p);
}
void callbackQueue(hsa_status_t status, hsa_queue_t *source, void *data) {
if (status != HSA_STATUS_SUCCESS) {
fprintf(stderr, "[%s:%d] GPU error in queue %p %d\n", __FILE__, __LINE__,
source, status);
abort();
}
}
void ATLGPUProcessor::createQueues(const int count) {
int *num_cus = reinterpret_cast<int *>(calloc(count, sizeof(int)));
hsa_status_t err;
/* Query the maximum size of the queue. */
uint32_t queue_size = 0;
err = hsa_agent_get_info(agent_, HSA_AGENT_INFO_QUEUE_MAX_SIZE, &queue_size);
ErrorCheck(Querying the agent maximum queue size, err);
if (queue_size > core::Runtime::getInstance().getMaxQueueSize()) {
queue_size = core::Runtime::getInstance().getMaxQueueSize();
}
/* Create queues for each device. */
int qid;
for (qid = 0; qid < count; qid++) {
hsa_queue_t *this_Q;
err =
hsa_queue_create(agent_, queue_size, HSA_QUEUE_TYPE_MULTI,
callbackQueue, NULL, UINT32_MAX, UINT32_MAX, &this_Q);
ErrorCheck(Creating the queue, err);
err = hsa_amd_profiling_set_profiler_enabled(this_Q, 1);
ErrorCheck(Enabling profiling support, err);
queues_.push_back(this_Q);
DEBUG_PRINT("Queue[%d]: %p\n", qid, this_Q);
}
free(num_cus);
}
void ATLCPUProcessor::createQueues(const int) {}
void ATLProcessor::destroyQueues() {
for (auto queue : queues_) {
hsa_status_t err = hsa_queue_destroy(queue);
ErrorCheck(Destroying the queue, err);
}
}
int ATLProcessor::num_cus() const {
hsa_status_t err;
/* Query the number of compute units. */
uint32_t num_cus = 0;
err = hsa_agent_get_info(
agent_, (hsa_agent_info_t)HSA_AMD_AGENT_INFO_COMPUTE_UNIT_COUNT,
&num_cus);
ErrorCheck(Querying the agent number of compute units, err);
return num_cus;
}