mirror of
https://github.com/autc04/Retro68.git
synced 2024-12-01 11:52:47 +00:00
725 lines
19 KiB
C++
725 lines
19 KiB
C++
/*
|
|
Copyright (c) 2014-2016 Intel Corporation. All Rights Reserved.
|
|
|
|
Redistribution and use in source and binary forms, with or without
|
|
modification, are permitted provided that the following conditions
|
|
are met:
|
|
|
|
* Redistributions of source code must retain the above copyright
|
|
notice, this list of conditions and the following disclaimer.
|
|
* Redistributions in binary form must reproduce the above copyright
|
|
notice, this list of conditions and the following disclaimer in the
|
|
documentation and/or other materials provided with the distribution.
|
|
* Neither the name of Intel Corporation nor the names of its
|
|
contributors may be used to endorse or promote products derived
|
|
from this software without specific prior written permission.
|
|
|
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
|
|
#ifndef OFFLOAD_ENGINE_H_INCLUDED
|
|
#define OFFLOAD_ENGINE_H_INCLUDED
|
|
|
|
#include <limits.h>
|
|
#include <bitset>
|
|
#include <list>
|
|
#include <set>
|
|
#include <map>
|
|
#include "offload_common.h"
|
|
#include "coi/coi_client.h"
|
|
|
|
#define SIGNAL_HAS_COMPLETED ((OffloadDescriptor *)-1)
|
|
const int64_t no_stream = -1;
|
|
|
|
// Address range
|
|
class MemRange {
|
|
public:
|
|
MemRange() : m_start(0), m_length(0) {}
|
|
MemRange(const void *addr, uint64_t len) : m_start(addr), m_length(len) {}
|
|
|
|
const void* start() const {
|
|
return m_start;
|
|
}
|
|
|
|
const void* end() const {
|
|
return static_cast<const char*>(m_start) + m_length;
|
|
}
|
|
|
|
uint64_t length() const {
|
|
return m_length;
|
|
}
|
|
|
|
// returns true if given range overlaps with another one
|
|
bool overlaps(const MemRange &o) const {
|
|
// Two address ranges A[start, end) and B[start,end) overlap
|
|
// if A.start < B.end and A.end > B.start.
|
|
return start() < o.end() && end() > o.start();
|
|
}
|
|
|
|
// returns true if given range contains the other range
|
|
bool contains(const MemRange &o) const {
|
|
return start() <= o.start() && o.end() <= end();
|
|
}
|
|
|
|
private:
|
|
const void* m_start;
|
|
uint64_t m_length;
|
|
};
|
|
|
|
// Data associated with a pointer variable
|
|
class PtrData {
|
|
public:
|
|
PtrData(const void *addr, uint64_t len) :
|
|
cpu_addr(addr, len), cpu_buf(0),
|
|
mic_addr(0), alloc_disp(0), mic_buf(0), mic_offset(0),
|
|
ref_count(0), is_static(false), is_omp_associate(false)
|
|
{}
|
|
|
|
//
|
|
// Copy constructor
|
|
//
|
|
PtrData(const PtrData& ptr):
|
|
cpu_addr(ptr.cpu_addr), cpu_buf(ptr.cpu_buf),
|
|
mic_addr(ptr.mic_addr), alloc_disp(ptr.alloc_disp),
|
|
mic_buf(ptr.mic_buf), mic_offset(ptr.mic_offset),
|
|
ref_count(ptr.ref_count), is_static(ptr.is_static),
|
|
is_omp_associate(ptr.is_omp_associate),
|
|
var_alloc_type(0)
|
|
{}
|
|
|
|
bool operator<(const PtrData &o) const {
|
|
// Variables are sorted by the CPU start address.
|
|
// Overlapping memory ranges are considered equal.
|
|
return (cpu_addr.start() < o.cpu_addr.start()) &&
|
|
!cpu_addr.overlaps(o.cpu_addr);
|
|
}
|
|
|
|
long add_reference() {
|
|
if (is_omp_associate || (is_static && !var_alloc_type)) {
|
|
return LONG_MAX;
|
|
}
|
|
#ifndef TARGET_WINNT
|
|
return __sync_fetch_and_add(&ref_count, 1);
|
|
#else // TARGET_WINNT
|
|
return _InterlockedIncrement(&ref_count) - 1;
|
|
#endif // TARGET_WINNT
|
|
}
|
|
|
|
long remove_reference() {
|
|
if (is_omp_associate || (is_static && !var_alloc_type)) {
|
|
return LONG_MAX;
|
|
}
|
|
#ifndef TARGET_WINNT
|
|
return __sync_sub_and_fetch(&ref_count, 1);
|
|
#else // TARGET_WINNT
|
|
return _InterlockedDecrement(&ref_count);
|
|
#endif // TARGET_WINNT
|
|
}
|
|
|
|
long get_reference() const {
|
|
if (is_omp_associate || (is_static && !var_alloc_type)) {
|
|
return LONG_MAX;
|
|
}
|
|
return ref_count;
|
|
}
|
|
|
|
public:
|
|
// CPU address range
|
|
const MemRange cpu_addr;
|
|
|
|
// CPU and MIC buffers
|
|
COIBUFFER cpu_buf;
|
|
COIBUFFER mic_buf;
|
|
|
|
// placeholder for buffer address on mic
|
|
uint64_t mic_addr;
|
|
|
|
uint64_t alloc_disp;
|
|
|
|
// additional offset to pointer data on MIC for improving bandwidth for
|
|
// data which is not 4K aligned
|
|
uint32_t mic_offset;
|
|
|
|
// if true buffers are created from static memory
|
|
bool is_static;
|
|
|
|
// true if MIC buffer created by omp_target_associate
|
|
bool is_omp_associate;
|
|
|
|
bool var_alloc_type;
|
|
mutex_t alloc_ptr_data_lock;
|
|
|
|
private:
|
|
// reference count for the entry
|
|
long ref_count;
|
|
};
|
|
|
|
typedef std::list<PtrData*> PtrDataList;
|
|
|
|
class PtrDataTable {
|
|
public:
|
|
typedef std::set<PtrData> PtrSet;
|
|
|
|
PtrData* find_ptr_data(const void *ptr) {
|
|
m_ptr_lock.lock();
|
|
PtrSet::iterator res = list.find(PtrData(ptr, 0));
|
|
|
|
m_ptr_lock.unlock();
|
|
if (res == list.end()) {
|
|
return 0;
|
|
}
|
|
return const_cast<PtrData*>(res.operator->());
|
|
}
|
|
|
|
PtrData* insert_ptr_data(const void *ptr, uint64_t len, bool &is_new) {
|
|
m_ptr_lock.lock();
|
|
std::pair<PtrSet::iterator, bool> res =
|
|
list.insert(PtrData(ptr, len));
|
|
|
|
PtrData* ptr_data = const_cast<PtrData*>(res.first.operator->());
|
|
m_ptr_lock.unlock();
|
|
|
|
is_new = res.second;
|
|
if (is_new) {
|
|
// It's necessary to lock as soon as possible.
|
|
// unlock must be done at call site of insert_ptr_data at
|
|
// branch for is_new
|
|
ptr_data->alloc_ptr_data_lock.lock();
|
|
}
|
|
return ptr_data;
|
|
}
|
|
|
|
void remove_ptr_data(const void *ptr) {
|
|
m_ptr_lock.lock();
|
|
list.erase(PtrData(ptr, 0));
|
|
m_ptr_lock.unlock();
|
|
}
|
|
private:
|
|
|
|
PtrSet list;
|
|
mutex_t m_ptr_lock;
|
|
};
|
|
|
|
// Data associated with automatic variable
|
|
class AutoData {
|
|
public:
|
|
AutoData(const void *addr, uint64_t len) :
|
|
cpu_addr(addr, len), ref_count(0)
|
|
{}
|
|
|
|
bool operator<(const AutoData &o) const {
|
|
// Variables are sorted by the CPU start address.
|
|
// Overlapping memory ranges are considered equal.
|
|
return (cpu_addr.start() < o.cpu_addr.start()) &&
|
|
!cpu_addr.overlaps(o.cpu_addr);
|
|
}
|
|
|
|
long add_reference() {
|
|
#ifndef TARGET_WINNT
|
|
return __sync_fetch_and_add(&ref_count, 1);
|
|
#else // TARGET_WINNT
|
|
return _InterlockedIncrement(&ref_count) - 1;
|
|
#endif // TARGET_WINNT
|
|
}
|
|
|
|
long remove_reference() {
|
|
#ifndef TARGET_WINNT
|
|
return __sync_sub_and_fetch(&ref_count, 1);
|
|
#else // TARGET_WINNT
|
|
return _InterlockedDecrement(&ref_count);
|
|
#endif // TARGET_WINNT
|
|
}
|
|
|
|
long nullify_reference() {
|
|
#ifndef TARGET_WINNT
|
|
return __sync_lock_test_and_set(&ref_count, 0);
|
|
#else // TARGET_WINNT
|
|
return _InterlockedExchange(&ref_count,0);
|
|
#endif // TARGET_WINNT
|
|
}
|
|
|
|
long get_reference() const {
|
|
return ref_count;
|
|
}
|
|
|
|
public:
|
|
// CPU address range
|
|
const MemRange cpu_addr;
|
|
|
|
private:
|
|
// reference count for the entry
|
|
long ref_count;
|
|
};
|
|
|
|
// Set of autimatic variables
|
|
typedef std::set<AutoData> AutoSet;
|
|
|
|
// Target image data
|
|
struct TargetImage
|
|
{
|
|
TargetImage(const char *_name, const void *_data, uint64_t _size,
|
|
const char *_origin, uint64_t _offset) :
|
|
name(_name), data(_data), size(_size),
|
|
origin(_origin), offset(_offset)
|
|
{}
|
|
|
|
// library name
|
|
const char* name;
|
|
|
|
// contents and size
|
|
const void* data;
|
|
uint64_t size;
|
|
|
|
// file of origin and offset within that file
|
|
const char* origin;
|
|
uint64_t offset;
|
|
};
|
|
|
|
typedef std::list<TargetImage> TargetImageList;
|
|
|
|
// dynamic library and Image associated with lib
|
|
struct DynLib
|
|
{
|
|
DynLib(const char *_name, const void *_data,
|
|
COILIBRARY _lib) :
|
|
name(_name), data(_data), lib(_lib)
|
|
{}
|
|
// library name
|
|
const char* name;
|
|
|
|
// contents
|
|
const void* data;
|
|
|
|
COILIBRARY lib;
|
|
};
|
|
typedef std::list<DynLib> DynLibList;
|
|
|
|
// Data associated with persistent auto objects
|
|
struct PersistData
|
|
{
|
|
PersistData(const void *addr, uint64_t routine_num,
|
|
uint64_t size, uint64_t thread) :
|
|
stack_cpu_addr(addr), routine_id(routine_num), thread_id(thread)
|
|
{
|
|
stack_ptr_data = new PtrData(0, size);
|
|
}
|
|
// 1-st key value - beginning of the stack at CPU
|
|
const void * stack_cpu_addr;
|
|
// 2-nd key value - identifier of routine invocation at CPU
|
|
uint64_t routine_id;
|
|
// 3-rd key value - thread identifier
|
|
uint64_t thread_id;
|
|
|
|
// corresponded PtrData; only stack_ptr_data->mic_buf is used
|
|
PtrData * stack_ptr_data;
|
|
// used to get offset of the variable in stack buffer
|
|
char * cpu_stack_addr;
|
|
};
|
|
|
|
typedef std::list<PersistData> PersistDataList;
|
|
|
|
// Data associated with stream
|
|
struct Stream
|
|
{
|
|
Stream(int device, int num_of_cpus) :
|
|
m_number_of_cpus(num_of_cpus), m_pipeline(0), m_last_offload(0),
|
|
m_device(device)
|
|
{}
|
|
~Stream() {
|
|
if (m_pipeline) {
|
|
COI::PipelineDestroy(m_pipeline);
|
|
}
|
|
}
|
|
|
|
COIPIPELINE get_pipeline(void) {
|
|
return(m_pipeline);
|
|
}
|
|
|
|
int get_device(void) {
|
|
return(m_device);
|
|
}
|
|
|
|
int get_cpu_number(void) {
|
|
return(m_number_of_cpus);
|
|
}
|
|
|
|
void set_pipeline(COIPIPELINE pipeline) {
|
|
m_pipeline = pipeline;
|
|
}
|
|
|
|
OffloadDescriptor* get_last_offload(void) {
|
|
return(m_last_offload);
|
|
}
|
|
|
|
void set_last_offload(OffloadDescriptor* last_offload) {
|
|
m_last_offload = last_offload;
|
|
}
|
|
|
|
static Stream* find_stream(uint64_t handle, bool remove);
|
|
|
|
static _Offload_stream add_stream(int device, int number_of_cpus) {
|
|
_Offload_stream result;
|
|
m_stream_lock.lock();
|
|
result = ++m_streams_count;
|
|
all_streams[m_streams_count] = new Stream(device, number_of_cpus);
|
|
m_stream_lock.unlock();
|
|
return(result);
|
|
}
|
|
|
|
static uint64_t get_streams_count() {
|
|
return m_streams_count;
|
|
}
|
|
|
|
typedef std::map<uint64_t, Stream*> StreamMap;
|
|
|
|
static uint64_t m_streams_count;
|
|
static StreamMap all_streams;
|
|
static mutex_t m_stream_lock;
|
|
|
|
int m_device;
|
|
|
|
// number of cpus
|
|
int m_number_of_cpus;
|
|
|
|
// The pipeline associated with the stream
|
|
COIPIPELINE m_pipeline;
|
|
|
|
// The last offload occured via the stream
|
|
OffloadDescriptor* m_last_offload;
|
|
|
|
// Cpus used by the stream
|
|
std::bitset<COI_MAX_HW_THREADS> m_stream_cpus;
|
|
};
|
|
|
|
typedef std::map<uint64_t, Stream*> StreamMap;
|
|
typedef std::bitset<COI_MAX_HW_THREADS> micLcpuMask;
|
|
|
|
// ordered by count double linked list of cpus used by streams
|
|
typedef struct CpuEl{
|
|
uint64_t count; // number of streams using the cpu
|
|
struct CpuEl* prev; // cpu with the same or lesser count
|
|
struct CpuEl* next; // cpu with the same or greater count
|
|
} CpuEl;
|
|
|
|
// class representing a single engine
|
|
struct Engine {
|
|
friend void __offload_init_library_once(void);
|
|
friend void __offload_fini_library(void);
|
|
|
|
#define CPU_INDEX(x) (x - m_cpus)
|
|
#define check_result(res, tag, ...) \
|
|
{ \
|
|
if (res == COI_PROCESS_DIED) { \
|
|
fini_process(true); \
|
|
exit(1); \
|
|
} \
|
|
if (res != COI_SUCCESS) { \
|
|
__liboffload_error_support(tag, __VA_ARGS__); \
|
|
exit(1); \
|
|
} \
|
|
}
|
|
|
|
int get_logical_index() const {
|
|
return m_index;
|
|
}
|
|
|
|
int get_physical_index() const {
|
|
return m_physical_index;
|
|
}
|
|
|
|
const COIPROCESS& get_process() const {
|
|
return m_process;
|
|
}
|
|
|
|
bool get_ready() {
|
|
return m_ready;
|
|
}
|
|
|
|
uint64_t get_thread_id(void);
|
|
|
|
// initialize device
|
|
void init(void);
|
|
|
|
// unload library
|
|
void unload_library(const void *data, const char *name);
|
|
|
|
// add new library
|
|
void add_lib(const TargetImage &lib)
|
|
{
|
|
m_lock.lock();
|
|
m_ready = false;
|
|
m_images.push_back(lib);
|
|
m_lock.unlock();
|
|
}
|
|
|
|
COIRESULT compute(
|
|
_Offload_stream stream,
|
|
const std::list<COIBUFFER> &buffers,
|
|
const void* data,
|
|
uint16_t data_size,
|
|
void* ret,
|
|
uint16_t ret_size,
|
|
uint32_t num_deps,
|
|
const COIEVENT* deps,
|
|
COIEVENT* event
|
|
);
|
|
|
|
#ifdef MYO_SUPPORT
|
|
// temporary workaround for blocking behavior for myoiLibInit/Fini calls
|
|
void init_myo(COIEVENT *event) {
|
|
COIRESULT res;
|
|
res = COI::PipelineRunFunction(get_pipeline(),
|
|
m_funcs[c_func_myo_init],
|
|
0, 0, 0, 0, 0, 0, 0, 0, 0,
|
|
event);
|
|
check_result(res, c_pipeline_run_func, m_index, res);
|
|
}
|
|
|
|
void fini_myo(COIEVENT *event) {
|
|
COIRESULT res;
|
|
res = COI::PipelineRunFunction(get_pipeline(),
|
|
m_funcs[c_func_myo_fini],
|
|
0, 0, 0, 0, 0, 0, 0, 0, 0,
|
|
event);
|
|
check_result(res, c_pipeline_run_func, m_index, res);
|
|
}
|
|
#endif // MYO_SUPPORT
|
|
|
|
//
|
|
// Memory association table
|
|
//
|
|
PtrData* find_ptr_data(const void *ptr) {
|
|
return m_ptr_set.find_ptr_data(ptr);
|
|
}
|
|
|
|
PtrData* find_targetptr_data(const void *ptr) {
|
|
return m_targetptr_set.find_ptr_data(ptr);
|
|
}
|
|
|
|
PtrData* insert_ptr_data(const void *ptr, uint64_t len, bool &is_new) {
|
|
return m_ptr_set.insert_ptr_data(ptr, len, is_new);
|
|
}
|
|
|
|
PtrData* insert_targetptr_data(const void *ptr, uint64_t len,
|
|
bool &is_new) {
|
|
return m_targetptr_set.insert_ptr_data(ptr, len, is_new);
|
|
}
|
|
|
|
void remove_ptr_data(const void *ptr) {
|
|
m_ptr_set.remove_ptr_data(ptr);
|
|
}
|
|
|
|
void remove_targetptr_data(const void *ptr) {
|
|
m_targetptr_set.remove_ptr_data(ptr);
|
|
}
|
|
|
|
//
|
|
// Automatic variables
|
|
//
|
|
AutoData* find_auto_data(const void *ptr) {
|
|
AutoSet &auto_vars = get_auto_vars();
|
|
AutoSet::iterator res = auto_vars.find(AutoData(ptr, 0));
|
|
if (res == auto_vars.end()) {
|
|
return 0;
|
|
}
|
|
return const_cast<AutoData*>(res.operator->());
|
|
}
|
|
|
|
AutoData* insert_auto_data(const void *ptr, uint64_t len) {
|
|
AutoSet &auto_vars = get_auto_vars();
|
|
std::pair<AutoSet::iterator, bool> res =
|
|
auto_vars.insert(AutoData(ptr, len));
|
|
return const_cast<AutoData*>(res.first.operator->());
|
|
}
|
|
|
|
void remove_auto_data(const void *ptr) {
|
|
get_auto_vars().erase(AutoData(ptr, 0));
|
|
}
|
|
|
|
//
|
|
// Signals
|
|
//
|
|
void add_signal(const void *signal, OffloadDescriptor *desc) {
|
|
m_signal_lock.lock();
|
|
m_signal_map[signal] = desc;
|
|
m_signal_lock.unlock();
|
|
}
|
|
|
|
OffloadDescriptor* find_signal(const void *signal, bool remove) {
|
|
OffloadDescriptor *desc = 0;
|
|
|
|
m_signal_lock.lock();
|
|
{
|
|
SignalMap::iterator it = m_signal_map.find(signal);
|
|
if (it != m_signal_map.end()) {
|
|
desc = it->second;
|
|
if (remove) {
|
|
it->second = SIGNAL_HAS_COMPLETED;
|
|
}
|
|
}
|
|
}
|
|
m_signal_lock.unlock();
|
|
|
|
return desc;
|
|
}
|
|
|
|
void complete_signaled_ofld(const void *signal) {
|
|
|
|
m_signal_lock.lock();
|
|
{
|
|
SignalMap::iterator it = m_signal_map.find(signal);
|
|
if (it != m_signal_map.end()) {
|
|
it->second = SIGNAL_HAS_COMPLETED;
|
|
}
|
|
}
|
|
m_signal_lock.unlock();
|
|
}
|
|
|
|
void stream_destroy(_Offload_stream handle);
|
|
|
|
void move_cpu_el_after(CpuEl* cpu_what, CpuEl* cpu_after);
|
|
void print_stream_cpu_list(const char *);
|
|
|
|
COIPIPELINE get_pipeline(_Offload_stream stream);
|
|
|
|
StreamMap get_stream_map() {
|
|
return m_stream_map;
|
|
}
|
|
|
|
// stop device process
|
|
void fini_process(bool verbose);
|
|
|
|
// list of stacks active at the engine
|
|
PersistDataList m_persist_list;
|
|
|
|
private:
|
|
Engine() : m_index(-1), m_physical_index(-1), m_process(0), m_ready(false),
|
|
m_proc_number(0), m_assigned_cpus(0), m_cpus(0), m_cpu_head(0)
|
|
{}
|
|
|
|
~Engine() {
|
|
m_ready = false;
|
|
for (StreamMap::iterator it = m_stream_map.begin();
|
|
it != m_stream_map.end(); it++) {
|
|
Stream * stream = it->second;
|
|
delete stream;
|
|
}
|
|
if (m_process != 0) {
|
|
fini_process(false);
|
|
}
|
|
if (m_assigned_cpus) {
|
|
delete m_assigned_cpus;
|
|
}
|
|
}
|
|
|
|
// set indexes
|
|
void set_indexes(int logical_index, int physical_index) {
|
|
m_index = logical_index;
|
|
m_physical_index = physical_index;
|
|
}
|
|
|
|
// set CPU mask
|
|
void set_cpu_mask(micLcpuMask *cpu_mask)
|
|
{
|
|
m_assigned_cpus = cpu_mask;
|
|
}
|
|
|
|
// start process on device
|
|
void init_process();
|
|
|
|
void load_libraries(void);
|
|
void init_ptr_data(void);
|
|
|
|
// performs library intialization on the device side
|
|
pid_t init_device(void);
|
|
|
|
private:
|
|
// get pipeline associated with a calling thread
|
|
COIPIPELINE get_pipeline(void);
|
|
|
|
// get automatic vars set associated with the calling thread
|
|
AutoSet& get_auto_vars(void);
|
|
|
|
// destructor for thread data
|
|
static void destroy_thread_data(void *data);
|
|
|
|
private:
|
|
typedef std::set<PtrData> PtrSet;
|
|
typedef std::map<const void*, OffloadDescriptor*> SignalMap;
|
|
|
|
// device indexes
|
|
int m_index;
|
|
int m_physical_index;
|
|
|
|
// cpu mask
|
|
micLcpuMask *m_assigned_cpus;
|
|
|
|
// number of COI pipes created for the engine
|
|
long m_proc_number;
|
|
|
|
// process handle
|
|
COIPROCESS m_process;
|
|
|
|
// If false, device either has not been initialized or new libraries
|
|
// have been added.
|
|
bool m_ready;
|
|
mutex_t m_lock;
|
|
|
|
// List of libraries to be loaded
|
|
TargetImageList m_images;
|
|
|
|
// var tables
|
|
PtrDataTable m_ptr_set;
|
|
PtrDataTable m_targetptr_set;
|
|
|
|
// signals
|
|
SignalMap m_signal_map;
|
|
mutex_t m_signal_lock;
|
|
|
|
// streams
|
|
StreamMap m_stream_map;
|
|
mutex_t m_stream_lock;
|
|
int m_num_cores;
|
|
int m_num_threads;
|
|
CpuEl* m_cpus;
|
|
CpuEl* m_cpu_head;
|
|
|
|
// List of dynamic libraries to be registred
|
|
DynLibList m_dyn_libs;
|
|
|
|
// constants for accessing device function handles
|
|
enum {
|
|
c_func_compute = 0,
|
|
#ifdef MYO_SUPPORT
|
|
c_func_myo_init,
|
|
c_func_myo_fini,
|
|
#endif // MYO_SUPPORT
|
|
c_func_init,
|
|
c_func_var_table_size,
|
|
c_func_var_table_copy,
|
|
c_func_set_stream_affinity,
|
|
c_funcs_total
|
|
};
|
|
static const char* m_func_names[c_funcs_total];
|
|
|
|
// device function handles
|
|
COIFUNCTION m_funcs[c_funcs_total];
|
|
|
|
// int -> name mapping for device signals
|
|
static const int c_signal_max = 32;
|
|
static const char* c_signal_names[c_signal_max];
|
|
};
|
|
|
|
#endif // OFFLOAD_ENGINE_H_INCLUDED
|