10 #ifndef EIGEN_CXX11_TENSOR_TENSOR_DEVICE_DEFAULT_H
11 #define EIGEN_CXX11_TENSOR_TENSOR_DEVICE_DEFAULT_H
46 #ifdef EIGEN_GPU_COMPILE_PHASE
48 for (
T* it = begin; it !=
end; ++it) {
52 std::fill(begin,
end, value);
55 template<
typename Type>
61 #if !defined(EIGEN_GPU_COMPILE_PHASE)
64 #elif defined(EIGEN_HIP_DEVICE_COMPILE)
74 #if !defined(EIGEN_GPU_COMPILE_PHASE) && !defined(SYCL_DEVICE_ONLY)
77 #elif defined(EIGEN_HIP_DEVICE_COMPILE)
87 #if !defined(EIGEN_GPU_COMPILE_PHASE) && !defined(SYCL_DEVICE_ONLY)
90 #elif defined(EIGEN_HIP_DEVICE_COMPILE)
104 #if !defined(EIGEN_GPU_COMPILE_PHASE)
108 #elif defined(EIGEN_HIP_DEVICE_COMPILE)
114 return EIGEN_CUDA_ARCH / 100;
#define EIGEN_DEVICE_FUNC
static const lastp1_t end
void * aligned_malloc(std::size_t size)
void aligned_free(void *ptr)
: TensorContractionSycl.h, provides various tensor contraction kernel for SYCL backend
std::ptrdiff_t l1CacheSize()
std::ptrdiff_t l3CacheSize()
void * allocate_temp(size_t num_bytes) const
void deallocate_temp(void *buffer) const
void memcpyDeviceToHost(void *dst, const void *src, size_t n) const
void memcpyHostToDevice(void *dst, const void *src, size_t n) const
void memset(void *buffer, int c, size_t n) const
size_t lastLevelCacheSize() const
Type get(Type data) const
void deallocate(void *buffer) const
void memcpy(void *dst, const void *src, size_t n) const
size_t numThreads() const
int majorDeviceVersion() const
void * allocate(size_t num_bytes) const
size_t firstLevelCacheSize() const
void fill(T *begin, T *end, const T &value) const