@@ -347,7 +347,7 @@ static enum ggml_status
347
347
ggml_backend_sycl_buffer_init_tensor (ggml_backend_buffer_t buffer,
348
348
ggml_tensor *tensor) try {
349
349
GGML_SYCL_DEBUG (" [SYCL] call %s" , __func__);
350
- debug_print_tensor ( " : tensor=" , tensor, " \n " );
350
+ GGML_SYCL_DEBUG ( " %s " , debug_get_tensor_str ( " : tensor=" , tensor, " \n " ). c_str () );
351
351
ggml_backend_sycl_buffer_context * ctx = (ggml_backend_sycl_buffer_context *)buffer->context ;
352
352
353
353
if (tensor->view_src != NULL ) {
@@ -384,7 +384,7 @@ static void ggml_backend_sycl_buffer_set_tensor(ggml_backend_buffer_t buffer,
384
384
const void *data, size_t offset,
385
385
size_t size) try {
386
386
GGML_SYCL_DEBUG (" [SYCL] call %s" , __func__);
387
- debug_print_tensor ( " : tensor=" , tensor);
387
+ GGML_SYCL_DEBUG ( " %s " , debug_get_tensor_str ( " : tensor=" , tensor). c_str () );
388
388
GGML_SYCL_DEBUG (" size=%zu offset=%zu\n " , size, offset);
389
389
ggml_backend_sycl_buffer_context * ctx = ( ggml_backend_sycl_buffer_context *)buffer->context ;
390
390
ggml_sycl_set_device (ctx->device );
@@ -412,7 +412,7 @@ static void ggml_backend_sycl_buffer_get_tensor(ggml_backend_buffer_t buffer,
412
412
void *data, size_t offset,
413
413
size_t size) try {
414
414
GGML_SYCL_DEBUG (" [SYCL] call %s" , __func__);
415
- debug_print_tensor ( " : tensor=" , tensor);
415
+ GGML_SYCL_DEBUG ( " %s " , debug_get_tensor_str ( " : tensor=" , tensor). c_str () );
416
416
GGML_SYCL_DEBUG (" size=%zu offset=%zu\n " , size, offset);
417
417
ggml_backend_sycl_buffer_context * ctx = ( ggml_backend_sycl_buffer_context *)buffer->context ;
418
418
@@ -443,8 +443,8 @@ ggml_backend_sycl_buffer_cpy_tensor(ggml_backend_buffer_t buffer,
443
443
ggml_tensor *dst) try {
444
444
bool is_cpy_supported = ggml_backend_buffer_is_sycl (src->buffer );
445
445
GGML_SYCL_DEBUG (" [SYCL] call %s" , __func__);
446
- debug_print_tensor ( " : dst=" , dst);
447
- debug_print_tensor ( " src=" , src);
446
+ GGML_SYCL_DEBUG ( " %s " , debug_get_tensor_str ( " : dst=" , dst). c_str () );
447
+ GGML_SYCL_DEBUG ( " %s " , debug_get_tensor_str ( " src=" , src). c_str () );
448
448
GGML_SYCL_DEBUG (" is_cpy_supported=%d\n " , is_cpy_supported);
449
449
if (is_cpy_supported) {
450
450
ggml_backend_sycl_buffer_context * src_ctx = (ggml_backend_sycl_buffer_context *)src->buffer ->context ;
@@ -524,7 +524,7 @@ catch (sycl::exception const &exc) {
524
524
static void ggml_backend_sycl_buffer_memset_tensor (ggml_backend_buffer_t buffer, ggml_tensor * tensor, uint8_t value,
525
525
size_t offset, size_t size) {
526
526
GGML_SYCL_DEBUG (" [SYCL] call %s" , __func__);
527
- debug_print_tensor ( " : tensor=" , tensor);
527
+ GGML_SYCL_DEBUG ( " %s " , debug_get_tensor_str ( " : tensor=" , tensor). c_str () );
528
528
GGML_SYCL_DEBUG (" size=%zu offset=%zu value=%u\n " , size, offset, value);
529
529
ggml_backend_sycl_buffer_context * ctx = (ggml_backend_sycl_buffer_context *) buffer->context ;
530
530
SYCL_CHECK (ggml_sycl_set_device (ctx->device ));
@@ -804,7 +804,7 @@ static enum ggml_status
804
804
ggml_backend_sycl_split_buffer_init_tensor (ggml_backend_buffer_t buffer,
805
805
ggml_tensor *tensor) try {
806
806
GGML_SYCL_DEBUG (" [SYCL] call %s" , __func__);
807
- debug_print_tensor ( " : tensor=" , tensor, " \n " );
807
+ GGML_SYCL_DEBUG ( " %s " , debug_get_tensor_str ( " : tensor=" , tensor, " \n " ). c_str () );
808
808
GGML_ASSERT (tensor->view_src == nullptr ); // views of split tensors are not supported
809
809
810
810
ggml_backend_sycl_split_buffer_context * ctx = (ggml_backend_sycl_split_buffer_context *)buffer->context ;
@@ -890,7 +890,7 @@ ggml_backend_sycl_split_buffer_set_tensor(ggml_backend_buffer_t buffer,
890
890
ggml_tensor *tensor, const void *data,
891
891
size_t offset, size_t size) try {
892
892
GGML_SYCL_DEBUG (" [SYCL] call %s" , __func__);
893
- debug_print_tensor ( " : tensor=" , tensor);
893
+ GGML_SYCL_DEBUG ( " %s " , debug_get_tensor_str ( " : tensor=" , tensor). c_str () );
894
894
GGML_SYCL_DEBUG (" size=%zu offset=%zu\n " , size, offset);
895
895
// split tensors must always be set in their entirety at once
896
896
GGML_ASSERT (offset == 0 );
@@ -946,7 +946,7 @@ ggml_backend_sycl_split_buffer_get_tensor(ggml_backend_buffer_t buffer,
946
946
const ggml_tensor *tensor, void *data,
947
947
size_t offset, size_t size) try {
948
948
GGML_SYCL_DEBUG (" [SYCL] call %s" , __func__);
949
- debug_print_tensor ( " : tensor=" , tensor);
949
+ GGML_SYCL_DEBUG ( " %s " , debug_get_tensor_str ( " : tensor=" , tensor). c_str () );
950
950
GGML_SYCL_DEBUG (" size=%zu offset=%zu\n " , size, offset);
951
951
// split tensors must always be set in their entirety at once
952
952
GGML_ASSERT (offset == 0 );
@@ -3816,7 +3816,7 @@ static void ggml_backend_sycl_set_tensor_async(ggml_backend_t backend,
3816
3816
const void *data, size_t offset,
3817
3817
size_t size) try {
3818
3818
GGML_SYCL_DEBUG (" [SYCL] call %s" , __func__);
3819
- debug_print_tensor ( " : tensor=" , tensor);
3819
+ GGML_SYCL_DEBUG ( " %s " , debug_get_tensor_str ( " : tensor=" , tensor). c_str () );
3820
3820
GGML_SYCL_DEBUG (" size=%zu offset=%zu\n " , size, offset);
3821
3821
ggml_backend_sycl_context * sycl_ctx = (ggml_backend_sycl_context *)backend->context ;
3822
3822
ggml_backend_buffer_t buf = tensor->view_src ? tensor->view_src ->buffer : tensor->buffer ;
@@ -3837,7 +3837,7 @@ static void ggml_backend_sycl_get_tensor_async(ggml_backend_t backend,
3837
3837
void *data, size_t offset,
3838
3838
size_t size) try {
3839
3839
GGML_SYCL_DEBUG (" [SYCL] call %s" , __func__);
3840
- debug_print_tensor ( " : tensor=" , tensor);
3840
+ GGML_SYCL_DEBUG ( " %s " , debug_get_tensor_str ( " : tensor=" , tensor). c_str () );
3841
3841
GGML_SYCL_DEBUG (" size=%zu offset=%zu\n " , size, offset);
3842
3842
ggml_backend_sycl_context * sycl_ctx = (ggml_backend_sycl_context *)backend->context ;
3843
3843
ggml_backend_buffer_t buf = tensor->view_src ? tensor->view_src ->buffer : tensor->buffer ;
@@ -3860,8 +3860,8 @@ static bool ggml_backend_sycl_cpy_tensor_async(ggml_backend_t backend,
3860
3860
bool is_cpy_supported = dst->buffer ->buft == ggml_backend_sycl_buffer_type (sycl_ctx->device ) &&
3861
3861
ggml_backend_buffer_is_sycl (src->buffer );
3862
3862
GGML_SYCL_DEBUG (" [SYCL] call %s" , __func__);
3863
- debug_print_tensor ( " : dst=" , dst);
3864
- debug_print_tensor ( " src=" , src);
3863
+ GGML_SYCL_DEBUG ( " %s " , debug_get_tensor_str ( " : dst=" , dst). c_str () );
3864
+ GGML_SYCL_DEBUG ( " %s " , debug_get_tensor_str ( " src=" , src). c_str () );
3865
3865
GGML_SYCL_DEBUG (" is_cpy_supported=%d\n " , is_cpy_supported);
3866
3866
if (is_cpy_supported) {
3867
3867
/*
0 commit comments