You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Yet it still uses the Vulkan Backend instead of the OpenCL
This is a problem for two reasons,
1: The Vulkan backend doesn't work for me (see log attached).
2: The OpenCL backend has the advantage (or at least did in the past) of spilling additional VRAM usage from the context into system memory, provided it doesn't fit into VRAM, which the Vulkan backend seems to lack (I'm presuming here since its not working on my machine right now, but if I recall right this is the case..). Either way Vulkan is not OpenCL so why are they treated the same?
First Bad Commit
No response
Relevant log output
ggml_vulkan: Found 1 Vulkan devices:
Vulkan0: AMD Radeon RX 7900 XTX (RADV NAVI31) (radv) | uma: 0 | fp16: 1 | warp size: 64
build: 3974 (958367bf) with gcc (GCC) 14.2.1 20240910 for x86_64-pc-linux-gnu
main: llama backend init
main: load the model and apply lora adapter, if any
llama_load_model_from_file: using device Vulkan0 (AMD Radeon RX 7900 XTX (RADV NAVI31)) - 24560 MiB free
llama_model_loader: loaded meta data with 33 key-value pairs and 292 tensors from /home/aeslampanah/Desktop/Projects/LLaMA/Meta-Llama-3.1-8B-Instruct-Q8_0.gguf (version GGUF V3 (latest))
llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
llama_model_loader: - kv 0: general.architecture str = llama
llama_model_loader: - kv 1: general.type str = model
llama_model_loader: - kv 2: general.name str = Meta Llama 3.1 8B Instruct
llama_model_loader: - kv 3: general.finetune str = Instruct
llama_model_loader: - kv 4: general.basename str = Meta-Llama-3.1
llama_model_loader: - kv 5: general.size_label str = 8B
llama_model_loader: - kv 6: general.license str = llama3.1
llama_model_loader: - kv 7: general.tags arr[str,6] = ["facebook", "meta", "pytorch", "llam...llama_model_loader: - kv 8: general.languages arr[str,8] = ["en", "de", "fr", "it", "pt", "hi", ...llama_model_loader: - kv 9: llama.block_count u32 = 32llama_model_loader: - kv 10: llama.context_length u32 = 131072llama_model_loader: - kv 11: llama.embedding_length u32 = 4096llama_model_loader: - kv 12: llama.feed_forward_length u32 = 14336llama_model_loader: - kv 13: llama.attention.head_count u32 = 32llama_model_loader: - kv 14: llama.attention.head_count_kv u32 = 8llama_model_loader: - kv 15: llama.rope.freq_base f32 = 500000.000000llama_model_loader: - kv 16: llama.attention.layer_norm_rms_epsilon f32 = 0.000010llama_model_loader: - kv 17: general.file_type u32 = 7llama_model_loader: - kv 18: llama.vocab_size u32 = 128256llama_model_loader: - kv 19: llama.rope.dimension_count u32 = 128llama_model_loader: - kv 20: tokenizer.ggml.model str = gpt2llama_model_loader: - kv 21: tokenizer.ggml.pre str = llama-bpellama_model_loader: - kv 22: tokenizer.ggml.tokens arr[str,128256] = ["!", "\"", "#", "$", "%", "&", "'", ...
llama_model_loader: - kv 23: tokenizer.ggml.token_type arr[i32,128256] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ...
llama_model_loader: - kv 24: tokenizer.ggml.merges arr[str,280147] = ["Ġ Ġ", "Ġ ĠĠĠ", "ĠĠ ĠĠ", "...llama_model_loader: - kv 25: tokenizer.ggml.bos_token_id u32 = 128000llama_model_loader: - kv 26: tokenizer.ggml.eos_token_id u32 = 128009llama_model_loader: - kv 27: tokenizer.chat_template str = {{- bos_token }}\n{%- if custom_tools ...llama_model_loader: - kv 28: general.quantization_version u32 = 2llama_model_loader: - kv 29: quantize.imatrix.file str = /models_out/Meta-Llama-3.1-8B-Instruc...llama_model_loader: - kv 30: quantize.imatrix.dataset str = /training_dir/calibration_datav3.txtllama_model_loader: - kv 31: quantize.imatrix.entries_count i32 = 224llama_model_loader: - kv 32: quantize.imatrix.chunks_count i32 = 125llama_model_loader: - type f32: 66 tensorsllama_model_loader: - type q8_0: 226 tensorsllm_load_vocab: control token: 128254 '<|reserved_special_token_246|>' is not marked as EOGllm_load_vocab: control token: 128249 '<|reserved_special_token_241|>' is not marked as EOGllm_load_vocab: control token: 128246 '<|reserved_special_token_238|>' is not marked as EOGllm_load_vocab: control token: 128243 '<|reserved_special_token_235|>' is not marked as EOGllm_load_vocab: control token: 128242 '<|reserved_special_token_234|>' is not marked as EOGllm_load_vocab: control token: 128241 '<|reserved_special_token_233|>' is not marked as EOGllm_load_vocab: control token: 128240 '<|reserved_special_token_232|>' is not marked as EOGllm_load_vocab: control token: 128235 '<|reserved_special_token_227|>' is not marked as EOGllm_load_vocab: control token: 128231 '<|reserved_special_token_223|>' is not marked as EOGllm_load_vocab: control token: 128230 '<|reserved_special_token_222|>' is not marked as EOGllm_load_vocab: control token: 128228 '<|reserved_special_token_220|>' is not marked as EOGllm_load_vocab: control token: 128225 '<|reserved_special_token_217|>' is not marked as EOGllm_load_vocab: control token: 128218 '<|reserved_special_token_210|>' is not marked as EOGllm_load_vocab: control token: 128214 '<|reserved_special_token_206|>' is not marked as EOGllm_load_vocab: control token: 128213 '<|reserved_special_token_205|>' is not marked as EOGllm_load_vocab: control token: 128207 '<|reserved_special_token_199|>' is not marked as EOGllm_load_vocab: control token: 128206 '<|reserved_special_token_198|>' is not marked as EOGllm_load_vocab: control token: 128204 '<|reserved_special_token_196|>' is not marked as EOGllm_load_vocab: control token: 128200 '<|reserved_special_token_192|>' is not marked as EOGllm_load_vocab: control token: 128199 '<|reserved_special_token_191|>' is not marked as EOGllm_load_vocab: control token: 128198 '<|reserved_special_token_190|>' is not marked as EOGllm_load_vocab: control token: 128196 '<|reserved_special_token_188|>' is not marked as EOGllm_load_vocab: control token: 128194 '<|reserved_special_token_186|>' is not marked as EOGllm_load_vocab: control token: 128193 '<|reserved_special_token_185|>' is not marked as EOGllm_load_vocab: control token: 128188 '<|reserved_special_token_180|>' is not marked as EOGllm_load_vocab: control token: 128187 '<|reserved_special_token_179|>' is not marked as EOGllm_load_vocab: control token: 128185 '<|reserved_special_token_177|>' is not marked as EOGllm_load_vocab: control token: 128184 '<|reserved_special_token_176|>' is not marked as EOGllm_load_vocab: control token: 128180 '<|reserved_special_token_172|>' is not marked as EOGllm_load_vocab: control token: 128179 '<|reserved_special_token_171|>' is not marked as EOGllm_load_vocab: control token: 128178 '<|reserved_special_token_170|>' is not marked as EOGllm_load_vocab: control token: 128177 '<|reserved_special_token_169|>' is not marked as EOGllm_load_vocab: control token: 128176 '<|reserved_special_token_168|>' is not marked as EOGllm_load_vocab: control token: 128175 '<|reserved_special_token_167|>' is not marked as EOGllm_load_vocab: control token: 128171 '<|reserved_special_token_163|>' is not marked as EOGllm_load_vocab: control token: 128170 '<|reserved_special_token_162|>' is not marked as EOGllm_load_vocab: control token: 128169 '<|reserved_special_token_161|>' is not marked as EOGllm_load_vocab: control token: 128168 '<|reserved_special_token_160|>' is not marked as EOGllm_load_vocab: control token: 128165 '<|reserved_special_token_157|>' is not marked as EOGllm_load_vocab: control token: 128162 '<|reserved_special_token_154|>' is not marked as EOGllm_load_vocab: control token: 128158 '<|reserved_special_token_150|>' is not marked as EOGllm_load_vocab: control token: 128156 '<|reserved_special_token_148|>' is not marked as EOGllm_load_vocab: control token: 128155 '<|reserved_special_token_147|>' is not marked as EOGllm_load_vocab: control token: 128154 '<|reserved_special_token_146|>' is not marked as EOGllm_load_vocab: control token: 128151 '<|reserved_special_token_143|>' is not marked as EOGllm_load_vocab: control token: 128149 '<|reserved_special_token_141|>' is not marked as EOGllm_load_vocab: control token: 128147 '<|reserved_special_token_139|>' is not marked as EOGllm_load_vocab: control token: 128146 '<|reserved_special_token_138|>' is not marked as EOGllm_load_vocab: control token: 128144 '<|reserved_special_token_136|>' is not marked as EOGllm_load_vocab: control token: 128142 '<|reserved_special_token_134|>' is not marked as EOGllm_load_vocab: control token: 128141 '<|reserved_special_token_133|>' is not marked as EOGllm_load_vocab: control token: 128138 '<|reserved_special_token_130|>' is not marked as EOGllm_load_vocab: control token: 128136 '<|reserved_special_token_128|>' is not marked as EOGllm_load_vocab: control token: 128135 '<|reserved_special_token_127|>' is not marked as EOGllm_load_vocab: control token: 128134 '<|reserved_special_token_126|>' is not marked as EOGllm_load_vocab: control token: 128133 '<|reserved_special_token_125|>' is not marked as EOGllm_load_vocab: control token: 128131 '<|reserved_special_token_123|>' is not marked as EOGllm_load_vocab: control token: 128128 '<|reserved_special_token_120|>' is not marked as EOGllm_load_vocab: control token: 128124 '<|reserved_special_token_116|>' is not marked as EOGllm_load_vocab: control token: 128123 '<|reserved_special_token_115|>' is not marked as EOGllm_load_vocab: control token: 128122 '<|reserved_special_token_114|>' is not marked as EOGllm_load_vocab: control token: 128119 '<|reserved_special_token_111|>' is not marked as EOGllm_load_vocab: control token: 128115 '<|reserved_special_token_107|>' is not marked as EOGllm_load_vocab: control token: 128112 '<|reserved_special_token_104|>' is not marked as EOGllm_load_vocab: control token: 128110 '<|reserved_special_token_102|>' is not marked as EOGllm_load_vocab: control token: 128109 '<|reserved_special_token_101|>' is not marked as EOGllm_load_vocab: control token: 128108 '<|reserved_special_token_100|>' is not marked as EOGllm_load_vocab: control token: 128106 '<|reserved_special_token_98|>' is not marked as EOGllm_load_vocab: control token: 128103 '<|reserved_special_token_95|>' is not marked as EOGllm_load_vocab: control token: 128102 '<|reserved_special_token_94|>' is not marked as EOGllm_load_vocab: control token: 128101 '<|reserved_special_token_93|>' is not marked as EOGllm_load_vocab: control token: 128097 '<|reserved_special_token_89|>' is not marked as EOGllm_load_vocab: control token: 128091 '<|reserved_special_token_83|>' is not marked as EOGllm_load_vocab: control token: 128090 '<|reserved_special_token_82|>' is not marked as EOGllm_load_vocab: control token: 128089 '<|reserved_special_token_81|>' is not marked as EOGllm_load_vocab: control token: 128087 '<|reserved_special_token_79|>' is not marked as EOGllm_load_vocab: control token: 128085 '<|reserved_special_token_77|>' is not marked as EOGllm_load_vocab: control token: 128081 '<|reserved_special_token_73|>' is not marked as EOGllm_load_vocab: control token: 128078 '<|reserved_special_token_70|>' is not marked as EOGllm_load_vocab: control token: 128076 '<|reserved_special_token_68|>' is not marked as EOGllm_load_vocab: control token: 128075 '<|reserved_special_token_67|>' is not marked as EOGllm_load_vocab: control token: 128073 '<|reserved_special_token_65|>' is not marked as EOGllm_load_vocab: control token: 128068 '<|reserved_special_token_60|>' is not marked as EOGllm_load_vocab: control token: 128067 '<|reserved_special_token_59|>' is not marked as EOGllm_load_vocab: control token: 128065 '<|reserved_special_token_57|>' is not marked as EOGllm_load_vocab: control token: 128063 '<|reserved_special_token_55|>' is not marked as EOGllm_load_vocab: control token: 128062 '<|reserved_special_token_54|>' is not marked as EOGllm_load_vocab: control token: 128060 '<|reserved_special_token_52|>' is not marked as EOGllm_load_vocab: control token: 128059 '<|reserved_special_token_51|>' is not marked as EOGllm_load_vocab: control token: 128057 '<|reserved_special_token_49|>' is not marked as EOGllm_load_vocab: control token: 128054 '<|reserved_special_token_46|>' is not marked as EOGllm_load_vocab: control token: 128046 '<|reserved_special_token_38|>' is not marked as EOGllm_load_vocab: control token: 128045 '<|reserved_special_token_37|>' is not marked as EOGllm_load_vocab: control token: 128044 '<|reserved_special_token_36|>' is not marked as EOGllm_load_vocab: control token: 128043 '<|reserved_special_token_35|>' is not marked as EOGllm_load_vocab: control token: 128038 '<|reserved_special_token_30|>' is not marked as EOGllm_load_vocab: control token: 128036 '<|reserved_special_token_28|>' is not marked as EOGllm_load_vocab: control token: 128035 '<|reserved_special_token_27|>' is not marked as EOGllm_load_vocab: control token: 128032 '<|reserved_special_token_24|>' is not marked as EOGllm_load_vocab: control token: 128028 '<|reserved_special_token_20|>' is not marked as EOGllm_load_vocab: control token: 128027 '<|reserved_special_token_19|>' is not marked as EOGllm_load_vocab: control token: 128024 '<|reserved_special_token_16|>' is not marked as EOGllm_load_vocab: control token: 128023 '<|reserved_special_token_15|>' is not marked as EOGllm_load_vocab: control token: 128022 '<|reserved_special_token_14|>' is not marked as EOGllm_load_vocab: control token: 128021 '<|reserved_special_token_13|>' is not marked as EOGllm_load_vocab: control token: 128018 '<|reserved_special_token_10|>' is not marked as EOGllm_load_vocab: control token: 128016 '<|reserved_special_token_8|>' is not marked as EOGllm_load_vocab: control token: 128015 '<|reserved_special_token_7|>' is not marked as EOGllm_load_vocab: control token: 128013 '<|reserved_special_token_5|>' is not marked as EOGllm_load_vocab: control token: 128011 '<|reserved_special_token_3|>' is not marked as EOGllm_load_vocab: control token: 128005 '<|reserved_special_token_2|>' is not marked as EOGllm_load_vocab: control token: 128004 '<|finetune_right_pad_id|>' is not marked as EOGllm_load_vocab: control token: 128002 '<|reserved_special_token_0|>' is not marked as EOGllm_load_vocab: control token: 128252 '<|reserved_special_token_244|>' is not marked as EOGllm_load_vocab: control token: 128190 '<|reserved_special_token_182|>' is not marked as EOGllm_load_vocab: control token: 128183 '<|reserved_special_token_175|>' is not marked as EOGllm_load_vocab: control token: 128137 '<|reserved_special_token_129|>' is not marked as EOGllm_load_vocab: control token: 128182 '<|reserved_special_token_174|>' is not marked as EOGllm_load_vocab: control token: 128040 '<|reserved_special_token_32|>' is not marked as EOGllm_load_vocab: control token: 128048 '<|reserved_special_token_40|>' is not marked as EOGllm_load_vocab: control token: 128092 '<|reserved_special_token_84|>' is not marked as EOGllm_load_vocab: control token: 128215 '<|reserved_special_token_207|>' is not marked as EOGllm_load_vocab: control token: 128107 '<|reserved_special_token_99|>' is not marked as EOGllm_load_vocab: control token: 128208 '<|reserved_special_token_200|>' is not marked as EOGllm_load_vocab: control token: 128145 '<|reserved_special_token_137|>' is not marked as EOGllm_load_vocab: control token: 128031 '<|reserved_special_token_23|>' is not marked as EOGllm_load_vocab: control token: 128129 '<|reserved_special_token_121|>' is not marked as EOGllm_load_vocab: control token: 128201 '<|reserved_special_token_193|>' is not marked as EOGllm_load_vocab: control token: 128074 '<|reserved_special_token_66|>' is not marked as EOGllm_load_vocab: control token: 128095 '<|reserved_special_token_87|>' is not marked as EOGllm_load_vocab: control token: 128186 '<|reserved_special_token_178|>' is not marked as EOGllm_load_vocab: control token: 128143 '<|reserved_special_token_135|>' is not marked as EOGllm_load_vocab: control token: 128229 '<|reserved_special_token_221|>' is not marked as EOGllm_load_vocab: control token: 128007 '<|end_header_id|>' is not marked as EOGllm_load_vocab: control token: 128055 '<|reserved_special_token_47|>' is not marked as EOGllm_load_vocab: control token: 128056 '<|reserved_special_token_48|>' is not marked as EOGllm_load_vocab: control token: 128061 '<|reserved_special_token_53|>' is not marked as EOGllm_load_vocab: control token: 128153 '<|reserved_special_token_145|>' is not marked as EOGllm_load_vocab: control token: 128152 '<|reserved_special_token_144|>' is not marked as EOGllm_load_vocab: control token: 128212 '<|reserved_special_token_204|>' is not marked as EOGllm_load_vocab: control token: 128172 '<|reserved_special_token_164|>' is not marked as EOGllm_load_vocab: control token: 128160 '<|reserved_special_token_152|>' is not marked as EOGllm_load_vocab: control token: 128041 '<|reserved_special_token_33|>' is not marked as EOGllm_load_vocab: control token: 128181 '<|reserved_special_token_173|>' is not marked as EOGllm_load_vocab: control token: 128094 '<|reserved_special_token_86|>' is not marked as EOGllm_load_vocab: control token: 128118 '<|reserved_special_token_110|>' is not marked as EOGllm_load_vocab: control token: 128236 '<|reserved_special_token_228|>' is not marked as EOGllm_load_vocab: control token: 128148 '<|reserved_special_token_140|>' is not marked as EOGllm_load_vocab: control token: 128042 '<|reserved_special_token_34|>' is not marked as EOGllm_load_vocab: control token: 128139 '<|reserved_special_token_131|>' is not marked as EOGllm_load_vocab: control token: 128173 '<|reserved_special_token_165|>' is not marked as EOGllm_load_vocab: control token: 128239 '<|reserved_special_token_231|>' is not marked as EOGllm_load_vocab: control token: 128157 '<|reserved_special_token_149|>' is not marked as EOGllm_load_vocab: control token: 128052 '<|reserved_special_token_44|>' is not marked as EOGllm_load_vocab: control token: 128026 '<|reserved_special_token_18|>' is not marked as EOGllm_load_vocab: control token: 128003 '<|reserved_special_token_1|>' is not marked as EOGllm_load_vocab: control token: 128019 '<|reserved_special_token_11|>' is not marked as EOGllm_load_vocab: control token: 128116 '<|reserved_special_token_108|>' is not marked as EOGllm_load_vocab: control token: 128161 '<|reserved_special_token_153|>' is not marked as EOGllm_load_vocab: control token: 128226 '<|reserved_special_token_218|>' is not marked as EOGllm_load_vocab: control token: 128159 '<|reserved_special_token_151|>' is not marked as EOGllm_load_vocab: control token: 128012 '<|reserved_special_token_4|>' is not marked as EOGllm_load_vocab: control token: 128088 '<|reserved_special_token_80|>' is not marked as EOGllm_load_vocab: control token: 128163 '<|reserved_special_token_155|>' is not marked as EOGllm_load_vocab: control token: 128001 '<|end_of_text|>' is not marked as EOGllm_load_vocab: control token: 128113 '<|reserved_special_token_105|>' is not marked as EOGllm_load_vocab: control token: 128250 '<|reserved_special_token_242|>' is not marked as EOGllm_load_vocab: control token: 128125 '<|reserved_special_token_117|>' is not marked as EOGllm_load_vocab: control token: 128053 '<|reserved_special_token_45|>' is not marked as EOGllm_load_vocab: control token: 128224 '<|reserved_special_token_216|>' is not marked as EOGllm_load_vocab: control token: 128247 '<|reserved_special_token_239|>' is not marked as EOGllm_load_vocab: control token: 128251 '<|reserved_special_token_243|>' is not marked as EOGllm_load_vocab: control token: 128216 '<|reserved_special_token_208|>' is not marked as EOGllm_load_vocab: control token: 128006 '<|start_header_id|>' is not marked as EOGllm_load_vocab: control token: 128211 '<|reserved_special_token_203|>' is not marked as EOGllm_load_vocab: control token: 128077 '<|reserved_special_token_69|>' is not marked as EOGllm_load_vocab: control token: 128237 '<|reserved_special_token_229|>' is not marked as EOGllm_load_vocab: control token: 128086 '<|reserved_special_token_78|>' is not marked as EOGllm_load_vocab: control token: 128227 '<|reserved_special_token_219|>' is not marked as EOGllm_load_vocab: control token: 128058 '<|reserved_special_token_50|>' is not marked as EOGllm_load_vocab: control token: 128100 '<|reserved_special_token_92|>' is not marked as EOGllm_load_vocab: control token: 128209 '<|reserved_special_token_201|>' is not marked as EOGllm_load_vocab: control token: 128084 '<|reserved_special_token_76|>' is not marked as EOGllm_load_vocab: control token: 128071 '<|reserved_special_token_63|>' is not marked as EOGllm_load_vocab: control token: 128070 '<|reserved_special_token_62|>' is not marked as EOGllm_load_vocab: control token: 128049 '<|reserved_special_token_41|>' is not marked as EOGllm_load_vocab: control token: 128197 '<|reserved_special_token_189|>' is not marked as EOGllm_load_vocab: control token: 128072 '<|reserved_special_token_64|>' is not marked as EOGllm_load_vocab: control token: 128000 '<|begin_of_text|>' is not marked as EOGllm_load_vocab: control token: 128223 '<|reserved_special_token_215|>' is not marked as EOGllm_load_vocab: control token: 128217 '<|reserved_special_token_209|>' is not marked as EOGllm_load_vocab: control token: 128111 '<|reserved_special_token_103|>' is not marked as EOGllm_load_vocab: control token: 128203 '<|reserved_special_token_195|>' is not marked as EOGllm_load_vocab: control token: 128051 '<|reserved_special_token_43|>' is not marked as EOGllm_load_vocab: control token: 128030 '<|reserved_special_token_22|>' is not marked as EOGllm_load_vocab: control token: 128117 '<|reserved_special_token_109|>' is not marked as EOGllm_load_vocab: control token: 128010 '<|python_tag|>' is not marked as EOGllm_load_vocab: control token: 128238 '<|reserved_special_token_230|>' is not marked as EOGllm_load_vocab: control token: 128255 '<|reserved_special_token_247|>' is not marked as EOGllm_load_vocab: control token: 128202 '<|reserved_special_token_194|>' is not marked as EOGllm_load_vocab: control token: 128132 '<|reserved_special_token_124|>' is not marked as EOGllm_load_vocab: control token: 128248 '<|reserved_special_token_240|>' is not marked as EOGllm_load_vocab: control token: 128167 '<|reserved_special_token_159|>' is not marked as EOGllm_load_vocab: control token: 128127 '<|reserved_special_token_119|>' is not marked as EOGllm_load_vocab: control token: 128105 '<|reserved_special_token_97|>' is not marked as EOGllm_load_vocab: control token: 128039 '<|reserved_special_token_31|>' is not marked as EOGllm_load_vocab: control token: 128232 '<|reserved_special_token_224|>' is not marked as EOGllm_load_vocab: control token: 128166 '<|reserved_special_token_158|>' is not marked as EOGllm_load_vocab: control token: 128130 '<|reserved_special_token_122|>' is not marked as EOGllm_load_vocab: control token: 128114 '<|reserved_special_token_106|>' is not marked as EOGllm_load_vocab: control token: 128234 '<|reserved_special_token_226|>' is not marked as EOGllm_load_vocab: control token: 128191 '<|reserved_special_token_183|>' is not marked as EOGllm_load_vocab: control token: 128064 '<|reserved_special_token_56|>' is not marked as EOGllm_load_vocab: control token: 128140 '<|reserved_special_token_132|>' is not marked as EOGllm_load_vocab: control token: 128096 '<|reserved_special_token_88|>' is not marked as EOGllm_load_vocab: control token: 128098 '<|reserved_special_token_90|>' is not marked as EOGllm_load_vocab: control token: 128192 '<|reserved_special_token_184|>' is not marked as EOGllm_load_vocab: control token: 128093 '<|reserved_special_token_85|>' is not marked as EOGllm_load_vocab: control token: 128150 '<|reserved_special_token_142|>' is not marked as EOGllm_load_vocab: control token: 128222 '<|reserved_special_token_214|>' is not marked as EOGllm_load_vocab: control token: 128233 '<|reserved_special_token_225|>' is not marked as EOGllm_load_vocab: control token: 128220 '<|reserved_special_token_212|>' is not marked as EOGllm_load_vocab: control token: 128034 '<|reserved_special_token_26|>' is not marked as EOGllm_load_vocab: control token: 128033 '<|reserved_special_token_25|>' is not marked as EOGllm_load_vocab: control token: 128253 '<|reserved_special_token_245|>' is not marked as EOGllm_load_vocab: control token: 128195 '<|reserved_special_token_187|>' is not marked as EOGllm_load_vocab: control token: 128099 '<|reserved_special_token_91|>' is not marked as EOGllm_load_vocab: control token: 128189 '<|reserved_special_token_181|>' is not marked as EOGllm_load_vocab: control token: 128210 '<|reserved_special_token_202|>' is not marked as EOGllm_load_vocab: control token: 128174 '<|reserved_special_token_166|>' is not marked as EOGllm_load_vocab: control token: 128083 '<|reserved_special_token_75|>' is not marked as EOGllm_load_vocab: control token: 128080 '<|reserved_special_token_72|>' is not marked as EOGllm_load_vocab: control token: 128104 '<|reserved_special_token_96|>' is not marked as EOGllm_load_vocab: control token: 128082 '<|reserved_special_token_74|>' is not marked as EOGllm_load_vocab: control token: 128219 '<|reserved_special_token_211|>' is not marked as EOGllm_load_vocab: control token: 128017 '<|reserved_special_token_9|>' is not marked as EOGllm_load_vocab: control token: 128050 '<|reserved_special_token_42|>' is not marked as EOGllm_load_vocab: control token: 128205 '<|reserved_special_token_197|>' is not marked as EOGllm_load_vocab: control token: 128047 '<|reserved_special_token_39|>' is not marked as EOGllm_load_vocab: control token: 128164 '<|reserved_special_token_156|>' is not marked as EOGllm_load_vocab: control token: 128020 '<|reserved_special_token_12|>' is not marked as EOGllm_load_vocab: control token: 128069 '<|reserved_special_token_61|>' is not marked as EOGllm_load_vocab: control token: 128245 '<|reserved_special_token_237|>' is not marked as EOGllm_load_vocab: control token: 128121 '<|reserved_special_token_113|>' is not marked as EOGllm_load_vocab: control token: 128079 '<|reserved_special_token_71|>' is not marked as EOGllm_load_vocab: control token: 128037 '<|reserved_special_token_29|>' is not marked as EOGllm_load_vocab: control token: 128244 '<|reserved_special_token_236|>' is not marked as EOGllm_load_vocab: control token: 128029 '<|reserved_special_token_21|>' is not marked as EOGllm_load_vocab: control token: 128221 '<|reserved_special_token_213|>' is not marked as EOGllm_load_vocab: control token: 128066 '<|reserved_special_token_58|>' is not marked as EOGllm_load_vocab: control token: 128120 '<|reserved_special_token_112|>' is not marked as EOGllm_load_vocab: control token: 128014 '<|reserved_special_token_6|>' is not marked as EOGllm_load_vocab: control token: 128025 '<|reserved_special_token_17|>' is not marked as EOGllm_load_vocab: control token: 128126 '<|reserved_special_token_118|>' is not marked as EOGllm_load_vocab: special tokens cache size = 256llm_load_vocab: token to piece cache size = 0.7999 MBllm_load_print_meta: format = GGUF V3 (latest)llm_load_print_meta: arch = llamallm_load_print_meta: vocab type = BPEllm_load_print_meta: n_vocab = 128256llm_load_print_meta: n_merges = 280147llm_load_print_meta: vocab_only = 0llm_load_print_meta: n_ctx_train = 131072llm_load_print_meta: n_embd = 4096llm_load_print_meta: n_layer = 32llm_load_print_meta: n_head = 32llm_load_print_meta: n_head_kv = 8llm_load_print_meta: n_rot = 128llm_load_print_meta: n_swa = 0llm_load_print_meta: n_embd_head_k = 128llm_load_print_meta: n_embd_head_v = 128llm_load_print_meta: n_gqa = 4llm_load_print_meta: n_embd_k_gqa = 1024llm_load_print_meta: n_embd_v_gqa = 1024llm_load_print_meta: f_norm_eps = 0.0e+00llm_load_print_meta: f_norm_rms_eps = 1.0e-05llm_load_print_meta: f_clamp_kqv = 0.0e+00llm_load_print_meta: f_max_alibi_bias = 0.0e+00llm_load_print_meta: f_logit_scale = 0.0e+00llm_load_print_meta: n_ff = 14336llm_load_print_meta: n_expert = 0llm_load_print_meta: n_expert_used = 0llm_load_print_meta: causal attn = 1llm_load_print_meta: pooling type = 0llm_load_print_meta: rope type = 0llm_load_print_meta: rope scaling = linearllm_load_print_meta: freq_base_train = 500000.0llm_load_print_meta: freq_scale_train = 1llm_load_print_meta: n_ctx_orig_yarn = 131072llm_load_print_meta: rope_finetuned = unknownllm_load_print_meta: ssm_d_conv = 0llm_load_print_meta: ssm_d_inner = 0llm_load_print_meta: ssm_d_state = 0llm_load_print_meta: ssm_dt_rank = 0llm_load_print_meta: ssm_dt_b_c_rms = 0llm_load_print_meta: model type = 8Bllm_load_print_meta: model ftype = Q8_0llm_load_print_meta: model params = 8.03 Bllm_load_print_meta: model size = 7.95 GiB (8.50 BPW) llm_load_print_meta: general.name = Meta Llama 3.1 8B Instructllm_load_print_meta: BOS token = 128000 '<|begin_of_text|>'llm_load_print_meta: EOS token = 128009 '<|eot_id|>'llm_load_print_meta: EOT token = 128009 '<|eot_id|>'llm_load_print_meta: EOM token = 128008 '<|eom_id|>'llm_load_print_meta: LF token = 128 'Ä'llm_load_print_meta: EOG token = 128008 '<|eom_id|>'llm_load_print_meta: EOG token = 128009 '<|eot_id|>'llm_load_print_meta: max token length = 256llm_load_tensors: ggml ctx size = 0.27 MiBllm_load_tensors: offloading 32 repeating layers to GPUllm_load_tensors: offloading non-repeating layers to GPUllm_load_tensors: offloaded 33/33 layers to GPUllm_load_tensors: CPU buffer size = 532.31 MiBllm_load_tensors: Vulkan0 buffer size = 7605.34 MiB.........................................................................................llama_new_context_with_model: n_ctx = 8192llama_new_context_with_model: n_batch = 4096llama_new_context_with_model: n_ubatch = 512llama_new_context_with_model: flash_attn = 0llama_new_context_with_model: freq_base = 500000.0llama_new_context_with_model: freq_scale = 1llama_kv_cache_init: Vulkan0 KV buffer size = 1024.00 MiBllama_new_context_with_model: KV self size = 1024.00 MiB, K (f16): 512.00 MiB, V (f16): 512.00 MiBllama_new_context_with_model: Vulkan_Host output buffer size = 0.49 MiBllama_new_context_with_model: Vulkan0 compute buffer size = 560.00 MiBllama_new_context_with_model: Vulkan_Host compute buffer size = 24.01 MiBllama_new_context_with_model: graph nodes = 1030llama_new_context_with_model: graph splits = 2common_init_from_params: warming up the model with an empty run - please wait ... (--no-warmup to disable)main: llama threadpool init, n_threads = 12system_info: n_threads = 12 (n_threads_batch = 12) / 24 | AVX = 1 | AVX_VNNI = 0 | AVX2 = 1 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | AMX_INT8 = 0 | FMA = 1 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | RISCV_VECT = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | n_ctx: 8192, add_bos: 1tokenize the promptprompt: "Narrative and narrator comments are bracketed [like this].
Spelling in narratives is correct."tokens: [ '<|begin_of_text|>':128000, 'Narr':64379, 'ative':1413, ' and':323, ' narrator':65271, ' comments':6170, ' are':527, ' bracket':32742, 'ed':291, ' [':510, 'like':4908, ' this':420, '].':27218, 'Sp':6540, 'elling':6427, ' in':304, ' narratives':58378, ' is':374, ' correct':4495, '.':13 ]recalculate the cached logits (check): embd_inp.size() 20, n_matching_session_tokens 0, embd_inp.size() 20, session_tokens.size() 0main: interactive mode on.Reverse prompt: 'Amir:'sampler seed: 17sampler params: repeat_last_n = 512, repeat_penalty = 1.000, frequency_penalty = 0.000, presence_penalty = 0.000 top_k = 40, tfs_z = 0.800, top_p = 0.950, min_p = 0.050, xtc_probability = 0.000, xtc_threshold = 0.100, typical_p = 1.100, temp = 0.800 mirostat = 0, mirostat_lr = 0.100, mirostat_ent = 5.000sampler chain: logits -> logit-bias -> penalties -> tail-free -> typical -> temp-ext -> dist generate: n_ctx = 8192, n_batch = 4096, n_predict = -2, n_keep = 20== Running in interactive mode. == - Press Ctrl+C to interject at any time. - Press Return to return control to the AI. - To return control without starting a new line, end your input with '/'. - If you want to submit another line, end your input with '\'.embd_inp.size(): 20, n_consumed: 0Narrative and narrator comments are bracketed [like this].Spelling in narratives is correct.eval: [ '<|begin_of_text|>':128000, 'Narr':64379, 'ative':1413, ' and':323, ' narrator':65271, ' comments':6170, ' are':527, ' bracket':32742, 'ed':291, ' [':510, 'like':4908, ' this':420, '].':27218, 'Sp':6540, 'elling':6427, ' in':304, ' narratives':58378, ' is':374, ' correct':4495, '.':13 ]n_past = 20n_remain: -3 Speval: [ ' Sp':3165 ]n_past = 21src/llama-sampling.cpp:91: GGML_ASSERT(cur_p->size > 0) failed
The text was updated successfully, but these errors were encountered:
Yeah, your issue with the Vulkan backend was unrelated to the backend itself, some sampling thing.
Not sure how you got to this point, but the current OpenCL backend is very fresh and just for Qualcomm phones and maybe Intel iGPUs. I don't think it supports AMD. But no clue why it goes to Vulkan even though you compiled for OpenCL.
@akumaburn you seem to have some stale old build somehow.
I know you're saying you rebuild everything but if you build it with GGML_VULKAN=OFF then the Vulkan backend will not be included. In your case it clearly is:
Name and Version
(base) [aeslampanah@a-pc llama.cpp]$ ./llama-cli --version
ggml_vulkan: Found 1 Vulkan devices:
Vulkan0: AMD Radeon RX 7900 XTX (RADV NAVI31) (radv) | uma: 0 | fp16: 1 | warp size: 64
version: 3974 (958367b)
built with gcc (GCC) 14.2.1 20240910 for x86_64-pc-linux-gnu
Operating systems
Linux
Which llama.cpp modules do you know to be affected?
llama-cli
Command line
Problem description & steps to reproduce
I've compiled llama.cpp like so:
Yet it still uses the Vulkan Backend instead of the OpenCL
This is a problem for two reasons,
1: The Vulkan backend doesn't work for me (see log attached).
2: The OpenCL backend has the advantage (or at least did in the past) of spilling additional VRAM usage from the context into system memory, provided it doesn't fit into VRAM, which the Vulkan backend seems to lack (I'm presuming here since its not working on my machine right now, but if I recall right this is the case..). Either way Vulkan is not OpenCL so why are they treated the same?
First Bad Commit
No response
Relevant log output
The text was updated successfully, but these errors were encountered: