commit 353ec251a42491f5192c48561da4b444ef67f23c main: seed = 1679424064 llama_model_load: loading model from 'models/30B/ggml-model-q4_0.bin' - please wait ... llama_model_load: n_vocab = 32000 llama_model_load: n_ctx = 512 llama_model_load: n_embd = 6656 llama_model_load: n_mult = 256 llama_model_load: n_head = 52 llama_model_load: n_layer = 60 llama_model_load: n_rot = 128 llama_model_load: f16 = 2 llama_model_load: n_ff = 17920 llama_model_load: n_parts = 4 llama_model_load: ggml ctx size = 20951.50 MB llama_model_load: memory_size = 1560.00 MB, n_mem = 30720 llama_model_load: loading model part 1/4 from 'models/30B/ggml-model-q4_0.bin' llama_model_load: ................................................................... done llama_model_load: model size = 4850.14 MB / num tensors = 543 llama_model_load: loading model part 2/4 from 'models/30B/ggml-model-q4_0.bin.1' llama_model_load: ................................................................... done llama_model_load: model size = 4850.14 MB / num tensors = 543 llama_model_load: loading model part 3/4 from 'models/30B/ggml-model-q4_0.bin.2' llama_model_load: ................................................................... done llama_model_load: model size = 4850.14 MB / num tensors = 543 llama_model_load: loading model part 4/4 from 'models/30B/ggml-model-q4_0.bin.3' llama_model_load: ................................................................... done llama_model_load: model size = 4850.14 MB / num tensors = 543 system_info: n_threads = 16 / 32 | AVX = 1 | AVX2 = 1 | AVX512 = 0 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | VSX = 0 | Calculating perplexity over 222 chunks 187.76 seconds per pass - ETA 11.58 hours [1]3.4156,[2]3.8599,[3]4.5575,[4]4.7039,[5]4.6240,[6]4.5976,[7]4.7593,[8]4.8232,[9]5.0702,[10]5.2904,[11]5.4586,[12]5.4983,[13]5.4620,[14]5.5447,[15]5.6865,[16]5.4533,[17]5.4181,[18]5.4289,[19]5.2077,[20]5.2249,[21]5.1701,[22]5.0011,[23]4.9683,[24]4.9033,[25]4.8899,[26]4.7535,[27]4.6014,[28]4.5092,[29]4.4205,[30]4.3012,[31]4.2311,[32]4.2518,[33]4.2055,[34]4.2351,[35]4.2495,[36]4.2717,[37]4.2679,[38]4.2686,[39]4.2935,[40]4.3320,[41]4.3484,[42]4.3794,[43]4.3606,[44]4.4113,[45]4.4339,[46]4.4205,[47]4.4471,[48]4.4424,[49]4.4481,[50]4.4255,[51]4.4396,[52]4.4414,[53]4.4844,[54]4.4764,[55]4.4573,[56]4.4129,[57]4.3826,[58]4.3983,[59]4.4146,[60]4.4471,[61]4.4477,[62]4.4956,[63]4.5113,[64]4.5225,[65]4.5492,[66]4.5420,[67]4.5540,[68]4.5721,[69]4.5946,[70]4.6250,[71]4.6474,[72]4.6820,[73]4.7222,[74]4.7268,[75]4.7373,[76]4.7432,[77]4.7495,[78]4.7362,[79]4.7617,[80]4.7632,[81]4.7720,[82]4.7732,[83]4.7426,[84]4.7365,[85]4.7406,[86]4.7331,[87]4.6804,[88]4.6282,[89]4.5911,[90]4.5699,[91]4.5698,[92]4.5624,[93]4.5619,[94]4.5618,[95]4.5845,[96]4.5811,[97]4.5740,[98]4.5720,[99]4.5594,[100]4.5547,[101]4.5669,[102]4.5622,[103]4.5723,[104]4.5747,[105]4.5771,[106]4.5881,[107]4.5820,[108]4.5851,[109]4.5819,[110]4.5798,[111]4.5944,[112]4.6139,[113]4.6145,[114]4.6098,[115]4.6126,[116]4.6039,[117]4.6062,[118]4.6273,[119]4.6455,[120]4.6756,[121]4.6886,[122]4.7086,[123]4.7391,[124]4.7558,[125]4.7553,[126]4.7903,[127]4.8212,[128]4.8456,[129]4.8352,[130]4.8412,[131]4.8416,[132]4.8395,[133]4.8318,[134]4.8415,[135]4.8423,[136]4.8388,[137]4.8394,[138]4.8314,[139]4.8271,[140]4.8288,[141]4.8068,[142]4.8057,[143]4.7834,[144]4.7698,[145]4.7648,[146]4.7594,[147]4.7667,[148]4.7717,[149]4.7708,[150]4.7730,[151]4.7787,[152]4.7772,[153]4.7715,[154]4.7719,[155]4.7805,[156]4.7810,[157]4.7975,[158]4.8019,[159]4.8074,[160]4.8155,[161]4.8307,[162]4.8131,[163]4.8068,[164]4.7907,[165]4.7707,[166]4.7531,[167]4.7290,[168]4.7073,[169]4.6974,[170]4.6923,[171]4.6757,[172]4.6677,[173]4.6575,[174]4.6385,[175]4.6238,[176]4.6143,[177]4.6019,[178]4.5868,[179]4.5767,[180]4.5720,[181]4.5588,[182]4.5465,[183]4.5406,[184]4.5427,[185]4.5355,[186]4.5387,[187]4.5466,[188]4.5437,[189]4.5594,[190]4.5607,[191]4.5777,[192]4.6006,[193]4.6131,[194]4.6226,[195]4.6423,[196]4.6540,[197]4.6711,[198]4.6835,[199]4.6865,[200]4.6713,[201]4.6573,[202]4.6581,[203]4.6498,[204]4.6437,[205]4.6452,[206]4.6440,[207]4.6398,[208]4.6402,[209]4.6428,[210]4.6493,[211]4.6585,[212]4.6661,[213]4.6754,[214]4.6797,[215]4.6826,[216]4.6956,[217]4.7105,[218]4.7233,[219]4.7254,[220]4.7244,[221]4.7224,[222]4.7241, Performance counter stats for './main -t 16 --perplexity -m models/30B/ggml-model-q4_0.bin -f ./models/wiki.test.raw.1404': 657691199.45 msec task-clock # 15.988 CPUs utilized 658325 context-switches # 1.001 /sec 226244 cpu-migrations # 0.344 /sec 16542626 page-faults # 25.153 /sec 2399687346212661 cycles # 3.649 GHz (16.67%) 1276014188975 stalled-cycles-frontend # 0.05% frontend cycles idle (16.67%) 1934998826371070 stalled-cycles-backend # 80.64% backend cycles idle (16.67%) 4342351759822173 instructions # 1.81 insn per cycle # 0.45 stalled cycles per insn (16.67%) 298417540497490 branches # 453.735 M/sec (16.67%) 499961519283 branch-misses # 0.17% of all branches (16.67%) 41135.305353635 seconds time elapsed 657623.473526000 seconds user 68.743263000 seconds sys