Table of Best Results #2755
SilvaRaulEnrique
started this conversation in
Ideas
Replies: 2 comments 4 replies
-
Running on an M2, built from scratch, I get this error: Anyone know why this is happening? |
Beta Was this translation helpful? Give feedback.
4 replies
-
"-gqa used to be needed when loading Llama2 70B model, but in the current code with the new GGUF format it is no longer used" |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
The idea of this post is construct and maintain colaborativaly a complete and actualizated best orders ranking list table to compile and executate, considering best models results for the more very long context with efectively complety coherent response, by example:
For NVidia 3090 (with CUDA and 24 GB RAM), for now the best orders are:
or?
./train-text-from-scratch
--vocab-model ServidorIA/models/s3nh/longchat-7b-v1.5-32k.ggmlv3.q8_0.bin
--ctx 64
--embd 256
--head 8
--layer 16
--checkpoint-out chk-Ultimas_Acordadas_y_Circulares-256x16.bin
--model-out ggml-Ultimas_Acordadas_y_Circulares-256x16-f32.bin
--train-data "Ultimas_Acordadas_y_Circulares.txt"
-t 4
-b 8
-n 32
--seed 1
--adam-iter 16
--print-details-interval 0
--predict 16
--use-flash
--mem-compute 8
Note: models TheBloke, and s3nh are in https://huggingface.co/
¿There is the best parameters or considerer "now" one best?, ¿and for another hardware configuration?
Beta Was this translation helpful? Give feedback.
All reactions