|
15 | 15 |
|
16 | 16 | """Global Default Configs for Models, Finetune Wrappers and Prompt Instructions Catalog.
|
17 | 17 |
|
18 |
| -These configs generally do not need to be accessed directly, but can be viewed, accessed and modified through |
| 18 | +These configs generally do not need to be accessed directly, but should be viewed, accessed and modified through |
19 | 19 | ModelCatalog and PromptCatalog classes.
|
20 | 20 |
|
21 | 21 | For customization, there is also the option in ModelCatalog to load a custom model catalog from json file, which
|
|
24 | 24 |
|
25 | 25 | global_model_repo_catalog_list = [
|
26 | 26 |
|
27 |
| - # embedding models |
| 27 | + {"model_name": "bling-tiny-llama-onnx", "model_family": "ONNXGenerativeModel", |
| 28 | + "model_category": "generative_local", "display_name": "llmware/bling-tiny-llama-onnx", |
| 29 | + "model_location": "llmware_repo","context_window": 2048, "instruction_following": False, |
| 30 | + "prompt_wrapper": "human_bot", "temperature": 0.0, "sample_default": False, "trailing_space": "", |
| 31 | + "hf_repo": "llmware/bling-tiny-llama-onnx", "custom_model_files": [], "custom_model_repo": "", |
| 32 | + "snapshot": True, "tokenizer_local": "tokenizer_tl.json", |
| 33 | + "fetch": {"snapshot": True, "module": "llmware.models", "method": "pull_snapshot_from_hf"}, |
| 34 | + "validation_files": ["model.onnx", "model.onnx.data"], |
| 35 | + "link": "https://huggingface.co/llmware/bling-tiny-llama-onnx"}, |
| 36 | + |
| 37 | + {"model_name": "bling-tiny-llama-ov", "model_family": "OVGenerativeModel", |
| 38 | + "model_category": "generative_local", "display_name": "bling-tiny-llama-ov", |
| 39 | + "model_location": "llmware_repo", |
| 40 | + "context_window": 2048, "instruction_following": False, "prompt_wrapper": "human_bot", |
| 41 | + "temperature": 0.0, "sample_default": False, "trailing_space": "", |
| 42 | + "tokenizer_local": "tokenizer_tl.json", |
| 43 | + "hf_repo": "llmware/bling-tiny-llama-ov", |
| 44 | + "custom_model_files": [], "custom_model_repo": "", |
| 45 | + "fetch": {"snapshot": True, "module": "llmware.models", "method": "pull_snapshot_from_hf"}, |
| 46 | + "validation_files": ["openvino_model.xml"], |
| 47 | + "link": "https://huggingface.co/llmware/bling-tiny-llama-ov"}, |
| 48 | + |
| 49 | + {"model_name": "bling-phi-3-ov", "model_family": "OVGenerativeModel", |
| 50 | + "model_category": "generative_local", "display_name": "bling-phi-3-ov", |
| 51 | + "model_location": "llmware_repo", |
| 52 | + "context_window": 4096, "instruction_following": False, "prompt_wrapper": "human_bot", |
| 53 | + "temperature": 0.0, "sample_default": False, "trailing_space": "", |
| 54 | + "tokenizer_local": "tokenizer_phi3.json", |
| 55 | + "hf_repo": "llmware/bling-phi-3-ov", |
| 56 | + "custom_model_files": [], "custom_model_repo": "", |
| 57 | + "fetch": {"snapshot": True, "module": "llmware.models","method": "pull_snapshot_from_hf"}, |
| 58 | + "validation_files": ["openvino_model.xml"], |
| 59 | + "link": "https://huggingface.co/llmware/bling-phi-3-ov"}, |
| 60 | + |
| 61 | + {"model_name": "bling-phi-3-onnx", "model_family": "ONNXGenerativeModel", |
| 62 | + "model_category": "generative_local", "display_name": "bling-phi-3-onnx", |
| 63 | + "model_location": "llmware_repo", |
| 64 | + "context_window": 4096, "instruction_following": False, "prompt_wrapper": "human_bot", |
| 65 | + "temperature": 0.0, "sample_default": False, "trailing_space": "", |
| 66 | + "tokenizer_local": "tokenizer_phi3.json", |
| 67 | + "hf_repo": "llmware/bling-phi-3-onnx", |
| 68 | + "custom_model_files": [], "custom_model_repo": "", |
| 69 | + "fetch": {"snapshot": True, "module": "llmware.models", "method": "pull_snapshot_from_hf"}, |
| 70 | + "validation_files": ["model.onnx", "model.onnx.data"], |
| 71 | + "link": "https://huggingface.co/llmware/bling-phi-3-onnx"}, |
| 72 | + |
| 73 | + {"model_name": "phi-3-onnx", "model_family": "ONNXGenerativeModel", |
| 74 | + "model_category": "generative_local", "display_name": "phi-3-onnx", |
| 75 | + "model_location": "llmware_repo", |
| 76 | + "context_window": 4096, "instruction_following": False, "prompt_wrapper": "human_bot", |
| 77 | + "temperature": 0.0, "sample_default": False, "trailing_space": "", |
| 78 | + "tokenizer_local": "tokenizer_phi3.json", |
| 79 | + "hf_repo": "llmware/phi-3-onnx", |
| 80 | + "custom_model_files": [], "custom_model_repo": "", |
| 81 | + "fetch": {"snapshot": True, "module": "llmware.models", "method": "pull_snapshot_from_hf"}, |
| 82 | + "validation_files": ["model.onnx", "model.onnx.data"], |
| 83 | + "link": "https://huggingface.co/llmware/phi-3-onnx"}, |
| 84 | + |
| 85 | + {"model_name": "phi-3-ov", "model_family": "OVGenerativeModel", |
| 86 | + "model_category": "generative_local", "display_name": "phi-3-ov", |
| 87 | + "model_location": "llmware_repo", |
| 88 | + "context_window": 4096, "instruction_following": False, "prompt_wrapper": "human_bot", |
| 89 | + "temperature": 0.0, "sample_default": False, "trailing_space": "", |
| 90 | + "tokenizer_local": "tokenizer_phi3.json", |
| 91 | + "hf_repo": "llmware/phi-3-ov", |
| 92 | + "custom_model_files": [], "custom_model_repo": "", |
| 93 | + "fetch": {"snapshot": True, "module": "llmware.models", "method": "pull_snapshot_from_hf"}, |
| 94 | + "validation_files": ["openvino_model.xml"], |
| 95 | + "link": "https://huggingface.co/llmware/phi-3-ov"}, |
| 96 | + |
| 97 | + # embedding models |
28 | 98 |
|
29 | 99 | {"model_name": "all-MiniLM-L6-v2", "display_name": "mini-lm-sbert", "model_family": "HFEmbeddingModel",
|
30 | 100 | "model_category": "embedding", "model_location": "hf_repo", "embedding_dims": 384, "context_window": 512,
|
|
0 commit comments