Skip to content

Commit e09b676

Browse files
committed
Merge branch 'concedo_experimental' into frankenstein_11
2 parents 97b78f0 + 292d743 commit e09b676

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

81 files changed

+15330
-13919
lines changed

.github/labeler.yml

Lines changed: 78 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,78 @@
1+
# https://github.com/actions/labeler
2+
3+
SYCL:
4+
- changed-files:
5+
- any-glob-to-any-file:
6+
- ggml-sycl.h
7+
- ggml-sycl.cpp
8+
- README-sycl.md
9+
Nvidia GPU:
10+
- changed-files:
11+
- any-glob-to-any-file:
12+
- ggml-cuda/**
13+
Vulkan:
14+
- changed-files:
15+
- any-glob-to-any-file:
16+
- ggml_vk_generate_shaders.py
17+
- ggml-vulkan*
18+
documentation:
19+
- changed-files:
20+
- any-glob-to-any-file:
21+
- docs/**
22+
- media/**
23+
testing:
24+
- changed-files:
25+
- any-glob-to-any-file:
26+
- tests/**
27+
build:
28+
- changed-files:
29+
- any-glob-to-any-file:
30+
- cmake/**
31+
- CMakeLists.txt
32+
- CMakePresets.json
33+
- codecov.yml
34+
examples:
35+
- changed-files:
36+
- any-glob-to-any-file: examples/**
37+
devops:
38+
- changed-files:
39+
- any-glob-to-any-file:
40+
- .devops/**
41+
- .github/**
42+
- ci/**
43+
python:
44+
- changed-files:
45+
- any-glob-to-any-file:
46+
- "**/*.py"
47+
- requirements/**
48+
- gguf-py/**
49+
- .flake8
50+
script:
51+
- changed-files:
52+
- any-glob-to-any-file:
53+
- scripts/**
54+
android:
55+
- changed-files:
56+
- any-glob-to-any-file:
57+
- examples/llama.android/**
58+
server:
59+
- changed-files:
60+
- any-glob-to-any-file:
61+
- examples/server/**
62+
ggml:
63+
- changed-files:
64+
- any-glob-to-any-file:
65+
- ggml.c
66+
- ggml.h
67+
- ggml-*.c
68+
- ggml-*.h
69+
- ggml-cuda/**
70+
nix:
71+
- changed-files:
72+
- any-glob-to-any-file:
73+
- "**/*.nix"
74+
- .github/workflows/nix-*.yml
75+
- .devops/nix/nixpkgs-instances.nix
76+
embedding:
77+
- changed-files:
78+
- any-glob-to-any-file: examples/embedding/

.github/workflows/labeler.yml

Lines changed: 17 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,17 @@
1+
name: "Pull Request Labeler"
2+
on:
3+
- pull_request_target
4+
5+
jobs:
6+
labeler:
7+
permissions:
8+
contents: read
9+
pull-requests: write
10+
runs-on: ubuntu-latest
11+
steps:
12+
- uses: actions/checkout@v4
13+
with:
14+
repository: "ggerganov/llama.cpp"
15+
- uses: actions/labeler@v5
16+
with:
17+
configuration-path: '.github/labeler.yml'

CMakeLists.txt

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -428,6 +428,8 @@ set_target_properties(ggml_v3 PROPERTIES POSITION_INDEPENDENT_CODE ON)
428428
add_library(common2
429429
common/common.cpp
430430
common/common.h
431+
common/sampling.cpp
432+
common/sampling.h
431433
common/grammar-parser.h
432434
common/grammar-parser.cpp
433435
examples/llava/llava.cpp

Makefile

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -62,9 +62,9 @@ endif
6262
CUBLASLD_FLAGS =
6363
CUBLAS_OBJS =
6464

65-
OBJS_FULL += ggml-alloc.o ggml-quants.o unicode.o unicode-data.o sgemm.o common.o grammar-parser.o
66-
OBJS_SIMPLE += ggml-alloc.o ggml-quants_noavx2.o unicode.o unicode-data.o sgemm_noavx2.o common.o grammar-parser.o
67-
OBJS_FAILSAFE += ggml-alloc.o ggml-quants_failsafe.o unicode.o unicode-data.o sgemm_failsafe.o common.o grammar-parser.o
65+
OBJS_FULL += ggml-alloc.o ggml-quants.o unicode.o unicode-data.o sgemm.o common.o sampling.o grammar-parser.o
66+
OBJS_SIMPLE += ggml-alloc.o ggml-quants_noavx2.o unicode.o unicode-data.o sgemm_noavx2.o common.o sampling.o grammar-parser.o
67+
OBJS_FAILSAFE += ggml-alloc.o ggml-quants_failsafe.o unicode.o unicode-data.o sgemm_failsafe.o common.o sampling.o grammar-parser.o
6868

6969
#lets try enabling everything
7070
CFLAGS += -pthread -s -Wno-deprecated -Wno-deprecated-declarations
@@ -497,6 +497,8 @@ llama.o: llama.cpp ggml.h ggml-alloc.h ggml-backend.h ggml-cuda.h ggml-metal.h l
497497
$(CXX) $(CXXFLAGS) -c $< -o $@
498498
common.o: common/common.cpp common/common.h common/log.h
499499
$(CXX) $(CXXFLAGS) -c $< -o $@
500+
sampling.o: common/sampling.cpp common/common.h common/sampling.h common/log.h
501+
$(CXX) $(CXXFLAGS) -c $< -o $@
500502
console.o: common/console.cpp common/console.h
501503
$(CXX) $(CXXFLAGS) -c $< -o $@
502504
grammar-parser.o: common/grammar-parser.cpp common/grammar-parser.h

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -58,7 +58,7 @@ when you can't use the precompiled binary directly, we provide an automated buil
5858
- For Arch Linux: Install `cblas` `openblas` and `clblast`.
5959
- For Debian: Install `libclblast-dev` and `libopenblas-dev`.
6060
- You can attempt a CuBLAS build with `LLAMA_CUBLAS=1`. You will need CUDA Toolkit installed. Some have also reported success with the CMake file, though that is more for windows.
61-
- For a full featured build (all backends), do `make LLAMA_OPENBLAS=1 LLAMA_CLBLAST=1 LLAMA_CUBLAS=1 LLAMA_VULKAN=1`
61+
- For a full featured build (all backends), do `make LLAMA_OPENBLAS=1 LLAMA_CLBLAST=1 LLAMA_CUBLAS=1 LLAMA_VULKAN=1`. (Note that `LLAMA_CUBLAS=1` will not work on windows, you need visual studio)
6262
- After all binaries are built, you can run the python script with the command `koboldcpp.py [ggml_model.bin] [port]`
6363

6464
- Note: Many OSX users have found that the using Accelerate is actually faster than OpenBLAS. To try, you may wish to run with `--noblas` and compare speeds.

0 commit comments

Comments
 (0)