]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commitdiff
chmod : make scripts executable (#2675)
authorCebtenzzre <redacted>
Wed, 23 Aug 2023 14:29:09 +0000 (10:29 -0400)
committerGitHub <redacted>
Wed, 23 Aug 2023 14:29:09 +0000 (17:29 +0300)
23 files changed:
ci/run.sh [changed mode: 0644->0755]
convert-falcon-hf-to-gguf.py [changed mode: 0644->0755]
convert-gptneox-hf-to-gguf.py [changed mode: 0644->0755]
convert-llama-7b-pth-to-gguf.py [changed mode: 0644->0755]
convert-llama-ggmlv3-to-gguf.py [changed mode: 0644->0755]
convert-llama-hf-to-gguf.py [changed mode: 0644->0755]
convert-lora-to-ggml.py
convert.py [changed mode: 0644->0755]
examples/embd-input/embd_input.py [changed mode: 0644->0755]
examples/embd-input/llava.py [changed mode: 0644->0755]
examples/embd-input/minigpt4.py [changed mode: 0644->0755]
examples/embd-input/panda_gpt.py [changed mode: 0644->0755]
examples/jeopardy/graph.py [changed mode: 0644->0755]
examples/jeopardy/jeopardy.sh [changed mode: 0644->0755]
examples/json-schema-to-grammar.py [changed mode: 0644->0755]
examples/make-ggml.py [changed mode: 0644->0755]
examples/reason-act.sh
examples/server-llama2-13B.sh [changed mode: 0644->0755]
examples/server/api_like_OAI.py
examples/server/chat-llama2.sh [changed mode: 0644->0755]
examples/server/chat.sh [changed mode: 0644->0755]
gguf.py [changed mode: 0644->0755]
scripts/get-wikitext-2.sh [changed mode: 0644->0755]

old mode 100644 (file)
new mode 100755 (executable)
old mode 100644 (file)
new mode 100755 (executable)
index b3e190a..50069db
@@ -1,3 +1,4 @@
+#!/usr/bin/env python3
 # HF falcon--> gguf conversion
 
 import gguf
old mode 100644 (file)
new mode 100755 (executable)
index a7cefc6..6eeff5b
@@ -1,3 +1,4 @@
+#!/usr/bin/env python3
 # HF gptneox--> gguf conversion
 
 import gguf
old mode 100644 (file)
new mode 100755 (executable)
index ab5c80b..f103f5f
@@ -1,3 +1,4 @@
+#!/usr/bin/env python3
 # 7b pth llama --> gguf conversion
 # Only models with a single datafile are supported, like 7B
 # HF files required in the model dir: config.json tokenizer_config.json tokenizer.json tokenizer.model
old mode 100644 (file)
new mode 100755 (executable)
index 86d4596..3bf9362
@@ -1,3 +1,4 @@
+#!/usr/bin/env python3
 import sys, struct, math, argparse
 from pathlib import Path
 
old mode 100644 (file)
new mode 100755 (executable)
index f8cfdaa..08fde23
@@ -1,3 +1,4 @@
+#!/usr/bin/env python3
 # HF llama --> gguf conversion
 
 import gguf
index b4999ff5a07c866b80f601d03fc7fe910ebdb7a3..04a7b8bbf4eaf803bd00fec983d8a75d160e40a8 100755 (executable)
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+#!/usr/bin/env python3
 import json
 import os
 import re
old mode 100644 (file)
new mode 100755 (executable)
index e720889..a701ab4
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+#!/usr/bin/env python3
 
 import gguf
 import argparse
old mode 100644 (file)
new mode 100755 (executable)
index be28966..f146acd
@@ -1,3 +1,4 @@
+#!/usr/bin/env python3
 import ctypes
 from ctypes import cdll, c_char_p, c_void_p, POINTER, c_float, c_int
 import numpy as np
old mode 100644 (file)
new mode 100755 (executable)
index bcbdd2b..06fad55
@@ -1,3 +1,4 @@
+#!/usr/bin/env python3
 import sys
 import os
 sys.path.insert(0, os.path.dirname(__file__))
old mode 100644 (file)
new mode 100755 (executable)
index 15c9b77..7b13e4a
@@ -1,3 +1,4 @@
+#!/usr/bin/env python3
 import sys
 import os
 sys.path.insert(0, os.path.dirname(__file__))
old mode 100644 (file)
new mode 100755 (executable)
index 0cfac5f..891ad7c
@@ -1,3 +1,4 @@
+#!/usr/bin/env python3
 import sys
 import os
 sys.path.insert(0, os.path.dirname(__file__))
old mode 100644 (file)
new mode 100755 (executable)
index 1b6c54b..8bc0706
@@ -1,3 +1,4 @@
+#!/usr/bin/env python3
 import matplotlib.pyplot as plt
 import os
 import csv
old mode 100644 (file)
new mode 100755 (executable)
old mode 100644 (file)
new mode 100755 (executable)
index 2dccc11..2a4cb65
@@ -1,3 +1,4 @@
+#!/usr/bin/env python3
 import argparse
 import json
 import re
old mode 100644 (file)
new mode 100755 (executable)
index f63d9fc..6a34eea
@@ -1,3 +1,4 @@
+#!/usr/bin/env python3
 """
 This script converts Hugging Face llama models to GGML and quantizes them.
 
index e7fe655dbcea3a724f575abc16a8ef6de8197b5f..046c48db584bc31f9c5fd57f84c21d84932a3511 100755 (executable)
@@ -1,4 +1,3 @@
-
 #!/bin/bash
 
 cd `dirname $0`
old mode 100644 (file)
new mode 100755 (executable)
index aa325a03ee444c9d141b1d35d4c44ce3979336ad..ed19237b0b3e5bb82aae4b45513ae7392f320fd3 100755 (executable)
@@ -1,3 +1,4 @@
+#!/usr/bin/env python3
 import argparse
 from flask import Flask, jsonify, request, Response
 import urllib.parse
old mode 100644 (file)
new mode 100755 (executable)
old mode 100644 (file)
new mode 100755 (executable)
diff --git a/gguf.py b/gguf.py
old mode 100644 (file)
new mode 100755 (executable)
index 4657467..9421080
--- a/gguf.py
+++ b/gguf.py
@@ -1,3 +1,4 @@
+#!/usr/bin/env python3
 import shutil
 import sys
 import struct
old mode 100644 (file)
new mode 100755 (executable)