+#!/usr/bin/env python3
# HF falcon--> gguf conversion
import gguf
+#!/usr/bin/env python3
# HF gptneox--> gguf conversion
import gguf
+#!/usr/bin/env python3
# 7b pth llama --> gguf conversion
# Only models with a single datafile are supported, like 7B
# HF files required in the model dir: config.json tokenizer_config.json tokenizer.json tokenizer.model
+#!/usr/bin/env python3
import sys, struct, math, argparse
from pathlib import Path
+#!/usr/bin/env python3
# HF llama --> gguf conversion
import gguf
-#!/usr/bin/env python
+#!/usr/bin/env python3
import json
import os
import re
-#!/usr/bin/env python
+#!/usr/bin/env python3
import gguf
import argparse
+#!/usr/bin/env python3
import ctypes
from ctypes import cdll, c_char_p, c_void_p, POINTER, c_float, c_int
import numpy as np
+#!/usr/bin/env python3
import sys
import os
sys.path.insert(0, os.path.dirname(__file__))
+#!/usr/bin/env python3
import sys
import os
sys.path.insert(0, os.path.dirname(__file__))
+#!/usr/bin/env python3
import sys
import os
sys.path.insert(0, os.path.dirname(__file__))
+#!/usr/bin/env python3
import matplotlib.pyplot as plt
import os
import csv
+#!/usr/bin/env python3
import argparse
import json
import re
+#!/usr/bin/env python3
"""
This script converts Hugging Face llama models to GGML and quantizes them.
-
#!/bin/bash
cd `dirname $0`
+#!/usr/bin/env python3
import argparse
from flask import Flask, jsonify, request, Response
import urllib.parse
+#!/usr/bin/env python3
import shutil
import sys
import struct