中文

Chat & support: my new Discord server

Want to contribute? TheBloke's Patreon page

Pankaj Mathur's Orca Mini 7B GPTQ

These files are GPTQ 4bit model files for Pankaj Mathur's Orca Mini 7B .

It is the result of quantising to 4bit using GPTQ-for-LLaMa .

Repositories available

Prompt template:

### System:
You are an AI assistant that follows instruction extremely well. Help as much as you can.

### User:
prompt

### Response:

or

### System:
You are an AI assistant that follows instruction extremely well. Help as much as you can.

### User:
prompt

### Input:
input

### Response:

How to easily download and use this model in text-generation-webui

Please make sure you're using the latest version of text-generation-webui

  • Click the Model tab .
  • Under Download custom model or LoRA , enter TheBloke/orca_mini_7B-GPTQ .
  • Click Download .
  • The model will start downloading. Once it's finished it will say "Done"
  • In the top left, click the refresh icon next to Model .
  • In the Model dropdown, choose the model you just downloaded: orca_mini_7B-GPTQ
  • The model will automatically load, and is now ready for use!
  • If you want any custom settings, set them and then click Save settings for this model followed by Reload the Model in the top right.
    • Note that you do not need to and should not set manual GPTQ parameters any more. These are set automatically from the file quantize_config.json .
  • Once you're ready, click the Text Generation tab and enter a prompt to get started!
  • How to use this GPTQ model from Python code

    First make sure you have AutoGPTQ installed:

    pip install auto-gptq

    Then try the following example code:

    from transformers import AutoTokenizer, pipeline, logging
    from auto_gptq import AutoGPTQForCausalLM, BaseQuantizeConfig
    import argparse
    
    model_name_or_path = "TheBloke/orca_mini_7B-GPTQ"
    model_basename = "orca-mini-7b-GPTQ-4bit-128g.no-act.order"
    
    use_triton = False
    
    tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
    
    model = AutoGPTQForCausalLM.from_quantized(model_name_or_path,
            model_basename=model_basename,
            use_safetensors=True,
            trust_remote_code=False,
            device="cuda:0",
            use_triton=use_triton,
            quantize_config=None)
    
    # Note: check the prompt template is correct for this model.
    prompt = "Tell me about AI"
    prompt_template=f'''USER: {prompt}
    ASSISTANT:'''
    
    print("\n\n*** Generate:")
    
    input_ids = tokenizer(prompt_template, return_tensors='pt').input_ids.cuda()
    output = model.generate(inputs=input_ids, temperature=0.7, max_new_tokens=512)
    print(tokenizer.decode(output[0]))
    
    # Inference can also be done using transformers' pipeline
    
    # Prevent printing spurious transformers error when using pipeline with AutoGPTQ
    logging.set_verbosity(logging.CRITICAL)
    
    print("*** Pipeline:")
    pipe = pipeline(
        "text-generation",
        model=model,
        tokenizer=tokenizer,
        max_new_tokens=512,
        temperature=0.7,
        top_p=0.95,
        repetition_penalty=1.15
    )
    
    print(pipe(prompt_template)[0]['generated_text'])
    

    Provided files

    orca-mini-7b-GPTQ-4bit-128g.no-act.order.safetensors

    This will work with AutoGPTQ, ExLlama, and CUDA versions of GPTQ-for-LLaMa. There are reports of issues with Triton mode of recent GPTQ-for-LLaMa. If you have issues, please use AutoGPTQ instead.

    It was created with group_size 128 to increase inference accuracy, but without --act-order (desc_act) to increase compatibility and improve inference speed.

    • orca-mini-7b-GPTQ-4bit-128g.no-act.order.safetensors
      • Works with AutoGPTQ in CUDA or Triton modes.
      • LLaMa models also work with [ExLlama]( https://github.com/turboderp/exllama} , which usually provides much higher performance, and uses less VRAM, than AutoGPTQ.
      • Works with GPTQ-for-LLaMa in CUDA mode. May have issues with GPTQ-for-LLaMa Triton mode.
      • Works with text-generation-webui, including one-click-installers.
      • Parameters: Groupsize = 128. Act Order / desc_act = False.

    Discord

    For further support, and discussions on these models and AI in general, join us at:

    TheBloke AI's Discord server

    Thanks, and how to contribute.

    Thanks to the chirper.ai team!

    I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training.

    If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects.

    Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits.

    Special thanks to : Luke from CarbonQuill, Aemon Algiz, Dmitriy Samsonov.

    Patreon special mentions : Pyrater, WelcomeToTheClub, Kalila, Mano Prime, Trenton Dambrowitz, Spiking Neurons AB, Pierre Kircher, Fen Risland, Kevin Schuppel, Luke, Rainer Wilmers, vamX, Gabriel Puliatti, Alex , Karl Bernard, Ajan Kanaga, Talal Aujan, Space Cruiser, ya boyyy, biorpg, Johann-Peter Hartmann, Asp the Wyvern, Ai Maven, Ghost , Preetika Verma, Nikolai Manek, trip7s trip, John Detwiler, Fred von Graf, Artur Olbinski, subjectnull, John Villwock, Junyu Yang, Rod A, Lone Striker, Chris McCloskey, Iucharbius , Matthew Berman, Illia Dulskyi, Khalefa Al-Ahmad, Imad Khwaja, chris gileta, Willem Michiel, Greatston Gnanesh, Derek Yates, K, Alps Aficionado, Oscar Rangel, David Flickinger, Luke Pendergrass, Deep Realms, Eugene Pentland, Cory Kujawski, terasurfer , Jonathan Leane, senxiiz, Joseph William Delisle, Sean Connelly, webtim, zynix , Nathan LeClaire.

    Thank you to all my generous patrons and donaters!

    Original model card: Pankaj Mathur's Orca Mini 7B

    orca_mini_7b

    An OpenLLaMa-7B model model trained on explain tuned datasets, created using Instructions and Input from WizardLM, Alpaca & Dolly-V2 datasets and applying Orca Research Paper dataset construction approaches.

    Dataset

    We build explain tuned WizardLM dataset ~70K , Alpaca dataset ~52K & Dolly-V2 dataset ~15K created using approaches from Orca Research Paper .

    We leverage all of the 15 system instructions provided in Orca Research Paper. to generate custom datasets, in contrast to vanilla instruction tuning approaches used by original datasets.

    This helps student model aka this model to learn thought process from teacher model, which is ChatGPT (gpt-3.5-turbo-0301 version).

    Please see below example usage how the System prompt is added before each instruction .

    Training

    The training configurations are provided in the table below.

    The training takes on 8x A100(80G) GPUs and lasts for around 7 Hours for cost of $84 using Lambda Labs

    We used DeepSpeed with fully sharded data parallelism, also know as ZeRO stage 3 by writing our own fine tunning scripts plus leveraging some of the model training code provided by amazing OpenAlpaca repo

    Here are some of params used during training:

    batch_size 32
    train_micro_batch_size_per_gpu 2
    gradient_accumulation_steps 2
    Learning rate 2e-5
    Max length 1024
    Epochs 3
    Optimizer AdamW

    Example Usage

    Below shows an example on how to use this model

    import torch
    from transformers import LlamaForCausalLM, LlamaTokenizer
    
    # Hugging Face model_path
    model_path = 'psmathur/orca_mini_7b'
    tokenizer = LlamaTokenizer.from_pretrained(model_path)
    model = LlamaForCausalLM.from_pretrained(
        model_path, torch_dtype=torch.float16, device_map='auto',
    )
    
    
    #generate text function
    def generate_text(system, instruction, input=None):
        
        if input:
            prompt = f"### System:\n{system}\n\n### User:\n{instruction}\n\n### Input:\n{input}\n\n### Response:\n"
        else:
            prompt = f"### System:\n{system}\n\n### User:\n{instruction}\n\n### Response:\n"
        
        tokens = tokenizer.encode(prompt)
        tokens = torch.LongTensor(tokens).unsqueeze(0)
        tokens = tokens.to('cuda')
    
        instance = {'input_ids': tokens,'top_p': 1.0, 'temperature':0.7, 'generate_len': 1024, 'top_k': 50}
    
        length = len(tokens[0])
        with torch.no_grad():
            rest = model.generate(
                input_ids=tokens, 
                max_length=length+instance['generate_len'], 
                use_cache=True, 
                do_sample=True, 
                top_p=instance['top_p'],
                temperature=instance['temperature'],
                top_k=instance['top_k']
            )    
        output = rest[0][length:]
        string = tokenizer.decode(output, skip_special_tokens=True)
        return f'[!] Response: {string}'
    
    # Sample Test Instruction Used by Youtuber Sam Witteveen https://www.youtube.com/@samwitteveenai
    system = 'You are an AI assistant that follows instruction extremely well. Help as much as you can.'
    instruction = 'Write a letter to Sam Altman, CEO of OpenAI, requesting him to convert GPT4 a private model by OpenAI to an open source project'
    print(generate_text(system, instruction))
    
    [!] Response:
    Dear Sam Altman,
    
    I am writing to request that you convert the GPT4 private model developed by OpenAI to an open source project. As a user of OpenAI, I have been waiting for the day when I can use the advanced natural language processing capabilities of GPT4 in a more open and accessible way.
    
    While OpenAI has made significant progress in developing AI applications, it has primarily focused on building private models that are not accessible to the general public. However, with the recent release of GPT-3, there is a growing demand for more open and accessible AI tools.
    
    Converting GPT4 to an open source project would allow for greater transparency, collaboration, and innovation. It would also help to build trust in the technology and ensure that it is used ethically and responsibly.
    
    I urge you to consider converting GPT4 to an open source project. This would be a significant contribution to the AI community and would help to create a more open and accessible future.
    
    Thank you for your consideration.
    
    Sincerely,
    
    [Your Name]
    

    P.S. I am #opentowork and #collaboration, if you can help, please reach out to me at psmathur.public@gmail.com

    Next Goals:

  • Try more data like actually using FLAN-v2, just like Orka Research Paper (I am open for suggestions)
  • Provide more options for Text generation UI. (may be https://github.com/oobabooga/text-generation-webui )
  • Provide 4bit GGML/GPTQ quantized model (may be TheBloke can help here)
  • Limitations & Biases:

    This model can produce factually incorrect output, and should not be relied on to produce factually accurate information. This model was trained on various public datasets. While great efforts have been taken to clean the pretraining data, it is possible that this model could generate lewd, biased or otherwise offensive outputs.

    Disclaimer:

    The license on this model does not constitute legal advice. We are not responsible for the actions of third parties who use this model. Please cosult an attorney before using this model for commercial purposes.

    Citiation:

    If you found wizardlm_alpaca_dolly_orca_open_llama_7b useful in your research or applications, please kindly cite using the following BibTeX:

    @misc{wizardlm_alpaca_dolly_orca_open_llama_7b,
      author = {Pankaj Mathur},
      title = {wizardlm_alpaca_dolly_orca_open_llama_7b: An explain tuned OpenLLaMA-7b model on custom wizardlm, alpaca, & dolly datasets},
      year = {2023},
      publisher = {GitHub, HuggingFace},
      journal = {GitHub repository, HuggingFace repository},
      howpublished = {\url{https://github.com/pankajarm/wizardlm_alpaca_dolly_orca_open_llama_7b}, \url{https://https://huggingface.co/psmathur/wizardlm_alpaca_dolly_orca_open_llama_7b}},
    }
    
    @software{openlm2023openllama,
      author = {Xinyang Geng and Hao Liu},
      title = {OpenLLaMA: An Open Reproduction of LLaMA},
      month = May,
      year = 2023,
      url = {https://github.com/openlm-research/open_llama}
    }
    
    @misc{openalpaca,
      author = {Yixuan Su and Tian Lan and Deng Cai},
      title = {OpenAlpaca: A Fully Open-Source Instruction-Following Model Based On OpenLLaMA},
      year = {2023},
      publisher = {GitHub},
      journal = {GitHub repository},
      howpublished = {\url{https://github.com/yxuansu/OpenAlpaca}},
    }
    
    @misc{alpaca,
      author = {Rohan Taori and Ishaan Gulrajani and Tianyi Zhang and Yann Dubois and Xuechen Li and Carlos Guestrin and Percy Liang and Tatsunori B. Hashimoto },
      title = {Stanford Alpaca: An Instruction-following LLaMA model},
      year = {2023},
      publisher = {GitHub},
      journal = {GitHub repository},
      howpublished = {\url{https://github.com/tatsu-lab/stanford_alpaca}},
    }