8000 GitHub - lorenzocesconetto/gemma-intro: Introduction to Gemma, the new open source LLM family by Google
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

lorenzocesconetto/gemma-intro

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Gemma Intro

Authors: Lorenzo Cesconetto, Pedro Gengo.

Running with llama.cpp

  • Pre-requisite: You must have llama.cpp already installed and setup.

  • Download from HF (please see HF cli download reference):

pip install "huggingface_hub[hf_transfer]" # hf_transfer enables faster download
HF_HUB_ENABLE_HF_TRANSFER=1 huggingface-cli download google/gemma-2b-it gemma-2b-it.gguf # must set HF_HUB_ENABLE_HF_TRANSFER=1 for a faster download
  • Quantized model:
./llama.cpp/quantize ../models_converted/gemma-2b-it.gguf ../models_quantized/gemma-2b-it-q8.gguf q8_0
  • Running:
./llama.cpp/main 2>/dev/null \
    --model ./models_quantized/gemma-2b-it-q8.gguf \
    --ctx-size 8192 \
    --batch-size 8 \
    --keep -1 \
    --n-predict -1 \
    --color \
    --interactive-first \
    --temp 0.1 \
    --reverse-prompt "<start_of_turn>user " \
    --in-suffix "<start_of_turn>model " \
    --in-prefix "<end_of_turn> "

About

Introduction to Gemma, the new open source LLM family by Google

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •  
0