| license: apache-2.0 | |
| library_name: transformers | |
| tags: | |
| - code | |
| - granite | |
| - llama-cpp | |
| - gguf-my-repo | |
| datasets: | |
| - codeparrot/github-code-clean | |
| - bigcode/starcoderdata | |
| - open-web-math/open-web-math | |
| - math-ai/StackMathQA | |
| metrics: | |
| - code_eval | |
| pipeline_tag: text-generation | |
| inference: true | |
| model-index: | |
| - name: granite-20b-code-base | |
| results: | |
| - task: | |
| type: text-generation | |
| dataset: | |
| name: MBPP | |
| type: mbpp | |
| metrics: | |
| - type: pass@1 | |
| value: 43.8 | |
| name: pass@1 | |
| - task: | |
| type: text-generation | |
| dataset: | |
| name: MBPP+ | |
| type: evalplus/mbppplus | |
| metrics: | |
| - type: pass@1 | |
| value: 51.6 | |
| name: pass@1 | |
| - task: | |
| type: text-generation | |
| dataset: | |
| name: HumanEvalSynthesis(Python) | |
| type: bigcode/humanevalpack | |
| metrics: | |
| - type: pass@1 | |
| value: 48.2 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 50.0 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 59.1 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 32.3 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 40.9 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 35.4 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 17.1 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 18.3 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 23.2 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 10.4 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 25.6 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 18.3 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 23.2 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 23.8 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 14.6 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 26.2 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 15.2 | |
| name: pass@1 | |
| - type: pass@1 | |
| value: 3.0 | |
| name: pass@1 | |
| # cobrakenji/granite-20b-code-base-Q3_K_M-GGUF | |
| This model was converted to GGUF format from [`ibm-granite/granite-20b-code-base`](https://huggingface.co/ibm-granite/granite-20b-code-base) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space. | |
| Refer to the [original model card](https://huggingface.co/ibm-granite/granite-20b-code-base) for more details on the model. | |
| ## Use with llama.cpp | |
| Install llama.cpp through brew. | |
| ```bash | |
| brew install ggerganov/ggerganov/llama.cpp | |
| ``` | |
| Invoke the llama.cpp server or the CLI. | |
| CLI: | |
| ```bash | |
| llama-cli --hf-repo cobrakenji/granite-20b-code-base-Q3_K_M-GGUF --model granite-20b-code-base.Q3_K_M.gguf -p "The meaning to life and the universe is" | |
| ``` | |
| Server: | |
| ```bash | |
| llama-server --hf-repo cobrakenji/granite-20b-code-base-Q3_K_M-GGUF --model granite-20b-code-base.Q3_K_M.gguf -c 2048 | |
| ``` | |
| Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well. | |
| ``` | |
| git clone https://github.com/ggerganov/llama.cpp && cd llama.cpp && make && ./main -m granite-20b-code-base.Q3_K_M.gguf -n 128 | |
| ``` | |