diff --git a/common/recursion.py b/common/recursion.py index 05792b0..c7ebe0c 100644 --- a/common/recursion.py +++ b/common/recursion.py @@ -17,7 +17,7 @@ def fab(n): if n <= 2: return n else: - return fab(n - 1) + fab(n - 2) + return fab(n - 1) + fab(n - 2) if __name__ == '__main__': diff --git a/examples/llm/alpaca_ft.ipynb b/examples/llm/alpaca_ft.ipynb index 05c813b..986279d 100644 --- a/examples/llm/alpaca_ft.ipynb +++ b/examples/llm/alpaca_ft.ipynb @@ -290,7 +290,7 @@ "name": "python", "nbconvert_exporter": "python", "pygments_lexer": "ipython3", - "version": "3.9.16" + "version": "3.9.16 (main, Mar 8 2023, 04:29:24) \n[Clang 14.0.6 ]" }, "orig_nbformat": 4, "vscode": { diff --git "a/images/\346\274\224\347\244\272.gif" "b/images/\346\274\224\347\244\272.gif" new file mode 100644 index 0000000..22bab31 Binary files /dev/null and "b/images/\346\274\224\347\244\272.gif" differ diff --git a/llmodels/vicuna/qk.py b/llmodels/vicuna/qk.py new file mode 100644 index 0000000..422a9f8 --- /dev/null +++ b/llmodels/vicuna/qk.py @@ -0,0 +1,22 @@ +#!/usr/bin/env python3 +# -*- coding:utf-8 -*- + +import torch +from transformers import LlamaForCausalLM, LlamaTokenizer +from transformers import pipeline + +BASE_MODEL = "../vicuna-13b" +model = LlamaForCausalLM.from_pretrained( + BASE_MODEL, + device_map="auto", + offload_folder="./data/vicuna", +) + +tokenizer = LlamaTokenizer.from_pretrained(BASE_MODEL) + +text = "Generative AI is" +pipe = pipeline("text-generation", model=model, tokenizer=tokenizer) +generated_text = pipe(text, max_length=50, do_sample=False, no_repeat_ngram_size=2)[0] +print(generated_text['generated_text']) + +