Skip to content

Instantly share code, notes, and snippets.

View vuiseng9's full-sized avatar

VS (Vui Seng Chua) vuiseng9

View GitHub Profile
@vuiseng9
vuiseng9 / a-to-z.md
Last active September 16, 2024 02:24
Set 1 Set 2 Set 3 Least Character Word
Apex Aurora Atom Apex
Blaze Blitz Bolt Bolt
Cipher Cascade Cobalt Cipher
Dynamo Drift Drift Drift
Echo Eclipse Ember Echo
Flux Flare Frost Flux
Gravitas Glide Grit Grit
Helix Horizon Haven Helix

Install

https://github.com/state-spaces/mamba

Run

from transformers import MambaConfig, MambaForCausalLM, AutoTokenizer, AutoModelForCausalLM
import torch
from functools import partial
from collections import OrderedDict, defaultdict
import os

Install

pip install transformers torch accelerate

Run

from transformers import AutoTokenizer, AutoModelForCausalLM, set_seed
import torch
from torch.profiler import profile, ProfilerActivity
$ python bench_linear.py --bs 1
BS:    1, Latency:    0.389 ms, IC:  4096, OC: 11008, Samples: 100, Warmup: 10

$ python bench_linear.py --bs 128
BS:  128, Latency:    3.640 ms, IC:  4096, OC: 11008, Samples: 100, Warmup: 10

$ python bench_linear.py --bs 1024
BS: 1024, Latency:   41.244 ms, IC:  4096, OC: 11008, Samples: 100, Warmup: 10
import intel_extension_for_pytorch # requried for XPU
import torch
from bigdl.llm.transformers import AutoModelForCausalLM
from transformers import AutoTokenizer, pipeline

# model_id = "facebook/opt-1.3b"
# model_id = "meta-llama/Llama-2-7b"
model_id = "meta-llama/Llama-2-7b-chat-hf"
prompt = "I love the Avengers,"
import warnings
from transformers import AutoTokenizer

class PromptCreator:
    def __init__(self, model_id):
        self.tokenizer = AutoTokenizer.from_pretrained(model_id)
        self.offset = len(self.tokenizer(self.tokenizer.special_tokens_map['bos_token'])['input_ids'])
        self.samples = [
                                {

Install

git clone https://github.com/bigcode-project/bigcode-evaluation-harness
pip install -e .

Deterministic Generation

mistralai/Mistral-7B-v0.1 should result in "pass@1": 0.29878 paper: 30.5%, 0.7% gap

accelerate launch $WORKDIR/main.py \
@vuiseng9
vuiseng9 / watched_jira.md
Created November 2, 2023 16:16
jira ql to filter watched issues
import os
from huggingface_hub import snapshot_download
REPO_ID = "repo_id"
LOCAL_ROOT= "/hf-model"
LOCAL_DIR = os.path.join(LOCAL_ROOT, os.path.basename(REPO_ID))
snapshot_download(repo_id=REPO_ID, local_dir=LOCAL_DIR, local_dir_use_symlinks=False)