Skip to content

IndexError in run_encoding.py #2

@amansahu278

Description

@amansahu278

Hello,
While working from the main branch in my fork,
I first generate the KV Cache and then I run the encoding where in I get the error

Traceback (most recent call last):
File "/home/am_sahu/./CacheGenExp/run_encoding.py", line 132, in
encode_function(
File "/home/am_sahu/./CacheGenExp/run_encoding.py", line 119, in encode_function
encode_input[l:l+1, i].to(torch.int16) )
IndexError: index 6478 is out of bounds for dimension 1 with size 6478

Here is the relevant output (which i consider is relevant)

Loading extension module torchac_backend...
You are using the default legacy behaviour of the <class 'transformers.models.llama.tokenization_llama.LlamaTokenizer'>. This is expected, and simply means that the legacy (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set legacy=False. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in huggingface/transformers#24565
100%|██████████| 2/2 [00:51<00:00, 25.93s/it]
Model and tokenizer loaded
Length of input: torch.Size([1, 6479])
TTFT: 1.7883199411444366
KV Cache generated
Using /home/am_sahu/.cache/torch_extensions/py310_cu121 as PyTorch extensions root...
Emitting ninja build file /home/am_sahu/.cache/torch_extensions/py310_cu121/torchac_backend/build.ninja...
Building extension module torchac_backend...
Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N)
ninja: no work to do.
Loading extension module torchac_backend...
75%|███████▌ | 3/4 [04:33<01:31, 91.03s/it]
Done with layer 0
Done with layer 1
Done with layer 2
Done with layer 3
Done with layer 4
Done with layer 5
Done with layer 6
Done with layer 7
Done with layer 8
Done with layer 9
Done with layer 10
Done with layer 11
Done with layer 12
Done with layer 13
Done with layer 14
Done with layer 15
Done with layer 16
Done with layer 17
Done with layer 18
Done with layer 19
Done with layer 20
Done with layer 21
Done with layer 22
Done with layer 23
Done with layer 24
Done with layer 25
Done with layer 26
Done with layer 27
Done with layer 28
Done with layer 29
Done with layer 30
Done with layer 31
Done with layer 32
Done with layer 33
Done with layer 34
Done with layer 35
Done with layer 36
Done with layer 37
Done with layer 38
Done with layer 39
Done with layer 40
Done with layer 41
Done with layer 42
Done with layer 43
Done with layer 44
Done with layer 45
Done with layer 46
Done with layer 47
Done with layer 48
Done with layer 49
Done with layer 50
Done with layer 51
Done with layer 52
Done with layer 53
Done with layer 54
Done with layer 55
Done with layer 56
Done with layer 57
Done with layer 58
Done with layer 59
Done with layer 60
Done with layer 61
Done with layer 62
Done with layer 63
Done with layer 0
Done with layer 1
Done with layer 2
Done with layer 3
Done with layer 4
Done with layer 5
Done with layer 6
Done with layer 7
Done with layer 8
Done with layer 9
Done with layer 10
Done with layer 11
Done with layer 12
Done with layer 13
Done with layer 14
Done with layer 15
Done with layer 16
Done with layer 17
Done with layer 18
Done with layer 19
Done with layer 20
Done with layer 21
Done with layer 22
Done with layer 23
Done with layer 24
Done with layer 25
Done with layer 26
Done with layer 27
Done with layer 28
Done with layer 29
Done with layer 30
Done with layer 31
Done with layer 32
Done with layer 33
Done with layer 34
Done with layer 35
Done with layer 36
Done with layer 37
Done with layer 38
Done with layer 39
Done with layer 40
Done with layer 41
Done with layer 42
Done with layer 43
Done with layer 44
Done with layer 45
Done with layer 46
Done with layer 47
Done with layer 48
Done with layer 49
Done with layer 50
Done with layer 51
Done with layer 52
Done with layer 53
Done with layer 54
Done with layer 55
Done with layer 56
Done with layer 57
Done with layer 58
Done with layer 59
Done with layer 60
Done with layer 61
Done with layer 62
Done with layer 63
Done with layer 0
Done with layer 1
Done with layer 2
Done with layer 3
Done with layer 4
Done with layer 5
Done with layer 6
Done with layer 7
Done with layer 8
Done with layer 9
Done with layer 10
Done with layer 11
Done with layer 12
Done with layer 13
Done with layer 14
Done with layer 15
Done with layer 16
Done with layer 17
Done with layer 18
Done with layer 19
Done with layer 20
Done with layer 21
Done with layer 22
Done with layer 23
Done with layer 24
Done with layer 25
Done with layer 26
Done with layer 27
Done with layer 28
Done with layer 29
Done with layer 30
Done with layer 31
Done with layer 32
Done with layer 33
Done with layer 34
Done with layer 35
Done with layer 36
Done with layer 37
Done with layer 38
Done with layer 39
Done with layer 40
Done with layer 41
Done with layer 42
Done with layer 43
Done with layer 44
Done with layer 45
Done with layer 46
Done with layer 47
Done with layer 48
Done with layer 49
Done with layer 50
Done with layer 51
Done with layer 52
Done with layer 53
Done with layer 54
Done with layer 55
Done with layer 56
Done with layer 57
Done with layer 58
Done with layer 59
Done with layer 60
Done with layer 61
Done with layer 62
Done with layer 63
Done with layer 0
Traceback (most recent call last):
File "/home/am_sahu/./CacheGenExp/run_encoding.py", line 132, in
encode_function(
File "/home/am_sahu/./CacheGenExp/run_encoding.py", line 119, in encode_function
encode_input[l:l+1, i].to(torch.int16) )
IndexError: index 6478 is out of bounds for dimension 1 with size 6478

The following are the commands and configurations i used
For generating the KV cache
python $GENCACHE_PATH/main.py \ --save_dir $GENCACHE_DATA_PATH/kvcache \ --path_to_context $GENCACHE_PATH/7k_prompts/1.txt \ && echo "KV Cache generated"

For Encoding
python $GENCACHE_PATH/run_encoding.py \ --output_path $GENCACHE_DATA_PATH/encoded \ --path_to_kv $GENCACHE_DATA_PATH/kvcache/test_kv_0.pkl \ --quantization_config $GENCACHE_PATH/config/quantization_7b.json \ && echo "Encoding done"

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions