Skip to content

Fixes in RL SCHEMA #31

@karthikviswanathn

Description

@karthikviswanathn
  1. Finding input prompt length in RL_SCHEMA after tokenizer fix
  2. Generate attention mask for generate_k_completions so that the later functions can use the mask.
  3. Fix the famous indexing issue:
student_output_scores, collected_exit_logits = student(completions['tokens'][:, :-1], 
                                                       prescribed_exit_layer_idxs = prescribed_exit_layers[:,1:]) 

Metadata

Metadata

Labels

No labels
No labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions