attention Building out an attention block in rtl starting assumptions: batch size of 1 sequence length of 4 dimension = 8 number system: int8 going for simplicity first