{"author":"chenyu","author_email":"chenyu@fastmail.com","author_time":1704834895,"commit_time":1704834895,"committer":"GitHub","committer_email":"noreply@github.com","hash":"f0d7ad8aaafdb1daae0f0939897a48515fa77027","message":"fix gpt2 attention with start_pos = 0 (#3061)\n\n* fix gpt2 attention with start_pos size 1\r\n\r\ntest cases taken from ll_transformer branch\r\n\r\n* fix interpreted","parents":["39b91131bc0ae0ba6184af0f007b454985a8e1c5"],"tree_hash":"33475185008acb8b638a96e7b1c6d972d00fde91"}