Remove the setting of _attn_implementation from llama_bidirectional_model
#3
by
nvidia-oliver-holworthy
- opened
Remove _attn_implementation from LlamaBidirectionalModel constructor.
Following the transformers 4.48.0 release which inclided a refactor of attention implementations. (transformers/pull/35235🚨All attention refactor🚨), In transformers versions 4.47.0 this line did not have an impact on the attention implementation. Following 4.48.0 this line was activated by the new attention initialization and activated "eager" attention instead of "sdpa".