Skip to content

Latest commit

History

History
23 lines (12 loc) 路 1.26 KB

CHANGELOG.md

File metadata and controls

23 lines (12 loc) 路 1.26 KB

Changelog

This file contains a high-level description of changes that were merged into the Inseq main branch since the last release. Refer to the releases page for an exhaustive overview of changes introduced at each release.

馃殌 Features

  • Added new models DbrxForCausalLM, OlmoForCausalLM, Phi3ForCausalLM, Qwen2MoeForCausalLM to model config.

馃敡 Fixes and Refactoring

  • Fix the issue in the attention implementation from #268 where non-terminal position in the tensor were set to nan if they were 0s (#269).

  • Fix the pad token in cases where it is not specified by default in the loaded model (e.g. for Qwen models) (#269).

  • Fix bug reported in #266 making value_zeroing unusable for SDPA attention. This enables using the method on models using SDPA attention as default (e.g. GemmaForCausalLM) without passing model_kwargs={'attn_implementation': 'eager'} (#267).

馃摑 Documentation and Tutorials

No changes

馃挜 Breaking Changes

No changes