v2.10.0
Huge update! This is likely going to be the last big 2.x update. This update greatly improves model implementation accuracy, and adds some of the newer Qwen models.
What's Changed
- Remove einsum in forward pass in AbstractAttention by @degenfabian in #783
- Colab compatibility bug fixes by @degenfabian in #794
- Remove einsum usage from create_alibi_bias function by @degenfabian in #781
- Actions token access by @bryce13950 in #797
- Remove einsum in apply_causal_mask in abstract_attention.py by @degenfabian in #782
- clarified arguments a bit for hook_points by @bryce13950 in #799
- Remove einsum in logit_attrs in ActivationCache by @degenfabian in #788
- Remove einsum in compute_head_results in ActivationCache by @degenfabian in #789
- Remove einsum usage in refactor_factored_attn_matrices in HookedTransformer by @degenfabian in #791
- Remove einsum usage in _get_w_in_matrix in SVDInterpreter by @degenfabian in #792
- Remove einsum usage in forward function of BertMLMHead by @degenfabian in #793
- Set default_prepend_bos to False in Bloom model configuration by @degenfabian in #806
- Remove einsum in complex_attn_linear by @degenfabian in #790
- Add a demo of collecting activations from a single location in the model. by @adamkarvonen in #807
- Add support for Qwen_with_Questions by @degenfabian in #811
- Added support for Qwen2.5 by @israel-adewuyi in #809
- Updated devcontainers to use python3.11 by @jonasrohw in #812
New Contributors
- @israel-adewuyi made their first contribution in #809
- @jonasrohw made their first contribution in #812
Full Changelog: v2.9.1...v2.10.0