Skip to content

[P1] Support more huggingface (transformer-based) models #46

@frankaging

Description

@frankaging

Descriptions:
Ideally, all the models listed here can be supported by this library without exposing the model details to the users of this library.

This requires we set up model folders for all model types and write config metadata for each of them annotating where to do interventions. This requires a lot of effort. This is a PR tracking the process towards the goal of supporting as many as we can.

Each model should take less than an hour to (1) configure and (2) write simple unit tests.

Here is the list of models that are in the pipeline to support (in order):

  • BERT-family
    • RoBERTa
    • DeBERTa
    • ELECTRA
  • xlm (multilingual model)
  • t5
  • Mistral
  • Mixtral (MoE, MixtralForCausalLM)
  • Phi
  • Mamba (but need to support recurrent interventions, not just layerwise interventions)
  • backpack-gpt2
  • please feel free to suggest other new models to support!

Metadata

Metadata

Labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions