Skip to content

Consolidate LLM attr logic #1420

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 8 commits into from

Conversation

craymichael
Copy link
Contributor

Summary: Add base class BaseLLMAttribution to consolidate repeat logic between perturbation/gradient-based LLM attr classes

Differential Revision: D65008854

@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D65008854

Summary:

Slight bug in the retry GitHub workflow causes failing workflows to rerun indefinitely. This _should_ fix it

Differential Revision: D65008676
Summary:

Add __call__ to TokenizerLike for transformers compatibility

Differential Revision: D64998805
Summary:

Use the __call__ method of tokenizers that returns a BatchEncoding with offsets. This allows us to grab text from the fully decoded string and not make assumptions about how many tokens correspond to a single string.

Differential Revision: D64998804
Summary:

visualize_image_attr_multiple can return a List[Axes], adds proper annotations to satisfy mypy


Why casting is necessary: numpy/numpy#24738

https://github.com/matplotlib/matplotlib/blob/v3.9.2/lib/matplotlib/pyplot.py#L1583C41-L1584C1

Differential Revision: D64998799
Summary:

Adds enough typing to get rid of
`captum/metrics/_core/infidelity.py:498: note: By default the bodies of untyped functions are not checked, consider using --check-untyped-defs  [annotation-unchecked]`

Differential Revision: D64998800
Summary:

Fix pyre/mypy errors in infidelity.py. Introduce new BaselineTupleType

Differential Revision: D64998803
Summary: Fix incorrect **kwargs annotation

Differential Revision: D65001879
Summary:

Add base class BaseLLMAttribution to consolidate repeat logic between perturbation/gradient-based LLM attr classes

Differential Revision: D65008854
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D65008854

craymichael added a commit to craymichael/captum that referenced this pull request Oct 26, 2024
Summary:

Add base class BaseLLMAttribution to consolidate repeat logic between perturbation/gradient-based LLM attr classes

Differential Revision: D65008854
craymichael added a commit to craymichael/captum that referenced this pull request Oct 26, 2024
Summary:

Add base class BaseLLMAttribution to consolidate repeat logic between perturbation/gradient-based LLM attr classes

Differential Revision: D65008854
@facebook-github-bot
Copy link
Contributor

This pull request has been merged in 638b920.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants