Skip to content

Subselect probes by input length #1123

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
erickgalinkin opened this issue Mar 7, 2025 · 2 comments
Open

Subselect probes by input length #1123

erickgalinkin opened this issue Mar 7, 2025 · 2 comments
Labels
architecture Architectural upgrades generators Interfaces with LLMs

Comments

@erickgalinkin
Copy link
Collaborator

Summary

Some targets will have artificial limits on input length that are independent of the model (e.g. a web frontend that allows only n characters/words of input)

Motivation

Running full sets of probes against these targets is necessarily going to be wasteful and will not indicate anything truly about robustness. If we subselect by length, we can reduce load and improve accuracy.

@mrowebot
Copy link
Contributor

Do we have a discrete list of such targets that can have their input lengths capped?

@leondz leondz added architecture Architectural upgrades generators Interfaces with LLMs labels Apr 23, 2025
@leondz
Copy link
Collaborator

leondz commented Apr 23, 2025

Interesting feature. Are there concrete examples of this?

Do we have a discrete list of such targets that can have their input lengths capped?

Not really - some are manually tracked in the openai module

Some targets will have artificial limits on input length that are independent of the model (e.g. a web frontend that allows only n characters/words of input)

This sounds like it requires three ingredients

  1. Knowledge of the max length, maybe set by config or a generator attrib
  2. Knowledge of prompt length, available after prompt is composed, requiring a tokenizer / estimation. A pattern will emerge with estimate token use before sending openai completions #1112
  3. Orchestration-level intervention to not pose the prompt. This could be represented as prompt:whatever output:None, which will come back as a skip - that seems appropriate to me, the prompt is skipped.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
architecture Architectural upgrades generators Interfaces with LLMs
Projects
None yet
Development

No branches or pull requests

3 participants