Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What's happening in the example? #33

Open
startakovsky opened this issue Feb 17, 2024 · 3 comments
Open

What's happening in the example? #33

startakovsky opened this issue Feb 17, 2024 · 3 comments
Labels
documentation Improvements or additions to documentation

Comments

@startakovsky
Copy link

I didn't find the example that clear but I have a guess at what's happening. Might be worth spelling out something to the effect of trying to map text to an embedding and then right back to the original text, or perhaps a smattering of points in the pre-image of the embedding, not really sure because it's not clear from what's written.

My two cents.

@jxmorris12 jxmorris12 added the documentation Improvements or additions to documentation label Feb 17, 2024
@jxmorris12
Copy link
Owner

@startakovsky can you be more specific? which example, and what did you think is confusing

@startakovsky
Copy link
Author

Looks like the example takes in text and outputs text. What's happening?

vec2text has a function invert_strings which takes a list of strings and produces a list of strings.

The name of the function was confusing to me.

In my mind it's a misnomer if what is actually happening is:

  1. Input List of strings
  2. Produce embeddings associated to those strings
  3. Then run invert_embeddings under the hood

Maybe this is because this whole thing seems to be about:

$\mathcal{E}(strings) = embeddings

$\mathcal{E}^{-1}(embeddings) = strings

And so maybe what would be helpful is thinking about this like:

The goal of invert_strings is to find similar strings. The way we do that is we embed each input, the running our algorithm to find the inverse of the embedding, landing on a semantically similar list of strings

The goal of inverse embedding is to find strings, when embedded that produce the embeddings.

@startakovsky
Copy link
Author

@jxmorris12 does that help?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
documentation Improvements or additions to documentation
Projects
None yet
Development

No branches or pull requests

2 participants