-
Notifications
You must be signed in to change notification settings - Fork 285
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[MIEB] Add new multimodal retrieval tasks #1523
Comments
Hey @izhx! Thanks for reaching out. Tagging @gowitheflow-1998 here as well. We're working on integrating MIEB docs with MTEB at the moment. I think the general steps are:
|
Thanks for reaching out @izhx - can you send the reference paper (can't seem to find the paper with that specific table) |
Hi, we will submit the paper to arxiv and open-source the models in about 10 days and are still finalizing the results. @KennethEnevoldsen In addition, I checked the |
Thanks for reaching out! Adding to @isaac-chung's comment, we welcome PRs both to improve the Any2AnyRetrieval Evaluator and add your specific tasks! We'll be happy to benchmark your model on all MIEB tasks on our end as well if you can PR your model implementation to here. An old doc for the full process can be found here. |
Hi, It appears that in However, the previous Therefore, I'm wondering if we should add an This is just an example, and also my current implementation. I look forward to everyone's discussion and suggestions for better solutions. |
of course, adding the ability to take in instructions (e.g, model-specific prompts triggered by is_query) has always been the plan since the start of MIEB. Although this ability is not optimized for a lot of the image-text models, especially ones that can't naturally do interleaved encodings (e.g, CLIP-based), I personally think this will be de facto for future models. At the moment, a few state-of-ther-art models have their own optimized formats. e.g., In general, I think it makes sense to add is_query if it doesn't affect other multi-modal models that don't benefit from it. Feel free to PR the solution if you have anything in mind! @izhx |
Thanks for the suggestions!
|
Hi, thanks for the cool MTEB toolkit.
We are currently preparing to release an embedding model for universal multimodal retrieval, along with our compiled evaluations. I noticed that you are also developing image extensions for MTEB. So I would like to inquire if you would be interested in incorporating our testing code into MTEB, perhaps as part of MIEB retrieval.
Our test is primarily divided into four parts: MTEB text retrieval, M-BEIR, ViDoRe, and a few additional it2it retrieval data. I guess many of them has already been incorporated into mteb.
Below is a preliminary model testing results table.
If you're interested, where could I find the docs to start with? Thanks a lot.
The text was updated successfully, but these errors were encountered: