isabelle_mohr's profile picture. MLE @JinaAI_ 🤖 Interested in all things Machine Learning!

Isabelle Mohr

@isabelle_mohr

MLE @JinaAI_ 🤖 Interested in all things Machine Learning!

Isabelle Mohr 已轉發

2025 could be the year of Deep(Re)Search. Test-time compute and reasoning model are transforming search systems now. With <think>, users have been educated to accept delayed gratification—longer waiting times in exchange for higher-quality, actionable results, much like the…

JinaAI_'s tweet image. 2025 could be the year of Deep(Re)Search. Test-time compute and reasoning model are transforming search systems now. With &amp;lt;think&amp;gt;, users have been educated to accept delayed gratification—longer waiting times in exchange for higher-quality, actionable results, much like the…

Isabelle Mohr 已轉發

Great work from MMTEB team! We have 3 contributors from @JinaAI_ ! @michael_g_u @jupyterjazz @isabelle_mohr

此推文已無法使用。

Isabelle Mohr 已轉發

Our submission to ECIR 2025 on jina-embeddings-v3 has been accepted! 🎉 At the ECIR Industry Day @jupyterjazz takes the stage to share how we train the latest version of our text embedding model. More details: ecir2025.eu


Isabelle Mohr 已轉發

Stop paying the OpenAI tax. The best AI devtools are actually open-source, free to use, and give you full control over your data and privacy. While proprietary AI dominated early headlines, the true revolution is happening in open source - where a flourishing ecosystem of…

avthars's tweet image. Stop paying the OpenAI tax. The best AI devtools are actually open-source, free to use, and give you full control over your data and privacy.

While proprietary AI dominated early headlines, the true revolution is happening in open source - where a flourishing ecosystem of…

I'm excited to attend @weaviate_io's AI Hack Night tonight!🚀#hacknightberlin


Isabelle Mohr 已轉發

Join our next ML reading group featuring VisRAG on Nov 29th:

No parsing or OCR; No multi-vector or late interaction! VisRAG from @TsinghuaNLP outperforms TextRAG by addressing RAG bottlenecks at both stages: achieving higher retrieval accuracy and better answer generation via multimodal reasoning. We're thrilled to invite @dgdsxyushi to…

JinaAI_'s tweet image. No parsing or OCR; No multi-vector or late interaction! VisRAG from @TsinghuaNLP outperforms TextRAG by addressing RAG bottlenecks at both stages: achieving higher retrieval accuracy and better answer generation via multimodal reasoning. We&apos;re thrilled to invite @dgdsxyushi to…


Isabelle Mohr 已轉發

We extended our priprint about late chunking, a novel method to make embeddings of chunks context-aware. We added: - Algorithm for long documents - Training method to make late chunking more effective - Comparison to Anthropic's contextual embedding arxiv.org/abs/2409.04701

michael_g_u's tweet image. We extended our priprint about late chunking, a novel method to make embeddings of chunks context-aware. We added:
- Algorithm for long documents
- Training method to make late chunking more effective
- Comparison to Anthropic&apos;s contextual embedding 
arxiv.org/abs/2409.04701

Want to learn more about Embeddings, Rerankers and ColBERT? Come to my talk on Thursday at @qdrant_engine's Vector Space Event 😎 more info here -> lu.ma/88rdjnhg


Isabelle Mohr 已轉發

With @isabelle_mohr at #bbuzz Thank you for being there for my session. 💙

sonam_pankaj_'s tweet image. With @isabelle_mohr at #bbuzz

Thank you for being there for my session. 💙

Isabelle Mohr 已轉發

Got my Weaviate👕 here ;) thanks @femke_plantinga and @philipvollet ! This afternoon my colleague @isabelle_mohr and @saahil will hold two presentations at Berlin Buzzwords, I’ll brief Jina CLIP and upcoming models @JinaAI_ . Come and join our presentation!

Weaviate goes Berlin Buzzwords 💚 Day 2 Come say hi, grab some holographic stickers and vector club shirts. Also, raffle giveaway at 13:30 today! See you there 🫶 @berlinbuzzwords @weaviate_io



Isabelle Mohr 已轉發

Clip your schedules for next week because Andreas and I will present our latest text&image embedding model with advanced text capabilities 😉 Paper: arxiv.org/abs/2405.20204 🤗: huggingface.co/jinaai/jina-cl… API: jina.ai/embeddings/

Get ready for the next MLOps Community Mini Summit! Join us on Wednesday, June 12th, at 17:00 UK time for "Fresh Data, Smart Retrieval: Milvus & Jina CLIP Explained."

mlopscommunity's tweet image. Get ready for the next MLOps Community Mini Summit!

Join us on Wednesday, June 12th, at 17:00 UK time for &quot;Fresh Data, Smart Retrieval: Milvus &amp;amp; Jina CLIP Explained.&quot;


This week I explored chunking methods: the Semantic Chunker from @llama_index on jinaai/wikisections dataset on Hugging Face. Varying the buffer size had pretty much no effect, while increasing the breakpoint percentile threshold increased chunking precision by a lot! @JinaAI_

isabelle_mohr's tweet image. This week I explored chunking methods: the Semantic Chunker from @llama_index on jinaai/wikisections dataset on Hugging Face. Varying the buffer size had pretty much no effect, while increasing the breakpoint percentile threshold increased chunking precision by a lot! @JinaAI_

Last night I had the pleasure of giving a talk together with @jupyterjazz at the Data Meetup Berlin hosted by Netlight! Love the knowledge sharing, and most importantly, to connect with so many passionate and interested people in the field. See you at the next one! #embeddings

isabelle_mohr's tweet image. Last night I had the pleasure of giving a talk together with @jupyterjazz  at the Data Meetup Berlin hosted by Netlight! Love the knowledge sharing, and most importantly, to connect with so many passionate and interested people in the field. See you at the next one!

#embeddings
isabelle_mohr's tweet image. Last night I had the pleasure of giving a talk together with @jupyterjazz  at the Data Meetup Berlin hosted by Netlight! Love the knowledge sharing, and most importantly, to connect with so many passionate and interested people in the field. See you at the next one!

#embeddings

I'll be giving a talk together with @jupyterjazz next week in Berlin about our German-English bilingual embedding model. If you wanna know how we trained this model and how to use it in a RAG pipeline, you better RSVP and attend! See ya there 🚀 meetup.com/data-meetup-be… @JinaAI_


Isabelle Mohr 已轉發

Saw @JinaAI_'s excellent long context (8192!) ColBERT earlier today? Eager to give long-document ColBERT a shot? New joint🫅colbert-ai and🪤RAGatouille release now supports any maximum length the underlying model can handle (& dynamically adjusts maxlen when encoding in-memory)

bclavie's tweet image. Saw @JinaAI_&apos;s excellent long context (8192!) ColBERT earlier today? Eager to give long-document ColBERT a shot?

New joint🫅colbert-ai and🪤RAGatouille release now supports any maximum length the underlying model can handle (&amp;amp; dynamically adjusts maxlen when encoding in-memory)

Isabelle Mohr 已轉發

@JinaAI_ embeddings now supported by sbert 2.3.0 :)

bo_wangbo's tweet image. @JinaAI_ embeddings now supported by sbert 2.3.0 :)

The long-awaited Sentence Transformers v2.3.0 is now released! It contains a ton of bug fixes, performance improvements, loading custom models, more efficient loading, a new strong loss function & more! Check out the release notes: github.com/UKPLab/sentenc… Or this 🧵below:



Isabelle Mohr 已轉發

A few days ago, @JinaAI_ released two new bilingual embedding models (German-English & Chinese-English), each supporting a max sequence length of 8K tokens! 🤯 ... and now you can use them with 🤗 Transformers.js, for cross-language retrieval, clustering, and so much more! 👇

xenovacom's tweet image. A few days ago, @JinaAI_ released two new bilingual embedding models (German-English &amp;amp; Chinese-English), each supporting a max sequence length of 8K tokens! 🤯

... and now you can use them with 🤗 Transformers.js, for cross-language retrieval, clustering, and so much more! 👇

Jina AI's open-source bilingual embedding models for German-English and Chinese-English are now on Hugging Face. We’re going to walk through installation and cross-language retrieval. jina.ai/news/jina-embe…



Isabelle Mohr 已轉發

We’re finally here with 2 new models, we call it bilingual embedding models, it allows you to perform monolingual and cross-lingual retrieval tasks, the future models are always X+EN, X is the main language and EN as the bridging language. Here are the first two: German-English…


Isabelle Mohr 已轉發

Our German-English and Chinese English embedding models are open-source now 🚀 huggingface.co/jinaai/jina-em… huggingface.co/jinaai/jina-em…

michael_g_u's tweet image. Our German-English and Chinese English embedding models are open-source now 🚀
huggingface.co/jinaai/jina-em…
huggingface.co/jinaai/jina-em…

We’re finally here with 2 new models, we call it bilingual embedding models, it allows you to perform monolingual and cross-lingual retrieval tasks, the future models are always X+EN, X is the main language and EN as the bridging language. Here are the first two: German-English…



Loading...

Something went wrong.


Something went wrong.