Multilingual Semantic Search
This talk will discuss why multilingual semantic search is amazing, how respective models are trained, and the new use cases this unlocks.
This talk will discuss why multilingual semantic search is amazing, how respective models are trained, and the new use cases this unlocks.
Connecting Large Language Models with embeddings and semantic search on your own data has become widely popular. But how does this work in other languages and across languages? Join me for this talk why multilingual semantic search is amazing, how respective models are trained, and new use-cases this unlocks.
Nils Reimers
Nils Reimers did his Ph.D. and Post-Doc at the TU Darmstadt, where he created the foundation on how to use transformer networks for semantic search. After his post-doc, he joined Hugging Face to work on self-supervised domain adaptation for semantic search. Last year, Nils joined Cohere.com as director of machine learning to work on large language models for text understanding, including search, classification and text aggregation.
Wikipedia has implemented a new policy that bans AI-generated article creation and substantial rewrites. The ban on LLM usage was voted by Wikipedia's editor community, who found themselves increasingly overwhelmed by the non-stop flood of AI slop making its way into the encyclopedia.
Granola recently raised a $125M Series C at a $1.5B valuation. In addition to the round announcement, Granola also shared the three new features coming to the app: personal and enterprise APIs, an updated MCP, and Spaces for sharing and collaboration.
Interloom has raised a $16.5M seed round to develop a platform that captures undocumented operational expertise and transforms it into a permanent context layer for AI agents. With its "Context Graph", Interloom aims to address the critical knowledge gap that affects enterprise AI deployment.
Mistral has launched Forge, a platform that lets enterprises train AI models from scratch on their own proprietary data for greater accuracy and control.
Cursor recently released Composer 2, a new in-house coding model that vastly improves its predecessor's performance. While Composer 2's benchmark scores may not be outstanding, Cursor is betting that the model's lower price point and native integration to the coding environment will drive adoption.
SF Bay Area media and education platform focused on AI and Data. As a voice of AI industry, Data Phoenix delivers news, practical knowledge, and helps companies be heard in the community.
Copyright © 2026 Data Phoenix. Published with Ghost and Data Phoenix.
Privacy Policy | Terms of Service | Cookie Preferences
Comments