Researcher Dominik Weckmüller discusses semantic search using embeddings to analyze text with geographic references. Topics include using deep learning models, creating embeddings, challenges in explainability, and the future of embeddings in different media and languages.
50:39
forum Ask episode
web_stories AI Snips
view_agenda Chapters
auto_awesome Transcript
info_circle Episode notes
question_answer ANECDOTE
Using HyperLogLog for Privacy
Dominik used HyperLogLog to count distinct social media users talking about urban green spaces while preserving privacy.
This helped understand park usage without compromising individual user data.
insights INSIGHT
Embeddings Capture Text Meaning
Embeddings are numerical representations of text that capture its meaning beyond keywords.
Similarity between embeddings lets you search large text databases by meaning, not just keyword matches.
insights INSIGHT
Semantic Search Without Keywords
Semantic search allows querying social media databases without predefined keywords, just using text and geographic references.
This flexibility is a major advantage over traditional topic-based keyword searches.
Get the Snipd Podcast app to discover more snips from this episode
This podcast episode is all about semantic search and using embeddings to analyse text and social media data.
Dominik Weckmüller, a researcher at the Technical University of Dresden, talks about his PhD research, where he looks at how to analyze text with geographic references.
He explains hyperloglog and embeddings, showing how these methods capture the meaning of text and can be used to search big databases without knowing the topics beforehand.
Here are the main points discussed:
Intro to Semantic Search and Hyperloglog: Looking at social media data by counting different users talking about specific topics in parks, while keeping privacy in mind.
Embeddings and Deep Learning Models: Turning text into numerical vectors (embeddings) to understand its meaning, allowing for advanced searches.
Application Examples: Using embeddings to search for things like emotions or activities in parks without needing predefined keywords.
Creating and Using Embeddings: Tools like transformers.js let you make embeddings on your computer, making it easy to analyze text.
Challenges and Innovations: Talking about how to explain the models, deal with long texts, and keep data private when using embeddings.
Future Directions: The potential for using embeddings with different media (like images and videos) and languages, plus the ongoing research in this fast-moving field.