Knowledge Graph Insights

Larry Swanson
undefined
10 snips
Jan 22, 2026 • 55min

Jim Hendler: Scaling AI and Knowledge with the Semantic Web – Episode 43

Jim Hendler, Tetherless World Professor and long-time Semantic Web researcher, reflects on early Semantic Web ideas and scaling knowledge on the web. He discusses SHOE, using URLs for interoperable ontologies, working with Tim Berners-Lee, DARPA efforts to scale knowledge representation, schema.org and Google’s influence, and how knowledge graphs fit with modern LLMs and future heterogeneous compute.
undefined
Jan 12, 2026 • 34min

Brad Bolliger: Pragmatic Semantic Modeling for Government Data – Episode 42

Brad Bolliger, an expert in AI and Data at Ernst & Young and co-chair of the NIEMOpen Technical Architecture Committee, shares insights on pragmatic semantic modeling. They discuss the impact of the Unified Justice Platform on entity resolution and the importance of the NIEM standard in improving public service collaboration. Brad emphasizes the need for tailored ontology management and the benefits of graph technology for entity-centric data modeling. The conversation also touches on crafting AI-ready data and the role of neuro-symbolic architectures in enhancing analytics.
undefined
9 snips
Dec 15, 2025 • 30min

Tara Raafat: Human-Centered Knowledge Graph and Metadata Leadership – Episode 41

Tara Raafat, Head of Metadata and Knowledge Graph Strategy at Bloomberg, shares her extensive experience in semantic technologies, emphasizing the human element in knowledge graph projects. She likens knowledge graphs to the DNA of enterprise information, crucial for AI readiness. Tara discusses the importance of engaging diverse stakeholders and assembling a star team of ontologists and engineers. She also explores the balance between theoretical models and practical constraints, and advocates for leveraging AI while keeping human oversight at the forefront.
undefined
Nov 3, 2025 • 32min

Alexandre Bertails: The Netflix Unified Data Architecture – Episode 40

Alexandre Bertails, a software engineer at Netflix's content engineering team, talks about pioneering the Unified Data Architecture (UDA) to enhance semantic interoperability through RDF. He explores the challenge of creating a singular schema for diverse internal needs while detailing the innovative Upper domain modeling language. Bertails highlights Upper's self-describing and self-governing traits, its compact design, and how Netflix operationalized these concepts to improve data management and accessibility across teams. Tune in for insights on making complex data relatable!
undefined
Oct 12, 2025 • 33min

Torrey Podmajersky: Aligning Language and Meaning in Complex Systems – Episode 39

Torrey Podmajersky, a seasoned UX/content strategist and president of Catbird Content, dives into the intricate world of semantics in design. Influenced by her father's philosophical pursuits, she explores how language shapes user experiences. Torrey highlights the importance of prelecting to bridge implicit knowledge gaps and the need for cross-functional teamwork in crafting effective product language. She also discusses the Cyc project and contrasts the decision-making processes of human writers with LLMs, revealing the profound impact of semantic design.
undefined
104 snips
Aug 20, 2025 • 39min

Casey Hart: The Philosophical Foundations of Ontology Practice – Episode 38

Casey Hart, Lead Ontologist for Ford and ontology consultant, blends a rich background in philosophy with practical ontology applications. He explores the philosophical foundations of ontology, revealing how concepts like metaphysics and epistemology shape knowledge graphs and AI. He shares insights from his time at Cycorp, discusses the complexities of defining AI beyond technology, and warns against inflated expectations surrounding AI in Silicon Valley. Hart emphasizes that ontologies are models that can oversimplify reality, making philosophical engagement crucial.
undefined
Aug 4, 2025 • 33min

Chris Mungall: Collaborative Knowledge Graphs in the Life Sciences – Episode 37

Chris Mungall Capturing knowledge in the life sciences is a huge undertaking. The scope of the field extends from the atomic level up to planetary-scale ecosystems, and a wide variety of disciplines collaborate on the research. Chris Mungall and his colleagues at the Berkeley Lab tackle this knowledge-management challenge with well-honed collaborative methods and AI-augmented computational tooling that streamlines the organization of these precious scientific discoveries. We talked about: his biosciences and genetics work at the Berkeley Lab how the complexity and the volume of biological data he works with led to his use of knowledge graphs his early background in AI his contributions to the gene ontology the unique role of bio-curators, non-semantic-tech biologists, in the biological ontology community the diverse range of collaborators involved in building knowledge graphs in the life sciences the variety of collaborative working styles that groups of bio-creators and ontologists have created some key lessons learned in his long history of working on large-scale, collaborative ontologies, key among them, meeting people where they are some of the facilitation methods used in his work, tools like GitHub, for example his group's decision early on to commit to version tracking, making change-tracking an entity in their technical infrastructure how he surfaces and manages the tacit assumptions that diverse collaborators bring to ontology projects how he's using AI and agentic technology in his ontology practice how their decision to adopt versioning early on has enabled them to more easily develop benchmarks and evaluations some of the successes he's had using AI in his knowledge graph work, for example, code refactoring, provenance tracking, and repairing broken links Chris's bio Chris Mungall is Department Head of Biosystems Data Science at Lawrence Berkeley National Laboratory. His research interests center around the capture, computational integration, and dissemination of biological research data, and the development of methods for using this data to elucidate biological mechanisms underpinning the health of humans and of the planet. He is particularly interested in developing and applying knowledge-based AI methods, particularly Knowledge Graphs (KGs) as an approach for integrating and reasoning over multiple types of data. Dr. Mungall and his team have led the creation of key biological ontologies for the integration of resources covering gene function, anatomy, phenotypes and the environment. He is a principal investigator on major projects such as the Gene Ontology (GO) Consortium, the Monarch Initiative, the NCATS Biomedical Data Translator, and the National Microbiome Data Collaborative project. Connect with Chris online LinkedIn Berkeley Lab Video Here’s the video version of our conversation: https://youtu.be/HMXKFQgjo5E Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 37. The span of the life sciences extends from the atomic level up to planetary ecosystems. Combine this scale and complexity with the variety of collaborators who manage information about the field, and you end up with a huge knowledge-management challenge. Chris Mungall and his colleagues have developed collaborative methods and computational tooling that enable the construction of ontologies and knowledge graphs that capture this crucial scientific knowledge. Interview transcript Larry: Hi everyone. Welcome to episode number 37 of the Knowledge Graph Insights podcast. I am really delighted today to welcome to the show Chris Mungall. Chris is a computational scientist working in the biosciences at the Lawrence Berkeley National Laboratory. Many people just call it the Berkeley Lab. He's the principal investigator in a group there, has his own lab working on a bunch of interesting stuff, which we're going to talk about today. So welcome, Chris, tell the folks a little bit more about what you're up to these days. Chris: Hi, Larry. It's great to be here. Yeah, so as you said, I'm here at Berkeley Lab. We're located in the Bay Area. We're just above UC Berkeley campus. We have a nice view of the San Francisco Bay looking into San Francisco, and so we're a national lab, so we're part of the Department of Energy National Lab system, and we have multiple different areas here in the lab looking at different aspects of science from physics, energy technologies, material science. I'm in the biosciences area, so we are really interested in how we can advance biological science in areas relevant to national scale challenges really in different areas like energy, the environment, health and bio-manufacturing. Chris: My own particular research is really focused on the role of genes and in particular the role of genes in complex systems. So this could be the genes that we have in our own cells, the genes in human beings, how they all work together to hopefully create a healthy human being. One part of my research also looks at the role of genes in the environment, and in particular the role of genes inside tiny old microbes that you'll find in the ocean water and in the soil. And how these genes all work together, both to help drive these microbial systems, help them work together and how they all work together really to drive ecosystems and biogeochemical cycles. Chris: So I think the overall aim is really just to get a picture of these genes and how they interact in these kind of complex systems and build up models of complex systems from scales right the way from atoms through the way through to organisms and indeed all the way to earth-scale systems. So my work is all computational. I don't have a wet lab. So one thing that we realized early on is just when you are sequencing these genomes and trying to interpret the genes, you're generating a lot of information and you need to be able to organize that somehow. And so that's how we arrived at working on knowledge graphs, basically to assemble all of this information together and to be able to use it in algorithms to help us interpret biological data and help us figure out the role of genes in these organisms. Larry: Yeah, many of the people I've talked to on this podcast, they come out of the semantic technology world and apply it in some place or another. It sounds like you came to this world because of the need to work with all the data you've got. What was your learning curve? Was it just another thing in your computational toolkit? Chris: Yeah, in some ways. In fact, my background is, if you go back far enough, my original background is more on the computational side and my undergrad was in AI, but this is back when AI meant good old-fashioned AI and symbolic reasoning and developing Prolog rules to reason about the world and so on. And at that time, I wasn't so interested in that side of AI. I really wanted to push forward with some of the more nascent neural network type approaches. But in those days, we didn't really have the computational power and I thought, "Well, maybe I really need to, I actually learned something about biological systems before trying to simulate them." So that's how I got involved in genomics. This was around about the time of just before the sequencing of the human genome, and I just got really interested in this area, a position came up here at Lawrence Berkeley National Laboratory, and I just got really involved in analyzing some of these genomes. Chris: And in doing this, I came across this project called the Gene Ontology that was developed by some of my colleagues originally in Cambridge and at Lawrence Berkeley National Laboratory. And the goal here was really as we were sequencing these genomes and we were figuring out there's 20,000 genes in the human genome, we discovered we had no way to really categorize what the functions of these different genes were. And if you think about it, there's multiple different ways that you can describe the function of any kind of machine, whether it's a molecular machine inside one of your cells or your car or your iPhone or whatever. You can describe it in terms of what the intent of that machine is. You can describe it in terms of where that machine is localized and what it does, and how that machine works as part of a larger ensemble of machines to achieve some larger objective. Chris: So my colleagues came up with this thing called the gene ontology, and I looked at that and I said, "Hey, I've got this background in symbolic reasoning and good old-fashioned AI. Maybe I could play a role in helping organize all of this information and figuring out ways to connect it together as part of a larger graph." We didn't call them knowledge graphs at this time, but we're essentially building knowledge graphs at the time and make use of, in those days quite early semantic web technologies. This is even before the development of all the web ontology language, but there was still this notion that we could use, we could use rules in combination with graphs to make inferences about things. And I thought, "Well, this seems like an ideal opportunity to apply some of this technology." Larry: That's interesting. It's funny we didn't plan this, but the episode right before you in the queue was of my friend Emeka Okoye. He's a guy who was building knowledge graphs in the late '90s, early 2000s, mostly the early 2000s before the term had been coined, and I think maybe even before a lot of the RDF and OWL and all that stuff was there. So you mentioned Prolog earlier, and what was your toolkit then, and how has it evolved up to the present? That's a huge question. Yeah. Chris: I didn't mean to get into my whole early days with Prolog. Yeah, I've definitely had some interest in applying a lot of these logic programming technologies. As you're aware,
undefined
13 snips
Jul 21, 2025 • 35min

Emeka Okoye: Exploring the Semantic Web with the Model Context Protocol – Episode 36

Emeka Okoye, a seasoned Knowledge Engineer and Semantic Architect with over 20 years in knowledge engineering, dives into the world of the Semantic Web. He shares insights on the transformative Model Context Protocol (MCP) and its impact on AI applications. Emeka discusses his RDF Explorer, a tool that allows developers easy access to semantic data without specialized language skills. The conversation also touches on the evolution of ontology engineering, his history in tech innovation in Nigeria, and the importance of making semantic technologies accessible globally.
undefined
Jul 6, 2025 • 32min

Tom Plasterer: The Origins of FAIR Data Practices – Episode 35

In this discussion, Tom Plasterer, Managing Director at XponentL Data and a leading expert in data strategy, delves into the origins and significance of FAIR data principles. He unpacks how the concept evolved from the semantic web to address the need for discoverable data in research and industry. Tom reveals the 15 facets of the FAIR acronym and emphasizes the critical role of knowledge graphs in implementing these standards. His journey from bioinformatics to data management showcases the importance of collaboration and shared terminology in enhancing data practices, especially in pharmaceuticals and life sciences.
undefined
14 snips
Jun 11, 2025 • 31min

Mara Inglezakis Owens: A People-Loving Enterprise Architect – Episode 34

Mara Inglezakis Owens, an enterprise architect at Delta Air Lines, blends her humanities background with digital anthropology to shape user-focused architecture. She discusses how mentoring shaped her approach and emphasizes the need for understanding actual stakeholder behaviors over self-reports. Mara also shares insights on justifying financial investments in her work, the significance of documentation in knowledge engineering, and lessons learned about embracing imperfection in digital systems design. Her human-centered focus exemplifies the evolution of enterprise architecture in modern businesses.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app