bolha.us is one of the many independent Mastodon servers you can use to participate in the fediverse.
We're a Brazilian IT Community. We love IT/DevOps/Cloud, but we also love to talk about life, the universe, and more. | Nós somos uma comunidade de TI Brasileira, gostamos de Dev/DevOps/Cloud e mais!

Server stats:

252
active users

#sparql

1 post1 participant0 posts today

The Model Context Protocol acts as a universal translator between LLMs and data sources, eliminating complex platform-specific requirements. Our new open source MCP server for ODBC (mcp-odbc-server) enables seamless integration of any ODBC-accessible data into RAG pipelines.

Read more in my latest newsletter!

linkedin.com/pulse/whats-model

#MCP#ODBC#SPARQL

Looking for some #SPARQL help, as I'm a newbie to this and my mental model of how triple stores work is incomplete.

I have two versions of the same set of terms in two Turtle files, and I want to load them up into a triple store and then compare them to see what changed: terms updated, added or deleted.

Linked Data Overheid, #LiDO (LD #Government) has been published as Free and Open Source Software.

LiDO is a framework for linking (NL and EU) #legislation, (NL, EU and ECHR) court decisions and parliamentary documents by using linked data technology.

The official production LiDO is here:
linkeddata.overheid.nl/front/p

The source code is here:
gitlab.com/koop/ld/lido

Snapshot of LiDO data is here:
data.overheid.nl/zoek?search=l

linkeddata.overheid.nlLiDO - HomeOmschrijving

wikidata + mediawiki = wikidata + provenance == wikiprov


by @beet_keeper

Today I want to showcase a Wikidata proof of concept that I developed as part of my work integrating Siegfried and Wikidata.

That work is wikiprov a utility to augment Wikidata results in JSON with the Wikidata revision history.

For siegfried it means that we can showcase the source of the results being returned by an identification without having to go directly back to Wikidata, this might mean more exposure for individuals contributing to Wikidata. We also provide access to a standard permalink where records contributing to a format identification are fixed at their last edit. Because Wikidata is more mutable than a resource like PRONOM this gives us the best chance of understanding differences in results if we are comparing siegfried+Wikidata results side-by-side.

I am interested to hear your thoughts on the results of the work. Lets go into more detail below.

📢 Creation of the @w3c Data Shapes #WorkingGroup which mission is to align data shapes standards with the latest versions of core Semantic Web standards, including support for #RDFstar
▶️ w3.org/groups/wg/data-shapes/

This group aims to update SHACL to align with #RDF & #SPARQL 1.2 versions, extend #SHACL specifications and develop new SHACL-based specs.

More info in the group's charter: w3.org/2024/12/data-shapes.htm

W3CData Shapes Working GroupThe mission of the Data Shapes Working Group is to update data shapes standards in line with the versions of core Semantic Web standards that cater for RDF-star and to extend the applications of data shapes with new packaging and use specifications.

Deadline for research paper submissions for ESWC 2025 is approaching! ...for contributions addressing theoretical, analytical, and empirical aspects of the Semantic Web, knowledge graphs and semantic technologies in general and their relation with AI.

abstract deadline: Dec 12, 2025
full paper deadline: Dec 19, 2025
2025.eswc-conferences.org/call

#eswc2025 @eswc_conf @fizise @AxelPolleres @albertmeronyo @pgroth @nfdi4culture #semanticweb #knowledgegraphs #ai #llms #sparql #rdf #owl #generativeAI

"A paper[1] presented at last week's EMNLP conference reports on a promising new AI-based tool (available at spinach.genie.stanford.edu/ ) to retrieve information from Wikidata using natural language questions. It can successfully answer complicated questions like the following:

"What are the musical instruments played by people who are affiliated with the University of Washington School of Music and have been educated at the University of Washington, and how many people play each instrument?"

The authors note that Wikidata is one of the largest publicly available knowledge bases [and] currently contains 15 billion facts, and claim that it is of significant value to many scientific communities. However, they observe that Effective access to Wikidata data can be challenging, requiring use of the SPARQL query language.

This motivates the use of large language models to convert natural language questions into SPARQL queries, which could obviously be of great value to non-technical users."

meta.wikimedia.org/wiki/Resear

spinach.genie.stanford.eduSpinach Wikidata
Replied in thread

My dive into #SPARQL and the #Wikidata environment continues and I just discovered some of the wonderful tools hosted on toolforge.org/.

Here is a map of all periodicals published in #Palestine (defined by a rectangular bounding box) before 1930: w.wiki/9u$o. Items on the map link to #Reasonator (reasonator.toolforge.org/), which provides an improved view of linked data available from Wikidata.

Replied in thread

1. Whistled Occitan : leveraging #Lingualibre speed to record #languages vocabularies we are collaborating with one of the last 5 practicians of Whistled #Occitan. Completmented with #Wikidata and #SPARQL, we successfully prototyped an interactive multimedia map with local toponyms, so the public can explore their ancestral land and hear place names being whistled to them in this endangered language. A working prototype, it works but also needs further care.

Continued thread

🤖🐲 User:Dragons_Bot to the rescue ! Doing clean ups !

Did you know ? Lingualibre has 219 #languages recorded, but a #SPARQL query will return 221 languages. Why ? Because Chinese, by example, is erroneously present twice 😲 :
- ❌ as Q130, iso: zho for #Chinese writing
- ✅ as Q113, iso `cmn`, for Chinese Mandarin
Tonight, I code a script to move all records toward `cmn`, on both #Lingualibre's items and #Commons' file wikipages. Fighto ! ò__ó

Continued thread

🤖🐲Another long day with User:Dragons_Bot!

Months ago, I activated several #SignLanguages on #LinguaLibre. People can video record signed words. While doing activity stats, a #SPARQL query shown missing data on 467 languages items. Dragons_Bot just fixed those. Will be useful for incoming 3rd recording type for #WhistledLanguages. 😉

Today, I use Lingualibre Wikibase as a calm pad for coding my bot. Some days, I will move to #Wikidata for live editing on languages. 🎉
en.wikipedia.org/wiki/Whistled