Status: early draft, far from usable. Feedback welcome.
Atomic Data is a proposed standard for modeling and exchanging data. It uses links to connect pieces of data, and therefore makes it easier to connect datasets to each other - even when these datasets exist on separate machines.
Atomic Data is especially suitable for knowledge graphs, distributed datasets, semantic data, p2p applications, decentralized apps, and data that is meant to be highly reusable. It is designed to be highly extensible, easy to use, and to make the process of domain specific standardization as simple as possible.
Atomic Data is Linked Data, as it is a more strict subset of RDF.
It is typed (you know if something is a
URL, etc.) and extensible through Atomic Schema, which means that you can define your own Classes, Properties and Datatypes.
Atomic Data has a standard for synchronizing data by communicating state changes, called Atomic Mutations.
You can use parts of Atomic Data separately, but the standard is designed as a full, integrated data management package that makes it easier to create, share and use structured data on the web.
- Atomic Data Core: the core model for typed, Linked Data
- Atomic Schema: defining properties, datatypes and classes
- Atomic Mutations: sharing state changes, verifying changes and collaboration
Linked data (RDF / the semantic web) enables us to use the web as a large, decentralized graph database. Using links everywhere in data has amazing merits: links remove ambiguity, they enable exploration, they enable connected datasets. Linked Data could help to democratize the web by decentralizing information storage, and giving people more control. The Solid Project by Tim Berners-Lee is a great example of why linked data can help to create a more decentralized web.
At Ontola, we've been working with linked data quite intensely for the last couple of years. We went all-in on RDF, and challenged ourselves to create software that communicates exclusively using it. That has been an inspiring, but also sometimes a frustrating journey. While building various production grade apps (e.g. our e-democracy platform Argu.co, which is used by various governments), we had to solve many problems. How to properly model data in RDF? How to deal with sequences? How do you deal with mutations? Converting to HTML? Typing? CORS? We tackled some of these problems by having a tight grip on the data that we create (e.g. we know the type of data, because we control the resources), and another part is creating new protocols, formats, tools, and libraries. But it took a long time, and it was hard. It's been almost 15 years since the introduction of linked data, and its adoption has been slow. We know that some of its merits are undeniable, and we truly want the semantic web to succeed. We believe the lack of growth partially has to do with a lack of tooling, but also with some problems that lie in the RDF data model.
Atomic Data aims to take the best parts from RDF, and learn from the past to make a more developer-friendly, performant and reliable data model to achieve a truly linked web.