Skip to main content

Data feeders


There are three ways to insert your page's data in Tonos: by using scripts, sitemaps, or webhooks. This page data is important to Tonos, because it is through this data that Tonos is able to recognize when you publish, update, or delete content.

NOTE

Your page needs to be publicly accesible in order for Tonos to be able to crawl its data.

Inserting page data through the script

This is the most used method for inserting page data in Tonos. Data such as the author, title, published-time, and tags, are received from Tonos. These properties should be located in the HTML body of your page along with the application script in order for this method to work.

Inserting page data through sitemaps

A sitemap is a file where you provide information about the pages, videos, and other files on your site, and the relationships between them. Search engines like Google read this file to crawl your site more efficiently. Sitemaps also help in understanding your website structure, making it easier to navigate your website. Tonos takes the links (<loc>) of your pages and crawls them one by one.

Inserting page data through a catalog crawler

Tonos through a catalog (a JSON file that contains a list of the objects which contain the properties Tonos takes from the page) crawls data from your web application. A template of what the catalog crawler stores is shown in the code snippet below.

JSON

{
"ContentId": "{The content Id}",
"Url": "{Page URL}",
"Title": "{title}",
"Description": "{Description}",
"PublishDate": "{Date}",
"Authors": [
"{Auth}",
"{Auth2}"
],
"Tags": [
"{Tag1}",
"{Tag2}"
],
"Sections": [
"{Section1}",
"{Section2}"
],
"Keywords": [
"{Keyword1}",
"{Keyword2}"
]
},

Inserting page data through webhooks

A webhook in web development is a method of augmenting or altering the behavior of a web page or web application with custom callbacks. In Tonos's case, a webhook is an endpoint where you, the user, insert your page data into Tonos manually.

Tonos crawler

Tonos has its own crawler which takes the link of the page you insert when creating your application in the admin page, and then proceeds to crawl all the pages that contain that URL.