TiddlyWiki / TiddlyWiki5

A self-contained JavaScript wiki for the browser, Node.js, AWS Lambda etc.
https://tiddlywiki.com/
Other
8.03k stars 1.19k forks source link

Structured data for search engines like google #1862

Open felixhayashi opened 9 years ago

felixhayashi commented 9 years ago

Since tiddlywiki has a very clean way of structuring data (title, tags, small content) it would make sense to give search engines information about the content by using meta-information and structured data markup (open graph etc.) so facebook, google etc. could potentially create rich snippets.

Just wanted to create this issue to create general awareness...

-Felix

pmario commented 9 years ago

If you create static content from TW, you can insert what ever you want, to make search engines happy. IMO they are much happier about static content, then web app content.

... Since empty TW, is already "kind of bloated" I'd like to avoid carrying more meta data for systems I'm not going to use. On the other hand I may want to include specific meta data for systems I am using.

... for me plugin territory.

tobibeer commented 9 years ago

@felixhayashi, did you figure out whether those APIs cater for having multiple "pages" in one, ...e.g. a SPA like TW? So long as google or others neither know nor care for indexing or opening tiddlers, that seems a difficult thing.

Or are you simply talking about some meta-data on a wiki-document-level?

felixhayashi commented 9 years ago

did you figure out whether those APIs cater for having multiple "pages" in one, ...e.g. a SPA like TW?

Well, theoretically it should work to mark areas of a SPA (in our case tiddlers) as being content snippets that deal with a certain topic. But not sure which "shared vocabularies" are especially designed for SPA. Here is a stackoverflow article that deals with schema.org http://webmasters.stackexchange.com/questions/68682/schema-org-webpage-structure-for-single-page-site

So long as google or others neither know nor care for indexing or opening tiddlers, that seems a difficult thing.

Yes, I thought about that too. We would need to explore further what content is actually accessed by a crawler. I know the google bot crawls after rendering and loading the js. so theoretically it should also follow all internal page links and index the content…