Open Qwlouse opened 11 years ago
pinae commented 10 days ago
I don't know how this should work since we don't want to generate files which are reachable for normal users.
Maybe we could use a api to supply the crawer with the information the user will get displayed if he enters a certain url.
We have RESTful URLs. Maybe that's enough for a crawler. We should test that first before diving into SE-APIs.
Our URLs do not work for indexing thus we are loading content only by ajax calls. This is not done by Crawlers. On the other Hand we need to talk about the fact that indexing may not be wanted....
If we don't want indexing we have to supply a robots.txt. I don't know the google crawler in detail but I think it might be capable of loading ajax.
Personally I think indexing would be a benefit.
google actually offers a special instructions for webmasters to allow for crawling and indexing dynamic pages:
https://developers.google.com/webmasters/ajax-crawling/
This would require that we change our urls from
hostname/#/foo.1
to
hostname/#!/foo.1
and have our backend deliver some static pages with content if the crawler asks for the special url:
hostname/?_escaped_fragment_=/foo.1
I think this should be doable. So if we want this...
As of yesterday we have changed the urls to hashbang in old browsers and without any # in new Browsers. So next we need to include the
<meta name="fragment" content="!">
into our pages and make the backend return static pages for "the ugly url":
hostname/?_escaped_fragment_=/foo.1
We really need to generate static pages to enable searchengines to scan the pages. At least we need to ensure that start page is well indexable.