Not so much an issue, but a request for help/direction/instructions on how to edit/change colours of the search? I am having real trouble getting to grips with this.
docusaurus.config.js ππ»
plugins: [
[require.resolve('@cmfcmf/docusaurus-search-local'), {
// whether to index docs pages
indexDocs: true,
// must start with "/" and correspond to the routeBasePath configured for the docs plugin
// use "/" if you use docs-only-mode
// (see https://v2.docusaurus.io/docs/2.0.0-alpha.70/docs-introduction#docs-only-mode)
docsRouteBasePath: '/docs',
// Whether to also index the titles of the parent categories in the sidebar of a doc page.
// 0 disables this feature.
// 1 indexes the direct parent category in the sidebar of a doc page
// 2 indexes up to two nested parent categories of a doc page
// 3...
//
// Do _not_ use Infinity, the value must be a JSON-serializable integer.
indexDocSidebarParentCategories: 0,
// whether to index blog pages
indexBlog: true,
// must start with "/" and correspond to the routeBasePath configured for the blog plugin
// use "/" if you use blog-only-mode
// (see https://v2.docusaurus.io/docs/2.0.0-alpha.70/blog#blog-only-mode)
blogRouteBasePath: '/blog',
// whether to index static pages
// /404.html is never indexed
indexPages: false,
// language of your documentation, see next section
language: "en",
// setting this to "none" will prevent the default CSS to be included. The default CSS
// comes from autocomplete-theme-classic, which you can read more about here:
// https://autocomplete.algolia.com/docs/autocomplete-theme-classic
// The default config is undefined (just that as text)
style: "none",
// lunr.js-specific settings
lunr: {
// When indexing your documents, their content is split into "tokens".
// Text entered into the search box is also tokenized.
// This setting configures the separator used to determine where to split the text into tokens.
// By default, it splits the text at whitespace and dashes.
//
// Note: Does not work for "ja" and "th" languages, since these use a different tokenizer.
tokenizerSeparator: /[\s\-]+/
}
}]
],
I have put the custom CSS variables in my custom.css & installed as per the algolia directions you noted in the link - but this approach is not working - do I need to swizzle the search component?
Hello @cmfcmf ππ»
Love the search feature! Works great!
Not so much an issue, but a request for help/direction/instructions on how to edit/change colours of the search? I am having real trouble getting to grips with this.
docusaurus.config.js ππ»
I have put the custom CSS variables in my custom.css & installed as per the algolia directions you noted in the link - but this approach is not working - do I need to swizzle the search component?
Many thanks for your help π
Wayne.