lucene-solr-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Thomas Werthmüller <>
Subject Solr: Highlighting word parts in excerpt does not work
Date Thu, 05 Apr 2012 06:34:25 GMT

I configured solr that also word parts are found. When is search "Monday"
or "Mond" the right document is found. This is done with the following
configuration in the schema.xml: <filter
class="solr.EdgeNGramFilterFactory" minGramSize="3" maxGramSize="30"/>.

Now, when I add hl=true to the query sting, the excerpt for "Monday" looks
good and the word is highlighted. When i search only with "Mond", the
document is found but no excerpt is returned because the query sting is not
the whole word.

I hope someone can give me a hint that also excerpts returned with word


<?xml version="1.0" encoding="UTF-8" ?>

 This is the Solr schema file. This file should be named "schema.xml" and
 should be in the conf directory under the solr home
 (i.e. ./solr/conf/schema.xml by default)
 or located where the classloader for the Solr webapp can find it.

 For more information, on how to customize this file, please see

<schema name="search-api-1.0" version="1.2">
    <!-- attribute "name" is the name of this schema and is only used for
display purposes.
         Applications should change this to reflect the nature of the
search collection.
         version="1.2" is Solr's version number for the schema syntax and
semantics.  It should
         not normally be changed by applications.
         1.0: multiValued attribute did not exist, all fields are
multiValued by nature
         1.1: multiValued attribute introduced, false by default
         1.2: omitTermFreqAndPositions attribute introduced, true by
default except for text fields.
         1.3: removed optional field compress feature
    <!-- field type definitions. The "name" attribute is
       just a label to be used by field definitions.  The "class"
       attribute and any other attributes determine the real
       behavior of the fieldType.
         Class names starting with "solr" refer to java classes in the
       org.apache.solr.analysis package.

    <!-- The StrField type is not analyzed, but indexed/stored verbatim.
       - StrField and TextField support an optional compressThreshold which
       limits compression (if enabled in the derived fields) to values which
       exceed a certain size (in characters).
    <fieldType name="string" class="solr.StrField" indexed="true"
stored="true" sortMissingLast="true" omitNorms="true"/>

    <!-- boolean type: "true" or "false" -->
    <fieldType name="boolean" class="solr.BoolField" indexed="true"
stored="true" sortMissingLast="true" omitNorms="true"/>

    <!-- The optional sortMissingLast and sortMissingFirst attributes are
         currently supported on types that are sorted internally as strings.
       - If sortMissingLast="true", then a sort on this field will cause
         without the field to come after documents with the field,
         regardless of the requested sort order (asc or desc).
       - If sortMissingFirst="true", then a sort on this field will cause
         without the field to come before documents with the field,
         regardless of the requested sort order.
       - If sortMissingLast="false" and sortMissingFirst="false" (the
         then default lucene sorting will be used which places docs without
         field first in an ascending sort and last in a descending sort.

    <!-- numeric field types that store and index the text
         value verbatim (and hence don't support range queries, since the
         lexicographic ordering isn't equal to the numeric ordering) -->
    <fieldType name="integer" class="solr.IntField" omitNorms="true"/>
    <fieldType name="long" class="solr.LongField" omitNorms="true"/>
    <fieldType name="float" class="solr.FloatField" omitNorms="true"/>
    <fieldType name="double" class="solr.DoubleField" omitNorms="true"/>

      These should only be used for compatibility with existing indexes
(created with older Solr versions)
      or if "sortMissingFirst" or "sortMissingLast" functionality is
needed. Use Trie based fields instead.

      Numeric field types that manipulate the value into
      a string value that isn't human-readable in its internal form,
      but with a lexicographic ordering the same as the numeric ordering,
      so that range queries work correctly.
    <fieldType name="sint" class="solr.SortableIntField"
sortMissingLast="true" omitNorms="true"/>
    <fieldType name="slong" class="solr.SortableLongField"
sortMissingLast="true" omitNorms="true"/>
    <fieldType name="sfloat" class="solr.SortableFloatField"
sortMissingLast="true" omitNorms="true"/>
    <fieldType name="sdouble" class="solr.SortableDoubleField"
sortMissingLast="true" omitNorms="true"/>

     Numeric field types that index each value at various levels of
     to accelerate range queries when the number of values between the range
     endpoints is large. See the javadoc for NumericRangeQuery for internal
     implementation details.

     Smaller precisionStep values (specified in bits) will lead to more
     indexed per value, slightly larger index size, and faster range
     A precisionStep of 0 disables indexing at different precision levels.
    <fieldType name="tint" class="solr.TrieIntField" indexed="true"
stored="true" precisionStep="8" omitNorms="true" positionIncrementGap="0"/>
    <fieldType name="tfloat" class="solr.TrieFloatField" indexed="true"
stored="true" precisionStep="8" omitNorms="true" positionIncrementGap="0"/>
    <fieldType name="tlong" class="solr.TrieLongField" indexed="true"
stored="true" precisionStep="8" omitNorms="true" positionIncrementGap="0"/>
    <fieldType name="tdouble" class="solr.TrieDoubleField" indexed="true"
stored="true" precisionStep="8" omitNorms="true" positionIncrementGap="0"/>

    <!-- The format for this date field is of the form
1995-12-31T23:59:59Z, and
         is a more restricted form of the canonical representation of
         The trailing "Z" designates UTC time and is mandatory.
         Optional fractional seconds are allowed: 1995-12-31T23:59:59.999Z
         All other components are mandatory.

         Expressions can also be used to denote calculations that should be
         performed relative to "NOW" to determine the value, ie...

                  ... Round to the start of the current hour
                  ... Exactly 1 day prior to now
                  ... 6 months and 3 days in the future from the start of
                      the current day

         Consult the DateField javadocs for more information.
    <fieldType name="date" class="solr.DateField" sortMissingLast="true"

    <!-- A Trie based date field for faster date range queries and date
faceting. -->
    <fieldType name="tdate" class="solr.TrieDateField" indexed="true"
stored="true" omitNorms="true" precisionStep="6" positionIncrementGap="0"/>

    <!-- solr.TextField allows the specification of custom text analyzers
         specified as a tokenizer and a list of token filters. Different
         analyzers may be specified for indexing and querying.

         The optional positionIncrementGap puts space between multiple
fields of
         this type on the same document, with the purpose of preventing
false phrase
         matching across fields.

         For more info on customizing your analyzer chain, please see

    <!-- One can also specify an existing Analyzer class that has a
         default constructor via the class attribute on the analyzer element
    <fieldType name="text_greek" class="solr.TextField">
      <analyzer class="org.apache.lucene.analysis.el.GreekAnalyzer"/>

    <!-- A text field that only splits on whitespace for exact matching of
words -->
    <fieldType name="text_ws" class="solr.TextField"
        <tokenizer class="solr.WhitespaceTokenizerFactory"/>

    <!-- A text field that uses WordDelimiterFilter to enable splitting and
matching of
        words on case-change, alpha numeric boundaries, and
non-alphanumeric chars,
        so that a query of "wifi" or "wi fi" could match a document
containing "Wi-Fi".
        Synonyms and stopwords are customized by external files, and
stemming is enabled.
        Duplicate tokens at the same position (which may result from
Stemmed Synonyms or
        WordDelim parts) are removed.

    This is the field type this module will (unless manually changed) use
    all fulltext fields. Some filters are commented out because they are not
    language-independent – i.e., SynonymFilterFactory, StopFilterFactory and
    SnowballPorterFilterFactory. Manually remove the XML comment surrounding
    those entries and configure them as necessary (editing synonyms.txt,
    stopwords.txt and protwords.txt, and selecting the language used for
    stemming) if you want to use them.
    <fieldType name="text" class="solr.TextField" indexed="true"
stored="true" multiValued="true" positionIncrementGap="100">
      <analyzer type="index">
        <charFilter class="solr.MappingCharFilterFactory"
        <tokenizer class="solr.WhitespaceTokenizerFactory"/>
        <filter class="solr.SynonymFilterFactory"
synonyms="index_synonyms.txt" ignoreCase="true" expand="false"/>
        <!-- Case insensitive stop word removal.
          add enablePositionIncrements=true in both the index and query
          analyzers to leave a 'gap' for more accurate phrase queries.
        <filter class="solr.StopFilterFactory"
                enablePositionIncrements="true" />
        <charFilter class="solr.HTMLStripCharFilterFactory" />
<filter class="solr.EdgeNGramFilterFactory" minGramSize="3"
        <filter class="solr.WordDelimiterFilterFactory"
        <filter class="solr.LowerCaseFilterFactory"/>
        <!-- <filter class="solr.SnowballPorterFilterFactory"
language="English" protected="protwords.txt"/> -->
        <filter class="solr.RemoveDuplicatesTokenFilterFactory"/>
      <analyzer type="query">
        <charFilter class="solr.MappingCharFilterFactory"
        <tokenizer class="solr.WhitespaceTokenizerFactory"/>
        <!-- <filter class="solr.SynonymFilterFactory"
synonyms="synonyms.txt" ignoreCase="true" expand="true"/> -->
        <filter class="solr.StopFilterFactory"
                enablePositionIncrements="true" />
        <filter class="solr.WordDelimiterFilterFactory"
        <filter class="solr.LowerCaseFilterFactory"/>
        <!-- <filter class="solr.SnowballPorterFilterFactory"
language="English" protected="protwords.txt"/> -->
        <filter class="solr.RemoveDuplicatesTokenFilterFactory"/>

    <!-- Edge N gram type - for example for matching against queries with
        KeywordTokenizer leaves input string intact as a single term.
    <fieldType name="edge_n2_kw_text" class="solr.TextField"
     <analyzer type="index">
       <tokenizer class="solr.KeywordTokenizerFactory"/>
       <filter class="solr.LowerCaseFilterFactory"/>
       <filter class="solr.EdgeNGramFilterFactory" minGramSize="2"
maxGramSize="25" />
     <analyzer type="query">
       <tokenizer class="solr.KeywordTokenizerFactory"/>
       <filter class="solr.LowerCaseFilterFactory"/>
   <!--  Setup simple analysis for spell checking -->

   <fieldType name="textSpell" class="solr.TextField"
       <tokenizer class="solr.StandardTokenizerFactory" />
       <filter class="solr.StopFilterFactory" ignoreCase="true"
       <filter class="solr.LengthFilterFactory" min="4" max="20" />
       <filter class="solr.LowerCaseFilterFactory" />
       <filter class="solr.RemoveDuplicatesTokenFilterFactory" />

    <!-- This is an example of using the KeywordTokenizer along
         With various TokenFilterFactories to produce a sortable field
         that does not include some properties of the source text
    <fieldType name="sortString" class="solr.TextField" indexed="true"
stored="true" sortMissingLast="true" omitNorms="true">
        <!-- KeywordTokenizer does no actual tokenizing, so the entire
             input string is preserved as a single token
        <tokenizer class="solr.KeywordTokenizerFactory"/>
        <!-- The LowerCase TokenFilter does what you expect, which can be
             when you want your sorting to be case insensitive
        <filter class="solr.LowerCaseFilterFactory" />
        <!-- The TrimFilter removes any leading or trailing whitespace -->
        <filter class="solr.TrimFilterFactory" />
        <!-- The PatternReplaceFilter gives you the flexibility to use
             Java Regular expression to replace any sequence of characters
             matching a pattern with an arbitrary replacement string,
             which may include back refrences to portions of the original
             string matched by the pattern.

             See the Java Regular Expression documentation for more
             infomation on pattern and replacement string syntax.

        <filter class="solr.PatternReplaceFilterFactory"
                pattern="(^\p{Punct}+)" replacement="" replace="all"

    <!-- A random sort type -->
    <fieldType name="rand" class="solr.RandomSortField" indexed="true" />

    <!-- since fields of this type are by default not stored or indexed,
any data added to
         them will be ignored outright
    <fieldtype name="ignored" stored="true" indexed="false"
class="solr.StrField" />


   <!-- Valid attributes for fields:
     name: mandatory - the name for the field
     type: mandatory - the name of a previously defined type from the
<types> section
     indexed: true if this field should be indexed (searchable or sortable)
     stored: true if this field should be retrievable
     compressed: [false] if this field should be stored using gzip
       (this will only apply if the field type is compressable; among
       the standard field types, only TextField and StrField are)
     multiValued: true if this field may contain multiple values per
     omitNorms: (expert) set to true to omit the norms associated with
       this field (this disables length normalization and index-time
       boosting for the field, and saves some memory).  Only full-text
       fields or fields that need an index-time boost need norms.

     The document IDs.
     - item_id contains the entity ID, e.g. a node's nid.
     - index_id is the machine name of the search index this entry belongs
     - id is a combination of the two, for uniquely identifying a document.
       The format used is "$index_id-$item_id".

   <field name="id"       type="string" indexed="true" stored="true"
required="true" />
   <field name="item_id"  type="string" indexed="true" stored="true"
required="true" />
   <field name="index_id" type="string" indexed="true" stored="true"
required="true" />

   <!-- Since sorting by ID is explicitly allowed, store item_id also in a
sortable way. -->
   <copyField source="item_id" dest="ss_search_api_id" />

   <!-- This field is used to build the spellchecker index. -->
   <field name="spell" type="textSpell" indexed="true" stored="true"
  <!-- Copy all text fields into our spell field for indexing. -->
   <copyField source="t_*" dest="spell"/>

   <!-- Dynamic field definitions.  If a field name is not found,
        will be used if the name matches any of the patterns.
        RESTRICTION: the glob-like pattern in the name attribute must have
        a "*" only at the start or the end.
        EXAMPLE:  name="*_i" will match any field ending in _i (like
myid_i, z_i)
        Longer patterns will be matched first.  if equal size patterns
        both match, the first appearing in the schema will be used.  -->

     We only use dynamic fields for storing all fields except the above
     since all others will change between indexes.
     We here define two dynamic fields for each used data type,  one
single- and
     one multi-valued.

     For customizing this, you can create an explicitly named field
     this behaviour in one instance. E.g., for indexing a node title not
     the normal "text" datatype but with "text_ws" (see definitions above),
     could add:
     <field name="t_title" type="text_ws" indexed="true" stored="true" />

   <!-- Fulltext fields (are always multiValued) -->
   <dynamicField name="t_*" type="text" termVectors="true" />
   <!-- These will be used for fields with type "String" or "URI". -->
   <dynamicField name="ss_*" type="sortString" multiValued="false"
termVectors="true" />
   <dynamicField name="sm_*" type="sortString" multiValued="true"
termVectors="true" />
   <!-- These will be used for fields with type "Integer" or "Duration". -->
   <dynamicField name="is_*" type="tlong" multiValued="false"
termVectors="true" />
   <dynamicField name="im_*" type="long" multiValued="true"
termVectors="true" />
   <!-- These will be used for fields with type "Decimal". -->
   <dynamicField name="fs_*" type="tdouble" multiValued="false"
termVectors="true" />
   <dynamicField name="fm_*" type="tdouble" multiValued="true"
termVectors="true" />
   <!-- These will be used for fields with type "Date". -->
   <dynamicField name="ds_*" type="tdate" multiValued="false"
termVectors="true" />
   <dynamicField name="dm_*" type="tdate" multiValued="true"
termVectors="true" />
   <!-- These will be used for fields with type "Boolean". -->
   <dynamicField name="bs_*" type="boolean" multiValued="false"
termVectors="true" />
   <dynamicField name="bm_*" type="boolean" multiValued="true"
termVectors="true" />

   <!-- These fields are used for facetting on string fields, so case and
        special characters are correctly displayed. -->
   <dynamicField name="f_ss_*" type="string" multiValued="false"
termVectors="true" />
   <dynamicField name="f_sm_*" type="string" multiValued="true"
termVectors="true" />
   <copyField source="ss_*" dest="f_ss_*" />
   <copyField source="sm_*" dest="f_sm_*" />

   <!-- The following causes solr to ignore any fields that don't already
match an existing
        field name or dynamic field, rather than reporting them as an error.
        Alternately, change the type="ignored" to some other type, e.g.
"text", if you want
        unknown fields indexed and/or stored by default. -->
   <dynamicField name="*" type="ignored" multiValued="true" />


 <!-- Field to use to determine and enforce document uniqueness.
      Unless this field is marked with required="false", it will be a
required field

 <!-- SolrQueryParser configuration: defaultOperator="AND|OR" -->
 <solrQueryParser defaultOperator="AND"/>


<?xml version="1.0" encoding="UTF-8" ?>
 Licensed to the Apache Software Foundation (ASF) under one or more
 contributor license agreements.  See the NOTICE file distributed with
 this work for additional information regarding copyright ownership.
 The ASF licenses this file to You under the Apache License, Version 2.0
 (the "License"); you may not use this file except in compliance with
 the License.  You may obtain a copy of the License at

 Unless required by applicable law or agreed to in writing, software
 distributed under the License is distributed on an "AS IS" BASIS,
 See the License for the specific language governing permissions and
 limitations under the License.
     For more details about configurations options that may appear in this
     file, see

     Specifically, the Solr Config can support XInclude, which may make it
easier to manage
     the configuration.  See
<config name="search-api-1.0">
  <!-- Set this to 'false' if you want solr to continue working after it has
       encountered an severe configuration error.  In a production
       you may want solr to keep working even if one handler is

       You may also set this to false using by setting the system property:


  <!-- lib directives can be used to instruct Solr to load an Jars
       and use them to resolve any "plugins" specified in your
solrconfig.xml or
       schema.xml (ie: Analyzers, Request Handlers, etc...).

       All directories and paths are resolved relative the instanceDir.

       If a "./lib" directory exists in your instanceDir, all files found
in it
       are included as if you had used the following syntax...

              <lib dir="./lib" />

       A dir option by itself adds any files found in the directory to the
       classpath, this is useful for including all jars in a directory.

  <lib dir="../../contrib/extraction/lib" />
       When a regex is specified in addition to a directory, only the files
in that
       directory which completely match the regex (anchored on both ends)
       will be included.

  <lib dir="../../dist/" regex="apache-solr-cell-\d.*\.jar" />
  <lib dir="../../dist/" regex="apache-solr-clustering-\d.*\.jar" />

       If a dir option (with or without a regex) is used and nothing is
       that matches, it will be ignored

  <lib dir="../../contrib/clustering/lib/downloads/" />
  <lib dir="../../contrib/clustering/lib/" />
  <lib dir="/total/crap/dir/ignored" />
       an exact path can be used to specify a specific file.  This will
       a serious error to be logged if it can't be loaded.
  <lib path="../a-jar-that-does-not-exist.jar" />

  <!-- Used to specify an alternate directory to hold all index data
       other than the default ./data under the Solr home.
       If replication is in use, this should match the replication
       Might be problematic for some installs, so we just keep the default.
  <!-- <dataDir>${}</dataDir> -->

  <!-- WARNING: this <indexDefaults> section only provides defaults for
index writers
       in general. See also the <mainIndex> section after that when
changing parameters
       for Solr's main Lucene index. -->
   <!-- Values here affect all index writers and act as a default unless
overridden. -->

    <!-- If both ramBufferSizeMB and maxBufferedDocs is set, then Lucene
will flush
     based on whichever limit is hit first.  -->

    <!-- Sets the amount of RAM that may be used by Lucene indexing
      for buffering added documents and deletions before they are
      flushed to the Directory.  -->
    <!-- <maxMergeDocs>2147483647</maxMergeDocs> -->

     Expert: Turn on Lucene's auto commit capability.  This causes
     segment flushes to write a new lucene index descriptor, enabling it to
     opened by an external IndexReader.  This can greatly slow down indexing
     speed.  NOTE: Despite the name, this value does not have any relation
     Solr's autoCommit functionality

     Expert: The Merge Policy in Lucene controls how merging is handled by
     Lucene.  The default in 2.3 is the LogByteSizeMergePolicy, previous
     versions used LogDocMergePolicy.

     LogByteSizeMergePolicy chooses segments to merge based on their size.
     Lucene 2.2 default, LogDocMergePolicy chose when to merge based on
     of documents

     Other implementations of MergePolicy must have a no-argument

     The Merge Scheduler in Lucene controls how merges are performed. The
     ConcurrentMergeScheduler (Lucene 2.3 default) can perform merges in the
     background using separate threads.  The SerialMergeScheduler (Lucene
     default) does not.

      This option specifies which Lucene LockFactory implementation to use.

      single = SingleInstanceLockFactory - suggested for a read-only index
               or when there is no possibility of another process trying
               to modify the index.
      native = NativeFSLockFactory  - uses OS native file locking
      simple = SimpleFSLockFactory  - uses a plain file for locking

      (For backwards compatibility with Solr 1.2, 'simple' is the default
       if not specified.)
    Controls how often Lucene loads terms into memory -->

    <!-- options specific to the main on-disk lucene index -->
    <!-- Deprecated -->

    <!-- inherit from indexDefaults <maxFieldLength>10000</maxFieldLength>

    <!-- If true, unlock any held write or commit locks on startup.
         This defeats the locking mechanism that allows multiple
         processes to safely access a lucene index, and should be
         used with care.
         This is not needed if lock type is 'none' or 'single'

    <!-- If true, IndexReaders will be reopened (often more efficient)
         of closed and then opened.  -->

    Controls how often Lucene loads terms into memory.  Default is 128 and
is likely good for most everyone. -->

        Custom deletion policies can specified here. The class must
        implement org.apache.lucene.index.IndexDeletionPolicy.

        The standard Solr IndexDeletionPolicy implementation supports
        index commit points on number of commits, age of commit point and
        optimized status.

        The latest commit point should always be preserved regardless
        of the criteria.
    <deletionPolicy class="solr.SolrDeletionPolicy">
      <!-- The number of commit points to be kept -->
      <str name="maxCommitsToKeep">1</str>
      <!-- The number of optimized commit points to be kept -->
      <str name="maxOptimizedCommitsToKeep">0</str>
          Delete all commit points once they have reached the given age.
          Supports DateMathParser syntax e.g.

          <str name="maxCommitAge">30MINUTES</str>
          <str name="maxCommitAge">1DAY</str>

    <!--  To aid in advanced debugging, you may turn on IndexWriter debug
      Setting to true will set the file that the underlying Lucene
      will write its debug infostream to.  -->
     <infoStream file="INFOSTREAM.txt">false</infoStream>


  <!-- Enables JMX if and only if an existing MBeanServer is found. Use this
    if you want to configure JMX through JVM parameters. Remove this to
    exposing Solr configuration and statistics to JMX.

If you want to connect to a particular server, specify the agentId
e.g. <jmx agentId="myAgent" />

If you want to start a new MBeanServer, specify the serviceUrl
e.g <jmx serviceUrl="service:jmx:rmi:///jndi/rmi://localhost:9999/solr"/>

For more details see
  <!-- <jmx /> -->

  <!-- the default high-performance update handler -->
  <updateHandler class="solr.DirectUpdateHandler2">
    <!-- A prefix of "solr." for class names is an alias that
         causes solr to search appropriate packages, including

    <!-- Perform a <commit/> automatically under certain conditions:
         maxDocs - number of updates since last commit is greater than this
         maxTime - oldest uncommited update (in ms) is this long ago
         Instead of enabling autoCommit, consider using "commitWithin"
         when adding documents.

    <!-- The RunExecutableListener executes an external command from a
      hook such as postCommit or postOptimize.
         exe - the name of the executable to run
         dir - dir to use as the current working directory. default="."
         wait - the calling thread waits until the executable returns.
         args - the arguments to pass to the program.  default=nothing
         env - environment variables to set.  default=nothing
    <!-- A postCommit event is fired after every commit or optimize command
    <listener event="postCommit" class="solr.RunExecutableListener">
      <str name="exe">solr/bin/snapshooter</str>
      <str name="dir">.</str>
      <bool name="wait">true</bool>
      <arr name="args"> <str>arg1</str> <str>arg2</str> </arr>
      <arr name="env"> <str>MYVAR=val1</str> </arr>
    <!-- A postOptimize event is fired only after every optimize command
    <listener event="postOptimize" class="solr.RunExecutableListener">
      <str name="exe">snapshooter</str>
      <str name="dir">solr/bin</str>
      <bool name="wait">true</bool>


  <!-- Use the following format to specify a custom IndexReaderFactory -
allows for alternate
       IndexReader implementations.

       ** Experimental Feature **
       Please note - Using a custom IndexReaderFactory may prevent certain
other features
       from working. The API to IndexReaderFactory may change without
warning or may even
       be removed from future releases if the problems cannot be resolved.

       ** Features that may not work with custom IndexReaderFactory **
       The ReplicationHandler assumes a disk-resident index. Using a custom
       IndexReader implementation may cause incompatibility with
ReplicationHandler and
       may cause replication to not work correctly. See SOLR-1366 for

  <indexReaderFactory name="IndexReaderFactory" class="package.class">
    Parameters as required by the implementation
  </indexReaderFactory >
  <!-- To set the termInfosIndexDivisor, do this: -->
  <!--<indexReaderFactory name="IndexReaderFactory"
    <int name="termInfosIndexDivisor">12</int>
  </indexReaderFactory >-->

    <!-- Maximum number of clauses in a boolean query... in the past, this
        range or prefix queries that expanded to big boolean queries -
built in Solr
        query parsers no longer create queries with this limitation.
        An exception is thrown if exceeded.  -->

    <!-- There are two implementations of cache available for Solr,
         LRUCache, based on a synchronized LinkedHashMap, and
         FastLRUCache, based on a ConcurrentHashMap.  FastLRUCache has
faster gets
         and slower puts in single threaded operation and thus is generally
         than LRUCache when the hit ratio of the cache is high (> 75%), and
may be
         faster under other scenarios on multi-cpu systems. -->
    <!-- Cache used by SolrIndexSearcher for filters (DocSets),
         unordered sets of *all* documents that match a query.
         When a new searcher is opened, its caches may be prepopulated
         or "autowarmed" using data from caches in the old searcher.
         autowarmCount is the number of items to prepopulate.  For LRUCache,
         the autowarmed items will be the most recently accessed items.
         class - the SolrCache implementation LRUCache or FastLRUCache
         size - the maximum number of entries in the cache
         initialSize - the initial capacity (number of entries) of
           the cache.  (seel java.util.HashMap)
         autowarmCount - the number of entries to prepopulate from
           and old cache.

    <!-- Cache used to hold field values that are quickly accessible
         by document id.  The fieldValueCache is created by default
         even if not configured here.

   <!-- queryResultCache caches results of searches - ordered lists of
         document ids (DocList) based on a query, a sort, and the range
         of documents requested.  -->

  <!-- documentCache caches Lucene Document objects (the stored fields for
each document).
       Since Lucene internal document ids are transient, this cache will
not be autowarmed.  -->

    <!-- If true, stored fields that are not requested will be loaded
      This can result in a significant speed improvement if the usual case
is to
      not load all stored fields, especially if the skipped fields are large
      compressed text fields.

    <!-- Example of a generic cache.  These caches may be accessed by name
         through SolrIndexSearcher.getCache(),cacheLookup(), and
         The purpose is to enable easy caching of user/application level
         The regenerator argument should be specified as an implementation
         of if autowarming is desired.  -->
    <cache name="myUserCache"

   <!-- An optimization that attempts to use a filter to satisfy a search.
         If the requested sort does not include score, then the filterCache
         will be checked for a filter matching the query. If found, the
         will be used as the source of document ids, and then the sort will
         applied to that.

   <!-- An optimization for use with the queryResultCache.  When a search
         is requested, a superset of the requested number of document ids
         are collected.  For example, if a search for a particular query
         requests matching documents 10 through 19, and queryWindowSize is
         then documents 0 through 49 will be collected and cached.  Any
         requests in that range can be satisfied via the cache.  -->

    <!-- Maximum number of documents to cache for any entry in the
         queryResultCache. -->

    <!-- a newSearcher event is fired whenever a new searcher is being
      and there is a current searcher handling requests (aka registered).
      It can be used to prime certain caches to prevent long request times
      certain requests.
    <!-- QuerySenderListener takes an array of NamedList and executes a
         local query request for each NamedList in sequence. -->
    <listener event="newSearcher" class="solr.QuerySenderListener">
      <arr name="queries">
        <lst><str name="q">solr</str> <str name="start">0</str> <str
        <lst><str name="q">rocks</str> <str name="start">0</str> <str
        <lst><str name="q">static newSearcher warming query from

    <!-- a firstSearcher event is fired whenever a new searcher is being
         prepared but there is no current registered searcher to handle
         requests or to gain autowarming data from. -->
    <listener event="firstSearcher" class="solr.QuerySenderListener">
      <arr name="queries">
        <lst><str name="q">solr rocks</str><str name="start">0</str><str
        <lst><str name="q">static firstSearcher warming query from

    <!-- If a search request comes in and there is no current registered
         then immediately register the still warming searcher and use it.
         "false" then all requests will block until the first searcher is
         warming. -->

    <!-- Maximum number of searchers that may be warming in the background
      concurrently.  An error is returned if this limit is exceeded.
      1-2 for read-only slaves, higher for masters w/o cache warming. -->


    Let the dispatch filter handler /select?qt=XXX
    handleSelect=true will use consistent error handling for /select and
    handleSelect=false will use solr1.1 style error formatting
  <requestDispatcher handleSelect="true" >
    <!--Make sure your system has some authentication before enabling
remote streaming!  -->
    <requestParsers enableRemoteStreaming="false"
multipartUploadLimitInKB="2048000" />

    <!-- Set HTTP caching related parameters (for proxy caches and clients).

         To get the behaviour of Solr 1.2 (ie: no caching related headers)
         use the never304="true" option and do not specify a value for
    <!-- <httpCaching never304="true"> -->
    <httpCaching lastModifiedFrom="openTime"
       <!-- lastModFrom="openTime" is the default, the Last-Modified value
            (and validation against If-Modified-Since requests) will all be
            relative to when the current Searcher was opened.
            You can change it to lastModFrom="dirLastMod" if you want the
            value to exactly corrispond to when the physical index was last

            etagSeed="..." is an option you can change to force the ETag
            header (and validation against If-None-Match requests) to be
            differnet even if the index has not changed (ie: when making
            significant changes to your config file)

            lastModifiedFrom and etagSeed are both ignored if you use the
            never304="true" option.
       <!-- If you include a <cacheControl> directive, it will be used to
            generate a Cache-Control header, as well as an Expires header
            if the value contains "max-age="

            By default, no Cache-Control header is generated.

            You can use the <cacheControl> option even if you have set
       <!-- <cacheControl>max-age=30, public</cacheControl> -->

  <!-- requestHandler plugins... incoming queries will be dispatched to the
     correct handler based on the path or the qt (query type) param.
     Names starting with a '/' are accessed with the a path equal to the
     registered name.  Names without a leading '/' are accessed with:
     If no qt is defined, the requestHandler that declares default="true"
     will be used.
  <requestHandler name="standard" class="solr.SearchHandler">
    <!-- default values for query parameters -->
    <lst name="defaults">
      <str name="echoParams">none</str>
      <str name="fl">item_id,score</str>
      <str name="q.op">AND</str>
      <str name="q.alt">*:*</str>

      <!-- By default don't spell check. -->
      <str name="spellcheck">false</str>
      <!-- Defaults for the spell checker when used. -->
      <str name="spellcheck.onlyMorePopular">true</str>
      <str name="spellcheck.extendedResults">false</str>
      <!-- The number of suggestions to return. -->
      <str name="spellcheck.count">1</str>

      <!-- Highlighting parameters -->
      <!-- By default, don't highlight. -->
      <str name="hl">false</str>
      <!-- Use the spellcheck field for highlighting, as that contains all
           data anyways. -->
      <str name="hl.fl">spell</str>
      <!-- Use special syntax for highlighting, so we can later strip other
           tags from the excerpt. -->
      <str name="hl.simple.pre">[HIGHLIGHT]</str>
      <str name="">[/HIGHLIGHT]</str>
      <!-- Use several shorter snippets and concatenate them. -->
      <str name="hl.snippets">3</str>
      <str name="hl.fragsize">70</str>
      <str name="hl.mergeContiguous">true</str>
    <arr name="last-components">

<!-- Please refer to for
details on configuring replication -->
<!-- remove the <lst name="master"> section if this is just a slave -->
<!-- remove  the <lst name="slave"> section if this is just a master -->
<requestHandler name="/replication" class="solr.ReplicationHandler" >
    <lst name="master">
      <str name="replicateAfter">commit</str>
      <str name="replicateAfter">startup</str>
      <str name="confFiles">schema.xml,stopwords.txt</str>
    <lst name="slave">
      <str name="masterUrl">http://localhost:8983/solr/replication</str>
      <str name="pollInterval">00:00:60</str>

  <!-- DisMaxRequestHandler allows easy searching across multiple fields
       for simple user-entered phrases.  It's implementation is now
       just the standard SearchHandler with a default query type
       of "dismax".
  <requestHandler name="dismax" class="solr.SearchHandler" default="true">
    <lst name="defaults">
     <str name="defType">dismax</str>
     <str name="echoParams">none</str>
     <float name="tie">0.01</float>
 <!--<str name="qf">
        text^0.5 features^1.0 name^1.2 sku^1.5 id^10.0 manu^1.1 cat^1.4
 <!--<str name="pf">
        text^0.2 features^1.1 name^1.5 manu^1.4 manu_exact^1.9
 <!--<str name="bf">
        popularity^0.5 recip(price,1,1000,1000)^0.3

     If you want to get more results back from Search API queries, just add
     additional fields here (after setting them to "stored" in the
     schema.xml). These will be returned alongside "id" and "score" in the
     search results.
     <str name="fl">item_id,score </str>
     If you want to allow results where not all keywords are matched,
     the following line and insert your own "minimum match" definition.
     <!-- <str name="mm">2&lt;-35%</str> -->
     <int name="ps">100</int>
     <str name="q.alt">*:*</str>

      <!-- Highlighting parameters -->
      <!-- By default, don't highlight. -->
      <str name="hl">false</str>
      <!-- Use the spellcheck field for highlighting, as that contains all
           data anyways. -->
      <str name="hl.fl">spell</str>
      <!-- Use special syntax for highlighting, so we can later strip other
           tags from the excerpt. -->
      <str name="hl.simple.pre">[HIGHLIGHT]</str>
      <str name="">[/HIGHLIGHT]</str>
      <!-- Use several shorter snippets and concatenate them. -->
      <str name="hl.snippets">3</str>
      <str name="hl.fragsize">70</str>
      <str name="hl.mergeContiguous">true</str>

      <!-- By default don't spell check. -->
      <str name="spellcheck">false</str>
      <!-- Defaults for the spell checker when used. -->
      <str name="spellcheck.onlyMorePopular">true</str>
      <str name="spellcheck.extendedResults">false</str>
      <!-- The number of suggestions to return. -->
      <str name="spellcheck.count">1</str>
    <arr name="last-components">

  <!-- Note how you can register the same handler multiple times with
       different names (and different init parameters)
  <requestHandler name="partitioned" class="solr.SearchHandler" >
    <lst name="defaults">
     <str name="defType">dismax</str>
     <str name="echoParams">explicit</str>
     <str name="qf">text^0.5 features^1.0 name^1.2 sku^1.5 id^10.0</str>
     <str name="mm">2&lt;-1 5&lt;-2 6&lt;90%</str>
     <!-- This is an example of using Date Math to specify a constantly
          moving date range in a config...
     <str name="bq">incubationdate_dt:[* TO NOW/DAY-1MONTH]^2.2</str>
    <!-- In addition to defaults, "appends" params can be specified
         to identify values which should be appended to the list of
         multi-val params from the query (or the existing "defaults").

         In this example, the param "fq=instock:true" will be appended to
         any query time fq params the user may specify, as a mechanism for
         partitioning the index, independent of any user selected filtering
         that may also be desired (perhaps as a result of faceted

         NOTE: there is *absolutely* nothing a client can do to prevent
         "appends" values from being used, so don't use this mechanism
         unless you are sure you always want it.
    <lst name="appends">
      <str name="fq">inStock:true</str>
    <!-- "invariants" are a way of letting the Solr maintainer lock down
         the options available to Solr clients.  Any params values
         specified here are used regardless of what values may be specified
         in either the query, the "defaults", or the "appends" params.

         In this example, the facet.field and facet.query params are fixed,
         limiting the facets clients can use.  Faceting is not turned on by
         default - but if the client does specify facet=true in the request,
         these are the only facets they will be able to see counts for;
         regardless of what other facet.field or facet.query params they
         may specify.

         NOTE: there is *absolutely* nothing a client can do to prevent
         "invariants" values from being used, so don't use this mechanism
         unless you are sure you always want it.
    <lst name="invariants">
      <str name="facet.field">cat</str>
      <str name="facet.field">manu_exact</str>
      <str name="facet.query">price:[* TO 500]</str>
      <str name="facet.query">price:[500 TO *]</str>

  <!-- The more like this handler offers many advantages over the standard
    when performing moreLikeThis requests.-->
  <requestHandler name="mlt" class="solr.MoreLikeThisHandler">
    <lst name="defaults">
      <str name="mlt.mintf">1</str>
      <str name="mlt.mindf">1</str>
      <str name="mlt.minwl">3</str>
      <str name="mlt.maxwl">50</str>
      <str name="mlt.maxqt">20</str>
      <str name="mlt.match.include">false</str>
      <str name="fl">item_id,score </str>
      <!-- Abort any searches longer than 1.5 seconds -->
      <!-- <int name="timeAllowed">1500</int> -->

   Search components are registered to SolrCore and used by Search Handlers

   By default, the following components are available:

   <searchComponent name="query"
class="org.apache.solr.handler.component.QueryComponent" />
   <searchComponent name="facet"
class="org.apache.solr.handler.component.FacetComponent" />
   <searchComponent name="mlt"
class="org.apache.solr.handler.component.MoreLikeThisComponent" />
   <searchComponent name="highlight"
class="org.apache.solr.handler.component.HighlightComponent" />
   <searchComponent name="stats"
class="org.apache.solr.handler.component.StatsComponent" />
   <searchComponent name="debug"
class="org.apache.solr.handler.component.DebugComponent" />

   Default configuration in a requestHandler would look like:
    <arr name="components">

    If you register a searchComponent to one of the standard names, that
will be used instead.
    To insert components before or after the 'standard' components, use:

    <arr name="first-components">

    <arr name="last-components">

   <!-- The spell check component can return a list of alternative spelling
  suggestions.  -->
  <searchComponent name="spellcheck" class="solr.SpellCheckComponent">

    <str name="queryAnalyzerFieldType">textSpell</str>

    <lst name="spellchecker">
      <str name="name">default</str>
      <str name="field">spell</str>
      <str name="spellcheckIndexDir">./spellchecker</str>
      <str name="buildOnOptimize">true</str>

    <!-- a spellchecker that uses a different distance measure -->
    <lst name="spellchecker">
      <str name="name">jarowinkler</str>
      <str name="field">spell</str>
      <str name="spellcheckIndexDir">./spellchecker2</str>
      <str name="buildOnOptimize">true</str>

    <!-- a file based spell checker
    <lst name="spellchecker">
      <str name="classname">solr.FileBasedSpellChecker</str>
      <str name="name">file</str>
      <str name="sourceLocation">spellings.txt</str>
      <str name="characterEncoding">UTF-8</str>
      <str name="spellcheckIndexDir">./spellcheckerFile</str>

  <queryConverter name="queryConverter"

  <!-- A request handler utilizing the spellcheck component.

  NOTE: This is purely as an example.  The whole purpose of the
  SpellCheckComponent is to hook it into the request handler that handles
  the standard or dismax SearchHandler) queries such that a separate
request is
  not needed to get suggestions.


<!-- <requestHandler name="/spell" class="solr.SearchHandler" lazy="true">
    <lst name="defaults">
      --><!-- omp = Only More Popular --><!--
      <str name="spellcheck.onlyMorePopular">false</str>
      --><!-- exr = Extended Results --><!--
      <str name="spellcheck.extendedResults">false</str>
      --><!--  The number of suggestions to return --><!--
      <str name="spellcheck.count">1</str>
    <arr name="last-components">
  </requestHandler> -->

  <searchComponent name="tvComponent"
  <!-- A Req Handler for working with the tvComponent.  This is purely as
an example.
  You will likely want to add the component to your already specified
request handlers. -->
<!-- <requestHandler name="tvrh"
    <lst name="defaults">
      <bool name="tv">true</bool>
    <arr name="last-components">
  </requestHandler> -->

  <!-- Clustering Component
       This relies on third party jars which are not included in the
       To use this component (and the "/clustering" handler),
       those jars will need to be downloaded, and you'll need to set the
       solr.cluster.enabled system property when running solr...
          java -Dsolr.clustering.enabled=true -jar start.jar
<!-- <searchComponent
    class="org.apache.solr.handler.clustering.ClusteringComponent" >
    --><!-- Declare an engine --><!--
    <lst name="engine">
      --><!-- The name, only one can be named "default" --><!--
      <str name="name">default</str>
           Class name of Carrot2 clustering algorithm. Currently available
algorithms are:

           * org.carrot2.clustering.lingo.LingoClusteringAlgorithm

           See for the
algorithm's characteristics.
           Overriding values for Carrot2 default algorithm attributes. For
a description
           of all available attributes, see:
           Use attribute key as name attribute of str elements below. These
can be further
           overridden for individual requests by specifying attribute key
as request
           parameter name and attribute value as parameter value.
      <str name="LingoClusteringAlgorithm.desiredClusterCountBase">20</str>
    <lst name="engine">
      <str name="name">stc</str>
  <requestHandler name="/clustering"
     <lst name="defaults">
       <bool name="clustering">true</bool>
       <str name="clustering.engine">default</str>
       <bool name="clustering.results">true</bool>
       --><!-- The title field --><!--
       <str name="carrot.title">name</str>
       <str name="carrot.url">id</str>
       --><!-- The field to cluster on --><!--
       <str name="carrot.snippet">features</str>
       --><!-- produce summaries --><!--
       <bool name="carrot.produceSummary">true</bool>
       --><!-- the maximum number of labels per cluster -->
       <!--<int name="carrot.numDescriptions">5</int>-->
       <!-- produce sub clusters --><!--
       <bool name="carrot.outputSubClusters">false</bool>
    <arr name="last-components">

  <!-- Solr Cell: -->
<!--<requestHandler name="/update/extract"
    <lst name="defaults">
      --><!-- All the main content goes into "text"... if you need to return
           the extracted text or do highlighting, use a stored field.
      <str name="fmap.content">text</str>
      <str name="lowernames">true</str>
      <str name="uprefix">ignored_</str>

      --><!-- capture link hrefs but ignore div attributes --><!--
      <str name="captureAttr">true</str>
      <str name="fmap.a">links</str>
      <str name="fmap.div">ignored_</str>

  <!-- A component to return terms and document frequency of those terms.
       This component does not yet support distributed search. -->
  <searchComponent name="termsComponent"

  <requestHandler name="/terms"
     <lst name="defaults">
      <bool name="terms">true</bool>
    <arr name="components">

  <!-- a search component that enables you to configure the top results for
       a given query regardless of the normal lucene scoring.-->
<!-- <searchComponent name="elevator" class="solr.QueryElevationComponent" >
    --><!-- pick a fieldType to analyze queries --><!--
    <str name="queryFieldType">string</str>
    <str name="config-file">elevate.xml</str>

  <!-- a request handler utilizing the elevator component -->
<!-- <requestHandler name="/elevate" class="solr.SearchHandler"
    <lst name="defaults">
      <str name="echoParams">explicit</str>
    <arr name="last-components">

  <!-- Update request handler.

       Note: Since solr1.1 requestHandlers require a valid content type
header if posted in
       the body. For example, curl now requires: -H 'Content-type:text/xml;
       The response format differs from solr1.1 formatting and returns a
standard error code.
       To enable solr1.1 behavior, remove the /update handler or change its
  <requestHandler name="/update" class="solr.XmlUpdateRequestHandler" />

<!--  <requestHandler name="/update/javabin"
class="solr.BinaryUpdateRequestHandler" />-->

   Analysis request handler.  Since Solr 1.3.  Use to return how a document
is analyzed.  Useful
   for debugging and as a token server for other types of applications.

   This is deprecated in favor of the improved
DocumentAnalysisRequestHandler and FieldAnalysisRequestHandler

   <requestHandler name="/analysis" class="solr.AnalysisRequestHandler" />

    An analysis handler that provides a breakdown of the analysis process
of provided documents.
    This handler expects a (single) content stream with the following

        <field name="id">1</field>
        <field name="name">The Name</field>
        <field name="text">The Text Value</field>

    Note: Each document must contain a field which serves as the unique
key. This key is used in the returned
    response to assoicate an analysis breakdown to the analyzed document.

    Like the FieldAnalysisRequestHandler, this handler also supports query
analysis by
    sending either an "analysis.query" or "q" request paraemter that holds
the query text to be analyized. It also
    supports the "analysis.showmatch" parameter which when set to true, all
field tokens that match the query
    tokens will be marked as a "match".
<!--  <requestHandler name="/analysis/document"
class="solr.DocumentAnalysisRequestHandler" />-->

    RequestHandler that provides much the same functionality as
analysis.jsp. Provides the ability
    to specify multiple field types and field names in the same request and
outputs index-time and
    query-time analysis for each of them.

    Request parameters are:
    analysis.fieldname - The field name whose analyzers are to be used
    analysis.fieldtype - The field type whose analyzers are to be used
    analysis.fieldvalue - The text for index-time analysis
    q (or analysis.q) - The text for query time analysis
    analysis.showmatch (true|false) - When set to true and when query
analysis is performed, the produced
                                      tokens of the field value analysis
will be marked as "matched" for every
                                      token that is produces by the query
<!--  <requestHandler name="/analysis/field"
class="solr.FieldAnalysisRequestHandler" />-->

  <!-- CSV update handler, loaded on demand -->
<!--  <requestHandler name="/update/csv" class="solr.CSVRequestHandler"
startup="lazy" />-->

   Admin Handlers - This will register all the standard admin
RequestHandlers.  Adding
   this single handler is equivalent to registering:

  <requestHandler name="/admin/luke"
class="org.apache.solr.handler.admin.LukeRequestHandler" />
  <requestHandler name="/admin/system"
class="org.apache.solr.handler.admin.SystemInfoHandler" />
  <requestHandler name="/admin/plugins"
 class="org.apache.solr.handler.admin.PluginInfoHandler" />
  <requestHandler name="/admin/threads"
 class="org.apache.solr.handler.admin.ThreadDumpHandler" />
  <requestHandler name="/admin/properties"
class="org.apache.solr.handler.admin.PropertiesRequestHandler" />
  <requestHandler name="/admin/file"
class="org.apache.solr.handler.admin.ShowFileRequestHandler" >

  If you wish to hide files under ${solr.home}/conf, explicitly register
the ShowFileRequestHandler using:
  <requestHandler name="/admin/file"
class="org.apache.solr.handler.admin.ShowFileRequestHandler" >
    <lst name="invariants">
     <str name="hidden">synonyms.txt</str>
     <str name="hidden">anotherfile.txt</str>
  <requestHandler name="/admin/"
class="org.apache.solr.handler.admin.AdminHandlers" />

  <!-- ping/healthcheck -->
  <requestHandler name="/admin/ping" class="PingRequestHandler">
    <lst name="defaults">
      <str name="qt">standard</str>
      <str name="q">solrpingquery</str>
      <str name="echoParams">all</str>

  <!-- Echo the request contents back to the client -->
<!-- <requestHandler name="/debug/dump" class="solr.DumpRequestHandler" >
    <lst name="defaults">
     <str name="echoParams">explicit</str> --><!-- for all params
(including the default etc) use: 'all' --><!--
     <str name="echoHandler">true</str>
  </requestHandler> -->

   <!-- Configure the standard fragmenter -->
   <!-- This could most likely be commented out in the "default" case -->
   <fragmenter name="gap" class="org.apache.solr.highlight.GapFragmenter"
    <lst name="defaults">
     <int name="hl.fragsize">100</int>

   <!-- A regular-expression-based fragmenter (e.g., for sentence
extraction) -->
   <fragmenter name="regex"
    <lst name="defaults">
      <!-- slightly smaller fragsizes work better because of slop -->
      <int name="hl.fragsize">70</int>
      <!-- allow 50% slop on fragment sizes -->
      <float name="hl.regex.slop">0.5</float>
      <!-- a basic sentence pattern -->
      <str name="hl.regex.pattern">[-\w ,/\n\"']{20,200}</str>

   <!-- Configure the standard formatter -->
   <formatter name="html" class="org.apache.solr.highlight.HtmlFormatter"
    <lst name="defaults">
     <str name="hl.simple.pre"><![CDATA[<em>]]></str>
     <str name=""><![CDATA[</em>]]></str>

  <!-- An example dedup update processor that creates the "id" field on the
       based on the hash code of some other fields.  This example has
       set to false since we are using the id field as the signatureField
and Solr
       will maintain uniqueness based on that anyway.

       You have to link the chain to an update handler above to use it ie:
         <requestHandler name="/update
           <lst name="defaults">
             <str name="update.processor">dedupe</str>
  <updateRequestProcessorChain name="dedupe">
      <bool name="enabled">true</bool>
      <str name="signatureField">id</str>
      <bool name="overwriteDupes">false</bool>
      <str name="fields">name,features,cat</str>
    <processor class="solr.LogUpdateProcessorFactory" />
    <processor class="solr.RunUpdateProcessorFactory" />

  <!-- queryResponseWriter plugins... query responses will be written using
    writer specified by the 'wt' request parameter matching the name of a
    The "default" writer is the default and will be used if 'wt' is not
    in the request. XMLResponseWriter will be used if nothing is specified
    The json, python, and ruby writers are also available by default.

    <queryResponseWriter name="xml"
class="org.apache.solr.request.XMLResponseWriter" default="true"/>
    <queryResponseWriter name="json"
    <queryResponseWriter name="python"
    <queryResponseWriter name="ruby"
    <queryResponseWriter name="php"
    <queryResponseWriter name="phps"

    <queryResponseWriter name="custom"

  <!-- XSLT response writer transforms the XML output by any xslt file found
       in Solr's conf/xslt directory.  Changes to xslt files are checked for
       every xsltCacheLifetimeSeconds.
  <queryResponseWriter name="xslt"
    <int name="xsltCacheLifetimeSeconds">5</int>

  <!-- example of registering a query parser
  <queryParser name="lucene"

  <!-- example of registering a custom function parser
  <valueSourceParser name="myfunc"
class="com.mycompany.MyValueSourceParser" />

  <!-- config for the admin interface -->

    <!-- configure a healthcheck file for servers behind a loadbalancer
    <healthcheck type="file">server-enabled</healthcheck>


  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message