SOLR-13952: Separate out Gradle-specific code from other (mostly test) changes and commit separately
This commit is contained in:
Родитель
c1c05b5175
Коммит
872e6b0a1d
|
@ -33,8 +33,6 @@ import org.apache.lucene.analysis.util.ResourceLoaderAware;
|
|||
import org.apache.lucene.analysis.util.TokenFilterFactory;
|
||||
import org.apache.lucene.search.PhraseQuery;
|
||||
|
||||
import static org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.*;
|
||||
|
||||
/**
|
||||
* Factory for {@link WordDelimiterFilter}.
|
||||
* <pre class="prettyprint">
|
||||
|
@ -76,31 +74,31 @@ public class WordDelimiterFilterFactory extends TokenFilterFactory implements Re
|
|||
super(args);
|
||||
int flags = 0;
|
||||
if (getInt(args, "generateWordParts", 1) != 0) {
|
||||
flags |= GENERATE_WORD_PARTS;
|
||||
flags |= org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.GENERATE_WORD_PARTS;
|
||||
}
|
||||
if (getInt(args, "generateNumberParts", 1) != 0) {
|
||||
flags |= GENERATE_NUMBER_PARTS;
|
||||
flags |= org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.GENERATE_NUMBER_PARTS;
|
||||
}
|
||||
if (getInt(args, "catenateWords", 0) != 0) {
|
||||
flags |= CATENATE_WORDS;
|
||||
flags |= org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.CATENATE_WORDS;
|
||||
}
|
||||
if (getInt(args, "catenateNumbers", 0) != 0) {
|
||||
flags |= CATENATE_NUMBERS;
|
||||
flags |= org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.CATENATE_NUMBERS;
|
||||
}
|
||||
if (getInt(args, "catenateAll", 0) != 0) {
|
||||
flags |= CATENATE_ALL;
|
||||
flags |= org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.CATENATE_ALL;
|
||||
}
|
||||
if (getInt(args, "splitOnCaseChange", 1) != 0) {
|
||||
flags |= SPLIT_ON_CASE_CHANGE;
|
||||
flags |= org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.SPLIT_ON_CASE_CHANGE;
|
||||
}
|
||||
if (getInt(args, "splitOnNumerics", 1) != 0) {
|
||||
flags |= SPLIT_ON_NUMERICS;
|
||||
flags |= org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.SPLIT_ON_NUMERICS;
|
||||
}
|
||||
if (getInt(args, "preserveOriginal", 0) != 0) {
|
||||
flags |= PRESERVE_ORIGINAL;
|
||||
flags |= org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.PRESERVE_ORIGINAL;
|
||||
}
|
||||
if (getInt(args, "stemEnglishPossessive", 1) != 0) {
|
||||
flags |= STEM_ENGLISH_POSSESSIVE;
|
||||
flags |= org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.STEM_ENGLISH_POSSESSIVE;
|
||||
}
|
||||
wordFiles = get(args, PROTECTED_TOKENS);
|
||||
types = get(args, TYPES);
|
||||
|
@ -162,17 +160,17 @@ public class WordDelimiterFilterFactory extends TokenFilterFactory implements Re
|
|||
|
||||
private Byte parseType(String s) {
|
||||
if (s.equals("LOWER"))
|
||||
return LOWER;
|
||||
return org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.LOWER;
|
||||
else if (s.equals("UPPER"))
|
||||
return UPPER;
|
||||
return org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.UPPER;
|
||||
else if (s.equals("ALPHA"))
|
||||
return ALPHA;
|
||||
return org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.ALPHA;
|
||||
else if (s.equals("DIGIT"))
|
||||
return DIGIT;
|
||||
return org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.DIGIT;
|
||||
else if (s.equals("ALPHANUM"))
|
||||
return ALPHANUM;
|
||||
return org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.ALPHANUM;
|
||||
else if (s.equals("SUBWORD_DELIM"))
|
||||
return SUBWORD_DELIM;
|
||||
return org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.SUBWORD_DELIM;
|
||||
else
|
||||
return null;
|
||||
}
|
||||
|
|
|
@ -54,6 +54,7 @@ import static org.apache.lucene.analysis.miscellaneous.WordDelimiterIterator.DEF
|
|||
* TODO: should explicitly test things like protWords and not rely on
|
||||
* the factory tests in Solr.
|
||||
*/
|
||||
@SuppressWarnings("deprecation")
|
||||
public class TestWordDelimiterFilter extends BaseTokenStreamTestCase {
|
||||
|
||||
/*
|
||||
|
|
|
@ -45,6 +45,7 @@ public class TestWordnetSynonymParser extends BaseTokenStreamTestCase {
|
|||
analyzer.close();
|
||||
|
||||
analyzer = new Analyzer() {
|
||||
@SuppressWarnings("deprecation")
|
||||
@Override
|
||||
protected TokenStreamComponents createComponents(String fieldName) {
|
||||
Tokenizer tokenizer = new MockTokenizer(MockTokenizer.WHITESPACE, false);
|
||||
|
|
|
@ -1,95 +0,0 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.lucene.util;
|
||||
|
||||
public class TestVirtualMethod extends LuceneTestCase {
|
||||
|
||||
private static final VirtualMethod<TestVirtualMethod> publicTestMethod =
|
||||
new VirtualMethod<>(TestVirtualMethod.class, "publicTest", String.class);
|
||||
private static final VirtualMethod<TestVirtualMethod> protectedTestMethod =
|
||||
new VirtualMethod<>(TestVirtualMethod.class, "protectedTest", int.class);
|
||||
|
||||
public void publicTest(String test) {}
|
||||
protected void protectedTest(int test) {}
|
||||
|
||||
static class TestClass1 extends TestVirtualMethod {
|
||||
@Override
|
||||
public void publicTest(String test) {}
|
||||
@Override
|
||||
protected void protectedTest(int test) {}
|
||||
}
|
||||
|
||||
static class TestClass2 extends TestClass1 {
|
||||
@Override // make it public here
|
||||
public void protectedTest(int test) {}
|
||||
}
|
||||
|
||||
static class TestClass3 extends TestClass2 {
|
||||
@Override
|
||||
public void publicTest(String test) {}
|
||||
}
|
||||
|
||||
static class TestClass4 extends TestVirtualMethod {
|
||||
}
|
||||
|
||||
static class TestClass5 extends TestClass4 {
|
||||
}
|
||||
|
||||
public void testGeneral() {
|
||||
assertEquals(0, publicTestMethod.getImplementationDistance(this.getClass()));
|
||||
assertEquals(1, publicTestMethod.getImplementationDistance(TestClass1.class));
|
||||
assertEquals(1, publicTestMethod.getImplementationDistance(TestClass2.class));
|
||||
assertEquals(3, publicTestMethod.getImplementationDistance(TestClass3.class));
|
||||
assertFalse(publicTestMethod.isOverriddenAsOf(TestClass4.class));
|
||||
assertFalse(publicTestMethod.isOverriddenAsOf(TestClass5.class));
|
||||
|
||||
assertEquals(0, protectedTestMethod.getImplementationDistance(this.getClass()));
|
||||
assertEquals(1, protectedTestMethod.getImplementationDistance(TestClass1.class));
|
||||
assertEquals(2, protectedTestMethod.getImplementationDistance(TestClass2.class));
|
||||
assertEquals(2, protectedTestMethod.getImplementationDistance(TestClass3.class));
|
||||
assertFalse(protectedTestMethod.isOverriddenAsOf(TestClass4.class));
|
||||
assertFalse(protectedTestMethod.isOverriddenAsOf(TestClass5.class));
|
||||
|
||||
assertTrue(VirtualMethod.compareImplementationDistance(TestClass3.class, publicTestMethod, protectedTestMethod) > 0);
|
||||
assertEquals(0, VirtualMethod.compareImplementationDistance(TestClass5.class, publicTestMethod, protectedTestMethod));
|
||||
}
|
||||
|
||||
@SuppressWarnings({"rawtypes","unchecked"})
|
||||
public void testExceptions() {
|
||||
// LuceneTestCase is not a subclass and can never override publicTest(String)
|
||||
expectThrows(IllegalArgumentException.class, () -> {
|
||||
// cast to Class to remove generics:
|
||||
publicTestMethod.getImplementationDistance((Class) LuceneTestCase.class);
|
||||
});
|
||||
|
||||
// Method bogus() does not exist, so IAE should be thrown
|
||||
expectThrows(IllegalArgumentException.class, () -> {
|
||||
new VirtualMethod<>(TestVirtualMethod.class, "bogus");
|
||||
});
|
||||
|
||||
// Method publicTest(String) is not declared in TestClass2, so IAE should be thrown
|
||||
expectThrows(IllegalArgumentException.class, () -> {
|
||||
new VirtualMethod<>(TestClass2.class, "publicTest", String.class);
|
||||
});
|
||||
|
||||
// try to create a second instance of the same baseClass / method combination
|
||||
expectThrows(UnsupportedOperationException.class, () -> {
|
||||
new VirtualMethod<>(TestVirtualMethod.class, "publicTest", String.class);
|
||||
});
|
||||
}
|
||||
|
||||
}
|
|
@ -107,6 +107,7 @@ public class ToParentBlockJoinSortField extends SortField {
|
|||
private FieldComparator<?> getStringComparator(int numHits) {
|
||||
return new FieldComparator.TermOrdValComparator(numHits, getField(), missingValue == STRING_LAST) {
|
||||
|
||||
@SuppressWarnings("deprecation")
|
||||
@Override
|
||||
protected SortedDocValues getSortedDocValues(LeafReaderContext context, String field) throws IOException {
|
||||
SortedSetDocValues sortedSet = DocValues.getSortedSet(context.reader(), field);
|
||||
|
@ -126,6 +127,7 @@ public class ToParentBlockJoinSortField extends SortField {
|
|||
|
||||
private FieldComparator<?> getIntComparator(int numHits) {
|
||||
return new FieldComparator.IntComparator(numHits, getField(), (Integer) missingValue) {
|
||||
@SuppressWarnings("deprecation")
|
||||
@Override
|
||||
protected NumericDocValues getNumericDocValues(LeafReaderContext context, String field) throws IOException {
|
||||
SortedNumericDocValues sortedNumeric = DocValues.getSortedNumeric(context.reader(), field);
|
||||
|
@ -144,6 +146,7 @@ public class ToParentBlockJoinSortField extends SortField {
|
|||
|
||||
private FieldComparator<?> getLongComparator(int numHits) {
|
||||
return new FieldComparator.LongComparator(numHits, getField(), (Long) missingValue) {
|
||||
@SuppressWarnings("deprecation")
|
||||
@Override
|
||||
protected NumericDocValues getNumericDocValues(LeafReaderContext context, String field) throws IOException {
|
||||
SortedNumericDocValues sortedNumeric = DocValues.getSortedNumeric(context.reader(), field);
|
||||
|
@ -162,6 +165,7 @@ public class ToParentBlockJoinSortField extends SortField {
|
|||
|
||||
private FieldComparator<?> getFloatComparator(int numHits) {
|
||||
return new FieldComparator.FloatComparator(numHits, getField(), (Float) missingValue) {
|
||||
@SuppressWarnings("deprecation")
|
||||
@Override
|
||||
protected NumericDocValues getNumericDocValues(LeafReaderContext context, String field) throws IOException {
|
||||
SortedNumericDocValues sortedNumeric = DocValues.getSortedNumeric(context.reader(), field);
|
||||
|
@ -186,6 +190,7 @@ public class ToParentBlockJoinSortField extends SortField {
|
|||
|
||||
private FieldComparator<?> getDoubleComparator(int numHits) {
|
||||
return new FieldComparator.DoubleComparator(numHits, getField(), (Double) missingValue) {
|
||||
@SuppressWarnings("deprecation")
|
||||
@Override
|
||||
protected NumericDocValues getNumericDocValues(LeafReaderContext context, String field) throws IOException {
|
||||
SortedNumericDocValues sortedNumeric = DocValues.getSortedNumeric(context.reader(), field);
|
||||
|
|
|
@ -83,6 +83,7 @@ public final class UpToTwoPositiveIntOutputs extends Outputs<Object> {
|
|||
}
|
||||
}
|
||||
|
||||
@SuppressWarnings("deprecation")
|
||||
private final static Long NO_OUTPUT = new Long(0);
|
||||
|
||||
private final boolean doShare;
|
||||
|
|
|
@ -42,6 +42,7 @@ import org.apache.lucene.search.SortField;
|
|||
*
|
||||
*
|
||||
*/
|
||||
@SuppressWarnings({"unchecked", "rawtypes", "deprecation"})
|
||||
public abstract class ValueSource {
|
||||
|
||||
/**
|
||||
|
|
|
@ -54,11 +54,13 @@ public class TotalTermFreqValueSource extends ValueSource {
|
|||
return name() + '(' + field + ',' + val + ')';
|
||||
}
|
||||
|
||||
@SuppressWarnings("rawtypes")
|
||||
@Override
|
||||
public FunctionValues getValues(Map context, LeafReaderContext readerContext) throws IOException {
|
||||
return (FunctionValues)context.get(this);
|
||||
}
|
||||
|
||||
@SuppressWarnings({"rawtypes", "unchecked"})
|
||||
@Override
|
||||
public void createWeight(Map context, IndexSearcher searcher) throws IOException {
|
||||
long totalTermFreq = 0;
|
||||
|
|
|
@ -102,6 +102,7 @@ public class WithinPrefixTreeQuery extends AbstractVisitingPrefixTreeQuery {
|
|||
/** Returns a new shape that is larger than shape by at distErr.
|
||||
*/
|
||||
//TODO move this generic code elsewhere? Spatial4j?
|
||||
@SuppressWarnings("deprecation")
|
||||
protected Shape bufferShape(Shape shape, double distErr) {
|
||||
if (distErr <= 0)
|
||||
throw new IllegalArgumentException("distErr must be > 0");
|
||||
|
|
|
@ -38,6 +38,7 @@ import org.apache.lucene.util.IOUtils;
|
|||
import org.apache.lucene.util.LuceneTestCase;
|
||||
import org.apache.lucene.util.TestUtil;
|
||||
|
||||
@SuppressWarnings("deprecation")
|
||||
public class TestWordBreakSpellChecker extends LuceneTestCase {
|
||||
private Directory dir;
|
||||
private Analyzer analyzer;
|
||||
|
|
|
@ -124,6 +124,7 @@ public class TikaEntityProcessor extends EntityProcessorBase {
|
|||
spatialMetadataField = context.getResolvedEntityAttribute("spatialMetadataField");
|
||||
}
|
||||
|
||||
@SuppressWarnings({"unchecked", "deprecation"})
|
||||
@Override
|
||||
public Map<String, Object> nextRow() {
|
||||
if(done) return null;
|
||||
|
|
|
@ -100,6 +100,7 @@ public class XPathEntityProcessor extends EntityProcessorBase {
|
|||
|
||||
}
|
||||
|
||||
@SuppressWarnings("deprecation")
|
||||
private void initXpathReader(VariableResolver resolver) {
|
||||
reinitXPathReader = false;
|
||||
useSolrAddXml = Boolean.parseBoolean(context
|
||||
|
@ -183,7 +184,7 @@ public class XPathEntityProcessor extends EntityProcessorBase {
|
|||
}
|
||||
}
|
||||
String url = context.getEntityAttribute(URL);
|
||||
List<String> l = url == null ? Collections.EMPTY_LIST : resolver.getVariables(url);
|
||||
List<String> l = url == null ? Collections.emptyList() : resolver.getVariables(url);
|
||||
for (String s : l) {
|
||||
if (s.startsWith(entityName + ".")) {
|
||||
if (placeHolderVariables == null)
|
||||
|
@ -224,7 +225,6 @@ public class XPathEntityProcessor extends EntityProcessorBase {
|
|||
readUsefulVars(r);
|
||||
}
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
private Map<String, Object> fetchNextRow() {
|
||||
Map<String, Object> r = null;
|
||||
while (true) {
|
||||
|
@ -283,6 +283,7 @@ public class XPathEntityProcessor extends EntityProcessorBase {
|
|||
|
||||
}
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
private void initQuery(String s) {
|
||||
Reader data = null;
|
||||
try {
|
||||
|
@ -355,6 +356,7 @@ public class XPathEntityProcessor extends EntityProcessorBase {
|
|||
}
|
||||
}
|
||||
|
||||
@SuppressWarnings({"unchecked", "rawtypes"})
|
||||
protected Map<String, Object> readRow(Map<String, Object> record, String xpath) {
|
||||
if (useSolrAddXml) {
|
||||
List<String> names = (List<String>) record.get("name");
|
||||
|
@ -391,7 +393,6 @@ public class XPathEntityProcessor extends EntityProcessorBase {
|
|||
|
||||
}
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
private Map<String, Object> readUsefulVars(Map<String, Object> r) {
|
||||
Object val = r.get(HAS_MORE);
|
||||
if (val != null)
|
||||
|
|
|
@ -585,6 +585,7 @@ public class XPathRecordReader {
|
|||
* records values. If a fields value is a List then they have to be
|
||||
* deep-copied for thread safety
|
||||
*/
|
||||
@SuppressWarnings("unchecked")
|
||||
private static Map<String, Object> getDeepCopy(Map<String, Object> values) {
|
||||
Map<String, Object> result = new HashMap<>();
|
||||
for (Map.Entry<String, Object> entry : values.entrySet()) {
|
||||
|
|
|
@ -41,7 +41,7 @@ public class TestWriterImpl extends AbstractDataImportHandlerTestCase {
|
|||
}
|
||||
|
||||
@Test
|
||||
@SuppressWarnings("unchecked")
|
||||
@SuppressWarnings({"unchecked", "rawtypes"})
|
||||
public void testDataConfigWithDataSource() throws Exception {
|
||||
List rows = new ArrayList();
|
||||
rows.add(createMap("id", "1", "desc", "one"));
|
||||
|
|
|
@ -36,6 +36,7 @@ import org.junit.Test;
|
|||
*
|
||||
* @since solr 1.3
|
||||
*/
|
||||
@SuppressWarnings({"rawtypes", "unchecked"})
|
||||
public class TestXPathEntityProcessor extends AbstractDataImportHandlerTestCase {
|
||||
boolean simulateSlowReader;
|
||||
boolean simulateSlowResultProcessor;
|
||||
|
@ -93,7 +94,6 @@ public class TestXPathEntityProcessor extends AbstractDataImportHandlerTestCase
|
|||
assertEquals("ü", l.get(2));
|
||||
}
|
||||
|
||||
@SuppressWarnings({"rawtypes", "unchecked"})
|
||||
@Test
|
||||
public void testMultiValuedWithMultipleDocuments() throws Exception {
|
||||
Map entityAttrs = createMap("name", "e", "url", "testdata.xml", XPathEntityProcessor.FOR_EACH, "/documents/doc");
|
||||
|
|
|
@ -29,6 +29,7 @@ import org.junit.Test;
|
|||
*
|
||||
* @since solr 1.3
|
||||
*/
|
||||
@SuppressWarnings({"rawtypes", "unchecked"})
|
||||
public class TestXPathRecordReader extends AbstractDataImportHandlerTestCase {
|
||||
@Test
|
||||
public void testBasic() {
|
||||
|
|
|
@ -93,6 +93,7 @@ public class TestZKPropertiesWriter extends AbstractDataImportHandlerTestCase {
|
|||
zkDir = null;
|
||||
}
|
||||
|
||||
@SuppressWarnings({"unchecked", "rawtypes"})
|
||||
@SuppressForbidden(reason = "Needs currentTimeMillis to construct date stamps")
|
||||
@Test
|
||||
public void testZKPropertiesWriter() throws Exception {
|
||||
|
|
|
@ -85,6 +85,7 @@ public class XLSXResponseWriter extends RawResponseWriter {
|
|||
}
|
||||
}
|
||||
|
||||
@SuppressWarnings("rawtypes")
|
||||
class XLSXWriter extends TabularResponseWriter {
|
||||
|
||||
static class SerialWriteWorkbook {
|
||||
|
@ -144,7 +145,6 @@ class XLSXWriter extends TabularResponseWriter {
|
|||
} catch (IOException e) {
|
||||
StringWriter sw = new StringWriter();
|
||||
e.printStackTrace(new PrintWriter(sw));
|
||||
String stacktrace = sw.toString();
|
||||
}finally {
|
||||
swb.dispose();
|
||||
}
|
||||
|
@ -232,6 +232,7 @@ class XLSXWriter extends TabularResponseWriter {
|
|||
|
||||
//NOTE: a document cannot currently contain another document
|
||||
List tmpList;
|
||||
@SuppressWarnings({"unchecked"})
|
||||
@Override
|
||||
public void writeSolrDocument(String name, SolrDocument doc, ReturnFields returnFields, int idx ) throws IOException {
|
||||
if (tmpList == null) {
|
||||
|
|
|
@ -23,7 +23,6 @@ import java.util.List;
|
|||
|
||||
import org.apache.solr.request.SolrQueryRequest;
|
||||
import org.apache.solr.response.SolrQueryResponse;
|
||||
import org.apache.tika.language.LanguageIdentifier;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
|
@ -44,12 +43,13 @@ public class TikaLanguageIdentifierUpdateProcessor extends LanguageIdentifierUpd
|
|||
super(req, rsp, next);
|
||||
}
|
||||
|
||||
@SuppressWarnings("deprecation")
|
||||
@Override
|
||||
protected List<DetectedLanguage> detectLanguage(Reader solrDocReader) {
|
||||
String content = SolrInputDocumentReader.asString(solrDocReader);
|
||||
List<DetectedLanguage> languages = new ArrayList<>();
|
||||
if (content.length() != 0) {
|
||||
LanguageIdentifier identifier = new LanguageIdentifier(content);
|
||||
org.apache.tika.language.LanguageIdentifier identifier = new org.apache.tika.language.LanguageIdentifier(content);
|
||||
// FIXME: Hack - we get the distance from toString and calculate our own certainty score
|
||||
Double distance = Double.parseDouble(tikaSimilarityPattern.matcher(identifier.toString()).replaceFirst("$1"));
|
||||
// This formula gives: 0.02 => 0.8, 0.1 => 0.5 which is a better sweetspot than isReasonablyCertain()
|
||||
|
|
|
@ -18,7 +18,6 @@ package org.apache.solr.response;
|
|||
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
import java.io.InputStream;
|
||||
import java.io.InputStreamReader;
|
||||
import java.io.StringWriter;
|
||||
import java.io.Writer;
|
||||
|
@ -58,6 +57,7 @@ import org.slf4j.LoggerFactory;
|
|||
|
||||
import static org.apache.solr.common.params.CommonParams.SORT;
|
||||
|
||||
@SuppressWarnings("rawtypes")
|
||||
public class VelocityResponseWriter implements QueryResponseWriter, SolrCoreAware {
|
||||
// init param names, these are _only_ loaded at init time (no per-request control of these)
|
||||
// - multiple different named writers could be created with different init params
|
||||
|
@ -128,7 +128,6 @@ public class VelocityResponseWriter implements QueryResponseWriter, SolrCoreAwar
|
|||
public void inform(SolrCore core) {
|
||||
// need to leverage SolrResourceLoader, so load init.properties.file here instead of init()
|
||||
if (initPropertiesFileName != null) {
|
||||
InputStream is = null;
|
||||
try {
|
||||
velocityInitProps.load(new InputStreamReader(core.getResourceLoader().openResource(initPropertiesFileName), StandardCharsets.UTF_8));
|
||||
} catch (IOException e) {
|
||||
|
@ -195,6 +194,7 @@ public class VelocityResponseWriter implements QueryResponseWriter, SolrCoreAwar
|
|||
}
|
||||
}
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
private VelocityContext createContext(SolrQueryRequest request, SolrQueryResponse response) {
|
||||
VelocityContext context = new VelocityContext();
|
||||
|
||||
|
|
|
@ -22,13 +22,14 @@
|
|||
|
||||
package org.apache.solr.handler.tagger;
|
||||
|
||||
import javax.xml.stream.XMLResolver;
|
||||
import javax.xml.stream.XMLStreamException;
|
||||
import javax.xml.stream.events.XMLEvent;
|
||||
import java.io.InputStream;
|
||||
import java.io.StringReader;
|
||||
|
||||
import com.ctc.wstx.stax.WstxInputFactory;
|
||||
import javax.xml.stream.XMLInputFactory;
|
||||
import javax.xml.stream.XMLResolver;
|
||||
import javax.xml.stream.XMLStreamException;
|
||||
import javax.xml.stream.events.XMLEvent;
|
||||
|
||||
import org.apache.commons.io.input.ClosedInputStream;
|
||||
import org.codehaus.stax2.LocationInfo;
|
||||
import org.codehaus.stax2.XMLInputFactory2;
|
||||
|
@ -46,11 +47,11 @@ import org.codehaus.stax2.XMLStreamReader2;
|
|||
public class XmlOffsetCorrector extends OffsetCorrector {
|
||||
|
||||
//TODO use StAX without hard requirement on woodstox. xmlStreamReader.getLocation().getCharacterOffset()
|
||||
|
||||
private static final XMLInputFactory2 XML_INPUT_FACTORY;
|
||||
//nocommit What's this about?
|
||||
private static final XMLInputFactory XML_INPUT_FACTORY;
|
||||
static {
|
||||
// note: similar code in Solr's EmptyEntityResolver
|
||||
XML_INPUT_FACTORY = new WstxInputFactory();
|
||||
XML_INPUT_FACTORY = XMLInputFactory2.newFactory();
|
||||
XML_INPUT_FACTORY.setXMLResolver(new XMLResolver() {
|
||||
@Override
|
||||
public InputStream resolveEntity(String publicId, String systemId, String baseURI, String namespace) {
|
||||
|
@ -59,7 +60,6 @@ public class XmlOffsetCorrector extends OffsetCorrector {
|
|||
});
|
||||
// TODO disable DTD?
|
||||
// XML_INPUT_FACTORY.setProperty(XMLInputFactory.IS_VALIDATING, Boolean.FALSE)
|
||||
XML_INPUT_FACTORY.configureForSpeed();
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -237,6 +237,7 @@ public interface Variable {
|
|||
DISKTYPE;
|
||||
|
||||
public final String tagName;
|
||||
@SuppressWarnings("rawtypes")
|
||||
public final Class type;
|
||||
public Meta meta;
|
||||
|
||||
|
@ -252,6 +253,7 @@ public interface Variable {
|
|||
final Variable impl;
|
||||
|
||||
|
||||
@SuppressWarnings({"unchecked", "rawtypes"})
|
||||
Type() {
|
||||
try {
|
||||
meta = Type.class.getField(name()).getAnnotation(Meta.class);
|
||||
|
@ -375,6 +377,7 @@ public interface Variable {
|
|||
@interface Meta {
|
||||
String name();
|
||||
|
||||
@SuppressWarnings("rawtypes")
|
||||
Class type();
|
||||
|
||||
String[] associatedPerNodeValue() default NULL;
|
||||
|
@ -399,6 +402,7 @@ public interface Variable {
|
|||
|
||||
String metricsKey() default NULL;
|
||||
|
||||
@SuppressWarnings("rawtypes")
|
||||
Class implementation() default void.class;
|
||||
|
||||
ComputedType[] computedValues() default ComputedType.NULL;
|
||||
|
|
|
@ -60,6 +60,7 @@ public class UpdateRequest extends AbstractUpdateRequest {
|
|||
* @deprecated Solr now always includes in the response the {@link #REPFACT}, this parameter
|
||||
* doesn't need to be explicitly set
|
||||
*/
|
||||
@Deprecated
|
||||
public static final String MIN_REPFACT = "min_rf";
|
||||
public static final String VER = "ver";
|
||||
public static final String OVERWRITE = "ow";
|
||||
|
|
|
@ -243,6 +243,7 @@ public class ZkStateReader implements SolrCloseable {
|
|||
* @return current configuration from <code>autoscaling.json</code>. NOTE:
|
||||
* this data is retrieved from ZK on each call.
|
||||
*/
|
||||
@SuppressWarnings("unchecked")
|
||||
public AutoScalingConfig getAutoScalingConfig(Watcher watcher) throws KeeperException, InterruptedException {
|
||||
Stat stat = new Stat();
|
||||
|
||||
|
|
|
@ -88,6 +88,7 @@ import static java.util.Collections.unmodifiableList;
|
|||
import static java.util.Collections.unmodifiableSet;
|
||||
import static java.util.concurrent.TimeUnit.NANOSECONDS;
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
public class Utils {
|
||||
public static final Function NEW_HASHMAP_FUN = o -> new HashMap<>();
|
||||
public static final Function NEW_LINKED_HASHMAP_FUN = o -> new LinkedHashMap<>();
|
||||
|
|
|
@ -40,6 +40,7 @@ import static java.nio.charset.StandardCharsets.UTF_8;
|
|||
import static java.util.Collections.unmodifiableList;
|
||||
import static java.util.Collections.unmodifiableSet;
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
public class ValidatingJsonMap implements Map<String, Object>, NavigableObject {
|
||||
|
||||
private static final String INCLUDE = "#include";
|
||||
|
|
|
@ -58,7 +58,7 @@ public class UsingSolrJRefGuideExamplesTest extends SolrCloudTestCase {
|
|||
private static final int NUM_INDEXED_DOCUMENTS = 3;
|
||||
private static final int NUM_LIVE_NODES = 1;
|
||||
|
||||
private Queue<String> expectedLines = new ArrayDeque();
|
||||
private Queue<String> expectedLines = new ArrayDeque<>();
|
||||
|
||||
@BeforeClass
|
||||
public static void setUpCluster() throws Exception {
|
||||
|
|
Загрузка…
Ссылка в новой задаче