Package org.apache.lucene.analysis

Examples of org.apache.lucene.analysis.Token.type()


    nextToken = tf.next(reusableToken);
    assertTrue("nextToken is null and it shouldn't be", nextToken != null);
    assertTrue(nextToken.term() + " is not equal to " + "j",
            nextToken.term().equals("j") == true);
    assertTrue(nextToken.getPositionIncrement() + " does not equal: " + 1, nextToken.getPositionIncrement() == 1);
    assertTrue(nextToken.type() + " is not equal to " + WikipediaTokenizer.CATEGORY, nextToken.type().equals(WikipediaTokenizer.CATEGORY) == true);
    assertTrue(nextToken.startOffset() + " does not equal: " + 132, nextToken.startOffset() == 132);
    assertTrue(nextToken.endOffset() + " does not equal: " + 133, nextToken.endOffset() == 133);

    nextToken = tf.next(reusableToken);
    assertTrue("nextToken is not null and it should be", nextToken == null);
View Full Code Here


    nextToken = tf.next(reusableToken);
    assertTrue("nextToken is null and it shouldn't be", nextToken != null);
    assertTrue(nextToken.term() + " is not equal to " + "j",
            nextToken.term().equals("j") == true);
    assertTrue(nextToken.getPositionIncrement() + " does not equal: " + 1, nextToken.getPositionIncrement() == 1);
    assertTrue(nextToken.type() + " is not equal to " + WikipediaTokenizer.CATEGORY, nextToken.type().equals(WikipediaTokenizer.CATEGORY) == true);
    assertTrue(nextToken.startOffset() + " does not equal: " + 132, nextToken.startOffset() == 132);
    assertTrue(nextToken.endOffset() + " does not equal: " + 133, nextToken.endOffset() == 133);

    nextToken = tf.next(reusableToken);
    assertTrue("nextToken is not null and it should be", nextToken == null);
View Full Code Here

    Token nextToken = tf.next(reusableToken);
    assertTrue("nextToken is null and it shouldn't be", nextToken != null);
    assertTrue(nextToken.term() + " is not equal to " + "a b c d",
            nextToken.term().equals("a b c d") == true);
    assertTrue(nextToken.getPositionIncrement() + " does not equal: " + 1, nextToken.getPositionIncrement() == 1);
    assertTrue(nextToken.type() + " is not equal to " + WikipediaTokenizer.CATEGORY, nextToken.type().equals(WikipediaTokenizer.CATEGORY) == true);
    assertTrue(nextToken.getFlags() + " does not equal: " + WikipediaTokenizer.UNTOKENIZED_TOKEN_FLAG, nextToken.getFlags() == WikipediaTokenizer.UNTOKENIZED_TOKEN_FLAG);
    assertTrue(nextToken.startOffset() + " does not equal: " + 11, nextToken.startOffset() == 11);
    assertTrue(nextToken.endOffset() + " does not equal: " + 18, nextToken.endOffset() == 18);
    nextToken = tf.next(reusableToken);
    assertTrue("nextToken is null and it shouldn't be", nextToken != null);
View Full Code Here

    Token nextToken = tf.next(reusableToken);
    assertTrue("nextToken is null and it shouldn't be", nextToken != null);
    assertTrue(nextToken.term() + " is not equal to " + "a b c d",
            nextToken.term().equals("a b c d") == true);
    assertTrue(nextToken.getPositionIncrement() + " does not equal: " + 1, nextToken.getPositionIncrement() == 1);
    assertTrue(nextToken.type() + " is not equal to " + WikipediaTokenizer.CATEGORY, nextToken.type().equals(WikipediaTokenizer.CATEGORY) == true);
    assertTrue(nextToken.getFlags() + " does not equal: " + WikipediaTokenizer.UNTOKENIZED_TOKEN_FLAG, nextToken.getFlags() == WikipediaTokenizer.UNTOKENIZED_TOKEN_FLAG);
    assertTrue(nextToken.startOffset() + " does not equal: " + 11, nextToken.startOffset() == 11);
    assertTrue(nextToken.endOffset() + " does not equal: " + 18, nextToken.endOffset() == 18);
    nextToken = tf.next(reusableToken);
    assertTrue("nextToken is null and it shouldn't be", nextToken != null);
View Full Code Here

    nextToken = tf.next(reusableToken);
    assertTrue("nextToken is null and it shouldn't be", nextToken != null);
    assertTrue(nextToken.term() + " is not equal to " + "a",
            nextToken.term().equals("a") == true);
    assertTrue(nextToken.getPositionIncrement() + " does not equal: " + 0, nextToken.getPositionIncrement() == 0);
    assertTrue(nextToken.type() + " is not equal to " + WikipediaTokenizer.CATEGORY, nextToken.type().equals(WikipediaTokenizer.CATEGORY) == true);
    assertTrue(nextToken.getFlags() + " equals: " + WikipediaTokenizer.UNTOKENIZED_TOKEN_FLAG + " and it shouldn't", nextToken.getFlags() != WikipediaTokenizer.UNTOKENIZED_TOKEN_FLAG);
    assertTrue(nextToken.startOffset() + " does not equal: " + 11, nextToken.startOffset() == 11);
    assertTrue(nextToken.endOffset() + " does not equal: " + 12, nextToken.endOffset() == 12);

    nextToken = tf.next(reusableToken);
View Full Code Here

    nextToken = tf.next(reusableToken);
    assertTrue("nextToken is null and it shouldn't be", nextToken != null);
    assertTrue(nextToken.term() + " is not equal to " + "a",
            nextToken.term().equals("a") == true);
    assertTrue(nextToken.getPositionIncrement() + " does not equal: " + 0, nextToken.getPositionIncrement() == 0);
    assertTrue(nextToken.type() + " is not equal to " + WikipediaTokenizer.CATEGORY, nextToken.type().equals(WikipediaTokenizer.CATEGORY) == true);
    assertTrue(nextToken.getFlags() + " equals: " + WikipediaTokenizer.UNTOKENIZED_TOKEN_FLAG + " and it shouldn't", nextToken.getFlags() != WikipediaTokenizer.UNTOKENIZED_TOKEN_FLAG);
    assertTrue(nextToken.startOffset() + " does not equal: " + 11, nextToken.startOffset() == 11);
    assertTrue(nextToken.endOffset() + " does not equal: " + 12, nextToken.endOffset() == 12);

    nextToken = tf.next(reusableToken);
View Full Code Here

    nextToken = tf.next(reusableToken);
    assertTrue("nextToken is null and it shouldn't be", nextToken != null);
    assertTrue(nextToken.term() + " is not equal to " + "b",
            nextToken.term().equals("b") == true);
    assertTrue(nextToken.getPositionIncrement() + " does not equal: " + 1, nextToken.getPositionIncrement() == 1);
    assertTrue(nextToken.type() + " is not equal to " + WikipediaTokenizer.CATEGORY, nextToken.type().equals(WikipediaTokenizer.CATEGORY) == true);
    assertTrue(nextToken.startOffset() + " does not equal: " + 13, nextToken.startOffset() == 13);
    assertTrue(nextToken.endOffset() + " does not equal: " + 14, nextToken.endOffset() == 14);

    nextToken = tf.next(reusableToken);
    assertTrue("nextToken is null and it shouldn't be", nextToken != null);
View Full Code Here

    nextToken = tf.next(reusableToken);
    assertTrue("nextToken is null and it shouldn't be", nextToken != null);
    assertTrue(nextToken.term() + " is not equal to " + "b",
            nextToken.term().equals("b") == true);
    assertTrue(nextToken.getPositionIncrement() + " does not equal: " + 1, nextToken.getPositionIncrement() == 1);
    assertTrue(nextToken.type() + " is not equal to " + WikipediaTokenizer.CATEGORY, nextToken.type().equals(WikipediaTokenizer.CATEGORY) == true);
    assertTrue(nextToken.startOffset() + " does not equal: " + 13, nextToken.startOffset() == 13);
    assertTrue(nextToken.endOffset() + " does not equal: " + 14, nextToken.endOffset() == 14);

    nextToken = tf.next(reusableToken);
    assertTrue("nextToken is null and it shouldn't be", nextToken != null);
View Full Code Here

        Token t1 = (Token) tokens1.get(i);
        Token t2 = (Token) tokens2.get(i);
        if (!(t1.term().equals(t2.term()))) throw new IllegalStateException("termText");
        if (t1.startOffset() != t2.startOffset()) throw new IllegalStateException("startOffset");
        if (t1.endOffset() != t2.endOffset()) throw new IllegalStateException("endOffset");
        if (!(t1.type().equals(t2.type()))) throw new IllegalStateException("type");
      }
      if (tokens1.size() != tokens2.size())   throw new IllegalStateException("size1=" + tokens1.size() + ", size2=" + tokens2.size());
    }

    catch (IllegalStateException e) {
View Full Code Here

    nextToken = tf.next(reusableToken);
    assertTrue("nextToken is null and it shouldn't be", nextToken != null);
    assertTrue(nextToken.term() + " is not equal to " + "c",
            nextToken.term().equals("c") == true);
    assertTrue(nextToken.getPositionIncrement() + " does not equal: " + 1, nextToken.getPositionIncrement() == 1);
    assertTrue(nextToken.type() + " is not equal to " + WikipediaTokenizer.CATEGORY, nextToken.type().equals(WikipediaTokenizer.CATEGORY) == true);
    assertTrue(nextToken.startOffset() + " does not equal: " + 15, nextToken.startOffset() == 15);
    assertTrue(nextToken.endOffset() + " does not equal: " + 16, nextToken.endOffset() == 16);

    nextToken = tf.next(reusableToken);
    assertTrue("nextToken is null and it shouldn't be", nextToken != null);
View Full Code Here

TOP
Copyright © 2018 www.massapi.com. All rights reserved.
All source code are property of their respective owners. Java is a trademark of Sun Microsystems, Inc and owned by ORACLE Inc. Contact coftware#gmail.com.