mirror of
https://gitflic.ru/project/openide/openide.git
synced 2026-04-30 02:09:59 +07:00
[spelling] @NotNull TokenConsumer consumer in Tokenizer
GitOrigin-RevId: ba7b161e92c8047602fe9fd993d48d2140c60179
This commit is contained in:
committed by
intellij-monorepo-bot
parent
5201d7904c
commit
14daf0f8c3
@@ -33,7 +33,7 @@ public class DocCommentTokenizer extends Tokenizer<PsiDocComment> {
|
||||
private static final Set<String> excludedTags = Set.of("author", "see", "by", "link");
|
||||
|
||||
@Override
|
||||
public void tokenize(@NotNull PsiDocComment comment, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PsiDocComment comment, @NotNull TokenConsumer consumer) {
|
||||
final CommentSplitter splitter = CommentSplitter.getInstance();
|
||||
|
||||
for (PsiElement el = comment.getFirstChild(); el != null; el = el.getNextSibling()) {
|
||||
|
||||
@@ -22,7 +22,7 @@ import java.util.Arrays;
|
||||
*/
|
||||
public class LiteralExpressionTokenizer extends EscapeSequenceTokenizer<PsiLiteralExpression> {
|
||||
@Override
|
||||
public void tokenize(@NotNull PsiLiteralExpression expression, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PsiLiteralExpression expression, @NotNull TokenConsumer consumer) {
|
||||
String text;
|
||||
if (!ExpressionUtils.hasStringType(expression)) {
|
||||
text = null;
|
||||
|
||||
@@ -25,7 +25,7 @@ import org.jetbrains.annotations.NotNull;
|
||||
public class MethodNameTokenizerJava extends NamedElementTokenizer<PsiMethod> {
|
||||
|
||||
@Override
|
||||
public void tokenize(@NotNull PsiMethod element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PsiMethod element, @NotNull TokenConsumer consumer) {
|
||||
if (element.isConstructor() || element.findDeepestSuperMethods().length > 0) return;
|
||||
|
||||
super.tokenize(element, consumer);
|
||||
|
||||
@@ -31,7 +31,7 @@ public class NamedElementTokenizer<T extends PsiNamedElement> extends Tokenizer<
|
||||
private final Tokenizer<PsiIdentifier> myIdentifierTokenizer = new PsiIdentifierTokenizer();
|
||||
|
||||
@Override
|
||||
public void tokenize(@NotNull T element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull T element, @NotNull TokenConsumer consumer) {
|
||||
PsiIdentifier psiIdentifier = PsiTreeUtil.getChildOfType(element, PsiIdentifier.class);
|
||||
if (psiIdentifier == null) return;
|
||||
|
||||
|
||||
@@ -10,7 +10,7 @@ import org.jetbrains.annotations.NotNull;
|
||||
|
||||
public class PsiIdentifierTokenizer extends Tokenizer<PsiIdentifier> {
|
||||
@Override
|
||||
public void tokenize(@NotNull PsiIdentifier element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PsiIdentifier element, @NotNull TokenConsumer consumer) {
|
||||
consumer.consumeToken(element, true, IdentifierSplitter.getInstance());
|
||||
}
|
||||
}
|
||||
|
||||
@@ -38,7 +38,7 @@ import java.util.List;
|
||||
public class PsiTypeTokenizer extends Tokenizer<PsiTypeElement> {
|
||||
|
||||
@Override
|
||||
public void tokenize(@NotNull PsiTypeElement element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PsiTypeElement element, @NotNull TokenConsumer consumer) {
|
||||
final PsiType type = element.getType();
|
||||
if (type instanceof PsiDisjunctionType) {
|
||||
tokenizeComplexType(element, consumer);
|
||||
|
||||
@@ -18,7 +18,7 @@ import java.util.List;
|
||||
public class JsonSpellcheckerStrategy extends SpellcheckingStrategy {
|
||||
private final Tokenizer<JsonStringLiteral> ourStringLiteralTokenizer = new Tokenizer<>() {
|
||||
@Override
|
||||
public void tokenize(@NotNull JsonStringLiteral element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull JsonStringLiteral element, @NotNull TokenConsumer consumer) {
|
||||
final PlainTextSplitter textSplitter = PlainTextSplitter.getInstance();
|
||||
if (element.textContains('\\')) {
|
||||
final List<Pair<TextRange, String>> fragments = element.getTextFragments();
|
||||
|
||||
@@ -19,7 +19,7 @@ public final class JsonPathSpellcheckerStrategy extends SpellcheckingStrategy {
|
||||
|
||||
private final Tokenizer<JsonPathStringLiteral> ourStringLiteralTokenizer = new Tokenizer<>() {
|
||||
@Override
|
||||
public void tokenize(@NotNull JsonPathStringLiteral element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull JsonPathStringLiteral element, @NotNull TokenConsumer consumer) {
|
||||
PlainTextSplitter textSplitter = PlainTextSplitter.getInstance();
|
||||
if (element.textContains('\\')) {
|
||||
List<Pair<TextRange, String>> fragments = element.getTextFragments();
|
||||
@@ -40,7 +40,7 @@ public final class JsonPathSpellcheckerStrategy extends SpellcheckingStrategy {
|
||||
|
||||
private final Tokenizer<JsonPathId> idLiteralTokenizer = new Tokenizer<>() {
|
||||
@Override
|
||||
public void tokenize(@NotNull JsonPathId element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull JsonPathId element, @NotNull TokenConsumer consumer) {
|
||||
PlainTextSplitter textSplitter = PlainTextSplitter.getInstance();
|
||||
consumer.consumeToken(element, textSplitter);
|
||||
}
|
||||
|
||||
@@ -33,7 +33,7 @@ public class GrDocCommentTokenizer extends Tokenizer<PsiDocComment> {
|
||||
private static final Set<String> excludedTags = ContainerUtil.immutableSet("author", "see", "by", "link");
|
||||
|
||||
@Override
|
||||
public void tokenize(@NotNull PsiDocComment comment, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PsiDocComment comment, @NotNull TokenConsumer consumer) {
|
||||
final CommentSplitter splitter = CommentSplitter.getInstance();
|
||||
|
||||
for (PsiElement el : comment.getChildren()) {
|
||||
|
||||
@@ -31,7 +31,7 @@ public class GroovySpellcheckingStrategy extends SpellcheckingStrategy {
|
||||
private final GrDocCommentTokenizer myDocCommentTokenizer = new GrDocCommentTokenizer();
|
||||
private final Tokenizer<PsiElement> myStringTokenizer = new EscapeSequenceTokenizer<>() {
|
||||
@Override
|
||||
public void tokenize(@NotNull PsiElement literal, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PsiElement literal, @NotNull TokenConsumer consumer) {
|
||||
String text = GrStringUtil.removeQuotes(literal.getText());
|
||||
if (!text.contains("\\")) {
|
||||
consumer.consumeToken(literal, PlainTextSplitter.getInstance());
|
||||
|
||||
@@ -14,7 +14,7 @@ public class ShIdentifierOwnerTokenizer extends Tokenizer<PsiNameIdentifierOwner
|
||||
public static final ShIdentifierOwnerTokenizer INSTANCE = new ShIdentifierOwnerTokenizer();
|
||||
|
||||
@Override
|
||||
public void tokenize(@NotNull PsiNameIdentifierOwner element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PsiNameIdentifierOwner element, @NotNull TokenConsumer consumer) {
|
||||
PsiElement identifier = element.getNameIdentifier();
|
||||
if (identifier == null) {
|
||||
return;
|
||||
|
||||
@@ -16,7 +16,7 @@ import org.jetbrains.annotations.NotNull;
|
||||
public class RestSpellcheckerStrategy extends SpellcheckingStrategy {
|
||||
private static final Tokenizer<PsiElement> REST_ELEMENT_TOKENIZER = new Tokenizer<>() {
|
||||
@Override
|
||||
public void tokenize(@NotNull PsiElement element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PsiElement element, @NotNull TokenConsumer consumer) {
|
||||
consumer.consumeToken(element, PlainTextSplitter.getInstance());
|
||||
}
|
||||
};
|
||||
|
||||
@@ -26,7 +26,7 @@ import java.util.List;
|
||||
public class PythonSpellcheckerStrategy extends SpellcheckingStrategy {
|
||||
private static class StringLiteralTokenizer extends Tokenizer<PyStringLiteralExpression> {
|
||||
@Override
|
||||
public void tokenize(@NotNull PyStringLiteralExpression element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PyStringLiteralExpression element, @NotNull TokenConsumer consumer) {
|
||||
final Splitter splitter = PlainTextSplitter.getInstance();
|
||||
for (PyStringElement stringElement : element.getStringElements()) {
|
||||
final List<TextRange> literalPartRanges;
|
||||
@@ -57,7 +57,7 @@ public class PythonSpellcheckerStrategy extends SpellcheckingStrategy {
|
||||
|
||||
private static class FormatStringTokenizer extends Tokenizer<PyStringLiteralExpression> {
|
||||
@Override
|
||||
public void tokenize(@NotNull PyStringLiteralExpression element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PyStringLiteralExpression element, @NotNull TokenConsumer consumer) {
|
||||
String stringValue = element.getStringValue();
|
||||
List<PyStringFormatParser.FormatStringChunk> chunks = PyStringFormatParser.parsePercentFormat(stringValue);
|
||||
Splitter splitter = PlainTextSplitter.getInstance();
|
||||
|
||||
@@ -22,7 +22,7 @@ import org.jetbrains.annotations.NotNull;
|
||||
public class CommentTokenizer extends Tokenizer<PsiComment> {
|
||||
|
||||
@Override
|
||||
public void tokenize(@NotNull PsiComment element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PsiComment element, @NotNull TokenConsumer consumer) {
|
||||
// doccomment chameleon expands as PsiComment inside PsiComment, avoid duplication
|
||||
if (element.getParent() instanceof PsiComment) return;
|
||||
consumer.consumeToken(element, CommentSplitter.getInstance());
|
||||
|
||||
@@ -32,7 +32,7 @@ class CustomFileTypeTokenizer extends Tokenizer<PsiElement> {
|
||||
}
|
||||
|
||||
@Override
|
||||
public void tokenize(@NotNull PsiElement element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PsiElement element, @NotNull TokenConsumer consumer) {
|
||||
CustomFileTypeLexer lexer = new CustomFileTypeLexer(mySyntaxTable);
|
||||
String text = element.getText();
|
||||
lexer.start(text);
|
||||
|
||||
@@ -13,7 +13,7 @@ public class PsiIdentifierOwnerTokenizer extends Tokenizer<PsiNameIdentifierOwne
|
||||
public static final PsiIdentifierOwnerTokenizer INSTANCE = new PsiIdentifierOwnerTokenizer();
|
||||
|
||||
@Override
|
||||
public void tokenize(@NotNull PsiNameIdentifierOwner element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PsiNameIdentifierOwner element, @NotNull TokenConsumer consumer) {
|
||||
PsiElement identifier = element.getNameIdentifier();
|
||||
if (identifier == null) {
|
||||
return;
|
||||
|
||||
@@ -35,7 +35,7 @@ public class SpellcheckingStrategy {
|
||||
new ExtensionPointName<>("com.intellij.spellchecker.support");
|
||||
public static final Tokenizer EMPTY_TOKENIZER = new Tokenizer() {
|
||||
@Override
|
||||
public void tokenize(@NotNull PsiElement element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull PsiElement element, @NotNull TokenConsumer consumer) {
|
||||
}
|
||||
|
||||
@Override
|
||||
|
||||
@@ -21,7 +21,7 @@ import org.jetbrains.annotations.NotNull;
|
||||
|
||||
public abstract class Tokenizer<T extends PsiElement> {
|
||||
|
||||
public abstract void tokenize(@NotNull T element, TokenConsumer consumer);
|
||||
public abstract void tokenize(@NotNull T element, @NotNull TokenConsumer consumer);
|
||||
|
||||
@NotNull
|
||||
public TextRange getHighlightingRange(PsiElement element, int offset, TextRange textRange) {
|
||||
|
||||
@@ -25,7 +25,7 @@ public class TokenizerBase<T extends PsiElement> extends Tokenizer<T> {
|
||||
}
|
||||
|
||||
@Override
|
||||
public void tokenize(@NotNull T element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull T element, @NotNull TokenConsumer consumer) {
|
||||
if (element instanceof PsiLanguageInjectionHost && InjectedLanguageUtil.hasInjections((PsiLanguageInjectionHost)element)) {
|
||||
return;
|
||||
}
|
||||
|
||||
@@ -163,7 +163,7 @@ public class XmlSpellcheckingStrategy extends SuppressibleSpellcheckingStrategy
|
||||
}
|
||||
|
||||
@Override
|
||||
public void tokenize(@NotNull XmlAttributeValue element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull XmlAttributeValue element, @NotNull TokenConsumer consumer) {
|
||||
PsiReference[] references = element.getReferences();
|
||||
for (PsiReference reference : references) {
|
||||
if (reference instanceof XmlEnumeratedValueReference) {
|
||||
|
||||
@@ -35,7 +35,7 @@ public class XmlTokenizerBase<T extends PsiElement> extends Tokenizer<T> {
|
||||
}
|
||||
|
||||
@Override
|
||||
public void tokenize(@NotNull T element, TokenConsumer consumer) {
|
||||
public void tokenize(@NotNull T element, @NotNull TokenConsumer consumer) {
|
||||
List<TextRange> excludeRanges = getSpellcheckOuterContentRanges(element);
|
||||
for (var spellcheckRange : getSpellcheckRanges(element)) {
|
||||
consumer.consumeToken(
|
||||
|
||||
Reference in New Issue
Block a user