Skip to content

Commit 4ebe774

Browse files
committed
isort
1 parent 352d774 commit 4ebe774

File tree

6 files changed

+15
-14
lines changed

6 files changed

+15
-14
lines changed

spacy/lang/ht/__init__.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,11 +5,11 @@
55
from ...language import BaseDefaults, Language
66
from .lemmatizer import HaitianCreoleLemmatizer
77
from .lex_attrs import LEX_ATTRS
8-
from .punctuation import TOKENIZER_PREFIXES, TOKENIZER_INFIXES, TOKENIZER_SUFFIXES
8+
from .punctuation import TOKENIZER_INFIXES, TOKENIZER_PREFIXES, TOKENIZER_SUFFIXES
99
from .stop_words import STOP_WORDS
1010
from .syntax_iterators import SYNTAX_ITERATORS
11-
from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS
1211
from .tag_map import TAG_MAP
12+
from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS
1313

1414

1515
class HaitianCreoleDefaults(BaseDefaults):

spacy/lang/ht/lemmatizer.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,8 @@
11
from typing import List, Tuple
22

3+
from ...lookups import Lookups
34
from ...pipeline import Lemmatizer
45
from ...tokens import Token
5-
from ...lookups import Lookups
66

77

88
class HaitianCreoleLemmatizer(Lemmatizer):

spacy/lang/ht/punctuation.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4,10 +4,10 @@
44
ALPHA_UPPER,
55
CONCAT_QUOTES,
66
HYPHENS,
7-
LIST_PUNCT,
8-
LIST_QUOTES,
97
LIST_ELLIPSES,
108
LIST_ICONS,
9+
LIST_PUNCT,
10+
LIST_QUOTES,
1111
merge_chars,
1212
)
1313

spacy/lang/ht/tag_map.py

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,20 +1,20 @@
11
from spacy.symbols import (
2-
NOUN,
3-
VERB,
4-
AUX,
52
ADJ,
6-
ADV,
7-
PRON,
8-
DET,
93
ADP,
10-
SCONJ,
4+
ADV,
5+
AUX,
116
CCONJ,
12-
PART,
7+
DET,
138
INTJ,
9+
NOUN,
1410
NUM,
11+
PART,
12+
PRON,
1513
PROPN,
1614
PUNCT,
15+
SCONJ,
1716
SYM,
17+
VERB,
1818
X,
1919
)
2020

spacy/lang/ht/tokenizer_exceptions.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
from spacy.symbols import ORTH, NORM
1+
from spacy.symbols import NORM, ORTH
22

33

44
def make_variants(base, first_norm, second_orth, second_norm):

spacy/tests/lang/ht/test_noun_chunks.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,5 @@
11
import pytest
2+
23
from spacy.tokens import Doc
34

45

0 commit comments

Comments
 (0)