Skip to content

Commit 9864066

Browse files
author
Dan Burger
committed
ASB-31862 replace do_verbose with logger.info
1 parent 894aaac commit 9864066

11 files changed

Lines changed: 235 additions & 328 deletions

File tree

bibcat/core/core_utils.py

Lines changed: 21 additions & 20 deletions
Original file line numberDiff line numberDiff line change
@@ -20,9 +20,13 @@
2020
from nltk.corpus import wordnet # type: ignore
2121

2222
from bibcat import config
23+
from bibcat.utils.logger_config import setup_logger
2324

2425
nlp = spacy.load(config.grammar.spacy_language_model)
2526

27+
logger = setup_logger(__name__)
28+
logger.setLevel(config.logging.level)
29+
2630

2731
# Determine if given text is important (e.g., is a keyword)
2832
def check_importance(
@@ -253,7 +257,7 @@ def is_pos_conjoined(word, pos):
253257

254258

255259
# Return boolean for if given word (NLP type word) is of given part of speech
256-
def is_pos_word(word, pos, keyword_objs=None, do_verbose=False): # noqa: C901
260+
def is_pos_word(word, pos, keyword_objs=None): # noqa: C901
257261
"""
258262
Method: is_pos_word
259263
WARNING! This method is *not* meant to be used directly by users.
@@ -271,11 +275,10 @@ def is_pos_word(word, pos, keyword_objs=None, do_verbose=False): # noqa: C901
271275
word_ancestors = list(word.ancestors) # All previous nodes leading to word
272276

273277
# Print some notes
274-
if do_verbose:
275-
print("Running is_pos_word for: {0}".format(word))
276-
print("dep_: {0}\npos_: {1}\ntag_: {2}".format(word_dep, word_pos, word_tag))
277-
print("Node head: {0}\nSentence: {1}".format(word.head, word.sent))
278-
print("Node lefts: {0}\nNode rights: {1}".format(list(word.lefts), list(word.rights)))
278+
logger.info("Running is_pos_word for: {0}".format(word))
279+
logger.info("dep_: {0}\npos_: {1}\ntag_: {2}".format(word_dep, word_pos, word_tag))
280+
logger.info("Node head: {0}\nSentence: {1}".format(word.head, word.sent))
281+
logger.info("Node lefts: {0}\nNode rights: {1}".format(list(word.lefts), list(word.rights)))
279282

280283
# Check if given word is of given part-of-speech
281284
# Identify roots
@@ -564,14 +567,13 @@ def is_pos_word(word, pos, keyword_objs=None, do_verbose=False): # noqa: C901
564567
raise ValueError("Err: {0} is not a recognized part of speech.".format(pos))
565568

566569
# Print some notes
567-
if do_verbose:
568-
print("Is pos={0}? {1}\n-".format(pos, check_all))
570+
logger.info("Is pos={0}? {1}\n-".format(pos, check_all))
569571
# Return the final verdict
570572
return check_all
571573

572574

573575
# Search text for given keywords and acronyms and return metric
574-
def search_text(text, keyword_objs, do_verbose=False):
576+
def search_text(text, keyword_objs):
575577
"""
576578
Method: search_text
577579
WARNING! This method is *not* meant to be used directly by users.
@@ -585,17 +587,16 @@ def search_text(text, keyword_objs, do_verbose=False):
585587
charspans_keywords += tmp_res[ii]["charspans"]
586588

587589
# Print some notes
588-
if do_verbose:
589-
# Extract global variables
590-
keywords = [item2 for item1 in keyword_objs for item2 in item1._keywords]
591-
acronyms = [
592-
item2 for item1 in keyword_objs for item2 in item1._acronyms_casesensitive + item1._acronyms_caseinsensitive
593-
]
594-
595-
print("Completed search_text().")
596-
print("Keywords={0}\nAcronyms={1}".format(keywords, acronyms))
597-
print("Boolean: {0}".format(check_keywords))
598-
print("Char. Spans: {0}".format(charspans_keywords))
590+
# Extract global variables
591+
keywords = [item2 for item1 in keyword_objs for item2 in item1._keywords]
592+
acronyms = [
593+
item2 for item1 in keyword_objs for item2 in item1._acronyms_casesensitive + item1._acronyms_caseinsensitive
594+
]
595+
596+
logger.info("Completed search_text().")
597+
logger.info("Keywords={0}\nAcronyms={1}".format(keywords, acronyms))
598+
logger.info("Boolean: {0}".format(check_keywords))
599+
logger.info("Char. Spans: {0}".format(charspans_keywords))
599600

600601
# Return boolean result
601602
return {"bool": check_keywords, "charspans": charspans_keywords}

0 commit comments

Comments
 (0)