2020from nltk .corpus import wordnet # type: ignore
2121
2222from bibcat import config
23+ from bibcat .utils .logger_config import setup_logger
2324
2425nlp = spacy .load (config .grammar .spacy_language_model )
2526
27+ logger = setup_logger (__name__ )
28+ logger .setLevel (config .logging .level )
29+
2630
2731# Determine if given text is important (e.g., is a keyword)
2832def check_importance (
@@ -253,7 +257,7 @@ def is_pos_conjoined(word, pos):
253257
254258
255259# Return boolean for if given word (NLP type word) is of given part of speech
256- def is_pos_word (word , pos , keyword_objs = None , do_verbose = False ): # noqa: C901
260+ def is_pos_word (word , pos , keyword_objs = None ): # noqa: C901
257261 """
258262 Method: is_pos_word
259263 WARNING! This method is *not* meant to be used directly by users.
@@ -271,11 +275,10 @@ def is_pos_word(word, pos, keyword_objs=None, do_verbose=False): # noqa: C901
271275 word_ancestors = list (word .ancestors ) # All previous nodes leading to word
272276
273277 # Print some notes
274- if do_verbose :
275- print ("Running is_pos_word for: {0}" .format (word ))
276- print ("dep_: {0}\n pos_: {1}\n tag_: {2}" .format (word_dep , word_pos , word_tag ))
277- print ("Node head: {0}\n Sentence: {1}" .format (word .head , word .sent ))
278- print ("Node lefts: {0}\n Node rights: {1}" .format (list (word .lefts ), list (word .rights )))
278+ logger .info ("Running is_pos_word for: {0}" .format (word ))
279+ logger .info ("dep_: {0}\n pos_: {1}\n tag_: {2}" .format (word_dep , word_pos , word_tag ))
280+ logger .info ("Node head: {0}\n Sentence: {1}" .format (word .head , word .sent ))
281+ logger .info ("Node lefts: {0}\n Node rights: {1}" .format (list (word .lefts ), list (word .rights )))
279282
280283 # Check if given word is of given part-of-speech
281284 # Identify roots
@@ -564,14 +567,13 @@ def is_pos_word(word, pos, keyword_objs=None, do_verbose=False): # noqa: C901
564567 raise ValueError ("Err: {0} is not a recognized part of speech." .format (pos ))
565568
566569 # Print some notes
567- if do_verbose :
568- print ("Is pos={0}? {1}\n -" .format (pos , check_all ))
570+ logger .info ("Is pos={0}? {1}\n -" .format (pos , check_all ))
569571 # Return the final verdict
570572 return check_all
571573
572574
573575# Search text for given keywords and acronyms and return metric
574- def search_text (text , keyword_objs , do_verbose = False ):
576+ def search_text (text , keyword_objs ):
575577 """
576578 Method: search_text
577579 WARNING! This method is *not* meant to be used directly by users.
@@ -585,17 +587,16 @@ def search_text(text, keyword_objs, do_verbose=False):
585587 charspans_keywords += tmp_res [ii ]["charspans" ]
586588
587589 # Print some notes
588- if do_verbose :
589- # Extract global variables
590- keywords = [item2 for item1 in keyword_objs for item2 in item1 ._keywords ]
591- acronyms = [
592- item2 for item1 in keyword_objs for item2 in item1 ._acronyms_casesensitive + item1 ._acronyms_caseinsensitive
593- ]
594-
595- print ("Completed search_text()." )
596- print ("Keywords={0}\n Acronyms={1}" .format (keywords , acronyms ))
597- print ("Boolean: {0}" .format (check_keywords ))
598- print ("Char. Spans: {0}" .format (charspans_keywords ))
590+ # Extract global variables
591+ keywords = [item2 for item1 in keyword_objs for item2 in item1 ._keywords ]
592+ acronyms = [
593+ item2 for item1 in keyword_objs for item2 in item1 ._acronyms_casesensitive + item1 ._acronyms_caseinsensitive
594+ ]
595+
596+ logger .info ("Completed search_text()." )
597+ logger .info ("Keywords={0}\n Acronyms={1}" .format (keywords , acronyms ))
598+ logger .info ("Boolean: {0}" .format (check_keywords ))
599+ logger .info ("Char. Spans: {0}" .format (charspans_keywords ))
599600
600601 # Return boolean result
601602 return {"bool" : check_keywords , "charspans" : charspans_keywords }
0 commit comments