Page MenuHomec4science

test_indexer_engine.py
No OneTemporary

File Metadata

Created
Sat, Oct 12, 11:23

test_indexer_engine.py

# -*- coding: utf-8 -*-
#
# This file is part of Invenio.
# Copyright (C) 2004, 2005, 2006, 2007, 2008, 2010, 2011, 2013, 2014, 2015 CERN.
#
# Invenio is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License as
# published by the Free Software Foundation; either version 2 of the
# License, or (at your option) any later version.
#
# Invenio is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Invenio; if not, write to the Free Software Foundation, Inc.,
# 59 Temple Place, Suite 330, Boston, MA 02111-1307, USA.
"""Unit tests for the indexing engine."""
from invenio.base.wrappers import lazy_import
from invenio.testsuite import InvenioTestCase, make_test_suite, run_test_suite
bibindex_engine = lazy_import('invenio.legacy.bibindex.engine')
load_tokenizers = lazy_import(
'invenio.legacy.bibindex.engine_utils.load_tokenizers')
list_union = lazy_import('invenio.legacy.bibindex.engine_utils.list_union')
get_values_recursively = lazy_import(
'invenio.legacy.bibindex.engine_utils.get_values_recursively')
class TestListSetOperations(InvenioTestCase):
"""Tests for list set operations."""
def test_list_union(self):
"""bibindex engine utils - list union."""
self.assertEqual([1, 2, 3, 4],
list_union([1, 2, 3],
[1, 3, 4]))
def test_list_unique(self):
"""bibindex engine - list unique."""
self.assertEqual([1, 2, 3],
bibindex_engine.list_unique([1, 2, 3, 3, 1, 2]))
class TestWashIndexTerm(InvenioTestCase):
"""Tests for washing index terms, useful for both searching and indexing."""
def test_wash_index_term_short(self):
"""bibindex engine - wash index term, short word."""
self.assertEqual("ellis",
bibindex_engine.wash_index_term("ellis"))
def test_wash_index_term_long(self):
"""bibindex engine - wash index term, long word."""
self.assertEqual(50 * "e",
bibindex_engine.wash_index_term(1234 * "e"))
def test_wash_index_term_case(self):
"""bibindex engine - wash index term, lower the case."""
self.assertEqual("ellis",
bibindex_engine.wash_index_term("Ellis"))
def test_wash_index_term_unicode(self):
"""bibindex engine - wash index term, unicode."""
self.assertEqual("ελληνικό αλφάβητο",
bibindex_engine.wash_index_term("Ελληνικό αλφάβητο"))
def test_wash_index_term_unicode_for_real(self):
"""bibindex engine - wash index term, unicode string."""
self.assertEqual("ellis",
bibindex_engine.wash_index_term(u"ellis"))
class TestGetWordsFromPhrase(InvenioTestCase):
"""Tests for getting words from phrase."""
def setUp(self):
self._TOKENIZERS = load_tokenizers()
def test_easy_phrase(self):
"""bibindex engine - getting words from `word1 word2' phrase."""
test_phrase = 'word1 word2'
l_words_expected = ['word1', 'word2']
tokenizer = self._TOKENIZERS["BibIndexDefaultTokenizer"]()
l_words_obtained = tokenizer.tokenize_for_words(test_phrase)
l_words_obtained.sort()
self.assertEqual(l_words_obtained, l_words_expected)
def test_stemming_phrase(self):
"""bibindex engine - getting stemmed words from l'anthropologie."""
test_phrase = "l'anthropologie"
l_words_not_expected = [
'anthropolog', 'l', "l'anthropolog", "l'anthropologi"]
l_words_expected = ['anthropologi', 'l', "l'anthropologi"]
tokenizer = self._TOKENIZERS["BibIndexDefaultTokenizer"]('en')
l_words_obtained = tokenizer.tokenize_for_words(test_phrase)
l_words_obtained.sort()
self.assertNotEqual(l_words_obtained, l_words_not_expected)
self.assertEqual(l_words_obtained, l_words_expected)
def test_remove_stopwords_phrase(self):
"""bibindex engine - test for removing stopwords from 'theory of'."""
test_phrase = 'theory of'
tokenizer = self._TOKENIZERS["BibIndexDefaultTokenizer"](
remove_stopwords='stopwords.kb')
words_obtained = tokenizer.tokenize_for_words(test_phrase)
words_expected = ['theory']
self.assertEqual(words_expected, words_obtained)
def test_stemming_and_remove_stopwords_phrase(self):
"""bibindex engine - test for removing stopwords and stemming from 'beams of photons'."""
test_phrase = 'beams of photons'
tokenizer = self._TOKENIZERS["BibIndexDefaultTokenizer"](
'en', remove_stopwords='stopwords.kb')
words_obtained = tokenizer.tokenize_for_words(test_phrase)
words_expected = ['beam', 'photon']
self.assertEqual(words_expected, words_obtained)
def test_dashed_phrase(self):
"""bibindex engine - getting words from `word1-word2' phrase."""
test_phrase = 'word1-word2'
l_words_expected = ['word1', 'word1-word2', 'word2']
tokenizer = self._TOKENIZERS["BibIndexDefaultTokenizer"]()
l_words_obtained = tokenizer.tokenize_for_words(test_phrase)
l_words_obtained.sort()
self.assertEqual(l_words_obtained, l_words_expected)
def test_arXiv_good(self):
"""bibindex engine - getting words from `arXiv:1007.5048' phrase."""
test_phrase = 'arXiv:1007.5048'
l_words_expected = [
'1007', '1007.5048', '5048', 'arxiv', 'arxiv:1007.5048']
tokenizer = self._TOKENIZERS["BibIndexDefaultTokenizer"]()
l_words_obtained = tokenizer.tokenize_for_words(test_phrase)
l_words_obtained.sort()
self.assertEqual(l_words_obtained, l_words_expected)
def test_arXiv_bad(self):
"""bibindex engine - getting words from `arXiv:1xy7.5z48' phrase."""
test_phrase = 'arXiv:1xy7.5z48'
l_words_expected = ['1xy7', '5z48', 'arxiv', 'arxiv:1xy7.5z48']
tokenizer = self._TOKENIZERS["BibIndexDefaultTokenizer"]()
l_words_obtained = tokenizer.tokenize_for_words(test_phrase)
l_words_obtained.sort()
self.assertEqual(l_words_obtained, l_words_expected)
class TestGetPairsFromPhrase(InvenioTestCase):
"""Tests for getting pairs from phrase."""
def setUp(self):
self._TOKENIZERS = load_tokenizers()
def test_remove_stopwords_phrase_first(self):
"""bibindex engine - getting pairs from phrase with stopwords removed first."""
test_phrase = 'Matrices on a point as the theory of everything'
tokenizer = self._TOKENIZERS["BibIndexDefaultTokenizer"](
remove_stopwords='stopwords.kb')
pairs_obtained = tokenizer.tokenize_for_pairs(test_phrase)
pairs_expected = ['matrices theory']
self.assertEqual(pairs_expected, pairs_obtained)
def test_remove_stopwords_phrase_second(self):
"""bibindex engine - getting pairs from phrase with stopwords removed second."""
test_phrase = 'Nonlocal action for long-distance'
tokenizer = self._TOKENIZERS["BibIndexDefaultTokenizer"](
remove_stopwords='stopwords.kb')
pairs_obtained = tokenizer.tokenize_for_pairs(test_phrase)
pairs_expected = ['nonlocal action', 'long distance', 'action long']
self.assertEqual(pairs_expected, pairs_obtained)
class TestGetWordsFromDateTag(InvenioTestCase):
"""Tests for getting words for date-like tag."""
def setUp(self):
self._TOKENIZERS = load_tokenizers()
def test_dateindex_yyyy(self):
"""bibindex engine - index date-like tag, yyyy."""
tokenizer = self._TOKENIZERS["BibIndexYearTokenizer"]()
self.assertEqual(["2010"],
tokenizer.get_words_from_date_tag("2010"))
def test_dateindex_yyyy_mm(self):
"""bibindex engine - index date-like tag, yyyy-mm."""
tokenizer = self._TOKENIZERS["BibIndexYearTokenizer"]()
self.assertEqual(["2010-03", "2010"],
tokenizer.get_words_from_date_tag("2010-03"))
def test_dateindex_yyyy_mm_dd(self):
"""bibindex engine - index date-like tag, yyyy-mm-dd."""
tokenizer = self._TOKENIZERS["BibIndexYearTokenizer"]()
self.assertEqual(["2010-03-08", "2010", "2010-03", ],
tokenizer.get_words_from_date_tag("2010-03-08"))
def test_dateindex_freetext(self):
"""bibindex engine - index date-like tag, yyyy-mm-dd."""
tokenizer = self._TOKENIZERS["BibIndexYearTokenizer"]()
self.assertEqual(["dd", "mon", "yyyy"],
tokenizer.get_words_from_date_tag("dd mon yyyy"))
class TestGetAuthorFamilyNameWords(InvenioTestCase):
"""Tests for getting family name words from author names."""
def setUp(self):
self._TOKENIZERS = load_tokenizers()
def test_authornames_john_doe(self):
"""bibindex engine - get author family name words for John Doe."""
tokenizer = self._TOKENIZERS["BibIndexAuthorTokenizer"]()
self.assertEqual(['doe', ],
tokenizer.get_author_family_name_words_from_phrase('John Doe'))
def test_authornames_doe_john(self):
"""bibindex engine - get author family name words for Doe, John."""
tokenizer = self._TOKENIZERS["BibIndexAuthorTokenizer"]()
self.assertEqual(['doe', ],
tokenizer.get_author_family_name_words_from_phrase('Doe, John'))
def test_authornames_campbell_wilson(self):
"""bibindex engine - get author family name words for Campbell-Wilson, D."""
tokenizer = self._TOKENIZERS["BibIndexAuthorTokenizer"]()
self.assertEqual(['campbell', 'wilson', 'campbell-wilson'],
tokenizer.get_author_family_name_words_from_phrase('Campbell-Wilson, D'))
class TestGetValuesFromRecjson(InvenioTestCase):
"""Tests for get_values_recursively function which finds values for tokenization in recjson record."""
@classmethod
def setUp(self):
self.dict1 = {'all': {'vehicles': {'cars': {'car1': ('flat tyre', 'windscreen'),
'car2': ('engine',)
},
'planes': ['Airplane', 'x,y - plane'],
'ufo': {}
},
'people': ['Frank', 'Theodor', 'Richard'],
'vikings': ['Odin', 'Eric'],
}
}
self.dict2 = {
'all': {'name': ([(['name1', 'name2', {'name3': 'name4'}], )], )}}
def test_dict1_all_fields(self):
"""bibindex termcollectors - get_field_values - complicated field."""
fields = self.dict1
phrases = []
get_values_recursively(fields['all'], phrases)
self.assertEqual(phrases, ['engine', 'flat tyre', 'windscreen', 'Airplane', 'x,y - plane',
'Odin', 'Eric', 'Frank', 'Theodor', 'Richard'])
def test_dict1_subfield(self):
"""bibindex termcollectors - get_field_values - simple field."""
fields = self.dict1
phrases = []
get_values_recursively(fields['all']['people'], phrases)
self.assertEqual(phrases, ['Frank', 'Theodor', 'Richard'])
def test_dict2_all_fields(self):
"""bibindex termcollectors - get_field_values - nested field."""
fields = self.dict2
phrases = []
get_values_recursively(fields['all'], phrases)
self.assertEqual(phrases, ['name1', 'name2', 'name4'])
TEST_SUITE = make_test_suite(TestListSetOperations,
TestWashIndexTerm,
TestGetWordsFromPhrase,
TestGetPairsFromPhrase,
TestGetWordsFromDateTag,
TestGetAuthorFamilyNameWords,
TestGetValuesFromRecjson,)
if __name__ == "__main__":
run_test_suite(TEST_SUITE)

Event Timeline