X-Git-Url: https://git.openstreetmap.org/nominatim.git/blobdiff_plain/e14e7c6235a40dbec451146bcb3aaec013d659c9..3206bf59df0213d24bd3e11df7dd2abaebf89911:/test/python/test_tools_database_import.py diff --git a/test/python/test_tools_database_import.py b/test/python/test_tools_database_import.py index 45324834..e370e084 100644 --- a/test/python/test_tools_database_import.py +++ b/test/python/test_tools_database_import.py @@ -80,39 +80,6 @@ def test_setup_extensions_old_postgis(temp_db_conn, monkeypatch): database_import.setup_extensions(temp_db_conn) -def test_install_module(tmp_path): - src_dir = tmp_path / 'source' - src_dir.mkdir() - (src_dir / 'nominatim.so').write_text('TEST nomiantim.so') - - project_dir = tmp_path / 'project' - project_dir.mkdir() - - database_import.install_module(src_dir, project_dir, '') - - outfile = project_dir / 'module' / 'nominatim.so' - - assert outfile.exists() - assert outfile.read_text() == 'TEST nomiantim.so' - assert outfile.stat().st_mode == 33261 - - -def test_install_module_custom(tmp_path): - (tmp_path / 'nominatim.so').write_text('TEST nomiantim.so') - - database_import.install_module(tmp_path, tmp_path, str(tmp_path.resolve())) - - assert not (tmp_path / 'module').exists() - - -def test_install_module_fail_access(temp_db_conn, tmp_path): - (tmp_path / 'nominatim.so').write_text('TEST nomiantim.so') - - with pytest.raises(UsageError, match='.*module cannot be accessed.*'): - database_import.install_module(tmp_path, tmp_path, '', - conn=temp_db_conn) - - def test_import_base_data(src_dir, temp_db, temp_db_cursor): temp_db_cursor.execute('CREATE EXTENSION hstore') temp_db_cursor.execute('CREATE EXTENSION postgis') @@ -171,23 +138,24 @@ def test_import_osm_data_default_cache(temp_db_cursor,osm2pgsql_options): def test_truncate_database_tables(temp_db_conn, temp_db_cursor, table_factory): - tables = ('word', 'placex', 'place_addressline', 'location_area', - 'location_area_country', 'location_property', + tables = ('placex', 'place_addressline', 'location_area', + 'location_area_country', 'location_property_tiger', 'location_property_osmline', 'location_postcode', 'search_name', 'location_road_23') for table in tables: - table_factory(table, content=(1, 2, 3)) + table_factory(table, content=((1, ), (2, ), (3, ))) + assert temp_db_cursor.table_rows(table) == 3 - database_import.truncate_data_tables(temp_db_conn, max_word_frequency=23) + database_import.truncate_data_tables(temp_db_conn) for table in tables: assert temp_db_cursor.table_rows(table) == 0 @pytest.mark.parametrize("threads", (1, 5)) -def test_load_data(dsn, src_dir, place_row, placex_table, osmline_table, word_table, - temp_db_cursor, threads): - for func in ('make_keywords', 'getorcreate_housenumber_id', 'make_standard_name'): +def test_load_data(dsn, src_dir, place_row, placex_table, osmline_table, + word_table, temp_db_cursor, threads): + for func in ('precompute_words', 'getorcreate_housenumber_id', 'make_standard_name'): temp_db_cursor.execute("""CREATE FUNCTION {} (src TEXT) RETURNS TEXT AS $$ SELECT 'a'::TEXT $$ LANGUAGE SQL """.format(func)) @@ -196,7 +164,33 @@ def test_load_data(dsn, src_dir, place_row, placex_table, osmline_table, word_ta place_row(osm_type='W', osm_id=342, cls='place', typ='houses', geom='SRID=4326;LINESTRING(0 0, 10 10)') - database_import.load_data(dsn, src_dir / 'data', threads) + database_import.load_data(dsn, threads) assert temp_db_cursor.table_rows('placex') == 30 assert temp_db_cursor.table_rows('location_property_osmline') == 1 + + +@pytest.mark.parametrize("languages", (None, ' fr,en')) +def test_create_country_names(temp_db_with_extensions, temp_db_conn, temp_db_cursor, + table_factory, tokenizer_mock, languages): + + table_factory('country_name', 'country_code varchar(2), name hstore', + content=(('us', '"name"=>"us1","name:af"=>"us2"'), + ('fr', '"name"=>"Fra", "name:en"=>"Fren"'))) + + assert temp_db_cursor.scalar("SELECT count(*) FROM country_name") == 2 + + tokenizer = tokenizer_mock() + + database_import.create_country_names(temp_db_conn, tokenizer, languages) + + assert len(tokenizer.analyser_cache['countries']) == 2 + + result_set = {k: set(v) for k, v in tokenizer.analyser_cache['countries']} + + if languages: + assert result_set == {'us' : set(('us', 'us1', 'United States')), + 'fr' : set(('fr', 'Fra', 'Fren'))} + else: + assert result_set == {'us' : set(('us', 'us1', 'us2', 'United States')), + 'fr' : set(('fr', 'Fra', 'Fren'))}