wip D2950
This commit is contained in:
@@ -1,13 +1,7 @@
|
||||
import logging
|
||||
|
||||
#import bson
|
||||
#from flask import current_app
|
||||
|
||||
from .routes import blueprint_search
|
||||
|
||||
log = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def setup_app(app, url_prefix: str =None):
|
||||
app.register_api_blueprint(
|
||||
blueprint_search, url_prefix=url_prefix)
|
||||
|
@@ -30,8 +30,9 @@ def index_node_save(node_to_index: dict):
|
||||
|
||||
|
||||
def index_node_delete(delete_id: str):
|
||||
|
||||
""" delete node using id"""
|
||||
from pillar.api.utils import algolia
|
||||
|
||||
try:
|
||||
algolia.index_node_delete(delete_id)
|
||||
except AlgoliaException as ex:
|
||||
|
@@ -1,10 +1,20 @@
|
||||
"""
|
||||
Define elasticsearch document mapping.
|
||||
|
||||
Elasticsearch consist of two parts.
|
||||
|
||||
Part 1: Define the documents in which you
|
||||
define who fields will be indexed.
|
||||
Part 2: Building elasticsearch json queries.
|
||||
|
||||
BOTH of these parts are equally importand to have
|
||||
a search API that returns relevant results.
|
||||
"""
|
||||
import logging
|
||||
|
||||
import elasticsearch_dsl as es
|
||||
from elasticsearch_dsl import analysis
|
||||
# from pillar import current_app
|
||||
|
||||
# define elasticsearch document mapping.
|
||||
|
||||
|
||||
log = logging.getLogger(__name__)
|
||||
@@ -100,7 +110,14 @@ class Node(es.DocType):
|
||||
index = 'nodes'
|
||||
|
||||
|
||||
def create_doc_from_user_data(user_to_index):
|
||||
def create_doc_from_user_data(user_to_index: dict) -> User:
|
||||
"""
|
||||
We are indexing a user object which identical between search backends
|
||||
|
||||
see celery.search_index_task.py
|
||||
|
||||
this functions returns a proper ElasticSearch document
|
||||
"""
|
||||
|
||||
if not user_to_index:
|
||||
return
|
||||
@@ -108,8 +125,8 @@ def create_doc_from_user_data(user_to_index):
|
||||
doc_id = str(user_to_index.get('objectID', ''))
|
||||
|
||||
if not doc_id:
|
||||
log.error('ID missing %s', user_to_index)
|
||||
return
|
||||
log.error('USER ID is missing %s', user_to_index)
|
||||
raise KeyError
|
||||
|
||||
doc = User(_id=doc_id)
|
||||
doc.objectID = str(user_to_index['objectID'])
|
||||
@@ -121,7 +138,14 @@ def create_doc_from_user_data(user_to_index):
|
||||
return doc
|
||||
|
||||
|
||||
def create_doc_from_node_data(node_to_index):
|
||||
def create_doc_from_node_data(node_to_index: dict) -> Node:
|
||||
"""
|
||||
We are indexing a Node object which identical between search backends
|
||||
|
||||
see celery.search_index_task.py
|
||||
|
||||
this functions returns a proper ElasticSearch document
|
||||
"""
|
||||
|
||||
if not node_to_index:
|
||||
return
|
||||
|
@@ -6,6 +6,7 @@ from elasticsearch.exceptions import NotFoundError
|
||||
from pillar import current_app
|
||||
from . import documents
|
||||
|
||||
log = logging.getLogger(__name__)
|
||||
|
||||
elk_hosts = current_app.config['ELASTIC_SEARCH_HOSTS']
|
||||
|
||||
@@ -14,13 +15,10 @@ connections.create_connection(
|
||||
sniff_on_start=True,
|
||||
timeout=20)
|
||||
|
||||
log = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def push_updated_user(user_to_index: dict):
|
||||
"""
|
||||
Push an update to the Elastic index when
|
||||
a user item is updated.
|
||||
Push an update to the Elastic index when a user item is updated.
|
||||
"""
|
||||
if not user_to_index:
|
||||
return
|
||||
@@ -30,11 +28,14 @@ def push_updated_user(user_to_index: dict):
|
||||
if not doc:
|
||||
return
|
||||
|
||||
log.debug('UPDATE USER %s', doc._id)
|
||||
log.debug('index update user elasticsearch %s', doc._id)
|
||||
doc.save()
|
||||
|
||||
|
||||
def index_node_save(node_to_index: dict):
|
||||
"""
|
||||
Push an update to the Elastic index when a node item is saved.
|
||||
"""
|
||||
|
||||
if not node_to_index:
|
||||
return
|
||||
@@ -44,13 +45,16 @@ def index_node_save(node_to_index: dict):
|
||||
if not doc:
|
||||
return
|
||||
|
||||
log.debug('CREATED ELK NODE DOC %s', doc._id)
|
||||
log.debug('index created node elasticsearch %s', doc._id)
|
||||
doc.save()
|
||||
|
||||
|
||||
def index_node_delete(delete_id: str):
|
||||
"""
|
||||
Delete node document from Elastic index useing a node id
|
||||
"""
|
||||
|
||||
log.debug('NODE DELETE INDEXING %s', delete_id)
|
||||
log.debug('index node delete %s', delete_id)
|
||||
|
||||
try:
|
||||
doc = documents.Node.get(id=delete_id)
|
||||
|
@@ -1,10 +1,5 @@
|
||||
import logging
|
||||
# import time
|
||||
|
||||
# from elasticsearch import helpers
|
||||
# import elasticsearch
|
||||
|
||||
# from elasticsearch.client import IndicesClient
|
||||
from typing import List
|
||||
|
||||
from elasticsearch.exceptions import NotFoundError
|
||||
from elasticsearch_dsl.connections import connections
|
||||
@@ -18,17 +13,14 @@ log = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class ResetIndexTask(object):
|
||||
"""
|
||||
Clear and build index / mapping
|
||||
"""
|
||||
""" Clear and build index / mapping """
|
||||
index_key = ''
|
||||
"""Key into the ELASTIC_INDICES dict in the app config."""
|
||||
|
||||
doc_types = []
|
||||
doc_types: List[type] = []
|
||||
name = 'remove index'
|
||||
|
||||
def __init__(self):
|
||||
|
||||
if not self.index_key:
|
||||
raise ValueError("No index specified")
|
||||
|
||||
@@ -48,10 +40,10 @@ class ResetIndexTask(object):
|
||||
try:
|
||||
idx.delete(ignore=404)
|
||||
log.info("Deleted index %s", index)
|
||||
except AttributeError:
|
||||
log.warning("Could not delete index '%s', ignoring", index)
|
||||
except NotFoundError:
|
||||
log.warning("Could not delete index '%s', ignoring", index)
|
||||
else:
|
||||
log.warning("Could not delete index '%s', ignoring", index)
|
||||
|
||||
# create doc types
|
||||
for dt in self.doc_types:
|
||||
|
@@ -1,7 +1,7 @@
|
||||
import logging
|
||||
import json
|
||||
from elasticsearch import Elasticsearch
|
||||
from elasticsearch_dsl import Search, Q
|
||||
import logging
|
||||
|
||||
from pillar import current_app
|
||||
|
||||
@@ -23,13 +23,9 @@ def add_aggs_to_search(search, agg_terms):
|
||||
|
||||
|
||||
def make_must(must: list, terms: dict) -> list:
|
||||
"""
|
||||
Given some term parameters
|
||||
we must match those
|
||||
"""
|
||||
""" Given term parameters append must queries to the must list """
|
||||
|
||||
for field, value in terms.items():
|
||||
|
||||
if value:
|
||||
must.append({'match': {field: value}})
|
||||
|
||||
@@ -54,8 +50,10 @@ def nested_bool(must: list, should: list, terms: dict) -> Search:
|
||||
|
||||
def do_search(query: str, terms: dict) -> dict:
|
||||
"""
|
||||
Given user input search for node/stuff
|
||||
Given user query input and term refinements
|
||||
search for public published nodes
|
||||
"""
|
||||
|
||||
should = [
|
||||
Q('match', name=query),
|
||||
|
||||
@@ -77,21 +75,19 @@ def do_search(query: str, terms: dict) -> dict:
|
||||
search = nested_bool(must, should, terms)
|
||||
add_aggs_to_search(search, node_agg_terms)
|
||||
|
||||
if current_app.config['DEBUG']:
|
||||
if log.isEnabledFor(logging.DEBUG):
|
||||
print(json.dumps(search.to_dict(), indent=4))
|
||||
|
||||
response = search.execute()
|
||||
|
||||
if current_app.config['DEBUG']:
|
||||
if log.isEnabledFor(logging.DEBUG):
|
||||
print(json.dumps(response.to_dict(), indent=4))
|
||||
|
||||
return response.to_dict()
|
||||
|
||||
|
||||
def do_user_search(query: str, terms: dict) -> dict:
|
||||
"""
|
||||
return user objects
|
||||
"""
|
||||
""" return user objects represented in elasicsearch result dict"""
|
||||
should = [
|
||||
Q('match', username=query),
|
||||
Q('match', full_name=query),
|
||||
@@ -101,7 +97,7 @@ def do_user_search(query: str, terms: dict) -> dict:
|
||||
Q('term', _type='user')
|
||||
]
|
||||
|
||||
# We got an id field. we MUST find it.
|
||||
# We most likely got and id field. we MUST find it.
|
||||
if len(query) == len('563aca02c379cf0005e8e17d'):
|
||||
must.append(Q('term', _id=query))
|
||||
|
||||
@@ -111,12 +107,12 @@ def do_user_search(query: str, terms: dict) -> dict:
|
||||
search = nested_bool(must, should, terms)
|
||||
add_aggs_to_search(search, user_agg_terms)
|
||||
|
||||
if current_app.config['DEBUG']:
|
||||
if log.isEnabledFor(logging.DEBUG):
|
||||
print(json.dumps(search.to_dict(), indent=4))
|
||||
|
||||
response = search.execute()
|
||||
|
||||
if current_app.config['DEBUG']:
|
||||
if log.isEnabledFor(logging.DEBUG):
|
||||
print(json.dumps(response.to_dict(), indent=4))
|
||||
|
||||
return response.to_dict()
|
||||
@@ -124,7 +120,8 @@ def do_user_search(query: str, terms: dict) -> dict:
|
||||
|
||||
def do_user_search_admin(query: str) -> dict:
|
||||
"""
|
||||
return users with all fields and aggregations
|
||||
return users seach result dict object
|
||||
search all user fields and provide aggregation information
|
||||
"""
|
||||
should = [
|
||||
Q('match', username=query),
|
||||
@@ -135,12 +132,12 @@ def do_user_search_admin(query: str) -> dict:
|
||||
search = Search(using=client)
|
||||
search.query = bool_query
|
||||
|
||||
if current_app.config['DEBUG']:
|
||||
if log.isEnabledFor(logging.DEBUG):
|
||||
log.debug(json.dumps(search.to_dict(), indent=4))
|
||||
|
||||
response = search.execute()
|
||||
|
||||
if current_app.config['DEBUG']:
|
||||
if log.isEnabledFor(logging.DEBUG):
|
||||
log.debug(json.dumps(response.to_dict(), indent=4))
|
||||
|
||||
return response.to_dict()
|
||||
|
@@ -11,37 +11,28 @@ log = logging.getLogger(__name__)
|
||||
blueprint_search = Blueprint('elksearch', __name__)
|
||||
|
||||
|
||||
terms = [
|
||||
'node_type', 'media',
|
||||
'tags', 'is_free', 'projectname',
|
||||
'roles',
|
||||
]
|
||||
|
||||
|
||||
def _valid_search() -> str:
|
||||
"""
|
||||
Returns search parameters, raising error when missing.
|
||||
"""
|
||||
|
||||
searchword = request.args.get('q', '')
|
||||
# if not searchword:
|
||||
# raise wz_exceptions.BadRequest(
|
||||
# 'You are forgetting a "?q=whatareyoulookingfor"')
|
||||
return searchword
|
||||
""" Returns search parameters """
|
||||
query = request.args.get('q', '')
|
||||
return query
|
||||
|
||||
|
||||
def _term_filters() -> dict:
|
||||
"""
|
||||
Check if frontent wants to filter stuff
|
||||
on specific fields AKA facets
|
||||
|
||||
return mapping with term field name
|
||||
and provided user term value
|
||||
"""
|
||||
|
||||
terms = [
|
||||
'node_type', 'media',
|
||||
'tags', 'is_free', 'projectname',
|
||||
'roles',
|
||||
]
|
||||
|
||||
parsed_terms = {}
|
||||
|
||||
for term in terms:
|
||||
parsed_terms[term] = request.args.get(term, '')
|
||||
|
||||
return parsed_terms
|
||||
return {term: request.args.get(term, '') for term in terms}
|
||||
|
||||
|
||||
@blueprint_search.route('/')
|
||||
@@ -54,12 +45,9 @@ def search_nodes():
|
||||
|
||||
@blueprint_search.route('/user')
|
||||
def search_user():
|
||||
|
||||
searchword = _valid_search()
|
||||
|
||||
terms = _term_filters()
|
||||
data = queries.do_user_search(searchword, terms)
|
||||
|
||||
return jsonify(data)
|
||||
|
||||
|
||||
|
@@ -21,7 +21,6 @@ def index_user_save(to_index_user: dict):
|
||||
|
||||
@skip_when_testing
|
||||
def index_node_save(node_to_index):
|
||||
|
||||
if not current_app.algolia_index_nodes:
|
||||
return
|
||||
current_app.algolia_index_nodes.save_object(node_to_index)
|
||||
|
Reference in New Issue
Block a user