v2.0 upgrade #10

Merged
steve merged 26 commits from v2.0 into master 2024-02-09 16:00:08 +00:00
4 changed files with 728 additions and 169 deletions
Showing only changes of commit 79fd75bc9d - Show all commits

View File

@ -22,3 +22,4 @@ import healthcareio.x12 as x12
import healthcareio.params as params import healthcareio.params as params
# from healthcareio import server # from healthcareio import server

View File

@ -26,6 +26,7 @@ from multiprocessing import Process
import transport import transport
from transport import providers from transport import providers
import jsonmerge import jsonmerge
# import plugins
import copy import copy

View File

@ -1,45 +1,180 @@
""" """
This class refactors the default parsing class (better & streamlined implementation) This class refactors the default parsing class (better & streamlined implementation)
The class will rely on the new plug/play architectural style perform parsing
""" """
from multiprocessing import Process, RLock from multiprocessing import Process, RLock
import os import os
import json import json
# from healthcareio.x12.util
from healthcareio import x12
import numpy as np
import transport
import copy
# from healthcareio.x12.util import file as File, document as Document
import time
class BasicParser (Process) :
def __init__(self,**_args):
super().__init__()
self._plugins = _args['plugins']
self._parents = _args['parents']
self._files = _args['files']
self._store = _args['store']
class parser (Process) : def apply(self,**_args):
_CONFIGURATION = {} _content = _args['content']
def __init__(self,path=None) : _filetype = _args['x12']
if not parser._CONFIGURATION : _doc = _args['document'] #{}
_path = path if path else os.sep.join([os.environ['HOME'],'.healthcareio/config.json']) _documentHandler = x12.util.document.Builder(plugins = self._plugins,parents=self._parents)
try:
for _row in _content :
# _data = None
_data,_meta = _documentHandler.bind(row=_row,x12=_filetype)
if _data and _meta :
_doc = _documentHandler.build(data=_data,document=_doc,meta=_meta,row=_row)
# print (['*** ',_doc])
pass
except Exception as e:
# #
# @TODO: Load custom configuration just in case we need to do further processing # Log something here ....
config = json.loads(open(path).read()) print (_row)
parser._CONFIGURATION = config['parser'] print (e)
# print (_row,_doc.keys())
pass
return _doc
def run(self):
_handleContent = x12.util.file.Content()
_handleDocument = x12.util.document.Builder(plugins = self._plugins,parents=self._parents)
_template = _plugins,_parents = x12.util.template(plugins=self._plugins)
for _absolute_path in self._files :
try:
_content = _handleContent.read(filename=_absolute_path)
_content,_filetype = _handleContent.split(_content)
# #
# do we have a custom configuration in this location # The first row is the header (it will be common to all claims)
# _header = copy.deepcopy(_template[_filetype])
_custompath = _path.replace('config.json','') _header = self.apply(content=_content[0],x12=_filetype, document=_header)
_custompath = _custompath if not _custompath.endswith(os.sep) else _custompath[:-1] _docs = []
_custompath = os.sep.join([_custompath,'custom'])
if os.exists(_custompath) :
files = os.listdir(_custompath)
if files :
_filename = os.sep.join([_custompath,files[0]])
_customconf = json.loads(open(_filename).read())
#
# merge with existing configuration
for _rawclaim in _content[1:] :
_document = copy.deepcopy(_header) #copy.deepcopy(_template[_filetype])
# _document = dict(_document,**_header)
if type(_absolute_path) == str:
_document['filename'] = _absolute_path
_doc = self.apply(content=_rawclaim,x12=_filetype, document=_document)
if _doc :
#
# @TODO: Make sure the test here is the existence of the primary key
# _doc = _handleDocument.merge(_doc,_header)
_docs.append(_doc)
else: else:
# print (['wtf ...',_rawclaim])
pass
#
# Let us submit the batch we have thus far
#
self.post(documents=_docs,type=_filetype)
except Exception as e:
print (e)
def post(self,**_args):
pass
class X12Parser(BasicParser):
def __init__(self,**_args):
super().__init__(**_args)
self._store = _args['store']
def post(self,**_args):
"""
Writing the files to a persistent storage in JSON format (hopefully)
"""
_documents = _args['documents']
if _documents :
_store = copy.copy(self._store,**{})
TABLE = 'claims' if _args['type'] in ['837','claims'] else 'remits'
_store['table'] = TABLE
_writer = transport.factory.instance(**_store)
_writer.write(_documents)
if getattr(_writer,'close') :
_writer.close()
def instance (**_args):
"""
:path
"""
# _files = x12.util.Files.get(_args['file'])
# #
# # We can split these files (multi-processing)
# #
# _jobCount = 1 if 'jobs' not in _args else int (_args['jobs'])
# _files = np.array_split(_files,_jobCount)
# PATH = os.sep.join([os.environ['HOME'],'.healthcareio','config.json'])
# if 'config' in _args :
# PATH = _args['config']
# f = open(PATH)
# _config = json.loads(f.read())
# f.close()
# jobs = []
# for _batch in _files :
# pthread = Parser(files=_batch,config=_config)
# pthread.start()
# jobs.append(pthread)
# time.sleep(1)
pass pass
#
# # class parser (Process) :
class getter : # _CONFIGURATION = {}
def value(self,) : # def __init__(self,path=None) :
pass # if not parser._CONFIGURATION :
class setter : # _path = path if path else os.sep.join([os.environ['HOME'],'.healthcareio/config.json'])
def files(self,files): # #
pass # # @TODO: Load custom configuration just in case we need to do further processing
# config = json.loads(open(path).read())
# parser._CONFIGURATION = config['parser']
# #
# # do we have a custom configuration in this location
# #
# _custompath = _path.replace('config.json','')
# _custompath = _custompath if not _custompath.endswith(os.sep) else _custompath[:-1]
# _custompath = os.sep.join([_custompath,'custom'])
# if os.exists(_custompath) :
# files = os.listdir(_custompath)
# if files :
# _filename = os.sep.join([_custompath,files[0]])
# _customconf = json.loads(open(_filename).read())
# #
# # merge with existing configuration
# else:
# pass
# #
# #
# class getter :
# def value(self,) :
# pass
# class setter :
# def files(self,files):
# pass

View File

@ -7,151 +7,573 @@ Contract:
""" """
import os import os
import importlib as IL import importlib as IL
# import imp
import sys import sys
from healthcareio import x12
class parser:
"""
This is a decorator used against the plugins to determine which function get applied against a given x12 element
The following are mandatory :
{element,map} or {element,columns},
- Providing the map attribute will invoke the built-in parser
- not providing the built-in parser will suggest that the code will handle parsing itself and thus must inform of the attributes expected
"""
def __init__(self,**_args):
self.element = _args['element']
self.x12 = _args['x12'] if 'x12' in _args else '*'
self.parent = _args['parent'] if 'parent' in _args else None
self.map = None if 'map' not in _args else _args['map']
self.cols = None if 'columns' not in _args else _args['columns']
self.anchor = None if 'anchor' not in _args else _args['anchor']
if 'field' in _args :
self.field = _args['field']
elif 'container' in _args :
self.container = _args['container']
def __call__(self, pointer):
def wrapper(**args): #(*args, **kwargs):
#
# @TODO: Log this in case we have an error
return pointer(**args) #(*args, **kwargs)
#
#-- These attributes will be read by the factory class to make sure the functions are assigned the data they are designed for
setattr(wrapper,'element',self.element)
setattr(wrapper,'x12',self.x12)
setattr(wrapper,'parent',self.parent)
_meta = {'element':self.element,'x12':self.x12,'parent':self.parent}
if self.cols :
setattr(wrapper,'columns',self.cols)
_meta['columns'] = self.cols
elif self.map :
setattr(wrapper,'map',self.map)
_meta['map'] = self.map
if hasattr(self,'container') :
setattr(wrapper,'container',self.container)
_meta['container'] = self.container
if hasattr(self,'field') :
setattr(wrapper,'field',self.field)
_meta['field'] = self.field
if hasattr(self,'anchor') and self.anchor:
_meta['anchor'] = self.anchor
setattr(wrapper,'anchor',self.anchor)
setattr(wrapper,'meta',_meta)
#
# clean up here ....
# 1. if anchor no field|containers
# 2. if parent no field or containers
# 3. field and containers can't be
return wrapper
DEFAULT_PLUGINS='healthcareio.x12.plugins.default' DEFAULT_PLUGINS='healthcareio.x12.plugins.default'
class MODE : def isplugin(_module,_name) :
TRUST,CHECK,TEST,TEST_AND_CHECK= [0,1,2,3]
def instance(**_args) :
pass
def has(**_args) :
""" """
This function will inspect if a function is valid as a plugin or not This function for a given module returns a True if a function is a plugin if
name : function name for a given file :_module module
path : python file to examine :_name a given name of a resource in a module
""" """
_pyfile = _args['path'] if 'path' in _args else ''
_name = _args['name']
# p = os.path.exists(_pyfile)
_module = {}
if os.path.exists(_pyfile):
_info = IL.utils.spec_from_file_location(_name,_pyfile)
if _info :
_module = IL.utils.module_from_spec(_info)
_info.load.exec(_module)
p = type(getattr(_module,_name)).__name__ =='function' # is a function
q = hasattr(getattr(_module,_name),'element') # has element {x12}
r = hasattr(getattr(_module,_name),'x12') # has an {x12} type i.e *,835,837
return p and q and r
def build_map (**_args):
"""
This function builds the map of {x12,element,pointer} to enable rapid access to parsing fucntions
:module module object as returned framework (importlib or imp)
"""
_plugins = [getattr(_args['module'],_name) for _name in dir(_args['module']) if isplugin(_args['module'],_name)]
_map = {}
_parents = {}
for _item in _plugins :
_id = _item.x12
if _id not in _map :
_map[_id] = {}
_element = _item.element
if _item.parent :
if _item.parent not in _parents :
_parents[_item.parent] = []
_parents[_item.parent].append (_item.element)
if type(_element) == list :
for _e in _element :
_map[_id][_e.strip()] = _item
else: else:
_module = sys.modules[DEFAULT_PLUGINS] _map[_id][_element]= _item
return hasattr(_module,_name) return _map,_parents
def get(**_args) :
"""
This function will inspect if a function is valid as a plugin or not
name : function name for a given file
path : python file to examine
"""
_pyfile = _args['path'] if 'path' in _args else ''
_name = _args['name']
# p = os.path.exists(_pyfile)
_module = {}
if os.path.exists(_pyfile):
_info = IL.utils.spec_from_file_location(_name,_pyfile)
if _info :
_module = IL.utils.module_from_spec(_info)
_info.load.exec(_module)
def instance(**_args):
else:
_module = sys.modules[DEFAULT_PLUGINS]
return getattr(_module,_name) if hasattr(_module,_name) else None
def test (**_args):
""" """
This function will test a plugin to insure the plugin conforms to the norm we are setting here This function returns an dictionary/map of functions available for execution
:pointer function to call The functions are organized as follows: {835:{BHT:_pointer,NM1:_pointer}}
:_args
- path provides the location (folder|file) to be extracted
""" """
_params = {}
try:
if 'pointer' in _args :
_caller = _args['pointer']
else:
_name = _args['name']
_path = _args['path'] if 'path' in _args else None
_caller = get(name=_name,path=_path)
_params = _caller()
# #
# the expected result is a list of field names [field_o,field_i] # Loading one of the default functions built-in ...
# #
return [_item for _item in _params if _item not in ['',None] and type(_item) == str] _module = IL.import_module(DEFAULT_PLUGINS)
except Exception as e : _map,_parents = build_map(module=_module)
return [] #
# loaded the common functions, now load domain specific ones one
#
_packages = ['remits','claims']
_handler = x12.util.document.Builder(parents={},plugins={})
for _x12 in _packages :
if hasattr(_module,_x12) :
_ixmap,_ixparents = build_map(module=getattr(_module,_x12))
_map = _handler.merge(_map,_ixmap)
_parents = _handler.merge(_parents,_ixparents)
#
# consolidate, the common elements across the ...
# We override any common processing element with the version specific element
_map['835'] = _handler.merge(_map['*'],_map['835'])
_map['837'] = _handler.merge(_map['*'],_map['837'])
if 'path' in _args:
#
# We can/will override the default modules given the user has provided a location
# _module = imp.load_source('udf',_args['path'])
_module = IL.machinery.SourcefileLoader('udf',_args['path']).load_module()
_udf_map,_udfp = build_map(module=_module)
_map = dict(_map,**_udf_map)
for key in _udfp:
if key not in _parents :
_parents[key] = _udfp[key]
else:
_parents[key] = _parents[key] + _udfp[key]
if 'filter' in _args :
return filter(elements = _args['filter'],plugins=_map)
# return _smap
return _map,_parents
def merge (_x,_y):
"""
This function will merge two objects _x, _y
"""
_zcols = list(set(_x.keys()) & set(_y.keys())) #--common columns
if _zcols :
_out = dict(_x,**{})
for _key in list(_y.keys()) :
if _key not in _zcols and _key:
_out[_key] = _y[_key]
else:
if type(_out[_key]) == list :
for value in _y[_key] :
if value not in _out[_key] :
_out[_key].append(value)
# _out[_key] += _y[_key]
elif type(_out[_key]) == dict:
_out[_key] = dict(_out[_key],**_y[_key])
else:
_out[_key] = _y[_key]
return _out
else:
return dict(_x,**_y)
# def template(**_args) :
# """
# This function generates an object template to be used in object assignment and export functionalities
# We chose to proceed in this manner so as to enforce consistency of the parser
# :plugins {*,837,835} with element and pointers associated
# """
# _plugins = _args['plugins']
# _object = {'837':{},'835':{}}
# for _x12 in _plugins :
# _pointers = _plugins[_x12]
# for _element in _pointers :
# _meta = _pointers[_element].meta
# _values = _meta['map'].values() if 'map' in _meta else _meta['columns']
# #
# # where do the attributes go ..
# #
# _attr = []
# for _item in list(_values) :
# if type(_item) == list :
# _attr = _attr + _item
# else:
# _attr.append(_item)
# _field = []
# if 'field' in _meta or 'container' in _meta :
# _field = _meta['field'] if 'field' in _meta else _meta['container']
# if 'anchor' in _meta : #-- No parents are expected
# _field = _meta['anchor'].values()
# elif _meta['parent'] :
# #
# # It means the attributes will be
# _parentPlug = filter(elements=[_meta['parent']],plugins=_plugins)
# _pid = list(_parentPlug.keys())[0]
# _parentMeta = _parentPlug[_pid][_meta['parent']].meta
# _attr = _attr + list(_parentMeta['map'].values()) if 'map' in _parentMeta else _parentMeta['columns']
# if 'anchor' in _parentMeta :
# _field = list(_parentMeta['anchor'].values())
# _field = [_field] if type(_field) == str else _field
# _attr = dict.fromkeys(_attr,'')
# if not _field :
# _info = (_attr)
# else:
# _info = (dict.fromkeys(_field,_attr))
# if _x12 == '*' :
# _object['837']= merge(_object['837'], _info)
# _object['835']= merge (_object['835'], _info)
# else:
# _object[_x12] = merge(_object[_x12],_info)
# return _object
def filter (**_args) :
_elements = _args['elements']
_plugins = _args['plugins']
_map = {}
for x12 in _plugins :
_item = _plugins[x12]
_found = list(set(_elements) & set(_item.keys()))
if _found :
_map[x12] = {key:_item[key] for key in _found }
return _map
def getTableName(**_args) :
_plugins = _args['plugins']
_meta = _args['meta']
_x12 = _meta['x12']
_foreignkeys = _args['tableKeys']
_attributes = list(_meta['map'].values()) if 'map' in _meta else _meta['columns']
if 'field' in _meta or 'container' in _meta:
_tableName = _meta['field'] if 'field' in _meta else _meta['container']
# _table = {_id:_attributes}
elif 'anchor' in _meta :
_tableName = _meta['anchor'].values()
# for _name in _meta['anchor'].values() :
# _table[_name] = _attributes
elif 'parent' in _meta and _meta['parent']:
#
# We can have a parent with no field/container/anchor
# We expect either a map or columns ...
#
_parentElement = _meta['parent']
_parentMeta = _plugins[_x12][_parentElement].meta
_parentTable = getTableName(plugins=_plugins,meta = _parentMeta,tableKeys=_foreignkeys)
_tableName = list(_parentTable.keys())[0]
# _table[_id] = _parentTable[_id] + _attributes
_attributes = _parentTable[_tableName] + _attributes
# print (_meta)
else:
#
# baseline tables have no parent, we need to determine the name
#
_tableName = 'claims' if _x12 == '837' else 'remits'
# print (_id,_attributes)
pass pass
def inspect(**_args): #
_mode = _args['mode'] # Are there any anchors
_name= _args['name'] if _x12 == '837':
_path= _args['path'] _keys = [_foreignkeys['claims']]
if _mode == MODE.CHECK : elif _x12 == '835' :
_doapply = [has] _keys = [_foreignkeys['remits']]
elif _mode == MODE.TEST :
_doapply = [test]
elif _mode == MODE.TEST_AND_CHECK :
_doapply = [has,test]
_status = True
_plugin = {"name":_name}
for _pointer in _doapply :
_plugin[_pointer.__name__] = _pointer(name=_name,path=_path)
if not _plugin[_pointer.__name__] :
_status = False
break
_plugin['loaded'] = _status
return _plugin
def load(**_args):
"""
This function will load all the plugins given an set of arguments :
path file
name list of functions to export
mode 1- CHECK ONLY, 2 - TEST ONLY, 3- TEST_AND_CHECK
"""
_path = _args ['path']
_names= _args['names']
_mode= _args ['mode'] if 'mode' in _args else MODE.TEST_AND_CHECK
_doapply = []
if _mode == MODE.CHECK :
_doapply = [has]
elif _mode == MODE.TEST :
_doapply = [test]
elif _mode == MODE.TEST_AND_CHECK :
_doapply = [has,test]
# _plugins = []
_plugins = {}
for _name in _names :
_plugin = {"name":_name}
if 'inspect' in _args and _args['inspect'] :
_plugin = inspect(name=_name,mode=_mode,path=_path)
else: else:
_plugin["method"] = "" _keys = list(set(_foreignkeys.values()))
_status = True _attr = []
_plugin['loaded'] = _status for _item in _attributes :
if _plugin['loaded'] : if type(_item) == list :
_plugin['pointer'] = get(name=_name,path=_path) _attr += _item
else: else:
_plugin['pointer'] = None _attr.append(_item)
_keys = list(set(_keys) - set(_attr))
_attr = _keys + _attr
# if 'container' in _meta and _meta['container'] == 'procedures' :
# print (_attributes)
_tableName = [_tableName] if type(_tableName) == str else _tableName
return dict.fromkeys(_tableName,_attr)
def _getTableName (**_args):
"""
This function provides a list of attributes associated with an entity
The function infers a relational structure from the JSON representation of a claim and plugin specifications
"""
_meta = _args['meta']
_xattr = list(_meta['map'].values()) if 'map' in _meta else _meta['columns']
_plugins = _args['plugins']
_foreignkeys = _args['tableKeys']
#
# Fix attributes, in case we have an index associated with multiple fields
#
_attr = []
if 'anchor' not in _meta and not _meta['parent']:
for _item in _xattr :
_attr += _item if type(_item) == list else [_item]
_name = None
_info = {}
_infoparent = {}
if 'field' in _meta :
_name = _meta['field']
elif 'container' in _meta :
_name = _meta['container']
elif 'anchor' in _meta :
# _plugins.append(_plugin) _name = list(_meta['anchor'].values())
_plugins[_name] = _plugin # if _name :
return _plugins # _name = _name if type(_name) == list else [_name]
# _info = dict.fromkeys(_name,_attr)
if _meta['parent'] :
_parentElement = filter(elements=[_meta['parent']],plugins=_plugins)
_x12 = list(_parentElement.keys())[0]
_id = list(_parentElement[_x12].keys())[0]
_infoparent = getTableName(meta = _parentElement[_x12][_id].meta,plugins=_plugins,tableKeys=_foreignkeys)
if _meta['x12'] == '*' :
_name = ['claims','remits'] if not _name else _name
_attr = list(set(_foreignkeys.values())) + _attr
else:
_name = 'claims' if _meta['x12'] == '837' and not _name else ('remits' if not _name and _meta['x12'] == '835' else _name)
_id = 'claims' if _meta['x12'] == '837' else 'remits'
if _id in _foreignkeys:
_attr = [_foreignkeys[_id]] + _attr
def parse(**_args): # if not _name :
""" # if _meta['x12'] == '*' :
This function will apply a function against a given function, and data # _name = ['claims','remits']
:row claim/remits pre-processed # else:
:plugins list of plugins # _name = 'claims' if _meta['x12'] == '837' else 'remits'
:conifg configuration associated with #
""" # Let us make sure we can get the keys associated here ...
_row = _args['row'] #
_document = _args['document'] # filter (elements = [])
_config = _args['config'] _name = _name if type(_name) == list else [_name]
""" _info = dict.fromkeys(_name,_attr)
"apply":"@path:name" if _infoparent:
""" _info = dict(_info,**_infoparent)
return _info
def getTableKeys(**_args):
_plugins=_args['plugins']
_pointer = filter(elements=['CLM'],plugins=_plugins)
_keys = {}
for _element in ['CLM','CLP'] :
_pointer = filter(elements=[_element],plugins=_plugins)
if not _pointer :
continue
_pointer = list(_pointer.values())[0]
_meta = _pointer[_element].meta
_name = _meta['map'][1] if 'map' in _meta else _meta['columns'][0]
_id = 'claims' if _element == 'CLM' else 'remits'
_keys[_id] = _name
return _keys
# print (list(_pointer.values())[0]['CLM'].meta)
# print (_pointer.values()[0].meta)
def sql (**_args):
_plugins = _args['plugins']
# _info = {'foreign':{},'keys':{'claims':None,'remits':None}}
_documentHandler = x12.util.document.Builder(plugins=_plugins,parents=_args['parents'])
_tableKeys = getTableKeys(plugins=_plugins)
_schemas = {}
for key in _plugins :
_mpointers = _plugins[key]
for _element in _mpointers :
_pointer = _mpointers[_element]
_meta = _pointer.meta
_info = getTableName(meta=_meta,plugins=_plugins,tableKeys=_tableKeys)
# _schemas = dict(_schemas,**_info)
if _info :
_schemas = _documentHandler.merge(_schemas,_info)
# print (_info)
return _schemas
# if not _info :
# print (_meta)
# continue
# if _meta['x12'] in ['837','837'] :
# _schema_id = 'claims' if _meta['x12'] == '837' else 'remits'
# _schema_id = [_schema_id]
# else:
# _schema_id = ['claims','remits']
# if _info :
# #
# # foreign tables need to be placed here
# for _id in _schema_id :
# if type(_info) == list :
# _schemas[_id]['attributes'] += _info
# else:
# _schemas[_id]['foreign'] = dict(_schemas[_id]['foreign'],**_info)
# else:
# #
# # This one goes to the main tables
# for _id in _schema_id :
# print (_info)
# _schemas[_id]['attributes'] += list(_info.values())
# DEFAULT_PLUGINS='healthcareio.x12.plugins.default'
# class MODE :
# TRUST,CHECK,TEST,TEST_AND_CHECK= [0,1,2,3]
# def instance(**_args) :
# pass
# def has(**_args) :
# """
# This function will inspect if a function is valid as a plugin or not
# name : function name for a given file
# path : python file to examine
# """
# _pyfile = _args['path'] if 'path' in _args else ''
# _name = _args['name']
# # p = os.path.exists(_pyfile)
# _module = {}
# if os.path.exists(_pyfile):
# _info = IL.utils.spec_from_file_location(_name,_pyfile)
# if _info :
# _module = IL.utils.module_from_spec(_info)
# _info.load.exec(_module)
# else:
# _module = sys.modules[DEFAULT_PLUGINS]
# return hasattr(_module,_name)
# def get(**_args) :
# """
# This function will inspect if a function is valid as a plugin or not
# name : function name for a given file
# path : python file to examine
# """
# _pyfile = _args['path'] if 'path' in _args else ''
# _name = _args['name']
# # p = os.path.exists(_pyfile)
# _module = {}
# if os.path.exists(_pyfile):
# _info = IL.utils.spec_from_file_location(_name,_pyfile)
# if _info :
# _module = IL.utils.module_from_spec(_info)
# _info.load.exec(_module)
# else:
# _module = sys.modules[DEFAULT_PLUGINS]
# return getattr(_module,_name) if hasattr(_module,_name) else None
# def test (**_args):
# """
# This function will test a plugin to insure the plugin conforms to the norm we are setting here
# :pointer function to call
# """
# _params = {}
# try:
# if 'pointer' in _args :
# _caller = _args['pointer']
# else:
# _name = _args['name']
# _path = _args['path'] if 'path' in _args else None
# _caller = get(name=_name,path=_path)
# _params = _caller()
# #
# # the expected result is a list of field names [field_o,field_i]
# #
# return [_item for _item in _params if _item not in ['',None] and type(_item) == str]
# except Exception as e :
# return []
# pass
# def inspect(**_args):
# _mode = _args['mode']
# _name= _args['name']
# _path= _args['path']
# if _mode == MODE.CHECK :
# _doapply = [has]
# elif _mode == MODE.TEST :
# _doapply = [test]
# elif _mode == MODE.TEST_AND_CHECK :
# _doapply = [has,test]
# _status = True
# _plugin = {"name":_name}
# for _pointer in _doapply :
# _plugin[_pointer.__name__] = _pointer(name=_name,path=_path)
# if not _plugin[_pointer.__name__] :
# _status = False
# break
# _plugin['loaded'] = _status
# return _plugin
# def load(**_args):
# """
# This function will load all the plugins given an set of arguments :
# path file
# name list of functions to export
# mode 1- CHECK ONLY, 2 - TEST ONLY, 3- TEST_AND_CHECK
# """
# _path = _args ['path']
# _names= _args['names']
# _mode= _args ['mode'] if 'mode' in _args else MODE.TEST_AND_CHECK
# _doapply = []
# if _mode == MODE.CHECK :
# _doapply = [has]
# elif _mode == MODE.TEST :
# _doapply = [test]
# elif _mode == MODE.TEST_AND_CHECK :
# _doapply = [has,test]
# # _plugins = []
# _plugins = {}
# for _name in _names :
# _plugin = {"name":_name}
# if 'inspect' in _args and _args['inspect'] :
# _plugin = inspect(name=_name,mode=_mode,path=_path)
# else:
# _plugin["method"] = ""
# _status = True
# _plugin['loaded'] = _status
# if _plugin['loaded'] :
# _plugin['pointer'] = get(name=_name,path=_path)
# else:
# _plugin['pointer'] = None
# # _plugins.append(_plugin)
# _plugins[_name] = _plugin
# return _plugins
# def parse(**_args):
# """
# This function will apply a function against a given function, and data
# :row claim/remits pre-processed
# :plugins list of plugins
# :conifg configuration associated with
# """
# _row = _args['row']
# _document = _args['document']
# _config = _args['config']
# """
# "apply":"@path:name"
# """
# _info = _args['config']['apply']
# _plug_conf = _args['config']['plugin'] if 'plugin' in _args['config'] else {}
# if _info.startswith('@') :
# _path = '' #-- get this from general configuration
# elif _info.startswith('!'):
# _path = _info.split('!')[0][1:]
# _name = _info.split(':')[-1]
# _name = _args['config']['apply'].split(_path)
_info = _args['config']['apply']
_plug_conf = _args['config']['plugin'] if 'plugin' in _args['config'] else {}
if _info.startswith('@') :
_path = '' #-- get this from general configuration
elif _info.startswith('!'):
_path = _info.split('!')[0][1:]
_name = _info.split(':')[-1]
_name = _args['config']['apply'].split(_path)