bug fixes and simplifying interfaces
This commit is contained in:
parent
ccc05acc01
commit
b239a5149f
2
setup.py
2
setup.py
|
@ -8,7 +8,7 @@ def read(fname):
|
||||||
return open(os.path.join(os.path.dirname(__file__), fname)).read()
|
return open(os.path.join(os.path.dirname(__file__), fname)).read()
|
||||||
args = {
|
args = {
|
||||||
"name":"data-transport",
|
"name":"data-transport",
|
||||||
"version":"1.4.1",
|
"version":"1.4.4",
|
||||||
"author":"The Phi Technology LLC","author_email":"info@the-phi.com",
|
"author":"The Phi Technology LLC","author_email":"info@the-phi.com",
|
||||||
"license":"MIT",
|
"license":"MIT",
|
||||||
"packages":["transport"]}
|
"packages":["transport"]}
|
||||||
|
|
|
@ -1,45 +1,24 @@
|
||||||
"""
|
"""
|
||||||
Data Transport - 1.0
|
Data Transport, The Phi Technology LLC
|
||||||
Steve L. Nyemba, The Phi Technology LLC
|
Steve L. Nyemba, steve@the-phi.com
|
||||||
|
|
||||||
This module is designed to serve as a wrapper to a set of supported data stores :
|
This library is designed to serve as a wrapper to a set of supported data stores :
|
||||||
- couchdb
|
- couchdb
|
||||||
- mongodb
|
- mongodb
|
||||||
- Files (character delimited)
|
- Files (character delimited)
|
||||||
- Queues (RabbmitMq)
|
- Queues (RabbmitMq)
|
||||||
- Session (Flask)
|
- Session (Flask)
|
||||||
- s3
|
- s3
|
||||||
|
- sqlite
|
||||||
The supported operations are read/write and providing meta data to the calling code
|
The supported operations are read/write and providing meta data to the calling code
|
||||||
Requirements :
|
Requirements :
|
||||||
pymongo
|
pymongo
|
||||||
boto
|
boto
|
||||||
couldant
|
couldant
|
||||||
The configuration for the data-store is as follows :
|
The configuration for the data-store is as follows :
|
||||||
couchdb:
|
e.g:
|
||||||
{
|
mongodb
|
||||||
args:{
|
provider:'mongodb',[port:27017],[host:localhost],db:<name>,doc:<_name>,context:<read|write>
|
||||||
url:<url>,
|
|
||||||
username:<username>,
|
|
||||||
password:<password>,
|
|
||||||
dbname:<database>,
|
|
||||||
doc:<document id>
|
|
||||||
}
|
|
||||||
}
|
|
||||||
RabbitMQ:
|
|
||||||
{
|
|
||||||
|
|
||||||
}
|
|
||||||
Mongodb:
|
|
||||||
{
|
|
||||||
args:{
|
|
||||||
host:<url>, #localhost:27017
|
|
||||||
username:<username>,
|
|
||||||
password:<password>,
|
|
||||||
dbname:<database>,
|
|
||||||
doc:<document id>s
|
|
||||||
|
|
||||||
}
|
|
||||||
}
|
|
||||||
"""
|
"""
|
||||||
__author__ = 'The Phi Technology'
|
__author__ = 'The Phi Technology'
|
||||||
import pandas as pd
|
import pandas as pd
|
||||||
|
@ -92,7 +71,15 @@ class factory :
|
||||||
PROVIDERS['sqlite3'] = PROVIDERS['sqlite']
|
PROVIDERS['sqlite3'] = PROVIDERS['sqlite']
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def instance(**args):
|
def instance(**_args):
|
||||||
|
if 'type' in _args :
|
||||||
|
#
|
||||||
|
# Legacy code being returned
|
||||||
|
return factory._instance(**_args);
|
||||||
|
else:
|
||||||
|
return instance(**_args)
|
||||||
|
@staticmethod
|
||||||
|
def _instance(**args):
|
||||||
"""
|
"""
|
||||||
This class will create an instance of a transport when providing
|
This class will create an instance of a transport when providing
|
||||||
:type name of the type we are trying to create
|
:type name of the type we are trying to create
|
||||||
|
@ -131,7 +118,7 @@ def instance(**_args):
|
||||||
"""
|
"""
|
||||||
|
|
||||||
provider = _args['provider']
|
provider = _args['provider']
|
||||||
context = _args['context']
|
context = _args['context']if 'context' in _args else None
|
||||||
_id = context if context in ['read','write'] else 'read'
|
_id = context if context in ['read','write'] else 'read'
|
||||||
if _id :
|
if _id :
|
||||||
args = {'provider':_id}
|
args = {'provider':_id}
|
||||||
|
@ -142,6 +129,7 @@ def instance(**_args):
|
||||||
args[key] = value
|
args[key] = value
|
||||||
#
|
#
|
||||||
#
|
#
|
||||||
|
|
||||||
args = dict(args,**_args)
|
args = dict(args,**_args)
|
||||||
|
|
||||||
# print (provider in factory.PROVIDERS)
|
# print (provider in factory.PROVIDERS)
|
||||||
|
@ -149,6 +137,7 @@ def instance(**_args):
|
||||||
pointer = factory.PROVIDERS[provider]['class'][_id]
|
pointer = factory.PROVIDERS[provider]['class'][_id]
|
||||||
else:
|
else:
|
||||||
pointer = sql.SQLReader if _id == 'read' else sql.SQLWriter
|
pointer = sql.SQLReader if _id == 'read' else sql.SQLWriter
|
||||||
|
|
||||||
return pointer(**args)
|
return pointer(**args)
|
||||||
|
|
||||||
return None
|
return None
|
|
@ -111,13 +111,47 @@ class DiskWriter(Writer):
|
||||||
pass
|
pass
|
||||||
finally:
|
finally:
|
||||||
DiskWriter.THREAD_LOCK.release()
|
DiskWriter.THREAD_LOCK.release()
|
||||||
class SQLiteReader (DiskReader):
|
class SQLite :
|
||||||
def __init__(self,**args):
|
def __init__(self,**_args) :
|
||||||
DiskReader.__init__(self,**args)
|
self.path = _args['database'] if 'database' in _args else _args['path']
|
||||||
self.path = args['database'] if 'database' in args else args['path']
|
self.conn = sqlite3.connect(self.path,isolation_level="IMMEDIATE")
|
||||||
self.conn = sqlite3.connect(self.path,isolation_level=None)
|
|
||||||
self.conn.row_factory = sqlite3.Row
|
self.conn.row_factory = sqlite3.Row
|
||||||
self.table = args['table']
|
self.fields = _args['fields'] if 'fields' in _args else []
|
||||||
|
def has (self,**_args):
|
||||||
|
found = False
|
||||||
|
try:
|
||||||
|
if 'table' in _args :
|
||||||
|
table = _args['table']
|
||||||
|
sql = "SELECT * FROM :table limit 1".replace(":table",table)
|
||||||
|
_df = pd.read_sql(sql,self.conn)
|
||||||
|
found = _df.columns.size > 0
|
||||||
|
except Exception as e:
|
||||||
|
pass
|
||||||
|
return found
|
||||||
|
def close(self):
|
||||||
|
try:
|
||||||
|
self.conn.close()
|
||||||
|
except Exception as e :
|
||||||
|
print(e)
|
||||||
|
def apply(self,sql):
|
||||||
|
try:
|
||||||
|
if not sql.lower().startswith('select'):
|
||||||
|
cursor = self.conn.cursor()
|
||||||
|
cursor.execute(sql)
|
||||||
|
cursor.close()
|
||||||
|
self.conn.commit()
|
||||||
|
else:
|
||||||
|
return pd.read_sql(sql,self.conn)
|
||||||
|
except Exception as e:
|
||||||
|
print (e)
|
||||||
|
class SQLiteReader (SQLite,DiskReader):
|
||||||
|
def __init__(self,**args):
|
||||||
|
super().__init__(**args)
|
||||||
|
# DiskReader.__init__(self,**args)
|
||||||
|
# self.path = args['database'] if 'database' in args else args['path']
|
||||||
|
# self.conn = sqlite3.connect(self.path,isolation_level=None)
|
||||||
|
# self.conn.row_factory = sqlite3.Row
|
||||||
|
self.table = args['table'] if 'table' in args else None
|
||||||
def read(self,**args):
|
def read(self,**args):
|
||||||
if 'sql' in args :
|
if 'sql' in args :
|
||||||
sql = args['sql']
|
sql = args['sql']
|
||||||
|
@ -135,7 +169,7 @@ class SQLiteReader (DiskReader):
|
||||||
except Exception as e :
|
except Exception as e :
|
||||||
pass
|
pass
|
||||||
|
|
||||||
class SQLiteWriter(DiskWriter) :
|
class SQLiteWriter(SQLite,DiskWriter) :
|
||||||
connection = None
|
connection = None
|
||||||
LOCK = Lock()
|
LOCK = Lock()
|
||||||
def __init__(self,**args):
|
def __init__(self,**args):
|
||||||
|
@ -143,12 +177,13 @@ class SQLiteWriter(DiskWriter) :
|
||||||
:path
|
:path
|
||||||
:fields json|csv
|
:fields json|csv
|
||||||
"""
|
"""
|
||||||
DiskWriter.__init__(self,**args)
|
# DiskWriter.__init__(self,**args)
|
||||||
self.table = args['table']
|
super().__init__(**args)
|
||||||
|
self.table = args['table'] if 'table' in args else None
|
||||||
|
|
||||||
self.conn = sqlite3.connect(self.path,isolation_level="IMMEDIATE")
|
# self.conn = sqlite3.connect(self.path,isolation_level="IMMEDIATE")
|
||||||
self.conn.row_factory = sqlite3.Row
|
# self.conn.row_factory = sqlite3.Row
|
||||||
self.fields = args['fields'] if 'fields' in args else []
|
# self.fields = args['fields'] if 'fields' in args else []
|
||||||
|
|
||||||
if self.fields and not self.isready():
|
if self.fields and not self.isready():
|
||||||
self.init(self.fields)
|
self.init(self.fields)
|
||||||
|
@ -185,7 +220,7 @@ class SQLiteWriter(DiskWriter) :
|
||||||
if not self.fields :
|
if not self.fields :
|
||||||
self.init(list(info.keys()))
|
self.init(list(info.keys()))
|
||||||
|
|
||||||
if type(info) == object :
|
if type(info) == dict :
|
||||||
info = [info]
|
info = [info]
|
||||||
elif type(info) == pd.DataFrame :
|
elif type(info) == pd.DataFrame :
|
||||||
info = info.to_dict(orient='records')
|
info = info.to_dict(orient='records')
|
||||||
|
@ -196,6 +231,8 @@ class SQLiteWriter(DiskWriter) :
|
||||||
cursor = self.conn.cursor()
|
cursor = self.conn.cursor()
|
||||||
sql = " " .join(["INSERT INTO ",self.table,"(", ",".join(self.fields) ,")", "values(:values)"])
|
sql = " " .join(["INSERT INTO ",self.table,"(", ",".join(self.fields) ,")", "values(:values)"])
|
||||||
for row in info :
|
for row in info :
|
||||||
|
print (row)
|
||||||
|
print (row.values())
|
||||||
stream =["".join(["",value,""]) if type(value) == str else value for value in row.values()]
|
stream =["".join(["",value,""]) if type(value) == str else value for value in row.values()]
|
||||||
stream = json.dumps(stream).replace("[","").replace("]","")
|
stream = json.dumps(stream).replace("[","").replace("]","")
|
||||||
|
|
||||||
|
|
|
@ -75,7 +75,15 @@ class SQLRW :
|
||||||
_info['securityLevel'] = 0
|
_info['securityLevel'] = 0
|
||||||
del _info['dbname']
|
del _info['dbname']
|
||||||
self.conn = _handler.connect(**_info)
|
self.conn = _handler.connect(**_info)
|
||||||
|
def has(self,**_args):
|
||||||
|
found = False
|
||||||
|
try:
|
||||||
|
table = _args['table']
|
||||||
|
sql = "SELECT * FROM :table LIMIT 1".replace(":table",table)
|
||||||
|
found = pd.read_sql(sql,self.conn).shape[0]
|
||||||
|
except Exception as e:
|
||||||
|
pass
|
||||||
|
return found
|
||||||
def isready(self):
|
def isready(self):
|
||||||
_sql = "SELECT * FROM :table LIMIT 1".replace(":table",self.table)
|
_sql = "SELECT * FROM :table LIMIT 1".replace(":table",self.table)
|
||||||
try:
|
try:
|
||||||
|
@ -201,8 +209,14 @@ class SQLWriter(SQLRW,Writer):
|
||||||
# values = [ "".join(["'",str(_row[key]),"'"]) if np.nan(_row[key]).isnumeric() else str(_row[key]) for key in _row]
|
# values = [ "".join(["'",str(_row[key]),"'"]) if np.nan(_row[key]).isnumeric() else str(_row[key]) for key in _row]
|
||||||
# print (values)
|
# print (values)
|
||||||
query = _sql.replace(":fields",",".join(fields)).replace(":values",values)
|
query = _sql.replace(":fields",",".join(fields)).replace(":values",values)
|
||||||
|
if type(info) == pd.DataFrame :
|
||||||
cursor.execute(query,_row.values())
|
_values = info.values.tolist()
|
||||||
|
elif type(info) == list and type(info[0]) == dict:
|
||||||
|
print ('........')
|
||||||
|
_values = [tuple(item.values()) for item in info]
|
||||||
|
else:
|
||||||
|
_values = info;
|
||||||
|
cursor.execute(query,_values)
|
||||||
|
|
||||||
|
|
||||||
pass
|
pass
|
||||||
|
@ -210,14 +224,23 @@ class SQLWriter(SQLRW,Writer):
|
||||||
_fields = ",".join(self.fields)
|
_fields = ",".join(self.fields)
|
||||||
# _sql = _sql.replace(":fields",_fields)
|
# _sql = _sql.replace(":fields",_fields)
|
||||||
# _sql = _sql.replace(":values",",".join(["%("+name+")s" for name in self.fields]))
|
# _sql = _sql.replace(":values",",".join(["%("+name+")s" for name in self.fields]))
|
||||||
_sql = _sql.replace("(:fields)","")
|
# _sql = _sql.replace("(:fields)","")
|
||||||
|
_sql = _sql.replace(":fields",_fields)
|
||||||
values = ", ".join("?"*len(self.fields)) if self._provider == 'netezza' else ",".join(["%s" for name in self.fields])
|
values = ", ".join("?"*len(self.fields)) if self._provider == 'netezza' else ",".join(["%s" for name in self.fields])
|
||||||
_sql = _sql.replace(":values",values)
|
_sql = _sql.replace(":values",values)
|
||||||
|
if type(info) == pd.DataFrame :
|
||||||
|
_info = info[self.fields].values.tolist()
|
||||||
|
elif type(info) == dict :
|
||||||
|
_info = info.values()
|
||||||
|
else:
|
||||||
|
# _info = []
|
||||||
|
|
||||||
# for row in info :
|
_info = pd.DataFrame(info)[self.fields].values.tolist()
|
||||||
# values = ["'".join(["",value,""]) if not str(value).isnumeric() else value for value in row.values()]
|
# for row in info :
|
||||||
|
|
||||||
cursor.executemany(_sql,info)
|
# if type(row) == dict :
|
||||||
|
# _info.append( list(row.values()))
|
||||||
|
cursor.executemany(_sql,_info)
|
||||||
|
|
||||||
# self.conn.commit()
|
# self.conn.commit()
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
|
@ -250,6 +273,13 @@ class BigQuery:
|
||||||
client = bq.Client.from_service_account_json(self.path)
|
client = bq.Client.from_service_account_json(self.path)
|
||||||
ref = client.dataset(self.dataset).table(table)
|
ref = client.dataset(self.dataset).table(table)
|
||||||
return client.get_table(ref).schema
|
return client.get_table(ref).schema
|
||||||
|
def has(self,**_args):
|
||||||
|
found = False
|
||||||
|
try:
|
||||||
|
found = self.meta(**_args) is not None
|
||||||
|
except Exception as e:
|
||||||
|
pass
|
||||||
|
return found
|
||||||
class BQReader(BigQuery,Reader) :
|
class BQReader(BigQuery,Reader) :
|
||||||
def __init__(self,**_args):
|
def __init__(self,**_args):
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue