bug fix around shape of candidate data to generate
This commit is contained in:
parent
c1a500fe4c
commit
553ee75a06
49
data/gan.py
49
data/gan.py
|
@ -166,7 +166,15 @@ class GNet :
|
|||
return _object
|
||||
def mkdir (self,path):
|
||||
if not os.path.exists(path) :
|
||||
os.mkdir(path)
|
||||
if os.sep in path :
|
||||
pass
|
||||
root = []
|
||||
for loc in path.split(os.sep) :
|
||||
root.append(loc)
|
||||
os.mkdir(os.sep.join(root))
|
||||
|
||||
else:
|
||||
os.mkdir(path)
|
||||
|
||||
|
||||
def normalize(self,**args):
|
||||
|
@ -520,8 +528,10 @@ class Predict(GNet):
|
|||
"""
|
||||
def __init__(self,**args):
|
||||
GNet.__init__(self,**args)
|
||||
self.generator = Generator(**args)
|
||||
self.values = args['values']
|
||||
self.generator = Generator(**args)
|
||||
self.values = args['values']
|
||||
self.ROW_COUNT = args['row_count']
|
||||
self.MISSING_VALUES = args['no_value']
|
||||
def load_meta(self, column):
|
||||
super().load_meta(column)
|
||||
self.generator.load_meta(column)
|
||||
|
@ -532,8 +542,8 @@ class Predict(GNet):
|
|||
model_dir = os.sep.join([self.train_dir,suffix+'-'+str(self.MAX_EPOCHS)])
|
||||
demo = self._LABEL #np.zeros([self.ROW_COUNT,self.NUM_LABELS]) #args['de"shape":{"LABEL":list(self._LABEL.shape)} mo']
|
||||
tf.compat.v1.reset_default_graph()
|
||||
z = tf.random.normal(shape=[self.BATCHSIZE_PER_GPU, self.Z_DIM])
|
||||
y = tf.compat.v1.placeholder(shape=[self.BATCHSIZE_PER_GPU, self.NUM_LABELS], dtype=tf.int32)
|
||||
z = tf.random.normal(shape=[self.ROW_COUNT, self.Z_DIM])
|
||||
y = tf.compat.v1.placeholder(shape=[self.ROW_COUNT, self.NUM_LABELS], dtype=tf.int32)
|
||||
if self._LABEL is not None :
|
||||
ma = [[i] for i in np.arange(self.NUM_LABELS - 2)]
|
||||
label = y[:, 1] * len(ma) + tf.squeeze(tf.matmul(y[:, 2:], tf.constant(ma, dtype=tf.int32)))
|
||||
|
@ -556,7 +566,7 @@ class Predict(GNet):
|
|||
labels = None
|
||||
|
||||
found = []
|
||||
|
||||
ratio = []
|
||||
for i in np.arange(CANDIDATE_COUNT) :
|
||||
if labels :
|
||||
f = sess.run(fake,feed_dict={y:labels})
|
||||
|
@ -569,10 +579,11 @@ class Predict(GNet):
|
|||
df = ( pd.DataFrame(np.round(f).astype(np.int32)))
|
||||
p = 0 not in df.sum(axis=1).values
|
||||
x = df.sum(axis=1).values
|
||||
print ( [np.sum(x),x.size])
|
||||
if np.divide( np.sum(x), x.size) :
|
||||
|
||||
if np.divide( np.sum(x), x.size) > .9 or p:
|
||||
ratio.append(np.divide( np.sum(x), x.size))
|
||||
found.append(df)
|
||||
if len(found) == NTH_VALID_CANDIDATE or i == CANDIDATE_COUNT:
|
||||
if i == CANDIDATE_COUNT:
|
||||
break
|
||||
else:
|
||||
continue
|
||||
|
@ -582,8 +593,9 @@ class Predict(GNet):
|
|||
#
|
||||
# In case we are dealing with actual values like diagnosis codes we can perform
|
||||
#
|
||||
|
||||
df = found[np.random.choice(np.arange(len(found)),1)[0]]
|
||||
INDEX = np.random.choice(np.arange(len(found)),1)[0]
|
||||
INDEX = ratio.index(np.max(ratio))
|
||||
df = found[INDEX]
|
||||
columns = self.ATTRIBUTES['synthetic'] if isinstance(self.ATTRIBUTES['synthetic'],list)else [self.ATTRIBUTES['synthetic']]
|
||||
|
||||
# r = np.zeros((self.ROW_COUNT,len(columns)))
|
||||
|
@ -592,9 +604,20 @@ class Predict(GNet):
|
|||
if len(found):
|
||||
print (len(found),NTH_VALID_CANDIDATE)
|
||||
# x = df * self.values
|
||||
|
||||
df = pd.DataFrame( df.apply(lambda row: self.values[np.random.choice(np.where(row != 0)[0],1)[0]] ,axis=1))
|
||||
#
|
||||
# let's get the missing rows (if any) ...
|
||||
#
|
||||
ii = df.apply(lambda row: np.sum(row) == 0 ,axis=1)
|
||||
if ii :
|
||||
#
|
||||
#@TODO Have this be a configurable variable
|
||||
missing = np.repeat(0, np.where(ii==1)[0].size)
|
||||
else:
|
||||
missing = []
|
||||
i = np.where(ii == 0)[0]
|
||||
df = pd.DataFrame( df.iloc.apply(lambda row: self.values[np.random.choice(np.where(row != 0)[0],1)[0]] ,axis=1))
|
||||
df.columns = columns
|
||||
df = df[columns[0]].append(pd.Series(missing))
|
||||
|
||||
|
||||
|
||||
|
|
|
@ -77,25 +77,23 @@ def generate(**args):
|
|||
df = args['data'] if not isinstance(args['data'],str) else pd.read_csv(args['data'])
|
||||
|
||||
column = args['column'] if (isinstance(args['column'],list)) else [args['column']]
|
||||
column_id = args['id']
|
||||
# column_id = args['id']
|
||||
#
|
||||
#@TODO:
|
||||
# If the identifier is not present, we should fine a way to determine or make one
|
||||
#
|
||||
# args['label'] = pd.get_dummies(df[column_id]).astype(np.float32).values
|
||||
bwrangler = Binary()
|
||||
# args['label'] = bwrangler.Export(df[[column_id]])
|
||||
_df = df.copy()
|
||||
for col in column :
|
||||
args['context'] = col
|
||||
args['column'] = col
|
||||
values = df[col].unique().tolist()
|
||||
# values.sort()
|
||||
args['values'] = values
|
||||
args['row_count'] = df.shape[0]
|
||||
#
|
||||
# we can determine the cardinalities here so we know what to allow or disallow
|
||||
handler = gan.Predict (**args)
|
||||
handler.load_meta(col)
|
||||
# handler.ROW_COUNT = df[col].shape[0]
|
||||
r = handler.apply()
|
||||
# print (r)
|
||||
_df[col] = r[col]
|
||||
|
|
Loading…
Reference in New Issue