]> AND Private Git Repository - predictops.git/blobdiff - predictops/learn/preprocessing.py
Logo AND Algorithmique Numérique Distribuée

Private GIT Repository
lightgbm is now working
[predictops.git] / predictops / learn / preprocessing.py
index 885aad3393979b897e3e0d8c40f3378dbba08e5a..9bc09ad2eca2759c22b6047c3ded8ab747e015de 100644 (file)
@@ -14,6 +14,7 @@ import pandas as pd
 fileConfig((Path.cwd() / 'config') / 'logging.cfg')
 logger = getLogger()
 
 fileConfig((Path.cwd() / 'config') / 'logging.cfg')
 logger = getLogger()
 
+
 class Preprocessing:
     '''
     Generate a pandas dataframe from a dictionary of features per datetime, which
 class Preprocessing:
     '''
     Generate a pandas dataframe from a dictionary of features per datetime, which
@@ -25,8 +26,8 @@ class Preprocessing:
      - NaN values are then filled with last known values.
     '''
 
      - NaN values are then filled with last known values.
     '''
 
-    def __init__(self, config_file = None,
-                 dict_features = None, dict_target = None):
+    def __init__(self, config_file=None,
+                 dict_features=None, dict_target=None):
         '''
         Constructor that defines all needed attributes and collects features.
         '''
         '''
         Constructor that defines all needed attributes and collects features.
         '''
@@ -35,9 +36,8 @@ class Preprocessing:
         self._start = datetime.strptime(self._config['DATETIME']['start'],
                                         '%m/%d/%Y %H:%M:%S')
         self._end = datetime.strptime(self._config['DATETIME']['end'],
         self._start = datetime.strptime(self._config['DATETIME']['start'],
                                         '%m/%d/%Y %H:%M:%S')
         self._end = datetime.strptime(self._config['DATETIME']['end'],
-                                        '%m/%d/%Y %H:%M:%S')
-        self._timestep = timedelta(hours =
-                                   self._config['DATETIME'].getfloat('hourStep'))
+                                      '%m/%d/%Y %H:%M:%S')
+        self._timestep = timedelta(hours=self._config['DATETIME'].getfloat('hourStep'))
         self._dict_features = dict_features
         self._dict_target = dict_target
 
         self._dict_features = dict_features
         self._dict_target = dict_target
 
@@ -46,26 +46,28 @@ class Preprocessing:
         self._datetimes = []
 
         self._features = set(chain.from_iterable([tuple(u.keys())
         self._datetimes = []
 
         self._features = set(chain.from_iterable([tuple(u.keys())
-                                                      for u in [*dict_features.values()]]))
+                                                  for u in [*dict_features.values()]]))
 
         #feature_files = Path.cwd() / 'config' / 'features'
 
         #feature_files = Path.cwd() / 'config' / 'features'
-        self._features = {feat : {'numerical': False, 'categorical': False}
+        self._features = {feat: {'numerical': False, 'categorical': False}
                           for feat in self._features}
 
         for feature in self._config['FEATURES']:
             if self._config['FEATURES'][feature]:
                 feature_file = self._config['FEATURE_CONFIG'][feature]
                 config = ConfigParser()
                           for feat in self._features}
 
         for feature in self._config['FEATURES']:
             if self._config['FEATURES'][feature]:
                 feature_file = self._config['FEATURE_CONFIG'][feature]
                 config = ConfigParser()
-                config.read(feature_file)
+                config.read(eval(feature_file))
                 for section in config:
                     if config.has_option(section, 'numerical'):
                 for section in config:
                     if config.has_option(section, 'numerical'):
-                        self._features[section]['numerical'] = config[section].getboolean('numerical')
-                        self._features[section]['categorical'] = config[section].getboolean('categorical')
+                        for feature in self._features:
+                            if feature.split('_')[0] == section:
+                                self._features[feature]['binary'] = config[section].getboolean('binary')
+                                self._features[feature]['categorical'] = config[section].getboolean('categorical')
+                                self._features[feature]['numerical'] = config[section].getboolean('numerical')
 
 
-        self._numerical_columns = [k for k in self._features if self._features[k]['numerical']]
+        self._binary_columns = [k for k in self._features if self._features[k]['binary']]
         self._categorical_columns = [k for k in self._features if self._features[k]['categorical']]
         self._categorical_columns = [k for k in self._features if self._features[k]['categorical']]
-
-
+        self._numerical_columns = [k for k in self._features if self._features[k]['numerical']]
 
     @property
     def start(self):
 
     @property
     def start(self):
@@ -75,7 +77,6 @@ class Preprocessing:
     def start(self, x):
         self._start = x
 
     def start(self, x):
         self._start = x
 
-
     @property
     def end(self):
         return self._end
     @property
     def end(self):
         return self._end
@@ -84,7 +85,6 @@ class Preprocessing:
     def end(self, x):
         self._end = x
 
     def end(self, x):
         self._end = x
 
-
     @property
     def timestep(self):
         return self._timestep
     @property
     def timestep(self):
         return self._timestep
@@ -93,7 +93,6 @@ class Preprocessing:
     def timestep(self, x):
         self._timestep = x
 
     def timestep(self, x):
         self._timestep = x
 
-
     def _fill_dict(self):
         '''
         Add datetime keys in the dated feature dictionary that are missing. The
     def _fill_dict(self):
         '''
         Add datetime keys in the dated feature dictionary that are missing. The
@@ -105,16 +104,16 @@ class Preprocessing:
         while current <= self._end:
             self._datetimes.append(current)
             if current not in self._dict_features:
         while current <= self._end:
             self._datetimes.append(current)
             if current not in self._dict_features:
-                self._dict_features[current] = {feature:np.NaN
+                self._dict_features[current] = {feature: np.NaN
                                                 for feature in self._features}
             else:
                                                 for feature in self._features}
             else:
-                null_dict = {feature:np.NaN
+                null_dict = {feature: np.NaN
                              for feature in self._features}
                 null_dict.update(self._dict_features[current])
                 self._dict_features[current] = null_dict
             current += self._timestep
         for k in self._dict_features:
                              for feature in self._features}
                 null_dict.update(self._dict_features[current])
                 self._dict_features[current] = null_dict
             current += self._timestep
         for k in self._dict_features:
-            null_dict = {feature:np.NaN
+            null_dict = {feature: np.NaN
                          for feature in self._features}
             null_dict.update(self._dict_features[k])
             self._dict_features[k] = null_dict
                          for feature in self._features}
             null_dict.update(self._dict_features[k])
             self._dict_features[k] = null_dict
@@ -122,8 +121,6 @@ class Preprocessing:
         self._full_dict = {k: self._dict_features[k]
                            for k in sorted(self._dict_features.keys())}
 
         self._full_dict = {k: self._dict_features[k]
                            for k in sorted(self._dict_features.keys())}
 
-
-
     @property
     def full_dict(self):
         '''
     @property
     def full_dict(self):
         '''
@@ -133,7 +130,6 @@ class Preprocessing:
             self._fill_dict()
         return self._full_dict
 
             self._fill_dict()
         return self._full_dict
 
-
     def _fill_nan(self):
         '''
         Fill NaN values, either by propagation or by interpolation (linear or splines)
     def _fill_nan(self):
         '''
         Fill NaN values, either by propagation or by interpolation (linear or splines)
@@ -150,7 +146,7 @@ class Preprocessing:
         elif self._config['PREPROCESSING']['fill_method'] == 'spline':
             self._dataframe[self._numerical_columns] =\
                 self._dataframe[self._numerical_columns].interpolate(method='spline',
         elif self._config['PREPROCESSING']['fill_method'] == 'spline':
             self._dataframe[self._numerical_columns] =\
                 self._dataframe[self._numerical_columns].interpolate(method='spline',
-                     order=self._config['PREPROCESSING'].getint('order'))
+                                                                     order=self._config['PREPROCESSING'].getint('order'))
 
         # For the categorical columns, NaN values are filled by duplicating
         # the last known value (forward fill method)
 
         # For the categorical columns, NaN values are filled by duplicating
         # the last known value (forward fill method)
@@ -173,21 +169,18 @@ class Preprocessing:
         self._dataframe = self._dataframe.drop(['row_ok'], axis=1)
         logger.info("Rows dropped")
 
         self._dataframe = self._dataframe.drop(['row_ok'], axis=1)
         logger.info("Rows dropped")
 
-
     def _add_history(self):
         '''
         Integrating previous nb of interventions as features
         '''
         logger.info("Integrating previous nb of interventions as features")
     def _add_history(self):
         '''
         Integrating previous nb of interventions as features
         '''
         logger.info("Integrating previous nb of interventions as features")
-        nb_lines = self._config['HISTORY_KNOWLEDGE'].getint('nb_lines')
-        for k in range(1,nb_lines+1):
-            name = 'history_'+str(nb_lines-k+1)
-            self._dataframe[name] = [np.NaN]*k + list(self._dict_target.values())[:-k]
+        nb_lines = eval(self._config['HISTORY_KNOWLEDGE']['nb_lines'])
+        for k in range(1, nb_lines + 1):
+            name = 'history_' + str(nb_lines - k + 1)
+            self._dataframe[name] = [np.NaN] * k + list(self._dict_target.values())[:-k]
             self._numerical_columns.append(name)
         self._dataframe = self._dataframe[nb_lines:]
 
             self._numerical_columns.append(name)
         self._dataframe = self._dataframe[nb_lines:]
 
-
-
     def _standardize(self):
         '''
         Normalizing numerical features
     def _standardize(self):
         '''
         Normalizing numerical features
@@ -197,26 +190,25 @@ class Preprocessing:
         self._dataframe[self._numerical_columns] =\
             preprocessing.scale(self._dataframe[self._numerical_columns])
 
         self._dataframe[self._numerical_columns] =\
             preprocessing.scale(self._dataframe[self._numerical_columns])
 
-
-
     def _one_hot_encoding(self):
         '''
         Apply a one hot encoding for category features
         '''
         logger.info("One hot encoding for categorical feature")
     def _one_hot_encoding(self):
         '''
         Apply a one hot encoding for category features
         '''
         logger.info("One hot encoding for categorical feature")
-
         # We store numerical columns
         df_out = pd.DataFrame()
         # We store numerical columns
         df_out = pd.DataFrame()
-        for col in  self._numerical_columns:
+        for col in self._numerical_columns:
+            df_out[col] = self._dataframe[col]
+        # Idem for binary features
+        for col in self._binary_columns:
             df_out[col] = self._dataframe[col]
         # The one hot encoding
         for col in self._categorical_columns:
             df_out[col] = self._dataframe[col]
         # The one hot encoding
         for col in self._categorical_columns:
-            pd1 = pd.get_dummies(self._dataframe[col],prefix=col)
+            pd1 = pd.get_dummies(self._dataframe[col], prefix=col)
             for col1 in pd1.columns:
                 df_out[col1] = pd1[col1]
         self._dataframe = df_out
 
             for col1 in pd1.columns:
                 df_out[col1] = pd1[col1]
         self._dataframe = df_out
 
-
     @property
     def dataframe(self):
         '''
     @property
     def dataframe(self):
         '''
@@ -236,7 +228,6 @@ class Preprocessing:
             self._one_hot_encoding()
         return self._dataframe
 
             self._one_hot_encoding()
         return self._dataframe
 
-
     @dataframe.setter
     def dataframe(self, df):
         self._dataframe = df
     @dataframe.setter
     def dataframe(self, df):
         self._dataframe = df