Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fixed module import issues and clean code #3

Open
wants to merge 2 commits into
base: master
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
13 changes: 2 additions & 11 deletions tl_algs/burak.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
import numpy as np
import pandas as pd
import json
from tl_algs import tl_alg
from . import tl_alg
from sklearn.cluster import KMeans
from sklearn.metrics.pairwise import euclidean_distances

Expand All @@ -27,7 +27,6 @@ def _kmeans_cluster(test_set_X, train_pool_X, train_pool_y, cluster_factor,
entry is the ith cluster of training and test instances.
"""

# master_X_df = train_pool_X.append(test_set_X)
master_X_df = pd.concat([train_pool_X, test_set_X])
num_clust = master_X_df.shape[0] // cluster_factor

Expand All @@ -46,8 +45,6 @@ def _kmeans_cluster(test_set_X, train_pool_X, train_pool_y, cluster_factor,
for i, clust in enumerate(X_test_clusters):
x_pool = pd.DataFrame(test_set_X.iloc[i, ]).transpose()
clusters[clust]['X_test'] = pd.concat([clusters[clust]['X_test'], x_pool])
# clusters[clust]['X_test'] = clusters[clust]['X_test']. \
# append(test_set_X.iloc[i, ])

# Populate clusters based on training data.
X_train_clusters = cluster_model.predict(train_pool_X)
Expand All @@ -56,10 +53,7 @@ def _kmeans_cluster(test_set_X, train_pool_X, train_pool_y, cluster_factor,
y_pool = pd.Series([train_pool_y.iloc[i]])
clusters[clust]['X_train'] = pd.concat([clusters[clust]['X_train'], x_pool])
clusters[clust]['y_train'] = pd.concat([clusters[clust]['y_train'], y_pool])
# clusters[clust]['X_train'] = clusters[clust]['X_train']. \
# append(train_pool_X.iloc[i, ])
# clusters[clust]['y_train'] = clusters[clust]['y_train'] \
# .append(pd.Series([train_pool_y.iloc[i]]))

# Remove clusters with no test instance.
to_remove = [
i for (i, d) in enumerate(clusters)
Expand Down Expand Up @@ -165,7 +159,6 @@ def burak_filter(self, test_set_X, test_set_domain, train_pool_X,
of which gives the confidence for the ith prediction.
predictions: List of class predictions.
"""
print("=====", len(train_pool_X), len(train_pool_y))
X_filtered, y_filtered = self.filter_instances(
train_pool_X,
train_pool_y,
Expand Down Expand Up @@ -221,8 +214,6 @@ class documentation for more information on the form of this method's
)
X_train_filtered = pd.concat([X_train_filtered, more_X_train])
y_train_filtered = pd.concat([y_train_filtered, more_y_train])
# X_train_filtered = X_train_filtered.append(more_X_train)
# y_train_filtered = y_train_filtered.append(more_y_train)

classifier = Base_Classifier(
random_state=rand_seed,
Expand Down
6 changes: 1 addition & 5 deletions tl_algs/ctl.py
Original file line number Diff line number Diff line change
@@ -1,9 +1,5 @@
import numpy as np
import pandas as pd
import json
import tl_alg
from . import tl_alg
from sklearn.cluster import KMeans
from sklearn.metrics.pairwise import euclidean_distances

class ClusterThenLabel(tl_alg.Base_Transfer):
"""
Expand Down
2 changes: 1 addition & 1 deletion tl_algs/peters.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
import numpy as np
import pandas as pd
from sklearn.metrics.pairwise import euclidean_distances
from tl_algs import tl_alg, burak
from . import tl_alg, burak


class Peters(tl_alg.Base_Transfer):
Expand Down
2 changes: 1 addition & 1 deletion tl_algs/tca.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@
import pandas as pd
import json
import da_tool.tca
from tl_algs import tl_alg
from . import tl_alg
from sklearn.cluster import KMeans
from sklearn.metrics.pairwise import euclidean_distances

Expand Down
2 changes: 1 addition & 1 deletion tl_algs/tca_plus.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
import numpy as np
import pandas as pd
import json
from tl_algs import tl_alg
from . import tl_alg
import da_tool.tca
from sklearn.metrics.pairwise import euclidean_distances
from scipy.stats import zscore
Expand Down
2 changes: 1 addition & 1 deletion tl_algs/tl_baseline.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
import numpy as np
import pandas as pd
import json
from tl_algs import tl_alg
from . import tl_alg


class Target_Baseline(tl_alg.Base_Transfer):
Expand Down
2 changes: 1 addition & 1 deletion tl_algs/tnb.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
import numpy as np
import pandas as pd
import json
from tl_algs import tl_alg
from . import tl_alg


def sim_minmax(column):
Expand Down
4 changes: 1 addition & 3 deletions tl_algs/trbag.py
Original file line number Diff line number Diff line change
@@ -1,9 +1,7 @@
import numpy as np
import pandas as pd
import json
from tl_algs import tl_alg
import voter
# from vuln_toolkit.common import vuln_metrics
from . import tl_alg, voter
from sklearn.dummy import DummyClassifier
from sklearn.metrics import f1_score

Expand Down