Skip to content

Commit bcd55d8

Browse files
committed
Files updated
1 parent fef9442 commit bcd55d8

32 files changed

+10
-35
lines changed
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.

scripts/Activity - 3.01 - NEWS Classification.py renamed to scripts/Activity - 01 - NEWS Classification.py

+1-2
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,6 @@
33
import pandas as pd
44
import seaborn as sns
55
import matplotlib.pyplot as plt
6-
get_ipython().run_line_magic('matplotlib', 'inline')
76
from nltk import word_tokenize
87
from nltk.corpus import stopwords
98
from nltk.stem import WordNetLemmatizer
@@ -18,7 +17,7 @@
1817

1918

2019

21-
data = pd.read_csv('data/news_political_dataset.csv')
20+
data = pd.read_csv('../data/news_political_dataset.csv')
2221
assert data.shape == tuple([69505, 3])
2322
print(data.sample(5))
2423

scripts/Exercise - 3.01 - Clustering.py renamed to scripts/Exercise - 01 - Clustering.py

-1
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,6 @@
99
import pandas as pd
1010
import numpy as np
1111
import matplotlib.pyplot as plt
12-
get_ipython().run_line_magic('matplotlib', 'inline')
1312
import re
1413
import string
1514
from nltk import word_tokenize

scripts/Exercise - 3.02 - KMeans.py renamed to scripts/Exercise - 02 - KMeans.py

-1
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,6 @@
33
import pandas as pd
44
from sklearn.datasets import fetch_20newsgroups
55
import matplotlib.pyplot as plt
6-
get_ipython().run_line_magic('matplotlib', 'inline')
76
import re
87
import string
98
from nltk import word_tokenize

scripts/Exercise - 3.03 - Logistic Regression.py renamed to scripts/Exercise - 03 - Logistic Regression.py

+1-2
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,6 @@
33

44
import pandas as pd
55
import matplotlib.pyplot as plt
6-
get_ipython().run_line_magic('matplotlib', 'inline')
76
import re
87
import string
98
from nltk import word_tokenize
@@ -17,7 +16,7 @@
1716

1817

1918

20-
review_data = pd.read_json('data/reviews_Musical_Instruments_5.json', lines=True)
19+
review_data = pd.read_json('../data/reviews_Musical_Instruments_5.json', lines=True)
2120
print(review_data[['reviewText', 'overall']].head())
2221

2322

scripts/Exercise - 3.04 - Naive Bayes.py renamed to scripts/Exercise - 04 - Naive Bayes.py

+1-5
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,7 @@
11
# coding: utf-8
22

3-
# In[1]:
4-
5-
63
import pandas as pd
74
import matplotlib.pyplot as plt
8-
get_ipython().run_line_magic('matplotlib', 'inline')
95
import re
106
import string
117
from nltk import word_tokenize
@@ -18,7 +14,7 @@
1814
warnings.filterwarnings('ignore')
1915

2016

21-
review_data = pd.read_json('data/reviews_Musical_Instruments_5.json', lines=True)
17+
review_data = pd.read_json('../data/reviews_Musical_Instruments_5.json', lines=True)
2218
print(review_data[['reviewText', 'overall']].head())
2319

2420
assert review_data.shape == tuple([10261, 9])

scripts/Exercise - 3.05 - KNN.py renamed to scripts/Exercise - 05 - KNN.py

+1-5
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,7 @@
11
# coding: utf-8
22

3-
# In[1]:
4-
5-
63
import pandas as pd
74
import matplotlib.pyplot as plt
8-
get_ipython().run_line_magic('matplotlib', 'inline')
95
import re
106
import string
117
from nltk import word_tokenize
@@ -21,7 +17,7 @@
2117
# In[2]:
2218

2319

24-
review_data = pd.read_json('data/reviews_Musical_Instruments_5.json', lines=True)
20+
review_data = pd.read_json('../data/reviews_Musical_Instruments_5.json', lines=True)
2521
print(review_data[['reviewText', 'overall']].head())
2622

2723

scripts/Exercise - 3.06 - Linear Regression.py renamed to scripts/Exercise - 06 - Linear Regression.py

+1-2
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,6 @@
55

66
import pandas as pd
77
import matplotlib.pyplot as plt
8-
get_ipython().run_line_magic('matplotlib', 'inline')
98
import re
109
import string
1110
from nltk import word_tokenize
@@ -21,7 +20,7 @@
2120
# In[7]:
2221

2322

24-
review_data = pd.read_json('data/reviews_Musical_Instruments_5.json', lines=True)
23+
review_data = pd.read_json('../data/reviews_Musical_Instruments_5.json', lines=True)
2524
print(review_data[['reviewText', 'overall']].head())
2625

2726

scripts/Exercise - 3.07 - Decision Tree.py renamed to scripts/Exercise - 07 - Decision Tree.py

+1-2
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,6 @@
55

66
import pandas as pd
77
import matplotlib.pyplot as plt
8-
get_ipython().run_line_magic('matplotlib', 'inline')
98
import re
109
import string
1110
from nltk import word_tokenize
@@ -21,7 +20,7 @@
2120
# In[2]:
2221

2322

24-
data_patio_lawn_garden = pd.read_json('data/reviews_Patio_Lawn_and_Garden_5.json', lines = True)
23+
data_patio_lawn_garden = pd.read_json('../data/reviews_Patio_Lawn_and_Garden_5.json', lines = True)
2524
print(data_patio_lawn_garden[['reviewText', 'overall']].head())
2625

2726

scripts/Exercise - 3.08 - Random Forest.py renamed to scripts/Exercise - 08 - Random Forest.py

+1-2
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,6 @@
55

66
import pandas as pd
77
import matplotlib.pyplot as plt
8-
get_ipython().run_line_magic('matplotlib', 'inline')
98
import re
109
import string
1110
from nltk import word_tokenize
@@ -21,7 +20,7 @@
2120
# In[3]:
2221

2322

24-
data_patio_lawn_garden = pd.read_json('data/reviews_Patio_Lawn_and_Garden_5.json', lines = True)
23+
data_patio_lawn_garden = pd.read_json('../data/reviews_Patio_Lawn_and_Garden_5.json', lines = True)
2524
assert data_patio_lawn_garden.shape == tuple([13272, 9])
2625
data_patio_lawn_garden[['reviewText', 'overall']].head()
2726

scripts/Exercise - 3.09 - XGBoost.py renamed to scripts/Exercise - 09 - XGBoost.py

+1-2
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,6 @@
55

66
import pandas as pd
77
import matplotlib.pyplot as plt
8-
get_ipython().run_line_magic('matplotlib', 'inline')
98
import re
109
import string
1110
from nltk import word_tokenize
@@ -21,7 +20,7 @@
2120
# In[2]:
2221

2322

24-
data_patio_lawn_garden = pd.read_json('data/reviews_Patio_Lawn_and_Garden_5.json', lines = True)
23+
data_patio_lawn_garden = pd.read_json('../data/reviews_Patio_Lawn_and_Garden_5.json', lines = True)
2524
assert data_patio_lawn_garden.shape == tuple([13272, 9])
2625
print(data_patio_lawn_garden[['reviewText', 'overall']].head())
2726

scripts/Exercise - 3.10 - Sampling.py renamed to scripts/Exercise - 10 - Sampling.py

+2-9
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44

55

66
import pandas as pd
7-
data = pd.read_excel('data/Online Retail.xlsx')
7+
data = pd.read_excel('../data/Online Retail.xlsx')
88
assert data.shape == tuple([541909, 8])
99
print(data.shape)
1010

@@ -56,11 +56,4 @@
5656
# In[48]:
5757

5858

59-
print(data_ugf_q2_sample.shape)
60-
61-
62-
# In[ ]:
63-
64-
65-
66-
59+
print(data_ugf_q2_sample.shape)

scripts/Exercise - 3.11 - Correlation.py renamed to scripts/Exercise - 11 - Correlation.py

-1
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,6 @@
55
import pandas as pd
66
import numpy as np
77
import matplotlib.pyplot as plt
8-
get_ipython().run_line_magic('matplotlib', 'inline')
98
import re
109
import string
1110
from nltk import word_tokenize

scripts/Exercise - 3.12 - PCA.py renamed to scripts/Exercise - 12 - PCA.py

-1
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,6 @@
88
import pandas as pd
99
import numpy as np
1010
import matplotlib.pyplot as plt
11-
get_ipython().run_line_magic('matplotlib', 'inline')
1211
import re
1312
import string
1413
from nltk import word_tokenize

0 commit comments

Comments
 (0)