rr commited on
Commit
48cf5e0
·
1 Parent(s): d14b206

Delete jithu.txt

Browse files
Files changed (1) hide show
  1. jithu.txt +0 -984
jithu.txt DELETED
@@ -1,984 +0,0 @@
1
- {
2
- "cells": [
3
- {
4
- "cell_type": "markdown",
5
- "metadata": {},
6
- "source": [
7
- "## K Means Clustering w elbow plot"
8
- ]
9
- },
10
- {
11
- "cell_type": "code",
12
- "execution_count": null,
13
- "metadata": {},
14
- "outputs": [],
15
- "source": [
16
- "#importing libraries\n",
17
- "from sklearn.cluster import KMeans\n",
18
- "import pandas as pd\n",
19
- "from sklearn.preprocessing import MinMaxScaler\n",
20
- "import matplotlib.pyplot as plt"
21
- ]
22
- },
23
- {
24
- "cell_type": "code",
25
- "execution_count": null,
26
- "metadata": {},
27
- "outputs": [],
28
- "source": [
29
- "#import dataset\n",
30
- "df = pd.read_csv(\"C:/Users/Balakrishnan/Desktop/6th sem/Data Mining/Datasets/income.csv\")\n",
31
- "df.head()"
32
- ]
33
- },
34
- {
35
- "cell_type": "code",
36
- "execution_count": null,
37
- "metadata": {},
38
- "outputs": [],
39
- "source": [
40
- "plt.scatter(df.Age, df.Income)\n",
41
- "plt.xlabel('Age')\n",
42
- "plt.ylabel('Income')"
43
- ]
44
- },
45
- {
46
- "cell_type": "code",
47
- "execution_count": null,
48
- "metadata": {},
49
- "outputs": [],
50
- "source": [
51
- "scaler = MinMaxScaler()\n",
52
- "scaler.fit(df[['Income']])\n",
53
- "df['Income'] = scaler.transform(df[['Income']])\n",
54
- "\n",
55
- "scaler.fit(df[['Age']])\n",
56
- "df['Age'] = scaler.transform(df[['Age']])"
57
- ]
58
- },
59
- {
60
- "cell_type": "code",
61
- "execution_count": null,
62
- "metadata": {},
63
- "outputs": [],
64
- "source": [
65
- "#Elbow plot\n",
66
- "sse = []\n",
67
- "k_rng = range(1, 10)\n",
68
- "for k in k_rng:\n",
69
- " km = KMeans(n_clusters = k)\n",
70
- " km.fit(df[['Age', 'Income']])\n",
71
- " sse.append(km.inertia_)"
72
- ]
73
- },
74
- {
75
- "cell_type": "code",
76
- "execution_count": null,
77
- "metadata": {},
78
- "outputs": [],
79
- "source": [
80
- "plt.xlabel('K')\n",
81
- "plt.ylabel('Sum of squared error')\n",
82
- "plt.plot(k_rng,sse)"
83
- ]
84
- },
85
- {
86
- "cell_type": "code",
87
- "execution_count": null,
88
- "metadata": {},
89
- "outputs": [],
90
- "source": [
91
- "km = KMeans(n_clusters = 3)\n",
92
- "y_predicted = km.fit_predict(df[['Age', 'Income']])\n",
93
- "y_predicted"
94
- ]
95
- },
96
- {
97
- "cell_type": "code",
98
- "execution_count": null,
99
- "metadata": {},
100
- "outputs": [],
101
- "source": [
102
- "km.cluster_centers_"
103
- ]
104
- },
105
- {
106
- "cell_type": "code",
107
- "execution_count": null,
108
- "metadata": {},
109
- "outputs": [],
110
- "source": [
111
- "df['cluster']=y_predicted\n",
112
- "df.head()"
113
- ]
114
- },
115
- {
116
- "cell_type": "code",
117
- "execution_count": null,
118
- "metadata": {},
119
- "outputs": [],
120
- "source": [
121
- "df1 = df[df.cluster == 0]\n",
122
- "df2 = df[df.cluster == 1]\n",
123
- "df3 = df[df.cluster == 2]\n",
124
- "\n",
125
- "plt.scatter(df1.Age, df1.Income, color = 'black')\n",
126
- "plt.scatter(df2.Age, df2.Income, color = 'orange')\n",
127
- "plt.scatter(df3.Age, df3.Income, color = 'maroon')\n",
128
- "plt.scatter(km.cluster_centers_[:,0], km.cluster_centers_[:,1], color = 'purple', marker = '*', label = 'centroid')\n",
129
- "plt.xlabel('Age')\n",
130
- "plt.ylabel('Income in $')"
131
- ]
132
- },
133
- {
134
- "cell_type": "code",
135
- "execution_count": null,
136
- "metadata": {},
137
- "outputs": [],
138
- "source": []
139
- },
140
- {
141
- "cell_type": "markdown",
142
- "metadata": {},
143
- "source": [
144
- "## Association Rule Mining"
145
- ]
146
- },
147
- {
148
- "cell_type": "code",
149
- "execution_count": null,
150
- "metadata": {},
151
- "outputs": [],
152
- "source": [
153
- "#!pip install --no-binary :all: mlxtend"
154
- ]
155
- },
156
- {
157
- "cell_type": "code",
158
- "execution_count": null,
159
- "metadata": {},
160
- "outputs": [],
161
- "source": [
162
- "import pandas as pd\n",
163
- "from mlxtend.preprocessing import TransactionEncoder\n",
164
- "from mlxtend.frequent_patterns import apriori"
165
- ]
166
- },
167
- {
168
- "cell_type": "code",
169
- "execution_count": null,
170
- "metadata": {},
171
- "outputs": [],
172
- "source": [
173
- "dataset = [[\"Milk\", \"Cola\", \"Beer\"], \n",
174
- " [\"Milk\", \"Pepsi\", \"Juice\"], \n",
175
- " [\"Milk\", \"Beer\"],\n",
176
- " [\"Cola\", \"Juice\"],\n",
177
- " [\"Milk\", \"Pepsi\", \"Beer\"]]\n",
178
- "dataset"
179
- ]
180
- },
181
- {
182
- "cell_type": "code",
183
- "execution_count": null,
184
- "metadata": {},
185
- "outputs": [],
186
- "source": [
187
- "var = TransactionEncoder()\n",
188
- "var_arr = var.fit(dataset).transform(dataset)\n",
189
- "df = pd.DataFrame(var_arr, columns = var.columns_)"
190
- ]
191
- },
192
- {
193
- "cell_type": "code",
194
- "execution_count": null,
195
- "metadata": {},
196
- "outputs": [],
197
- "source": [
198
- "df"
199
- ]
200
- },
201
- {
202
- "cell_type": "code",
203
- "execution_count": null,
204
- "metadata": {},
205
- "outputs": [],
206
- "source": [
207
- "#from mlxtend.frequent_patterns import apriori\n",
208
- "fq_is = apriori(df, min_support = 0.5, use_colnames = True)\n",
209
- "fq_is"
210
- ]
211
- },
212
- {
213
- "cell_type": "code",
214
- "execution_count": null,
215
- "metadata": {},
216
- "outputs": [],
217
- "source": [
218
- "from mlxtend.frequent_patterns import association_rules\n",
219
- "rules_ap = association_rules(fq_is, metric = \"confidence\", min_threshold = 0.01)\n",
220
- "rules_ap"
221
- ]
222
- },
223
- {
224
- "cell_type": "code",
225
- "execution_count": null,
226
- "metadata": {},
227
- "outputs": [],
228
- "source": []
229
- },
230
- {
231
- "cell_type": "markdown",
232
- "metadata": {},
233
- "source": [
234
- "## Decision Tree"
235
- ]
236
- },
237
- {
238
- "cell_type": "code",
239
- "execution_count": null,
240
- "metadata": {},
241
- "outputs": [],
242
- "source": [
243
- "import pandas as pd\n",
244
- "from sklearn.preprocessing import LabelEncoder\n",
245
- "from sklearn import tree"
246
- ]
247
- },
248
- {
249
- "cell_type": "code",
250
- "execution_count": null,
251
- "metadata": {},
252
- "outputs": [],
253
- "source": [
254
- "df = pd.read_csv('C:/Users/Balakrishnan/Desktop/6th sem/Data Mining/Datasets/lung cancer.csv')\n",
255
- "df.head()"
256
- ]
257
- },
258
- {
259
- "cell_type": "code",
260
- "execution_count": null,
261
- "metadata": {},
262
- "outputs": [],
263
- "source": [
264
- "df.isnull().any()"
265
- ]
266
- },
267
- {
268
- "cell_type": "code",
269
- "execution_count": null,
270
- "metadata": {},
271
- "outputs": [],
272
- "source": [
273
- "le_GENDER = LabelEncoder()\n",
274
- "le_LUNG_CANCER = LabelEncoder()\n",
275
- "\n",
276
- "df['Gender'] = le_GENDER.fit_transform(df['GENDER'])\n",
277
- "df['LungCancer'] = le_LUNG_CANCER.fit_transform(df['LUNG_CANCER'])\n",
278
- "\n",
279
- "df.head()\n"
280
- ]
281
- },
282
- {
283
- "cell_type": "code",
284
- "execution_count": null,
285
- "metadata": {},
286
- "outputs": [],
287
- "source": [
288
- "df = df.drop(['GENDER','LUNG_CANCER'], axis = 1)\n",
289
- "df.head()"
290
- ]
291
- },
292
- {
293
- "cell_type": "code",
294
- "execution_count": null,
295
- "metadata": {},
296
- "outputs": [],
297
- "source": []
298
- },
299
- {
300
- "cell_type": "code",
301
- "execution_count": null,
302
- "metadata": {},
303
- "outputs": [],
304
- "source": [
305
- "inputs = df.drop('LungCancer', axis = 1)\n",
306
- "target = df['LungCancer']\n",
307
- "inputs.head()"
308
- ]
309
- },
310
- {
311
- "cell_type": "code",
312
- "execution_count": null,
313
- "metadata": {},
314
- "outputs": [],
315
- "source": [
316
- "print(target)"
317
- ]
318
- },
319
- {
320
- "cell_type": "code",
321
- "execution_count": null,
322
- "metadata": {},
323
- "outputs": [],
324
- "source": [
325
- "from sklearn import tree\n",
326
- "model = tree.DecisionTreeClassifier()\n",
327
- "model.fit(inputs, target)\n",
328
- "model.score(inputs, target)\n"
329
- ]
330
- },
331
- {
332
- "cell_type": "code",
333
- "execution_count": null,
334
- "metadata": {},
335
- "outputs": [],
336
- "source": [
337
- "print(model.predict([[63,1,2,2,1,2,1,1,2,1,2,1,1,2,2]]))"
338
- ]
339
- },
340
- {
341
- "cell_type": "code",
342
- "execution_count": null,
343
- "metadata": {},
344
- "outputs": [],
345
- "source": [
346
- "tree.plot_tree(model) "
347
- ]
348
- },
349
- {
350
- "cell_type": "code",
351
- "execution_count": null,
352
- "metadata": {},
353
- "outputs": [],
354
- "source": [
355
- "#fig, axes = plt.subplots(nrows = 1, ncols = 1, figsize = (2,2), dpi = 300)\n",
356
- "#tree.plot_tree(model,\n",
357
- " # feature_names = 'AGE''SMOKING''YELLOW_FINGERS''ANXIETY''PEER_PRESSURE''CHRONIC DISEASE''FATIGUE''ALLERGY''WHEEZING''ALCOHOL CONSUMING''COUGHING''SHORTNESS OF BREATH''SWALLOWING DIFFICULTY''CHEST PAIN''GENDER',\n",
358
- " # class_names = ['true', 'false'],\n",
359
- " # filled =True, rounded=True,\n",
360
- " # )\n",
361
- " "
362
- ]
363
- },
364
- {
365
- "cell_type": "markdown",
366
- "metadata": {},
367
- "source": [
368
- "## SVM bank loan"
369
- ]
370
- },
371
- {
372
- "cell_type": "code",
373
- "execution_count": null,
374
- "metadata": {},
375
- "outputs": [],
376
- "source": [
377
- "# libraries\n",
378
- "import pandas as pd\n",
379
- "import numpy as np\n",
380
- "import matplotlib.pyplot as plt\n",
381
- "import seaborn\n",
382
- "import sklearn\n",
383
- "from sklearn import model_selection\n",
384
- "from sklearn.model_selection import train_test_split\n",
385
- "from sklearn import svm\n",
386
- "from sklearn.svm import SVC\n",
387
- "from sklearn import metrics\n",
388
- "from sklearn.metrics import confusion_matrix\n",
389
- "from sklearn.metrics import classification_report\n",
390
- "from sklearn.metrics import roc_curve, auc "
391
- ]
392
- },
393
- {
394
- "cell_type": "code",
395
- "execution_count": null,
396
- "metadata": {},
397
- "outputs": [],
398
- "source": [
399
- "bl = pd.read_csv(\"C:/Users/Balakrishnan/Desktop/6th sem/Data Mining/Datasets/bankloan.csv\")\n",
400
- "bl = pd.DataFrame(bl)"
401
- ]
402
- },
403
- {
404
- "cell_type": "code",
405
- "execution_count": null,
406
- "metadata": {},
407
- "outputs": [],
408
- "source": [
409
- "bl_data = bl.drop(['address','ed','debtinc','employ'],1)\n",
410
- "bl_data.head(3)"
411
- ]
412
- },
413
- {
414
- "cell_type": "code",
415
- "execution_count": null,
416
- "metadata": {},
417
- "outputs": [],
418
- "source": [
419
- "x = bl_data.drop('default', axis = 1)\n",
420
- "y = bl_data.default\n",
421
- "x.head()"
422
- ]
423
- },
424
- {
425
- "cell_type": "code",
426
- "execution_count": null,
427
- "metadata": {},
428
- "outputs": [],
429
- "source": [
430
- "y.head()"
431
- ]
432
- },
433
- {
434
- "cell_type": "code",
435
- "execution_count": null,
436
- "metadata": {},
437
- "outputs": [],
438
- "source": [
439
- "x_train, x_test, y_train, y_test = train_test_split(x, y, test_size=0.4, random_state=0)"
440
- ]
441
- },
442
- {
443
- "cell_type": "code",
444
- "execution_count": null,
445
- "metadata": {},
446
- "outputs": [],
447
- "source": [
448
- "# building an SVM having penalty 10 and gamma auto optimized\n",
449
- "svct = SVC(kernel = 'linear',\n",
450
- " C = 10, gamma = 'auto', \n",
451
- " probability = True).fit(x_train, y_train)\n",
452
- "\n",
453
- "print(svct)"
454
- ]
455
- },
456
- {
457
- "cell_type": "code",
458
- "execution_count": null,
459
- "metadata": {},
460
- "outputs": [],
461
- "source": [
462
- "y_pred = svct.predict(x_test)\n",
463
- "y_pred"
464
- ]
465
- },
466
- {
467
- "cell_type": "code",
468
- "execution_count": null,
469
- "metadata": {},
470
- "outputs": [],
471
- "source": [
472
- "cnf = confusion_matrix(y_test, y_pred)\n",
473
- "cnf"
474
- ]
475
- },
476
- {
477
- "cell_type": "code",
478
- "execution_count": null,
479
- "metadata": {},
480
- "outputs": [],
481
- "source": [
482
- "# ROC Curve with AUC Plotting\n",
483
- "# finding probabilities for 0 and 1 classses\n",
484
- "preds1 = svct.predict_proba(x_test)[:,1]\n",
485
- "print(preds1)"
486
- ]
487
- },
488
- {
489
- "cell_type": "code",
490
- "execution_count": null,
491
- "metadata": {},
492
- "outputs": [],
493
- "source": [
494
- "from sklearn.metrics import roc_curve, auc \n",
495
- "fpr1, tpr1, thresholds1 = metrics.roc_curve(y_test, preds1)"
496
- ]
497
- },
498
- {
499
- "cell_type": "code",
500
- "execution_count": null,
501
- "metadata": {},
502
- "outputs": [],
503
- "source": [
504
- "fpr1"
505
- ]
506
- },
507
- {
508
- "cell_type": "code",
509
- "execution_count": null,
510
- "metadata": {},
511
- "outputs": [],
512
- "source": [
513
- "tpr1"
514
- ]
515
- },
516
- {
517
- "cell_type": "code",
518
- "execution_count": null,
519
- "metadata": {},
520
- "outputs": [],
521
- "source": [
522
- "thresholds1"
523
- ]
524
- },
525
- {
526
- "cell_type": "code",
527
- "execution_count": null,
528
- "metadata": {},
529
- "outputs": [],
530
- "source": [
531
- "df1 = pd.DataFrame(dict(fpr = fpr1, tpr = tpr1))\n",
532
- "auc = metrics.auc(fpr1, tpr1)"
533
- ]
534
- },
535
- {
536
- "cell_type": "code",
537
- "execution_count": null,
538
- "metadata": {},
539
- "outputs": [],
540
- "source": [
541
- "# plotting\n",
542
- "plt.figure()\n",
543
- "plt.plot(fpr1, tpr1, color = 'darkorange', label = 'ROC CURVE(area = %0.2f)' % auc)\n",
544
- "plt.plot([0,1], [0,1], color = 'navy')\n",
545
- "plt.xlim([0.0, 1.0])\n",
546
- "plt.ylim([0.0, 1.05])\n",
547
- "plt.xlabel('False Positive Rate')\n",
548
- "plt.ylabel('True Positive Rate')\n",
549
- "plt.title('ROC example')\n",
550
- "plt.legend(loc = 'lower right')\n",
551
- "plt.show()\n"
552
- ]
553
- },
554
- {
555
- "cell_type": "markdown",
556
- "metadata": {},
557
- "source": [
558
- "## SVM Iris"
559
- ]
560
- },
561
- {
562
- "cell_type": "code",
563
- "execution_count": null,
564
- "metadata": {},
565
- "outputs": [],
566
- "source": [
567
- "from sklearn import svm\n",
568
- "import pandas as pd\n"
569
- ]
570
- },
571
- {
572
- "cell_type": "code",
573
- "execution_count": null,
574
- "metadata": {},
575
- "outputs": [],
576
- "source": [
577
- "df = pd.read_csv(\"C:/Users/Balakrishnan/Desktop/6th sem/Data Mining/Datasets/iris_df.csv\")\n",
578
- "df.head(5)"
579
- ]
580
- },
581
- {
582
- "cell_type": "code",
583
- "execution_count": null,
584
- "metadata": {},
585
- "outputs": [],
586
- "source": [
587
- "df.columns = ['X4','X3','X1','X2','Y']\n",
588
- "df = df.drop(['X4','X3'],1)\n",
589
- "df.head()"
590
- ]
591
- },
592
- {
593
- "cell_type": "code",
594
- "execution_count": null,
595
- "metadata": {},
596
- "outputs": [],
597
- "source": [
598
- "from sklearn.model_selection import train_test_split\n",
599
- "from sklearn import svm"
600
- ]
601
- },
602
- {
603
- "cell_type": "code",
604
- "execution_count": null,
605
- "metadata": {},
606
- "outputs": [],
607
- "source": [
608
- "x = df.drop(\"Y\", axis = 1)\n",
609
- "y = df.Y\n",
610
- "\n",
611
- "support = svm.SVC()\n",
612
- "\n",
613
- "\n",
614
- "trainx, testx, trainy, testy = train_test_split(x,y, test_size = 0.3)\n",
615
- "support.fit(trainx, trainy)\n",
616
- "print('Accuracy:/n', support.score(testx, testy))\n",
617
- "\n",
618
- "pred = support.predict(testx)"
619
- ]
620
- },
621
- {
622
- "cell_type": "markdown",
623
- "metadata": {},
624
- "source": [
625
- "## Linear regression single variable"
626
- ]
627
- },
628
- {
629
- "cell_type": "code",
630
- "execution_count": null,
631
- "metadata": {},
632
- "outputs": [],
633
- "source": [
634
- "import pandas as pd\n",
635
- "import numpy as np\n",
636
- "from sklearn import linear_model\n",
637
- "import matplotlib.pyplot as plt"
638
- ]
639
- },
640
- {
641
- "cell_type": "code",
642
- "execution_count": null,
643
- "metadata": {},
644
- "outputs": [],
645
- "source": [
646
- "df = pd.read_csv(\"C:/Users/Balakrishnan/Desktop/6th sem/Data Mining/Datasets/house_price.csv\")\n",
647
- "df.head()"
648
- ]
649
- },
650
- {
651
- "cell_type": "code",
652
- "execution_count": null,
653
- "metadata": {},
654
- "outputs": [],
655
- "source": [
656
- "plt.xlabel('area')\n",
657
- "plt.ylabel('price')\n",
658
- "plt.scatter(df.area,df.price,color = 'blue')"
659
- ]
660
- },
661
- {
662
- "cell_type": "code",
663
- "execution_count": null,
664
- "metadata": {},
665
- "outputs": [],
666
- "source": [
667
- "reg = linear_model.LinearRegression()\n",
668
- "reg.fit(df[['area']], df.price)"
669
- ]
670
- },
671
- {
672
- "cell_type": "code",
673
- "execution_count": null,
674
- "metadata": {},
675
- "outputs": [],
676
- "source": [
677
- "# prediction\n",
678
- "reg.predict([[3100]])"
679
- ]
680
- },
681
- {
682
- "cell_type": "code",
683
- "execution_count": null,
684
- "metadata": {},
685
- "outputs": [],
686
- "source": [
687
- "reg.coef_"
688
- ]
689
- },
690
- {
691
- "cell_type": "code",
692
- "execution_count": null,
693
- "metadata": {},
694
- "outputs": [],
695
- "source": [
696
- "reg.intercept_"
697
- ]
698
- },
699
- {
700
- "cell_type": "code",
701
- "execution_count": null,
702
- "metadata": {},
703
- "outputs": [],
704
- "source": [
705
- "plt.xlabel('area')\n",
706
- "plt.ylabel('price')\n",
707
- "plt.scatter(df.area,df.price,color='green')\n",
708
- "plt.plot(df.area,reg.predict(df[['area']]), color = 'red')"
709
- ]
710
- },
711
- {
712
- "cell_type": "markdown",
713
- "metadata": {},
714
- "source": [
715
- "## Linear Regression Multiple Variable\n"
716
- ]
717
- },
718
- {
719
- "cell_type": "code",
720
- "execution_count": null,
721
- "metadata": {},
722
- "outputs": [],
723
- "source": [
724
- "# pandas numpy\n",
725
- "from sklearn import linear_model"
726
- ]
727
- },
728
- {
729
- "cell_type": "code",
730
- "execution_count": null,
731
- "metadata": {},
732
- "outputs": [],
733
- "source": [
734
- "df = pd.read_csv(\"C:/Users/Balakrishnan/Desktop/6th sem/Data Mining/Datasets/house_pricemodel.csv\")\n",
735
- "df.head()"
736
- ]
737
- },
738
- {
739
- "cell_type": "code",
740
- "execution_count": null,
741
- "metadata": {},
742
- "outputs": [],
743
- "source": [
744
- "df.isnull()"
745
- ]
746
- },
747
- {
748
- "cell_type": "code",
749
- "execution_count": null,
750
- "metadata": {},
751
- "outputs": [],
752
- "source": [
753
- "df.bedrooms.median()\n",
754
- "df.bedrooms = df.bedrooms.fillna(df.bedrooms.median())\n",
755
- "df"
756
- ]
757
- },
758
- {
759
- "cell_type": "code",
760
- "execution_count": null,
761
- "metadata": {},
762
- "outputs": [],
763
- "source": [
764
- "reg = linear_model.LinearRegression()\n",
765
- "reg.fit(df.drop('price', axis = 1), df.price)"
766
- ]
767
- },
768
- {
769
- "cell_type": "code",
770
- "execution_count": null,
771
- "metadata": {},
772
- "outputs": [],
773
- "source": [
774
- "reg.coef_"
775
- ]
776
- },
777
- {
778
- "cell_type": "code",
779
- "execution_count": null,
780
- "metadata": {},
781
- "outputs": [],
782
- "source": [
783
- "reg.predict([[3200,3,10]])"
784
- ]
785
- },
786
- {
787
- "cell_type": "markdown",
788
- "metadata": {},
789
- "source": [
790
- "## SImple Logistic Regression"
791
- ]
792
- },
793
- {
794
- "cell_type": "code",
795
- "execution_count": null,
796
- "metadata": {},
797
- "outputs": [],
798
- "source": [
799
- "# pandas, matplotlib\n",
800
- "from sklearn.model_selection import train_test_split\n",
801
- "from sklearn.linear_model import LogisticRegression"
802
- ]
803
- },
804
- {
805
- "cell_type": "code",
806
- "execution_count": null,
807
- "metadata": {},
808
- "outputs": [],
809
- "source": [
810
- "df = pd.read_csv(\"C:/Users/Balakrishnan/Desktop/6th sem/Data Mining/Datasets/insurance_age.csv\")\n",
811
- "df.head()"
812
- ]
813
- },
814
- {
815
- "cell_type": "code",
816
- "execution_count": null,
817
- "metadata": {},
818
- "outputs": [],
819
- "source": [
820
- "plt.scatter(df.age,df.bought_insurance, marker ='*', color = 'purple')"
821
- ]
822
- },
823
- {
824
- "cell_type": "code",
825
- "execution_count": null,
826
- "metadata": {},
827
- "outputs": [],
828
- "source": [
829
- "x_train, x_test, y_train, y_test = train_test_split(df[['age']], df.bought_insurance, train_size=0.8)"
830
- ]
831
- },
832
- {
833
- "cell_type": "code",
834
- "execution_count": null,
835
- "metadata": {},
836
- "outputs": [],
837
- "source": [
838
- "model = LogisticRegression()\n",
839
- "model.fit(x_train, y_train)"
840
- ]
841
- },
842
- {
843
- "cell_type": "code",
844
- "execution_count": null,
845
- "metadata": {},
846
- "outputs": [],
847
- "source": [
848
- "y_predicted = model.predict(x_test)\n",
849
- "model.predict_proba(x_test)"
850
- ]
851
- },
852
- {
853
- "cell_type": "code",
854
- "execution_count": null,
855
- "metadata": {},
856
- "outputs": [],
857
- "source": [
858
- "model.score(x_test, y_test)"
859
- ]
860
- },
861
- {
862
- "cell_type": "code",
863
- "execution_count": null,
864
- "metadata": {},
865
- "outputs": [],
866
- "source": [
867
- "y_predicted"
868
- ]
869
- },
870
- {
871
- "cell_type": "code",
872
- "execution_count": null,
873
- "metadata": {},
874
- "outputs": [],
875
- "source": [
876
- "model.predict([[45]])"
877
- ]
878
- },
879
- {
880
- "cell_type": "markdown",
881
- "metadata": {},
882
- "source": [
883
- "## Gradient descent simple"
884
- ]
885
- },
886
- {
887
- "cell_type": "code",
888
- "execution_count": null,
889
- "metadata": {},
890
- "outputs": [],
891
- "source": [
892
- "import numpy as np\n",
893
- "import matplotlib.pyplot as plt"
894
- ]
895
- },
896
- {
897
- "cell_type": "code",
898
- "execution_count": null,
899
- "metadata": {},
900
- "outputs": [],
901
- "source": [
902
- "def gradient_descent(x,y):\n",
903
- " m_curr = b_curr = 0\n",
904
- " iterations = 10000\n",
905
- " n = len(x) # length of x\n",
906
- " learning_rate = 0.001\n",
907
- " #learning rate = .00001\n",
908
- " \n",
909
- " \n",
910
- " \n",
911
- " for i in range(iterations):\n",
912
- " y_predicted = m_curr * x + b_curr\n",
913
- " cost = (1/n) * sum([val**2 for val in (y-y_predicted)])\n",
914
- " md = -(2/n)*sum(x*(y-y_predicted)) # m derivative\n",
915
- " bd = -(2/n)*sum(y-y_predicted) # b derivative\n",
916
- " m_curr = m_curr - learning_rate * md\n",
917
- " b_curr = b_curr - learning_rate * bd\n",
918
- " #print (\"m {}, b {}, iteration {}\".format(m_curr,b_curr, i))\n",
919
- " print (\"m {}, b {}, cost {} iteration {}\".format(m_curr,b_curr,cost, i)) # each iteration to print the values"
920
- ]
921
- },
922
- {
923
- "cell_type": "code",
924
- "execution_count": null,
925
- "metadata": {},
926
- "outputs": [],
927
- "source": [
928
- "x = np.array([1,2,3,4,5])\n",
929
- "y = np.array([5,7,9,11,13])"
930
- ]
931
- },
932
- {
933
- "cell_type": "code",
934
- "execution_count": null,
935
- "metadata": {},
936
- "outputs": [],
937
- "source": [
938
- "gradient_descent(x,y)"
939
- ]
940
- },
941
- {
942
- "cell_type": "code",
943
- "execution_count": null,
944
- "metadata": {},
945
- "outputs": [],
946
- "source": []
947
- },
948
- {
949
- "cell_type": "code",
950
- "execution_count": null,
951
- "metadata": {},
952
- "outputs": [],
953
- "source": []
954
- },
955
- {
956
- "cell_type": "code",
957
- "execution_count": null,
958
- "metadata": {},
959
- "outputs": [],
960
- "source": []
961
- }
962
- ],
963
- "metadata": {
964
- "kernelspec": {
965
- "display_name": "Python 3",
966
- "language": "python",
967
- "name": "python3"
968
- },
969
- "language_info": {
970
- "codemirror_mode": {
971
- "name": "ipython",
972
- "version": 3
973
- },
974
- "file_extension": ".py",
975
- "mimetype": "text/x-python",
976
- "name": "python",
977
- "nbconvert_exporter": "python",
978
- "pygments_lexer": "ipython3",
979
- "version": "3.7.6"
980
- }
981
- },
982
- "nbformat": 4,
983
- "nbformat_minor": 4
984
- }