1. Tomasz Drabas, Denny Lee Packt Publishing Ltd, 27-Feb-2017 2. Data Analysis with Python and PySpark Jonathan Rioux Simon and Schuster, 12-Apr-2022 3. PySpark Cookbook: Over 60 recipes for implementing big data processing and analytics using Apache Spark and Python Denny Lee, Tomasz Drabas Packt Publishing Ltd, 29-Jun-2018 4. Machine Learning with PySpark: With Natural Language Processing and Recommender Systems Pramod Singh Apress, 14-Dec-2018 5. Learning Spark: Lightning-Fast Big Data Analysis Holden Karau, Andy Konwinski, Patrick Wendell, Matei Zaharia "O'Reilly Media, Inc.", 28-Jan-2015 6. Advanced Analytics with PySpark Akash Tandon, Sandy Ryza, Sean Owen, Uri Laserson, Josh Wills "O'Reilly Media, Inc.", 14-Jun-2022 7. PySpark Recipes: A Problem-Solution Approach with PySpark2 Raju Kumar Mishra Apress, 09-Dec-2017 8. Learn PySpark: Build Python-based Machine Learning and Deep Learning Models Pramod Singh Apress, 06-Sept-2019 9. Learning Spark Jules S. Damji, Brooke Wenig, Tathagata Das, Denny Lee "O'Reilly Media, Inc.", 16-Jul-2020 10. Advanced Analytics with Spark: Patterns for Learning from Data at Scale Sandy Ryza, Uri Laserson, Sean Owen, Josh Wills "O'Reilly Media, Inc.", 12-Jun-2017 11. Applied Data Science Using PySpark: Learn the End-to-End Predictive Model-Building Cycle Ramcharan Kakarla, Sundar Krishnan, Sridhar Alla Apress, 2021 12. Essential PySpark for Scalable Data Analytics: A beginner's guide to harnessing the power and ease of PySpark 3 Sreeram Nudurupati Packt Publishing Ltd, 29-Oct-2021 13. Spark: The Definitive Guide: Big Data Processing Made Simple Bill Chambers, Matei Zaharia "O'Reilly Media, Inc.", 08-Feb-2018 14. Spark for Python Developers Amit Nandi Packt Publishing, 24-Dec-2015 15. Frank Kane's Taming Big Data with Apache Spark and Python Frank Kane Packt Publishing Ltd, 30-Jun-2017 16. Stream Processing with Apache Spark: Mastering Structured Streaming and Spark Streaming Gerard Maas, Francois Garillot "O'Reilly Media, Inc.", 05-Jun-2019 17. Data Analytics with Spark Using Python Jeffrey Aven Addison-Wesley Professional, 18-Jun-2018 18. Graph Algorithms: Practical Examples in Apache Spark and Neo4j Mark Needham, Amy E. Hodler "O'Reilly Media, Inc.", 16-May-2019 19. Spark in Action: Covers Apache Spark 3 with Examples in Java, Python, and Scala Jean-Georges Perrin Simon and Schuster, 12-May-2020 20. Mastering Spark with R: The Complete Guide to Large-Scale Analysis and Modeling Javier Luraschi, Kevin Kuo, Edgar Ruiz "O'Reilly Media, Inc.", 07-Oct-2019 21. High Performance Spark: Best Practices for Scaling and Optimizing Apache Spark Holden Karau, Rachel Warren "O'Reilly Media, Inc.", 25-May-2017 22. Apache Spark in 24 Hours, Sams Teach Yourself Jeffrey Aven Sams Publishing, 31-Aug-2016
Sunday, February 12, 2023
PySpark Books (2023 Feb)
Download Books
Hands-on 5 Regression Algorithms Using Scikit-Learn
Download Code and Data
What is Regression? When the targets are real numbers and we are trying the establish a relationship between a target and a predictor, the problem is called a “regression problem”. Example 1: Salary vs Years of Experience Example 2: Weight vs Height Regression: Predicting Bengaluru Housing Prices 1. Linear Regression (Ordinary Least Squares algorithm) 2. Polynomial Regression 3. Linear Regression using Stochastic Gradient Descent 4. Regression using Support Vector Machines 5. Regression using Decision Trees Linear Regression (Ordinary Least Squares algorithm) 1: In Linear Regression, you try to fit a line to the data. Basic Idea Behind Ordinary Least Squares Algorithm: How much predictions are deviating from the actual data? Mapping errors on the graph: >>> import numpy as np >>> from sklearn.linear_model import LinearRegression >>> X = np.array([[1, 1], [1, 2], [2, 2], [2, 3]]) >>> # y = 1 * x_0 + 2 * x_1 + 3 >>> y = np.dot(X, np.array([1, 2])) + 3 >>> reg = LinearRegression().fit(X, y) >>> reg.score(X, y) 1.0 >>> reg.coef_ array([1., 2.]) >>> reg.intercept_ 3.0... >>> reg.predict(np.array([[3, 5]])) array([16.]) Ref: scikit-learn.org Which attributes to transform during EDA? 1. Check if you model requires numerical features and if you can make the attributes numerical. For ex, for the problem of predicting housing prices, we can convert BHK column to floating point numbers: 2 BHK -> 2 2 BHK + Study -> 2.5 3 BHK -> 3 3 BHK + Servent -> 3.5 2. What if the ‘bhk’ attribute is not given? >>> pandas_df.dropna(subset = [‘bhk’]) If we have engineered all the features, can we drop null records from all the features? >>> pandas_df.dropna(inplace = True) 2. Polynomial Regression What if your data is actually more complex than a simple straight line? Generating Polynomial Features Generate a new feature matrix consisting of all polynomial combinations of the features with degree less than or equal to the specified degree. For example, if an input sample is two dimensional and of the form [a, b], the degree-2 polynomial features are [1, a, b, a^2, ab, b^2]. include_bias: bool, default=True ::: If True (default), then include a bias column, the feature in which all polynomial powers are zero (i.e. a column of ones - acts as an intercept term in a linear model). >>> import numpy as np >>> from sklearn.preprocessing import PolynomialFeatures >>> X = np.arange(6).reshape(3, 2) >>> X array([[0, 1], [2, 3], [4, 5]]) >>> poly = PolynomialFeatures(2) >>> poly.fit_transform(X) array([[ 1., 0., 1., 0., 0., 1.], [ 1., 2., 3., 4., 6., 9.], [ 1., 4., 5., 16., 20., 25.]]) Building the Polynomial Regression Model >>> from sklearn.preprocessing import PolynomialFeatures >>> poly_features = PolynomialFeatures(degree=2, include_bias=False) >>> X_poly = poly_features.fit_transform(X) >>> X[0] array([-0.75275929]) >>> X_poly[0] Array([-0.75275929, 0.56664654]) X_poly now contains the original feature of X plus the square of this feature. Now you can fit a LinearRegression model to this extended training data: >>> lin_reg = LinearRegression() >>> lin_reg.fit(X_poly, y) >>> lin_reg.intercept_, lin_reg.coef_ (array([ 1.78134581]), array([[ 0.93366893, 0.56456263]]))Tags: Machine Learning,Technology,3. Linear Regression using Stochastic Gradient Descent
What’s Gradient Descent? Gradient Descent is a very generic optimization algorithm capable of finding optimal solutions to a wide range of problems. The general idea of Gradient Descent is to tweak parameters iteratively in order to minimize a cost function. Suppose you are lost in the mountains in a dense fog; you can only feel the slope of the ground below your feet. A good strategy to get to the bottom of the valley quickly is to go downhill in the direction of the steepest slope. This is exactly what Gradient Descent does: it measures the local gradient of the error function with regards to the parameter vector θ, and it goes in the direction of descending gradient. Once the gradient is zero, you have reached a minimum! Concretely, you start by filling θ with random values (this is called random initialization), and then you improve it gradually, taking one baby step at a time, each step attempting to decrease the cost function (e.g., the MSE), until the algorithm converges to a minimum (see the figure below). Solving the problem of Linear Regression (Using SGD) Here are the high level steps that we take in implementing a simple and naive Linear Regression model using SGD: 1. Random Initialization: Initialize the model with a line along the x-axis. 2. Calculate the error function for this line. 3. By doing minor changes (d(slope) and d(intercept)) in slope and intercept, adjust the linear model to reduce the error function. 4. Repeat steps (2) and (3) until convergence. Code from sklearn.linear_model import SGDRegressor sgd_reg = SGDRegressor(n_iter=50, penalty=None, eta0=0.1) sgd_reg.fit(X, y.ravel()) >>> sgd_reg.intercept_, sgd_reg.coef_ (array([ 4.18380366]), array([ 2.74205299]))4. Regression using Support Vector Machines
We start with explaining what SVM is and then move on to using it for regression: The fundamental idea behind SVMs is best explained with some pictures. Figures below shows part of the iris dataset. The two classes can clearly be separated easily with a straight line (they are linearly separable). The left plot shows the decision boundaries of three possible linear classifiers. The model whose decision boundary is represented by the dashed line is so bad that it does not even separate the classes properly. The other two models work perfectly on this training set, but their decision boundaries come so close to the instances that these models will probably not perform as well on new instances. In contrast, the solid line in the plot on the right represents the decision boundary of an SVM classifier; this line not only separates the two classes but also stays as far away from the closest training instances as possible. You can think of an SVM classifier as fitting the widest possible street (represented by the parallel dashed lines) between the classes. This is called large margin classification. And the circled points are your ‘support vectors’. SVM Regression As we mentioned earlier, the SVM algorithm is quite versatile: not only does it support linear and nonlinear classification, but it also supports linear and nonlinear regression. The trick is to reverse the objective: Instead of trying to fit the largest possible street between two classes while limiting margin violations, SVM Regression tries to fit as many instances as possible on the street while limiting margin violations (i.e., instances off the street). The width of the street is controlled by a hyperparameter ϵ. Figure below shows two linear SVM Regression models trained on some random linear data, one with a large margin (ϵ = 1.5) and the other with a small margin (ϵ = 0.5).5. Regression using Decision Trees
First we would explain what Decision Trees are and how they work. Binary decision trees operate by subjecting attributes to a series of binary (yes / no) decisions. Each decision leads to one of two possibilities. Each decision leads to another decision or it leads to prediction. How a Binary Decision Tree Generates Predictions? When an observation or row is passed to a nonterminal node, the row answers the node’s question. If it answers yes, the row of attributes is passed to the leaf node below and to the left of the current node. If the row answers no, the row of attributes is passed to the leaf node below and to the right of the current node. The process continues recursively until the row arrives at a terminal (that is, leaf) node where a prediction value is assigned to the row. The value assigned by the leaf node is the mean of the outcomes of the all the training observations that wound up in the leaf node. Below is the Decision Tree for Iris Dataset. Simple Psuedo Code for ‘Regression Using Decision Tree’ Only For The Purpose of Demonstration. Step 1: Find avarage value for interval of x and y. Let us call these values XA abd YA. Step 2: Split the curve into two by drawing a vertical line. Step 3: For x < XA, choose the average values of (x, y) from left side, drawing a horizontal line passing from this point on the left side. Step 4: For x > XA, choose the average values of (x, y) from right side, drawing a horizontal line passing from this point on the left side. Repeat steps (1) to (4) for (n-1) times where n is the depth you want in your decision tree. Moving on to Regression. Below is our sample data: Block diagram of depth 1 tree for simple problem Comparison of predictions and actual values versus attribute for simple example Notice how the predicted value for each region is always the average target value of the instances in that region. The algorithm splits each region in a way that makes most training instances as close as possible to that predicted value. DecisionTreeRegressor using sklearn from sklearn.tree import DecisionTreeRegressor tree_reg = DecisionTreeRegressor(max_depth=2) tree_reg.fit(X, y)References
1. Linear Regression (Ordinary Least Squares algorithm) 1.1. linear-regression-theory 1.2. penalized linear regression 2, 3, 4: Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems Book by Aurelien Geron 5: Machine Learning in Python (Essential Techniques For Predictive Analysis) By: Michael Bowles
Thursday, February 9, 2023
Machine Learning Books (Mar 2020)
Download Books
Putting the books listed below into three categories based on complexity
I: Mathematical Theory
-
1. Deep Learning
Book by Aaron Courville, Ian Goodfellow, and Yoshua Bengio -
5. Pattern Recognition and Machine Learning
Book by Christopher Bishop -
8. Understanding Machine Learning: From Theory to Algorithms
Textbook by Shai Ben-David and Shai Shalev-Shwartz
II: Mix of Theory and Applied Study
-
2. Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow
Concepts, Tools, and Techniques to Build Intelligent Systems
Book by Aurelien Geron
III: Applied Study
-
6. The Hundred-Page Machine Learning Book
Book by Andriy Burkov -
9. Machine Learning for Absolute Beginners: A Plain English Introduction
Book by O. Theobald -
52. Machine Learning in Python (Essential Techniques For Predictive Analysis)
By: Michael Bowles -
53. Fifty Algorithms Every Programmer Should Know (2e)
By: Imran Ahmad (PhD) -
54. Applied Machine Learning and AI for Engineers
Solve Business Problems That Can't Be Solved Algorithmically (Release 1)
Jeff Prosise
O’Reilly Media, Inc. (2022)
Tags: List of Books,Machine Learning,All
1. Deep Learning Book by Aaron Courville, Ian Goodfellow, and Yoshua Bengio 2. Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow Concepts, Tools, and Techniques to Build Intelligent Systems Book by Aurelien Geron 3. The Elements of Statistical Learning Book by Jerome H. Friedman, Robert Tibshirani, and Trevor Hastie 4. An Introduction to Statistical Learning: With Applications in R Book 5. Pattern Recognition and Machine Learning Book by Christopher Bishop 6. The Hundred-Page Machine Learning Book Book by Andriy Burkov 7. Deep Learning with Python Book by François Chollet 8. Understanding Machine Learning: From Theory to Algorithms Textbook by Shai Ben-David and Shai Shalev-Shwartz 9. Machine Learning for Absolute Beginners: A Plain English Introduction Book by O. Theobald 10. Python Machine Learning Book by Sebastian Raschka 11. Artificial Intelligence: A Modern Approach Textbook by Peter Norvig and Stuart J. Russell 12. Introduction to Machine Learning Textbook by Ethem Alpaydın 13. Machine Learning: A Probabilistic Perspective Textbook by Kevin P. Murphy 14. Machine Learning for Hackers Book by Drew Conway and John Myles White 15. Programming Collective Intelligence Book: O'Reilly 16. Machine Learning For Dummies Book by John Mueller and Luca Massaron 17. Bayesian Reasoning and Machine Learning Book by David Barber 18. Reinforcement Learning: An Introduction Book by Andrew Barto and Richard S. Sutton 19. Learning from Data: A Short Course Book by Hsuan-Tien Lin, Malik Magdon-Ismail, and Yaser Abu-Mostafa 20. Machine Learning in Action Book by Peter Harrington 21. Machine Learning: The Art and Science of Algorithms that Make Sense of Data Book by Peter Flach 22. Introduction to Machine Learning with Python A Guide for Data Scientists Author(s): Andreas C. Müller, Sarah Guido Publisher: O’Reilly Media, Year: 2016 23. Fundamentals of Machine Learning for Predictive Data Analytics: Algorithms, Worked Examples, and Case Studies Textbook by Aoife D'Arcy, Brian Mac Namee, and John D. Kelleher 24. Mining of Massive Datasets Book by Anand Rajaraman and Jeffrey Ullman 25. Foundations of Machine Learning Textbook by Afshin Rostamizadeh, Ameet Talwalkar, and Mehryar Mohri 26. Superintelligence: Paths, Dangers, Strategies Book by Nick Bostrom 27. Make Your Own Neural Network: A Gentle Journey Through the Mathematics ... Book by Tariq Rashid 28. Python Machine Learning: Machine Learning and Deep Learning with Python, Scikit-learn, and TensorFlow 2, 3rd Edition Book by Sebastian Raschka and Vahid Mirjalili 29. Machine Learning: An Algorithmic Perspective Book by Stephen Marsland 30. The Master Algorithm: How the Quest for the Ultimate Learning Machine Will Remake Our World Book by Pedro Domingos 31. Grokking Deep Learning Book by Andrew W. Trask 32. Advances in Financial Machine Learning Book by Marcos Lopez de Prado 33. Machine Learning: A Guide to Current Research Book by Tom M. Mitchell 34. Pattern Classification Book by David G. Stork, Peter E. Hart, and Richard O. Duda 35. Building Machine Learning Systems with Python - Second Edition Book by Luis Pedro Coelho and Willi Richert 36. Bayesian Methods for Hackers: Probabilistic Programming and Bayesian Inference Book by Cameron Davidson-Pilon 37. Information Theory, Inference and Learning Algorithms Textbook by David J. C. MacKay 38. Probabilistic Graphical Models: Principles and Techniques Book by Daphne Koller and Nir Friedman 39. Interpretable Machine Learning Book by Christoph Molnar 40. The Book of Why: The New Science of Cause and Effect Book by Dana Mackenzie and Judea Pearl 41. Fundamentals of Deep Learning: Designing Next-Generation Machine Intelligence Algorithms Book by Nicholas Locascio and Nikhil Buduma 42. Deep Reinforcement Learning Hands-On: Apply Modern RL Methods, with Deep Q-networks, Value Iteration, Policy Gradients, TRPO, AlphaGo Zero and More 43. Think Stats Book by Allen B. Downey 44. Gaussian Processes for Machine Learning Book by Carl Edward Rasmussen and Christopher K. I. Williams 45. Data Mining: Practical Machine Learning Tools and Techniques Book 46. Machine Learning with R Book by Brett Lantz 47. Python Data Science Handbook: Essential Tools for Working with Data Book by Jake VanderPlas 48. Real world machine learning: video edition Book by Henrik Brink, Joseph Richards, and Mark Fetherolf 49. Machine Learning Algorithms: Popular Algorithms for Data Science and Machine Learning Book by Giuseppe Bonaccorso 50. Machine Learning: A Bayesian and Optimization Perspective Book by Sergios Theodoridis 51. Mathematics for Machine Learning Textbook by A. Aldo Faisal, Cheng Soon Ong, and Marc Peter Deisenroth 52. Machine Learning in Python (Essential Techniques For Predictive Analysis) By: Michael Bowles 53. Fifty Algorithms Every Programmer Should Know (2e) By: Imran Ahmad (PhD) 54. Applied Machine Learning and AI for Engineers Solve Business Problems That Can't Be Solved Algorithmically (Release 1) Jeff Prosise O’Reilly Media, Inc. (2022)
Wednesday, February 8, 2023
Spark SQL in Images
Tags: Spark,Technology,1. Spark's components
2. Spark SQL Architecture
3. SQL Data Types
4. Spark's context objects
5. File Formats Supported By Spark
6. SQL Workflow
7. Catalyst Optimizer
Below steps explain the workflow of the catalyst optimizer: 1. Analyzing a logical plan with the metadata 2. Optimizing the logical plan 3. Creating multiple physical plans 4. Analyzing the plans and finding the most optimal physical plan 5. Converting the physical plan to RDDs
A Solved Exercise in RDD Filter and Join Operations (Interview Preparation)
Download Code and Data
Problem Statement: Consider the Universal Identity Number data scenario with two datasets UIN Customer data and Bank account linking data. UIN Card data (UINCardData.csv): Schema Details: UIN, MobileNumber,Gender,SeniorCitizens,Income Bank account link data (BankAccountLink.csv): Schema Details: MobileNumber, LinkedtoBankAccount, BankAccountNumber Requirement Join both datasets and find the UIN number that is not linked with the Bank Account number. Print UIN number and BankAccountNumber. Save the final output to a specified HDFS directory.
In [39]:
from pyspark import SparkContext
sc = SparkContext.getOrCreate()
In [73]:
uin = sc.textFile("./UINCardData.csv")
In [74]:
uin.take(5)
Out[74]:
['UIN00001,982120000,Male,N,65000', 'UIN00002,982120001,Male,N,35000', 'UIN00003,982120002,Male,N,34000', 'UIN00004,982120003,Male,Y,44000', 'UIN00005,982120004,Male,N,54000']
In [75]:
from collections import namedtuple
UINCARD = namedtuple('UINCARD', ['UIN', 'MobileNumber'])
In [76]:
uin = uin.map(lambda line:line.split(",")).map(lambda r : (r[1], UINCARD(r[0], r[1])))
In [77]:
uin.take(5)
Out[77]:
[('982120000', UINCARD(UIN='UIN00001', MobileNumber='982120000')), ('982120001', UINCARD(UIN='UIN00002', MobileNumber='982120001')), ('982120002', UINCARD(UIN='UIN00003', MobileNumber='982120002')), ('982120003', UINCARD(UIN='UIN00004', MobileNumber='982120003')), ('982120004', UINCARD(UIN='UIN00005', MobileNumber='982120004'))]
In [90]:
bankacc = sc.textFile("BankAccountLink.csv")
In [91]:
print(bankacc.count())
bankacc.take(5)
500
Out[91]:
['982120000,Y,20004562111', '982120001,Y,20004562112', '982120002,Y,20004562113', '982120003,Y,20004562114', '982120004,Y,20004562115']
In [92]:
bankacc = bankacc.filter(lambda r: r.split(",")[1] == 'N')
In [94]:
print(bankacc.count())
bankacc.take(5)
53
Out[94]:
['982120006,N,20004552111', '982120018,N,20004552112', '982120019,N,20004552113', '982120020,N,20004552114', '982120021,N,20004552115']
In [95]:
BANKACC = namedtuple('BANKACC', ['MobileNumber', 'LinkedtoBankAccount', 'BankAccountNumber'])
In [96]:
bankacc = bankacc.map(lambda line:line.split(",")).map(lambda r : (r[0], BANKACC(r[0], r[1], r[2])))
In [97]:
bankacc.take(5)
Out[97]:
[('982120006', BANKACC(MobileNumber='982120006', LinkedtoBankAccount='N', BankAccountNumber='20004552111')), ('982120018', BANKACC(MobileNumber='982120018', LinkedtoBankAccount='N', BankAccountNumber='20004552112')), ('982120019', BANKACC(MobileNumber='982120019', LinkedtoBankAccount='N', BankAccountNumber='20004552113')), ('982120020', BANKACC(MobileNumber='982120020', LinkedtoBankAccount='N', BankAccountNumber='20004552114')), ('982120021', BANKACC(MobileNumber='982120021', LinkedtoBankAccount='N', BankAccountNumber='20004552115'))]
In [98]:
result = uin.join(bankacc)
In [99]:
result.collect()
Out[99]:
[('982120006', (UINCARD(UIN='UIN00007', MobileNumber='982120006'), BANKACC(MobileNumber='982120006', LinkedtoBankAccount='N', BankAccountNumber='20004552111'))), ('982120020', (UINCARD(UIN='UIN00021', MobileNumber='982120020'), BANKACC(MobileNumber='982120020', LinkedtoBankAccount='N', BankAccountNumber='20004552114'))), ('982120048', (UINCARD(UIN='UIN00049', MobileNumber='982120048'), BANKACC(MobileNumber='982120048', LinkedtoBankAccount='N', BankAccountNumber='20004552124'))), ('982120049', (UINCARD(UIN='UIN00050', MobileNumber='982120049'), BANKACC(MobileNumber='982120049', LinkedtoBankAccount='N', BankAccountNumber='20004552125'))), ('982120297', (UINCARD(UIN='UIN00298', MobileNumber='982120297'), BANKACC(MobileNumber='982120297', LinkedtoBankAccount='N', BankAccountNumber='20004552912'))), ('982120301', (UINCARD(UIN='UIN00302', MobileNumber='982120301'), BANKACC(MobileNumber='982120301', LinkedtoBankAccount='N', BankAccountNumber='20004552916'))), ('982120313', (UINCARD(UIN='UIN00314', MobileNumber='982120313'), BANKACC(MobileNumber='982120313', LinkedtoBankAccount='N', BankAccountNumber='20004552928'))), ('982120324', (UINCARD(UIN='UIN00325', MobileNumber='982120324'), BANKACC(MobileNumber='982120324', LinkedtoBankAccount='N', BankAccountNumber='20004552939'))), ('982120018', (UINCARD(UIN='UIN00019', MobileNumber='982120018'), BANKACC(MobileNumber='982120018', LinkedtoBankAccount='N', BankAccountNumber='20004552112'))), ('982120019', (UINCARD(UIN='UIN00020', MobileNumber='982120019'), BANKACC(MobileNumber='982120019', LinkedtoBankAccount='N', BankAccountNumber='20004552113'))), ('982120021', (UINCARD(UIN='UIN00022', MobileNumber='982120021'), BANKACC(MobileNumber='982120021', LinkedtoBankAccount='N', BankAccountNumber='20004552115'))), ('982120022', (UINCARD(UIN='UIN00023', MobileNumber='982120022'), BANKACC(MobileNumber='982120022', LinkedtoBankAccount='N', BankAccountNumber='20004552116'))), ('982120030', (UINCARD(UIN='UIN00031', MobileNumber='982120030'), BANKACC(MobileNumber='982120030', LinkedtoBankAccount='N', BankAccountNumber='20004552118'))), ('982120033', (UINCARD(UIN='UIN00034', MobileNumber='982120033'), BANKACC(MobileNumber='982120033', LinkedtoBankAccount='N', BankAccountNumber='20004552121'))), ('982120046', (UINCARD(UIN='UIN00047', MobileNumber='982120046'), BANKACC(MobileNumber='982120046', LinkedtoBankAccount='N', BankAccountNumber='20004552122'))), ('982120047', (UINCARD(UIN='UIN00048', MobileNumber='982120047'), BANKACC(MobileNumber='982120047', LinkedtoBankAccount='N', BankAccountNumber='20004552123'))), ('982120052', (UINCARD(UIN='UIN00053', MobileNumber='982120052'), BANKACC(MobileNumber='982120052', LinkedtoBankAccount='N', BankAccountNumber='20004552128'))), ('982120300', (UINCARD(UIN='UIN00301', MobileNumber='982120300'), BANKACC(MobileNumber='982120300', LinkedtoBankAccount='N', BankAccountNumber='20004552915'))), ('982120309', (UINCARD(UIN='UIN00310', MobileNumber='982120309'), BANKACC(MobileNumber='982120309', LinkedtoBankAccount='N', BankAccountNumber='20004552924'))), ('982120310', (UINCARD(UIN='UIN00311', MobileNumber='982120310'), BANKACC(MobileNumber='982120310', LinkedtoBankAccount='N', BankAccountNumber='20004552925'))), ('982120312', (UINCARD(UIN='UIN00313', MobileNumber='982120312'), BANKACC(MobileNumber='982120312', LinkedtoBankAccount='N', BankAccountNumber='20004552927'))), ('982120315', (UINCARD(UIN='UIN00316', MobileNumber='982120315'), BANKACC(MobileNumber='982120315', LinkedtoBankAccount='N', BankAccountNumber='20004552930'))), ('982120316', (UINCARD(UIN='UIN00317', MobileNumber='982120316'), BANKACC(MobileNumber='982120316', LinkedtoBankAccount='N', BankAccountNumber='20004552931'))), ('982120317', (UINCARD(UIN='UIN00318', MobileNumber='982120317'), BANKACC(MobileNumber='982120317', LinkedtoBankAccount='N', BankAccountNumber='20004552932'))), ('982120318', (UINCARD(UIN='UIN00319', MobileNumber='982120318'), BANKACC(MobileNumber='982120318', LinkedtoBankAccount='N', BankAccountNumber='20004552933'))), ('982120320', (UINCARD(UIN='UIN00321', MobileNumber='982120320'), BANKACC(MobileNumber='982120320', LinkedtoBankAccount='N', BankAccountNumber='20004552935'))), ('982120325', (UINCARD(UIN='UIN00326', MobileNumber='982120325'), BANKACC(MobileNumber='982120325', LinkedtoBankAccount='N', BankAccountNumber='20004552940'))), ('982120031', (UINCARD(UIN='UIN00032', MobileNumber='982120031'), BANKACC(MobileNumber='982120031', LinkedtoBankAccount='N', BankAccountNumber='20004552119'))), ('982120051', (UINCARD(UIN='UIN00052', MobileNumber='982120051'), BANKACC(MobileNumber='982120051', LinkedtoBankAccount='N', BankAccountNumber='20004552127'))), ('982120053', (UINCARD(UIN='UIN00054', MobileNumber='982120053'), BANKACC(MobileNumber='982120053', LinkedtoBankAccount='N', BankAccountNumber='20004552129'))), ('982120055', (UINCARD(UIN='UIN00056', MobileNumber='982120055'), BANKACC(MobileNumber='982120055', LinkedtoBankAccount='N', BankAccountNumber='20004552131'))), ('982120298', (UINCARD(UIN='UIN00299', MobileNumber='982120298'), BANKACC(MobileNumber='982120298', LinkedtoBankAccount='N', BankAccountNumber='20004552913'))), ('982120299', (UINCARD(UIN='UIN00300', MobileNumber='982120299'), BANKACC(MobileNumber='982120299', LinkedtoBankAccount='N', BankAccountNumber='20004552914'))), ('982120304', (UINCARD(UIN='UIN00305', MobileNumber='982120304'), BANKACC(MobileNumber='982120304', LinkedtoBankAccount='N', BankAccountNumber='20004552919'))), ('982120306', (UINCARD(UIN='UIN00307', MobileNumber='982120306'), BANKACC(MobileNumber='982120306', LinkedtoBankAccount='N', BankAccountNumber='20004552921'))), ('982120307', (UINCARD(UIN='UIN00308', MobileNumber='982120307'), BANKACC(MobileNumber='982120307', LinkedtoBankAccount='N', BankAccountNumber='20004552922'))), ('982120314', (UINCARD(UIN='UIN00315', MobileNumber='982120314'), BANKACC(MobileNumber='982120314', LinkedtoBankAccount='N', BankAccountNumber='20004552929'))), ('982120319', (UINCARD(UIN='UIN00320', MobileNumber='982120319'), BANKACC(MobileNumber='982120319', LinkedtoBankAccount='N', BankAccountNumber='20004552934'))), ('982120029', (UINCARD(UIN='UIN00030', MobileNumber='982120029'), BANKACC(MobileNumber='982120029', LinkedtoBankAccount='N', BankAccountNumber='20004552117'))), ('982120032', (UINCARD(UIN='UIN00033', MobileNumber='982120032'), BANKACC(MobileNumber='982120032', LinkedtoBankAccount='N', BankAccountNumber='20004552110'))), ('982120050', (UINCARD(UIN='UIN00051', MobileNumber='982120050'), BANKACC(MobileNumber='982120050', LinkedtoBankAccount='N', BankAccountNumber='20004552126'))), ('982120054', (UINCARD(UIN='UIN00055', MobileNumber='982120054'), BANKACC(MobileNumber='982120054', LinkedtoBankAccount='N', BankAccountNumber='20004552130'))), ('982120056', (UINCARD(UIN='UIN00057', MobileNumber='982120056'), BANKACC(MobileNumber='982120056', LinkedtoBankAccount='N', BankAccountNumber='20004552132'))), ('982120296', (UINCARD(UIN='UIN00297', MobileNumber='982120296'), BANKACC(MobileNumber='982120296', LinkedtoBankAccount='N', BankAccountNumber='20004552911'))), ('982120302', (UINCARD(UIN='UIN00303', MobileNumber='982120302'), BANKACC(MobileNumber='982120302', LinkedtoBankAccount='N', BankAccountNumber='20004552917'))), ('982120303', (UINCARD(UIN='UIN00304', MobileNumber='982120303'), BANKACC(MobileNumber='982120303', LinkedtoBankAccount='N', BankAccountNumber='20004552918'))), ('982120305', (UINCARD(UIN='UIN00306', MobileNumber='982120305'), BANKACC(MobileNumber='982120305', LinkedtoBankAccount='N', BankAccountNumber='20004552920'))), ('982120308', (UINCARD(UIN='UIN00309', MobileNumber='982120308'), BANKACC(MobileNumber='982120308', LinkedtoBankAccount='N', BankAccountNumber='20004552923'))), ('982120311', (UINCARD(UIN='UIN00312', MobileNumber='982120311'), BANKACC(MobileNumber='982120311', LinkedtoBankAccount='N', BankAccountNumber='20004552926'))), ('982120321', (UINCARD(UIN='UIN00322', MobileNumber='982120321'), BANKACC(MobileNumber='982120321', LinkedtoBankAccount='N', BankAccountNumber='20004552936'))), ('982120322', (UINCARD(UIN='UIN00323', MobileNumber='982120322'), BANKACC(MobileNumber='982120322', LinkedtoBankAccount='N', BankAccountNumber='20004552937'))), ('982120323', (UINCARD(UIN='UIN00324', MobileNumber='982120323'), BANKACC(MobileNumber='982120323', LinkedtoBankAccount='N', BankAccountNumber='20004552938'))), ('982120326', (UINCARD(UIN='UIN00327', MobileNumber='982120326'), BANKACC(MobileNumber='982120326', LinkedtoBankAccount='N', BankAccountNumber='20004552941')))]
In [ ]: