AI vs. IoT: What’s the Difference?

No alt text provided for this image

While Artificial Intelligence (AI) and the Internet of Things (IoT) are both hot topics, they are not the same. They have differences but at the same time are connected and related. Artificial intelligence is a type of science that works to imitate intelligent behavior in computers.  Internet of things is the internet-working of devices like homes, sensors, cars, and home appliances that can communicate together and often with the external environment like other cars, devices and human beings. 

Some of the differences between AI and IoT include interaction with cloud computing, scalability, cost, and ability to learn from data. For example, with cloud computing, IoT generates significant amounts of data and cloud computing provides a pathway for that data. On the other hand, AI intersects with cloud computing in the sense that it allows the devices to act and react in a way more similar to the human experience. 

In terms of learning from data, in IoT, there can be multiple sensors, and each has some set of processes where identical information is shared on the internet, but in AI, the system actual learns from the activities or errors occurring to try to evolve into a better version of itself. As it relates to cost, IoT generally costs much less than $50K USD with all components involved from hardware to infrastructure, whereas with AI the charges are typically are calculated for each case and can vary substantially based on complexity and industry. 

No alt text provided for this image

IoT focuses on connecting machines and making use of the collected data while AI is about mimicking intelligent behavior in machines. As the devices powered by IoT continue to grow, AI can help by dealing with the big data by making sense of it. That being said, IoT can exist without AI. And, AI can exist without IoT. But, data is only useful to humans if it creates insights that can be acted upon. Using IoT and AI together create connected intelligence. 

A use case of IoT and AI working together is Tesla Motors self-driving cars. In this example, the car is the “thing,” and the power of AI is used to predict the behavior of a car in a variety of environments. The Telsa cars operate as a network meaning that when one car learns something, all the cars can then learn something.  

No alt text provided for this image

There are several data scientists that believe the future of IoT is in the AI. Undoubtably, when the two are combined the value delivered can increase for the customer, as well as the organization.

#BigData #IoT #AIA

Cure Cancer: AI and Machine Learning

There are several ways that machine learning tools can be used on existing data sets to potentially discover a cure for cancer. First, anybody can download the tools for free nearly anywhere in the world with a consistent internet connection. One of my favorite programs is R that works on both Windows and Mac machines and installs in a matter of minutes. I particularly like R because of the machine learning libraries in it that can be leveraged in programming. While I previously shared some general machine learning algorithms, in this post, I am going to take it a little deeper for those that do have a technical background and want to expand their toolkit and experiment with some of these machine learning techniques.

The first step is understanding what variables you might have access to as it relates to cancer and the nature of those variables. A variety of both structured and unstructured data can be combined in frameworks like Hadoop to prepare the data for analysis. If you want to leverage different machine learning techniques, it is useful to understand how trees work because with decision trees there is not the assumption of linearity which is helpful when trying to glean insights through non-linear data analysis.

No alt text provided for this image

Classification trees help separate data into classes that belong to the response variable. If the target variable has more than two categories, different variants of the algorithm are leveraged, but overall classification trees are useful when the target variable is categorical (like yes/no).  On the other hand, regression trees or prediction trees can be useful when a response variable is numeric or continuous. The target variable determines whether or not to use classification or regression tree. Conditional logistic regression can be useful in tackling sparse data type issues.

   The advantages of decision trees include fast computations, invariance under the monotone transformation of variables, an easy extension to categorical outcomes, resistance to irrelevant variables, one tuning parameter, ability to handle missing data and outputs that can be easily understood by non-technical audiences. The disadvantages can include accuracy since the function needs to involve higher order interactions and variance since each split depends on previous splits and small changes can cause big changes in the decision tree. Some important definitions to understand include:

  • Root is the topmost node of the tree
  • Edge is the link between two nodes
  • Child is a node that has a parent node
  • Parent is a node that has an edge to a child node
  • Leaf is a node that does not have a child node in the tree
  • Height is the length of the longest path to a leaf
  • Depth is the length of the path to its root
No alt text provided for this image

Let’s start with considering the existing prostate cancer data set available in R. The data represents a population of 97 males. This is a good data set to illustrate how easily different tree growth algorithms and classification techniques can be used to predict tumor spread in males. In this specific example, the measures for prediction are PSA, the size of the prostate, benign prostatic hyperplasia, Gleason score, and capsular penetration. Therefore, to better understand and predict the tumor spread (seminar vesicle invation=svi) the following variables were used for the tree growth algorithms: log of benign prostatic hyperplasia amount (lbph), log of prostate-specific antigen (lpsa), Gleason score (gleason), log of capsular penetration (lcp) and log of cancer volume (lcavol).

Here is a quick program that I wrote in R to better understand this data set:

R Script

# Loading the proper libraries to conduct this analysis on the prostate cancer dataset in R
# Setting up the classification tree
# Lets look at the results
# Plotting the results
# Making the plot tree
plot(classification,uniform=T,main="Classification tree for prostate cancer")
text(classification,use.n = T, all=T, cex=.8)
# Making the tree
# Looking at the results
plot(regression,uniform=T,main="Regression tree for prostate cancer")
text(regression,use.n = T, all=T,cex=.8)
# Now doing the conditional inference tree
# Lets look at the results
# Plotting the results
plot(conditional,main="Conditional inference tree for prostate cancer")

This script resulted in the following information:

> printcp(classification)


Classification tree:

rpart(formula = svi ~ lbph + lpsa + gleason + lcp + lcavol, data = Prostate,

    method = "class")


Variables actually used in tree construction:

[1] lcp


Root node error: 21/97 = 0.21649


n= 97


       CP nsplit rel error  xerror    xstd

1 0.52381      0   1.00000 1.00000 0.19316

2 0.01000      1   0.47619 0.80952 0.17831

Regression tree:

rpart(formula = svi ~ lbph + lpsa + gleason + lcp + lcavol, data = Prostate, 
    method = "anova")

Variables actually used in tree construction:
[1] lcp  lpsa

Root node error: 16.454/97 = 0.16962

n= 97 

       CP nsplit rel error  xerror    xstd
1 0.45551      0   1.00000 1.00780 0.14079
2 0.21489      1   0.54449 0.68052 0.15327
3 0.01000      2   0.32960 0.53091 0.11726

> conditional

Conditional inference tree with 3 terminal nodes

Response:  svi 
Inputs:  lbph, lpsa, gleason, lcp, lcavol 
Number of observations:  97 

1) lcp <= 1.7492; criterion = 1, statistic = 43.496
  2) lpsa <= 2.972975; criterion = 1, statistic = 20.148
    3)*  weights = 66 
  2) lpsa > 2.972975
    4)*  weights = 18 
1) lcp > 1.7492
  5)*  weights = 13 
No alt text provided for this image

Note that head node is the seminal vesicle invasion which shows the tumor spread. The cross-validation results show there is only one split in the three with a relative value for the first split of .80952 and a standard deviation of .17831. The log of capsular penetration was used to split the tree when the log of capsular penetration at <1.791. There were three leaf nodes in the regression tree algorithm because the script split the data set three times. The relative error for the first split was 0.68052, and a standard deviation of 0.15327 and at the second split the relative error is 0.53091 and a standard deviation of 0.11726. The tree was split at the first log of capsular penetration at <1.791 and the log of the prostate-specific antigen at < 2.973. The conditional tree algorithm produced a split at <1.749 of the log of capsular penetration at the 0.001 significance level and <2.973 for the log of prostate-specific antigen also at the 0.001 significance level.

In this particular example, the condition tree growth algorithm produced more useful information than the classification and regression tree growth algorithm. That being said, while sometimes the language as it relates to machine learning is complicated to understand, it really just comes down to using the right variables as input and testing different machine learning algorithms relative to the problem being solved. Testing different machine learning algorithms boils down to running a few lines of code in R, python or your favorite programming language.

Clinical data around pathology related detail, tumor evolution and cell-level information is being generated at exponentially increasing levels. Many of these data sets are starting to be available online for analysis. The type of algorithms used in this example could be used on these big data sets to accelerate the discovery of a cure for cancer. But, it is not going to happen without individuals that are willing to embrace these types of tools for analysis.

#BigData #AI #Oncology #MachineLearning

Ensembles and Random Forest Analysis: How it Works

Ensemble methods can use multiple machine learning algorithms to predict performance. Ensemble is essentially about combining methods to have better predictions.  For example, in terms of logistic regression with ensemble classification, if the first classifier is a base classifier and the second is a corrector classifier, then the first does the initial classification, and the predicted class is then fed into the feature of the second classifier.  The second classifier can either result in a classification which is identical to the first or can correct the prediction if more accuracy is found.  The base classifier helps with the initial prediction of the target class.  The corrector classifier attempts to correct any errors in the prediction of the base classifier by focusing on the decision boundary of the base classifier.  For example, a choice of the base classifier could be logistic regression.  Logistic regression is a parametric discriminative classifier that can be used for training.  Also, for a corrector, the k-nearest neighbors can be the parametric classifier which would take the average of k nearest training data to make the prediction.

Random Forest is a type of ensemble method that performs both regression and classification with the use of multiple decision trees.  The technique is often referred to as Bootstrap Aggregation.  The Bootstrap Aggregation method involves training each decision tree on a different random.  The sampling in this instance occurs through replacement.

AI versus Big Data: What’s the Difference?

Artificial intelligence is fueled by computers, big data, and algorithms. Big data is the input for business intelligence capabilities. Big data represents the large volume of data that often needs to go through a data quality process of cleansing before it can be turned into business insights. Artificial intelligence, on the other hand, occurs when computers act on that data input. Artificial intelligence changes behavior based on findings and then modifies the approach. Big data analytics are more about looking for a given piece of data to produce insight versus having the computer act on the results that are found. Big data analytics produces insights through the identification of patterns through things like the sequential analysis, leveraging technologies like Hadoop that can analyze both structured and unstructured data. While artificial intelligence can also be based off structured and unstructured data, with artificial intelligence, the computer learns from that big data and keeps collecting it and then acting upon it.

Industry examples of how big data is being leveraged in artificial intelligence range from consumer goods to the creative arts to media. For example, in consumer goods, Hello Barbie runs off of machine learning where the microphone on Barbie’s necklace records what the child says and analyzes it to determine a fitting response. The server gets the response back to Barbie in under a second. In the creative arts, music-generating algorithms are being used from newspapers and speeches to create themes for new lyrics and help musicians better understand target audiences to increase record sales. In media, the BBC project, Talking with Machines lets listeners engage in conversation with their smart devices to insert their perspective to become part of the story creation.

Artificial intelligence influences big data analytics and vice-versa. Artificial Intelligence uses big data to run algorithms, like machine learning algorithms. In machine learning algorithms, training and test datasets are used for the analysis.  Big data analytics can be useful to prepare those test and training datasets for machine learning. Also, access to big data allows artificial intelligence to continue to learn more additional data sources. Machine learning algorithms can reproduce behaviors based on big data that is feeding processors that it puts through a trial and error type of algorithms. 

Essentially big data is what can teach artificial intelligence, and the rise of artificial intelligence is complementary to the exponential growth of big data. Understanding the basics of how big data and artificial intelligence intersect is important as they are both here to stay and have the potential to boost, not only revenue but innovative and creative capabilities for businesses.

#AI #BigData

Will AI Replace Humans?

Should artificial intelligence be used as a tool to support or replace decision makers? After all, decision making relates to reasoning. Fifty-two million American jobs will be displaced or changed due to automation by 2030. While the changing nature of work causes some anxiety, the machines are just acting human, not actually human. And, while technology eliminates some jobs, it does not eliminate work.

Artificial intelligence can be used as a tool to support decision makers. But, technology empowered by artificial intelligence definitely does not eliminate the need for governance and ethics as it relates to the social good. Humans have the unique ability to create a vision and plan to achieve it. The strength of artificial intelligence lies in data processing, not in complex judgment and decision making. However, artificial intelligence is complementary to complex decision making.  Organizations should be asking themselves how computers can support humans in solving complex problems? For example, AI for Good is an initiative that focuses on the United Nations platform sharing the beneficial use of artificial intelligence projects to solve some of society’s biggest challenges. 

No alt text provided for this image

While the goals of some other artificial intelligence initiatives are to generate a software program that can solve complex problems and moderate itself with thoughts and emotions similar to human intelligence, it is important to understand the limitations of this scientific pursuit. There are many philosophical challenges in executing this intent from how to freedom is defined to how values are determined and understanding is measured. The challenge in pursuing these type of initiatives is in the programming. Artificially intelligent systems create their own rules upon existing rules and cannot deviate from them or make random decisions which in turn makes it difficult for the system to gain understanding similar to the human experience. Numerous studies have shown that free will influences mental processing and intelligence. In terms of artificial intelligence breakthroughs, although there have been some wins like in 2012 scenario when one of Google’s supercomputers scanned 10 million YouTube videos and learned to identify a cat with 75% accuracy, a four-year-old does that experience flawlessly, and it is not exactly tackling the issues of culture, vision or values as it relates to complex decision making. In summary, we are far cry from what you might believe from watching an episode of West World.

No alt text provided for this image

Humans have a history of adapting and thriving when new kinds of work have emerged in society so even defining the human experience is a changing goal post for any programmer to master. A simple walk through an art museum reflects how complicated the human perception of reality over time is to mimic let alone predict.

One of the challenges that artificial intelligence faces is that is it developing at a rate faster than some social systems which is why there is increasing interest in ethics and public policy as it relates to artificial intelligence. Also, some of the data input being used to drive machine learning programs is not reflective of the communities that the programs ultimately seek to serve. However, much of this issue is simply reflective of an age-old data quality issue in programming that poor data sources result in weak data outputs that in turn can lead to poor decision making. Bad data has been estimated to cost the United States $3 trillion per year.

No alt text provided for this image

One potential takeaway from this rapidly evolving digital economy is that a purpose driven life is uniquely human. And, the purpose each of us finds in terms of living a meaningful existence, comes from a complex understanding of where we’ve been and where we are going, along with some seemingly random but transformative events along the way. Regardless, those that embrace artificial intelligence to tackle problems with purpose are likely to create more impact than those that reject these innovative technologies.

#Purpose #HumanExperience #AIforGood

The Future of Artificial Intelligence

One of the future challenges of big data analytics for artificial intelligence includes the role that it plays relative to human judgment.  For example, it was found that human parole boards do much worse than machines powered by artificial intelligence in calculating which dangerous prisoners should be entered back into society.  Similarly, skilled pathologists were not able to beat machine learning algorithms as it related to the diagnosis of breast cancer. Banks are currently delivering advice to wealthy clients using artificial intelligence from a Singapore-based technology company.  In a William Grove study covering 136 different research studies, expert judgment was found to be better than machine learning equivalents in only eight studies.

Business can access more data than ever before but research has found that organizations struggle still to see the bigger picture in terms of organizational priorities. Proper framing and focus relative to the problem that is attempted to be solved has found to be critical earlier on in the process as large data sets continue to grow. Organizations must take a step back to think about what is needed before digging deeper and getting lost in the weeds.

A vital concern as big data continues to grow is that there is still a gap in the literature as it relates to the role of leadership in big data governance effectiveness.  Also, with more pressure likely to increase from the public as the nature of work continues to change, those serving in governing roles are under more scrutiny than ever before.  Despite numerous peer review research findings that have indicated how essential high-level support is to effective big data governance, many governing bodies still do have the necessary knowledge to govern effectively.  However, it does not take a technical mind to understand that the output of machine learning programs is only as good as the inputs.  Inclusion is critical from the beginning especially in representing marginalized voices.  While the future of artificial intelligence certainly will lead to business efficiencies and improved customer service experience, there still is a role for a human touch reflective of our shared vision and values.

#AI #BigData #MachineLearning