machine learning
Good Article on Why AI Projects Fail

Today I ran across this article that was very good as it focused on lessons learned, which potentially helps everyone interested in these topics. It contained a good mix of problems at a non-technical level.
Below is the link to the article, as well as commentary on the Top 3 items listed from my perspective.
https://www.cio.com/article/3429177/6-reasons-why-ai-projects-fail.html
Item #1:
The article starts by discussing how the “problem” being evaluated was misstated using technical terms. It led me to believe that at least some of these efforts are conducted “in a vacuum.” That was a surprise given the cost and strategic importance of getting these early-adopter AI projects right.
In Sales and Marketing you start the question, “What problem are we trying to solve?” and evolve that to, “How would customers or prospects describe this problem in their own words?” Without that understanding, you can neither initially vet the solution nor quickly qualify the need for your solution when speaking with those customers or prospects. That leaves a lot of room for error when transitioning from strategy to execution.
Increased collaboration with Business would likely have helped. This was touched on at the end of the article under “Cultural challenges,” but the importance seemed to be downplayed. Lessons learned are valuable – especially when you are able to learn from the mistakes of others. To me, this should have been called out early as a major lesson learned.
Item #2:
This second area had to do with the perspective of the data, whether that was the angle of the subject in photographs (overhead from a drone vs horizontal from the shoreline) or the type of customer data evaluated (such as from a single source) used to train the ML algorithm.
That was interesting because it appears that assumptions may have played a part in overlooking other aspects of the problem, or that the teams may have been overly confident about obtaining the correct results using the data available. In the examples cited those teams did figure those problems out and took corrective action. A follow-on article describing the process used to make their root cause determination in each case would be very interesting.
As an aside, from my perspective, this is why Explainable AI is so important. There are times that you just don’t know what you don’t know (the unknown unknowns). Being able to understand why and on what the AI is basing its decisions should help with providing better quality curated data up-front, as well as being able to identify potential drifts in the wrong direction while it is still early enough to make corrections without impacting deadlines or deliverables.
Item #3:
This didn’t surprise me but should be a cause for concern as advances are made at faster rates and potentially less validation is made as organizations race to be first to market with some AI-based competitive advantage. The last paragraph under ‘Training data bias’ stated that based on a PWC survey, “only 25 percent of respondents said they would prioritize the ethical implications of an AI solution before implementing it.”
Bonus Item:
The discussion about the value of unstructured data was very interesting, especially when you consider:
- The potential for NLU (natural language understanding) products in conjunction with ML and AI.
- This is a great NLU-pipeline diagram from North Side Inc. in Canada, one of the pioneers in this space.
- The importance of semantic data analysis relative to any ML effort.
- The incredible value that products like MarkLogic’s database or Franz’s AllegroGraph provide over standard Analytics Database products.
- I personally believe that the biggest exception to assertion this will be from GPU databases (like OmniSci) that easily handle streaming data, can accomplish extreme computational feats well beyond those of traditional CPU based products, and have geospatial capabilities that provide an additional dimension of insight to the problem being solved.
Update: This is a link to a related article that discusses trends in areas of implementation, important considerations, and the potential ROI of AI projects: https://www.fastcompany.com/90387050/reduce-the-hype-and-find-a-plan-how-to-adopt-an-ai-strategy
This is definitely an exciting space that will experience significant growth over the next 3-5 years. The more information, experiences, and lessons learned shared the better it will be for everyone.
Discussions that Seed the Roots of Creativity
A few months ago I purchased Fitbit watches for my children and myself. My goals were twofold. First, I was hoping that they would motivate all of us to be more active. Second, I wanted to foster a sense of competition (including fair play and winning) within my children. Much of their pre-High School experiences focused on “participation,” as many schools feel that competition is bad. Unfortunately, competition is everywhere in life, so if don’t play to win you may not get the opportunity to play at all.
It is fun seeing them push to be the high achiever for the day, and to continually push themselves to do better week-by-week and month-by-month. I believe this creates a wonderful mindset that makes you want to do more, learn more, achieve more, and make an overall greater impact with everything they do. People who do that are also more interesting to spend time with, so that is a bonus.
Recently my 14 year-old son and I went for a long walk at night. It was a cold, windy, and fairly dark night. We live in fairly rural area so it is not uncommon to see and hear various wild animals on a 3-4 mile walk. I’m always looking for opportunities to teach my kids things in a way that is fun and memorable, and in a way that they don’t realize they are being taught. Retention of the concepts is very high when I am able to make it relevant to something we are doing.
That night we started talking about the wind. It was steady with occasional gusts, and at times it changed direction slightly. I pointed out the movement on bushes and taller grass on the side of the road. We discussed direction, and I told him to think about the wind like an invisible arrow, and then explained how those arrows traveled in straight lines or vectors until they met some other object. We discussed which object would “win,” and how the force of one object could impact another object. My plan was to discuss Newton’s three laws of motion.
My son asked if that is why airplanes sometimes appear to be flying at an angle but are going straight. He seemed to be grasping the concept. He then asked me if drones would be smart enough to make those adjustments, which quickly led to me discussing the use potential future of “intelligent” AI-based drones by the military. When he was 9 he wanted to be a Navy SEAL, but once he saw how much work that was he decided that he would rather be Transformer (which I explained was not a real thing). My plan was to use this example to discuss robotics and how you might program a robot to do various tasks, and then move to how it could learn from the past tasks and outcomes. I wanted him to logically break down the actions and think about managing complexity. But, no such luck that night.
His mind jumped to “Terminator” and “I, Robot.” I pointed out that Science Fiction does occasionally become Science Fact, which makes this type of discussion even more interesting. I also pointed out that there is spectrum between the best possible outcome – Utopia, and the worst possible outcome – Dystopia, and asked him what he thought could happen if machines could learn and become smarter on their own.
His response was that things would probably fall somewhere in the middle, but that there would be people at each end trying to pull the technology in their direction. That seemed like a very enlightened estimation. He asked me what I thought and I replied that I agreed with him. I then noted how some really intelligent guys like Stephen Hawking and Elon Musk are worried about the dystopian future and recently published a letter to express their concerns about potential pitfalls of AI (artificial intelligence). This is where the discussion became really interesting…
We discussed why you would want a program or a robot to learn and improve – so that it could continue to become better and more efficient, just like a person. We discussed good and bad, and how difficult it could be to control something that doesn’t have morals or understand social mores (which he felt if this robot was smart enough to learn on its own that it would also learn those things based on observations and interactions). That was an interesting perspective.
I told him about my discussions with his older sister, who wants to become a Physician, about how I believe that robotics, nanotechnology, and pharmacology will become the future of medicine. He and I took the logical next step and thought about a generic but intelligent medicine that identified and fixed problems independently, and then sent the data and lessons learned for others to learn from.
I’m sure that we will have an Internet of Things (IoT) discussion later, but for now I will tie this back to our discussion and Fitbit wearable technology.
After the walk I was thinking about what just happened, and was pleased because it seemed to spark some genuine interest in him. I’m always looking for that perfect recipe for innovation, but it is elusive and so far lacks repeatability. It may be possible to list many of the “ingredients” (intelligence, creativity, curiosity, confidence (to try and accept and learn from failure), multi-disciplinary experiences and expertise) and “measurements” (such as a mix of complementary skills, a mix of roles, and a special environment (i.e., strives to learn and improve, rewards both learning and success but doesn’t penalize failure, and creates a competitive environment that understands that in most cases the team is more important than any one individual)).
That type of environment is magical when you can create it, but it takes so much more than just having people and a place that seem to match the recipe. A critical “activation” component or two is missing. Things like curiosity, creativity, ingenuity, and a bit of fearlessness.

I tend to visualize things, so while I was thinking about this I pictured a tree with multiple “brains” (my mental image looked somewhat like broccoli) that had visible roots. Those roots were creative ideas that went off in various directions. Trees with more roots that were bigger and went deeper would stand out in a forest of regular trees.
Each major branch (brain/person) would have a certain degree of independence, but ultimately everything on the tree worked as a system. To me, this description makes so much more sense than the idea of a recipe, but it still doesn’t bring me closer to being able map the DNA of this imaginary tree.
At the end of our long walk it seemed that I probably learned as much as my son did. We made a connection that will likely lead to more walks and more discussions.
And in a strange way, I can thank the purchase of these Fitbit watches for being the motivation for an activity that led to this amazing discussion. From that perspective alone this was money well spent.
Ideas are sometimes Slippery and Hard to Grasp
I started this blog the goal of it becoming an “idea exchange,” as well a way to pass along lessons learned to help others. Typical guidance for a blog is to focus on one thing only and do it well in order to develop a following. That is especially important if you want to monetize the blog, but that is not and has not been my goal.
One of the things that has surprised me is how different the comments and likes are for each post. Feedback from the last post was even more diverse and surprising than usual. It ranged from comments about “Siri vs Google,” to feedback about Sci-Fi books and movies, to Artificial Intelligence.
I asked a few friends for feedback and received something very insightful (Thanks Jim). He stated that he found the blog interesting, but wasn’t sure what the objective was. He went on to identify several possible goals for the last post. Strangely enough, his comments mirrored the type of feedback that I received. That pointed out an area for improvement to me, and I appreciated that, as well as the wisdom of focusing on one thing. Who knows, maybe in the future…
This also reminded me of a white paper written 12-13 years ago by someone I used to work with. It was about how Bluetooth was going to be the “next big thing.” He had read an IEEE paper or something and saw potential for this new technology. His paper provided the example of your toaster and coffee maker communicating so that your breakfast would be ready when you walk into the kitchen in the morning.
At that time I had a couple of thoughts. Who cared about something that only had a 20-30 foot range when WiFi was becoming popular and had much greater range? In addition, a couple of years earlier I had a tour of the Microsoft “House of the Future,” in which everything was automated and key components communicated with each other. But everything in the house was all hardwired or used WiFi – not Bluetooth. It was easy to dismiss his assertion because it seemed to lack pragmatism, and the value of the idea was difficult to quantify given the use case provided.
Looking back now I view that white paper as having insight (if it were visionary he would have come out with the first Bluetooth speakers, or car interface, or even phone earpiece and gotten rich), but it failed to present use cases that were easy enough to understand yet different enough from what was available at the time to demonstrate the real value of the idea. His expression of idea was not tangible enough and therefore too slippery to be easily grasped and valued.
I’m a huge believer that good ideas sometimes originate where you least expect them. Often those ideas are incremental in nature – seemingly simple and sometimes borderline obvious, often building on some other idea or concept. An idea does not need to be unique in order to be important or valuable, but it does need to be presented in a way that is easy to understand the benefits, differentiation, and value. That is just good communication.
One of the things I miss most from when my consulting company was active was the interaction between a couple of key people (Jason and Peter) and myself. Those guys were very good at taking an idea and helping build it out. This worked well because we had some overlapping expertise and experiences as well as skills and perspectives that were more complementary in nature. That diversity increased the depth and breadth to our efforts to develop and extend those ideas by asking the tough questions early and ensuring that we could convince each other of the value.
Our discussions were creative and highly collaborative as well as a lot of fun. Each of us improved from them, and the outcome was usually something viable from a commercial perspective. As a growing and profitable small business you need to constantly innovate to differentiate yourself from your competition. Our discussions were driven as much by necessity as they were by intellectual curiosity, and I personally believe that this was part of the magic.
So, back to the last post. I view various technologies as building blocks. Some are foundational and others are complementary. To me, the key is not viewing those various technologies as competing with each other. Instead, I look for potential value created by integrating them with each other. That may not always possible and does not always lead to something better, but occasionally it does so to me it is a worthwhile exercise. With regard to voice technology, I do believe that we will see more, better and smarter applications of it – especially as realtime systems become more complex due to the use of an increasing number of specialized component systems and sensors.
While today’s smartphones would not pass the Turing Test or proposed alternatives, they are an improvement over more simplistic voice translation tools available just a few years ago. Advancement requires the tools to understand context in order to make inferences. This brings you closer to machine learning, and big data (when done right) significantly increases that potential.
Ultimately, this all leads back to Artificial Intelligence (at least in my mind). It’s a big leap from a simple voice translation tool to AI, but when viewed as building blocks it is not such a stretch.
Now think about creating an interface (API) that allows one smart device to communicate with another in a manner akin to the collaborative efforts described above with my old team. It’s not simply having a front-end device exchanging keywords or queries with a back-end device. Instead, it is two or more devices and/or systems having a “discussion” about what is being requested, looking at what each component “knows,” asking clarifying questions and making suggestions, and then finally taking that multi-dimensional understanding of the problem to determine what is really needed.
So, possibly not true AI, but a giant leap forward from what we have today. That would help turn the science fiction of the past into science fact in the near future. The better the understanding and inferences by the smart system, the better the results.
I also believe that the unintended consequences of these new smart systems is that as they become more human-like in their approach the more likely they will be to make errors like a human. Hopefully they will be able to back test recommendations to validate and minimize errors. If they are intelligent enough to monitor results and make suggestions about corrective actions when they determine that the recommendation is not having the optimal desired results would make them even “smarter.” Best of all there won’t be an ego creating a distortion filter on the results. Or maybe there will…
A lot of the building blocks required to create these new systems are available today. But, it takes both vision and insight to see that potential, translate ideas from slippery and abstract to tangible and purposeful, and then start building something really cool. As that happens we will see a paradigm shift in how we interact with computers and how they interact with us. That will lead us to the systematic integration that I wrote about in a big data / nanotechnology post.
So, what is the real objective of my blog? To get people thinking about things in a different way, to foster collaboration and partnerships between businesses and educational institutions in order to push the limits of technology, and to foster discussion about what others believe the future of computing and smart devices will look like. I’m confident that I will see these types of systems in my lifetime, and believe in the possibility of a lot of this occurring within the next decade.
What are your thoughts?
The Future of Smart Interfaces
Recently I was helping one of my children research a topic for a school paper. She was doing well, but the results she was getting were overly broad. So, I taught her some “Google-Fu,” explaining how you can structure queries in ways that yield better results. She replied that search engines should be smarter than that. I explained that sometimes the problem is that search engines look at your past searches and customize results as an attempt to appear smarter or to motivate someone to do or believe something.
Unfortunately, those results can be skewed and potentially lead someone in the wrong direction. It was a good reminder that getting the best results from search engines often requires a bit of skill and query planning, as well as occasional third-party validation.
Then the other day I saw this commercial from Motel 6 (“Gas Station Trouble”) where a man has problems getting good results from his smart phone. That reminded me of seeing someone speak to their phone, getting frustrated by the responses received. His questions went something like this:
“Siri, I want to take my wife to dinner tonight, someplace that is not too far away, and not too late. And she likes to have a view while eating so please look for something with a nice view. Oh, and we don’t want Italian food because we just had that last night.”
Just as amazing as the question being asked was watching him ask it over and over again in the exact same way, each time becoming even more frustrated. I asked myself, “Are smartphones making us dumber?” Instead of contemplating that question I began to think about what future smart interfaces would or could be like.
I grew up watching Sci-Fi computer interfaces like “Computer” on Star Trek (1966), “HAL” on 2001 : A Space Odyssey (1968), “KITT” from Knight Rider (1982), and “Samantha” from Her (2013). These interfaces had a few things in common:
- They responded to verbal commands;
- They were interactive – not just providing answers, but also asking qualifying questions and allowing for interrupts to drill-down or enhance the search (e.g., with pictures or questions that resembled verbal Venn diagrams);
- They often provided suggestions for alternate queries based on intuition. That would have been helpful for the gentleman trying to find a restaurant.
Despite having 50 years of science fiction examples, we are still a long way off from realizing that goal of a truly intelligent interface. Like many new technologies, they were originally envisioned by science fiction writers long before they appeared in science.
There seems to be a spectrum of common beliefs about modern interfaces. On one end there are products that make visualization easy, facilitating understanding, refinement and drill-down of data sets. Tableau is a great example of this type of easy to use interface. At the other end of the spectrum the emphasis is on back-end systems – robust computer systems that digest huge volumes of data and return the results to complex queries within seconds. Several other vendors offer powerful analytics platforms. In reality, you really need a strong front-end and back-end if you want to achieve the full potential of either.
But, there is so much more potential…
I predict that within the next 3 – 5 years we will see business and consumer interface examples (powered by Natural Language Processing, or NLP) that are closer to the verbal interfaces from those familiar Sci-Fi shows (albeit with limited capabilities and no flashing lights).
Within the next 10 years I believe we will have computer interfaces that intuit our needs and facilitate the generation of correct answers quickly and easily. While this is unlikely to be at the level of “The world’s first intelligent Operating System” envisioned in the movie “Her,” and probably won’t even be able to read lips like “HAL,” it should be much more like HAL and KITT than like Siri (from Apple) or Cortana (from Microsoft).
Siri was groundbreaking consumer technology when it was introduced. Cortana seems to have taken a small leap ahead. While I have not mentioned Google Now, it is somewhat of a latecomer to this consumer smart interface party, and in my opinion is behind both Siri and Cortana.
So, what will this future smart interface do? It will need to be very powerful, harnessing a natural language interface on the front-end with an extremely flexible and robust analytics interface on the back-end. The language interface will need to take a standard question (in multiple languages and dialects) – just as if you were asking a person, deconstruct it using Natural Language Processing, and develop the proper query based on the available data. That is important but only gets you so far.
Data will come from many sources – things that we consider today with relational, object, graph, and NoSQL databases. There will be structured and unstructured data that must be joined and filtered quickly and accurately. In addition, context will be more important than ever. Pictures and videos could be scanned for facial recognition, location (via geotagging), and in the case of videos analyze speech. Relationships will be identified and inferred based on a variety of sources, using both data and metadata. Sensors will collect data from almost everything we do and (someday) wear, which will provide both content and context.
The use of Stylometry will identify outside content likely related to the people involved in the query and provide further context about interests, activities, and even biases. This is how future interfaces will truly understand (not just interpret), intuit (so it can determine what you really want to know), and then present results that may be far more accurate than we are used to today. Because the interface is interactive in nature it will provide the ability to organize and analyze subsets of data quickly and easily.
So, where do I think that this technology will originate? I believe that it will be adapted from video game technology. Video games have consistently pushed the envelope over the years, helping drive the need for higher bandwidth I/O capabilities in devices and networks, better and faster graphics capabilities, and larger and faster storage (which ultimately led to flash memory and even Hadoop). Animation has become very lifelike and games are becoming more responsive to audio commands. It is not a stretch of the imagination to believe that this is where the next generation of smart interfaces will be found (instead of from the evolution of current smart interfaces).
Someday it may no longer be possible to “tweak” results through the use or omission of keywords, quotation marks, and flags. Additionally, it may no longer be necessary to understand special query languages (SQL, NoSQL, SPARQL, etc.) and syntax. We won’t have to worry as much about incorrect joins, spurious correlations and biased result sets. Instead, we will be given the answers we need – even if we don’t realize that this was what we needed in the first place. At that point computer systems may appear nearly omniscient.
When this happens parents will no longer need to teach their children “Google-Fu.” Those are going be interesting times indeed.
Genetics, Genomics, Nanotechnology, and more
Science has been interesting to me for most of my lifetime, but it wasn’t until my first child was born that I shifted from “interested” to “involved.” My eldest daughter was diagnosed with Systemic Onset Juvenile Idiopathic Arthritis (SoJIA – originally called Juvenile Rheumatoid Arthritis, or JRA) when she was 15 months old, which also happened to be about six months into the start of my old Consulting company and in the middle of a very critical Y2K ERP system upgrade and rehosting project. It was definitely a challenging time in my life.
At that time there was very little research on JRA because it was estimated there were only 30,000 children affected by the disease and the implication was that funding research would not have a positive ROI. As an aside, this was also a few years before the breakthroughs of biological medicines like Enbrel for use on children.
One of the things that I learned was that this disease could be horribly debilitating. Children often had physical deformities as a result of this disease. Even worse, the systemic type that my daughter has could result in premature death. As a first-time parent it was extremely difficult to imagine that type of life for your child.
Luckily, the company that I had just started was taking off so I decided to finds ways to personally make a tangible difference for all children with this disease. We decided to take 50% of our net profits and use them to fund medical research. We had a goal of funding $1 million in research and finding a cure for Juvenile Arthritis within the next 5-7 years.
As someone new to “major gifts” and philanthropy I quickly learned that some forms of gifts were more beneficial than others. While most organizations wanted you to start a fund (which we did), the impact from that tended to be more long-term and less immediate. I met someone who was passionate, knowledgeable, and successful in her field who showed me a different and better approach (here’s a post that describes that in more detail).
I no longer wanted to blindly give money and hope that it was used quickly and properly. Rather, I wanted to treat these donations like investments in a near-term cure. In order to be successful, I needed to understand research from both medical and scientific perspectives in these areas. That began a new journey of research and independent learning in completely new areas.
There was a lot going on in the fields of Genetics and Genomics at the time (here’s a good explanation of the difference between the two). My interest and efforts in this area led to a position on the Medical and Scientific Advisory Committee with the Arthritis Foundation. With the exception of me, the other members were talented and successful physicians who were also involved with medical research. We met quarterly, and I did ask questions and make suggestions that made a difference. But, unlike everyone else on the committee, I needed to study and prepare for 40+ hours for each call to ensure that had enough of an understanding to add value and not be a distraction.
A few years later we did work for a Nanotechnology company (more info here for those interested). The Chief Scientist wasn’t that interested in explaining what they did until I described some of our research projects on gene expression. He then went into great detail about what they were doing and how he believed it would change what we do in the future. I saw that and agreed, but also started thinking of the potential for leveraging nanotechnology with medicine.
I was listening to the “TED Radio Hour” while driving today and heard a segment about entrepreneur Richard Resnick. It was exciting because it got me thinking about this again, and this is a topic that I haven’t thought about much for the past few years (the last time was contemplating how new analytics products could be useful in this space).
There are efforts going on today with custom, personalized medicines that target only specific genes for a very specific outcome. The genetic modifications being performed on plants today will likely be performed on humans in the near future (I would guess within 10-15 years). The body is an incredibly adaptive organism, so it will be very challenging to implement anything that is consistently safe and effective long-term. But, that day will come.
It’s not a huge leap to go from genetically modified “treatment cells” to true nanotechnology (and not just extremely small particles). Just think, machines that can be designed to work independently within us, do what they are programmed to do, and more importantly identify and understand adaptations (i.e., artificial intelligence) as they occur and alter their approach and treatment plan accordingly. To me, this is extremely exciting. It’s not that I want to live to be 100+ years old – because I don’t. But, being able to do things that have a positive impact on the quality of life for children and their families is a worthy goal from my perspective.
My advice is to always continue learning, keep an open mind, and see what you can personally do to make a difference. You will never know unless you try.
Note: Updated to fix and remove dead links.