Sections

When Analytics Fail

Nick Petri

Image provided by: ETF Prodigy

After hearing the redundancy of my title, Analyst on OpenView’s Research and Analytics team, most people naturally assume I love numbers. Often I tell them I do. Numbers boil down an infinitely complex world into a manageable format that can be manipulated, modeled, and mined for insight. Accurate numbers, properly used, can teach you a ton about how the world works, and in turn, the adjustments you make based on that analysis can affect the world in tangible and profound ways.

But numbers can’t answer everything, and if you follow them blindly, you can be badly misled.

Too many executives, whether “numbers people” or not, see statistical analysis as a necessary stamp of approval to any business decision.  But the simple truth is that there isn’t a mathematical answer to every question, and if you force one using data that doesn’t accurately describe the real world, no amount of color-coded charts, quadratic regressions, and chi-squared tests will yield an ounce of insight.

These situations are incredibly common in the business world. Unlike physicists, chemists, or biologists, business people are unable to run tightly controlled experiments. We can’t scour the earth for cool problems with clean statistical answers, like academics can. The problems come to us, and they’re usually messy and fraught with biases.

While the statistical discipline has very good tools for identifying relationships within a data set, it’s blind to the quality of the data itself. Here are three common problems that can render a statistically significant conclusion dead wrong:

  • Self-Selection: Survey or interview respondents often respond because they have good things to say about the subject. Likewise, firms prefer to disclose financial information when it’s favorable or improving. This type of problem paints a rosier picture, on average, than is really the case.
  • Missing Data: More information is almost always available on larger companies and more recent events. When there’s missing information, do you come up with a proxy, estimate the missing fields, or exclude those entries altogether? Any solution you choose will introduce a layer of noise into the equation.
  • Tenuous Proxies: Say we’re trying to measure the impact of marketing spend on customer acquisition for a particular industry. Since neither variable is public, we’ll have to use proxies and assumptions to estimate them. It’s important to remember that the ultimate conclusion isn’t measuring the actual variables, but their proxies: the strength of the conclusion relies heavily on how close the two are. Since we don’t have the actual variables, this can be very difficult to measure.

The result is a difficult balancing act: you have to tolerate moderate levels of bias within the data to reach a conclusion, but as the problems pile up, the validity of that conclusion declines no matter how statistically significant it may appear.

Still, an analyst is being paid to solve a problem, and to some, an inconclusive result is the worst kind of failure. It can be tempting to present a conclusion as a home run despite obvious flaws in the methodology. The best analysts properly communicate their level of confidence in the results, and know when to admit that the data at hand is inconclusive, even if it isn’t what their stakeholders were hoping to hear. Doing otherwise can be extremely destructive.

So do I love numbers? I’d say our relationship is rocky. I certainly respect them as a powerful tool to understand the world around me, but am realistic about their limitations. Statistical conclusions are only as good as the data that goes into them, and there isn’t good data for every problem.

Comments

Tags:

awareness

Data Analysts And Scientists More Important Than Ever For The Enterprise

Daniel Newman

The business world is now firmly in the age of data. Not that data wasn’t relevant before; it was just nowhere close to the speed and volume that’s available to us today. Businesses are buckling under the deluge of petabytes, exabytes, and zettabytes. Within these bytes lie valuable information on customer behavior, key business insights, and revenue generation. However, all that data is practically useless for businesses without the ability to identify the right data. Plus, if they don’t have the talent and resources to capture the right data, organize it, dissect it, draw actionable insights from it and, finally, deliver those insights in a meaningful way, their data initiatives will fail.

Rise of the CDO

Companies of all sizes can easily find themselves drowning in data generated from websites, landing pages, social streams, emails, text messages, and many other sources. Additionally, there is data in their own repositories. With so much data at their disposal, companies are under mounting pressure to utilize it to generate insights. These insights are critical because they can (and should) drive the overall business strategy and help companies make better business decisions. To leverage the power of data analytics, businesses need more “top-management muscle” specialized in the field of data science. This specialized field has lead to the creation of roles like Chief Data Officer (CDO).

In addition, with more companies undertaking digital transformations, there’s greater impetus for the C-suite to make data-driven decisions. The CDO helps make data-driven decisions and also develops a digital business strategy around those decisions. As data grows at an unstoppable rate, becoming an inseparable part of key business functions, we will see the CDO act as a bridge between other C-suite execs.

Data skills an emerging business necessity

So far, only large enterprises with bigger data mining and management needs maintain in-house solutions. These in-house teams and technologies handle the growing sets of diverse and dispersed data. Others work with third-party service providers to develop and execute their big data strategies.

As the amount of data grows, the need to mine it for insights becomes a key business requirement. For both large and small businesses, data-centric roles will experience endless upward mobility. These roles include data anlysts and scientists. There is going to be a huge opportunity for critical thinkers to turn their analytical skills into rapidly growing roles in the field of data science. In fact, data skills are now a prized qualification for titles like IT project managers and computer systems analysts.

Forbes cited the McKinsey Global Institute’s prediction that by 2018 there could be a massive shortage of data-skilled professionals. This indicates a disruption at the demand-supply level with the needs for data skills at an all-time high. With an increasing number of companies adopting big data strategies, salaries for data jobs are going through the roof. This is turning the position into a highly coveted one.

According to Harvard Professor Gary King, “There is a big data revolution. The big data revolution is that now we can do something with the data.” The big problem is that most enterprises don’t know what to do with data. Data professionals are helping businesses figure that out. So if you’re casting about for where to apply your skills and want to take advantage of one of the best career paths in the job market today, focus on data science.

I’m compensated by University of Phoenix for this blog. As always, all thoughts and opinions are my own.

For more insight on our increasingly connected future, see The $19 Trillion Question: Are You Undervaluing The Internet Of Things?

The post Data Analysts and Scientists More Important Than Ever For the Enterprise appeared first on Millennial CEO.

Comments

About Daniel Newman

Daniel Newman serves as the Co-Founder and CEO of EC3, a quickly growing hosted IT and Communication service provider. Prior to this role Daniel has held several prominent leadership roles including serving as CEO of United Visual. Parent company to United Visual Systems, United Visual Productions, and United GlobalComm; a family of companies focused on Visual Communications and Audio Visual Technologies. Daniel is also widely published and active in the Social Media Community. He is the Author of Amazon Best Selling Business Book "The Millennial CEO." Daniel also Co-Founded the Global online Community 12 Most and was recognized by the Huffington Post as one of the 100 Business and Leadership Accounts to Follow on Twitter. Newman is an Adjunct Professor of Management at North Central College. He attained his undergraduate degree in Marketing at Northern Illinois University and an Executive MBA from North Central College in Naperville, IL. Newman currently resides in Aurora, Illinois with his wife (Lisa) and his two daughters (Hailey 9, Avery 5). A Chicago native all of his life, Newman is an avid golfer, a fitness fan, and a classically trained pianist

When Good Is Good Enough: Guiding Business Users On BI Practices

Ina Felsheim

Image_part2-300x200In Part One of this blog series, I talked about changing your IT culture to better support self-service BI and data discovery. Absolutely essential. However, your work is not done!

Self-service BI and data discovery will drive the number of users using the BI solutions to rapidly expand. Yet all of these more casual users will not be well versed in BI and visualization best practices.

When your user base rapidly expands to more casual users, you need to help educate them on what is important. For example, one IT manager told me that his casual BI users were making visualizations with very difficult-to-read charts and customizing color palettes to incredible degrees.

I had a similar experience when I was a technical writer. One of our lead writers was so concerned with readability of every sentence that he was going through the 300+ page manuals (yes, they were printed then) and manually adjusting all of the line breaks and page breaks. (!) Yes, readability was incrementally improved. But now any number of changes–technical capabilities, edits, inserting larger graphics—required re-adjusting all of those manual “optimizations.” The time it took just to do the additional optimization was incredible, much less the maintenance of these optimizations! Meanwhile, the technical writing team was falling behind on new deliverables.

The same scenario applies to your new casual BI users. This new group needs guidance to help them focus on the highest value practices:

  • Customization of color and appearance of visualizations: When is this customization necessary for a management deliverable, versus indulging an OCD tendency? I too have to stop myself from obsessing about the font, line spacing, and that a certain blue is just a bit different than another shade of blue. Yes, these options do matter. But help these casual users determine when that time is well spent.
  • Proper visualizations: When is a spinning 3D pie chart necessary to grab someone’s attention? BI professionals would firmly say “NEVER!” But these casual users do not have a lot of depth on BI best practices. Give them a few simple guidelines as to when “flash” needs to subsume understanding. Consider offering a monthly one-hour Lunch and Learn that shows them how to create impactful, polished visuals. Understanding if their visualizations are going to be viewed casually on the way to a meeting, or dissected at a laptop, also helps determine how much time to spend optimizing a visualization. No, you can’t just mandate that they all read Tufte.
  • Predictive: Provide advanced analytics capabilities like forecasting and regression directly in their casual BI tools. Using these capabilities will really help them wow their audience with substance instead of flash.
  • Feature requests: Make sure you understand the motivation and business value behind some of the casual users’ requests. These casual users are less likely to understand the implications of supporting specific requests across an enterprise, so make sure you are collaborating on use cases and priorities for substantive requests.

By working with your casual BI users on the above points, you will be able to collectively understand when the absolute exact request is critical (and supports good visualization practices), and when it is an “optimization” that may impact productivity. In many cases, “good” is good enough for the fast turnaround of data discovery.

Next week, I’ll wrap this series up with hints on getting your casual users to embrace the “we” not “me” mentality.

Read Part One of this series: Changing The IT Culture For Self-Service BI Success.

Follow me on Twitter: @InaSAP

Comments

How AI Can End Bias

Yvonne Baur, Brenda Reid, Steve Hunt, and Fawn Fitter

We humans make sense of the world by looking for patterns, filtering them through what we think we already know, and making decisions accordingly. When we talk about handing decisions off to artificial intelligence (AI), we expect it to do the same, only better.

Machine learning does, in fact, have the potential to be a tremendous force for good. Humans are hindered by both their unconscious assumptions and their simple inability to process huge amounts of information. AI, on the other hand, can be taught to filter irrelevancies out of the decision-making process, pluck the most suitable candidates from a haystack of résumés, and guide us based on what it calculates is objectively best rather than simply what we’ve done in the past.

In other words, AI has the potential to help us avoid bias in hiring, operations, customer service, and the broader business and social communities—and doing so makes good business sense. For one thing, even the most unintentional discrimination can cost a company significantly, in both money and brand equity. The mere fact of having to defend against an accusation of bias can linger long after the issue itself is settled.

Beyond managing risk related to legal and regulatory issues, though, there’s a broader argument for tackling bias: in a relentlessly competitive and global economy, no organization can afford to shut itself off from broader input, more varied experiences, a wider range of talent, and larger potential markets.

That said, the algorithms that drive AI don’t reveal pure, objective truth just because they’re mathematical. Humans must tell AI what they consider suitable, teach it which information is relevant, and indicate that the outcomes they consider best—ethically, legally, and, of course, financially—are those that are free from bias, conscious or otherwise. That’s the only way AI can help us create systems that are fair, more productive, and ultimately better for both business and the broader society.

Bias: Bad for Business

When people talk about AI and machine learning, they usually mean algorithms that learn over time as they process large data sets. Organizations that have gathered vast amounts of data can use these algorithms to apply sophisticated mathematical modeling techniques to see if the results can predict future outcomes, such as fluctuations in the price of materials or traffic flows around a port facility. Computers are ideally suited to processing these massive data volumes to reveal patterns and interactions that might help organizations get ahead of their competitors. As we gather more types and sources of data with which to train increasingly complex algorithms, interest in AI will become even more intense.

Using AI for automated decision making is becoming more common, at least for simple tasks, such as recommending additional products at the point of sale based on a customer’s current and past purchases. The hope is that AI will be able to take on the process of making increasingly sophisticated decisions, such as suggesting entirely new markets where a company could be profitable, or finding the most qualified candidates for jobs by helping HR look beyond the expected demographics.

As AI takes on these increasingly complex decisions, it can help reduce bias, conscious or otherwise. By exposing a bias, algorithms allow us to lessen the impact of that bias on our decisions and actions. They enable us to make decisions that reflect objective data instead of untested assumptions; they reveal imbalances; and they alert people to their cognitive blind spots so they can make more accurate, unbiased decisions.

Imagine, for example, a major company that realizes that its past hiring practices were biased against women and that would benefit from having more women in its management pipeline. AI can help the company analyze its past job postings for gender-biased language, which might have discouraged some applicants. Future postings could be more gender neutral, increasing the number of female applicants who get past the initial screenings.

AI can also support people in making less-biased decisions. For example, a company is considering two candidates for an influential management position: one man and one woman. The final hiring decision lies with a hiring manager who, when they learn that the female candidate has a small child at home, assumes that she would prefer a part-time schedule.

That assumption may be well intentioned, but it runs counter to the outcome the company is looking for. An AI could apply corrective pressure by reminding the hiring manager that all qualifications being equal, the female candidate is an objectively good choice who meets the company’s criteria. The hope is that the hiring manager will realize their unfounded assumption and remove it from their decision-making process.

At the same time, by tracking the pattern of hiring decisions this manager makes, the AI could alert them—and other people in HR—that the company still has some remaining hidden biases against female candidates to address.

Look for Where Bias Already Exists

In other words, if we want AI to counter the effects of a biased world, we have to begin by acknowledging that the world is biased. And that starts in a surprisingly low-tech spot: identifying any biases baked into your own organization’s current processes. From there, you can determine how to address those biases and improve outcomes.

There are many scenarios where humans can collaborate with AI to prevent or even reverse bias, says Jason Baldridge, a former associate professor of computational linguistics at the University of Texas at Austin and now co-founder of People Pattern, a startup for predictive demographics using social media analytics. In the highly regulated financial services industry, for example, Baldridge says banks are required to ensure that their algorithmic choices are not based on input variables that correlate with protected demographic variables (like race and gender). The banks also have to prove to regulators that their mathematical models don’t focus on patterns that disfavor specific demographic groups, he says. What’s more, they have to allow outside data scientists to assess their models for code or data that might have a discriminatory effect. As a result, banks are more evenhanded in their lending.

Code Is Only Human

The reason for these checks and balances is clear: the algorithms that drive AI are built by humans, and humans choose the data with which to shape and train the resulting models. Because humans are prone to bias, we have to be careful that we are neither simply confirming existing biases nor introducing new ones when we develop AI models and feed them data.

“From the perspective of a business leader who wants to do the right thing, it’s a design question,” says Cathy O’Neil, whose best-selling book Weapons of Math Destruction was long-listed for the 2016 National Book Award. “You wouldn’t let your company design a car and send it out in the world without knowing whether it’s safe. You have to design it with safety standards in mind,” she says. “By the same token, algorithms have to be designed with fairness and legality in mind, with standards that are understandable to everyone, from the business leader to the people being scored.” (To learn more from O’Neil about transparency in algorithms, read Thinkers in this issue.)

Don’t Do What You’ve Always Done

To eliminate bias, you must first make sure that the data you’re using to train the algorithm is itself free of bias, or, rather, that the algorithm can recognize bias in that data and bring the bias to a human’s attention.

SAP has been working on an initiative that tackles this issue directly by spotting and categorizing gendered terminology in old job postings. Nothing as overt as No women need apply, which everyone knows is discriminatory, but phrases like outspoken and aggressively pursuing opportunities, which are proven to attract male job applicants and repel female applicants, and words like caring and flexible, which do the opposite.

Once humans categorize this language and feed it into an algorithm, the AI can learn to flag words that imply bias and suggest gender-neutral alternatives. Unfortunately, this de-biasing process currently requires too much human intervention to scale easily, but as the amount of available de-biased data grows, this will become far less of a limitation in developing AI for HR.

Similarly, companies should look for specificity in how their algorithms search for new talent. According to O’Neil, there’s no one-size-fits-all definition of the best engineer; there’s only the best engineer for a particular role or project at a particular time. That’s the needle in the haystack that AI is well suited to find.

Look Beyond the Obvious

AI could be invaluable in radically reducing deliberate and unconscious discrimination in the workplace. However, the more data your company analyzes, the more likely it is that you will deal with stereotypes, O’Neil says. If you’re looking for math professors, for example, and you load your hiring algorithm with all the data you can find about math professors, your algorithm may give a lower score to a black female candidate living in Harlem simply because there are fewer black female mathematicians in your data set. But if that candidate has a PhD in math from Cornell, and if you’ve trained your AI to prioritize that criterion, the algorithm will bump her up the list of candidates rather than summarily ruling out a potentially high-value hire on the spurious basis of race and gender.

To further improve the odds that AI will be useful, companies have to go beyond spotting relationships between data and the outcomes they care about. It doesn’t take sophisticated predictive modeling to determine, for example, that women are disproportionately likely to jump off the corporate ladder at the halfway point because they’re struggling with work/life balance.

Many companies find it all too easy to conclude that women simply aren’t qualified for middle management. However, a company committed to smart talent management will instead ask what it is about these positions that makes them incompatible with women’s lives. It will then explore what it can change so that it doesn’t lose talent and institutional knowledge that will cost the company far more to replace than to retain.

That company may even apply a second layer of machine learning that looks at its own suggestions and makes further recommendations: “It looks like you’re trying to do X, so consider doing Y,” where X might be promoting more women, making the workforce more ethnically diverse, or improving retention statistics, and Y is redefining job responsibilities with greater flexibility, hosting recruiting events in communities of color, or redesigning benefits packages based on what similar companies offer.

Context Matters—and Context Changes

Even though AI learns—and maybe because it learns—it can never be considered “set it and forget it” technology. To remain both accurate and relevant, it has to be continually trained to account for changes in the market, your company’s needs, and the data itself.

Sources for language analysis, for example, tend to be biased toward standard American English, so if you’re building models to analyze social media posts or conversational language input, Baldridge says, you have to make a deliberate effort to include and correct for slang and nonstandard dialects. Standard English applies the word sick to someone having health problems, but it’s also a popular slang term for something good or impressive, which could lead to an awkward experience if someone confuses the two meanings, to say the least. Correcting for that, or adding more rules to the algorithm, such as “The word sick appears in proximity to positive emoji,” takes human oversight.

Moving Forward with AI

Today, AI excels at making biased data obvious, but that isn’t the same as eliminating it. It’s up to human beings to pay attention to the existence of bias and enlist AI to help avoid it. That goes beyond simply implementing AI to insisting that it meet benchmarks for positive impact. The business benefits of taking this step are—or soon will be—obvious.

In IDC FutureScapes’ webcast “Worldwide Big Data, Business Analytics, and Cognitive Software 2017 Predictions,” research director David Schubmehl predicted that by 2020 perceived bias and lack of evidentiary transparency in cognitive/AI solutions will create an activist backlash movement, with up to 10% of users backing away from the technology. However, Schubmehl also speculated that consumer and enterprise users of machine learning will be far more likely to trust AI’s recommendations and decisions if they understand how those recommendations and decisions are made. That means knowing what goes into the algorithms, how they arrive at their conclusions, and whether they deliver desired outcomes that are also legally and ethically fair.

Clearly, organizations that can address this concern explicitly will have a competitive advantage, but simply stating their commitment to using AI for good may not be enough. They also may wish to support academic efforts to research AI and bias, such as the annual Fairness, Accountability, and Transparency in Machine Learning (FATML) workshop, which was held for the third time in November 2016.

O’Neil, who blogs about data science and founded the Lede Program for Data Journalism, an intensive certification program at Columbia University, is going one step further. She is attempting to create an entirely new industry dedicated to auditing and monitoring algorithms to ensure that they not only reveal bias but actively eliminate it. She proposes the formation of groups of data scientists that evaluate supply chains for signs of forced labor, connect children at risk of abuse with resources to support their families, or alert people through a smartphone app when their credit scores are used to evaluate eligibility for something other than a loan.

As we begin to entrust AI with more complex and consequential decisions, organizations may also want to be proactive about ensuring that their algorithms do good—so that their companies can use AI to do well. D!

Read more thought provoking articles in the latest issue of the Digitalist Magazine, Executive Quarterly.


About the Authors:

Yvonne Baur is Head of Predictive Analytics for Sap SuccessFactors solutions.

Brenda Reid is Vice President of Product Management for Sap SuccessFactors solutions.

Steve Hunt is Senior Vice President of Human Capital Management Research for Sap SuccessFactors solutions.

Fawn Fitter is a freelance writer specializing in business and technology.

Comments

Tags:

Big Data: Better Than Big Muscles at Kinduct

Stephanie Overby

Travis McDonough has always been looking for a competitive edge. As an amateur athlete “on the small side,” he sought other ways—exercise, nutrition, strategy—to get ahead.

Today McDonough is the of CEO of Kinduct, a provider of cloud-based software that analyzes data from wearables, electronic medical records, computer vision solutions, and more to assess and make recommendations about physical human performance. Kinduct provides 100 professional sports organizations, including the five major sports leagues in North America, with intelligence to make decisions about their athletes and training programs.

Digital Fills a Gap

A chiropractor by training, McDonough owned and operated a network of sports rehabilitation clinics, where he found that patients retained only a fraction of what they were instructed to do through text or conversation. “As we treated athletes, we realized there was a gaping hole in the industry for technology [to fill],” he says.

McDonough first launched a company to create 3D videos designed to help his athlete patients better understand their injuries and the resulting therapy. The videos, delivered by text or e-mail, would illustrate what happens inside the human body when it experiences whiplash, for example.

“We quickly realized we couldn’t just be a content company and push information without understanding more about the athlete,” he says. Athletes and their trainers collected a massive amount of individual health and performance data that was available to be tapped from electronic medical records, wearable devices, and computer vision-based tracking systems that measure and record information such as how fast an athlete is running or jumping. “We needed to be agnostic and aggressive consumers of all kinds of data sources in order to push more targeted programs to our clients,” he says. So McDonough recruited his brother’s brother-in-law (vice president of product, Dave Anderson) to develop software to make sense of it all.

Innovate a Better Athlete

The software is suited for healthcare and military applications: the Canadian Armed Forces uses it to deliver exercise, wellness, and nutrition programs to its troops. But McDonough knew that the world of professional sports would provide his most eager customers.

Professional sports teams use Kinduct’s analytics to reduce injury and win more games.

“The sports world is willing to embrace innovation more quickly than other markets, like healthcare, that are slower-moving. And that’s where our passion lives. Many of us are sports fanatics and have been athletes,” says McDonough of the company’s 70 employees. Kinduct’s first customers were National Hockey League (NHL) teams, followed in short order by the National Basketball Association (NBA).

For its professional sports clients, Kinduct has uncovered more than 100 novel correlations. Most are closely guarded secrets, but several have become public. The company found, for example, that when a basketball player’s sleep falls below a certain threshold, there is a strong correlation with reduced free throw percentages two days later. That discovery led one NBA team (McDonough won’t say which) to focus on getting players to bed on time and making travel schedule changes to enable the requisite rest.

Kinduct software also found correlations for hockey teams. It demonstrated to a leading hockey team that better grip strength was likely to lead to harder and faster shots on goal. Moreover, when the system ingested three years of historical computer vision information, it found that a player’s ability to slow down dramatically affects the chances of soft tissue injuries, which are costly to professional sports teams and athletes. The software can send an alert when it spots a trend that could predict the possibility of such an injury.

We’re in this to go big. That means carrying a burn rate, hiring aggressively, and investing in research.

The software “will never replace the experts in the trenches,” says McDonough. “But we are able to arm coaches and trainers with the intelligence necessary to make more informed decisions. Technology will never replace the power of a good relationship.”

Think a Few Plays Ahead

Kinduct is based in McDonough’s hometown of Halifax, Nova Scotia, which boasts five universities, strong government subsidies, a low cost of living, and, for Kinducts’s predominantly U.S.-based customers, a favorable currency exchange rate. Despite these advantages, Halifax isn’t widely known for its digital innovators. “We’ve got a huge chip on our shoulder,” says McDonough. “We want to prove that we’re just as capable of becoming a global success as companies elsewhere,” such as Silicon Valley or London.

The Kinduct platform can help athletes or medical patients improve their condition or performance.

Nevertheless, McDonough spends significant time in Silicon Valley meeting with investors and looking at potential U.S. expansion (Kinduct closed a US$9 million Series A investment led by Intel Capital in October). “There’s a huge benefit to growing in Nova Scotia,” he says, “but we also need to be in the epicenter of the tech space.”

McDonough has big ideas for Kinduct’s future, thanks to the explosion of health- and fitness-tracking devices. “We can pull all the data in and, when we see a negative pattern, provide the user with the exact roadmap they need to follow to improve their condition or performance,” he says. “That’s equally as useful to a professional football player or an Olympic athlete as it is to someone recovering from a knee replacement or living with type 2 diabetes.”

Kinduct has 16 projects underway to measure the impact of the platform in helping individuals manage conditions like peripheral vascular disease and cognitive decline. “We want to show how the platform can empower and engage patients,” says McDonough.

Go Big or Go Home

Meanwhile, however, McDonough intends “to dominate the sports space. That is our bubble wrap of credibility, and we can leverage that to do other things.”

Focus was never a strong suit for McDonough, who struggled with dyslexia and ADD as a kid. “Thank God for sport, which helped to channel my energy,” he says. But that wandering mind, he says, has also been an asset. “Like a lot of ADD sufferers, I have a lot of imagination,” he says. For balance, he’s hired a leadership team that keeps him grounded, and he has assembled a board of experienced business and technology leaders. “They have the institutional knowledge in how to scale,” he says.

McDonough is blunt: right now, he’d rather be innovative than profitable. “We’re in this to go big. That means carrying a burn rate, hiring aggressively, and investing in research,” he says. “We’re lucky enough to be in locker rooms with these teams and close to some of the best in the business in terms of medicine and training and data science. That’s helping us to produce our future roadmap.” D!

Read more thought provoking articles in the latest issue of the Digitalist Magazine, Executive Quarterly.

Comments