Smartphones and tablets, along with apps connected to new cloud-computing platforms, are revolutionizing the workplace. We’re still early in this workplace transformation, and the tools so familiar to us will be around for quite sometime. The leaders, managers, and organizations that are using new tools sooner will quickly see how tools can drive cultural changes — developing products faster, with less bureaucracy and more focus on what’s important to the business.
If you’re trying to change how work is done, changing the tools and processes can be an eye-opening first step.
Many of the companies I work with are creating new productivity tools, and every company starting now is using them as a first principle. Companies run their business on new software-as-a-service tools. The basics of email and calendaring infrastructure are built on the tools of the consumerization of IT. Communication and work products between members of the team and partners are using new tools that were developed from the ground up for sharing, collaboration and mobility.
Some of the exciting new tools for productivity that you can use today include: Quip,Evernote, Box and Box Notes, Dropbox, Slack, Hackpad, Asana, Pixxa Perspective, Haiku Deck, and more below. This list is by no means exhaustive, and new tools are showing up all the time. Some tools take familiar paradigms and pivot them for touch and mobile. Others are hybrids of existing tools that take a new view on how things can be more efficient, streamlined, or attuned to modern scenarios. All are easily used via trials for small groups and teams, even within large companies.
Tools drive cultural change
Tools have a critical yet subtle impact on how work gets done. Tools can come to define the work, as much as just making work more efficient. Early in the use of new tools there’s a combination of a huge spike in benefit, along with a temporary dip in productivity. Even with all the improvements, all tools over time can become a drag on productivity as the tools become the end, rather than the means to an end. This is just a natural evolution of systems and processes in organizations, and productivity tools are no exception. It is something to watch for as a team.
The spike comes from the new ways information is acquired, shared, created, analyzed and more. Back when the PC first entered the workplace, it was astounding to see the rapid improvements in basic things like preparing memos, making “slides,” or the ability to share information via email.
There’s a temporary dip in productivity as new individual and organizational muscles are formed and old tools and processes are replaced across the whole team. Everyone individually — and the team has a whole — feels a bit disrupted during this time. Things rapidly return to a “new normal,” and with well-chosen tools and thoughtfully-designed processes, this is an improvement.
As processes mature or age, it is not uncommon for those very gains to become burdensome. When a new lane opens on a highway, traffic moves faster for awhile, until more people discover the faster route, and then it feels like things are back where they started. Today’s most common tools and processes have reached a point where the productivity increases they once brought feel less like improvements and more like extra work that isn’t needed. All too often, the goals have long been lost, and the use of tools is on autopilot, with the reason behind the work simply “because we always did it that way.”
New tools are appearing that offer new ways to work. These new ways are not just different — this is not about fancier reports, doing the old stuff marginally faster, or bigger spreadsheets. Rather, these new tools are designed to solve problems faced by today’s mobile and continuous organization. These tools take advantage of paradigms native to phones and tablets. Data is stored on a cloud. Collaboration takes place in real time. Coordination of work is baked into the tools. Work can be accessed from a broad range of computing devices of all types. These tools all build on the modern SaaS model, so they are easy to get, work outside your firewall and come with the safety and security of cloud-native companies.
The cultural changes enabled by these tools are significant. While it is possible to think about using these tools “the same old way,” you’re likely to be disappointed. If you think a new tool that is about collaboration on short-lived documents will have feature parity with a tool for crafting printed books, then you’re likely to feel like things are missing. If you’re looking to improve your organizational effectiveness at communication, collaboration and information sharing, then you’re also going to want to change some of the assumptions about how your organization works. The fact that the new tools do some things worse and other things differently points to the disruptive innovation that these products have the potential to bring — the “Innovator’s Dilemma” is well known to describe the idea that disruptive products often feel inferior when compared to entrenched products using existing criteria.
Overcoming traps and pitfalls
Based on seeing these tools in action and noticing how organizations can re-form around new ways of working, the following list compiles some of the most common pitfalls addressed by new tools. In other words, if you find yourself doing these things, it’s time to reconsider the tools and processes on your team, and try something new.
Some of these will seem outlandish when viewed through today’s concept. As a person who worked on productivity tools for much of my career, I think back to the time when it was crazy to use a word processor for a college paper; or when I first got a job, and typing was something done by the “secretarial pool.” Even the use of email in the enterprise was first ridiculed, and many managers had assistants who would print out email and then type dictated replies (no, really!). Things change slowly, then all of a sudden there are new norms.
In our Harvard Business School class, “Digital Innovation,” we crafted a notion of “doing it wrong,” and spent a session looking at disruption in the tools of the workplace. In that spirit, “you’re doing it wrong,” if you:
- Spend more time summarizing or formatting a document than worrying about the actual content. Time and time again, people over-invest in the production qualities of a work product, only to realize that all that work was wasted, as most people consume it on a phone or look for the summary. This might not be new, but it is fair to say that the feature sets of existing tools and implementation (both right for when they were created, I believe) would definitely emphasize this type of activity.
- Aim to “complete” a document, and think your work is done when a document is done. The modern world of business and product development knows that you’re never done with a product, and that is certainly the case for documents that are steps along the way. Modern tools assume that documents continue to exist but fade in activity — the value is in getting the work out there to the cloud, and knowing that the document itself is rarely the end goal.
- Figure out something important with a long email thread, where the context can’t be shared and the backstory is lost. If you’re collaborating via email, you’re almost certainly losing important context, and not all the right folks are involved. A modern collaboration tool like Slack keeps everything relevant in the tool, accessible by everyone on the team from everywhere at any time, but with a full history and search.
- Delay doing things until someone can get on your calendar, or you’re stuck waiting on someone else’s calendar. The existence of shared calendaring created a world of matching free/busy time, which is great until two people agree to solve an important problem — two weeks from now. Modern communication tools allow for notifications, fast-paced exchange of ideas and an ability to keep things moving. Culturally, if you let a calendar become a bottleneck, you’re creating an opening for a competitor, or an opportunity for a customer or partner to remain unhappy. Don’t let calendaring become a work-prevention tool.
- Believe that important choices can be distilled down into a one-hour meeting. If there’s something important to keep moving on, then scheduling a meeting to “bring everyone together” is almost certainly going to result in more delays (in addition to the time to get the meeting going in the first place). The one-hour meeting for a challenging issue almost never results in a resolution, but always pushes out the solution. If you’re sharing information all along, and the right people know all that needs to be known, then the modern resolution is right there in front of you. Speaking as a person who almost always shunned meetings to avoid being a bottleneck, I think it’s worth considering that the age-old technique of having short and daily sync meetings doesn’t really address this challenge. Meetings themselves, one might argue, are increasingly questionable in a world of continuously connected teams.
- Bring dead trees and static numbers to the table, rather than live, onscreen data. Live data analysis was invented 20 years ago, but too many still bring snapshots of old data to meetings which then too often digress into analyzing the validity of numbers or debating the slice/view of the data, further delaying action until there’s an update. Modern tools like Tidemark and Apptio provide real-time and mobile access to information. Meetings should use live data, and more importantly, the team should share access to live data so everyone is making choices with all the available information.
- Use the first 30 minutes of a meeting recreating and debating the prior context that got you to a meeting in the first place. All too often, when a meeting is scheduled far in advance, things change so much that by the time everyone is in the room, the first half of the hour (after connecting projectors, going through an enterprise log-on, etc.) is spent with everyone reminding each other and attempting to agree on the context and purpose of the gathering. Why not write out a list of issues in a collaborative document like Quip, and have folks share thoughts and data in real time to first understand the issue?
- Track what work needs to happen for a project using analog tools. Far too many projects are still tracked via paper and pen which aren’t shared, or on whiteboards with too little information, or in a spreadsheet mailed around over and over again. Asana is a simple example of an easy-to-use and modern tool that decreases (to zero) email flow, allows for everyone to contribute and align on what needs to be done, and to have a global view of what is left to do.
- Need to think which computer or device your work is “on.” Cloud storage from Box,Dropbox, OneDrive and others makes it easy (and essential) to keep your documents in the cloud. You can edit, share, comment and track your documents from any device at any time. There’s no excuse for having a document stuck on a single computer, and certainly no excuse risking the use of USB storage for important work.
- Use different tools to collaborate with partners than you use with fellow employees. Today’s teams are made up of vendors, contractors, partners and customers all working together. Cloud-based tools solve the problem of access and security in modern ways that treat everyone as equals in the collaboration process. There’s a huge opportunity to increase the effectiveness of work across the team by using one set of tools across organizational boundaries.
Many of these might seem far-fetched, and even heretical to some. From laptops to color printing to projectors in conference rooms to wireless networking to the Internet itself, each of those tools were introduced to skeptics who said the tools currently in use were “good enough,” and the new tools were slower, less efficient, more expensive, or just superfluous.
The teams that adopt new tools and adapt their way of working will be the most competitive and productive teams in an organization. Not every tool will work, and some will even fail. The best news is that today’s approach to consumerization makes trial easier and cheaper than at any other time.
If you’re caught in a rut, doing things the old way, the tools are out there to work in new ways and start to change the culture of your team.
–Steven Sinofsky @stevesi
This article originally appeared on <re/code>.
One of the most difficult stages in growing your own skillset is when you have to hire someone for a job you can’t actually do yourself. Whether you’re a founder of a new company, or just growing a company or team, at some point the skills needed for a growing organization exceed your own experience.
Admitting that you don’t really have the skills the business requires is the first, and most difficult step. This is especially true as an engineer where there’s a tendency to think we can just figure things out. It is not uncommon to go through a thought process that basically boils down to: coding must be the hardest job, so all the other jobs can be done by someone with coding skills.
Fight the fear, let go of control, and make moves towards a well-rounded organization.
If you’ve ever tried some simple home repairs or paint touchup you know this logic doesn’t work—you only need to spend an hour watching some cable TV DiY show and you can see how the people with skills are always unraveling the messes created by those who thought they could improvise. The software equivalent can sometimes be seen as a developer attempting to design the user interaction flow in a paint program or PowerPoint. Sure it can be done by a developer (and there are talented developers who can of course do it all), but he or she can quickly reach their limits, and so will the user interaction.
Open up your engineer’s mind to embrace the truth that every other discipline or function you will ever collaborate with has a deep set of skills experiences that you lack. Relative to engineering, the “softer” skills often pose the biggest eye-opening surprise to engineers. Until you’ve seen the magic worked by those skilled in marketing, communications, sales, business development or a host of other disciplines you might not appreciate the levels of success you can achieve by turning over the task to trained professionals.
I have seen this first-hand many times. Most recently it occurred while working with the a16z portfolio company Local Motion when it came time to do some of the early announcements around the fleet-management company. The co-founders possess engineering and design backgrounds from elite institutions, and built the product themselves, hardware and software. Both are experienced mountaineers, and so they have this engrained sense of self-sufficiency, which is valuable both for building companies and scaling mountains.
When it came time to work with the industry press to tell the story of their company, in some ways they had to suppress their self-sufficient instincts. The founders were self-aware enough to know they had not done this before and agreed to enlist the help of those who have depth and breadth of experience. The pros showed up and spent time learning the team, the business, and the story (professionals do that!). They came back with a plan, roles, responsibilities, and defined what success would look like. It was amazing to watch how the founders absorbed and learned at each step all those things which they had not personally experienced before.
This sounds easy and pretty obvious. But if you put yourself in their shoes you know that this is not just bringing in a hired gun to get some press, rather this is hiring on a new member of the team and a new founding member of the family. What is vital to keep in mind, is that this kind of work is as important as every line of code and every circuit board. The lesson of letting go and letting professionals do their work is clear: delegating is never easy for most, but is spectacularly difficult if you don’t know what the other person is going to do and when the outcome matters a whole lot. Still, you need to let the specialists into your carefully engineered world.
There are moments of terror. You’re watching people talk about your product using tools and techniques you are unfamiliar with to connect with your potential customers. Even though it is a product, you are apt to feel as though this is a discussion about yourself. You question every step. You doubt the skills of the person you hired. You are certain everything will go wrong.
It is at that point—right when you start to panic and think that unless you do this yourself things will fail—that you need to let go. You need to say “yes” to hiring a person to do the work, and then let them do their best work.
Just keep reminding yourself that you’ve never done the job before, and that you’re role is to hire someone who knows more than you. Even when you’ve wrapped your head around that, there are a few ways you can get tripped up when you are The key to success here is avoiding these mistakes when you are in the hiring process:
- Asking candidates to teach you. A good candidate will of course know more than you. Their interview is not a time for them to teach you what they do for a living. The interview is for you to learn the specifics of a given candidate, not the job function. The best bet is to do your homework. If you’re hiring your first sales leader then use your network and talk to some subject matter experts and learn the steps of the role ahead of time.
- Expecting a candidate to know or create your strategy. It is fine to expect engineering candidates to know the tools and techniques you use. You wouldn’t expect an engineering candidate to know your unannounced product, of course. It is equally challenging to expect a new marketing person to have a marketing plan for your product. Even if you ask them to brainstorm for hours, keep in mind the inputs into the process—they only know the specifics you have provided them. For example, don’t expect a marketing candidate to magically come up with the right pricing strategy for your product without a chance to really dive in. On the other hand, you can expect a candidate to walk you through in extreme detail their most recent work on a similar topic. You can get to their thought process and how they worked through the details of the problem domain.
- Interviewing too many folks. You will always hear stories about the best hire ever after seeing 100 people. Those stories are legendary. On the other hand, you rarely hear the stories that start with “we could not find the perfect QA leader so we waited and waited until we had a quality crisis.” Yet these latter stories happen far too often. Again, you should not compromise, but if after bringing a dozen or more people through a process you are still searching, consider the patterns you’re seeing and why this is happening. A good practice if you’ve not found right hire after going through a lot of folks is to bring in a new point of view. Consider recruiting the help of a search firm, a board member, or a subject matter advisor to get you over the first hire in a new job function. Do you need the help of a search firm? Would you benefit from the help of a board member or subject matter advisor to get you over the first hire in a new job function?
These add up to the quest for the perfect hire. When it comes to engineering you give yourself a lot of leeway because you feel you can direct a less experienced person and because you can gauge more easily what they know and don’t know. When it comes to other roles you become more reluctant to let go of a dream candidate. This almost always nets out costing you time, and in a new effort time is money. That isn’t saying to settle, but it is saying to use the same techniques of approximation you naturally use when hiring people in your comfort zone.
The most difficult part of hiring for a job you don’t know first-hand is the human side. Every growing organization needs diversity because every product and service is used by a diverse group of people. The different job functions often bring with them diversity of personality types that add to the challenges of hiring. The highly analytical developer looking to hire a strong qualitative thinker for marketing, or the highly empathetic sales leader, is often going to face a challenge just making the human connection.
This human connection is a two-way street. Embrace it. Recognize the leap each of you are taking. Realize that the interpersonal skills required to call on customers every day are just different than the interpersonal skills used when hacking. The challenge of making that human connection is one for the person doing the hiring to overcome. Often that’s the biggest opportunity for personal growth when hiring people to do a job you can’t.
–Steven Sinofsky (@stevesi)
Note: A form of this post originally appeared on FastCo.
Hunter Walk shared some thoughts on notifications and the challenges he and (certainly through twitter) many people see. Many of the companies I’ve met with see design challenges in how much and when to offer up notifications. There’s a long history of trying different approaches and modalities to notifications and so it seems worth some additional perspective for those familiar with what we see today on modern mobile platforms.
Notifications are one of those features where everyone has an opinion, and rightfully so. The feature is so visible and for just about everyone seems so close to being helpful but yet always off by just a little. There’s a general UX principle that is worth considering, which is anytime you push some feature on your customer you really want it to be right (correct, useful, helpful) for him/her 100% of the time. If not, chances are your customer will recall the negatives of the feature far more than the positives. This applies to notifications, autocorrect, form completion, and more. If you find yourself putting a lot of design energy into how your customer can undo or dismiss your best guess at what was intended, then you’re probably being too aggressive.
Anytime you push some feature on your customer you really want it to be right for him/her 100% of the time.
In some ways, today’s Notification Centers are the extreme case of “we’re collectively going to be wrong so often that we’re just going to put stuff in one place” or “there’s just so much we app developers have to tell you that the platforms are squeezing it all into one place to avoid cluttering up the platform itself”.
It is as if it isn’t enough we have to manage all of our apps, bookmarks, and preferences, we now have to manage all the ways apps tell us stuff we might want to know. Hunter’s raised the complaint (to a chorus of agreement) that after a two weeks, you have to go in and turn off notifications on a new app. Of course this improves battery life, reduces chatter, and then as some noted causes you to start to ignore the app.
What’s an app developer to do?
What to do?
In the PC era a lot of effort went into honing the design of verbs and interaction. It took a decade to develop the right approaches to menus, toolbars, status bars, panes, and more. That’s because many apps were essentially a large set of verbs this was the big design challenge. The rough equivalent to this design challenge is the role of notifications in mobile. That’s because in a mobile world many apps exist to be essentially a stream of information.
Notifications suffer a clear tension between platform pm and the app pm.
Platform pm wants to contain apps to the app experience, extending the walled-garden such that apps don’t interfere with other apps. By definition a notification is a way for one app to interfere with other apps. Platform pm sees notifications as necessary far less frequently than app pm might. This leads to a set of APIs that offer a clear, albeit limited, view of what a notification is and what it can do. This seems reasonable and we all want the platform folks maintaining a global view of consistency in approach and control.
App pm sees the world through the lens of their app. The assumption is that someone downloading and using an app has made a choice to count on that app for the purpose it was designed, and so whether it is an airline app alerting you to a flight (or a potential discount on a future flight), a bank with a balance alert (or advertising a new bank feature), or a communication tool letting you know of some inbound message (or alerting you that a friend is now on line) the app pm sees any of these as worthwhile reasons to interfere with your flow or context. On all platforms, apps often design their own type of notifications that get used while you are using the app (for in app purchase or feature advertising) because the platform is not rich enough. All this seems legitimate, and certainly at the time of initial designs.
Over time the initial designs from both parties tend to lead to an expansion in the ability to interrupt you. Each subsequent release of a platform almost always adds more capabilities to enhance and customize notifications in an effort to offer more while also trying to keep the noise in the system manageable. Each app adds more and more notifications in an effort to more deeply engage with customers and likely to encourage customers to use more surface area of the app.
Many who use iOS 7 have spent quite a bit of time mired notification customization. Here is an overview of the iOS 7 features, both the notifications and notification center, worth a look if you’re on Android or not sure of the impressive depth that Apple designed for notification. Android is probably not quite at the same level of consistency and control, though as you might expect there are several apps that can help you customize notifications of other apps.
At the extreme we end up with two core design challenges.
Notification spam. This one is easy. Too many apps just think too much of what is going on is important to you. Like too much of any design the burden falls to app product managers to just be more thoughtful. Like so many elements of any platform, when there is a view that making money depends on getting folks to use more of a product or spend more time in a product, the platform starts to look a bit like “surface area to be exploited”. Like the old Start Menu and desktop in Windows, the more places an app can “infuse” itself and invade your space the better. On Android we see this in the share item menu as another bit of surface area to be gamed or exploited.
Notification action. The most common issue with notification is that your flow is interrupted and then you seem to be pulled into a state of distraction until you deal with the inbound notice. We each have our own human based algorithms for how to cope. We always jump on SMS. We (almost) always ignore a ringing phone. We wish we could find a way for some app or another to stop bugging us so we uninstall it.
On iOS there is very little you can do to a notification other than dismiss it or just jump directly to the app or place in the app that generated the notification. Modal or must-act notifications are generally discouraged. The resulting notification center then turns into a list to read that spans a bunch of apps and for some ends up to be a list of stuff you’ve already seen popup or in context or just a reminder to get to the app.
On Android, the design takes a different approach which is to enable notifications that can take actions. This is where the elegance of notifications is really stretched, but in a way that many find appealing. For example, when you receive a new mail message the gmail notification lets you archive it based on reading the initial content or various SMS clients offer you the ability to reply.
On Windows Phone, one has the additional option of pivoting notifications by person on your home screen so you can glance and see that there is activity by person. This has a natural appeal when there are a small set of folks you care deeply about but as a general purpose mechanism it might not scale particularly well.
The core challenge with offering verbs with notifications is almost “classic” in that one can never off the right set of verbs because eventually the design turns into attempting to implement the a substantial number of features of the app in the notification. Mail is a great challenge: delete, file, reply, flag, etc. all become possible verbs. Each usage pattern in aggregate leads to the whole mail experience. The more users you have the larger the group of customers that don’t like the subset of verbs you picked.
Ultimately taking action based on a notification turns into a bit of a frustration in that the notification centers essentially offer a new way to launch all your apps. What was a nice feature turns into a level of indirection almost all the time.
Therein is the opportunity. In a world where many people are almost constantly glancing at their phones and wanting to know more about what is going on in their digital lives and a world where almost every app represents an endless stream of information along with in-app notifications, it seems that notifications need a different level of semantics.
For example, with just a few friends Facebook always has something new to see so why notify you of the obvious. For many, Twitter is essentially a notification engine. Mail certainly is a constant stream, arguably of decreasing importance. In other words, it isn’t even clear what makes sense to notify you about when the natural behavior is to periodically launch apps to see what’s new within the app context and apps are generating new information all the time.
Similarly, if most everyone knows that when you are talking to another human you both have to turn your phones upside down to avoid being distracted (or sharing private information), then there’s a good chance we’ve collectively missed the mark notifications. The iOS “do not disturb” is an awesome feature but yet it seems to undo all the work in both the notification center and in the apps.
My view is that a feature that requires us to customize it before it becomes useful or less annoying is defaulted the wrong way. Of course this is literally impossible with a product used by more than a few people, since any design at all will have both critics and shortcomings. However, it is possible to default to “out of the way” and then provide a mechanism for people to decide what they might want to be notified about once a usage pattern is established.
For example, I might assert that for an app like mail, sms, Facebook, or Twitter the simple iOS badge is enough. We are all in and out of these apps enough during the day that a specific notification is redundant with the in-app notifications already there.
Each app can almost certainly step back and either know a priori or offer a mechanism that puts people in control of their experience with notifications. It is almost certainly the case that if we’re bouncing in and out of apps all the time but really do want to know if SMS comes from a loved one amongst the 100′s of SMS many get each day, that is likely the way to design a feature.
It is easy to imagine using more context (loved the twitter suggestion to not notify while driving/moving fast). It is easy to imagine more machine learning applied to notifications. But I think we can start from a fresh perspective that the mechanisms provided are just being over-used to begin with when we look at modern usage patterns.
–Steven Sinofsky (@stevesi)
How often do you hear things like “let’s ban email”, “no more attachments”, “death to PowerPoint decks”, “we’re going paperless”, “meeting free friday” or one of dozens of “bans” designed to do away with something that has become annoying or inefficient in the workplace? If you’re around long enough you can see just about anything cross over from innovative new tool to candidate to be banned. The problem is that banning a tool (or process) in an attempt at simplification never solves the problem. Rather, one should to look at a different approach, an approach that focuses on the work not the tool or process.
What’s the problem?
It is well understood that new technologies go through an adoption curve. In the classic sense it is a normal distribution as described by researchers in the 1950′s. More recently and generally cited in the software world is Geoffrey Moore’s Crossing the Chasm which describes a slightly different path. These models all share a common view of a group of early adopters followed by a growing base of users of a technology.
While adoption is great, we are all too used to experiencing excess enthusiasm for new technologies. As a technology spreads, so does the enthusiasm. Invariably some folks use the technology to the the point of abusing it. From reply all to massive attachments to elaborate scorecards with more dimensions than anyone can understand, the well-intentioned enthusiastic user turns a game-changing tool into a distraction or worse.
Just as with adoption curves, once can create a conceptual “irritation curve” and overlay it with adoption. Of course what is pictured below is not based on any data or specific to any technology, but consistent with our collective anecdotal point of view.
The key is that at some point the adoption of a new product crosses the chasm and becomes widely used within a company. While there is a time delay, sometimes years, at some point the perceived “abuse” of the technology causes a cross-over where for some set of people the irritation outpaces the utility. Just as there are early adopters, there are also irritation canaries who are the first to feel the utility of the new technology declining with increased usage.
We see this same dynamic not just for tools, but for business processes as well. That status report, dashboard, or checkin mail all start off as well-intentioned and then after some period of time the “just one more thing” or spreading over-usage at all levels of a team turn a positive into a burden.
Then at some point people start to reject the tool or process. Some even call for an outright ban or elimination.
What’s the solution?
The way to break the cycle is to dive into the actual work and not the tool. Historically, tools fade away when the work process changes.
It is tough to find examples of popular tools and processes that were simply banned that did not make a comeback. Companies that ban meetings or email on fridays just have more meetings and email on monday-thursday. I’ve personally seen far too many examples of too much information crammed on to a page (smaller fonts or margins anyone) or slides that need to be printed rather than projected in an effort to squeeze more on a page when there are forced limits on story-telling.
On the other hand, from voice mail to fax machines to pagers to typewriters to voice calls we have examples of tools that achieve high and subsequently irritating usage levels can and do go away because new tools take over. If you were around for any of those then you know that people called for them to be banned and yet they continued, until one day we all just stopped using them.
A favorite historical example is a company that told me they removed all the typewriters when PCs were introduced. The company was trying to save time because typewriters were much more difficult to use than PCs with printers (of course!). The problem was immediately seen by those responsible for the workflows in the company–all of a sudden no one could fill out an expense report, transfer to another department, or pay an invoice. All of these work processes, the blizzard of paperwork that folks thought were caused by typewriters, were rendered inoperable. These processes all required a typewriter to fill out the form and the word processors had no way of navigating pre-printed forms in triplicate. Of course what needed to happen was not a pre-printed form that worked in a word processor (what the administrative folks asked for), but a rethinking of the workflow that could be enabled by new tools (what management needed to do).
This sort of rethinking of work is what is so exciting right now. It is fair to say that the established, and overloaded, desktop work-processes and tools of the past 20 years are being disrupted by a new generation of tools. In addition to re-imagining how work can be done to avoid the problems of the past, these tools are built on a modern, mobile, cloud, and social infrastructure.
For example, Tom Preston-Werner, co-founder of GitHub, tells a great story about the motivations for GitHub that echoes my own personal experience. As software projects grew the communication of code changes/checkins generate an overwhelming blizzard of mail. Rather than just shut down these notifications and hope for the best, what was needed was a better tool so he invented one.
At Asana, Dustin Moskovitz, tells of their goal to eliminate email for a whole set of tracking and task management efforts. We’ve all seen examples of the collaborative process playing out poorly by using email. There’s too much email and no ability to track and manage the overall work using the tool. Despite calls to ban the process, what is really needed is a new tool. So Asana is one of many companies working to build tools that are better suited to the work than one we currently all collectively seem to complain about.
Just because a tool is broadly deployed doesn’t mean it is the right or best way to work.
We’re seeing new tools that are designed from the ground up to enable new ways of working and these are based on the learning from the past two decades of tool abuse.
What are some warning signs for teams and managers?
It is easy to complain about a tool. Sometimes the complaints are about the work itself and the tool is just the scapegoat. There’s value in looking at tool usage or process creation from a team or management perspective. My own experience is that the clarion calls to ban a tool or process have some common warning signs that are worth keeping an eye out for as the team might avoid the jump to banning something, which we know won’t work.
- Who is setting expectations for work product / process? If management is mandating the use of a tool the odds of a rebellion against it go up. As a general rule, the more management frames the outcome and the less the mechanism for the outcome the more tolerance there will be for the tool. Conversely, if the team comes up with a way of working that is hard for outsiders to follow or understand, it is likely to see pushback from partners or management. However, if it is working and the goal is properly framed then it seems harmless to keep using a tool. Teams should be allowed to use or abuse tools as they see fit so long as the work is getting done, no matter how things might look from outside.
- Does the work product benefit the team doing the work or the person asking? A corollary to above is the tool or process that is mandated but seems to have no obvious benefit is usually a rebellion in-waiting. Document production is notorious for this. From status reports to slides to spreadsheets, the specification by management to create ever more elaborate “work products” for the benefit of management invariably lead to a distaste for the tool. It is always a good idea for management to reduce the need to create work, tools, and processes where the benefit accrues to management exclusively. Once again, the members of the team will likely start to feel like banning the use of the tool is the only way to ease the overload or tax.
- Do people get evaluated (explicitly or implicitly) on the quality of the work product/process or the end-result? A sure-fire warning sign to the looming distaste of a tool or process is when a given work product becomes a goal or is itself measured. Are people measured by the completion of a report? Does someone look at how many email notifications get generated by someone? Does someone get kudos for completing a template about the group’s progress? All of these are tools that might be considered valuable in the course of achieving the actual goals of the team, but are themselves the path along the way. Are your status reports getting progressively more elaborate? Are people creating email rules to shunt email notifications to a folder? Are people starting to say “gosh I must have missed that”? All of those are warning signs that there is an impending pushback against the tool or process.
- What doesn’t get done if you just stop? The ultimate indicator for a need to change a tool or process is to play out what would happen if you really did ban it. We all know that banning email is really impractical. There are simply too many exceptions and that is exactly the point. Many tools can have a role in the modern workplace. Banning a tool in isolation of the work never works. Taking a systematic look at the work required that uses a tool, those that use the tool, and those that benefit from the output is the best way to approach the desire to use the most appropriate toolset in the workplace.
What tools need to change in your organization? What work needs to change so that the team doesn’t need to rely on inappropriate or inefficient tools?
PS: As I finished writing this post, this Forrester report came across twitter: Reality Check: Enterprise Social Does Not Stem Email Overload.
For me, 1984 was the year of Van Halen’s wonderful [sic] album, The Right Stuff, and my second semester of college. It would also prove to be a time of enlightenment for me and computing. On this 30th anniversary of the Apple Macintosh on January 25 and the Superbowl commercial on January 22. I wanted to share my own story of the way the introduction of the Macintosh profoundly changed my path in life.
Perhaps a bit indulgent, bit it seemed worth a little backstory. I think everyone from back then is feeling a bit of nostalgia over the anniversary of the commercial, the product, and what was created.
High School, pre-Macintosh
Like many Dungeons and Dragons players my age, my first exposure to post-Pong computing was an Atari 800 that my best friend was lucky enough to have (our high school was not one to have an Apple ][ which hadn’t really made it to suburban Orlando). While my friends were busy listening to the Talking Heads, Police, and B-52s, I was busy teaching myself to program on the Atari. Even though it had the 8K BASIC cartridge it lacked tape storage. Every time I went over to use the computer I had to start over. Thinking about business at an early age (I suppose) I would continue to code and refine what I thought would be a useful program for our family business, the ability to compute sales tax on purchases from different states. Enter the total sale, compute the sales tax for a state by looking up the rate in a table.
My father, an entrepreneur but hardly a technologist, was looking to buy a computer to “automate” our family business. In 1981, he characteristically dove head first into computing and bought an Osborne I. For a significant amount of money ($1,795, or $4,600 today) we owned an 8 bit CPU and two 90K floppy drives and all (five) of the business programs one could ever need.
I started to write a whole business suite for the business (inventory, customers, orders) in BASIC which is what my father had hoped I would conjure up (in between SATs and college prep). Well that was a lot harder than I thought it would be (so were the SATs). Then I discovered dBase II and something called a “database” that made little sense to me in the abstract (and would only come to mean something much later in my education). In a short time I was able to create a character-based system that would be used to run the family business.
To go to college I had a matching Osborne I with a 300b modem so I could do updates and bug fixes (darn that shipping company–they changed the rate on COD shipments right during midterms which I had hard-coded!).
College Fall Semester
I loaded up the Osborne I and my Royal typewriter/daisy wheel/parallel port “letter quality” printer and was off to sunny Ithaca.
Computer savvy Cornell issued us our “BITNET electronic mail accounts”, mine was TGUJ@CORNELLA.EDU. Equal parts friendly, memorable, and useful and no one knew what to do with them. The best part was email ID came printed on a punch card. As a user of an elite Osborne I felt I went back in time when I had to log on to the mainframe from a VT100 terminal. The only time I ever really used TGUJ was to apply for a job with Computer Services.
I got a job working for the computer services group as a Student Terminal Operator (STO). I had two 4 hour shifts. One was in the main computer science major “terminal room” in Upson Hall featuring dozens of VT100 terminals. The other shift was Friday night (yes, you read that correctly) at the advanced “lab” featuring SGI graphics workstations, IBM PC XTs, an Apple Lisa, peripherals like punch card machines, and a 5′ tall high-speed printer. For the latter, I was responsible for changing the ribbon, a task that required me to put on a mask and plastic arm-length gloves.
It turned out that Friday night was all about people coming in to write papers on the few IBM/MS-DOS PCs using WordPerfect. These were among the few PCs available for general purpose use. I spent most of the time dealing with graduate students writing dissertations. My primary job was keeping track of the keyboard templates that were absolutely required to use WordPerfect. This experience would later make me appreciate the Mac that much more.
In the computer science department I had a chance to work on a Xerox Star and Alto (see below) along with Sun Workstations, microVAX mini, and so on. The resources available were an incredible blessing to the curious. The computing world was a cacophony of tools and platforms with the vast majority of campus not yet tapping into the power of computing and those that did were using what was most readily accessible. Cornell was awash in the sea of different computing platforms, and to my context that just seemed normal, like there were a lot of different types of cars. This was especially apparent from my vantage point in the computer facilities.
One experience with a new, top-secret, computer was about to change all that.
I ended up getting to use a new computer from an unidentified company. One night after my shift, a fellow STO dragged me back to Upson Hall and took me into a locked room in the basement. There I was able to see and use a new computer. It was a wooden box attached to a wall with an actual chain. It had a mouse, which used on the Xerox and Sun workstations. It had a bitmap screen like a workstation. It had an “interface” like the Xerox. There was a menu bar across the top and a desktop of files and folders. It seemed small and much more quiet than the dorm-refrigerator sized units I was used to hearing.
What was really magical about it was that it had a really easy to use painting program that we all just loved. It had a “word processor”. It was much easier to use than the Xerox which had special keys and a somewhat overloaded desktop metaphor. It crashed a lot even after a short time using it. It also started up pretty quickly. Most everything we did with it felt new and different compared to all the other computers we used.
The end of the semester and exams approached. The few times and couple of hours I had to play with this computer were exciting. In the sea of computing options, it was definitely the most exciting thing I had experienced. Perhaps being chained to the wall added to the excitement, but there was something that really resonated with us. When I try to remember the specifics, I mostly recall an emotional buzz.
My computing world was filled with diversity, and complexity, which left me unprepared for the way the world was going to change in just the next six weeks.
To think about Apple’s commercial, one really has think about the context of the start of the year 1984. The Orwellian dialog was omnipresent. Of course as freshman in college we had just finished our obligatory compare/contrast the dystopian messages in Animal Farm, Brave New World, and 1984 not to mention the Cold War as front and center dialog at every turn. The country emerging from recession gave us all a contrasting optimism.
At the same time, IBM was omnipresent. IBM was synonymous with computing. Sure the Charlie Chaplin ads were great, but the image of computing to almost everyone was that of the IBM mainframe (CORNELLA was located out by the Ithaca airport). While IBM was almost literally the pillar of innovation (just a couple of years later, scientists at IBM would spell IBM with Xenon atoms), there was also great deal of distrust given the tenor of the time. The thought of a globally dominant company, a computer company, was uncomfortable to those familiar with fellow Cornellian Kurt Vonnegut’s omnipresent RAMJAC.
Then the Apple commercial ran. It was truly mesmerizing (far more so to me than the Superbowl). It took me about one second to stitch together all that was going on right before my eyes.
Apple was introducing a new computer.
It was going to be a lot different from the IBM PC.
The world was not going to be like 1984.
And most importantly, the computer I had just been playing with weeks earlier was, in fact, the Apple Macintosh.
I was so excited to head back to the terminal rooms and talk about this with my fellow STOs and to use the new Apple Macintosh.
Upon returning to the terminal room in Upson, Macs had already started to replace VT100s. First just a couple and then over time, terminal access moved to an emulation program on Macs (rumor had it that the Macs were actually cheaper than terminals!).
My Friday night shift was transformed. Several Macs were added to the lab. I had to institute a waiting list. Soon only the stalwarts were using the PCs. I started to see a whole new crowd on those lonely computer nights.
I saw seniors in Arts & Sciences preparing resumes and printing them on the ImageWriter (note, significantly easier to change the ribbon, which I had to do quite often every night). Those in the Greek System came by for help making signs for parties. Students discovered their talent with MacPaint pixel art and fat bits. All over campus signs changed overnight from misaligned stencils to ImageWriter printouts testing the limits of font faces per page.
I have to admit, however, I spent an inordinate amount of time attempting to recover documents that were lost to memory corruption bugs on the original MacWrite. The STOs all developed a great trouble shooting script and signs were posted with all sorts of guesses (no more than 4 fonts per document, keep documents under 5 pages, don’t use too many carriage returns). We anxiously awaited updates and students would often wait in line to update their “MacWrite disks” when word spread of an update (hey, there was no Internet download).
In short order, Macintosh swept across campus. Cornell along with many schools was part of Apple’s genius campaign on campuses. While I still had my Osborne, I was using Macintosh more often than not.
The next couple of years saw an explosion of use of Macintosh across campus. The next incoming class saw many students purchasing a Mac at the start of college. Research funds were buying Macs. Everywhere you looked they were popping up on desks. There was even a dedicated store just off campus that sold and serviced Macs. People were changing their office furniture and layout to support using a mouse. Computer labs were being rearranged to support local printers and mice. The campus store started stocking floppy disks, which became a requirement for most every class.
Document creation had moved from typewriters and limited use of WordPerfect to near ubiquitous use of MacWrite practically by final exams that Spring. Later, Microsoft Mac Word, which proved far more robust became the standard.
The Hotel School’s business students were using Microsoft Mac Excel almost immediately.
The Chemistry department made a wholesale switch to Macintosh. The software was a huge driver of this. It is hard to explain how difficult it was to prepare a chemistry journal article before Macintosh (the department employed a full time molecular draftsman to prepare manuscripts). The introduction of ChemDraw was a turning point for publishing chemists (half my major was chemistry).
It was in the Chemistry department where I found a home for my fondness of Macintosh and an incredibly supportive faculty (especially Jon Clardy). The research group had a little of everything, including MS-DOS PCs with mice which were quite a novelty. There were also Macs with external hard drives.
I also had access to MacApp and the tools (LightSpeed Pascal) to write my own Mac software. Until then all my programming had been on PCs (and mainframes, and Unix). I had spent two summers as an intern (at Martin Marietta, the same company dBase programmer Wayne Ratliff worked!) hacking around MS-DOS writing utilities to do things that were as easy as drag and drop on a Mac or just worked with MacWrite and Mac Excel. As fun as learning K&R, C, and INT 21h was, the Macintosh was calling.
My first project was porting a giant Fortran program (Molecular Mechanics) to the Mac. Surprisingly it worked (perhaps today, equally surprising was the existence of a Fortran compiler). It cemented the lab’s view that the Macs could also be for work, not just document creation. Next up I just started exploring the visualizations available on the Mac. Programming graphics was all new to me. Programming an object-oriented event loop seemed mysterious and indirect to me compared to INT 21h or stdio.
But within a few hacking sessions (fairly novel to the chemistry department) the whole thing came together. Unlike all of the previous systems I used, the elegance of the Mac was special. I felt like the more I used it the more it all made sense. When I would bury myself in Unix systems programming it seemed more like a series of things, tricks, you needed to know. Macintosh felt like a system. As I learned more I felt like I was able to guess how new things would work. I felt like the bugs in my programs were more my bugs and not things I misunderstood.
The proof of this was that through the Spring semester my senior year I was able to write a program that visualized the periodic table of the elements using dozens of different variables. It was a way to explore periodicity of the elements. I wrote routines for an X-Y plot, bar charts, text tables, and the pièce de résistance was a 2.5-dimensional perspective of the periodic table showing a single property (commonly used to illustrate the periodic nature of electron affinity). I had to ask a lot of friends who were taking computer graphics on SGIs for help! Still, not only had I just been able to program another new OS (by then this was my 5th or 6th) but I was able to program a graphical user interface for the first time.
MacMendeleev was born.
The geek in all of us has that special moment when at once you feel empowered and marvel at a system. That day in the spring of 1987 when I rendered a perspective drawing from my own code on a system that I had seen go from a chained down plywood box to ubiquity across campus was magical. Even my final report for the project was, to me, a work of art.
The geek in all of us has that special moment when at once you feel empowered and marvel at a system.
It wasn’t just the programming that was possible. It wasn’t just the elegance and learnability of the system. It wasn’t even the ubiquity that the Macintosh achieved on campus. It was all of those. Most of all it represented a tool that allowed me to realize some of my own potential. I was awful at Chemistry. Yet with Macintosh I was able to contribute to the department and probably showed a professor or two that in spite of my lack of actual chemistry aptitude I could do something (and dang, my lab reports looked amazing!). I was, arguably, able to learn some chemistry.
I achieved with Macintosh what became one of the most important building blocks in my education.
I’m forever thankful for the empowerment that came from using a “bicycle of the mind”.
I’m forever thankful for the empowerment that came from using a “bicycle of the mind”.
What came next
Graduate school diverged in terms of computing. We used DEC VMS, though SmallTalk was our research platform. So much of the elegance of the Macintosh OS (MacApp and Lisa before that) was much clearer to me as I studied the nuances of object-oriented programming.
I used my Macintosh II to write papers, make diagrams, and remote into the microVAX at my desk. I also used Macintosh to create a resume for Microsoft with a copy of Microsoft Word I won at an ACM conference for my work on MacMendeleev.
I also used Macintosh to create a resume for Microsoft with a copy of Microsoft Word…
When I made it to Microsoft I found a great many shared the same experience. I met folks who worked on Mac Excel and also had Macs in boxes chained to tables. I met folks who wrote some of those Macintosh programs I used in college. So many of the folks in the “Apps” team I was hired into that year grew up on that unique mixture of Mac and Unix (Microsoft used Xenix back then). We all became more than converts to MS-DOS and Windows (3.0 was being developed when I landed at Microsoft).
There’s no doubt our collective experiences contribute to the products we each work on. Wikipedia even documents the influence of MacApp on MFC (my first product contribution), which was by design (and also by design was where not to be influenced). It is wonderful to think that through tools like MFC and Visual Basic along with ubiquitous computing, Windows brought to so many young programmers that same feeling of mastery and empowerment that I felt when I first used Macintosh.
Fast-forwarding, I can’t help but think about today’s college students having grown up hacking the web but recently exposed as programmers to mobile platforms. The web to them is like the Atari was to me—programmable, understandable, and fun. The ability to take your ideas, connect them to the Internet, touch your creation, and make your own experience must feel like building a Macintosh program from scratch felt like to me. The unique combination of mastery of the system, elegance of design, and empowerment is what separates a technology from a movement.
Macintosh certainly changed my path in life…
For me, Macintosh was an early contributor to my learning, skills, and ultimately my self-confidence. Macintosh certainly changed my professional path in life. For sure, 1984 was not at all like 1984 for me.
Yes, of course I’m a PC (and definitely a Surface). Nothing contributed more to my professional life than the PC!
PS: How far have we come? Check out this Computer Chronicles from 1985 where the new Macintosh is discussed.
I was talking with a founder/CEO of an enterprise startup about what it is like to disrupt a sizable incumbent. In the case we were talking about the disrupting technology was losing traction and the incumbent was regaining control of the situation, back off their heels, and generally felt like they had fended off the “attack” on a core business. This causes a lot of consternation at the disrupting startup as deals aren’t won, reviews and analyst reports swing the wrong way, and folks start to question the direction. If there really is a product/market fit, then hold on and persevere because almost always the disruption is still going to happen. Let’s look at why.
The most important thing to realize about a large successful company reacting to a disruptive market entry is that every element of the company just wants to return to “normal” as quickly as possible. It is that simple.
Every action about being disrupted is dictated by a desire to avoid changing things and to maintain the status quo.
If the disruption is a product feature, the motion is figuring out how to tell customers the feature isn’t that important (best case) or how to quickly add something along the lines of the feature and move on (worst case). If the disruption is a pricing change then every effort is about how to “manage customers” without actually changing the price. If the disruption is a new and seemingly important adjacent product, then the actions focus on how to point out that such a product isn’t really necessary. Across the spectrum of potential activities, it is why the early competitive responses are often dismissive or outwardly ignore the challenger. Aside from the normal desire to avoid validating a new market entry by commenting, it takes a lot of time for a large enterprise to go through the work to formulate a response and gain consensus. Therefore an articulate way of changing very little has a lot of appeal.
Status quo is the ultimate goal of the incumbent.
Once a disruptive product gains enough traction that a more robust response is required, the course of action is almost always one that is designed to reduce changes to plans, minimize effort overall, and to do just enough to “tie”. Why is that? Because in a big company “versus” a small company, enterprise customers tend to see “a tie as a win to the incumbent”. Customers have similar views about having their infrastructure disrupted and wish to minimize change, so goals are aligned. The idea of being able to check off that a given scenario is handled by what you already own makes things much easier.
Keep in mind that in any organization, large or small, everyone is at or beyond capacity. There’s no bench, no free cycles. So any change in immediate work necessarily means something isn’t going to get done. In a large organization these challenges are multiplied by scale. People worry about their performance reviews; managers worry about the commitments to other groups; sales people worry about quarterly quotas. All of these worries are extremely difficult to mitigate because they cross layers of managers and functions.
As much as a large team or leader would like to “focus” or “wave a wand” to get folks to see the importance of a crisis, the reality of doing so is itself a massive change effort that takes a lot of time.
This means that the actions taken often follow a known pattern:
- Campaign. The first thing that takes place is a campaign of words and positioning. The checklist of features, the benefits of the existing product, the breadth of features of the incumbent compared to the new product, and so on. If the new product is cheaper, then the focus turns to value. Almost always the campaign emphasizes the depth, breadth, reliability, and comfort of the incumbent’s offer. A campaign might also be quite negative and focus on a fear, compatibility with existing infrastructure, or conventional wisdom weakness of a disruptor, or the might introduce a pretty big leap of repositioning of the incumbent product. A good example of this is how on-premises server products have competed with SaaS by highlighting the lack of flexibility or potential security issues around the cloud. This approach is quick to wind up and easy to wind down. Once it starts to work you roll it out all over the world and execute. Once the deals are won back then the small tiger team that created the campaign goes back to articulating the product as originally intended, aka normal.
- Partnership. Quite often there can be a competitive response of best practices or a third-party tool/add-on that appears to provide some similar functionality. The basic idea is to use someone else to offer the benefit articulated by a disruptive product. Early in the SaaS competition, the on-premises companies were somewhat quick to partner with “hosting” companies who would simply build out a dedicated rack of servers and run the traditional software “as a service”. This repotting plants approach to SaaS has the benefit that once the immediate crisis is mitigated, either the need to actually offer and support the partnership ends or the company just becomes committed to this new sales channel for existing products. Again, everything else continues as it was.
- Special effort. Every once in a while the pressure is so great internally to compete that the engineering team signs up for a “one off” product change or special feature. Because the engineering team was already booked, a special effort is often something carefully negotiated and minimized in scope and effort. Engineering minimizes it internally to avoid messing up dependencies and other features. Sales will be specific in what they expect the result to do because while the commitment is being made they will likely begin to articulate this to red-hot customer situations. At the extreme, it is not uncommon for the engineering team to suggest to the sales organization that a consultant or third-party can use some form of extensibility in the product to implement something that looks like the missing work. The implications of doing enterprise work in a way that minimizes impact is that, well, the impact is minimized. Without the proper architecture or an implementation at the right level in the stack, the effort ultimately looks incomplete or like a one-off. Almost all the on-premise products attempting to morph into cloud products exhibit this in the form of features that used to be there simply not being available in the “SaaS version”. With enough wins, it is almost likely that the special effort feature doesn’t ever get used. Again, the customer is just as likely to be happy with the status quo.
All of these typical responses have the attribute that they can be ignored by the vast majority of resources on a business. Almost no one has to change what they are doing while the business is responding to a disruptive force. Large incumbents love when they can fend off competitors with minimal change.
Large incumbents love when they can fend off competitors with minimal change.
Once the initial wave of competitive wins settles in and the disruptive products lose, there is much rejoicing. The teams just get back to what they were doing and declare victory. Since most of the team didn’t change anything, folks just assume that this was just another competitor with inferior products, technology, approaches that their superior product fended off. Existing customers are happy. All is good.
Or is it?
This is exactly where the biggest opportunity exists for a disruptive market entry. The level of complacency that settles into an incumbent after the first round of victories is astounding. There’s essentially a reinforcing feedback loop because there was little or no dip in revenue (in fact if revenue was growing before then it still is), product usage is still there, customers go back to asking for features the same as they were before, sales people are making quota, and so on. Things went back to normal for the incumbent.
In fact, just about every disruption happens this way–the first round or first approaches don’t quite take hold.
Why is this?
- Product readiness can improve. Obviously the most common is that the disruptive product simply isn’t ready. The feature set, scale, enterprise controls, or other attributes are deficient. A well-run new product will have done extensive early customer work knowing what is missing and will balance launching with these deficiencies and with the ability to continue to develop the product. In a startup environment, a single company rarely gets a second shot with customers so calibrating readiness is critical. Relative to the broader category of disruption, the harsh reality is that if the disruptor’s idea or approach is the right one but the entry into the market was premature, the learning will apply to the next entry. That’s why the opportunity for disruption is still there. It is why time to market is not always the advantage and being able to apply learning from failures (your own or another entry) can be so valuable.
- Missing ingredient gets added. Often a disruptive product makes a forward-looking bet on some level of enterprise infrastructure or capability as a requirement for the new product to take hold. The incumbent latches on to this missing ingredient and uses it to create an overall state of lack of readiness. If there’s one thing that disruptors know, it is not to bet against Moore’s law. If your product takes more compute, more storage, or more bandwidth, these are most definitely short-term issues. Obviously there’s no room for sloppy work, but by and large time is on your side. So much of the disruption around mobile computing was slowed down by the enterprise issues around managing budgets and allocation of “mobile phones”. Companies did not see it as likely that even better phones would become essential for life outside of work and overwhelm the managed phone process. Similarly, the lack of high-speed mobile networks was seen as a barrier, but all the while the telcos are spending billions to build them out.
- Conventional wisdom will change. One of the most fragile elements of change are the mindsets of those that need to change. This is even more true in enterprise computing. In a world where the average tenure of a CIO is constantly under pressure, where budgets are always viewed with skepticism, and where the immediate needs far exceed resources and time, making the wrong choice can be very costly. Thus the conventional wisdom plays an important part in the timeline for a disruption taking hold. From the PC to the GUI to client/server, to the web, to the cloud, to acceptance of open source each of these went through a period where conventional wisdom was that these were inappropriate for the enterprise. Then one day we all wake up to a world where the approach is required for the enterprise. The new products that are forward-looking and weather the negatives wishing to maintain the status quo get richly rewarded when the conventional wisdom changes.
- Legacy products can’t change. Ultimately the best reason to persevere is because the technology products you’re disrupting simply aren’t going to be suited to the new world (new approach, new scenarios, new technologies). When you re-imagine how something should be, you have an inherent advantage. The very foundation of technology disruption continues to point out that incumbents with the most to lose have the biggest challenges leading through generational changes. Many say the enterprise software world, broadly speaking, is testing these challenges today.
All of these are why disruption has the characteristic of seeming to take a much longer time to take hold than expected, but when it does take hold it happens very rapidly. One day a product is ready for primetime. One day a missing ingredient is ubiquitous. One day conventional wisdom just changes. And legacy products really struggle to change enough (sometimes in business or sometimes in technology) to be “all in” players in the new world.
Of course all this hinges on an idea plus execution of a disruptive idea. All the academic theory and role-playing in the world cannot offer wisdom on knowing if you’re on to something. That’s where the team and entrepreneur’s intuition, perseverance, and adaptability to new data are the most valuable assets.
The opportunity and ability to disrupt the enterprise takes patience and more often than not several attempts, by one or more players learning and adjusting the overall approach. The intrinsic strengths of the incumbent means that new products can usually be defended against for a short time. At the same time the organization and operation of a large and successful company also means that there is near certainty that a subsequent wave of disruption will be stronger, better, and more likely to take hold simply because of the desire for the incumbent to get back to “normal”.
–Steven Sinofsky (@stevesi)
I love the Consumer Electronics Show. Maybe I’m numb from decades of attending it. Maybe I’m just too much of a fan of watching stuff get made. Maybe I just like long lines and the potential for airborne illness. Really what I love is the technology industry and that every year we get together and demo new products, share works in progress, and take chances on offering products people don’t yet (or ever) know they want. CES 2014 was an exceptionally unique year and one that I think will be remembered as the start of a new era, much how the 1970 show changed TV with the introduction of the VCR or the 1981 show changed music with the CD player.
CES 2014 was an exceptionally unique year and one that I think will be remembered as the start of a new era.
But wait, you ask “What product was launched at CES 2014?” The answer is “None”. Instead, this is a year in which every product is about software, and every product assumed that the computer involved would be based on modern mobile platforms, and most everything connected to a cloud service. As an industry we’re not there yet, as we will talk about below, some offerings still cling to previous models of accessing computing and we’re likely to see much changing of the guard as breakthrough products emerge.
The ubiquity of the modern mobile platform, smartphones and tablets, might seem obvious to all of us in computing proper, but it took the better part of a decade for it to go from a section of the show to a big presence to woven into the fabric of every exhibitor. Likewise, software has gone from “content” to “console games” to “pc applications that get thrown in with a device” to the raison d’être or differentiation of consumer electronics.
So put aside the lines, the endless sameness of non-differentiated products, the puzzling keynotes, or even the absence of Apple and Google, and consider the over 3200 companies of all sizes showing off products of all kinds. For me, I think back to when I was a kid and the excitement around what was next came at the yearly Auto Show or reading about the historical World’s Fair Expos. It is hard to avoid concluding that CES is our era’s expression of the future—transportation, healthcare, communication, entertainment, and more are represented by the innovation on display at CES.
It is hard to avoid concluding that CES is our era’s expression of the future—transportation, healthcare, communication, entertainment, and more are represented by the innovation on display at CES.
Me, I’m just excited to get to go to the show and systematically walk up and down every aisle exploring what is there to see. My one set of eyes and one post can’t compete with the likes of the professional tech press that push out hundreds of posts during the week or with the amazingly thorough coverage of “best of” done by many.
Instead, I offer these observations or themes from a product development perspective—what would I be looking at as a product manager or engineer. As I’ve said in this blog many times, learning comes from observing and sharing. Product plans come from many points of view and sources coming together in the context of a company. I cover a lot, but there is more. It was a great show for learning and thinking about the next phase of our industry.
This report looks at themes covering embedded smarts, healthcare devices, communication wearables, screens (4K, curved, skinny), less futzing, and overall trends up/down.
First, a bit of humility
One thing required when looking at new products and technologies is humility. Even though many would like to differ, CES is not a shopping mall where you go to find the new big thing to buy or use right away. This is counter-intuitive because a lot of the products at CES are new and for sale. But in practice, they have not been used and many times not even released to reviewers yet. So you want to step back as you read about the products and not look through the lens of “would I buy and use this today” and instead think about the context overall. As part of that I like to remind myself of a few things about what we see:
- Companies aren’t dumb. A lot of times when a product is first seen something jumps out at you as totally wrong. Keep in mind many of the products are not about the use cases for today, but for use cases yet to be seen. The most classic example is the Walkman— a “tape recorder” that didn’t record. Or more recently, a digital camera that is bigger, heavier, costlier, and worse than a film camera. Sometimes the new use cases aren’t even obvious to the companies yet, and this is even more true today as many “hardware” companies are moving forward rapidly with hardware or the supply chain is making available new components because it can, neither really having software that can implement new use cases.
- Limitations seen in less than one minute are known by the product people. Every product has issues, limitation, constraints. Walking up to a brand new product and thinking you’re the first person to notice such is usually a mistake. While the person at the booth might know the FAQ, it is a good idea to assume the product folks back at HQ actually know the limitations. I can’t count how many times people commented on the battery life of one of the wearables with screens—as though the people developing them would not like to have a month of battery life or were not aware of the trade-off between weight and battery life.
- Iteration is baked into the product you are seeing. Even though the product is for sale, it might not be done yet. It will get smaller, faster, cheaper, power efficient, lighter, and more feature rich. It will do so quickly. Many of those plans are in place. Because so much of the hardware is now subject to Moore’s Law, it is already happening and you can just wait—the price of 4K displays will drop rapidly and because of 1080P volume the price is already spectacular compared to what we’ve come to expect based on previous generations. For software, we all know updates and features are part of the plan. There’s no guarantee things will go in the “right” direction for every product but iteration will happen. Because there are many players, keep in mind that iteration by one player becomes learning for another player so there is ample opportunity for changes in leadership. We all know in technology, first mover advantage is not necessarily an advantage. Multi-party, iteration is the reason.
- Core competency matters. With so many devices doing so many things and so many products incorporating features from other products for differentiation, it is important to focus on the core competency of a product. There’s a good chance a product will try to do too much or for that matter all the products will try to differentiate themselves based on some peripheral features. Don’t lose sight that TVs should have a good picture, fitness bands should measure your fitness well, scales should be fast and easy to read, speakers should sound good and so on.
- Everything has depth and experts. Every year I get surprised by some product that I never thought of and think how amazing that idea is, and then I see 3 more of them on the show floor. It is easy to forget that inside the CE industry there are many industries. Within those industries are people who spend their careers mastering something that, to the uninitiated, might seem narrow. I saw a modern blood sugar monitor (see below) this year that was totally unique. Then I saw two more. These experts are all feeding off many of the same inputs and so one should expect some degree of convergent innovation. Said another way, in the context of a broad show like CES, something that I think is really cool might not actually be all that innovative to those in the field with some domain knowledge.
Let’s look at some themes and within them put on our product manager hats and see how what we observed might influence our own choices in products design. I’m going to take the observations from the show floor and project forward a bit as that’s what product management needs to do with the data when there are technology bets to be made, products to design, and specs to write.
Intel kicked off the show with a keynote declaring that all devices need to be smart. Walking around the show floor showed that this advice has already been taken to heart. While smart TVs are the most obviously visible (and also a holdover from the past two or so years), we also saw smart cars, smart healthcare devices, smart fitness monitors, smart watches, smart home appliances, smart projectors, and more. Smart was everywhere. Should it be?
Smart can mean anything from a touch-based user interface replacing the existing mechanical UI to taking a formerly mechanical device and embedding an entire OS with app ecosystem into the device.
Moore’s Law is an important contributor to this trend. Previous views of smart devices would have meant connecting the hardware device to a PC, with all of the costs, size, power that this entails. Home automation that used to take a PC now just connects devices with Wi-Fi to a cloud service, for example. A home blood pressure monitor would have stored some number of readings until you connected it with a serial cable to a PC and now it just sends those over Wi-Fi to a cloud service. TVs would have been connected to a PC that presented a full PC experience through an alternate user interface that today can offer this same type of functionality through an entirely embedded solution. Now it is both feasible and economic to include an ARM-based computing platform and either a Linux or Android OS driving the “smarts”.
But is this always right? The product manager view might be that it is time to look at use cases and scenarios and step back. While the hardware side is possible, the software might not be delivering the right experience. The truth is, some devices should be dumb. And that’s ok. The internet of things does not need to recreate the challenges of the internet of PCs. A single general purpose approach used everywhere might not be the best approach compared to tailored devices working with a very rich mobile device and cloud services.
The truth is, some devices should be dumb. And that’s ok.
One reason for this is that there can only be so many app ecosystems. It simply won’t be possible for apps to be delivered reliably and in a feature complete manner across all of the various smart devices. While today it might be possible for a streaming music service to be omnipresent on every possible smart device from a watch to a car to a TV to a refrigerator to a treadmill (and a phone and a tablet), down the road the user experience for that streaming app will have become rich enough that the primary use case will drive the expected experience which won’t be duplicated across devices, whether that is because the devices vary in capabilities, screen sizes, or just human interaction or just because there are too many different platforms.
Two examples help to reinforce this product challenge.
- Screens / TV. We all want lots of stuff on our big screens. We want streaming video, live broadcast television, music (maybe), and perhaps some web services like messaging. But these are all sophisticated experiences (finding the video, dealing with TV signals/guides/DVR, managing playlists, different apps), and so it means they likely demand (or will demand) a rich interaction model connected to services. Good news! We already have this interaction model on our modern mobile tablets and phones. Why try to duplicate this with the added complexity and variety of TVs? Rather a device like Chromecast or Apple TV shows how you can use the TV simply as a “dumb screen” which becomes far more manageable, the UI is far better, and is a much better overall experience. These solutions, where the screen is dumb and the mobile device serves as the gateway to the dumb screen seem to put the code in the right place and reduce complexity and increase simplicity for the use case. It is worth asking if this Twitter client on a TV will ever match what you can do on your mobile device in your hand while watching the show? That’s not to say there won’t be other use cases integrating apps into TV, but just showing subsets of the mobile apps side by side doesn’t seem right.
- Autos. From Audi to Volvo we saw smarts added to cars. This was added in the form of a screen, a telemetry platform, and apps. What is different about this compared to TVs is that we don’t want cars to be dumb. We want cars to be smart about being cars (safety, maintenance, better driving and accident avoidance). Like TVs, however, it isn’t clear that we want to put the equivalent of a unique mobile platform in every car brand. Is there any chance the mapping app in a car will be on par with the mapping app in my mobile device? Wouldn’t I rather have the same ability to send my mobile screen to the car screen that I get with Chromecast or Apple TV? Perhaps having a protocol that supports touch in that scenario is very helpful too. In the meantime the smarts of the car can focus on the things the car needs to do, and perhaps even recognize the best way to have a user experience and manage those would be with an app and cloud service? Ultimately, the way cars are made means that the technology choices are out of date by the time the car makes it to market and if you own the car for 5 years then those technology choices are really dated and perhaps the overall resale value of the car declines.
Will this UX really be right today or in 8 years?
The fact that all the screens and cars are making bets on technologies that are just capable of being used helps us all—this is not the time to be cynical but the time to learn. These products are not done yet and we can’t highlight the greatness of the Lean Startup and MVP and then be critical of bringing to market products that might not quite be done—that’s where reviews, experts, and frankly store return policies can help. As a product manager you want to ask yourself about the trajectory and likelihood of success of an approach down the road when the work all comes together. These new products show exciting scenarios but maybe there are better ways to implement them.
The advances in sensors have been breathtaking thanks to technologies like MEMS and others. Combining those with the ability to embed and whole OS and connectivity to cloud services in what used to be basic diagnostic equipment is a revolution in healthcare. Here too we saw many new and breakthrough products. As a telemetry nut, obsessive compulsive, and geek these are some of the most exciting products ever. One thing that made this CES seem so new and fresh is that this feels like a renaissance in consumer electronics. Devices you buy at reasonable price points, solve specific problems like an appliance, and just work for a scenario. In most ways, these new devices are starting to deliver now.
Basic body telemetry like weight, blood pressure, composition and more can now be easily measured, tracked over time, and even shared easily with care givers or compared with a circle of friends. Stepping on a scale every morning is quietly making a bar chart, setting alerts, and trending your data. And even better, such devices are learning from past designs and becoming easier to setup and use. No longer do you need a PC, EXE, and USB cable. Instead the device is paired over Bluetooth with a dedicated app and you’re up and running with a great UI in no time. Basic scenarios like maintaining compliance with medication are made easier by smart pill boxes that alert you wirelessly on your mobile device to take medicine. Overkill? Perhaps, but compliance rates are still not where they need to be. And combine this with easy measurement of blood pressure and you can see how putting smart in the right place, cloud services and mobile apps to make things accessible can be such a huge advance.
Three healthcare products that demonstrate this include:
- Head injury. Much has been written about the rise in head injury in sports and long term risk associated with cumulative concussion, particularly football. Reebok with the Checklight is one of many companies with a product designed to measure cumulative head impact using accelerometers. The packaging is very user friendly as you can see (and it won a best of CES award). The basics of the device are cool—a red light goes off when a certain level of cumulative concussion risk has been reached. Other variants of similar devices have different form factors (helmet integrated, mouth guard) and can even report real time to a mobile app. The telemetry, use case, and execution are all coming together at CES 2014.
- UV exposure. The JUNE UV detection bracelet by netamo simply measures cumulative sun exposure and integrates with a mobile app, again with a simple UI on the device and a data connection to a mobile app. There’s a lot to like about this for folks who work or play outdoors and want to mitigate the risk of skin disease or damage. The app/service provides advice and a suggested “routine” for your skin based on data.
- Blood glucose. Those that have been touched by diabetes (perhaps one of the more insidious diseases in the developing world, costing the US an estimated $245B a year in healthcare and related costs) know the complexity and challenges of testing and managing glucose levels. The YoFi Meter, http://www.yofimeter.com/, is a very smart device (see the above discussion). It combines a glucose test strip dispenser/reader, a lancet dispenser, along with a simplified tracking interface on a touch screen and an integrated 3G connection to a cloud data service. This is a device that takes compliance to a new level. At first I might have thought this device is too smart, and then after talking to the designer I learned of many use cases where a companion mobile phone isn’t available or possible (for example, students must be tested by the nurse who doesn’t have time to call up parents with real time data and a phone might not be available in school).
Each of these three devices shows how telemetry and mobile apps/cloud services can dramatically change the basics of healthcare for a scenario.
There are challenges we will all need to deal with however. These challenges are not new to those who already work with data. Data does not always lead to actionable next steps and sometimes more data leads to more ambiguity. These three devices show how the reality is that science is not yet caught up to being able to present us with all this data.
In the case of concussion and head injury, right now the data is unclear on how much cumulative impact over what period of time is “safe”. So while it can be measured, exactly when and how to act is not clear, particularly for children. It is easy to see how the debate will quickly move to one of acceptable levels. So more studies over a longer period of time will be needed which for this type of measurement will take decades given that the measurement is just now available. Science is hard. Glucose measurement is the other end of the spectrum. For diabetics the data and management is well understood, but compliance is challenging or at least not super convenient. The advances are amazing and ready now! Sun exposure is one that becomes interesting only because the data basically says to minimize exposure as much as possible—in other words there’s not really an acceptable level of UV light (i.e. SPF 40, see http://www.aad.org/media-resources/stats-and-facts/prevention-and-care/sunscreens).
Together these show the opportunities and challenges in the healthcare telemetry space. In any product design, the ability to measure something and present the measurement to a customer is not the same as being able to provide reliable and actionable information. It is critical in the design of a product to be clear on what to do when the product tells you something, lest the dreaded “Check Engine’ syndrome.
One might even offer the world’s first connected toothbrush:
While many healthcare devices are wearable, the broader category itself exploded this year as has been well documented…everywhere. When it comes to sophisticated wearable communication devices, this is a year of learning products. Most are not ready for primetime or broad usage, though many will find niches with early adopters or enthusiasts.
This shouldn’t be news to anyone. Consider as an example post-VHS digital formats for movies forays into optical media (LaserDisc anyone), the path from first products to broadly used products is often one with many twists and turns in basic technology and scenarios. In addition that path from the first component sized DVD player to the 6″ round portable DVD player or integrated flat screen DVD player took some time. Innovation does not happen all at once, even though we often remember it as punctuated moments in time.
There’s no need to document the dozens of communication wearables on display. Most shared the same basic characteristics, with Pebble being the established player that has already earned an enthusiastic base of early adopters. These pair with a mobile device, share notifications, and permit some level of interactivity and apps/ecosystem. Some do less and trade off towards a longer battery life by doing less. Others try to subsume the mobile device entirely and act as a phone (see below).
The primary “cause” of this is that the ability to miniaturize the hardware platform and squeeze a full software platform on the device has surpassed the ability to build a software experience and use case. These devices, by and large, are currently in the “because we can” phase of innovation. That’s not bad and in fact when software meets hardware it is often a necessary ordering.
The primary challenge, at least from my perspective, is that no one has arrived at a new use case. We’re simply looking to move some use cases of the mobile device to a wrist based device. But the device on the wrist is “less of everything”. Taking a disruption point of view, this isn’t disruptive yet. As often discussed, the first PC-derived tablets were more PCs without keyboards than they were a new set of use cases (pen based drawing/notetaking notwithstanding). It wasn’t until the iPad presented a new set of mobile scenarios and capabilities and the hardware was better able to meet the scenarios that a device without a keyboard was able to define a new use case.
Absent a use case, the dialog around wearables will just bounce around the constraints of screen size and battery life. You can only do so much with a tiny screen and a wrist sized device can only operate a screen for so long. While one likes to be notified with a UX based on a glance, it turns out this is pretty much what mobile phone designers work to do all the time and with a lot more screen real estate and elaborate UX. On those platforms the debate is an endless one around how much can you do to a notification and what are the verbs that act on it. Is a new text just read, read and reply with a canned set of replies (and can/how might those be customized), or full messaging capabilities? Do those choices extend to custom messaging apps like WhatsApp or Skype? Who will write those apps? When those apps have new features do they carry over to the wrist?
Here is one example of a fitness watch that is also full smartphone on a wrist, including a pull out Bluetooth headset.
This is a complex set of questions. From a product manager perspective, they all boil down to defining use cases and scenarios for why a device should exist. Is it a companion? Is it a replacement? What does it do uniquely such that I’d be willing to forgo other functions I already have? Disruption theory says that it is totally ok for a new product to do less, so long as it is so good at something that people want more as that’s the whole point of being disruptive.
It is still early. It is too early to judge these devices as what is possible and for most of us too early to be adopting these devices. They are the stuff of Star Trek, which by that theory of innovation only means it is a matter of time.
Screens: 4K, curved, and skinny
If you’ve seen just one article on CES then it is certain you know that new TVs on display were both 4K and curved. Cutting to the chase, if you buy a TV in 2015 odds are it will be 4K. The rapid march to 4K is more massive than anything we’ve seen in screens. Moore’s Law is our friend here and at some point the entire supply chain will just convert to the mechanics of making 4K and it becomes essentially non-economic to maintain the old processes and supply chain. We’ve seen this with memory, storage, processors, and screens. Silicon based innovation lends itself to rapid and whole movement of products. That’s good for all of us.
Cutting to the chase, if you buy a TV in 2015 odds are it will be 4K.
For screens, 4K has two unique elements:
- Extremely rapid cost reduction. Competition is fierce and the economics of the processes will likely drive 4K screens to “acceptable” consumer prices much more quickly than the move to flat screen or 1080P. During the show Dell announced a 4K monitor for $699. My early adopter 15″ VGA LCD costs $1999 (and weighed more than the Dell will). Yay for consumers!
- Content will appear. While we can all bemoan the hype and failure cycle of 3D at home, which included a lack of content, there were already significant content deals for 4K, notably Netflix. While you need 15MB bandwidth for 4K the content will be there. Rest assured, the rest of the content industry heard this and so I suspect we will see brilliant 4K content of some form very rapidly. Again, yay for consumers!!
If you have any doubts, once you see a 4K screen you will want one. Put aside all the arguments about physics, optics, and more, it just feels right. In practice, what you really want are high gamut and 60fps, so let’s hope these attributes and benefits become clear to consumers. Again, this shows the value of reviews, community, and expertise in the adoption of CE.
Curved screens were somewhat of a surprise to many attendees I believe. In booth after booth people had somewhat puzzled looks at them and there seemed to be a broad effort to quiz the booth staff with “so what good is it”. Most of the time we all got the general answers about immersive experience or less reflection. Each of these to some degree are true (especially reflectivity in many situations).
Again, as product managers we see a hardware technology appear because it can but the use case hasn’t yet been determined. Like the first color computer screens that many argued against claiming software was inherently black and white, curved screens are about new use cases not just curving a football game. One view around CES and you can easily see scenarios such as signage that become incredibly cool. Today signs that are interactive are much cooler than static signs (or menus and more). Signs that need to be on curved surfaces are static and boring. Maybe curved displays will be a niche at home and find themselves useful only for commercial signs. I’m going to bet on the creativity of content and product people to develop new use cases and before we know it curved screens will be ubiquitous as “flat screens”.
Finally, this year saw a great many more wide-aspect ratio screens at 21:9. For the most part, the mass market of screens are made in a small set of aspect ratios depending on mass adoption. Like film was historically, there are both benefits to this along with those that want to experiment with alternate aspect ratios. The iOS tablet world is 4:3 and the Windows/Android/HD world is 16:9. The ultra-wide 21:9 seems rather appealing for a number of use cases, including side by side and multiple inputs. If you combine the ultra-wide screen, more pixels, and curved display you recreate a developer workstation or Bloomberg terminal but with a single screen which can mean less space, easier ergonomics, and perhaps less power. Again, it seems like the use cases are going to quickly follow the technical ability to make the screen.
The product manager view of screens is to consider what you app or content can do when being projected on or making content for these new capabilities. As we saw with Retina pixel density, these changes can happen quickly and getting left behind is not always a good spot to be.
The evolution of most CE devices is often to more features and customization, and over time this can be viewed negatively. Certainly at some point this complexity makes products unappealing for many. The industry has a great many enthusiasts who love to customize and tweak. Analogous to the auto shows, some people used to love to look under the hood and adjust the engine.
Back in the heyday of Auto Shows (they are still huge, but CES and tech have eclipsed those shows in media coverage in my biased view), the talk was about components of cars. This dialog was broad and understood. Average consumers knew about horsepower, disc brakes, electronic fuel injection. Today cars are about design, convenience, and use-centric concepts like capacity and MPG. CES, this year in particular, has transitioned to talking and showing more about use cases and less about how products are built.
In almost all devices you have to look hard to find gigahertz or megabytes. You see tasks or uses much more up front. This isn’t always the case and often the first questions are about specs. Still, I would say a lot of “progress”.
Some examples of this jumped out at me, particularly in the gaming world. The gaming world has traditionally been split between consoles representing the true CE experience and the gaming PC which defined the ultimate enthusiast experience when it came to moding your gaming PC. For gamers or those that want to just play games this is a banner time with an explosion in gaming options. Many believe the usage in phones and tablets will dominate with casual games available in app stores. The new consoles from Sony and Microsoft promise to bring gaming to new technical levels with their advanced PC componentry in a true CE package. Finally, at CES we saw the SteamOS powered devices (PCs) and an example of a more state of the art or “modern” PC.
Steam Machine. SteamOS promises to bring the simplicity of consoles with the power of PC gaming. Some critics are saying it brings neither and is in-between. But the popularity of the Steam platform is significant with millions of intensely active members. The product manager question is whether Steam disrupts PC (or console) gaming or simply extends the life of a gaming platform that while popular is being squeezed between consoles and mobile devices. Is the SteamOS powered device re-imagination of PC gaming or an appealing convergence of the PC with consoles (see http://blog.learningbyshipping.com/2014/01/07/the-four-stages-of-disruption/)?
A Steam Machine is an Intel-based device meeting a set of baseline specs, combined with the Steam Controller and SteamOS. The debate among gamers is about the specs and capabilities of the underlying hardware along with the lack of ability to mod the devices. The Steam Machines themselves represent a broad range of “sealed case” form factors, most of which existed as Windows game PCs in various forms.
Razer Project Christine. Razer introduced Christine, which is a highly stylized modular PC. While the idea behind a modular from factor has been tried several times before, the combination of hardware interop, industrial design, and openness to accessorization (my own word) are at a unique point in time. Razer has a very active customer based that thrives the combination of gaming and accessories for gaming. It might just be that this hassle-free notion of moding a gaming PC will appeal to a broad set of PC gaming customers. In many ways this is an attempt to disrupt the PC gamer, while maintaining a commitment to customization. Project Christine beat out Steam Powered for CES Best of Show in the category.
Reduced futzing is really enabled across a broad range of CE devices because of mobile apps, WiFi/Bluetooth connectivity, and cloud services. What used to be elaborate setup and configuration is now enabled via simple apps that connect to devices over wireless protocols. The rich UX afforded by devices replaces single line LED displays or embedded web server experiences. The ability to save/restore data and settings in the cloud replaces sharing via dedicated (and awkward) subset experiences for social networks. From WiFi access points to scales to cameras, we will spend (and tolerate) less time futzing and more time using CE devices.
From a product manager perspective what excites me about these two innovations and the broader theme is the move “up the stack”. Our computing industry has broadly moved to modern platforms for both hardware and software and seeing gaming move in this direction is critical to the health of this style of rich interactive gaming. It is also a natural maturing of a technology area and to resist the change essentially guarantees disruption. The market for people willing to devote time to futzing is shrinking, no matter how much we (having built more PCs than I can count) enjoyed it. There are simply too many options for how to spend more time gaming and less time futzing. Just like people want to use cars to get around, not stare under the hood and fix them before going somewhere, the move up the stack is relentless for most every consumer.
Trending up and down
To wrap up a quick look at the year over year view of CES and what is on the move taking up more floor space and mind share and what is taking less beyond the items mentioned above.
Android. There was a lot more Android this year than last year. Android of course is particularly popular among wearables and TVs where there is no third party option to use iOS. The inexpensive Android tablets we all heard about from the holidays were on display where you could see the dozens of OEMs packaging every conceivable screen size and spec into tablets. One note is that there was a complete absence of 4:3 Android devices, which I find interesting given the competitive nature of things. What this feels like is a reaction to avoid being like Apple, when in practice it might be that for some device sizes the squarer aspect ratio might be more convenient.
Chromebooks. Building off what we might have read as momentum in the US over the holidays was a broader presence of Chromebooks. This year saw an all-in-one along with several lighter and thinner (and still inexpensive) clamshell formats.
Phablets. It was interesting to see the number of show attendees using their really big phones (or small tablets). I would do a quick badge check and noted that more often than not the phablet user was not an employee of Samsung or LG, but just a user. I think this is a trend worth watching. If you have only one device the tradeoff towards a bigger screen becomes interesting.
WWAN connectivity. Looking back personally, I was totally wrong on WWAN connectivity. I did not see a market where the carriers would make it so easy and relatively inexpensive to add another device to a data plan. From the glucose meter to auto fleet tracking to wearables, SIM cards were everywhere. This in spite of the fact that the hardware costs are real and the data plan is real. Many products, particularly those using WWAN for sending telemetry to a cloud service, will include a SIM and fixed data plan as part of the device price or part of the service plan, which is super cool.
3D Printers. The 3D printers are all remarkable. It is hard to overstate how much this will disrupt so many fields. There’s a lot of talk about a drone delivering a product, but what about just downloading it and printing it at home. The use cases for 3D printers are not at the broad consumer level but soon will be.
Touch screens. All the major screen makers also showed screens that were touch capable. The most common touch detector was an IR field. All of these were hooked up to PCs. Every screen is a touchscreen or it feels broken. Also pictured below is Panasonic’s 4K portable running Windows.
PCs. While Microsoft, HP, Dell, and others having stopped exhibiting on the main show floor can easily be responsible for the lack of PCs, the reality is that PCs are part of the fabric of the technology world, but not the front and center consumer electronics device. The reality of the floor is that PCs were not part of the use cases or scenarios and the modern mobile platforms have taken over. This image below sort of summed it up. Here we had RCA showing their new PCs (fairly thick and heavy laptops) in a booth with toasters and dorm refrigerators. Yet we all know, today, the products at the show were built using PCs, the businesses are run with PCs, and even the show itself would not be possible without PCs. The focus is just changing.
Discs. There are no optical discs to be found, for storage or content. Content comes from the cloud. Content is stored in the cloud. This was the last show with discs, and they were practically not there. One might say the same for spinning media and even the satellite tuners were offering solid state storage options to reduce noise, size, and thermals.
Projectors. Another reminder of Moore’s Law as applied to glass is that absence of projectors. While there are quite a few very tiny portable projectors, by and large projectors have been replaced by simpler and brighter on wall displays (soon to be curved and 4K).
Wires. There’s almost no wire at CES anymore. Even the Monster cable booth was mostly about cleaning screens and wireless headsets more than wire. As a person who has no wired communications at home, I can relate.
Home Theater. After a decade or more of “home theater”, the drive to simplicity, the role of streaming content, unified HDMI, and the preference for mobile, has all but eliminated the idea of a complex, multi-component, theater. Today’s theater experience is so much different than emulating a movie theater in a dedicated room at home using a projector and stack of 1000 watt components. People want to watch TV but also interact at the same time and that leads to a very different environment. With the ability to watch anything instantly on a tablet, a home theater has some strong competition for attention and use cases. The lack of a need for a home media library of discs or drives also alters the need for a dedicated room. Live events and film fans still will have their dedicated experiences, but the equipment is drastically simpler.
Remote controls. The world of complex infrared controls at home is being disrupted by mobile devices. Even infrared, while simple and low cost, feels like it will be disrupted by Bluetooth or even Wi-Fi.
For those in technology, CES is really the greatest show on earth. We’re all privileged to be part of the technology industry in such a pivotal time. Next year is going to be even better. I’m counting on it.
—Steven Sinofsky (@stevesi)
Tiny laptop level power supply that works by the magic of physics from Finsix in Menlo Park, CA:
A few hundred thousand dollars worth of add-ons to convert a Canon C500 into a studio broadcast camera:
Sony’s full family of Xperia Android devices is also water-safe. These are all worth a second look I believe.
Example use case for 4K screens with 4 up 1080P security cameras (also on the floor, 4K recording cameras):
Case for iPhone that embeds the ability to take infrared images. Another example of “possible, but still developing the core use case”. This was developed by FLIR, an existing maker of IR cameras:
Bluetooth headphones plus earmuffs in one.
Example use case for curved signage. Wouldn’t this be a more interesting booth with a moving sign: