Published/updated: June 2011
Let me put my cards on the table. I do not, repeat not, subscribe to the view that users always know whatís best for them in a work environment. Indeed over the years I have come to the conclusion that generally they donít.
Why? Well for two reasons. Firstly, users are generally driven by want more than need, and want is in turn driven primarily by factors that are very subjective and not particularly business-oriented - status, image, personal interest, religion (with a small írí), etc.
Secondly, users generally donít know what they donít know. They are unaware, for example, of how important requirements such as cost management, security, compliance, return on investment and future-proofing are to making sensible technology-related decisions. They are even less aware of the factors that enable these such as consistent policy, efficient and effective processes, coherent infrastructure and properly managed support.
So, I feel compelled to challenge people when they use the terms "consumerisation" and "productivity" as if they are synonymous. Letís define these two terms so you can see what I mean.
When people talk about the "consumerisation of IT", they are generally referring to the phenomenon of users making their own decisions on technology and services. In practice, this is centred on the elements of IT that are most accessible to users, i.e. devices (notebooks, handhelds, slates, etc) and publically available services (e.g. social media, communication and collaboration services). What these have in common is that users are able to start making use of these without involvement of IT professionals, though, of course, the IT department is often leaned on to provide connectivity to the corporate network and bail users out when things go wrong.
Productivity is about the business getting the best return from resources consumed (money, man hours, management bandwidth, operational bandwidth, etc) whilst ensuring the business is properly protected from exposure to unnecessary risk. I know some talk about it in the context of individuals being able to make the best use of their time, but this only matters if the net contribution from any productivity improvement action is positive at an overall business level. At the risk of repeating myself, but in the interests of getting absolute clarity on this, productivity can only be sensibly assessed, monitored and managed at an aggregate level - focusing on individual employees can be very misleading.
Now if we bring these two concepts together, it becomes very clear very quickly that "consumerisation" and "productivity" do not always go hand in hand. As a really simple example, if a free-for-all on mobile devices leads to some notional productivity improvement from employees being able to get their e-mail on their personal iPhone or Android device, this is only useful if there isnít a negative impact elsewhere. If the result is that IT needs to spend additional time on management, support and remedial work that detracts from their ability to service critical processes or business development activity, then the result could easily be an overall net loss in business productivity (not to mention the probable increase in various risks).
As I am in that kind of mood, there is one more notion I would like to challenge. This is the principle that consumerisation somehow means that we should give users a totally free hand to choose their own tech, even when itís funded by the business. I often hear people say that we shouldnít force equipment and applications on people. Itís better to give employees an allowance for kit such as notebooks, slates and mobiles, and let them install whatever software they think is best for them to do their jobs efficiently. This is re-inforced by the premise that Generation Y types coming into the workplace will "demand" it anyway, so you might as well "get with the programme".
Quite apart from the obvious dodgy reasoning that for the first time in human history, it apparently makes sense for inexperienced teenagers and young adults with a lot of growing up to do to dictate business policy, we must again come back to the principle that users really arenít that informed and objective when it comes to tech decision-making. Some might think that the latest premium bling available in the High Street, or the budget option if thatís all they can afford, might make them more productive, but the chances are that something provided by the business that has been selected in an informed and sensible manner will actually do the job much better.
But the genie is now out of the bottle. The trouble is that too many tech-related decisions in the past made by IT departments have been very one-sided. The emphasis has often been on meeting the corporate requirement and making the life of IT professionals as easy as possible, with not enough consideration given to user preference. This history and the associated user frustration with IT, together with the ease of access to technology we all now have as consumers, means the phenomenon of consumerisation is probably here to stay in one form or another. If you try to stop it, youíll simply drive activity underground. The imperative is therefore to manage it, bearing in mind all of the factors we have discussed.
In my next article on the matter, Iíll pick this up and suggest some ways of thinking through it to drive genuine productivity while avoiding the pitfalls of natural chaos.
This is part one of a two part series. Part 2 here.
CLICK HERE TO VIEW ORIGINAL PUBLISHED ON
By Dale Vile & Jack Vile
By Dale Vile & Jack Vile
Making the right technology investments in todayís fast moving digital age can mean the difference between success and failure in pretty much any industry. But how well are decisions are actually being made in this space? ...more
By Dale Vile
Some argue that IT operations doesnít matter anymore; itís all about developers. Our aim in this paper is to re-balance the discussion based on research in which feedback was gathered from over 400 European IT professionals. ...more
By Dale Vile & Jack Vile
We often hear that cloud computing dramatically reduces the need for in-house IT teams, and might even lead to their ultimate demise. The research reported here provides a very different view based on analysis of real business objectives and actual experience. ...more
By Dale Vile & Jack Vile
As technology becomes smarter, more opportunities arise to exploit AI, machine learning and other forms of intelligent systems to drive efficiency and transformation. But whatís the impact on IT teams? ...more
By Tony Lock
Have we all been caught asleep at the capacity planning wheel? Business users today want, and expect new IT services to be delivered in the blink of an eye, the necessary resources provisioned instantly, and changes made ďon demandĒ. ...more
By Dale Vile and Tony Lock
Itís easy to be caught out by a cyber attack or internal mistake that leads to your customersí data or important intellectual property ending up on the black market. Making sure your business is adequately protected and is able to respond effectively ...more
By Dale Vile, Tony Lock & Jack Vile
Application programming interfaces (APIs) have been around for decades. In the early days of IT they were primarily used to give programmers convenient access to libraries of prebuilt functions. As systems became more distributed, APIs found their place ...more