High-Performance Product Management

The article High-Performance Product Management: The Impact of Structure, Process, Competencies, and Role Definition in the Journal of Product Innovation Management is interesting for many reasons.  The least of them is because it provides a very good history of research in product management.  More importantly, it combines qualitative interviews with factor analysis and maximum likelihood estimation to develop and test a model for improving performance in product management.

The paper identifies several key factors that potentially impact product management performance. A set of qualitative interviews is conducted to develop hypotheses related to constructs that may drive product management performance. These hypotheses are used to develop a causal model for product management performance that includes constructs related to roles and responsibilities, organization structure, and marketing processes related to product management. An empirical survey of 198 product managers from a variety of industries is conducted to test the causal model. The results of the causal model suggest that performance of a product management organization is driven by structural barriers in the organization, the quality of marketing processes, roles and responsibilities, and knowledge and competencies. The findings suggest that structural boundaries and interfaces are the biggest impediment to effective product management, followed by clarity of roles and responsibilities. The research highlights the importance of organization structure and effective human resource practices in improving product management performance.

Below is what I learned from it:


The overall model proposed for Product Management Excellence is:

Overall recommendation from the model and associated analysis are:

  1. Remove Organizational Barriers and Connect Silos: For example, between existing product management vs. new product management residing in different organizations)
  2. Do not expect product managers to learn on the job: Develop and provide formal training to augment knowledge / competencies / soft skills.
  3. Define Authority and Responsibilities Clearly: In many organizations, product managers have no clear authority and vaguely defined roles. This makes is difficult to deliver results.
  4. Allow Product Managers to focus on Strategy & Planning: 42% of product managers surveyed believed that they spent most of their time on tactical activities and coordination. They had no time to make strategic decisions.
  5. Institute Quality Product Management Processes: From Market Requirement Documents to SKU Planning, lack of effective processes increases non-value-added work and reduces effectiveness

How to get employee engagement in R&D strategy

Time and again, I have found that most employees do not understand or even know about the company strategy.  Corporate Executve Board has some good data in Get Your Frontline Onboard: Communicate, Clarify and Cascade CEB Views – Finance and Strategy:

A surprising number of employees don’t know what their company’s strategy is. A study by the International Association of Business Communicators found that only one in three companies say their employees understand and live the strategy. Robert Kaplan and David Norton, the founders of the Balanced Scorecard, found the situation to be worse. They found only 5% of employees understand company strategy. Without understanding, execution is impossible. Therefore, communication is critical, not only to promote understanding but to help employees appreciate how the strategy relates to what they do.

Of the three Cs, communicate and clarify have been discussed pretty thoroughly.  I want to reemphasize Cascade.  That is the crucial, and pretty much the most difficult portion of getting employee engagement.  Most front-line employees do not or can not figure out what they can do to help with company strategy.  Management needs to make the effort to actually help employees understand what behavior is expected of them.  Sending employees the strategy and asking them to follow it is not it.

A funny anecdote: The presidents of a company actually tried to enforce strategy buy-in.  The strategy was very generic (Reduce Costs and Increase Revenues).  At first, they communicated and celebrated the strategy and assumed everyone would follow it.  That did not happen.  Then the president set up mandatory meetings between mid-level managers and senior executives to get buy-in.  That did not work either because the mid-level managers did not know what they were supposed to do (nor did the senior executives).  These mandatory meetings became question answer sessions that generated no results.  The president than decreed that the mandatory meetings will only be about how mid-level managers would implement the strategy.  No questions will be answered.  That failed as well.  The strategy and the entire effort was then dropped.

I will leave you with the recommendations on CEB on cascade:

To help employees take ownership over their role in the execution, communications about strategy should always be accompanied by goals and metrics. These should be goals and objectives that employees can relate to and can be integrated with their daily tasks. Also, be sure to give them visibility into the goals that everyone up the line is trying to achieve as well so they understand how what they are doing contributes to the larger objectives. Ultimately, front line employees need to know:
What I need to do – goals and tasks
Why I need to do this – the value it provides the customer, the employee, the department and the organization
Don’t create too many goals. Prioritize to make it more manageable. If employees are overwhelmed by the scope of the strategy, or the number of goals they need to achieve, they are less like to perform well.

Social Networking Helps R&D Collaboration?

The article Frontiers of Collaboration: The Evolution of Social Networking in Knowledge@Wharton discusses how social networking (Wiki, Blogs, tweets, etc.) is replacing Knowledge Management and improving communication at the same time

Weinberger began the session by asking panelists what made the introduction of social networking tools different from previous technological endeavors to improve communication and collaboration. One significant issue discussed was how social networking compared with knowledge management (KM). KM systems first appeared on the scene about 20 years ago and once represented the frontier, embodying companies’ most innovative ideas for integrating internal access to disparate information in order to improve communication, collaboration and business processes.

before social networking tools enabled quick and casual communication, many bloggers in corporate organizations had “some KM tool where you captured the knowledge in the tool’s silo and assigned all sorts of tags, folders and so on to it. You would then pass the blog to your manager for him or her to [learn from] what you were writing.

Social networking is easing some of the frustration users in many organizations have encountered with traditional KM systems. Through use of Twitter and other tools, more of the intellectual capital that KM systems once guarded is flowing freely, in real time, inside and outside organizations. If an employee needs to find expertise or share information, he or she doesn’t have to work within the rigid confines of a KM system, or even the confines of his or her organization. Instead, the employee can use social media to collaborate with others and to find answers more quickly and put relevant advice into practice.

Clearly, social networking can add value to R&D community.  However, I am not sure if I would agree with this as whole heartedly as the authors / speakers about replacing KM.

One problem with social networking is the volume of information that can become available and time/effort needed to find the right information.  Consider if all employees in a 10,000 person R&D house started tweeting what they were doing – the signal to noise ratio would be terrible.  Add to this personal tweets and the entire system would become unmanageable.  Furthermore, how does one control the flow of information and ensure that proprietary information is not accidentally leaked?  The panelist thought that all that the risks outweigh benefits –  I am not sure I agree:

Fitton, whose consulting firm focuses on helping companies to use micro-blogging in a business environment, suggested that companies may find the “messy and random serendipity” of Twitter and other social networks to be more efficient than lumbering KM systems and processes. “It brings an infusion of humanity to business,”

There might be value to social networking in building R&D communities and helping virtual teams collaborate effectively, but the idea that real R&D knowledge can be shared effectively through micro-blogging sounds a bit simplistic.  R&D teams do not just need access to knowledge, they need access to the right type of knowledge at the right time.  If one is designing a new cell phone and has a question about what impact human body will have on reception, it does not help to go search through blogs, nor does one have time to do mass mailing / tweets to request help, weed out the responses and find the right person.

Underlying assumption for social networking in R&D world is that someone has the right information available at their finger-tips and are willing and able to stop what they are doing to provide that information back.  How likely is that going to be?  Not to mention the constraints that social networking tools like tweeter add to communication… Unlike the speaker, I am not sure that constraint is actually valuable:

But does the 140-character limit for posts to Twitter enable engagement, or is it “a sign of triviality?” asked Weinberger. “Constraints breed invention,” replied Shellen. Douglas added that communities using Twitter, Google Wave and other tools are creating their own etiquette. Panelists agreed that both the creation of etiquette for particular conversations and the sheer ability to engage in several discussions at once would be difficult using blogs and older forms of web content sharing programs.

There are more problems with sharing:

Lippe noted that, in the legal field, “there’s already a structure of knowledge, and most knowledge repositories and structures of the collaborative web have existed for multiple generations. So, the question is, how do you tap into them?” One core structure is attorney-client privilege, which Lippe said “has long preceded the information confidentiality and security regime that we all have now. It creates the structure of what you can and cannot share.” In the legal universe, he added, the messy serendipity of “horizontal” social networking cannot solve the hardest problems.

Just to be clear, I do not think that social networking does not have value.  If used properly, it can help companies build focused virtual R&D communities across geographical and cultural boundaries.  However, R&D managers will need to do a lot of work structuring and managing the flow of information so it be value-added.  Furthermore, objectives of social networking should be crafted very carefully and monitored consistently to ensure that it is indeed delivering results.

Three Lessons for Sustainable Scenario Planning

I have always been a fan of scenario planning.  It really does provide great insights into R&D strategy and allows organization to develop robust R&D plans.  The article Six Lessons for Sustainable Scenario Planning talks about how interest in scenario planning is increasing because of the turbulent economy:

One business discipline that generated a huge amount of interest during the recession was scenario planning. We wrote about it for Bloomberg Businessweek and advised many companies on it. The Corporate Strategy Board (CSB) ran a series of meetings around the globe on scenario planning where clients exchanged ideas and talked about how to implement the most successful practices we saw in our client networks. These discussions led to the six lessons below.

I will make it a bit simpler than the six lessons in the article – three points to keep in mind while setting up or maintaining a scenario-based planning process.  I had trouble with Scenario Planning at one of the organizations.  These three are my lessons learned from the experience.:

  1. Formalize Scenario Planning: Have clear ownership / accountability for scenario planning. Ensure that the responsibility / authority are clearly defined and delineated from other ongoing efforts.  Finally, define and expect clear deliverables and results from the scenario planning exercise.
  2. Develop and use Actionable and Plausible Scenarios: It is not easy to devise scenarios that engender useful discussion and lead to robust plans.  On the other hand, one of the biggest benefits of scenario planning is the discussion around assumptions of different scenarios.  All scenarios are based on assumptions.  Organizations should make these assumptions known (explicitly) and allow some discussion on them.  However, the assumptions discussion should be managed effectively and stopped at some stage – otherwise the scenario-based planning discussion never actually happens.  The CIA actually publishes very good global geopolitical scenarios that can be used as a foundation.  However, scenarios that you would have to use will depend on the level at which you are doing strategic planning…
  3. Integrate Scenarios into overall planning and risk management: Once results of scenario analysis are know, use them to drive strategic planning and integrate them into risk management process.  Nothing drives implementation as much as results…

Managing Project Execution Risks (wonkish)

The Project Management journal has an article called Managing risk symptom: A method to identify major risks of serious problem projects in SI environment using cyclic causal model.  The article lays out an interesting framework for managing project execution risks in large system integration (SI) environment.  Some of the concepts are work remembering.

Serious problem projects (SPPs) often occur, particularly in a system integration environment, and it is difficult to prevent them, since the relationships among phenomena that occur throughout the project life cycle are extremely complicated. Our goal is to make it easier to identify major risks by distinguishing phenomena that are sources of future SPPs from phenomena observed in actual field projects. By choosing several events whose causal relation is known to be cyclic, we constructed a causal model and clarified that it can contribute to the easier recognition of SPPs empirically, by analyzing actual SPP cases.

The overall message is to anticipate major problem spirals by the analyzing events, understanding if the problems are root cause of a death spiral or a derivative of the death spiral and then taking effective action not only to mitigate the problem (event) but also the underlying death spiral.

The paper is a bit difficult to read – probably because I am not familiar with Japanese project management terminology and because of Japanese English….  However here are the major take aways:


Risk events have different consequences depending on the development phase.  The article divides the project into three pages upper (proposal / award), middle (early development), lower (detailed development and launch).

The article lays out a model with three types of consequences of risks based on the phase (Devil Spiral and Death Spiral):
When an event occurs or it is anticipated, the article suggests to map them to the model based on the phase of the project and then determine if it is a Derivative event – a result of the spiral (as per the article a case that is derived from a death spiral) or an Accelerating event – a root cause that accelerates the spiral.  Once done, the idea is to actively manage events, understand the nature of the spiral and take counter measure to prevent the spirals from accelerating.
Finally, here is an example of the completed analysis:

Confirmation bias in R&D management

Here is a bit of a philosophical problem that I have been thinking about for quite some time.  In the scientific world, there are all kinds of checks on proposals/decisions/results before they are accepted.  In fact, skepticism is actually somewhat welcomed. Why are R&D management decisions not subject to similar level of scrutiny?  Time and again I have found that decisions of senior R&D executives are not challenged and debated.  If innovation can only happen when when there is questioning of status quo in R&D, why not the same for R&D management innovation?

The article Confirmation bias in science: how to avoid it summarizes the problem pretty effectively (albeit in the context of scientific research):

One of the most common arguments against a scientific finding is confirmation bias: the scientist or scientists only look for data that confirms a desired conclusion. Confirmation bias is remarkably common—it is used by psychics, mediums, mentalists, and homeopaths, just to name a few.

The article had three interesting examples of confirmation bias. The one that is most applicable to R&D management and organizational pride comes from 18th century France – where the need to maintain national pride and a belief that all is well led to an amazing acceptance of bad research / decision:

… Prosper-René Blondlot announced the discovery of N-rays. He was immediately famous in France, and very shortly afterwards, researchers from around the world confirmed that they too had seen N-rays. N-rays were an ephemeral thing: observed only as a corona around an electric discharge from certain crystals. They were only observed by the human eye, making them difficult to quantify.

But not everyone was convinced. Many researchers outside of France were suspicious of the number of claims coming from French labs for the properties of N-rays. In the end, an American scientist Robert Wood visited the lab of Blondlot to see it for himself. During one of the experiments he surreptitiously removed the crystal that supposedly generated the N-rays, after which Blondlot failed to notice the absence of N-rays. The N-rays failed to vanish when their source was removed.

From an observation of many firms from my management consulting days, I find that confirmation bias is even stronger in R&D management.  In fact, many senior managers seem to surround themselves with people who actually do nothing but confirm their decisions.  Below are what I think are root causes that encourage confirmation bias in R&D management and some thoughts on what could be done about them.  I welcome any comments and criticism.
First, the process of scientific critique takes a very long time. For example, from the same Ars Techica article, the evaluation of research took 24 times as long as the work it self:

… the total amount of time coding the model? Maybe 24 hours, total. OK, call it 36 hours with some debugging. Running the code to get results? Maybe a minute per parameter set, so let’s call it a month. So that’s 32 days from around 730 total. What was all the rest of that time devoted to? Trying to anticipate every possible objection to our approach. Checking if those objections were valid. Trying to find examples of physically realistic parameters to test our model with. Seeing if the code was actually modeling what we thought it was. Making sure that our assumptions were valid. In summary, we were trying to prove ourselves wrong.

This is not practical in R&D management world.  Clearly, if it takes two years to decide a course of action, no action can be taken.  This problem has traditionally meant that management decisions can not actually be discussed or questioned.  However, I am not sure that is accurate (more about it below).

Furthermore, scientific research review is easier because the experts in the area naturally form communities along the lines of discipline.  It is always possible to find the expert with right expertise if one searches long enough:

The question session was fast and lively. And, yes, after the session, a senior scientist approached me and told me in no uncertain terms why our idea would not work—that sound you heard was me falling down the hole in our model. He was, and still is, right.

R&D management, on the other hand reaches across disciplines and there are no experts that can question results.  More importantly, all disciplines traditionally report their needs, requirements and results in a jargon specific to them.  The only person who is authorized to bridge across the jargons is the senior manager. This authority and visibility gives senior managers a unique vantage and makes it difficult for anyone else to question their decisions.

Furthermore, scientific work / decisions can be replicated by others and results tested / verified.  This is not true in R&D management world.  Decisions have long-term consequences and once made, there is hardly ever a way to test what would have happened if some other decision was made (because economic and competitive landscape changes fundamentally by the time results of decisions are visible.  This makes it difficult  for anyone to question and or critique R&D management decisions.

Finally, the consequences of failed scientific work are somewhat limited – only the lives of researchers are directly impacted.  The consequences of failed R&D management decisions are often much larger and can have significant impact on thousands of lives.  This pressure along with lack of sufficient ability to measure the effectiveness of decisions encourages R&D managers to surround themselves with people who confirm their decisions…

So what can be done about the confirmation bias:

  1. Encourage constructive criticism of R&D management decisions.  Even if the time-frame for questioning is much shorter than scientific work – an hour or a week, the fact that others view points are on the table will will actually have value in itself.  This is even more important in the new world where decisions are impacting incredibly complex systems that no one person can understand.
  2. Implement processes, tools and systems to make information necessary to make R&D management decision more broadly available: Even though disciplines participating in R&D and R&D management have all specific jargons, they are still tied together by a common thread of achieving desired objectives.  It is important to leverage this common thread and set up tools to actually elucidate the information that will let everyone – not just the R&D manager  – see the data required to make effective decisions. This will have an added advantage of validating data and making sure there are no errors.
  3. Quantify gut feelings that lead to decisions: In the end, since R&D management is always based on intuition – since no one can actually foresee the future when the results of those decisions are available.  This has traditionally meant that these qualitative decisions are not quantified in any way.  Standardized checklists are an easy way to quantify what the gut feel has been.
  4. Document decisions: Once the decisions are quantified, it is easy to document them in the tools and systems we talked about in step 2. If the decisions are easily accessible, it makes it possible to learn from them and understand why things worked or did not work.  It also makes it possible to recover or redirect if things indeed do go wrong.
  5. Develop intermediate milestones, inchstones or check points: If the only way to check the results of the decision is at the end, there is no way to recover or redirect.  By putting in place intermediate check points, especially based on key assumptions identified (step 3) and documented (step 4), R&D managers can improve their chances of success.
  6. Develop dashboards to monitor results of decisions: Combine systems in Step 2 with check points in Step 5 to develop dashboards that quickly show if things are not working – giving advance warnings to prevent catastrophic failures…

Again, I welcome any criticism (constructive or otherwise)!

How to keep your top talent

Here is what the pointy haired boss suggests:

Here are three ways to keep the top talentfrom the corporate executive board:

  1. Get to know the top talent
  2. Don’t mistake current level of performance with future potential
  3. Differentially reward top talent

Here is a bonus also from CEB – things to keep in mind for motivating your teams:

His core aim is to clearly communicate a consistent vision and then drive accountability for executing it. He’s done this by avoiding five dysfunctions on his staff that aligns well with Lencioni’s work. Lencioni’s five dysfunctions are: 1. Absence of trust 2. Fear of conflict 3. Lack of commitment 4. Avoidance of accountability 5. Inattention to results

Strategic considerations for teaming, alliances and collaborations

Management Science has a cool (at least I think so) paper on Cross-Function and Same-Function Alliances: How Does Alliance Structure Affect the Behavior of Partnering Firms:

Firms collaborate to develop and deliver new products. These collaborations vary in terms of the similarity of the competencies that partnering firms bring to the alliance. In same-function alliances, partnering firms have similar competencies, whereas in cross-function alliances, partners have very different competencies.

This is very important in co-development.  If a company in consumer electronics is co-designing a new device with a PCB manufacturer, the alliance is likely to be same-functional. Good news is that alliance between firms with similar competencies seem to work well (with caveats – see below).

On examining managers’ view of these alliances, we find that, on average, same-function alliances are expected to perform better than cross-function alliances, holding fixed the level of inputs. 

However, if the same consumer electronics firm wanted to work with a new company on wireless power, a brand new technology, the alliance might be cross-functional.  Many R&D managers are apprehensive of collaboration with dissimilar firms.  The paper uses game theory to come up with a very interesting finding that cross-functional collaboration leads to increased investments:

partners in cross-function alliances may invest more in their alliances than those in same-function alliances.

And that multiple partners are not a problem in cross-functional collaboration, but they are if the firms collaborating have similar competencies.  This is very important in Aerospace & Defense world as many government contracts do indeed have several same-functional parters:

It is also often believed that increasing the number of partnering firms is not conducive for collaborative effort. Our analysis shows that this belief is correct for same-function alliances, but not for cross-function alliances. 

Finally, a somewhat straight forward learning – once the firms have learned from each other and become more similar in competency, they do stop investing the way they used to in the cross-functional stage:

We extend our model to consider alliances where firms have an opportunity to learn from their partners and later leverage this knowledge outside the scope of their alliance. Though such learning increases the resources committed by alliance partners in the learning phase, it decreases investment in the subsequent competition and also dampens the overall investment across the two stages. 

Too many metrics?

The article The Only KPIs Your Firm Will Ever Need in AccountingWEB.com discusses why too many metrics are not value adding:

Measurement for measurement sake’s is senseless, as quality pioneer Philip Crosby understood when he uttered, ‘Building a better scale doesn’t change your weight.’

They seem to contrast it from the McKinsey maxim:

“What you can measure you can manage.”

 The example they have is of Continental turn around, where Mr. Bethune focused on just three high-level metrics:

  1. On-time arrival
  2. Lost luggage
  3. Customer complaints
Clearly, these are important for an airline.  However they are not the only metrics that need to be monitored in an operating airline.  The article makes sense and it is important for Managers to have dashboards that have unique metrics – Key Predictive Indicators.  However, in most cases, KPI need to be broken down into constituents that can be controlled to improve efficiencies.  So it is not that KPI replace detailed metrics – they provide a way to consolidate many metrics into meaningful indicators and help managers easily detect and eliminate problems.

R&D Portfolio Management case study – Microsoft Kin

It is not very  often do we get a look inside R&D management processes and tools at giants like Microsoft and Toyota. So, it is good to learn as much as we can when information becomes available.  I am studying new public disclosures on Toyota’s R&D process and will post about it soon.  The topic of interest today is Microsoft and its killing of Kin product line with only 10k units sold.  This was a big failure – the acquisition of Danger alone is reported to have been around $500M, which does not include the cost of developing the product line and associated software. Lets start off with a quick background from a great article in Ars Technica:

Microsoft’s ambitions with the KIN were sound. As much as the iPhone and, lately, Android handsets garner all the press attention, smartphones represent only a minority of phone sales—a growing minority, but a minority all the same. There are many, many people who don’t have a smartphone, and don’t even particularly want one, and they easily outnumber smartphone users.

Redmond wanted to be a part of this broader market. The company was already a big player in the smartphone market with Windows Mobile; the KIN was a product of its ambitions beyond that space. So rather than starting from scratch, in 2008 Microsoft bought Danger, the company behind the T-Mobile Sidekick line.

To a certain extent Microsoft succeeded in this new device.  Clearly they had great start from Danger and their cloud computing platform.  The idea of social networking focused phone for tweens was also great.  As the AnandTech article points out the phone did have some very good features:

KIN included a notable number of features Microsoft and its Danger team executed better than anyone else in the smartphone market today.

Amongst the notable features are innovative form factor, good usability, great battery life and aforementioned social media integration and very innovative packaging.  So why did Kin fail?  I guess the problems are related to a broad failure of R&D management processes:

  1. Portfolio Management: Executive sponsorship critical to R&D project funding
  2. Acquisition Integration: Not invented here
  3. Product Management: Positioning product as alternate to smart phones but the same cost
  4. Project Management: Significant development delays
  5. Overall R&D management: Unclear strategy, ambiguous goals 

Lets look at each one of these factors in detail:

Several posts such a Engadget and Mini-Microsoft have pointed out that Executive sponsorship is a critical part of Microsoft’s R&D portfolio management.  The Project Pink (which later became Kin) was sponsored by J. Allard, while Andy Lees sponsored a somewhat competing project of Windows Phone 7.  It is a bit strange to prioritize product portfolios based on executive sponsorship and leads to significant problems.  From engadget:

To get anywhere, a project inside Microsoft needs an executive sponsor, and for Pink, Allard had been that guy from day one. It was his baby. Of course, Allard was a visionary, an idea man; Lees — like most Microsoft execs — is a no-nonsense numbers guy, and to put it bluntly, he didn’t like that Pink existed. To quote our sources, Lees was “jealous,” and he was likely concerned that Kin was pulling mindshare (and presumably resources) from Windows Mobile’s roadmap. With enough pressure, Lees ended up getting his way; Pink fell under his charge and Allard was forced into the background

Having two competing priorities is not uncommon in R&D portfolios.  However, alignment of priorities and project pipeline gets done well in advance of launch (at early stages and continuously during portfolio reviews) in most companies with effective portfolio management processes.  In case of Microsoft however, the two projects ended being misaligned strategically, along market niches and through release schedule.  Apparently, Lees, the executive in-charge of Windows Phone 7 ended up re-aligning scopes only partially – which hurt overall results:

Having Lees in control changed everything, if for no other reason than he didn’t care about the project at all. This was right around the time that Windows Phone 7 was rebooting, and Pink didn’t fit in his game plan; to him, it was little more than a contractual obligation to Verizon, a delivery deadline that needed to be met. Pink — Allard’s vision of it, anyhow — was re-scoped, retooled, and forced onto a more standardized core that better fit in with the Windows Phone roadmap, which in turn pushed back the release date. Ironically, because they had to branch off so early, Kin would ultimately end up with an operating system that shares very little with the release version of Windows Phone 7 anyway.

Having acquired technology integrated into new products is not uncommon either.  However, there did not seem to be adequate integration of Danger into Microsoft.  The rejection of acquired technology from Danger and the move to enforce Windows Phone 7 structure on to a completely different OS ended up delaying the project by more than 18 months:

This move allegedly set the release of the devices back 18 months, during which time Redmond’s carrier partner became increasingly frustrated with the delays.

Since Windows Phone 7 is a smart phone OS and requires associated expensive hardware, this added to costs of the phone. Such a big delay in launch of the device soured relationship with the launch partner Verizon and reduced their appetite to subsidize the phone and service.

Apparently when it came time to actually bring the Kins to market, Big Red had soured on the deal altogether and was no longer planning to offer the bargain-basement pricing deals it first had tendered. The rest, as they say, is history — though we don’t think even great prices could have accounted for what was fundamentally a flawed product. Our source says that the fallout from this troubled partnership is that Microsoft has backed away from Verizon as a Windows Phone 7 launch partner, claiming that the first handsets you see won’t be offered on the CDMA carrier — rather that we should expect GSM partners to get first crack.

Product management processes ensure that product is aligned with the target market – in terms of price, functionality and usability.  Due to portfolio management failures, product management failed as well:

Some suggest that the KIN really failed because teenagers all want iPhones. There’s certainly some truth in that—iPhones are certainly aspirational goods—but iPhones are expensive. The comparison is made because the KIN was fundamentally priced like an iPhone—but it was never meant to be. Had it been priced like a Sidekick, as it should have been, and as Verizon initially set out to do, it would have substantially undercut the iPhone and been a better fit for the Facebook generation to boot. It wouldn’t do everything the iPhone could do, but it wouldn’t be operating in the same market anyway.

Furthermore, the schedule slips led to a very incomplete feature set that did not include a calender, instant messaging etc.:

That brings me to what else was lacking that was rather glaring – a calendar. With the right execution, the KIN could have perfectly integrated the Facebook event calendar, invitations, and exchange or Google calendars. Instead, the KIN has absolutely no planning tools or event notifications.

Nor was the data plan priced for the target market:

For starters, the devices lacked a realistic pricing structure – despite not quite being a smartphone, Verizon priced the data plans for the KIN as if they were, at $29.99 per month. There’s since been discussion that Verizon originally intended heavily reduced pricing for the KINs, but soured on the deal when Microsoft delayed release. At the right price, the KINs could have been a compelling alternative to the dying breed of featurephones. It’s hard to argue that there isn’t a niche that the KIN could have filled at the bottom, yet above boring featurephones. At $10 per month or less for data, the KIN would’ve been a much more successful sell.

Add to this mix strategic direction problems: First Microsoft could not get itself to support competitive social networking sites such as Flickr – which actually reduced the value of the product

The giants of the social networking space include Facebook and Twitter, for which Kin offered at least fair support. But rather than support Flickr for images and (Google-owned) YouTube for video, Microsoft plugged in its Windows Live services for these media. Kin also lacked established functionality such as a calendar and instant messaging as well as support for fast-growing services embraced by social networkers such as Foursquare.

Also attempting multiple very conflicting goals (compete with iPhone and be cheap like sidekick) led to muddled execution:

The heart of both Microsoft’s and Google’s mobile operating system strategy is to have diverse handsets running its software. Still, both companies look at the level of integration Apple can achieve with the iPhone and are drawn to have a heavier hand in the design of handsets. This sort of licensor regret is part of what drove Google to create the Nexus One and likely also contributed to Microsoft’s decision to create the Kin handsets. 

It appears that in absence of clear portfolio goals and metrics, there is a lot of politics – which further reduces efficiency and efficacy, drives morale down and leads to rumors of layoff:

But wait, there’s more — the Kin team is being refocused onto the WP7 project, but that’s not the only shakeup going on. Our source said there had been rumblings that Steven Sinofsky — president of the Windows and Windows Live groups — is making a play for the entire mobile division as well in an attempt to bring a unified, Windows-centric product line to market.

I hope you are with me: portfolio management problems along with acquisition challenges problems led to project management problems and resulted in very large schedule slips. Those schedule slips along with feature creep and problems with product management.  Unclear strategic direction and management in-infighting tied everything together and ensured that the project failed.

Many lessons to be learned here!  What do you think?