BIM_PLMA year ago I wrote a blog post questioning if the construction industry would learn from PLM practices in relation to BIM.
In that post, I described several lessons learned from other industries. Topics like:

  • Working on a single, shared repository of on-line data (the Digital Mock Up).  Continuity of data based on a common data model – not only 3D
  • It is a mindset. People need to learn to share instead of own data
  • Early validation and verification based on a virtual model. Working in the full context
  • Planning and anticipation for service and maintenance during the design phase. Design with the whole lifecycle in mind (and being able to verify the design)

The comments to that blog post already demonstrated that the worlds of PLM and BIM are not 100 percent comparable and that there are some serious inhibitors preventing them to come closer. One year later, let´s see where we are:

BIM moving into VDC (or BLM ?)

discussThe first trend that becomes visible is that people in the construction industry start to use more and more the term Virtual Design and Construction (VDC)  instead of BIM (Building Information Model or Building Information Management?).

The good news here is that there is less ambiguity with the term VDC instead of BIM. Does this mean many BIM managers will change their job title? Probably not as most construction companies are still in the learning phase what a digital enterprise means for them.

Still Virtual Design and Construction focuses a lot on the middle part of the full lifecycle of a construction. VDC does not necessary connect the early concept phase and for sure almost neglects the operational phase. The last phase is often ignored as construction companies are not thinking (yet) about Repair & Maintenance contracts (the service economy).

And surprisingly, last week I saw a blog post from Dassault Systemes, where Dassault introduced the word BLM (Building Lifecycle Management).  Related to this blog post also some LinkedIn discussions started.  BLM, according to Dassault Systemes, is the combination of BIM and PLM – read this post here.

The challenge however for construction companies is to, what are the related data sets they require and how can you create this continuity of data. This brings us to one of the most important inhibitors.

 

Data Ownership

imageWhere in other industries a clear product data owner exists, the ownership of data in EPC (Engineering, Procurement, Construction) companies, typical for the construction industry or oil & gas industry is most of the times on purpose vague.

First of all the owner of a construction often does not know which data could be relevant to maintain. And secondly, as soon as the owner asks for more detailed information, he will have to pay for that, raising the costs, which not directly flow back to benefits, only later during the FM (Facility Management) /Operational stage.

And let´s imagine the owner could get the all the data required. Next the owner is at risk, as potentially having the information might makes you liable for mistakes and claims.

From discussion with construction owners I learned their policy is not to aim for the full dataset related to a construction. It reduces the risk to be liable. Imagine Boeing and Airbus would follow this approach. This brings us to another important inhibitor.

A risk shifting business

imageThe construction industry on its own is still a risk shifting business, where each party tries to pass the risk of cost of failure to another stakeholder in the pyramid. The most powerful owners / operators of the construction industry quickly play down the risk to their contractors and suppliers. And these companies then then distribute the risk further down to their subcontractors.

If you do not accept the risk, you are no longer in the game. This is different from other industries and I have seen this approach in a few situations.

imageFor example, I was dealing with an EPC company that wanted to implement PLM. The company expected that the PLM implementer would take a large part of the risk for the implementation. As they were always taking the risk too for their big customers when applying for a project. Here there was a clash of cultures, as PLM implementers learned that the risk of a successful PLM implementation is vague as many soft values define the success. It is not a machine or platform that has to work after some time.

imageAnother example was related to requirements management. Here the EPC company wanted to become clear and specific to their customer. However their customer reacted very strange. Instead of being happy that the EPC company invested in more upfront thinking and analysis, the customer got annoyed as they were not used to be specific so early in the process. They told the EPC company, “if you have so many questions, probably you do not understand the business”.

So everyone in the EPC business is pushed to accept a higher risk and uncertainty than other industries. However, the big reward is that you are allowed to have a cost of failure above 15 – 20 percent without feeling bad. Which this percentage you would be out of business in other industries. And this brings us to another important inhibitor.

Accepted high cost of failure

No_roiAs the industry accepts this high cost of failure, companies are not triggered to work different or to redesign their processes in order to lower the inefficiencies. The UK government mandates BIM Level 2 for their projects starting in 2016 and beyond, to reduce costs through inefficiencies.

But will the UK government invest to facilitate and aim for data ownership? Probably not, as the aim of governments is not to be extreme economical. Being not liable has a bigger value than being more efficient for governments as I learned. Being more efficient is the message to the outside world to keep the taxpayer satisfied.

It is hard to change this way of thinking. It requires a cultural change through the whole value chain. And cultural change is the “worst” thing that can happen to a company. The biggest inhibitor.

Cultural change

imageCultural change is a point that touches all industries and there is no difference between the construction industry and for example a classical discrete manufacturing company. Because of global competition and comparable products other industries have been forced already to work different, in order to survive (and are still challenged)

The cultural change lies in people. We (the older generation) are educated and brought up in classical engineering models that reflect the post second world war best practices. Being important in a process is your job justification and job guarantee.

New paradigms, based on a digital world instead of a document-shifting world, need to be defined and matured and will make many classical data processing jobs redundant. Read this interesting article from the Economist: The Onrushing Wave

This is a challenge for every company. The highest need to implement this cultural change is ironically for those countries with the highest legacy: Western Europe / the United-States.

As these countries also have the highest labor cost, the impact of, keep on doing the old stuff, will reduce their competitiveness. The impact for construction companies is less, as the construction industry is still a local business, as at the end resources will not travel the globe to execute projects.

However cheaper labor costs become more and more available in every country. If companies want to utilize them, they need to change the process. They need shift towards more thinking and knowledge in the early lifecycle to avoid the need for high qualified people to be in the field to the fix errors.

Sharing instead of owning

imageFor me the major purpose of PLM is to provide an infrastructure for people to share information in such a manner that others, not aware of the information, can still easily find and use the information in a relevant context of their activities. The value: People will decide on actual information and no longer become reactive on fixing errors due to lack of understanding the context.

The problem for the construction industry is that I have not seen any vendor focusing on sharing the big picture. Perhaps the BLM discussion will be a first step. For the major tool providers, like Autodesk and Bentley, their business focus is on the continuity of their tools, not on the continuity of data.

Last week I noticed a cloud based Issue Management solution, delivered by Kubus. Issue Management is one of the typical and easy benefits a PLM infrastructure can deliver. In particular if issues can be linked to projects, construction parts, processes, customers. If this solution becomes successful, the extension might be to add more data elements to the cloud solution. Main question will remain: Who owns the data ? Have a look:

Cloud based Issues Management

 

For continuity of data, you need standards and openness – IFC is one of the many standards needed in the full scope of collaboration. Other industries are further developed in their standards driven by end-user organizations instead of vendors. Companies should argue with their vendors that openness is a right, not a privilege.

Conclusion

A year ago, I was more optimistic about the construction industry adopting PLM practices. What I have learned this year, and based on feedback from others, were are not at the turning point yet. Change is difficult to achieve from one day to the other. Meanwhile, the whole value chain in the construction industry has different objectives. Nobody will take the risk or can afford the risk.

I remain interested to see where the construction industry is heading.

What do you think will 2015 be the year of a breakthrough?

coopIn the past two years, I have been heavily involved in PLM Proof of Concepts sitting at both sides of the table. Supporting companies in their PLM selection, supporting a vendor explaining their value to the customer and supporting implementers assisting them with industry knowledge, all in the context of a PLM selection process.

The Proof of Concept is crucial in a PLM selection process as it is the moment where the first glimpse of reality comes to the table.

Different size of companies, different consultants all have a different view on the importance of the Proof of Concept. Let me share you my thoughts after a quick recap on the PLM selection process.

The PLM selection process

1. Build a vision

visionIt is important that a company understands what they want to achieve in the next five to ten years, before starting a PLM selection process. Implementing PLM means a business transformation, even if you are a small company. If the management does not understand a vision is required, there is a potential risk upcoming, as PLM without a change in the way people work, will not deliver the expected results.

2. Issue an RFI to potential candidates

rfi-plmOnce you have a PLM vision, it is time to get in touch with potential suppliers. The RFI (Request for Information) phase is the phase where you can educate yourself better by challenging the suppliers to work with you on the future solutions.

3. Discuss with selected candidates

discussFrom the RFI responses you understand which companies are attractive because they match your vision, your budget or industry. Have a first interaction with the selected companies and let them demo their standard environment targeted to your vision.

4. POC

test conceptIn this stage, you check with the preferred companies their ability to deliver and your ability to work together. The POC phase should give you the understanding of the scope for the upcoming PLM project and help you to understand who and how the project can be executed. More details about this step below.

5. RFP

No_roiAlthough some companies start with an RFP before the POC, for me it makes most sense to verify the details after you have a proper understanding of the To-Be solution. The RFP is often the base for the contractual scope and therefore should be as accurate as possible

In the past, I wrote in more detail about the PLM selection process. Two posts:  PLM selection: Don’t do this and PLM selection: Do this. Have a read if you want to understand this part in more depth. Now let´s focus on the POC .

POC targets

  • As described before, the target of the Proof of Concept should be to get a better understanding of the potential To-Be processes and obtain an impression of the capabilities of the implementer and the preferred PLM software.

The result should be that you have more realistic expectations of what can be achieved and the challenges your company will face.

  • From there, you can evaluate the risks, address them and build an achievable roadmap to implement. It is important that the focus is not just on the cost of the implementation.
  • To sell PLM inside your company, you need to realign with the vision and explain, to all people involved,the value of “Why PLM”.

Explaining the value is complex, as not everyone needs the same message. The management will focus on business benefits where users will focus how it impacts their daily life.  If you forget to explain the value, the PLM projects, it is considered again as just another software purchase.

POC DO’s

businessMake sure the Proof of Concept is driven by validating future business scenarios, focusing on the To-Be solution. The high-level scenarios should be demonstrated and explained to the business people. In this stage, it is important people realize the benefits and the value of the new processes.

sales eventThe POC is also an internal sales event. The goal should be to get more enthusiastic and supportive business people in your company for the upcoming PLM project. Identify the champions you will need to lean on during the implementation.

balanceTest the implementer. To my opinion the critical success of a PLM implementation depends on the implementation team, not on the software. Therefore, the POC phase is the best moment to learn if you can work with the implementer. Do they know your business? Do they have experience with your business? The more you are aligned, the higher the chance you will be successful as a team

commitShow commitment to engage. Often I have seen POC engagements where the company demanded the implementer or vendor a Proof of Concept for free. This creates an unbalanced situation during the Proof of Concept as the vendor or implementer can not invest time and resources in the process as expected without any commitment from the company. By paying a certain fee for the POC, a company can demonstrate to the implementer /vendor that this POC is valuable for you and you can request the same response from them.

POC DON’Ts

no detailsThe Proof of Concept is not a detailed function/feature check to identify each mouse-click or option in the system. During the implementation, these details might come up. It is important in a Proof of Concept to understand the big picture and not to get lost in the details. As human beings we tend to focus on what does not work, not realizing that probably over eighty-ninety percent works according the needs

ultimateDo not expect the ultimate To-Be scenario demonstrated during the Proof of Concept. The Proof of Concept is a learning stage for both the company and the implementer to imagine the best possible scenario. PLM systems are generic and likely they will not provide a similar configuration and functionality matching your environment. At this stage validate if the primary capabilities are there and if there are gaps.

plm vendorDo not run a POC with a vendor (only). This might be one of the most critical points for a POC. A PLM software vendor’s target is to sell their software and for that reason they often have dedicated presales teams that will show you everything in a smooth manner, overwhelming you with all the beauty of the software. However after the POC this team is gone and you will have to align yourself again with the implementation partner, trying to match again your business needs and their understanding.

imageRealize – you get what you are asking for. This is more a Do-and-Don’t message packed together. A Proof of Concept phase is a point where companies get to know each other. If you are not focused, do not expect the implementer / vendor to be committed. A PLM implementation is not product. It is a business transformation supported by products and services. Do not treat PLM implementers and vendors in the same way, as your customers treat you (in case you deliver products).

Conclusion

There are still many more thoughts about the Proof of Concept . Ideally you run two POCs in parallel, either with two implementers of the preferred software (if possible) or with two different implementers representing different software.

Ideally, as I know it is a challenge, especially for small and medium-sized businesses, where people are running to keep the business on-going.

Still remember, PLM is a business transformation, targeting to improve your business in the upcoming five to ten years, avoiding you are running out of business.

Your thoughts ?

As a bonus a short anecdote that I posted in 2010 still relevant:

plm heaven or hell

Some time ago a Christian PLM Sales professional died (let’s call him Jack) and according to his believe he faced Saint Peter at the gates of Heaven and Hell.
Saint Peter greeted Jack and said: “Jack, with the PLM Sales you have done good and bad things to the world. For that reason, I cannot decide if you should go to Heaven or to Hell. Therefore, I allow you to make the choice yourself”.

Jack replied: “But Saint Peter, how can I make such an important decision for the rest of my eternal life. It is too difficult!”

Saint Peter replied: “No problem Jack, take a look at Heaven and Hell, take your time and then tell me your decision.”

Jack entered Heaven and he was surprised about the quietness and green atmosphere there. Angels were singing, people were eating from golden plates with the best food ever, people were reading poetry and everything was as peaceful as you could imagine. In the distance, he could see God surrounded by some prophets talking about the long-term future. After some time, Jack had seen it and went to Hell to have a view there.

And when he opened the gates of Hell, he was astonished. Everywhere he looked there were people partying, having fun. It reminded him off these sales kick-offs, he had in the past, exotic places with lots of fun. In the distance, he could see the Devil as DJ playing the latest dance music – or was it DJ Tiësto?

Jack did not hesitate and ran back to Saint Peter, no time to lose. “Saint Peter,” he said “I want to go to Hell, no doubt. And pity I did not know it before”

“So be it, ” said Saint Peter “go for it.”

And then once Jack entered Hell, it was suddenly all fire around him, people were screaming of pain and suffering and also Jack felt the first flames.

“Devil!!”  He screamed “what happened to what I have seen before?”

With a sarcastic voice, the devil replied: “That? That was a proof of concept.”

Shaping the PLM platform of the Future

2050In this post my observations from the PDT 2014 Europe conference which was hosted in the Microsoft Conference center in Paris and organized by Eurostep and CIMdata.

It was the first time I attended this event. I was positively surprised about the audience and content. Where other PLM conferences were often more focusing on current business issues, here a smaller audience (130 persons) was looking into more details around the future of PLM. Themes like PLM platforms, the Circular Economy, Open Standards and longevity of data were presented and discussed here.

The emergence of the PLM platform

SNAGHTML149e44b9Pieter Bilello from CIMdata kicked off with his presentation: The emergence of the PLM platform. Peter explained we have to rethink our PLM strategy for two main reasons:

1.  The product lifecycle will become more and more circular due to changing business models and in parallel the different usage/availability of materials will have an impact how we design and deliver products

2.  The change towards digital platforms at the heart of our economy (The Digital Revolution as I wrote about also in previous posts) will impact organizations dramatically.

Can current processes and tools support today’s complexity. And what about tomorrow? According to a CIMdata survey there is a clear difference in profit and performance between leaders and followers, and the gap is increasing faster. “Can you afford yourself to be a follower ?” is a question companies should ask themselves.

Rethinking PLM platform does not bring the 2-3 % efficiency benefit but can bring benefits from 20 % and more.

Peter sees a federated platform as a must for companies to survive. I in particular likes his statement:

The new business platform paradigm is one in which solutions from multiple providers must be seamlessly deployed using a resilient architecture that can withstand rapid changes in business functions and delivery modalities

Industry voices on the Future PLM platform

Auto

SNAGHTML14a2180eSteven Vetterman from ProSTEP talked about PLM in the automotive industry. Steven started describing the change in the automotive industry, by quoting Heraclitus Τα πάντα ρεί – the only constant is change. Steven described two major changes in the automotive industry:

1.  The effect of globalization, technology and laws & ecology

2.  The change of the role of IT and the impact of culture & collaboration

Interesting observation is that the preferred automotive market will shift to the BRIC countries. In 2050 more than 50 % of the world population (estimate almost 10 billion people at that time) will be living in Asia, 25 percent in Africa. Europe and Japan are aging. They will not invest in new cars.

For Steven, it was clear that current automotive companies are not yet organized to support and integrate modern technologies (systems engineering / electrical / software) beyond mechanical designs. Neither are they open for a true global collaboration between all players in the industry. Some of the big automotive companies are still struggling with their rigid PLM implementation. There is a need for open PLM, not driven from a single PLM system, but based on a federated environment of information.

Aero

Yves Baudier spoke on behalf of the aerospace industry about the standardization effort at their Strategic Standardization Group around Airbus and some of its strategic suppliers, like Thales, Safran, BAE systems and more. If you look at the ASD Radar, you might get a feeling for the complexity of standards that exist and are relevant for the Airbus group.

standards at airbus

It is a complex network of evolving standard all providing (future) benefits in some domains. Yves was talking about the through Lifecycle support which is striving for data creation once and reuse many times during the lifecycle. The conclusion from Yves, like all the previous speakers is that: The PLM Platform of the Future will be federative, and standards will enable PLM Interoperability

Energy and Marine

SNAGHTML14a7edf3Shefali Arora from Wärtsilä spoke on behalf of the energy and marine sector and gave an overview of the current trends in their business and the role of PLM in Wärtsilä. With PLM, Wärtsilä wants to capitalize on its knowledge, drive costs down and above all improve business agility. As the future is in flexibility. Shefali gave an overview of their PLM roadmap covering the aspects of PDM (with Teamcenter), ERP (SAP) and a PLM backbone (Share-A-space). The PLM backbone providing connectivity of data between all lifecycle stages and external partners (customer / suppliers) based on the PLCS standard. Again another session demonstrating the future of PLM is in an open and federated environment

Intermediate conclusion:
The future PLM platform is a federated platform which adheres to standards provides openness of interfaces that permit the platform to be reliable over multiple upgrade cycles and being able to integrate third-parties (Peter Bilello)

Systems Engineering

imageThe afternoon session I followed the Systems Engineering track. Peter Bilello gave an overview of Model-Based Systems engineering and illustrated based on a CIMdata survey that even though many companies have a systems engineering strategy in place it is not applied consistently. And indeed several companies I have been dealing with recently expressed their desire to integrate systems engineering into their overall product development strategy. Often this approach is confused by believing requirements management and product development equal systems engineering. Still a way to go.

Dieter Scheithauer presented his vision that Systems Engineering should be a part of PLM, and he gave a very decent, academic overview how all is related. Important for companies that want to go into that direction, you need to understand where you aiming at. I liked his comparison of a system product structure and a physical product structure, helping companies to grab the difference between a virtual, system view and a physical product view:

system and product

More Industry voices

Construction industry

imageThe afternoon session started with Christophe Castaing, explaining BIM (Building Information Modeling) and the typical characteristics of the construction industry. Although many construction companies focus on the construction phase, for 100 pieces of information/exchange to be managed during the full life cycle only 5 will be managed during the initial design phase (BIM), 20 will be managed during the construction phase (BAM) and finally 75 will be managed during the operation phase (BOOM). I wrote about PLM and BIM last year: Will 2014 become the year the construction industry will discover PLM?

Christophe presented the themes from the French MINnD project, where the aim is starting from an Information Model to come to a platform, supporting and integrated with the particular civil and construction standards, like IFC. CityGml but also PLCS standard (isostep ISO 10303-239

Consumer Products

Amir Rashid described the need for PLM in the consumer product markets stating the circular economy as one of the main drivers. Especially in consumer markets, product waste can be extremely high due to the short lifetime of the product and everything is scrapped to land waste afterward. Interesting quote from Amir: Sustainability’s goal is to create possibilities not to limit options. He illustrated how Xerox already has sustainability as part of their product development since 1984. The diagram below demonstrates how the circular economy can impact all business today when well-orchestrated.

circular economy

SNAGHTML14b000f6Marc Halpern closed the tracks with his presentation around Product Innovation Platforms, describing how Product Design and PLM might evolve in the upcoming digital era. Gartner believes that future PLM platforms will provide insight (understand and analyze Big Data), Adaptability (flexible to integrate and maintain through an open service oriented architecture), promoting reuse (identifying similarity based on metadata and geometry), discovery (the integration of search analysis and simulation) and finally community (using the social paradigm).

If you look to current PLM systems, most of them are far from this definition, and if you support Gartner’s vision, there is still a lot of work for PLM vendor to do.

Interesting Marc also identified five significant risks that could delay or prevent from implementing this vision:

  • inadequate openness (pushing back open collaboration)
  • incomplete standards (blocking implementation of openness)
  • uncertain cloud performance (the future is in cloud services)
  • the steep learning curve (it is a big mind shift for companies)
  • Cyber-terrorism (where is your data safe?)

After Marc´s session there was an interesting panel discussion with some the speakers from that day, briefly answering discussing questions from the audience. As the presentations have been fairly technical, it was logical that the first question that came up was: What about change management?

A topic that could fill the rest of the week but the PDT dinner was waiting – a good place to network and digest the day.

DAY 2

imageDay 2 started with two interesting topics. The first presentation was a joined presentation from Max Fouache (IBM) and Jean-Bernard Hentz (Airbus – CAD/CAM/PDM R&T and IT Backbones). The topic was about the obsolescence of information systems: Hardware and PLM applications. As in the aerospace industry some data needs to be available for 75 years. You can imagine that during 75 years a lot can change to hardware and software systems. At Airbus, there are currently 2500 applications, provided by approximate 600 suppliers that need to be maintained. IBM and Airbus presented a Proof of Concept done with virtualization of different platforms supporting CATIA V4/V5 using Linux, Windows XP, W7, W8 which is just a small part of all the data.

The conclusion from this session was:

To benefit from PLM of the future, the PLM of the past has to be managed. Migration is not the only answer. Look for solutions that exist to mitigate risks and reduce costs of PLM Obsolescence. Usage and compliance to Standards is crucial.

Standards

Next Howard Mason, Corporate Information Standards Manager took us on a nice journey through the history of standards developed in his business. I loved his statement: Interoperability is a right, not a privilege

imageIn the systems engineering track Kent Freeland talked about Nuclear Knowledge Management and CM in Systems Engineering. As this is one of my favorite domains, we had a good discussion on the need for pro-active Knowledge Management, which somehow implies a CM approach through the whole lifecycle of a plant. Knowledge management is not equal to store information in a central place. It is about building and providing data in context that it can be used.

Ontology for systems engineering

Leo van Ruijven provided a session for insiders: An ontology for Systems Engineering based on ISO 15926-11. His simplified approach compared to the ISO 15288 lead to several discussion between supporters and opponents during lunch time.

Master Data Management

imageAfter lunch time Marc Halpern gave his perspective on Master Data Management, a new buzz-word or discipline need to orchestrate enterprise collaboration.

Based on the type of information companies want to manage in relation to each other supported by various applications (PLM, ERP, MES, MRO, …) this can be a complex exercise and Marc ended with recommendations and an action plan for the MDM lead. In my customer engagements I also see more and more the digital transformation leads to MDM questions. Can we replace Excel files by mastered data in a database?

SNAGHTML14c68ed3

Almost at the end of the day I was speaking about the PDM platform of the people targeted for the people from the future. Here I highlighted the fundamental change in skills that’s upcoming. Where my generation was trained to own and capture information as much as possible information in your brain (or cabinet), future generations are trained and skilled in finding data and building information out of it. Owning (information) is not crucial for them. Perhaps as the world is moving fast. See this nice YouTube movie at the end.

image

Ella Jamsin ended the conference on behalf of the Ellen MacArthur Foundation explaining the need to move to a circular economy and the PLM should play a role in that. No longer is PLM from cradle-to-grave but PLM should support the lifecycle from cradle-to-cradle.

Unfortunate I could not attend all sessions as there were several parallel sessions. Neither have I written about all sessions I attended. The PDT Europe conference, a conference for people who mind about the details around the PLM future concepts and the usage of standards, is a must for future strategists.

changeBusiness is changing and becoming digital as you might have noticed. If you haven´t noticed it, you might be disconnected from the world or work in a stable silo. A little bit simplified and provocative otherwise you would not read further.

The change towards digital also has its effect on how PLM is evolving. Initially considered as an extension of PDM, managing engineering data, slowly evolving to an infrastructure to support the whole product lifecycle.

The benefits from a real PLM infrastructure are extremely high as it allows people to work smarter, identify issues earlier and change from being reactive towards proactive. In some industries, this change in working is they only way to stay in business. Others with still enough margin will not act.

Note: I am talking about a PLM infrastructure as I do not believe in a single PLM system anymore. For me PLM is supported through a collection of services across the whole product lifecycle, many potentially in one system or platform.

Changing from an engineering-centric system towards an infrastructure across the departmental silos is the biggest challenge for PLM. PLM vendors and ERP vendors with a PLM offering are trying provide this infrastructure and mainly fight against Excel. As an Excel file can easy pass the border from one department to the other. No vision needed for Excel.

A PLM infrastructure however requires a vision. A company has to look at its core business processes and decide on which information flows through the organization or even better their whole value chain.

Building this vision, understanding this vision and then being able to explain the vision is a challenge for all companies. Where sometime even management says

“Why do we need to have a vision, just fix the problem”

also people working in departments are not looking forward to change their daily routines because they need to share information. Here you here statements like

“Why people feel the need to look at the big picture. I want to have my work done.”

So if current businesses do not change, will there be a change?

Innovation

Here I see the digital world combined with search-based applications coming up. Search based applications allow companies to index their silos and external sources and get an understanding of the amount of data there exists. And from these results learn that there is a lot of duplicated data or invalid information at different places.

This awareness might create the understanding that if instead of having hundred thousands of Excels in the organization, it would be better to have the data inside a database, uniquely stored and connected to other relevant information.

I have described this process in my past three posts and agreed the context remains complex if you are not involved or interested in it.

To grasp this concept first of all, you need to have the will to understand it. There is a lot of strategic information out there from companies like Accenture, Capgemini and more.

Next if you want to understand it in a more down-to-earth manner it is important to listen and talk with your peers from other companies, other industries. This is currently happening all around the world and I invite you to participate.

Here is a list of events that I am attending or planned to attend but too far away:

October 7th Stockholm – ENOVIA user conference

imageHere I will participate as a panel member in the discussion around the concept of zero files. Here we want to explain and discuss to the audience what a data-centric approach means for an organization. Also, customers will share their experiences. This conference is focusing on the ENOVIA community – you can still register here

October 14th-15th Paris – Product Data Technology 2014 conference

pdteuropeHere I will speak about the PLM future (based on data) and what PLM should deliver for the future generations. This conference is much broader and addresses all PLM related topics in a broader perspective

October 28 Product Innovation conference in San Diego

picongressI have always enjoyed participating to this conference as like the PDT2014 it brings people together for networking and discussions. Mostly on business topics not on IT-issues.

November 26 Infuseit seminar in Copenhagen

imageRelative new in the Nordics Infuseit, a PLM consultancy company, is able to attract an audience that wants to work on understanding the PLM future. Instead of listening to presenters, here you are challenged to to discuss and contribute to build a common opinion. I will be there too.

 

Conclusion: It is time to prepare yourself for the change – it is happening and be educated an investment that will be rewarding for your company

What do you think – Is data-centric a dream ?

2050This is for the moment the last post about the difference between files and a data-oriented approach. This time I will focus on the need for open exchange standards and the relation to proprietary systems. In my first post, I explained that a data-centric approach can bring many business benefits and is pointing to background information for those who want to learn more in detail. In my second post, I gave the example of dealing with specifications.

It demonstrated that the real value for a data-centric approach comes at the moment there are changes of the information over time. For a specification that is right the first time and never changes there is less value to win with a data-centric approach. Moreover, aren’t we still dreaming that we do everything right the first time.

The specification example was based on dealing with text documents (sometimes called 1D information). The same benefits are valid for diagrams, schematics (2D information) and CAD models (3D information)

1D,2D,3D …..

1DThe challenge for a data-oriented approach is that information needs to be stored in data elements in a database, independent of an individual file format. For text, this might be easy to comprehend. Text elements are relative simple to understand. Still the OpenDocument standard for Office documents is in the background based on a lot of technical know-how and experience to make it widely acceptable. For 2D and 3D information this is less obvious as this is for the domain of the CAD vendors.

CAD vendors have various reasons not to store their information in a neutral format.

  • First of all, and most important for their business, a neutral format would reduce the dependency on their products. Other vendors could work with these formats too, therefore reducing the potential market capture. You could say that in a certain manner the Autodesk 2D format for DXF (and even DWG) have become a neutral format for 2D data as many other vendors have applications that read and write back information in the DXF-data format. So far DXF is stored in a file but you could store DXF data also inside a database and make it available as elements.
  • This brings us to the second reason why using neutral data formats are not that evident for CAD vendors. It reduces their flexibility to change the format and optimize it for maximal performance. Commercially the significant, immediate disadvantage of working in neutral formats is that it has not been designed for particular needs in an individual application and therefore any “intelligent” manipulations on the data are hard to achieve..

3dThe same reasoning can be applied to 3D data, where different neutral formats exist (IGES, STEP, …. ). It is very difficult to identify a common 3D standard without losing many benefits that an individual 3D CAD format brings currently. For example, CATIA is handling 3D CAD data in a complete different way as Creo does, and again handled different compared to NX, SolidWorks, Solid Edge and Inventor. Even some of them might use the same CAD kernel.

However, it is not only about the geometry anymore; the shapes represent virtual objects that have metadata describing the objects. In addition other related information exists, not necessarily coming from the design world, like tasks (planning), parts (physical), suppliers, resources and more

PLM, ERP, systems and single source of truth

This brings us in the world of data management, in my world mainly PLM systems and ERP systems. An ERP system is already a data-centric application, the BOM is already available as metadata as well as all the scheduling and interaction with resources, suppliers and financial transactions. Still ERP systems store a lot of related documents and drawings, containing content that does not match their data model.

PLM systems have gradually becoming more and more data centric as the origin was around engineering data, mostly stored in files. In a data-centric approach, there is the challenge to exchange data between a PLM system and an ERP system. Usually there is a need to share information between two systems, mainly the items. Different definitions of an item on the PLM and ERP side make it hard to exchange information from one system to the other. It is for that reason why there are many discussions around PLM and ERP integration and the BOM.

ebom_mbom_problem

In the modern data-centric approach however we should think less and less in systems and more and more in business processes performed on actual data elements. This requires a company-wide, actually an enterprise-wide or industry-wide data definition of all information that is relevant for the business processes. This leads into Master Data Management, the new required skill for enterprise solution architects

black holeThe data-centric approach creates the impression that you can achieve a single source of the truth as all objects are stored uniquely in a database. SAP solves the problem by stating everything fits in their single database. To my opinion this is more a black hole approach: Everything gets inside, but even light cannot escape. Usability and reuse of information that was stored with the intention not to be found is the big challenge here.

Other PLM and ERP vendors have different approaches. Either they choose for a service bus architecture where applications in the background link and synchronize common data elements from each application. Therefore, there is some redundancy, however everything is connected. More and more PLM vendors focus on building a platform of connected data elements, where on top applications will run, like the 3DExperience platform from Dassault Systèmes.

androidAs users we are more and more used to platforms as Google, Apple provide these platforms already in the cloud for common use on our smartphones. The large amount of apps run on shared data elements (contacts, locations …) and store additional proprietary data.

Platforms, Networks and standards

And here we enter an interesting area of discussion. I think it is a given that a single database concept is a utopia. Therefore, it will be all about how systems and platforms communicate with each other to provide in the end the right information to the user. The systems and platforms need to be data-centric as we learned from the discussion around the document (file centric) or data-centric approach.

In this domain, there are several companies already active for years. Datamation from Dr. Kais Al-Timimi in the UK is such a company. Kais is a veteran in the PLM and data modeling industry, and they provide a platform for data-centric collaboration. This quote from one of his presentations, illustrates we share the same vision:

“……. the root cause of all interoperability and data challenges is the need to transform data between systems using different, and often incompatible, data models.

It is fundamentally different from the current Application Centric Approach, in that data is SHARED, and therefore, ‘NOT OWNED’ by the applications that create it.

This means in a Data Centric Approach data can deliver MORE VALUE, as it is readily sharable and reusable by multiple applications. In addition, it removes the overhead of having to build and maintain non-value-added processes, e.g. to move data between applications.”

Another company in the same domain is Eurostep, who are also focusing on business collaboration between in various industries. Eurostep has been working with various industry standards, like AP203/214, PLCS and AP233. Eurostep has developed their Share-A-space platform to enable a data-centric collaboration.

ISO-BIMThis type of data collaboration is crucial for all industries. Where the aerospace and automotive industry are probably the most mature on this topic, the process industry and construction industry are currently also focusing on discovering data standards and collaboration models (ISO 15926 / BIM). It will be probably the innovators in these industries that clear the path for others. For sure it will not come from the software vendors as I discussed before.

Conclusion

If you reach this line, it means the topic has been interesting in depth for you. In the past three post starting from the future trend, an example and the data modeling background, I have tried to describe what is happening in a simplified manner.

If you really want to dive into the PLM for the future, I recommend you visit the upcoming PDT 2014 conference in Paris on October 14 and 15. Here experts from different industries will present and discuss the future PLM platform and its benefits. I hope to meet you there.

pdteurope

 

Some more to read:

https://us.sogeti.com/wp-content/uploads/2014/04/PLM-Systems-White-Paper.pdf

dataIn my previous post, I talked about the unstoppable trend towards digital information and knowledge based on data becoming the new business paradigm.

Building knowledge based on information extracted from data, instead of working with documents and people, who need to manipulate these documents.

Moreover, the reasons to move towards a digital data-oriented approach are the immense business benefits it can bring to an organization. Having online visibility on information in context of other information from different stakeholders allows companies to be more proactive.

A proactive company will react faster to the market or customer. This will reduce waste and resources (materials / people) and therefore in the end be more competitive. This is all described in my first post, with relevant links to various global references.

In this post, I want to describe in an example what the differences are between a document-oriented and a data-oriented approach and how it affects people and business. This might give you an impression of the expected business benefits.

The ultimate goal behind a data-oriented approach is to have a single version of the truth for a product, project or plant. This can be realized by treating information as data elements in various connected database, where on demand reports or dashboards can be created based on actual information, instead of documents generated by duplicating data in new systems and locations. Digital data will provide paperless processes accessible almost anywhere around the world.

push or pull

As an example, I will explain the difference between document-centric and data-centric when dealing with specifications.

The specification

Everyone knows the challenge with specifications. Most of the time in printed documents describing how a product or service should work from the client point of view. There are two principles behind specifications:

  • Complexity. The more complex product or service is, the bigger chance that specifications are not complete or hundred percent understood, leading to an iterative change process. The challenge here is to manage the change and the consistency of the full specifications.
  • Industry and margins. In a repetitive business, for example, the automotive or other mass consumer products, products can be quiet complex and once sold hard to maintain and repair. In a competitive business, an error in the field can consume a lot of the expected profit. In the construction industry, where most of the time single projects are executed by a chain of disciplines, the industry (still) accepts the costs overrun and the high costs of fixing issues in the field, instead of being clearer upfront during the design and planning.

Let’s stay with the example in the middle of complexity and industry volume. In color the various stages of the process.

The document based specification – 1

pdfWhen the document based specification arrives, the company has to get an understanding of the content. The project manager has a first read through the document (100+ pages) and decides to send the document (it is a pdf) to sales, engineering, legal and planning. Engineering decides to distribute the document internally to mechanical, electrical and quality (for compliance).

The project manager stresses everyone on a weekly base to deliver the responses and tries to understand if the answers will come in time. There are some meetings needed with the stakeholders as the whole understanding needs to be consistent. Based on several iterations a response is compiled.

The data-oriented approach – 1

ReqStructureWhen the document based specification arrives, the project leader first stores the document as a reference in the PLM system and extracts all the customer requirements as data elements in the system. While extracting the requirements, the projects manager groups them into digital folders (functional / non-functional, contractual, regulations, etc.) and assigns them to the relevant stakeholders, who get notified by the system. Each of the persons assigned, again the engineering manager has distributed the discipline specific requirements internally.

The project manager watches the progress of the requirements analyses which are around a virtual model. There is still a need for meetings with the stakeholders to agree on the solution approach. Everything is stored and visible online in the system. This visibility has helped some of the stakeholders to be better-aligned upfront. In the end, the response is generated and converted to the customer’s format.

Not much benefit for step 1

imageIf you compare the two approaches, there is mainly one person happy: the project manager. Instead of spending time to collect the status of all information, direct visibility on the response helps him/her to prioritize of focus where attention is needed, instead of discovering it on a weekly base.

There is some small benefit from the virtual model as other stakeholders can have a better understanding of the actual progress.

However, for the rest, all stakeholders are complaining. It is difficult to work. (Fl)Excel was much easier. Moreover, thinking about a virtual model takes time as we are not used working in this way. Typically something for aerospace you might think.

And now the benefits come – step 2

The customer has placed the order, and the project has started. The design has started, and people start to discover discrepancies or ambiguous demands that need to be negotiated with the customer. Is it part of the project and if not, should it become part of the project and at which costs (for whom)

The document oriented approach – 2

searchSeveral engineers are now discussing with the counterparts at the customer the detailed interpretation of the requirements, either through face-to-face meetings or emails. Changes are collected and sent to the project manager, who tries to understand what has changed and how to merge it in an on-going specification document. To avoid many revisions, he/she tries to update the document on a bi-weekly base, send it to the internal stakeholders for review and with their feedback generates a specification document for the customer that supposed to cover the latest agreements.

Unfortunate not all changes have reached the document as some of the stakeholders were busy and forgot to include some of the changes agreed with the customer as they were in a lost email. Also, a previous change of a requirement was overwritten as an update from quality used the old data. Finally, some design solutions were changed, which raised the costs. And not sure if the product with all its changes will be compliant after delivery. However, luckily nobody noticed so far, not even the customer

The data-oriented approach – 2

whyworryThanks to the virtual model and the relations between all the requirements, any change in a requirement gets notified in the system. When a requirement is further clarified, it is updated in the system. When a requirement needs to be changed, it is clear what the impact of this change is. A change workflow assures that decisions are made visible and approved. Potentially changes that lead to more work were quoted to the customer for acceptance. Luckily the compliancy engineer noted that the change of materials used would lead to a compliancy issue. On a bi-weekly base, the project manager generates an agreed specification for the customer based on the data in the system.

Benefits are growing.

imageThe project manager remains the happiest person and is even happier as less discussion is needed about who changed what and why. Alternatively, discussions about changes that should exist and cannot be found. The time saved by the project manager could be used to collaborate even better with the teams (without annoying them) or perhaps a second project to manage in parallel.

Other stakeholders start to enjoy the data-oriented approach too. Less ambiguity on their side too, fewer iterations because changes were not apparent. As all information is related to the virtual model online, the actual status is clear when making a decision. Less fixing afterwards and luckily still project meeting between the stakeholders to synchronize. The PLM system does not eliminate communication; it provides a reliable baseline of the truth. No need (and option) to look in your archives.

At this stage, benefits start to become clear. Fewer iterations and better decisions will have an impact on the costs and project stress. Still a complaint from the engineers might be that they need to do too much upfront thinking although some years later they might discover that this will be their main job. Fixing issues from the past have diminished.

And then the ultimate benefits

Now the project has reached the physical state. It is manufactured or under commissioning.

The document oriented approach – 3

No_roiIn the document oriented approach, many issues might pop-up because they have not been considered in the early phase, or they got lost during document exchanges. Does the product work as specified? Is the building certified as specified?

The customer is king and for manufacturing companies this might lead to product recalls or launch delays. In the construction world, people in the field, will fix the issues by using skilled resources and creating a waste of materials and/or resources.

Data handover to the owner is a nightmare for a the project-centric delivery. Several people have been searching for documents, specifications and emails to build and compile the required documents for handover.

The data-oriented approach – 3

sel_aIn the data-centric approach, the behavior of the physical product works as expected as most of the issues have been solved in the virtual model. When testing the product it works as specified as the specifying requirements have always been linked to the product. Moreover, they have been agreed and approved by the relevant stakeholders. Where relevant, the customer has paid for the extra work specified.

The handover process was not so stressful as before with the document-oriented approach. As the required information was known and specified upfront related to the requirements, the maturity process of the virtual model assured this data exists in the system. Now the as-built information matches the as-specified information. What a relief.

Conclusion

imageIt is clear that the significant benefits can be found in step 3. I wrote the comparison in an extreme manner, knowing that reality lies in the middle. Excellent people can comprehend and fix more upfront because of their experience. Building the ultimate virtual model is not yet an easy achievement either.

The savings in materials and required resources are significant in a data-oriented approach. The time savings and the quality enhancements might change your company into a market-leader. The cost savings achieved through a pro-active approach will make your margin growing (unless competition does the same) and enable you to innovate.

One final remark on business change

ideaIf business change could be achieved by selecting the right tool or system without a business change, you will never get a competitive advantage. As your competitor can buy these too.

However, if you change to a data-centric approach, it will be a though change process and therefore once implemented you will leave competitors behind that keep on hanging on the past.

data futureMy holidays are over. After reading and cycling a lot, it is time to focus again on business and future. Those of you who have followed my blog the past year must have noticed that I have been talking on a regular base about business moving to a data-oriented approach instead of a document / file-based approach. I wrote an introduction to this topic at the beginning of this year: Did you notice PLM has been changing?

It is part of a bigger picture, which some people might call the Second Machine Age, Industry 4.0, The Third Wave or even more disturbing The onrushing wave.

This year I have had many discussions around this topic with companies acting in various industries; manufacturing, construction, oil & gas, nuclear and general EPC-driven companies. There was some commonality in all these discussions:

  • PLUS: Everyone believes it is a beautiful story and it makes sense
  • MINUS: Almost nobody wants to act upon it as it is an enormous business change and to change the way a company works you need C-level understanding
  • PLUS: Everyone thinks the concept is clear to them
  • MINUS: Few understand what it means to work data-oriented and what the impact on their business would be

Therefore, what I will try to do in the upcoming blog posts (two-three-four ??) is to address the two negative observations and how to make them more precise.

What is data / information / knowledge?

dataData for me is a collection of small artifacts (numbers, characters, lines, sound bits, …) which have no meaning at all. This could be bundled together as a book, a paper drawing, a letter but also bundled together as a digital format like an eBook, a CAD file, an email and even transmission bytes of a network / internet provider can be considered as data.

informationData becomes significant once provided in the context of each other or in the context of other data. At that time, we start calling it information. For that reason, a book or a drawing provides information as the data has been structured in such a manner to become meaningful. The data sent through the network cable only becomes information when it is filtered and stripped from the irrelevant parts.

knowledgeInformation is used to make decisions based on knowledge. Knowledge is the interpretation of information, which combined in a particular way, helps us to make decisions. And the more decisions we make and the more information we have about the results of these decisions, either by us or other, it will increase our knowledge.

 

Data and big data

Now we have some feeling about data, information and knowledge. For academics, there is room to discuss and enhance the definition. I will leave it by this simple definition.

Big data is the term for all digital data that is too large to handle in a single data management system, but available and searchable through various technologies. Data can come from any source around the world as through the internet an infrastructure exists to filter and search for particular data.

big dataBy analyzing and connecting the data coming from these various sources, you can generate information (placing the data in context) and build knowledge. As it is an IT-driven activity, this can be done in the background and give almost actual data to any person. This is a big difference with information handling in the old way, where people have to collect and connect manual the data.

The power of big data applies to many business areas. If you know how your customers are thinking and associating their needs to your products, you can make them better and more targeted to your potential market. Or, if you know how your products are behaving in the field during operation (Internet of Things) you can provide additional services, instant feedback and be more proactive. Plus the field data once analyzed provide actual knowledge helping you to make better products or offer more accurate services.

Wasn’t there big data before?

clip_image012Yes, before the big data era there was also a lot of information available. This information could be stored in “analogue” formats ( microfiche, paper, clay tablets, papyrus) or in digital formats, better known as files or collections of files (doc, pdf, CAD-files, ZIP….).

Note the difference. Here I am speaking about information as the data is contained in these formats.

You have to open or be in front of information container first, before seeing the data. In the digital world, this is often called document management, content management. The challenge of these information containers is that you need to change the whole container version once you modify one single piece of data inside it. And each information container holds duplicated information from a data element. Therefore, it is hard to manage a “single version of the truth” approach.

And here comes the data-oriented approach

The future is about storing all these pieces of data inside connected data environments, instead of storing a lot of data inside a (versioned) information container (a file / a document).

Managing these data elements in the context of each other allow people to build information from any viewpoint – project oriented, product oriented, manufacturing oriented, service oriented, etc.

The data remains unique, therefore supporting much closer the single version of the truth approach. Personally I consider the single version of the truth as a utopia, however reducing the amount of duplicated data by having a data-oriented approach will bring a lot more efficiency.

In my next post, I will describe an example of a data-oriented approach and how it impacts business, both from the efficiency point of view and from the business transformation point of view. As the data-oriented approach can have immense benefits . However, they do not come easy. You will have to work different.

Some more details

An important point to discuss is that this data-oriented approach requires a dictionary, describing the primary data elements used in a certain industry. The example below demonstrates a high-level scheme for a plant engineering environment.

plant data model

Data standards exist in almost any industry or they are emerging and crucial for the longevity and usage of the data. I will touch it briefly in one of the upcoming posts, however, for those interested in this topic in relation to PLM, I recommend attending the upcoming PDT Europe. If you look at the agenda there is a place to learn and discuss a lot about the future of PLM.

pdteurope

I hope to see you there.

imageLast week I attended the PI Apparel conference in London. It was the second time this event was organized and approximate 100 participants were there for two full days of presentations and arranged network meetings. Last year I was extremely excited about this event as the different audience, compare to classical PLM events, and was much more business focused.

Read my review from last year here: The weekend after PI Apparel 2013

This year I had the feeling that the audience was somewhat smaller, missing some of the US representatives and perhaps there was a slightly more, visible influence from the sponsoring vendors. Still an enjoyable event and hopefully next year when this event will be hosted in New York, it will be as active as last year.

Here are some of my observations.

Again the event had several tracks in parallel beside the keynotes, and I look forward in the upcoming month to see the sessions I could not attend. Obvious where possible I followed the PLM focused sessions.


clip_image002First keynote came from Micaela le Divelec Lemmi, Executive Vice President and Chief Corporate Operations Officer of Gucci. She talked us through the areas she is supervising and gave some great insights. She talked about how Gucci addresses sustainability through risk and cost control. Which raw materials to use, how to ensure the brands reputation is not at risk, price volatility and the war on talent. As Gucci is a brand in the high-end price segment, image and reputation are critical, and they have the margins to assure it is managed. Micaela spoke about the short-term financial goals that a company as Gucci has related to their investors. Topics she mentioned (I did not write them down as I was tweeting when I heard them) were certainly worthwhile to consider and discuss in detail with a PLM consultant.

clip_image003

Micaela further described Gucci´s cooperate social responsibility program with a focus on taking care of the people, environment and culture. Good to learn that human working conditions and rights are a priority even for their supply chain. Although it might be noted that 75 % of Gucci´s supply chain is in Italy. One of the few brands that still has the “Made in Italy” label.

My conclusion was that Micaela did an excellent PR job for Gucci, which you would expect for a brand with such a reputation. Later during the conference we had a discussion would other brands with less exclusivity and more operating in the mass consumer domain be able to come even close to such programs?


clip_image005Next Göktug and Hakan gave us their insights deploying their first PLM system at the AYDINLI group.

The company is successful in manufacturing and selling licensed products from Pierre Cardin, Cacharel and US Polo Association mainly outside the US and Western Europe.

Their primary focus was to provide access to the most accurate and most updated information from one source. In parallel, standardization of codes and tech packs was a driver. Through standardization quality and (re)use could be improved, and people would better understand the details. Additional goals are typical PLM goals: following the product development stages during the timeline, notify relevant users about changes in the design, work on libraries and reuse and integrate with SAP.

Interesting Hakan mentioned that in their case SAP did not recommend to use their system for the PLM related part due to lack of knowledge of the apparel industry. A wise decision which would need followup for other industries.

In general the PLM implementation described by Göktug and Hakan was well phased and with a top-down push to secure there is no escape to making the change. As of all PLM implementations in apparel they went live in their first phase rather fast as the complex CAD integrations from classical PLM implementations were not needed here.


Next I attended the Infor session with the title: Work the Way you Live: PLM built for the User. A smooth marketing session with a function / feature demo demonstrating the flexibility and configuration capabilities of the interface. Ease of use is crucial in the apparel industry, where Excel is still the biggest competitor. Excel might satisfy the needs from the individual, it lacks the integration and collaboration aspect a PLM system can offer.


clip_image007More interesting was the next session that I attended from Marcel Oosthuis, who was responsible as Process Re-Engineering Director (read PLM leader). Marcel described how they had implemented PLM at Tommy Hilfiger, and it was an excellent story (perhaps too good to be true).

I believe larger companies with the right focus and investment in PLM resources can achieve this kind of results. The target for Tommy Hilfiger´s PLM implementation was beyond 1000 users, therefore, a serious implementation.

Upfront the team defined first what the expected from the PLM system to select (excellent !!). As the fashion industry is fast, demanding and changing all the time, the PLM system needs to be Swift, Flexible and Prepared for Change. This was not a classical PLM requirement.

In addition, they were looking for a high-configurable system, providing best practices and a vendor with a roadmap they could influence. Here I got a little more worried as high-configurable and best practices not always match the prepared for change approach. A company might be tempted to automate the way they should work based on the past (best practices from the past)

It was good to hear that Marcel did not have to go into the classical ROI approach for the system. His statement, which I fully endorse that it is about the capability to implement new and better processes. They are often not comparable with the past (and nobody measured the past)

Marcel described how the PLM team (eight people + three external from the PLM vendor) made sure that the implementation was done with the involvement of the end users. End user adoption was crucial as also key user involvement when building and configuring the system.

It was one of the few PLM stories where I hear how all levels of the organization were connected and involved.


imageNext Sue Butler, director from Kurt Salmon, described how to maximize ROI from your PLM investment. It is clear that many PLM consultants are aligned, and Sue brought up all the relevant points and angles you needed to look at for successful PLM implementation.

Main points: PLM is about changing the organization and processes, not about implementing a tool. She made a point that piloting the software is necessary as part of the learning and validation process. I agree on that under the condition that it is an agile pilot which does not take months to define and perform. In that case, you might be already locked in into the tool vision too much – focus on the new processes you want to achieve.

Moreover, because Sue was talking about maximize ROI from a PLM implementation, the topics focus on business areas that support evolving business processes and measure (make sure you have performance metrics) came up.


imageThe next session Staying Ahead of the Curve through PLM Roadmap Reinvention conducted by Austin Mallis, VP Operations, Fashion Avenue Sweater Knits, beautifully completed previous sessions related to PLM.

Austin nicely talked about setting the right expectations for the future (There is no perfect solution / Success does not mean stop / Keeping the PLM vision / No True End). In addition, he described the human side of the implementation. How to on-board everyone (if possible) and admitting you cannot get everyone on-board for the new way of working.


imageNext in row was my presentation with potential the longest title: “How to transform your Business to ensure you Benefit from the Value PLM can deliver”.

Luckily the speakers before me that day already addressed many of the relevant topics, and I could focus on three main thoughts completing the story:

1. Who decides on PLM and Why?

I published the results from a small survey I did a month ago via my blog (A quick PLM survey). See the main results below.

clip_image010

It was interesting to observe that both the management and the users in the field are the majority demanding for PLM. Consultants have some influence and PLM vendors even less. The big challenge for a company is that the management and consultants often talk about PLM from a strategic point of view, where the PLM vendor and the users in the field are more focused on the tool(s).

From the expectations you can see the majority of PLM implementations is about improving collaboration, next time to market, increase quality and centralizing and managing all related information.

2. Sharing data instead of owning data

(You might have read about it several times in my blog) and the trend that we move to platforms with connected data instead of file repositories. This should have an impact on your future PLM decisions.

3. Choosing the right people

The third and final thought was about choosing the right people and understanding the blocker. I elaborated on that topic already before in my recent blog post: PLM and Blockers

My conclusions for the day were:

A successful PLM implementation requires a connection in communication and explanation between all these levels. These to get a company aligned and have an anchored vision before even starting to implement a system (with the best partner)


imageThe day was closed by the final keynote of the day from Lauren Bowker heading T H E U N S E E N. She and her team are exploring the combinations of chemistry and materials to create new fashion artifacts. Clothes and materials that change color based on air vent, air pollution or brain patterns. New and inspiring directions for the fashion lovers.

Have a look here: http://seetheunseen.co.uk/


The morning started with Suzanne Lee, heading BioCouture who is working on various innovative methodologies to create materials for the apparel industry by using all kind of live micro-organisms like bacteria, fungi and algae and using materials like cellulose, chitin and protein fibers, which all can provide new possibilities for sustainability, comfort, design, etc. Suzanne´s research is about exploring these directions perhaps shaping some new trends in the 5 – 10 years future ahead. Have a look into the future here:


clip_image012Renate Eder took us into the journey of visualization within Adidas, with her session: Utilizing Virtualization to Create and Sell Products in a Sustainable Manner.

It was interesting to learn that ten years ago she started the process of having more 3D models in the sales catalogue. Where classical manufacturing companies nowadays start from a 3D design, here at Adidas at the end of the sales cycle 3D starts. Logical if you see the importance and value 3D can have for mass market products.

Adidas was able to get 16000 in their 3D catalogue thanks to the work from 60 of their key suppliers who were fully integrated in the catalogue process. The benefit from this 3D catalogue was that their customers, often the large stores, need lesser samples, and the savings are significant here (plus a digital process instead of transferring goods).

Interesting discussion during the Q&A part was that the virtual product might even look more perfect than the real product, demonstrating how lifelike virtual products can be.

And now Adidas is working further backwards from production patterns (using 3D) till at the end 3D design. Although a virtual 3D product cannot 100 % replace the fit and material feeling, Renate believes that also introducing 3D during design can reduce the work done during pilots.


Finally for those who stayed till the end there was something entirely different. Di Mainstone elaborating on her project: Merging Architecture & the Body in Transforming the Brooklyn Bridge into a Playable Harp. If you want something entirely different, watch here:

Conclusion

The apparel industry remains an exciting industry to follow. For some of the concepts – being data-centric, insane flexible, continuous change and rapid time to market are crucial here.

This might lead development of PLM vendors for the future, including using it based on cloud technology.

From the other side, the PLM markets in apparel is still very basic and learning, see this card that I picked up from one of the vendors. Focus on features and functions, not touching the value (yet)

clip_image014

image

Friends, this is the first evening that there is no soccer on television for two weeks. So I have time to write something.

Currently, I am preparing my session for PI Apparel 2014 in London on 15/16 July. Last year´s PI Apparel was a discovery for me as the audience was so different compared to classical PLM conferences. Of course, the products might be not as complex, but the time to market needs and, therefore, the need to work as fast and concurrent as possible is a huge differentiator. See my post from last year´s conference here: The weekend after PI Apparel 2013

In a way, PLM for apparel companies is more data-centric than some of the original industries where PLM was born. In the traditional way, file management and document sharing were the initial drivers.

At this years conference, I will talk about the needed change in the way people work that comes with a PLM implementation.  I will share the full story plus my observations in my next post end of July.

Before that, I have a question to all readers of this blog who are working for a company that has implemented or starts implementing PLM to answer two questions from the survey below. The answers will help me to confirm my prejudgments or change my mind.

So if you have some time between the soccer matches, please respond to the survey below if you qualify:

 

Click here to answer a quick survey before PI Apparel

 

Thanks and enjoy the upcoming matches

image

image

Two weeks ago I attended the Nobletek PLM forum in Belgium, where a group of experts, managers and users discussed topics related to my favorite theme: “Is PLM changing? “

Dick Terleth (ADSE) lead a discussion with title “PLM and Configuration Management as a proper profession” or "How can the little man grow?". The context of the discussion was related to the topic: “How is it possible that the benefits of PLM (and Configuration Management) are not understood at C-level?” or with other words: “Why is the value for Configuration Management and PLM not obvious?”.

In my previous post, PLM is doomed unless …., I quoted Ed Lopategui (www.eng-eng.com), who commented that being a PLM champion (or a Configuration Management expert as Dick Terleth would add) is bad for your career. Dick Terleth asked the same question, showing pictures of the self-assured accountant and the Configuration Management or PLM professional. (Thanks Dick for the pictures). Which job would you prefer?

image

The PLM ROI discussion

No_roiA first attempt to understand the difference could be related to the ROI discussion, which seems to be only applicable for PLM. Apparently ERP and financial management systems are a must for companies. No ROI discussion here. Persons who can control/report the numbers seem to have the company under control. For the CEO and CFO the value of PLM is often unclear. And to make it worse, PLM vendors and implementers are fighting for their unique definition of PLM so we cannot blame companies to be confused. This makes it clear that if you haven´t invested significant time to understand PLM, it will be hard to see the big picture. And at C-level people do not invest significant time to understand the topic. It is the C-level´s education, background or work experience that make him/her decide.

So if the C-level is not educated on PLM, somebody has to sell the value to them. Oleg Shilovitsky wrote about it recently in his post Why is it hard to sell PLM ROI and another respected blogger, Joe Barkai, sees the sun come up behind the cloud, in his latest post PLM Service Providers Ready To Deliver Greater Value. If you follow the posts of independent PLM bloggers (although who is 100 % independent), you will see a common understanding that implementing PLM currently requires a business transformation as old processes were not designed for a modern infrastructure and digital capabilities.

PLM is about (changing) business processes

imageBack to the Nobletek PLM forum. Douglas Noordhoorn, the moderator of the forum challenged the audience stating that PLM has always been there (or not there – if you haven´t discovered it). It is all about managing the product development processes in a secure way. Not talking about “Best Practices” but “Good practices." Those who had a proper education in the aerospace industry learned that good processes are crucial to deliver planes that can fly and are reliable.

Of course, the aerospace industry is not the same as other industries. However, more and more other industries in my network, like Nuclear new build, the construction industry or other Engineering, Procurement and Construction companies want to learn from aerospace and automotive good practices. They realize they are losing market share due to the fact that the cost of failure combined with relative high labor costs makes them too expensive. But from where to they get their proper good practices education?

The PLM professional?

myplmAnd this was an interesting point coming up from the Nobletek forum. There is no proper, product agnostic education for PLM (anymore). If you study logistics, you will learn a lot about various processes and how they can be optimized for a certain scenario. When you study engineering, there is a lot of focus on engineering disciplines and methods. But there is no time to educate engineers in-depth to understand the whole product development process and how to control it. Sometimes I give a guest lecture to engineering classes. It is never an important part of the education.

To become a PLM professional

imageFor those who never had any education in standard engineering processes, there is Frank Watts Engineering control book, which probably would be a good base. But it is not the PLM professional only that should be aware, of the good practices. Moreover, all companies manufacturing products, plants or buildings should learn these basics. As a side step, it would make a discussion around BIM more clear. At this time, manufacturing companies are every time discovering their good practices in the hard way.

And when this education exists, companies will be aware that it is not only about the tools, but it is the way the information is flowing through the organization. Even there is a chance that somewhere at C-level someone has been educated and understands the value. For ERP everyone agrees. For PLM, it remains a labyrinth of processes designed by companies learning on the job currently. Vendors and implementers pushing what they have learned. Engineering is often considered as a hard-to-manage discipline. As a SAP country manager once said to me: “Engineers are actually resources that do not want be managed, but we will get them …..”

And then the future ……

PLM bookI support the demand for a better education in engineering processes especially for industries outside aerospace or automotive. I doubt if it will have a significant impact although it might create the visibility and understanding for PLM at C-level. No need anymore for the lone ranger who fights for PLM. Companies will have better educated people that understand the need for good practices that exist. These good practices will be the base for companies when discussing with PLM vendors and implementers. Instead of vendors and implementers pushing their vision, you can articulate, and follow your vision.

However, we need a new standard book too. We are currently in the middle of a big change. Thanks to modern technology and connectivity the world is changing. I wrote and spoke about it in: Did you notice PLM is changing?

doc2dataThis is a change of generations and concepts which have not been foreseen by Frank Watts and others. What will be the new standard for data-centric companies instead of document based control?

The digital revolution is here (Industry 4.0), and here (digital revolution), and here (the third industrial revolution).

 

This awareness needs to become visible at C-level.
Who will educate them ??

 

spain_nl

Now back to soccer – 4 years ago Spain-The Netherlands was the last match – the final. Now it is the first match for them – will the Dutch change the game ?

Follow

Get every new post delivered to your Inbox.

Join 363 other followers

%d bloggers like this: