You are currently browsing the category archive for the ‘Digital enterprise’ category.

After the series about “Learning from the past,” it is time to start looking towards the future.  I learned from several discussions that I am probably working most of the time with advanced companies. I believe this would motivate companies that lag behind even to look into the future even more.

If you look into the future for your company, you need new or better business outcomes. That should be the driver for your company. A company does not need PLM or a Digital Twin. A company might want to reduce its time to market, improve collaboration between all stakeholders. These objectives can be realized by different ways of working and an IT-infrastructure to allow these processes to become digital and connected.

That is the “game”. Coming back to the future of PLM.  We do not need a discussion about definitions; I leave this to the academics and vendors. We will see the same applies to the concept of a Digital Twin.

My statement: The digital twin is not new. Everybody can have their own digital twin as long as you interpret the definition differently. Does this sound like the PLM definition?

The definition

I like to follow the Gartner definition:

A digital twin is a digital representation of a real-world entity or system. The implementation of a digital twin is an encapsulated software object or model that mirrors a unique physical object, process, organization, person, or other abstraction. Data from multiple digital twins can be aggregated for a composite view across a number of real-world entities, such as a power plant or a city, and their related processes.

As you see, not a narrow definition. Now we will look at the different types of interpretations.

Single-purpose siloed Digital Twins

  1. Simple – data only

One of the most straightforward applications of a digital twin is, for example, my Garmin Connect environment. When cycling, my device registers performance parameters (speed, cadence, power, heartbeat, location). After every trip, I can analyze my performance. I can see changes in my overall performance; compare my performance with others in my category (weight, age, sex).

Based on that, I can decide if I want to improve my performance. My personal business goal is to maintain and improve my overall performance, knowing I cannot stop aging by upgrading my body.

On November 4th, 2020, I am participating in the (almost virtual) Digital Twin conference organized by Bits&Chips in the Netherlands. In the context of human performance, I look forward to Natal van Riel’s presentation: Towards the metabolic digital twin – for sure, this direction is not simple. Natal is a full professor at the Technical University in Eindhoven, the “smart city” in the Netherlands

  1. Medium – data and operating models

Many connected devices in the world use the same principle. An airplane engine, an industrial robot, a wind turbine, a medical device, and a train carriage; all track the performance based on this connection between physical and virtual, based on some sort of digital connectivity.

The business case here is also monitoring performance, predict maintenance, and upgrade the product when needed.

This is the domain of Asset Lifecycle Management, a practice that exists for decades. Based on financial and performance models, the optimal balance between maintaining and overhaul has to be found. Repairs are disruptive and can be extremely costly. A manufacturing site that cannot produce can costs millions per day. Connecting data between the physical and the virtual model allows us to have real-time insights and be proactive. It becomes a digital twin.

  1. Advanced – data and connected 3D model

The ditial twin we see the most in marketing videos is a virtual twin, using a 3D-representation for understanding and navigation.  The 3D-representation provides a Virtual Reality (VR) environment with connected data. When pointing at the virtual components, information might appear, or some animation takes place.

Building such a virtual representation is a significant effort; therefore, there needs to be a serious business case.

The simplest business case is to use the virtual twin for training purposes. A flight simulator provides a virtual environment and behavior as-if you are flying in the physical airplane – the behavior model behind the simulator should match as good as possible the real behavior. However, as it is a model, it will never be 100 % reality and requires updates when new findings or product changes appear.

A virtual model of a platform or plant can be used for training on Standard Operating Procedures (SOPs). In the physical world, there is no place or time to conduct such training. Here the complexity might be lower. There is a 3D Model; however, serious updates can only be expected after a major maintenance or overhaul activity.

These practices are not new either and are used in places where the physical training cannot be done.

More challenging is the Augmented Reality (AR) use case. Here the virtual model, most of the time, a lightweight 3D Model, connects to real-time data coming from other sources. For example, AR can be used when an engineer has to service a machine. The AR-environment might project actual data from the machine, indicate service points and service procedures.

The positive side of the business case is clear for such an opportunity, ensuring service engineers always work with the right information in a real-time context. The main obstacle for implementing AR, in reality, is the access to data, the presentation of the data and keeping the data in the AR-environment matching the reality.

And although there are 3D Models in use, they are, to my knowledge, always created in siloes, not yet connected to their design sources.Have a look at the Digital Twin conference from Bits&Chips, as mentioned before.

Several of the cases mentioned above will be discussed here. The conference’s target is to share real cases concluded by Q & A sessions, crucial for a virtual event.

Connected Virtual Twins along the product lifecycle

So far, we have been discussing the virtual twin concept, where we connect a product/system/person in the physical world to a virtual model. Now let us zoom in on the virtual twins relevant for the early parts of the product lifecycle, the manufacturing twin, and the development twin. This image from Siemens illustrates the concept:

On slides they imagine a complete integrated framework, which is the future vision. Let us first zoom in on the individual connected twins.

The digital production twin

This is the area of virtual manufacturing and creating a virtual model of the manufacturing plant. Virtual manufacturing planning is not a new topic. DELMIA (Dassault Systèmes) and Tecnomatix (Siemens) are already for a long time offering virtual manufacturing planning solutions.

At that time, the business case was based on the fact that the definition of a manufacturing plant and process done virtually allows you to optimize the plant before investing in physical assets.

Saving money as there is no costly prototype phase to optimize production. In a virtual world, you can perform many trade-off studies without extra costs. That was the past (and for many companies still the current situation).

With the need to be more flexible in manufacturing to address individual customer orders without increasing the overhead of delivering these customer-specific solutions, there is a need for a configurable plant that can produce these individual products (batch size 1).

This is where the virtual plant model comes into the picture again. Instead of having a virtual model to define the ultimate physical plant, now the virtual model remains an active model to propose and configure the production process for each of these individual products in the physical plant.

This is partly what Industry 4.0 is about. Using a model-based approach to configure the plant and its assets in a connected manner. The digital production twin drives the execution of the physical plant. The factory has to change from a static factory to a dynamic “smart” factory.

In the domain of Industry 4.0, companies are reporting progress. However, to my experience, the main challenge is still that the product source data is not yet built in a model-based, configurable manner. Therefore, requiring manual rework. This is the area of Model-Based Definition, and I have been writing about this aspect several times. Latest post: Model-Based: Connecting Engineering and Manufacturing

The business case for this type of digital twin, of course, is to be able to customer-specific products with extremely competitive speed and reduced cost compared to standard. It could be your company’s survival strategy. As it is hard to predict the future, as we see from COVID-19, it is still crucial to anticipate the future, instead of waiting.

The digital development twin

Before a product gets manufactured, there is a product development process. In the past, this was pure mechanical with some electronic components. Nowadays, many companies are actually manufacturing systems as the software controlling the product plays a significant role. In this context, the model-based systems engineering approach is the upcoming approach to defining and testing a system virtually before committing to the physical world.

Model-Based Systems Engineering can define a single complex product and perform all kinds of analysis on the system even before there is a physical system in place.  I will explain more about model-based systems engineering in future posts. In this context, I want to stress that having a model-based system engineering environment combined with modularity (do not confuse it with model-based) is a solid foundation for dealing with unique custom products. Solutions can be configured and validated against their requirements already during the engineering phase.

The business case for the digital development twin is easy to make. Shorter time to market, improved and validated quality, and reduced engineering hours and costs compared to traditional ways of working. To achieve these results,  for sure, you need to change your ways of working and the tools you are using. So it won’t be that easy!

For those interested in Industry 4.0 and the Model-Based System Engineering approach, join me at the upcoming PLM Road Map 2020 and PDT 2020 conference on 17-18-19 November. As you can see from the agenda, a lot of attention to the Digital Twin and Model-Based approaches.

Three digital half-days with hopefully a lot to learn and stay with our feet on the ground.  In particular, I am looking forward to Marc Halpern’s keynote speech: Digital Thread: Be Careful What you Wish For, It Just Might Come True

Conclusion

It has been very noisy on the internet related to product features and technologies, probably due to COVIC-19 and therefore disrupted interactions between all of us – vendors, implementers and companies trying to adjust their future. The Digital Twin concept is an excellent framing for a concept that everyone can relate to. Choose your business case and then look for the best matching twin.

I believe we are almost at the end of learning from the past. We have seen how, from an initial serial CAD-driven approach with PDM, we evolved to PLM-managed structures, the EBOM and the MBOM. Or to illustrate this statement, look at the image below, where I use a Tech-Clarity image from Jim Brown.

The image on the right describes perfectly the complementary roles of PLM and ERP. The image on the left shows the typical PDM-approach. PDM feeding ERP in a linear process. The image on the right, I believe it is from 2004, shows the best practice before digital transformation. PLM is supporting product innovation in an iterative approach, pushing released information to ERP for execution.

As I think in images, I like the concept of a circle for PLM and an arrow for ERP. I am always using those two images in discussions with my customers when we want to understand if a particular activity should be in the PLM or ERP-domain.

Ten years ago, the PLM-domain was conceptually further extended by introducing support for products in operations and service. Similar to the EBOM (engineering) and the MBOM (manufacturing), the SBOM (service) was introduced to support product information for products in operation. In theory a full connected cicle.

Asset Lifecycle Management

At the same time, I was promoting PLM-practices for owners/operators to enhance Asset Lifecycle Management. My first post from June 2010 was called: PLM for Asset Lifecycle Management and Asset Development introduces this approach.

Conceptually the SBOM and Asset Lifecycle Management have a lot in common. There is a design product, in this case, an asset (plant, machine) running in the field, and we need to make sure operators have the latest information about the asset. And in case of asset changes, which can be a maintenance operation, a repair or complete overall, we need to be sure the changes are based on the correct information from the as-built environment. This requires full configuration management.

Asset changes can be based on extensive projects that need to be treated like new product development projects, with a staged approach that can take weeks, months, sometimes years. These activities are typical activities performed in PLM-systems, not in MRO-systems that are designed to manage the actual operation. Again here we see the complementary roles of PLM (iterative) and MRO (execution).

Since 2008, I have worked a lot in this environment, mainly in the nuclear and process industry. If you want to learn more about this aspect of PLM, I recommend looking at the PLMpartner website, where Bjørn Fidjeland, in cooperation with SharePLM, published a course on Plant Information Management. We worked together in several projects and Bjørn has done a great effort to describe the logical model to be used instead of a function-feature story.

Ten years ago, we were not calling this concept the “Digital Twin,” as the aim was to provide end-to-end support of asset information from engineering, procurement, and construction towards operation in a coordinated manner. The breaking point in the relation between the EPCs and Owner/Operators is the data-handover – how much of your IP can/do you expose and what is needed. Nowadays, we would call striving for end-to-end data continuity the Digital Thread.

Hot from the press in this context, CIMdata just published a commentary Managing the Digital Thread in Global Value Chains describing Eurostep’s ShareAspace capabilities and experiences in managing an end-to-end information flow (Digital Thread) in a heterogeneous environment based on exchange standards like ISO 10303-239 PLCS.  Their solution is based on what I consider a more modern approach for managing digital continuity compared to the traditional approach I described before. Compare the two images in this paragraph. The first image represents the old/current way with a disconnected handover, the second represents ShareAspace connected approach based on a real digital thread.

The Service BOM

As discussed with Asset Lifecycle Management, there is a disconnect between the engineering disciplines and operations in the field, looking from the point of view of an Asset owner/operator.

Now when we look from the perspective of a manufacturing company that produces assets to be serviced, we can identify a different dataflow and a new structure, the Service BOM (SBOM).

The SBOM provides information on how a product needs to be serviced. What are the parts that require service, and what are the service kits that are possible for that product? For that reason, service engineering should be done in parallel to product engineering. When designing a product, the engineer needs to identify which the wearing parts (always require service in time) and which parts might be serviceable.

There are different ways to look at the SBOM. Conceptually, the SBOM could be created in close relation with the EBOM. At the moment you define your product, you also should specify how the product will be services. See the image below

From this example, it is clear that part standardization and modularization have a considerable benefit for services downstream. What if you have only one serviceable part that applies to many products? The number of parts to have in stock will be strongly reduced instead of having many similar parts that only fit in a single product?

Depending on the type of product, the SBOM can be generic, serving many products in the field. In that case, the company has to deal with catalogs, to be defined in PLM. Or the SBOM can be aligned with the As-Built of a capital product in the field. In that case, the concepts of Asset Lifecycle Management apply. Click on the image to see a clear picture.

The SBOM on its own,  in such an environment, will have links to specific documents, service instructions, operating manuals.

If your PLM-system allows it, extending the EBOM and MBOM with an SBOM is not a complex effort. What is crucial to understand is that the SBOM has its own lifecycle, which can even last longer than the active product sold. So sometimes, manufacturing specifications, related to service parts need to be maintained too, creating a link between the SBOM and potential MBOM(s).

ECM = Enterprise Change Management

When I discussed ECM in my previous post in the context of Engineering Change Management, I got the feedback that nowadays, everyone talks about Enterprise Change Management. Engineering Change Management is old school.

In the past, and even in a 2014 benchmark, a customer had two change management systems. One in PLM and one in ERP, and companies were looking into connecting these two processes. Like the BOM-interaction between PLM and ERP, this is technology-wise, never a real problem.

The real problem in such situations was to come to a logical flow of events. Many times the company insisted that every change should start from the ERP-system as we like to standardize. This means that even an engineering change had to be registered first in the ERP-system

Luckily the reach of PLM has grown. PLM is no longer the engineering tool (IT-system thinking). PLM has become the information backbone for product information all along the product lifecycle. Having the MBOM and SBOM available through a PLM-infrastructure allows organizations to streamline their processes.

Aras – digital thread through connected structures

And in this modern environment, enterprise change management might take place mostly in a PLM-infrastructure. The PLM-infrastructure providing a digital thread, as the Aras picture above illustrates, provides the full traceability to support configuration management.

However, we still have to remember that configuration management and engineering change management, first of all, are based on methodology and processes. Next, the combination of tools to be used will vary.

I like to conclude this topic with a quote from Lee Perrin’s comment on my previous blog post

I would add that aerospace companies implemented CM, to avoid fatal consequences to their companies, but also to their flying customers.

PLM provides the framework within which to carry out Configuration Management. CM can indeed be carried out without PLM, as was done in the old paper-based days. As you have stated, PLM makes the whole CM process much more efficient. I think more transparent too.

Conclusion

After nine posts around the theme Learning from the past to understand the future, I walked through the history of CAD, PDM and PLM in a fast mode, pointing to practices and friction points. In the blogging space, it is hard to find this information as most blog posts are coming from software vendors explaining why their tool is needed. Hopefully, these series have helped many of you to understand a broader context. Now I want to focus on the future again in my upcoming blog posts.

Still, feel free to contact me and discuss methodology topics.

Picture by Christi Wijnen – a good friend and photographer in the Netherlands

In the series learning from the past to understand the future, we have almost reached the current state of PLM before digitization became visible. In the last post, I introduced the value of having the MBOM preparation inside a PLM-system, so manufacturing engineering can benefit from early visibility and richer product context when preparing the manufacturing process.

Does everyone need an MBOM?

It is essential to realize that you do not need an EBOM and a separate MBOM in case of an Engineering To Order primary process. The target of ETO is to deliver a unique customer product with no time to lose. Therefore, engineering can design with a manufacturing process in mind.

The need for an MBOM comes when:

  • You are selling a specific product over a more extended period of time. The engineering definition, in that case, needs to be as little as possible dependent on supplier-specific parts.
  • You are delivering your portfolio based on modules. Modules need to be as long as possible stable, therefore independent of where they are manufactured and supplier-specific parts. The better you can define your modules, the more customers you can reach over time.
  • You are having multiple manufacturing locations around the world, allowing you to source locally and manufacture based on local plant-specific resources. I described these options in the previous post

The challenge for all companies that want to move from ETO to BTO/CTO is the fact that they need to change their methodology – building for the future while supporting the past. This is typically something to be analyzed per company on how to deal with the existing legacy and installed base.

Configurable EBOM and MBOM

In some previous posts, I mentioned that it is efficient to have a configurable EBOM. This means that various options and variants are managed in the same EBOM-structure that can be filtered based on configuration parameters (date effectivity/version identifier/time baseline). A configurable EBOM is often called a 150 % EBOM

The MBOM can also be configurable as a manufacturing plant might have almost common manufacturing steps for different product variants. By using the same process and filtered MBOM, you will manufacture the specific product version. In that case, we can talk about a 120 % MBOM

Note: the freedom of configuration in the EBOM is generally higher than the options in the configurable MBOM.

The real business change for EBOM/MBOM

So far, we have discussed the EBOM/MBOM methodology. It is essential to realize this methodology only brings value when the organization will be adapted to benefit from the new possibilities.

One of the recurring errors in PLM implementations is that users of the system get an extended job scope, without giving them the extra time to perform these activities. Meanwhile, other persons downstream might benefit from these activities. However, they will not complain. I realized that already in 2009, I mentioned such a case: Where is my PLM ROI, Mr. Voskuil?

Now let us look at the recommended business changes when implementing an EBOM/MBOM-strategy

  1. Working in a single, shared environment for engineering and manufacturing preparation is the first step to take.

Working in a PLM-system is not a problem for engineers who are used to the complexity of a PDM-system. For manufacturing engineers, a PLM-environment will be completely new. Manufacturing engineers might prepare their bill of process first in Excel and ultimately enter the complete details in their ERP-system. ERP-systems are not known for their user-friendliness. However, their interfaces are often so rigid that it is not difficult to master the process. Excel, on the other side, is extremely flexible but not connected to anything else.

And now, this new PLM-system requires people to work in a more user-friendly environment with limited freedom. This is a significant shift in working methodology. This means manufacturing engineers need to be trained and supported  over several months. Changing habits and keep people motivated takes energy and time. In reality, where is the budget for these activities?  See my 2016 post: PLM and Cultural Change Management – too expensive?

  1. From sequential to concurrent

Once your manufacturing engineers are able to work in a PLM-environment, they are able to start the manufacturing definition before the engineering definition is released. Manufacturing engineers can participate in design reviews having the information in their environment available. They can validate critical manufacturing steps and discuss with engineers potential changes that will reduce the complexity or cost for manufacturing. As these changes will be done before the product is released, the cost of change is much lower. After all, having engineering and manufacturing working partially in parallel will reduce time to market.

Reducing time to market by concurrent engineering

One of the leading business drivers for many companies is introducing products or enhancements to the market. Bringing engineering and manufacturing preparation together also means that the PLM-system can no longer be an engineering tool under the responsibility of the engineering department.

The responsibility for PLM needs to be at a level higher in the organization to ensure well-balanced choices. A higher level in the organization automatically means more attention for business benefits and less attention for functions and features.

From technology to methodology – interface issues?

The whole EBOM/MBOM-discussion often has become a discussion related to a PLM-system and an ERP-system. Next, the discussion diverted to how these two systems could work together, changing the mindset to the complexity of interfaces instead of focusing on the logical flow of information.

In an earlier PI Event in München 2016, I lead a focus group related to the PLM and ERP interaction. The discussion was not about technology, all about focusing on what is the logical flow of information. From initial creation towards formal usage in a product definition (EBOM/MBOM).

What became clear from this workshop and other customer engagements is that people are often locked in their siloed way of thinking. Proposed information flows are based on system capabilities, not on the ideal flow of information. This is often the reason why a PLM/ERP-interface becomes complicated and expensive. System integrators do not want to push for organizational change, they prefer to develop an interface that adheres to the current customer expectations.

SAP has always been promoting that they do not need an interface between engineering and manufacturing as their data management starts from the EBOM. They forgot to mention that they have a difficult time (and almost no intention) to manage the early ideation and design phase. As a Dutch SAP country manager once told me: “Engineers are resources that do not want to be managed.” This remark says all about the mindset of ERP.

After overlooking successful PLM-implementations, I can tell the PLM-ERP interface has never been a technical issue once the methodology is transparent. A company needs to agree on logical data flow from ideation through engineering towards design is the foundation.

It is not about owning data and where to store it in a single system. It is about federated data sets that exist in different systems and that are complementary but connected, requiring data governance and master data management.

The SAP-Siemens partnership

In the context of the previous paragraph, the messaging around the recently announced partnership between SAP and Siemens made me curious. Almost everyone has shared an opinion about the partnership. There is a lot of speculation, and many questions were imaginarily answered by as many blog posts in the field. Last week Stan Przybylinski shared CIMdata’s interpretations in a webinar Putting the SAP-Siemens Partnership In Context, which was, in my opinion, the most in-depth analysis I have seen.

For what it is worth, my analysis:

  • First of all, the partnership is a merger of slide decks at this moment, aiming to show to a potential customer that in the SAP/Siemens-combination, you find everything you need. A merger of slides does not mean everything works together.

  • It is a merger of two different worlds. You can call SAP a real data platform with connected data, where Siemens offering is based on the Teamcenter backbone providing a foundation for a coordinated approach. In the coordinated approach, the data flexibility is lower. For that reason, Mendix is crucial to make Siemens portfolio behave like a connected platform too.
    You can read my doubts about having a coordinated and connected system working together (see image above). It was my #1 identified challenge for this decade: PLM 2020 – PLM the next decade (before COVID-19 became a pandemic and illustrated we need to work connected)
  • The fact that SAP will sell TC PLM and Siemens will sell SAP PPM seems like loser’s statement, meaning our SAP PLM is probably not good enough, or our TC PPM capabilities are not good enough. In reality, I believe they both should remain, and the partnership should work on logical data flows with data residing in two locations – the federated approach. This is how platforms reside next to each other instead of the single black hole.

  • The fact that standard interfaces will be developed between the two systems is a subtle sales argument with relatively low value. As I wrote in the “from technology to methodology”-paragraph, the challenges are in the organizational change within companies. Technology is not the issue, although system integrators also need to make a living.
  • What I believe makes sense is that both SAP and Siemens, have to realize their Industry 4.0 end-to-end capabilities. It is a German vision now for several years and it is an excellent vision to strive for. Now it is time to build the two platforms working together. This will be a significant technical challenge mainly for Siemens as its foundation is based on a coordinated backbone.
  • The biggest challenge, not only for this partnership, is the organizational change within companies that want to build an end-to-end connected solution. In particular, in companies with a vast legacy, the targeted industries by the partnership, the chasm between coordinated legacy data and intended connected data is enormous. Technology will not fix it, perhaps smoothen the pain a little.

 

Conclusion

With this post, we have reached the foundation of the item-centric approach for PLM, where the EBOM and MBOM are managed in a real-time context. Organizational change is the biggest inhibitor to move forward. The SAP-Siemens partnership is a sales/marketing approach to create a simplified view for the future at C-level discussions.
Let us watch carefully what happens in reality.

Next time potentially the dimension of change management and configuration management in an item-centric approach.
Or perhaps Martijn Dullaart will show us the way before, following up on his tricky poll question

 

This time a short post (for me) as I am in the middle the series “Learning from the past to understand the future” and currently collecting information for next week’s post. However, recently Rob Ferrone, the original Digital Plumber, pointed me to an interesting post from Scott Taylor, the Data Whisperer.

In code: The Virtual Dutchman discovered the Data Whisperer thanks to the original Digital Plumber.

Scott’s article with the title: “Data Management Hasn’t Failed, but Data Management Storytelling Has” matches precisely the discussion we have in the PLM community.

Please read his article, and just replace the words Data Management by PLM, and it could have been written for our community. In a way, PLM is a specific application of data management, so not a real surprise.

Scott’s conclusions give food for thought in the PLM community:

To win over business stakeholders, Data Management leadership must craft a compelling narrative that builds urgency, reinvigorates enthusiasm, and evangelizes WHY their programs enable the strategic intentions of their enterprise. If the business leaders whose support and engagement you seek do not understand and accept the WHY, they will not care about the HOW. When communicating to executive leadership, skip the technical details, the feature functionality, and the reference architecture and focus on:

  • Establishing an accessible vocabulary
  • Harmonizing to a common voice
  • Illuminating the business vision

When you tell your Data Management story with that perspective, it can end happily ever after.

It all resonates well with what I described in the PLM ROI Myth – it is clear that when people hear the word Myth, they have a bad connotation, same btw for PLM.

The fact that we still need to learn storytelling is because most of us are so much focused on technology and sometimes on discovering the new name for PLM in the future.

Last week I pointed to a survey from the PLMIG (PLM Interest Group) and XLifcycle, inviting you to help to define the future definition of PLM.

You are still welcome here: Towards a digital future: the evolving role of PLM in the future digital world.

Also, I saw a great interview with Martin Eigner on Minerva PLM TV interview by Jennifer Moore. Martin is well known in the PLM world and has done foundational work for our community

. According to Jennifer, he is considered as The Godfather of PLM.  This tittle fits nicely in today’s post. Those who have seen his presentations in recent years will remember Martin is talking about SysLM (System Lifecycle Management) as the future for PLM.

It is an interesting recording to watch – click on the image above to see it. Martin explains nicely why we often do not get the positive feedback from PLM implementations – starting at minute 13 for those who cannot wait.

In the interview, you will discover we often talk too much about our discipline capabilities where the real discussion should be talking business. Strategy and objectives are discussed and decided at the management level of a company. By using storytelling, we can connect to these business objectives.

The end result will be more likely that a company understands why to invest significantly in PLM as now PLM is part of its competitiveness and future continuity.

Conclusion

I shared links to two interesting posts from the last weeks. Studying them will help you to create a broader view. We have to learn to tell the right story. People do not want PLM – they have personal objectives. Companies have business objectives, and they might lead to the need for a new and changing PLM. Connecting to the management in an organization, therefore, is crucial.

Next week again more about learning from the past to understand the future

I usually write a post after participating in a PLM conference. Last week, I participated in TECHNIA’s PLM Innovation Forum, which was a 100 % virtual event with over 1500 registered participants from 58 countries. These numbers show the power of a virtual conference during these difficult times. It is an excellent option for a sustainable future – less travel to be there.

The additional beauty of this event is that, although the live sessions are over, all the content will be available until May 31st. You can still join!

It was (and is) a well-organized and massive event with over 70 sessions; the majority pre-recorded. As you can imagine 70 live sessions in two days would be too massive to grasp. Today the Friday after the event, I have been watching other sessions that have my interest, and it felt like another conference day.

TECHNIA, globally the largest Dassault Systèmes (DS)  implementer after DS themselves as Jonas Geyer, Technia’s CEO,  mentioned in his introduction speech, illustrated the breadth of their industry and technology skills complementary or based on the 3DEXPERIENCE platform.

TECHNIA was supported by Dassault Systèmes Execs and subject experts. In addition, a larger group of companies and interest groups supported the conference, even our humble PLM Green Alliance as you can see in the image above.

I followed the full two live days in real-time, meanwhile man sitting in my virtual booth to chat with virtual visitors. To my surprise, the anxiety during the conference felt like a physical conference – you get energized.

The positive point for me,  no finger food or a standing lunch and decent coffee when needed. The point to enhance and learn for this type of event, is to make the booth a little more human – perhaps supported by video?

At the end,  a great event, and if you are interested in the Dassault Systèmes/TECHNIA combined offering, supported by customer stories, take the chance till the end of May to register and browse the rich content.

 

Now I will share some of my picks from the live event. Another post will come based on my additional discoveries and networking discussions.

 

 

The B.CONNECT project

Fabien Hoefer and Philip Haller both from B.Braun, a medical device, and pharmaceutical company, with a wide range of products.  Their massive PLM-project, approx. sixty persons involved was driven by the fact that every product has a lot of related data stored in different silos that it becomes impossible to have the correct understanding and status and to maintain it for the product and service lifecycle, on average, 10 – 15 years.

Their target is a real PLM-platform implementation connecting the people, the processes, data, and systems. Their aim is really about the “connected” approach, a characteristic of a digital company.

As you can still watch the presentation, look at the following topics discussed:

  • focus on product archetypes instead of division (portfolio management)
  • data templates based on classification, global and specific data sets (data governance)
  • the need to have a Master Data Management in place (data governance)
  • the unique product identifier (remember the FFF-discussion in my blog)
  • data-driven documentation (a perfect example of a digital PLM implementation)
  • platform strategy (one application for one capability in a heterogeneous systems environment)
  • Ownership of the PLM implementation at board level (it is not an engineering tool)
  • in the Q&A – the mix of waterfall & agile – the hybrid approach (as in the medical world the validation of the system is required – a point we missed in the SmarTeam FDA toolkit – validation of a system is needed when the system/processes change)

In the Q&A session, it was clear that the big elephant in the room, the migration, has been identified, but no answers yet. See my presentation to understand the reference to the elephant.  I am curious about B. Braun’s approach, given my experience with PLM digital transformations. Will it be entirely digital or hybrid.

Looking forward to learning more from Fabien or Philip.

Business drivers for Sustainable Manufacturing

This session, presented by Hannes Lindfred from TECHNIA, was one of my favorite presentations,  as it links tightly to what we want to achieve with the PLM Green Alliance.

The subtitle of the presentation says it all: “How PLM can support Supply chain transparency, Circular economy, and System oriented product development”.

In a relaxed and entertaining manner, he explained the concepts and the needs of a circular economy, combined with examples from reality. In particular, I liked his closing statement linking the potential of digitization, modern PLM, and the circular economy. We have to learn to think and act circular. Highly recommended to watch!

Leading PLM Trends & Potential Disruptors

A PLM conference would not be a PLM-conference if Peter Bilello from CIMdata would not be speaking. We share a lot of insights related to digital transformation and the understanding it requires the involvement of PLM. However, it is not the traditional PLM that is needed.

PLM needs to be rethought, think about the concept of a Product Innovation Platform. A digital platform is required if we want end-to-end digitalization; otherwise, we keep working in optimized silos.

Peter shared some survey results (see below) from early this year. It illustrates that most companies currently invest in traditional PDM aspects. Restating the need for our PLM communities to learn and educate and rethink aspects of PLM and learn to communicate them.

Remarkably similar to some of the aspects I explained in my: From Coordinated to Connected presentations. Changing to data, changing workforce, changing processes meaning systems thinking. Another plea for everyone to invest in learning. See his concluding remarks:

The closing Q&A session was interesting, addressing additive manufacturing, the graph database, and potential PLM disruptors coming from outside the traditional PLM space.

I recommend, pay attention to the closing questions – so many good points to put PLM in perspective.

From Coordinated to Connected & Sustainable

Of course, I recommend you watch my presentation. It is one of the few opportunities to hear in a short time all the thoughts and concepts that I developed over the past 5 – 6 years. It saves you reading all my blog posts, which are less structured than this presentation.

I recommend you to watch this presentation in the context of Peter Bilello’s presentation as there are a lot of similarities, told in different words.

After my presentation, I appreciated the Q&A part, as it allowed me to point to some more of the related topics: Legacy CAD-issues – the incompatibility of the past and future data, Management vision and the Perception of ROI.

 

Professional PLM
Raise your standards and your horizons

An interesting presentation to watch, after seeing Peter Bilello’s presentation and my presentation,  is the one given by Roger Tempest. Roger is another veteran in the PLM-world and co-founder of the PLM Interest group. For many years Roger is striving to get the PLM professional recognized and certified. We both share the experience that being a PLM consultant is not a profession to become wealthy.

One of the reasons might be that the scope of PLM and what is the required skill level is not precise. PLM considered as an engineering tool and PLM having so many diverse definitions.

The challenge of Roger’s approach is that it tries to capture people within a standardized PLM framework, which becomes apparent in the Q&A session. Currently, he is in the stage of building a steering group, “looking for companies that are fairly committed to PLM”. So which companies are the ones interested in PLM to commit time and resources to build a professional PLM body? This can be only academic people and PLM Vendors/Implementers. The last group will probably not likely agree on standardization.

Also related to the question about the different industries and maturity levels for companies came with an unsatisfactory answer. He talks about “absolute” PLM and no need to compare PLM with other industries. Here I believe there is such a fundamental difference in the meaning of PLM when talking to the traditional manufacturing companies as compared to high-tech/software-driven industries. I inserted here Marc Halpern’s maturity/technology diagram that I have been referencing in my presentation too.

The final question about vendors joining the PLM standardization group seems to be a utopia. As I expressed in my presentation, referring to Marc Halpern’s business maturity diagram, the vendors show us the vision of various business aspects related to PLM.

Marc already indicated this is the phase of the Product Innovation Platform.

As long as the professional PLM organization is focusing on defining the standard, I foresee the outside world will move faster and be more diverse than a single PLM expert can handle. A typical issue with many other standards as you can see below.

What’s Next

I hope to see and participate more in virtual PLM conferences as it allows much larger audiences to connect compared to traditional conferences. However, there are things to improve, and therefore I want to propose some enhancements:

Make sure during the “live” sessions, there is the experience of “being live and connected”. Even when streaming a pre-recorded lecture, always follow-up immediately with a live Q&A session. I found the Q&A sessions very educative as they clarify or put the presentation in a broader context.

The current virtual booth as only a chat room is too primitive – it reminded me of the early days of internet communication – discussion groups in ASCII-terminal mode through Compuserve (remember). A booth could become a virtual meeting space on its own – all, of course, depending on the amount of bandwidth available. The feeling of “The Doctor is in”

It is great that the content is available for 30 days, and I agree there is a need for a time limit on the content; otherwise, the conference becomes more a library. What I would like to see after the “live” days to still have a kind of place for sharing. What are your favorite presentations, and why should others look at it?

 

Conclusion

A great event and learning experience for me. Virtual conferences are the future for sure, and I encourage others to develop this type of conferences related to PLM further. It is a way to share knowledge and discuss topics in a sustainable manner. In the upcoming 30 days, I will come back to the conference one more time, based on interesting topics discovered or discussion related to the content. 

Meanwhile, I encourage you too – if you are still in lockdown and if there is time to study – this is one of these unique opportunities.

 

Life goes on, and I hope you are all staying safe while thinking about the future. Interesting in the context of the future, there was a recent post from Lionel Grealou with the title: Towards PLM 4.0: Hyperconnected Asset Performance Management Framework.

Lionel gave a kind of evolutionary path for PLM. The path from PLM 1.0 (PDM) ending in a PLM 4.0 definition.  Read the article or click on the image to see an enlarged version to understand the logical order. Interesting to mention that PLM 4.0 is the end target, for sure there is a wishful mind-mapping with Industry 4.0.

When seeing this diagram, it reminded me of Marc Halpern’s diagram that he presented during the PDT 2015 conference. Without much fantasy, you can map your company to one of the given stages and understand what the logical next step would be. To map Lionel’s model with Marc’s model, I would state PLM 4.0 aligns with Marc’s column Collaborating.

In the discussion related to Lionel’s post, I stated two points. First, an observation that most of the companies that I know remain in PLM 1.0 or 2.0, or in Marc’s diagram, they are still trying to reach the level of Integrating.

Why is it so difficult to move to the next stage?

Oleg Shilovitsky, in a reaction to Lionel’s post, confirmed this. In Why did manufacturing stuck in PLM 1.0 and PLM 2.0? Oleg points to several integration challenges, functional and technical. His take is that new technologies might be the answer to move to PLM 3.0, as you can read from his conclusion.

What is my conclusion?

There are many promising technologies, but integration is remaining the biggest problem for manufacturing companies in adopting PLM 3.0. The companies are struggling to expand upstream and downstream. Existing vendors are careful about the changes. At the same time, very few alternatives can be seen around. Cloud structure, new data management, and cloud infrastructure can simplify many integration challenges and unlock PLM 3.0 for future business upstream and especially downstream. Just my thoughts…

Completely disconnected from Lionel’s post,  Angad Sorte from Plural Nordic AS wrote a LinkedIn post: Why PLM does not get attention from your CEO. Click on the image to see an enlarged version, that also neatly aligns with Industry 4.0. Coincidence, or do great minds think alike? Phil Collins would sing: It is in the air tonight

Angad’s post is about the historical framing of PLM as a system, an engineering tool versus a business strategy. Angrad believes once you have a clear definition, it will be easier to explain the next steps for the business. The challenge here is: Do we need, or do we have a clear definition of PLM? It is a topic that I do not want to discuss anymore due to a variety of opinions and interpretations.  An exact definition will never lead to a CEO stating, “Now I know why we need PLM.”

I believe there are enough business proof points WHY companies require a PLM-infrastructure as part of a profitable business. Depending on the organization, it might be just a collection of tools, and people do the work. Perhaps this is the practice in small enterprises?

In larger enterprises, the go-to-market strategy, the information needs, and related processes will drive the justification for PLM. But always in the context of a business transformation. Strategic consultancy firms are excellent in providing strategic roadmaps for their customers, indicating the need for a PLM-infrastructure as part of that.

Most of the time, they do not dive more in-depth as when it comes to implementation, other resources are needed.

What needs to be done in PLM 1.0 to 4.0 per level/stage is well described in all the diagrams on a high-level. The WHAT-domain is the domain of the PLM-vendors and implementers. They know what their tools and skillsets can do, and they will help the customer to implement such an environment.

The big illusion of all the evolutionary diagrams is that it gives a false impression of evolution.  Moving to the next level is not just switching on new or more technology and involve more people.

So the big question is HOW and WHEN to make progress.

HOW to make progress

In the past four years, I have learned that digital transformation in the domain of PLM is NOT an evolution. It is disruptive as the whole foundation for PLM changes. If you zoom in on the picture on the left, you will see the data model on the left, and the data model on the right is entirely different.

On the left side of the chasm, we have a coordinated environment based on data-structures (items, folders, tasks) to link documents.

On the right side of the chasm, we have a connected environment based on federated data elements and models (3D, Logical, and Simulation models).

I have been discussing this topic in the past two years at various PLM conferences and a year ago in my blog: The Challenges of a connected ecosystem for PLM

If you are interested in learning more about this topic, register for the upcoming virtual PLM Innovation Forum organized by TECHNIA. Registration is for free, and you will be able to watch the presentation, either live or recorded for 30 days.

At this moment, the detailed agenda has not been published, and I will update the link once the session is visible.  My presentation will not only focus on the HOW to execute a digital transformation, including PLM can be done, but also explain why NOW is the moment.

NOW to make progress

When the COVID19-related lockdown started, must of us thought that after the lockdown, we will be back in business as soon as possible. Now understanding the impact of the virus on our society, it is clear that we need to re-invent ourselves for a sustainable future, be more resilient.

It is now time to act and think differently as due to the lockdown, most of us have time to think.  Are you and your company looking forward to creating a better future? Or will you and your company try to do the same non-sustainable rat race of the past and being caught by the next crises.

McKinsey has been publishing several articles related to the impact of COVID19 and the article: Beyond coronavirus: The path to the next normal very insightful

As McKinsey never talks about PLM, therefore I want to guide you to think about more sustainable business.

Use a modern PLM-infrastructure, practices, and tools to remain competitive, meanwhile creating new or additional business models. Realizing concepts as digital twins, AR/VR-based business models require an internal transition in your company, the jump from coordinated to connected. Therefore, start investigating, experimenting in these new ways of working, and learn fast. This is why we created the PLM Green Alliance as a platform to share and discuss.

If you believe there is no need to be fast, I recommend you watch Rebecka Carlsson’s presentation at the PLMIF event. The title of her presentation: Exponential Tech in Sustainability. Rebecca will share insights for business development about how companies can upgrade to new business models based on the new opportunities that come with sustainability and exponential tech.

The reason I recommend her presentation because she addresses the aspect of exponential thinking nicely. Rebecka states we are “programmed” to think local-linear as mankind. Exponential thinking goes beyond our experience. Something we are not used doing until with the COVID19-virus we discovered exponential growth of the number of infections.

Finally, and this I read this morning, Jan Bosch wrote an interesting post: Why Agile Matters, talking about the fact that during the design and delivery of the product to the market, the environment and therefore the requirements might change. Read his post, unless as Jan states:

Concluding, if you’re able to perfectly predict the optimal set of requirements for a system or product years ahead of the start of production or deployment and if you’re able to accurately predict the effect of each requirement on the user, the customer and the quality attributes of the system, then you don’t need Agile.

What I like about Jan’s post is the fact that we should anticipate changing requirements. This statement combined with Rebecka’s call for being ready for exponential change, with an emerging need for sustainability, might help you discuss in your company how a modern New Product Introduction process might look like, including requirements for a sustainable future that might come in later (per current situation) or can become a practice for the future

Conclusion

Now is the disruptive moment to break with the old ways of working.  Develop plans for the new Beyond-COVID19-society.  Force yourselves to work in more sustainable modes (digital/virtual), develop sustainable products or services (a circular economy), and keep on learning. Perhaps we will meet virtually during the upcoming PLM Innovation Forum?

Note: You have reached the end of this post, which means you took the time to read it all. Now if you LIKE or DISLIKE the content, share it in a comment. Digital communication is the future. Just chasing for Likes is a skin-deep society. We need arguments.
Looking forward to your feedback.

Meanwhile, two weeks of a partial lockdown have passed here in the Netherlands, and we have at least another 3 weeks to go according to the Dutch government. The good thing in our country, decisions, and measures are made based on the advice of experts as we cannot rely on politicians as experts.

I realize that despite the discomfort for me, for many other people in other countries, it is a tragedy. My mental support to all of you, wherever you are.

So what has happened since Time to Think (and act differently)?

All Hands On Deck

In the past two weeks, it has become clear that a global pandemic as this one requires an “All Hands On Deck” mentality to support the need for medical supplies and in particular respiration devices, so-called ventilators. Devices needed to save the lives of profoundly affected people. I have great respect for the “hands” that are doing the work in infectious environments.

Due to time pressure, innovative thinking is required to reach quick results in many countries. Companies and governmental organizations have created consortia to address the urgent need for ventilators. You will not see so much PR from these consortia as they are too busy doing the real work.

Still, you see from many of the commercial participants their marketing messages, why, and how they contribute to these activities.

One of the most promoted capabilities is PLM collaboration on the cloud as there is a need for real-time collaboration between people that are under lockdown. They have no time setting-up environments and learning new tools to use for collaboration.

For me, these are grand experiments, can a group of almost untrained people corporate fast in a new environment.

For sure, offering free cloud software, PLM, online CAD or 3D Printing, seems like a positive and compassionate gesture from these vendors. However, this is precisely the wrong perception in our PLM-world – the difficulty with PLM does not lie necessary in the tools.

 

It is about learning to collaborate outside your silo.

Instead of “wait till I am done” it should become “this is what I have so far – use it for your progress”. This is a behavior change.

Do we have time for behavioral changes at this moment? Time will tell if the myth will become a reality so fast.

A lot of thinking

The past two weeks were weeks of thinking and talking a lot with PLM-interested persons along the globe using virtual meetings.

As long as the lockdowns will be there I keep on offering free of charge PLM coaching for individuals who want to understand the future of PLM.

Through all these calls, I really became THE VirtualDutchman in many of these meetings (thanks Jagan for the awareness).

I realized that there is a lot of value in virtual meetings, in particular with the video option on. Although I believe video works well when you had met before as most of my current meetings were with people, I have met before face-to-face. Hence, you know each other facial expressions already.

I am a big fan of face-to-face meetings as I learned in the past 20 years that despite all the technology and methodology issues, the human factor is essential. We are not rational people; we live and decide by emotions.

Still, I conclude that in the future, I could do with less travel, as I see the benefits from current virtual meetings.

Less face-to-face meetings will help me to work on a more sustainable future as I am aware of the impact flying has on the environment. Also, talking with other people, there is the notion that after the lockdowns, virtual conferencing might become more and more a best practice. Good for the climate, the environment, and time savings – bad for traditional industries like aircraft carriers, taxis, and hotels. I will not say 100 % goodbye but reduce.

A Virtual PLM conference!

I was extremely excited to participate in the upcoming PLM Innovation Forum (PLMIF) starting on April 28th, organized by TECHNIA. I have been visiting the event in the past a few times in Stockholm. It was a great place to meet many of the people from my network.

This time I am even more excited as the upcoming PLMIF will be a VIRTUAL conference with all the aspects of a real conference – read more about the conference here.

There will be an auditorium where lectures will be given, there are virtual booths, and it will be a place to network virtually. In my next post, I hope to zoom in on the conference.

Sustainability, a circular economy, and modern PLM should go together. Since 2014, these topics have been on the agenda of the joint CIMdata Roadmap/PDT conferences. Speakers like Amir Rashid KTH Sweden, Ken Webster Ellen MacArthur Foundation, and many others have been talking about the circular economy.

The Scandinavian mindset for an inclusive society for people and the environment for sure, has influenced the agenda. The links above lead to some better understanding of what is meant by a circular economy and a sustainable future, as also the short YouTube movie below:

The circular economy is crucial for a sustainable future. Therefore, I am looking forward to participating in the upcoming PLM Innovation Forum on April 28th, where it will be all about digitalization for sustainable product development and manufacturing. Hopefully, with the right balance towards the WHY-side of our brain, not so much about WHAT.

You are welcomed to register for free here: the virtual PLM Innovation Forum – we might meet there (virtually).

The PLM Green Alliance

The PLM Green Alliance had been announced some months ago, started by Rich McFall and supported by  Bjorn Fidjeland,  Oleg Shilovitsky, and me.

It was the first step to proactively bringing people together to discuss topics like reducing our carbon footprint, sharing and brainstorming about innovations that will lead to a sustainable future for ourselves and our children, grand-grand-children. The idea behind the PLM Green Alliance is that a proactive approach is much cheaper in the long term as we can still evaluate and discuss options.

This brings me back to the All hands On Deck approach we currently use for fighting the COVID-19 virus.

In a crisis mode, the damage to the people and the economy is severe. Besides, in a crisis mode, a lot of errors will be made, but don’t blame or joke about these people that are trying. Without failure, there is no learning.

We are in a potential time of disruption as the image shows below, but we do not have the complete answers for the future

Think about how you could pro-actively work on a sustainable future for all of us. This will be my personal target, combined with explaining and coaching companies related to topics of modern PLM, during the current lockdown and hopefully long after. The PLM Green Alliance is eager to learn from you and your companies where they are contributing to a more sustainable and greener future.

Do not feel your contribution is not needed, as according to research done by the Carr Center’s Erica Chenoweth: The ‘3.5% rule’: How a small minority can change the world. It could be an encouragement to act instead of watching who will determine your future.

Conclusion

While learning to live in a virtual world, we might be realizing that the current crisis is an opportunity to switch faster to a more sustainable and inclusive society. For PLM moving to data-driven, cloud-based environments, using a Model-Based approach along the whole lifecycle, is a path to reduce friction when delivering innovations. From years to weeks? Something we wished to have today already. Stay safe!

People, wherever you are, we are in a kind of lockdown. Some countries more restricted than others. Still, the challenge will be for most of us how to survive in two perhaps three months of being locked in your home and make the best of it. As I am not a virus expert, I will not give you any recommendations on this topic. As a PLM geek, I want to share with you the opportunities I see for the upcoming months.

A crisis is an opportunity

Most of us should be lucky that we do not live in the same situation as twenty years ago. At that time, internet connectivity was expensive and slow. Meaning working from home would mean isolation from the rest of the world. The positive point now is that we can be connected virtually without travel, without face-to-face meetings, and we are pushed to do so. This external push is an interesting point for me.

The traditional attitude for my PLM engagements was that face-to-face meetings are crucial for creating a human connection and trust. Now I ask myself is this a behavior of the past that should become obsolete in the future. Probably we cannot afford this approach anymore in the future if we take sustainability and the environment into consideration. We live now in a globally connected world, but should we act still in the old way?

Perhaps not. Let’s look at some of the examples that it is time to shift behaviors.

We might think in the Western world we know it all due to our dominance in the past hundred years. However, when you study history, you will see civilizations come to power and after hundreds of years, they lose power because they kill themselves internally. Apparently, a typical human property that will not disappear – still interesting to analyze when considering a globally connected world. Where is the point of gravity today?

Interestingly, the ancient Chinese population already knew that a crisis was an opportunity, as I am being told. The Chinese characters for crisis mean danger and opportunity, respectively, according to Wiki – see the image above. Joe Barkai was one of the first in my network that took action to explain that instead of focusing on the loss of what is happening now, we should take the opportunity to be better prepared for the future. You can read his post here: The Corona virus and your company’s brand. And these kinds of messages are popping up more frequently now. Let’s stay safe while thinking and preparing for the future.

Now a PLM related example.

Remember what the FFF is happening?

Two-three weeks ago, we had a vivid discussion in our PLM and CM community based on the famous FFF mnemonic.  What the FFF is happening was a post sharing my point of view, and there were a lot of reactions from different people.

The purpose of my post was to explain that the whole discussion was based on paradigms that drawings are defining the part. Because of that, we have a methodology to decide if YES or NO we need a new part number or revision. To me, this practice should no longer be a discussion.

A part has a unique identifier, and a document has a unique identifier. In PLM-systems, the information is managed by relations, no longer by identifiers – who knows the exact unique identifier? In a PLM-system information is connected, and the attributes of the part and document will tell you the details of the type of information. “Intelligent or meaningful” identifiers are in such an environment no longer relevant. Think about that…..

In the comments of my post, Jesse Leal was confirming this statement:

This in contrary to Joe Brouwer, who you might have noticed, always is spitting his opinion that the good old days of the draftsman are gone, Boeing made a tremendous mistake and that PLM is fake. This all combined with hyperlinks to his products and opinions. The comment below says it all:

Two points to observe in this response:

Hey, Bob, send me the new digital identifier”.

This statement assumes that if a person needs to retrieve information from someone else, they need to contact this person (Bob).

Bob then needs to drop his current work and answer to the response and send the latest version of a drawing?  This is old school. In a PLM-system,  information should be connected, and if Bob has released his latest drawing (no matter if it is FFF), any user could find the latest approved version, not even having to look at the identifier (which could be meaningless) but by following the relations between products, parts, and documents.

This is PLM!

One of the benefits, Bob does not get disturbed during the day by these kinds of questions and can focus on his critical work as an expert.

Second, if you need to sit with a designer to understand PLM, then you are probably talking with the wrong person. Designers work in the context of PDM. When we speak about PLM, we are talking about a broader scope beyond engineering and design.

This is a common mistake in a lot of marketing stories. Companies that focus on the design space only, some EBOM-integrations with CAD-systems, are most of the time focusing on PDM.  When Agile PLM came out (later Oracle E9) and later Aras without CAD-integrations, these companies were focusing on the flow of information inside the company, not necessarily driven by CAD. Of course, the traditional PLM companies combine CAD integration with other capabilities. Dassault Systèmes, Siemens, and PTC all have a strong relationship with their native CAD-systems. However, their offerings go way beyond CAD-integrations e.g. end-to-end governance, change processes and an item-centric backbone.

The diagram above explains the basics for the future. In a push-mode, the person in the middle has the responsibility to distribute information and ensure it remains accurate for all stakeholders. This makes this person crucial (good job security) but extremely inefficient compared to people working in the pull-mode, being responsible for getting the accurate data themselves. It may be clear the pull-mode is the model of a digital enterprise.

So if you have the time now, take this time to rethink how well your company is ready for a digital future. Companies that currently rely on Bob are in trouble as Bob is currently sitting at home. Companies that have learned to shift from the push-mode to the pull-mode could continue working as planned, as they do not need Bob. And don’t worry about your job. If you are in Bob’s position you will lose your job over time. However, when you keep on evolving, learning and adding value to your company, you will be always needed – don’t lock yourself in.

If you want to be inspired more in this area, read Jan Bosch’s post: This is not the end . Here Jan mentions the opportunity to move to digital practices (and more) – get out of our traditional patterns

 

What can you do?

Even though COVID-19 has, and will have, a dramatic impact on our society, this is also the moment to rewire some of our processes.  Because there was never time to think and act due to the running business. It reminded me of the financial crisis in 2008, when the market for PLM vendors was terrible, no significant sales for them as companies could not invest.

However, for me, 2008 was an extremely busy year,  thanks to all kinds of regulations from governments. There was time and budget to support employees to raise their skills and PLM was one of these domains. That year I conducted many workshops. It was also the year that I started my blog virtualdutchman.com.

Now we are in a similar situation and probably worse as now we are locked to our homes. However, we are also better connected. Imagine this situation without the internet. Now we can learn even better.

So let’s benefit from this connectivity and use the lockdown time to learn, think, and discuss with peers. Challenge and involve the management of your company how they see and lead to the future.


In that context, I am happy to spend on average one day per week on free conference calls if you need clarification or support for your PLM-related ideas.

Contact me through a personal message on LinkedIn, and we will find a way to connect.

 

Conclusion

This decade will be decisive for many of us. At the beginning of this year, I wrote PLM 2020- The next decade (4 challenges). With my narrow PLM-mind, I overlooked viruses. Bill Gates did not do that, as you can see from his 2015 TED talk: The next outbreak? We’re not ready.  Bill also explains that our traditional thinking patterns should change in a globally connected world.

I wish you all the time to think and educate yourself and prepare for a changed future. Stay safe inside, stay healthy, knowing for some of you it will be a big challenge.

At the beginning of this week, I was attending the 9th edition of the PI conference in London. Where it started as a popular conference with 300 – 400 attendees at its best, we were now back to a smaller number of approximately 100 attendees.

It illustrates that PLM as a standalone topic is no longer attracts a broad audience as Marketkey (the organization of the conference) confirms. The intention is that future conferences will be focusing on the broader scope of PLM, where business transformation will be one of the main streams.

In this post, I will share my highlights of the conference, knowing that other sessions might have been valuable too, but I had to make a choice.

It is about people

Armin Prommersberger, CTO from DIRAC and the chairman of the conference, made a great point: “What we will discuss in the upcoming two days, it is all about people not about technology.”

I am not sure if this opening has influenced the mood of the conference, as when I look back to what was the central theme: It is all about how we deal with people when explaining, implementing and justifying PLM.

AI at the Forefront of a Digital Transformation

Muhannad Alomari from R2 Data Labs as a separate unit within Rolls Royce to explore and provide data innovation started with his keynote speech sharing the AI initiatives within his team.

He talked about several projects where AI will become crucial.

For example, the EHM program related to engine behavior. How to detect anomalies, how to establish predictive maintenance and maximize the time an airplane engine is in operation. Interesting to mention is that Muhannad explained that most simulation models are based on simplified simulation models, not accurate enough to discover anomalies.

Modeling in the PLM world with feedback from reality

Machine learning and feedback loops are crucial to optimize the models both for the discovery of irregularities and, of course, to improve understanding of the engine behavior and predict maintenance. Currently, maintenance is defined based on the worst-case scenario for the engine, which in reality, of course, will not be the case for most engines. There is a lot (millions) to gain here for a company.

Interesting to mention is that Muhannad gave a realistic view of the current status of Artificial Intelligence (AI). AI is currently still dumb – it is a set of algorithms that need to be adapted whenever new patterns are discovered. Deep learning is still not there – currently, we still need human beings for that.

This was in contrast with the session from Kalypso later with the title: Supercharge your PLM with advanced analytics. It was a typical example of where a realistic story (R2 Data Labs) shows such a big difference with what is sold by PLM vendors or implementers. Kalypso introduced Product Lifecycle Intelligence (PLI) – you can see the dream on the left (click on the image to enlarge).

Combine PLM with Analytics, and you have Intelligence.  My main comment is, knowing from the field the first three phases in most companies have a lack of data quality and consistency. Therefore any “Intelligence” probably will be based on unreliable sources. Not an issue if you are working in the domain of politics, however when it comes to direct cost and quality implications, it can be a significant risk. We still have a way to go before we have a reliable PLM data backbone for analytics.

 

Keeping PLM Momentum after a Successful Campaign

Susanna Mäentausta from Kemira in Finland gave an exciting update of their PLM project. Where in 2019, she shared with us their PLM roadmap (see my 2019 post: The weekend after PI PLMx London 2019); this time, Susanna shared with us how they are keeping the PLM momentum.

Often PLM implementations are started based on a hypothetical business case (I talked about this in my post The PLM ROI Myth). But then, when you implement PLM, you need to take care you provide proof points to motivate the management. And this is exactly what the PLM team in Kemira has been doing. Often management believes that after the first investment, the project is done (“We bought the software – so we are done”) however the business and process change that will deliver the value is not reported.

Susanna shared with us how they defined measurable KPIs for two reasons.  First, to motivate the management that there are business progress and benefits, however, it is a journey. And secondary the facts are used to kill the legends that “Before PLM we were much faster or efficient.” These types of legends are often expressed loudly by persons who consider PLM as an overhead (killing their freedom) instead of a way to be more efficient in business. In the end, for a company, the business is more important than the person’s belief.

On the question for Susanna, what she would have done better with hindsight, she answered: “Communicate, communicate, communicate.” A response I fully support as often PLM teams are too busy completing their day-to-day work, that there is no spare time for communication. Crucial to achieving a business change.

My agreement: PLM needs facts based during implementation and support combined with the understanding we are dealing with people and their emotions too. Both need full attention.

Acceleration Digitalization at Stora Enso

Samuli Savo, Chief Digital Officer at Stora Enso, explained the principles of innovation, related to digitalization at his company. Stora Enso, a Swedish/Finish company, historically one of the largest forestry companies in the world as well as one of the most significant paper and packaging producers, is working on a transformation to become the renewable materials company. For me, he made two vital points on how Stora Enso’s digitalization’s journey is organized.

He pleads for experimentation funded by corporate as in the experimental stage, as it does not make sense to have a business case. First DO and then ANALYZE, where many companies have to policy first to ANALYZE and then DO, killing innovative thinking.

The second point was the active process to challenge startups to solve business challenges they foresee and, combined with a governance process for startups, allow these companies to be supported and become embedded within member companies of the Combient Foundry, like Stora Enso. By doing such in a structured way, the outcome must lead to innovation.

I was thinking about the hybrid enterprise model that I have been explaining in the past. Great story.

Cyber-security and Future Mobility

Out of interest, I followed the session from Madeline Cheah, Cybersecurity Innovation Lead at HORIBA MIRA. She gave an excellent and well-structured overview. Madeline leads the cybersecurity research program. Part of this job is investigating ways to prevent vehicles from being attacked.  In particular, when it comes to connected and autonomous vehicles. How to keep them secure.

She discussed the known gaps are and the cybersecurity implications of future mobility so extensive that I even doubted will there ever be an autonomous vehicle on the road. So much to define and explore. She looked at it from the perspective of the Internet of Everything, where Everything is divided into Things, Data, Processes, and People. Still, a lot of work to do, see image below

Good Times Ahead: Delay Mitigation Through a Plan for Every Part

Ian Quest, director at Quick Release, gave an overview of what their company aims to be. You could translate it as the plumbers of the automotive industry Where in the ideal world information should be flowing from design to release, there are many bottlenecks, leakages, hiccups that need to be resolved as the image shows.

Where their customers often do not have the time and expertise to fix these issues, Quick Release brings in various skillsets and common sense. For example, how to deal with the Bill of Materials, Configuration Management, and many other areas that you need to address with methodology first instead of (vendor-based) technology. I believe there is a significant need for this type of company in the PLM-domain.

The second part, presented by Nick Solly, with a focus on their QRonos tool, was perhaps a little too much a focus on the capabilities of the tool. Ian Quest, in his introduction,  already made the correct statement:

The QRonos tool, which is more or less a reporting tool, illustrates again that when people care about reliable data (planning, tasks, parts, deliverables, …..), you can improve your business significantly by creating visibility to delays or bottlenecks. The value lies in measurable activities and from there, learn to predict or enhance – see R2 Labs, Kemira and the PLI dream.

Conclusion

It is clear that a typical PLM conference is no longer a technology festival – it is about people. People are trying to change or improve their business. Trying to learn from each other, knowing that the technical concepts and technology are there.

I am looking forward to the upcoming PI events where this change will become more apparent.

 

Last week I shared my thoughts related to my observation that the ROI of PLM is not directly visible or measurable, and I explained why. Also, I explained that the alignment of an organization requires a myth to make it happen. A majority of readers agreed with these observations. Some others either misinterpreted the headlines or twisted the story in favor of their opinion.

A few came from Oleg Shilovitsky and as Oleg is quite open in his discussions, it allows me to follow-up on his statements. Other people might share similar thoughts but they haven’t had the time or opportunity to be vocal. Feel free to share your thoughts/experiences too.

Some misinterpretations from Oleg’s post: PLM circa 2020 – How to stop selling Myths

  • The title “How to stop selling Myths” is the first misinterpretation.
    We are not selling myths – more below.
  • “Jos Voskuil’s recommendation is to create a myth. In his PLM ROI Myths article, he suggests that you should not work on a business case, value, or even technology” is the second misinterpretation, you still need a business case, you need value and you need technology.

And I got some feedback from Lionel Grealou, who’s post was a catalyst for me to write the PLM ROI Myth post. I agree I took some shortcuts based on his blog post. You can read his comments here. The misinterpretation is:

  • “Good luck getting your CFO approve the business change or PLM investment based on some “myth” propaganda :-)” as it is the opposite, make your plan, support your plan with a business case and then use the myth to align

I am glad about these statements as they allow me to be more precise, avoiding misperceptions/myth-perceptions.

A Myth is bad

Some people might think that a myth is bad, as the myth is most of the time abstract.  I think these people do not realize that there a lot of myths that they are following; it is a typical social human behavior to respond to myths. Some myths:

  • How can you be religious without believing in myths?
  • In this country/world, you can become anything if you want?
  • In the past, life was better
  • I make this country great again

The reason human beings need myths is that without them, it is impossible to align people around abstract themes. Try for each of the myths above to create an end-to-end logical story based on factual and concrete information. Impossible!

Read Yuval Harari’s book Sapiens about the power of myths. Read Steven Pinker’s book Enlightenment Now to understand that statistics show a lot of current myths are false. However, this does not mean a myth is bad. Human beings are driven by social influences and myths – it is our brain.

Unless you have no social interaction, you might be immune to myths. With brings me to quoting Oleg once more time:

“A long time ago when I was too naive and too technical, I thought that the best product (or technology) always wins. Well… I was wrong. “

I went through the same experience, having studied physics and mathematics makes you think extremely logical. Something I enjoyed while developing software. Later, when I started my journey as the virtualdutchman mediating in PLM implementations, I discovered logical alone does not work in businesses. The majority of decisions are done based on “gut feelings” still presented as reasonable cases.

Unless you have an audience of Vulcans, like Mr. Spock, you need to deal with the human brain. Consider the myth as the envelope to pass the PLM-project to the management. C-level acts by myths as so far I haven’t seen C-level management spending serious time on understanding PLM. I will end with a quote from Paul Empringham:

I sometimes wish companies would spend 6 months+ to educate themselves on what it takes to deliver incremental PLM success BEFORE engaging with software providers

You don’t need a business case

Lionel is also skeptical about some “Myth-propaganda” and I agree with him. The Myth is the envelope, inside needs to be something valuable, the strategy, the plan, and the business case. Here I want to stress one more time that most business cases for PLM are focusing on tool and collaboration efficiency. And from there projecting benefits. However, how well can we predict the future?

If you implement a process, let’s assume BOM-collaboration done with Excel by BOM-collaboration based on an Excel-on-the-cloud-like solution, you can measure the differences, assuming you can measure people’s efficiency. I guess this is what Oleg means when he explains OpenBOM has a real business case.

However, if you change the intent for people to work differently, for example, consult your supplier or manufacturing earlier in the design process, you touch human behavior. Why should I consult someone before I finish my job, I am measured on output not on collaboration or proactive response? Here is the real ROI challenge.

I have participated in dozens of business cases and at the end, they all look like the graph below:

The ROI is fantastic – after a little more than 2 years, we have a positive ROI, and the ROI only gets bigger. So if you trust the numbers, you would be a fool not to approve this project. Right?

And here comes the C-level gut-feeling. If I have a positive feeling (I follow the myth), then I will approve. If I do not like it, I will say I do not trust the numbers.

Needless to say that if there was a business case without ROI, we do not need to meet the C-level. Unless, and it happens incidental, at C-level, there was already a decision we need PLM from Vendor X because we played golf together, we are condemned together or we believe the same myths.

In reality, the old Gartner graph from realized benefits says it all. The impact of culture, processes, and people can make or break a plan.

You do not need an abstract story for PLM

Some people believe PLM on its own is a myth. You just need the right technology and people will start using it, spreading it out and see how we have improved business. Sometimes email is used as an example. Email is popular because you can with limited effort, collaborate with people, no matter where they are. Now twenty years later, companies are complaining about the lack of traceability, the lack of knowledge and understanding related to their products and processes.

PLM will always have the complexity of supporting traceability combined with real-time collaboration. If you focus only on traceability, people will complain that they are not a counter clerk. If you focus solely on collaboration, you miss the knowledge build-up and traceability.

That’s why PLM is a mix of governance, optimized processes to guarantee quality and collaboration, combined with a methodology to tune the existing processes implemented in tools that allow people to be confident and efficient. You cannot translate a business strategy into a function-feature list for a tool.

Conclusion

Myths are part of the human social alignment of large groups of people. If a Myth is true or false, I will not judge. You can use the Myth as an envelope to package your business case. The business case should always be a combination of new ways of working (organizational change), optimized processes and finally, the best tools. A PLM tool-only business case is to my opinion far from realistic

 

Now preparing for PI PLMx London on 3-4 February – discussing Myths, Single BOMs and the PLM Green Alliance

%d bloggers like this: