You are currently browsing the category archive for the ‘Platform’ category.

PDT Europe is over, and it was this year a surprising aligned conference, showing that ideas and concepts align more and more for modern PLM. Håkan Kårdén opened the conference mentioning the event was fully booked, about 160 attendees from over 19 countries. With a typical attendance of approx. 120 participants, this showed the theme of the conference: Continuous Transformation of PLM to support the Lifecycle Model-Based Enterprise was very attractive and real. You can find a history of tweets following the hashtag #pdte17

Setting the scene

Peter Bilello from CIMdata kicked-off by bringing some structure related to the various Model-Based areas and Digital Thread. Peter started by mentioning that technology is the least important issue as organization culture, changing processing and adapting people skills are more critical factors for a successful adoption of modern PLM. Something that would repeatedly be confirmed by other speakers during the conference.

Peter presented a nice slide bringing the Model-Based terminology together on one page. Next, Peter took us through various digital threads in the different stages of the product lifecycle. Peter concluded with the message that we are still in a learning process redefining optimal processes for PLM, using Model-Based approaches and Digital Threads and thanks (or due) to digitalization these changes will be rapid. Ending with an overall conclusion that we should keep in mind:


It isn’t about what we call digitalization; It is about delivering value to customers and all other stakeholders of the enterprise

Next Marc Halpern busted the Myth of Digital Twins (according to his session title) and looked into realistic planning them. I am not sure if Marc smashed some of the myths although it is sure Digital Twin is at the top of the hype cycle and we are all starting to look for practical implementations. A digital twin can have many appearances and depends on its usage. For sure it is not just a 3D Virtual model.

There are still many areas to consider when implementing a digital twin for your products. Depending on what and how you apply the connection between the virtual and the physical model, you have to consider where your vendor really is in maturity and avoid lock in on his approach. In particular, in these early stages, you are not sure which technology will last longer, and data ownership and confidentially will play an important role. And opposite to quick wins make sure your digital twin is open and use as much as possible open standards to stay open for the future, which also means keep aiming for working with multiple vendors.

Industry sessions

Next, we had industry-focused sessions related to a lifecycle Model-Based enterprise and later in the afternoon a session from Outotec with the title: Managing Installed Base to Unlock Service opportunities.

The first presentation from Väino Tarandi, professor in IT in Construction at KTH Sweden presented his findings related to BIM and GIS in the context of the lifecycle, a test bed where PLCS meets IFC. Interesting as I have been involved in BIM Level 3 discussions in the UK, which was already an operational challenge for stakeholders in the construction industry now extended with the concept of the lifecycle. So far these projects are at the academic level, and I am still waiting for companies to push and discover the full benefits of an integrated approach.

Concepts for the industrial approach could be learned from Outotec as you might understand later in this post. Of course the difference is that Outotec is aiming for data ownership along the lifecycle, where in case of the construction industries, each silo often is handled by a different contractor.

Fredrik Ekström from Swedish Transport Administration shared his challenges of managing assets for both road and railway transport – see image on the left. I have worked around this domain in the Netherlands, where asset management for infrastructure and asset management for the rail infrastructure are managed in two different organizations. I believe Fredrik (and similar organizations) could learn from the concepts in other industries. Again Outotec’s example is also about having relevant information to increase service capabilities, where the Swedish Transport Administration is aiming to have the right data for their services. When you look at the challenges reported by Fredrik, I assume he can find the answers in other industry concepts.

Outotec’s presentation related to managing installed base and unlock service opportunities explained by Sami Grönstrand and Helena Guiterrez was besides entertaining easy to digest content and well-paced. Without being academic, they explained somehow the challenges of a company with existing systems in place moving towards concepts of a digital twin and the related data management and quality issues. Their practical example illustrated that if you have a clear target, understanding better a customer specific environment to sell better services, can be achieved by rational thinking and doing, a typical Finish approach. This all including the “bi-modal approach” and people change management.

Future Automotive

Ivar Hammarstadt, Senior Analyst Technology Intelligence for Volvo Cars Corporation entertained us with a projection toward the future based on 160 years of automotive industry. Interesting as electrical did not seem to be the only way to go for a sustainable future depending on operational performance demands.

 

Next Jeanette Nilsson and Daniel Adin from Volvo Group Truck shared their findings related to an evaluation project for more than one year where they evaluated the major PLM Vendors (Dassault Systemes / PTC / Siemens) on their Out-of-the-box capabilities related to 3D product documentation and manufacturing.

They concluded that none of the vendors were able to support the full Volvo Truck complexity in a OOTB matter. Also, it was a good awareness project for Volvo Trucks organization to understand that a common system for 3D geometry reduces the need for data transfers and manual data validation. Cross-functional iterations can start earlier, and more iterations can be performed. This will support a shortening of lead time and improve product quality. Personally, I believe this was a rather expensive approach to create awareness for such a conclusion, pushing PLM vendors in a competitive pre-sales position for so much detail.

Future Aerospace

Kenny Swope from Boeing talked us through the potential Boeing journey towards a Model-Based Enterprise. Boeing has always been challenging themselves and their partners to deliver environments close to what is possible. Look at the Boeing journey and you can see that already in 2005 they were aiming for an approach that most of current manufacturing enterprises cannot meet. And now they are planning their future state.

To approach the future state Boeing aims to align their business with a single architecture for all aspects of the company. Starting with collecting capabilities (over 400 in 6 levels) and defining value streams (strategic/operational) the next step is mapping the capabilities to the value streams.  Part of the process would be to look at the components of a value stream if they could be fulfilled by a service. In this way you design your business for a service-oriented architecture, still independent from any system constraints. As Kenny states the aerospace and defense industry has a long history and therefore slow to change as its culture is rooted in the organization. It will be interesting to learn from Kenny next hear how much (mandatory) progress towards a model-based enterprise has been achieved and which values have been confirmed.

Gearing up for day 2

Martin Eigner took us in high-speed mode through his vision and experience working in a bi-modular approach with Aras to support legacy environments and a modern federated layer to support the complexity of a digital enterprise where the system architecture is leading. I will share more details on these concepts in my next post as during day 2 of PDT Europe both Marc Halpern and me were talking related to this topic, and I will combine it in a more extended story.

The last formal presentation for day one was from Nigel Shaw from Eurostep Ltd where he took us through the journey of challenges for a model-based enterprise. As there will not be a single model that defines all, it will be clear various models and derived models will exist for a product/system.  Interesting was Nigel’s slide showing the multiple models disciplines can have from an airplane (1948). Similar to the famous “swing” cartoon, used to illustrate that every single view can be entirely different from the purpose of the product.

Next are these models consistent and still describing the same initial specified system. On top of that, even the usage of various modeling techniques and tools will lead to differences in the system. And the last challenge on top is managing the change over the system’s lifecycle. From here Nigel stepped into the need for digital threads to govern relations between the various views per discipline and lifecycle stage, not only for the physical and the virtual twin.  When comparing the needs of a model-based enterprise through its lifecycle, Nigel concluded that using PLCS as a framework provides an excellent fit to manage such complexity.

Finally, after a panel discussion, which was more a collection of opinions as the target was not necessary to align in such a short time, it was time for the PDT dinner always an excellent way to share thoughts and verify them with your peers.

Conclusion

Day 1 was over before you knew it without any moment of boredom and so I hope is also this post. Next week I will close reviewing the PDT conference with some more details about my favorite topics.

 

Advertisements

At this moment there are two approaches to implement PLM. The most common practice is item-centric and model-centric will be potentially the best practice for the future. Perhaps your company still using a method from the previous century called drawing-centric. In that case, you should read this post with even more attention as there are opportunities to improve.

 

The characteristics of item-centric

In an item-centric approach, the leading information carrier is an item also known as a part. The term part is sometimes confusing in an organization as it is associated with a 3D CAD part. In SAP terminology the item is called Material, which is sometimes confusing for engineering as they consider Material the raw material. Item-centric is an approach where items are managed and handled through the whole lifecycle. In theory, an item can be a conceptual item (for early estimates), a design item (describing the engineering intent), a manufacturing item (defining how an item is consumed) and potentially a service item.

The picture below illustrates the various stages of an item-centric approach. Don’t focus on the structure, it’s an impression.

It is clear these three structures are different and can contain different item types. To read more about the details for an EBOM/MBOM approach read these post on my blog:

Back to item-centric. This approach means that the item is the leading authority of the product /part. The id and revision describe the unique object in the database, and the status of the item tells you in the current lifecycle stage for the item. In some cases, where your company makes configurable products also the relation between two items can define effectivity characteristics, like data effectivity, serial number effectivity and more. From an item structure, you can find its related information in context. The item points to the correct CAD model, the assembly or related manufacturing drawings, the specifications. In case of an engineering item, it might point towards approved manufacturers or approved manufacturing items.

Releasing an item or a BOM means the related information in context needs to validated and frozen too. In case your company works with drawings for manufacturing, these drawings need to be created, correct and released, which sometimes can be an issue due to some last-minute changes that can happen. The above figure just gives an impression of the potential data related to an item. It is important to mention that reports, which are also considered documents, do not need an approval as they are more a snapshot of the characteristics at that moment of generation.

The advantages of an item-centric approach are:

  • End-to-end traceability of information
  • Can be implemented in an evolutionary approach after PDM-ERP without organizational changes
  • It enables companies to support sharing of information
  • Sharing of information forces companies to think about data governance
    (not sure if a company wants to invest on that topic)

The main disadvantages of an item-centric approach are:

  • Related information on the item is not in context and therefore requires its own management and governance to ensure consistency
  • Related information is contained in documents, where availability and access is not always guaranteed

Still, the item-centric approach brings big benefits to a company that was working in a classical drawing-driven PDM-ERP approach. An additional remark needs to be made that not every company will benefit from an item-centric approach as typically Engineering-to-Order companies might find this method creating too much overhead.

The characteristics of Model-Centric

A model-centric approach is considered the future approach for modern enterprises as it brings efficiency, speed, multidisciplinary collaboration and support for incremental innovation in an agile way. When talking about a model-centric approach, I do not mean a 3D CAD model-centric approach. Yes, in case the product is mature, there will be a 3D Model serving as a base for the physical realization of the product.

However, in the beginning, the model can be still a functional or logical model. In particular, for complex products, model-based systems engineering might be the base for defining the solution. Actually, when we talk about products that interact with the outside world through software, we tend to call them systems. This explains that model-based systems engineering is getting more and more a recommended approach to make sure the product works as expected, fulfills all the needs for the product and creates a foundation for incremental innovation without starting from scratch.

Where the model-based architecture provides a framework for all stakeholders, the 3D CAD model will be the base for a digital thread towards manufacturing. Linking parameters from the logical and functional model towards the physical model a connection is created without the need to create documents or input-files for other disciplines. Adding 3D Annotations to the 3D CAD model and manufacturing process steps related to the model provides a direct connection to the manufacturing process.

The primary challenge of this future approach is to have all these data elements (requirements, functions, components, 3D design instances, manufacturing processes & resources to be connected in a federated environment (the product innovation platform). Connecting, versioning and baselining are crucial for a model-centric approach. This is what initiatives like Industry 4.0 are now exploring through demonstrators, prototypes to get a coherent collection of managed data.

Once we are able to control this collection of managed data concepts of digital twin or even virtual twin can be exploited linking data to a single instance in the field.

Also, the model can serve as the foundation for introduction incremental innovation, bringing in new features.  As the model-based architecture provides direct visibility for change impact (there are no documents to study), it will be extremely lean and cost-efficient to innovate on an existing product.

Advantages of model-centric

  • End-to-end traceability of all data related to a product
  • Extremely efficient in data-handling – no overhead on data-conversions
  • Providing high-quality understanding of the product with reduced effort compared to drawing-centric or item-centric approaches
  • It is scalable to include external stakeholders directly (suppliers/customers) leading to potential different, more beneficial business models
  • Foundation for Artificial Intelligence at any lifecycle step.

Disadvantages of model-centric

  • It requires a fundamentally different way of working compared to past. Legacy departments, legacy people, and legacy data do not fit directly into the model-centric approach. A business transformation is required, not evolution.
  • It is all about sharing data, which requires an architecture that is built to share information across Not through a service bus but as a (federated) platform of information.
    A platform requires a strong data governance, both from the dictionary as well as authorizations which discipline is leading/following.
  • There is no qualified industrial solution from any vendor yet at this time. There is advanced technology, there are demos, but to my knowledge, there is no 100% model-centric enterprise yet. We are all learning. Trying to distinguish reality from the hype.

 

Conclusions

The item-centric approach is the current best practice for most PLM implementations. However, it has the disadvantage that it is not designed for a data-driven approach, the foundation of a digital enterprise. The model-centric approach is new. Some facets already exist. However, for the total solution companies, vendors, consultants, and implementers are all learning step-by-step how it all connects. The future of model-centric is promising and crucial for survival.

Do you want to learn where we are now related to a model-centric approach?
Come to PDT2017 in Gothenburg on 18-19th October and find out more from the experts and your peers.

Last week I published a dialogue I had with Flip van der Linden, a fellow Dutchman and millennial, eager to get a grip on current PLM. You can read the initial post here: A PLM dialogue.  In the comments, Flip continued the discussion (look here).  I will elaborate om some parts of his comments and hope some others will chime in. It made me realize that in the early days of blogging and LinkedIn, there were a lot of discussions in the comments. Now it seems we become more and more consumers or senders of information, instead of having a dialogue. Do you agree? Let me know.

Point 1

(Flip) PLM is changing – where lies the new effort for (a new generation of) PLM experts.  I believe a huge effort for PLM is successful change management towards ‘business Agility.’ Since a proper response to an ECR/ECO would evidently require design changes impacting manufacturing and even after-sales and/or legal.  And that’s just the tip of the iceberg.

 

You are right, the main challenge for future PLM experts is to explain and support more agile processes, mainly because software has become a major part of the solution. The classical, linear product delivery approach does not match the agile, iterative approach for software deliveries. The ECR/ECO process has been established to control hardware changes, in particular because there was a big impact on the costs. Software changes are extremely cheap and possible fast, leading to different change procedures. The future of PLM is about managing these two layers (hardware/software) together in an agile way. The solution for this approach is that people have to work in multi-disciplinary teams with direct (social) collaboration and to be efficient this collaboration should be done in a digital way.

A good article to read in this context is Peter Bilello’s article: Digitalisation enabled by product lifecycle management.

 

(Flip) What seems to be missing is an ‘Archetype’ of the ideal transformed organization. Where do PLM experts want to go with these businesses in practice? Personally, I imagine a business where DevOps is the standard, unique products have generic meta-data, personal growth is an embedded business process and supply chain related risks are anticipated on and mitigated through automated analytics. Do you know of such an evolved archetypal enterprise model?

I believe the ideal archetype does not exist yet. We are all learning, and we see examples from existing companies and startups pitching their story for a future enterprise. Some vendors sell a solution based on their own product innovation platform, others on existing platforms and many new vendors are addressing a piece of the puzzle, to be connected through APIs or Microservices. I wrote about these challenges in Microservices, APIs, Platforms and PLM Services.  Remember, it took us “old PLM experts” more than 10-15 years to evolve from PDM towards PLM, riding on an old linear trajectory, caught up by a new wave of iterative and agile processes. Now we need a new generation of PLM experts (or evolving experts) that can combine the new concepts and filter out the nonsense.

Point 2

(Flip) But then given point 2: ‘Model-based enterprise transformations,’ in my view, a key effort for a successful PLM expert would also be to embed this change mgt. as a business process in the actual Enterprise Architecture. So he/she would need to understand and work out a ‘business-ontology’ (Dietz, 2006) or similar construct which facilitates at least a. business processes, b. Change (mgt.) processes, c. emerging (Mfg.) technologies, d. Data structures- and flows, e. implementation trajectory and sourcing.

And then do this from the PLM domain throughout the organization per optimization.  After all a product-oriented enterprise revolves around the success of its products, so eventually, all subsystems are affected by the makeup of the product lifecycle. Good PLM is a journey, not a trip. Or, does a PLM expert merely facilitates/controls this enterprise re-design process? And, what other enterprise ontologism tools and methods do you know of?

Only this question could be a next future blog post. Yes, it is crucial to define a business ontology to support the modern flow of information through an enterprise. Products become systems, depending on direct feedback from the market. Only this last sentence already requires a redefinition of change processes, responsibilities. Next, the change towards data-granularity introduces new ways of automation, which we will address in the upcoming years. Initiatives like Industry 4.0 / Smart Manufacturing / IIoT all contribute to that. And then there is the need to communicate around a model instead of following the old documents path. Read more about it in Digital PLM requires a Model-Based Enterprise. To close this point:  I am not aware of anyone who has already worked and published experiences on this topic, in particular in the context of PLM.

 

Point 3

(Flip) Where to draw the PLM line in a digital enterprise? I personally think this barrier will vanish as Product Lifecycle Management (as a paradigm, not necessarily as a software) will provide companies with continuity, profitability and competitive advantage in the early 21st century. The PLM monolith might remain, but supported by an array of micro services inside and outside the company (next to IoT, hopefully also external data sets).

I believe there is no need to draw a PLM line. As Peter’s article: Digitalisation enabled by product lifecycle management already illustrated there is a need for a product information backbone along the whole (circular) lifecycle, where product information can interact with other enterprise platforms, like CRM, ERP and MES and BI services. Sometimes we will see overlapping functionality, sometimes we will see the need to bridge the information through Microservices. As long as these bridges are data-driven and do not need manual handling/transformation of data, they fit in the future, lean digital enterprise.

Conclusion:

This can be an ongoing dialogue, diving into detailed topics of a modern PLM approach. I am curious to learn from my readers, how engaged they are in this topic? Do you still take part in PLM dialogues or do you consume? Do you have “tips and tricks” for those who want to shape the future of PLM?


Let your voice be heard! (and give Flip a break)

 

observationComing back from holiday (a life without PLM), it is time to pick up blogging again. And like every start, the first step is to make a status where we are now (with PLM) and where PLM is heading. Of course, it remains a nopinion based on dialogues I had this year.

First and perhaps this is biased, there is a hype in LinkedIn groups or in the blogs that I follow, a kind of enthusiasm coming from OnShape and Oleg Shilovitsky´s new company OpenBOM: the hype of cloud services for CAD/Data Management and BOM management.

servicesTwo years ago I discussed at some PLM conferences, that PLM should not necessary be linked to a single PLM system. The functionality of PLM might be delivered by a collection of services, most likely cloud-based, and these services together providing support for the product lifecycle. In 2014 I worked with Kimonex, an Israeli startup that developed a first online BOM solution, targeting the early design collaboration. Their challenge was to find customers that wanted to start this unknown journey. Cloud-based meant real-time collaboration, and this is also what Oleg wrote about last week: Real-time collaborative edit is coming to CAD & PLM

Real-time collaboration is one of the characteristics of a digital enterprise, where thanks to the fact information is stored as data, information can flow rapidly through an organization. Data can be combined and used by anyone in the organization in a certain context. This approach removes the barriers between PLM and ERP. To my opinion, there is no barrier between PLM and ERP. The barrier that companies create exists because people believe PLM is a system, and ERP is a system. This is the way of (system) thinking is coming from the previous century.

So is the future about cloud-based, data-driven services for PLM?

To my opinion systems are still the biggest inhibitor for modern PLM. Without any statistical analysis based on my impressions and gut feelings, this is what I see:

  • The majority of companies that say the DO PLM, actually do PDM. They believe it is PLM because their vendor is a PLM company and they have bought a PLM system. However, in reality, the PLM system is still used as an advanced PDM system by engineering to push (sometimes still manual) at the end information into the well-established ERP system. Check with your company which departments are working in the PLM system – anyone beyond engineering ?
  • There is a group of companies that have implemented PLM beyond their engineering department, connecting to their suppliers in the sourcing and manufacturing phase. Most of the time the OEMs forces their suppliers to deliver data in an old-fashioned way or sometimes more advanced integrated in the OEM environment. In this case, the supplier has to work in two systems: their own PDM/PLM environment and the OEM environment. Not efficient, still the way traditional PLM vendors promote partner / supply chain integration .

This is an area where you might think that a services-based environment like OnShape or OpenBOM might help to connect the supply chain. I think to so too. Still, before that we reach this stage there are some hurdles to overcome:

Persistence of data

owning or sharingThe current generation in management of companies older than 20 years grew up with the fact that “owning data” is the only way to stay relevant in business. Even open innovation is a sensitive topic. What happens with data your company does not own because it is in the cloud in an environment you do not own (but share ?) . As long as companies insist on owning the data, a service-based PLM environment will not work.

A nice compromise at this time is ShareAspace from EuroStep. I wrote about ShareAspace last year when I attended sessions from Volvo Car (The weekend after PI Munich 2016) ShareAspace was used as a middleware to map and connect between two PLM/PDM environments. In this way, persistence of data remains. The ShareAspace data model is based on PLCS, which is a standard in the core industries. And standards are to my opinion the second hurdle for a services-based approach

Standards

A standard is often considered as overhead, and the reason for that is that often a few vendors dominate the market in a certain domain and therefore become THE standard. Similar to persistence of data, what is the value of data that you own but that you only can get access to through a vendor´s particular format?

Good for the short-term, but what about the long-term. (Most of the time we do not think about the long-term and consider interoperability problems as a given). Also, a services based PLM environment requires support for standards to avoid expensive interfaces and lack of long-term availability. Check in your company how important standards are when selecting and implementing PLM.

Conclusion

There is a nice hype for real-time collaboration through cloud solutions. For many current companies not good enough as there is a lot of history and the mood to own data. Young companies that discover the need for a modern services-based solution might be tempted to build such an environment. For these companies, the long-term availability might be a topic to study

Note: I just realized if you are interested in persistency and standards you should attend PDT 2016 on 9 & 1o November in Paris. Another interesting post just published from Lionel Grealou : Single Enterprise BOM: Utopia vs Dystopia is also touching this subject

NL-PLMAs a genuine Dutchman, I was able to spend time last month in the Netherlands, and I attended two interesting events: BIMOpen2015, where I was invited to speak about what BIM could learn from PLM (see Dutch review here) and the second event: Where engineering meets supply chain organized by two startup companies located in Yes!Delft an incubator place working close to the technical university of Delft (Dutch announcement here)

Two different worlds and I realized later, they potential have the same future. So let’s see what happened.

BIMopen 2015

bimopenBIMopen 2015 had the theme: From Design to Operations and the idea of the conference was to bring together construction companies (the builders) and the facility managers (the operators) and discuss the business value they see from BIM.

First I have to mention that BIM is a confusing TLA like PLM. So many interpretations of what BIM means. For me, when I talk about BIM I mean Building Information Management. In a narrower meaning, BIM is often considered as a Building Information Model – a model that contains all multidisciplinary information. The last definition does not deal with typical lifecycle operations, like change management, planning, and execution.

The BIMopen conference started with Ellen Joyce Dijkema from BDO consultants who addressed the cost of failure and the concepts of lean. Thinking. The high cost of failure is known and accepted in the construction industry, where at the end of the year profitability can be 1 % of turnover (with a margin of +/- 3 % – so being profitable is hard).

Lean thinking requires a cultural change, which according to Ellen Joyce is an enormous challenge, where according to a study done by Prof Dr. A. Cozijnsen there is only 19 % of chance this will be successful, compared to 40 % chance of success for new technology and 30 % of chance for new work processes.

succes

It is clear changing culture is difficult and in the construction industry it might be even harder. I had the feeling a large part of the audience did not grasp the opportunity or could find a way to apply it to their own world.

My presentation about what BIM could learn from PLM was similar. Construction companies have to spend more time on upfront thinking instead of fixing it later (costly). In addition thinking about the whole lifecycle of a construction, also in operations can bring substantial revenue for the owner or operator of a construction. Where traditional manufacturing companies take the entire lifecycle into account, this is still not understood in the construction industry.

This point was illustrated by the fact that there was only one person in the audience with the primary interest to learn what BIM could contribute to his job as facility manager and half-way the conference he still was not convinced BIM had any value for him.

PLMandBIM

A significant challenge for the construction industry is that there is no end-to-end ownership of data, therefore having a single company responsible for all the relevant and needed data does not exist. Ownership of data can result in legal responsibility at the end (if you know what to ask for) and in a risk shifting business like the construction industry companies try to avoid responsibility for anything that is not directly related to the primary activities.

Some larger companies during the conference like Ballast Nedam and HFB talked about the need to have a centralized database to collect all the data related to a construction (project). They were building these systems themselves, probably because they were not aware of PLM systems or did not see through the first complexity of a PLM system, therefore deciding a standard system will not be enough.

whyworryI believe this is short-term thinking as with a custom system you can get quick results and user acceptance (it works the way the user is asking for) however custom systems have always been a blockage for the future after 10-15 years as they are developed with a mindset from that time.

If you want to know, learn more about my thoughts have a look at 2014 the year the construction industry did not discover PLM. I will write a new post at the end of the year with some positive trends. Construction companies start to realize the benefits of a centralized data-driven environment instead of shifting documents and risks.

The cloud might be an option they are looking for. Which brings me to the second event.

Engineering meets Supply Chain

This was more an interactive workshop / conference where two startups KE-Works and TradeCloud illustrated the individual value of their solution and how it could work in an integrated way. I had been in touch with KE-Works before because they are an example of the future trend, platform-thinking. Instead of having one (or two) large enterprise system(s), the future is about connecting data-centric services, where most of them can run in the cloud for scalability and performance.

KE-Works provides a real-time workflow for engineering teams based on knowledge rules. Their solution runs in the cloud but connects to systems used by their customers. One of their clients Fokker Elmo explained how they want to speed up their delivery process by investing in a knowledge library using KE-works knowledge rules (an approach the construction industry could apply too)

image

In general if you look at what KE-works does, it is complementary to what PLM-systems or platforms do. They add the rules for the flow of data, where PLM-systems are more static and depend on predefined processes.

tradecloudTradeCloud provides a real-time platform for the supply chain connecting purchasing and vendors through a data-driven approach instead of exchanging files and emails. TradeCloud again is another example of a collection of dedicated services, targeting, in this case, the bottom of the market. TradeCloud connects to the purchaser’s ERP and can also connect to the vendor’s system through web services.

The CADAC group, a large Dutch Autodesk solution provided also showed their web-services based solution connecting Autodesk Vault with TradeCloud to make sure the right drawings are available. The name of their solution, the “Cadac Organice Vault TradeCloud Adapter” is more complicated than the solution itself.

observationWhat I saw that afternoon was three solutions providers connected using the cloud and web services to support a part of a company’s business flow. I could imagine that adding services from other companies like OnShape (CAD in the cloud), Kimonex (BOM Management for product design in the cloud) and probably 20 more candidates can already build and deliver a simplified business flow in an organization without having a single, large enterprise system in place that connects all.

The Future

InnovDilemmaI believe this is the future and potential a breakthrough for the construction industry. As the connections between the stakeholders can vary per project, having a configurable combination of business services supported by a cloud infrastructure enables an efficient flow of data.

As a PLM expert, you might think all these startups with their solutions are not good enough for the real world of PLM. And currently they are not – I agree. However disruption always comes unnoticed. I wrote about it in 2012 (The Innovators Dilemma and PLM)

Conclusion

Innovation happens when you meet people, observe and associate in areas outside your day-to-day business. For me, these two events connected some of the dots for the future. What do you think? Will a business process based on connected services become the future?

Sometimes we have to study careful to see patterns have a look here what is possible according to some scientists (click on the picture for the article)

 

image

PDT2015-1In this post observations from the PDT 2015 conference which took place in the IVA Conference Center, part of the Royal Swedish Academy of Engineering Services in Stockholm.

The conference was hosted by Eurostep supported by CIMdata, Airbus, Siemens Energy and Volvo AB.

For me, the PDT conference is interesting because there is a focus on architecture and standards flavored with complementary inspiring presentations. This year there were approximate 110 participants from 12 countries coming from different industries listening to 25 presentations spread over two days.

Some highlights

cimdataPeter Bilello from CIMdata kicked off the conference with his presentation: The Product Innovation Platform: What’s Missing.

Peter explained how the joined vision from CIMdata, Gartner and IDC related to a product innovation platform is growing.

The platform concept is bringing PLM to the enterprise level as a critical component to support innovation. The main challenge is to make the complex simple – easier said than done, but I agree this is the real problem of all the software vendors.

Peter showed an interesting graph based on a survey done by CIMdata, showing two trends.

  • The software and technology capabilities are closing more and more the gap with the vision (a dream can come true)
  • The gap between the implemented capabilities and the technical possible capabilities is growing too. Of course, there is a difference between the leaders and followers.

Peter described the three success factors determining if a platform can be successful:

  • Connection: how easy is it for others to connect and plug into the platform to participate as part of the platform. Translated to capabilities this requires the platform to support open standards to connect external data sources as you do not want to build new interfaces for every external source. Also, the platform provider should provide an integration API with a low entry level to get the gravity (next point)
  • Gravity: how well does the platform attract participants, both producers, and consumers. Besides a flexible and targeted user interfaces, there must be an infrastructure that allows companies to model the environment in such a manner that it supports experts creating the data, but also support consumers in data, who are not able to navigate through details and want a consumer-friendly environment.
  • Flow: how well does the platform support the exchange and co-creation of value. The smartphone platforms are extremely simple compared to a business platform as the dimension of lifecycle status and versioning is not there. A business platform needs to have support for versioning and status combined with relating the information in the right context. Here I would say only the classical PLM vendors have in-depth experience with that.

Having read these three bullet points and taking existing enterprise software vendors for PLM, ERP, and other “platforms” in mind, you see there is still a way to go before we have a “real” platform available.

According to Peter, companies should start with anchoring the vision for a business innovation platform in their strategic roadmap. It will be an incremental journey anyway. How clear the vision is connected to business execution in reality differentiates leaders and followers.

gartner

 

Next Marc Halpern from Gartner elaborated on enabling Product Innovation Platforms. Marc started to say that the platform concept is still the process of optimizing PLM.

Marc explained the functional layers making up a product innovation platform, see below

 

Gartner-platform layering

According to Marc, in 2017 the major design, PLM and business suite vendors will all offer product innovation platforms, where certain industries are more likely to implement product innovation platforms faster than others.

Marc stressed that moving to a business innovation platform is a long, but staged, journey. Each stage of the journey can bring significant value.

Gartner has a 5-step maturity model based on the readiness of the organization. Moving from reactive, repeatable, integrating towards collaborating and ultimately orchestrating companies become business ready for PDM first, next PLM and the Product Innovation Platform at the end. You cannot skip one of these steps according to Marc. I agree, PLM implementations in the past failed because the company was dreaming that the PLM system would solve the business readiness of the organization.

Marc ended with a case study and the conclusions were not rocket science.

The importance of change management, management understanding and commitment, and business and IT joined involvement. A known best practice, still we fail in many situations to act accordingly, due to underestimation of the effort. See also my recent blog post: The importance of change management for PLM.

peepoople logoNext session from Camilla Wirseen was a real revelation. Her presentation:  We are all Peepoople – innovation from the bottom of the pyramid.

She described how Anders Wilhemson, original a professor in architecture, focused on solving a global, big problem addressing 2.5 billion people in the world. These 2.5 billion persons, the poorest of the world, lack sanitation, which results in a high death rate for children (every 15 seconds a child dies because of contaminated water). Also the lack of safe places for sanitation lead to girls dropping out of school and women and children being at risk for rape when going to toilet places.

The solution is a bag, made of high-performance biodegradable plastics combined with chemicals, already in the bag, processing the feces to kill potential diseases and make the content available as fertilizer for the agricultural industry.

The plastic bag might not be new, but adding the circular possibilities to it, make it a unique approach to creating a business model providing collection and selling of the content again. For the poorest every cent they can earn makes a different.

peepoople statement

Currently in initial projects the Peepoo system has proven its value: over 95 % user acceptance. It is the establishment that does not want to introduce Peepoo on a larger scale. Apparently they never realized themselves the problems with sanitation.

Peepoo is scaling up and helping the bottom of our society. And the crazy fact is that it was not invented by engineers but by an architect. This is challenging everyone to see where you can contribute to a better world. Have a look at peepoople.cominnovation with an enormous impact!

volvologoNext Volvo Cars and Volvo Trucks presented similar challenges: How to share product data based on external collaboration. The challenge of Volvo Cars is that it has gone through different ownerships and they require a more and more flexible infrastructure to share data. It is not about data pushing to a supplier anymore, it is about integrating partners where you have to share a particular part of your IP with the partner. And where the homegrown KPD system is working well for internal execution, it was never designed for partner sharing and collaboration. Volvo Cars implemented a Shared Technology Control application outside the firewall based on Share-A-space, where inside and outside data is mapped and connected. See their summary below. A pragmatic approach which is bringing direct benefits.

clip_image002[10]

Concluding from the Volvo sessions: Apparently it ‘s hard to extend an existing system or infrastructure for secure collaboration with an external partner. The complexity of access right, different naming conventions, etc. Instead of that it is more pragmatic to have an intermediate system in the middle, like Share-A-space, that connects both worlds. The big advantage of Share-A-space is that the platform is based on the ISO 10303 (PLCS) standard and, therefore, has one of the characteristics of a real platform: openness based on standards.

awesomegroupJonas Hammerberg from the Awesome Group closed day one with an inspiring and eye-opening presentation: Make PLM – The Why and How with Gamification FUN.

Jonas started to describe the behavioral drivers new generations have based on immediate feedback for the feeling of achievement, pride and status and being in a leading environment combined with the feelings of being in a group feeling friendship, trust, and love.

Current organizations are not addressing these different behaviors, it leads to disengagement at the office / work floor as Jonas showed from a survey held in Sweden – see figure. The intrinsic motivation is missing. One of the topics that concerns me the most when seeing current PLM implementations.

engagement

The Awesome group has developed apps and plug-ins for existing software, office and PLM bring in the feelings of autonomy, mastery and purpose to the individual performing in teams. Direct feedback and stimulating team and individual performance as part of the job.

By doing so the organization also gets feedback on the behavior, activity, collaboration and knowledge sharing of individuals and how this related to their performance. An interesting concept to be implemented in situations where gamification makes sense.

clip_image002[12]Owe Lind and Magnus Lidström from Scania talked about their Remote Diagnostics approach where diagnostic readings can be received from a car through a mobile phone network either to support preventive maintenance or actual diagnostics on the road and provide support.

Interesting Owe and Magnus were not using the word IoT (Internet of Things) at all, a hype related to these capabilities. Have a look here on YouTube

clip_image002[14]There was no chance to fall asleep after lunch, where Robin Teigland from the Stockholm School of Economics took us in a whirlwind through several trends under the title: The Third Revolution – exploring new forms of value creation through doing more with less.

The decomposition of traditional business into smaller and must faster communities undermine traditional markets. Also concepts like Uber, Bitcoin becoming a serious threat. The business change as a result of connectivity and communities leading to more and more networks of skills bringing together knowledge to design a car (Local Motors), funding (Kickstarter) – and it is all about sharing knowledge instead of keeping it inside – sharing creates the momentum in the world. You can look at Robin’s presentation(s) at Slideshare here.

future quote

All very positive trends for the future, however, a big threat to the currently established companies. Robin named it the Third Revolution which is in line with what we are discussing in our PLM world, although some of us call it even the Fourth Revolution (Industry 4.0).

image

EignerProfessor Martin Eigner from the Technical University of Kaiserslautern brought us back to reality in his presentation: Industry 4.0 or Industrial Internet: What is the impact for PLM?

Martin stood at the base for what we call PLM and already for several years he is explaining to us that the classical definition for PLM is too narrow. More and more we are developing systems instead of products. Therefore, he prefers the abbreviation SysLM, which is more than 3 characters and therefore probably hard to accept by the industry.

PDMtoSysLM

System development and, therefore, multidisciplinary development of systems introduces a new complexity. Traditional change management for Mechanical CAD (ECO/ECR) is entirely different from how software change management is handled (baselines / branches related to features). The way systems are designed, require a different methodology where systems engineering is an integral part of the development process, see Model-Based Systems Engineering (MBSE).

Next Martin discussed 4 potential IT-architectures where, based on the “products” and business needs, a different balance of PLM, ALM or ERP activities is required.

Martin’s final point was about the need for standards support these architectures, bringing together OSLC, PCLS, etc.
Standards are necessary for fast and affordable integrations and data exchange.

imageMy presentation: The Perfect Storm or a fatal Tsunami was partly summarizing topics from the conference and, in addition, touching on two topics.

The first topic is related to big data and analytics.  Many are trying to get a grip on big data with analytics. However, the real benefit of big data comes when you are able to apply algorithms to it. Gartner just made an interesting statement related to big data (below) and Marc Halpern added to this quote that there is an intrinsic need for data standards in order to apply algorithms.

Gartner algorithms

When algorithms can be used, classical processes like ECO, ECR or managers might become obsolete and even a jobs like an accountant is at risk. This as predicted in article in the Economist in February 2014 – the onrushing Wave

The second topic, where I believe we are still hesitating too long at management level, is making decisions, to anticipate the upcoming digital wave and all of its side effects. We see a huge wave coming. If we do not mobilize the people, this wave might be a tsunami for those still at the seaside

Conclusion: PDT2015 was an inspiring, well-balanced conference with excellent opportunity to network with all people attending. For those interested in the details of the PLM future and standards an ideal opportunity to get up to date. And next the challenge: Make it happen at your company!

.. if you reach this point, my compliments for your persistency to read it all. Too long for a blog post and even here I had to strip

 

clip_image002Three weeks ago there was the Product Innovation conference in Düsseldorf. In my earlier post (here) I described what I experienced during this event. Now, after all the information is somehow digested, here a more high-level post, describing the visible change in business and how it relates to PLM. Trying to describe this change in non-academic wording but in images. Therefore, I described the upcoming change in the title: from linear to circular and fast.

 

Let me explain this image step by step

In the middle of the previous century, we were thinking linear in education and in business. Everything had a predictable path and manufacturing companies were pushing their products to the market. First local, later in time, more global. Still the delivery process was pretty linear:

clip_image003

This linear approach is reflected in how organizations are structured, how they are aligned to the different steps of the product development and manufacturing process. Below a slide I used at the end of the nineties to describe the situation and the pain; lack of visibility what happens overall.

clip_image005

It is discouraging to see that this situation still exists in many companies.

At the end of the nineties, early 2000, PLM was introduced, conceptually managing the whole lifecycle. In reality, it was mainly a more tight connection between design and manufacturing preparation, pushing data into ERP. The main purpose was managing the collaboration between different design disciplines and dispersed teams.

Jim Brown (Tech-Clarity) wrote at that time a white paper, which is still valid for many businesses, describing the complementary roles of PLM and ERP. See the picture below:

clip_image007

Jim introduced the circle and the arrow. PLM: a circle with iterations, interacting with ERP: the arrow for execution. Here visual it became already clear an arrow does not have the same behavior as a circle. The 100 % linearity in business was gone.

Let´s have a closer look at the PLM circle

This is how PLM is deployed in most organizations:

clip_image009Due to the implementation of siloed systems for PDM, ERP, SCM and more, the flow of information is disconnected when moving from the design domain to the execution domain.

Information is pushed in the ERP system as disconnected information, no longer managed and connected to its design intent.

Next, the ERP system is most of the time not well-equipped for managing after sales and services content. Another disconnect comes up.

Yes, spare parts could be ordered through ERP, but issues appearing at the customer base are not stored in ERP, often stored in a separate system again (if stored beyond email).

The result is that when working in the concept phase, there is no information available for R&D to have a good understanding of how the market or customers work with their product. So how good will it be? Check in your company how well your R&D is connected with the field?

And then the change started …

This could have stayed reality for a long time if there were not a huge business change upcoming. The world becomes digital and connected. As a result, local inefficiencies or regional underperformance will be replaced by better-performing companies. The Darwin principle. And most likely the better performing companies are coming from the emerging markets as there they do not suffer from the historical processes and “knowledge of the past”. They can step into the digital world much faster.

clip_image011In parallel with these fast growing emerging markets, we discovered that we have to reconsider the ways we use our natural resources to guarantee a future for next generations. Instead of spilling resources to deliver our products, there is a need to reuse materials and resources, introducing a new circle: the circular economy.

The circular economy can have an impact on how companies bring products to the market. Instead of buying products (CAPEX) more and more organizations (and modern people) start using products or services in a rental model (OPEX). No capital investment anymore, pay as you go for usage or capacity.

clip_image013This, however, has an impact how traditional companies are organized – you need to be connected to your customers or you are out of business – a commodity.

The digital and connected world can have a huge impact on the products or services available in the near future. You are probably familiar with the buzz around “The Internet of Things” or “Smart and Connected”.

No longer are products depending on mechanical behavior only, more and more products are relying on electrical components with adaptive behavior through software. Devices that connect with their environment report back information to the manufacturer. This allows companies to understand what happens with their products in the field and how to react on that.

Remember the first PLM circle?
Now we can create continuity of data !

PLM_flowCombine the circular economy, the digital and connected world and you will discover everything can go much faster. A crucial inhibitor is how companies can reorganize themselves around this faster changing, circular approach. Companies need to understand and react to market trends in the fastest and adequate way. The future will be probably about lower volumes of the same products, higher variability towards the market and most likely more and more combining products with services (the Experience Model). This requires a flexible organization and most likely a new business model which will differ from the sequential, hierarchical organizations that we know at this moment.

The future business model ?

The flexibility in products and services will more and more come from embedded software or supported by software services. Software services will be more and more cloud based, to avoid IT-complexity and give scalability.

Software development and integration with products and services are already a challenge for classical mechanical companies. They are struggling to transform their mechanical-oriented design process towards support for software. In the long-term, the software design process could become the primary process, which would mean a change from (sequential – streamlined) lean towards (iterative – SCRUM) agile.

Once again, we see the linear process becoming challenged by the circular iterations.

This might be the end of lean organizations, potentially having to mix with agile conepts..

clip_image015If it was a coincidence or not, I cannot judge, however during the PI Conference I learned about W.L. Gore & Associates, with their unique business model supporting this more dynamic future. No need to have a massive organization re-org to align the business, as the business is all the time aligning itself through its employees.

Last weekend, I discovered Semco Partners in the newspaper and I am sure there are more companies organizing themselves to become reactive instead of linear – for sure in high-tech world.

Conclusion:

Linearity is disappearing in business, it is all about reactive, multidisciplinary teams within organizations in order to support customers and their fast changing demands.

Fast reactions need new business organizations models (flexible, non-hierarchical) and new IT-support models (business information platforms – no longer PLM/ERP system thinking)

What do you think ? The end of linear ?

 

I have talked enough about platforms recently. Still if you want to read more about it:

Cimdata: Business strategy and platformization position paper

Engineering.com: Prod. Innovation Platform PlugnPlay in next generation PLM

Gartner: Product Innovation Platforms

VirtualDutchman: Platform, Backbone, Service Bus or BI

%d bloggers like this: