Welcome!

@CloudExpo Authors: Liz McMillan, Elizabeth White, Matt Brickey, Rene Buest, Mehdi Daoudi

Related Topics: @DevOpsSummit, Open Source Cloud, @CloudExpo

@DevOpsSummit: Blog Post

Debunking the OSLC Link-Only Myth | @DevOpsSummit #API #Agile #DevOps

Open Services for Lifecycle Collaboration is an open community for creating specifications for integrating lifecycle activities

Most large organizations require dozens and sometimes hundreds of specialized software tools to manage the lifecycle of the physical products or software applications they create. It isn't hard to imagine the monumental waste these organizations incur in attempting to manually coordinate the efforts of the teams that use these many disparate tools to create a single product. Open Services for Lifecycle Collaboration (OSLC) is an open community for creating specifications for integrating lifecycle activities across tools to address this problem. Now imagine how much more happy and productive an organization would be if all those tools could integrate via standard interfaces. In broad strokes, this is the goal of the OSLC community.

The technical foundation for OSLC is based on the W3C Linked Data method of publishing and accessing information and inspired by how the World Wide Web works. This "Linked Data" foundation is often misinterpreted to mean that data should only be linked across systems and never replicated or modified. The reality is that OSLC fully supports Create, Read, Update and Delete (CRUD) operations on lifecycle artifacts such as stories, defects and tests in a way that enables them to be integrated across systems through replication. Those artifacts, however, should of course be linked with one another across systems using OSLC links for navigation and traceability across systems. See the OSLC Primer for more detail.

While replication is fully supported by the technical underpinnings of OSLC, it is a widely held best practice within the community that data should only be replicated sparingly and only when necessary. Replicating data is inherently complicated and can lead to data inconsistencies if not done properly with a sophisticated toolset that must be correctly configured. However, the "C" in OSLC stands for "Collaboration" and the point of OSLC is to facilitate collaboration in the ways that create the most business value rather than to dictate a specific integration philosophy. And this means that replicating data with proxy object representations is OK when necessary to achieve the business outcome. In this case, a "proxy object" means a representation of a lifecycle artifact stored in another repository. Proxy objects should only contain values needed for local context and have an OSLC reference to the original object to apply services (e.g., CRUD).

Consider, for example, a software delivery team where the developers are using an Agile planning tool for software project management and the testers on the QA team have selected a specialized quality management tool to manage their work. The developers and testers require integration so that testers know what features and stories to test, while developers know which identified defects must be fixed.

To satisfy this integration need, Stories in the Agile planning tool can be represented as proxy objects (Requirements) within the quality management tool. Links are used to connect these two artifacts across systems for traceability. This allows the tester using the quality management tool to read the requirement and create the corresponding test cases to test it. Furthermore, the internal link between the requirement and the test case within the quality management tool enables traceability and reporting so the tester can verify that tests have corresponding requirements and vice versa.

Now when the tester inevitably finds a defect in the software, that defect report is filed in the tester's quality management system of choice and linked to the corresponding test case. But this doesn't get the defect fixed. The next step is for a proxy representation of the defect in the quality management tool to be created in the Agile planning tool and linked back to the original defect. This enables the developer to schedule the defect to be corrected in an upcoming Agile sprint.

As this example illustrates, addressing the business need for integration can require a combination of linking and replication techniques. While the information that is replicated should be minimized as much as possible, this approach is fully supported by OSLC. While Linked Data techniques are a crucial part of the solution and a foundation of OSLC, the notion that OSLC is only about links is just a myth.

More Stories By Wesley Coelho

Wesley Coelho is the Senior Director of Business Development at Tasktop Technologies. At Tasktop, he has taken a leading role in creating an extensive network of ALM integration partners including ten OEM distribution partnerships with companies such as IBM, HP, and CA Technologies.

Prior to Tasktop, Wesley led a software development team pioneering a new generation of eCommerce technology at Elastic Path Software. Wesley holds an undergraduate degree in Commerce from the University of Victoria and an M.Sc. in Computer Science from the University of British Columbia (Canada).

More Stories By Rainer Ersch

Rainer Ersch is a Senior Research Scientist with more than 33 years industry experience. He has worked as a director of SW Development Tools Enterpise Communications as well as serving as a consultant and coach for System and Software Development Environments. He has been active with the OSLC community as a member of the Steering Committee for over two years and as the workgroup lead of the ALM-PLM Interoperability Workgroup.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@CloudExpo Stories
"We focus on SAP workloads because they are among the most powerful but somewhat challenging workloads out there to take into public cloud," explained Swen Conrad, CEO of Ocean9, Inc., in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
"We are still a relatively small software house and we are focusing on certain industries like FinTech, med tech, energy and utilities. We help our customers with their digital transformation," noted Piotr Stawinski, Founder and CEO of EARP Integration, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
"I think DevOps is now a rambunctious teenager – it’s starting to get a mind of its own, wanting to get its own things but it still needs some adult supervision," explained Thomas Hooker, VP of marketing at CollabNet, in this SYS-CON.tv interview at DevOps Summit at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
"We've been engaging with a lot of customers including Panasonic, we've been involved with Cisco and now we're working with the U.S. government - the Department of Homeland Security," explained Peter Jung, Chief Product Officer at Pulzze Systems, in this SYS-CON.tv interview at @ThingsExpo, held June 6-8, 2017, at the Javits Center in New York City, NY.
"We're here to tell the world about our cloud-scale infrastructure that we have at Juniper combined with the world-class security that we put into the cloud," explained Lisa Guess, VP of Systems Engineering at Juniper Networks, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
"With Digital Experience Monitoring what used to be a simple visit to a web page has exploded into app on phones, data from social media feeds, competitive benchmarking - these are all components that are only available because of some type of digital asset," explained Leo Vasiliou, Director of Web Performance Engineering at Catchpoint Systems, in this SYS-CON.tv interview at DevOps Summit at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
Your homes and cars can be automated and self-serviced. Why can't your storage? From simply asking questions to analyze and troubleshoot your infrastructure, to provisioning storage with snapshots, recovery and replication, your wildest sci-fi dream has come true. In his session at @DevOpsSummit at 20th Cloud Expo, Dan Florea, Director of Product Management at Tintri, provided a ChatOps demo where you can talk to your storage and manage it from anywhere, through Slack and similar services with...
As enterprise cloud becomes the norm, businesses and government programs must address compounded regulatory compliance related to data privacy and information protection. The most recent, Controlled Unclassified Information and the EU’s GDPR have board level implications and companies still struggle with demonstrating due diligence. Developers and DevOps leaders, as part of the pre-planning process and the associated supply chain, could benefit from updating their code libraries and design by in...
"Peak 10 is a hybrid infrastructure provider across the nation. We are in the thick of things when it comes to hybrid IT," explained Michael Fuhrman, Chief Technology Officer at Peak 10, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
SYS-CON Events announced today that Calligo, an innovative cloud service provider offering mid-sized companies the highest levels of data privacy and security, has been named "Bronze Sponsor" of SYS-CON's 21st International Cloud Expo ®, which will take place on Oct 31 - Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Calligo offers unparalleled application performance guarantees, commercial flexibility and a personalised support service from its globally located cloud plat...
"We were founded in 2003 and the way we were founded was about good backup and good disaster recovery for our clients, and for the last 20 years we've been pretty consistent with that," noted Marc Malafronte, Territory Manager at StorageCraft, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
"We are an IT services solution provider and we sell software to support those solutions. Our focus and key areas are around security, enterprise monitoring, and continuous delivery optimization," noted John Balsavage, President of A&I Solutions, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
There is a huge demand for responsive, real-time mobile and web experiences, but current architectural patterns do not easily accommodate applications that respond to events in real time. Common solutions using message queues or HTTP long-polling quickly lead to resiliency, scalability and development velocity challenges. In his session at 21st Cloud Expo, Ryland Degnan, a Senior Software Engineer on the Netflix Edge Platform team, will discuss how by leveraging a reactive stream-based protocol,...
"We are focused on SAP running in the clouds, to make this super easy because we believe in the tremendous value of those powerful worlds - SAP and the cloud," explained Frank Stienhans, CTO of Ocean9, Inc., in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
"DivvyCloud as a company set out to help customers automate solutions to the most common cloud problems," noted Jeremy Snyder, VP of Business Development at DivvyCloud, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
DevOps at Cloud Expo, taking place October 31 - November 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with 21st Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to w...
"As we've gone out into the public cloud we've seen that over time we may have lost a few things - we've lost control, we've given up cost to a certain extent, and then security, flexibility," explained Steve Conner, VP of Sales at Cloudistics,in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
Internet of @ThingsExpo, taking place October 31 - November 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with 21st Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. The Internet of Things (IoT) is the most profound change in personal and enterprise IT since the creation of the Worldwide Web more than 20 years ago. All major researchers estimate there will be tens of billions devic...
"The Striim platform is a full end-to-end streaming integration and analytics platform that is middleware that covers a lot of different use cases," explained Steve Wilkes, Founder and CTO at Striim, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
"We want to show that our solution is far less expensive with a much better total cost of ownership so we announced several key features. One is called geo-distributed erasure coding, another is support for KVM and we introduced a new capability called Multi-Part," explained Tim Desai, Senior Product Marketing Manager at Hitachi Data Systems, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.