Arnaud’s Open blog

Opinions on open source and standards

More on Linked Data and IBM

For those technically inclined, you can learn more about IBM’s interest in Linked Data as an application integration model and the kind of standard we’d like the W3C Linked Data Platform WG to develop by reading a paper I presented earlier this year at the WWW2012 Linked Data workshop titled: “Using read/write Linked Data for Application Integration — Towards a Linked Data Basic Profile”.

Here is the abstract:

Linked Data, as defined by Tim Berners-Lee’s 4 rules [1], has
enjoyed considerable well-publicized success as a technology for
publishing data in the World Wide Web [2]. The Rational group in
IBM has for several years been employing a read/write usage of
Linked Data as an architectural style for integrating a suite of
applications, and we have shipped commercial products using this
technology. We have found that this read/write usage of Linked
Data has helped us solve several perennial problems that we had
been unable to successfully solve with other application
integration architectural styles that we have explored in the past.
The applications we have integrated in IBM are primarily in the
domains of Application Lifecycle Management (ALM) and
Integration System Management (ISM), but we believe that our
experiences using read/write Linked Data to solve application
integration problems could be broadly relevant and applicable
within the IT industry.
This paper explains why Linked Data, which builds on the
existing World Wide Web infrastructure, presents some unique
characteristics, such as being distributed and scalable, that may
allow the industry to succeed where other application integration
approaches have failed. It discusses lessons we have learned along
the way and some of the challenges we have been facing in using
Linked Data to integrate enterprise applications.
Finally, we discuss several areas that could benefit from
additional standard work and discuss several commonly
applicable usage patterns along with proposals on how to address
them using the existing W3C standards in the form of a Linked
Data Basic Profile. This includes techniques applicable to clients
and servers that read and write linked data, a type of container
that allows new resources to be created using HTTP POST and
existing resources to be found using HTTP GET (analogous to
things like Atom Publishing Protocol (APP) [3]).

The full article can be found as a PDF file: Using read/write Linked Data for Application Integration — Towards a Linked Data Basic Profile

September 11, 2012 Posted by | linkeddata, standards | , , , | Leave a comment

Linked Data

Several months ago I edited my “About” text on this blog to add that: “After several years focusing on strategic and policy issues related to open source and standards, including in the emerging markets, I am back to more technical work.”

One of the projects that I have been working on in this context is Linked Data.

It all started over a year ago when I learned from the IBM Rational team that Linked Data was the foundation of Open Services for Lifecycle Collaboration Lifecycle (OSLC) which Rational uses as their platform for application integration. The Rational team was very pleased with the direction they were on but reported challenges in using Linked Data. They were looking for help in addressing these.

Fundamentally, the crux of the challenges they faced came down to a lack of formal definition of Linked Data. There is plenty of documentation out there on Linked Data but not everyone has the same vision or definition. The W3C has a growing collection of standards related to the Semantic Web but not everyone agrees on how they should be used and combined, and which one applies to Linked Data.

The problem with how things stand isn’t so much that there isn’t a way to do something. The problem is rather that, more often than not, there are too many ways. This means users have to make choices all the time. This makes starting to use Linked Data difficult for beginners and it hinders interoperability because different users make different choices.

I organized a teleconference with the W3C Team in which we explained what IBM Rational was doing with Linked Data and the challenges they were facing. The W3C team was very receptive to what we had to say and offered to organize a workshop to discuss our issues and see who else would be interested.

The Linked Enterprise Data Patterns Workshop took place on December 6 and 7, 2011 and was well attended. After a day and a half of presentations and discussions the participants found themselves largely agreeing and unanimously concluded that: the W3C should create a Working Group to create a Recommendation that formally defines a “Linked Data Platform”.

The workshop was followed by a submission by IBM and others of the Linked Data Basic Profile and the launch by W3C of the Linked Data Platform (LDP) Working Group (WG) which I co-chair.

You can learn more about this effort and IBM’s position by reading the “IBM on the Linked Data Platform” interview the W3C posted on their website and reading the “IBM lends support for Linked Data standards through W3C group” article I published on the Rational blog.

On a personal level, I’ve known about the W3C Semantic Web activities since my days as a W3C Team Member but I had never had the opportunity to work in this space before so I’m very excited about this project. I’m also happy to be involved again with the W3C where I still count many friends. 🙂

I will try to post updates on this blog as the WG makes progress.

September 10, 2012 Posted by | ibm, linkeddata, standards | , , , | Leave a comment

Open standards and globalization

Among the standards principles IBM announced on September 23rd, there is one that is particularly dear to me (surely because of my current responsibilities at IBM but also because of my background with W3C). This principle is what we refer to as the principle of “Global Application”. It reads:

Encourage emerging and developed economies to both adopt open global standards and to participate in the creation of those standards.

Despite what the OOXML proponents have been claiming for the sake of their own benefit, having multiple standards for the same task doesn’t do anybody any good. Multiple standards means market fragmentation and duplication of efforts. Market fragmentation and duplication of efforts mean less choice and higher cost.

As we move forward we must learn from the past while not letting it get in our way. We must ensure that standards are developed in such a way that all stakeholders can participate and feel compelled to do so. This is essential for all requirements to be addressed but also for everybody to have a sense of ownership. Both of these elements are key to the adoption of the standard by all.

I consider the case of the Uniform Office Format (UOF) a perfect example of our failure to do just that. What was it that led China to create its own format rather than work with us on expanding ODF so that it addresses their needs? Their work started with a fork from OpenOffice mind you. So, why weren’t they at the table with us?

We need to understand what went wrong and ensure that this doesn’t happen again. For everybody’s benefit. Failure to so will result in more pain for everybody, just like the pain we are experiencing with UOF.

The situation with UOF is now that China is trying to gain support from vendors like IBM. These vendors would like to play in the Chinese market but they have already heavily invested in ODF and are understandly not too keen on the idea of spending resources on UOF. They would rather see China adopt ODF. But ODF doesn’t quite fit China’s needs. So, efforts are being made towards a possible convergence of the formats but these are merely damage control that remain costly for all.

And this is not all. The Global Application principle cannot be separated from the principle of “Implementability” which reads:

Collaborate with standards bodies and developer communities to ensure that open software interoperability standards are freely available and implementable.

Indeed, one of the major barriers to global adoption by developing countries of the so called “international standards” is the toll on implementing them. Whether it is about paying just to access the document or about paying royalties to foreign companies for patents that read on the standard, the price tag this constitutes is just not acceptable to emerging countries. They already face enough challenges otherwise.

The European Commission as well as countries like India are trying to move the ball by developing policies that restrict public procurement to “open standards” which they define as being royalty free. This is provoking reactions from various organizations that want to stop this movement. Their main contention appears to be that we’ve been developing standards for decades on a RAND basis and adopting a royalty free only policy will rule out hundreds of existing standards and products. I say: tough!

It’s about time that we recognize that the way we’ve been doing standards isn’t going to work anymore. And we just cannot expect the world to be shackled by the way we’ve been doing things in the past.

Traditionally, IT standards have for the most part been developed by the western world and then pushed onto the rest of the world. A RAND based system might have been fine in an environment where the odds were balanced by the fact that all parties had more or less similar stakes in the game. But this doesn’t work when you add a bunch of new players who find themselves at the table empty handed.

So, it’s not surprise that the rest of the world is telling us “No, thanks”. Can we really blame them?

Those that cling onto the old ways are part of the past. The future simply cannot be based on a grossly unbalanced system that gives a hudge advantage to some parties. Getting rid of the toll on implementating standards is the price to pay to see them globally adopted. Failures to recognize that simple fact and attempts to derail the trends set by the European Commission and the likes are simply a waste of time.

October 30, 2008 Posted by | standards | , , , | 1 Comment

Rave: IBM support

I don’t usually talk much about what IBM does but today I’m going to.

People I work with know that I’m usually pretty vocal about what I think the company doesn’t do well but if there is one thing that IBM does really well in my experience is support.

I’ve been at IBM for close to 9 years now and this has been true all along. Whenever some piece of material breaks down it gets fixed extremely quickly and painlessly.

I just had a couple of such experiences. Last week a disk on one of my computers started to fail. I called support in the afternoon and the very next morning a new disk was delivered to my door. I work from home.

Most defective items have to be returned. But IBM support makes that really easy. With the new item comes a label to ship the defective part back. All you have to do is place the old part in the package, stick the label on top of the old one, and order the pick up either by phone or over the web. In some cases, the box even contains a piece of tape for you to use. Can’t be easier.

Last Friday my son managed to pour a glass of water over my keyboard. I called support and this Monday morning I received a new keyboard.

In cases when the whole computer has to be sent, you first receive a shipping box – overnight – to send it in. Typically two days later you get it back repaired. The turnover can’t be better.

I should point out that this is no special services for employees. This is the exact same service anybody gets.

Anyway, I can’t say I have much experience with many other companies when it comes to support but I have never seen anything getting close to what IBM does in this regard.

I’ve actually realized that I like that so much that it’s often something I bring up when people ask me what it’s like to work at IBM.

So, it’s only fair that I should give the company credit for that. 🙂

September 15, 2008 Posted by | ibm | , | Leave a comment

Let’s be clear: The Apache Software Foundation does NOT support OOXML.

OK, I’ll admit that nobody has claimed otherwise. Yet.

But in these days and age you are never too prudent. It wouldn’t surprise me to see this or other similar fancy claim being published eventually.

Indeed, in its desperate and last minute attempts to convince National Bodies around the world that OOXML is happening anyway so they might as well support it as an ISO standard, Microsoft is eager to claim support by as many companies and organizations as possible.

As evidence, in its latest OOXML propaganda open letter Microsoft lists IBM among other companies as having “already adopted (or announced adoption of) Open XML in their products”. This, despite a clear explanation of the contrary by Rob Weir, published two months ago! Does anyone believe they haven’t seen it or heard about this? I sure don’t. And if there was any room for misunderstanding Bob Sutor’s statement filled that in.

A colleague in a foreign country even reported that in a National Body meeting he had been confronted by a representative from Microsoft who was trying to silence him via intimidation and insistence that IBM supported OOXML contrary to what he was saying.

Microsoft’s oversight of IBM’s denials is clearly not accidental. It is part of a well crafted and continuous disingenuous plan to convince NBs at all cost. There is already so much evidence of Microsoft going far beyond what most would consider normal lobbying behavior it is sickening. For one, I’m not ready to forget the case of the NGOs in India. Talk about dirty practices.

But what really is at the bottom of Microsoft’s claims is that basically any software that handles XML supports OOXML. While technically this is true to a certain degree, such a bold claim without any further qualification is pure misinformation. Obviously, one of the advantages of using XML is to make your format, whatever it is, easier to handle, it’s one of the fundamental benefits of using XML. But as I previously touched on in my entry on XML vs Open, there is a big difference between being able to handle XML files at the XML level and truly supporting the particular format at hand.

Supporting OOXML. cannot be merely declared on the sole basis that a software can read OOXML files, or store OOXML files. If that were the case, then any XML parser could be said to support OOXML and the Apache Software Foundation could be said to support OOXML because its XML parser, Xerces, can read OOXML files (one would actually have to unzip them first but it’s not like Microsoft would stop at that kind of detail). But it takes much more than that to really support OOXML.

One has to understand the actual structure beyond the XML representation and the semantic associated to each and every piece of data found in an OOXML file. That’s what the 6000+ pages of the specification are supposed to define, unfortunately they do that extremely poorly.

The good news is that I don’t think Microsoft is fooling that many people. Based on my own observation of Microsoft representatives and the way they talk to people they seem to be completely oblivious to the fact that they appear as if they think the people they are talking to are too stupid to see through their tired arguments. I’ve got news for them: people aren’t that stupid. Thankfully. And I’m hopeful the results at the end of the month will make that clear.

The other good news is that whether OOXML gets approved or not, I believe Microsoft will pay a high price for all of its mischief and its image will come out of this badly damaged, something they can only blame themselves for.

In the meanwhile, don’t take for granted any claims of support for OOXML from Microsoft. The fact that Microsoft claims IBM has adopted OOXML can only make one wonder about all the other companies they list…

March 19, 2008 Posted by | standards | , , , , | 15 Comments