TweetFollow Us on Twitter

OpenDoc vs OLE
Volume Number:10
Issue Number:8
Column Tag:New Technologies

OpenDoc vs. OLE

When elephants fight, only the ants get hurt

By Jeff Alger, Consulting Methodologist

A monolith of American industry threatens to unilaterally impose a standard by virtue of its sheer size and dominance of the market. Quickly a consortium develops to counter the move; a rival standard is proposed, the hype wars begin. The lumbering giant points to the fait accompli: its technology is already in use throughout the world and will soon be a standard part of its operating systems on millions of machines. The upstarts point to technical advantages of their proposal and darkly hint at the threat to democracy and family values that will loom if the industry titan is allowed to further consolidate its grip. The press eats it up, sensing a David vs. Goliath story in the making. The battleground: component object-oriented software. The industry titan: Microsoft. The upstart consortium: Apple and IBM? Wait a minute. Am I the only one that sees something weird about this alignment? Age-old IBM a frightened upstart? Entrepreneurial Microsoft the lumbering giant? What’s going on here?

Unless you’ve been hiding underground the past few months - not a bad idea, actually - you’ve been ducking blows in the OLE (“Object Linking and Embedding”) vs. OpenDoc slugfest. Both address roughly the same issues: how to combine object-oriented software components into heterogenous documents, how to exchange data between programs, how to store and retrieve objects, how to distribute them across a network, how to allow objects to interoperate across programming languages and machines. Given these stakes, the battle may well be what both sides are claiming: the key to computing in the second half of the 90’s. But sorting out which is better and for what purposes is like trying to get directions to the nearest town from two people who are rolling on the ground trading rabbit punches. Let’s step back to a safe distance and take a dispassionate look at what’s really going on.

Of Objects and Markets

The basic technology at issue can be divided into basic four tiers of standards.

1. A standard for the binary representation of objects.

2. A standard mechanism for storing and retrieving documents containing arbitrary collections of objects.

3. A standard for how to allow users to combine arbitrary user-visible objects in their documents.

4. A standard mechanism for distributing and sharing objects between applications and over a network of dissimilar machines.

The first issue has to do with how objects are stored in memory. Microsoft’s standard is known as the Common Object Model (COM). The standard adopted for OpenDoc is IBM’s System Object Model (SOM). There is more than simple storage at issue. How do we allow different programming languages to create and access common objects? For that matter, just what is an object, anyway, and where should the line be drawn between an object at run time and the program source code used to create it? These are fundamental issues and the winner, if there is a single winner, of the current object wars will have a huge influence on object-oriented language designers for many years to come.

Older concepts of file systems and databases just can’t cope with the storage problems we face today. Both OLE and OpenDoc contain new standards, along with libraries that implement them, for storing and retrieving what are known as “compound documents,” files that contain nested collections of diverse sorts of objects. This, too, is likely to give the winner - if there is a single winner - a lot of throw weight in battles over other issues in the next generation of object systems.

Ah, yes, let’s not forget the users. They don’t directly see binary objects or compound document files or programming languages. To them an object is something they can point to on the screen or on hard copy and give a name. They rightly expect objects to be consistent and well-behaved no matter where they appear. They get frustrated at the arbitrary distinctions we’ve traditionally drawn between data that belongs in one sort of application and data that belongs somewhere else. “Wouldn’t it be nice,” Joe User might well say, “if I could simply copy or drag anything I can select into any other document in any other program and have it still work?” Indeed. That is the stuff of the user interface and portions of OLE and OpenDoc.

Finally, the arcane world of distributed objects. Neither side really has them yet, but in the struggle for control of tomorrow’s operating systems this is shaping up as potentially the most important part of the war. In a world of multiple machines and multiple virtual memory spaces, how do we tell one object from another? How are they created and destroyed? Do we store objects in their own little mini programs with separate address spaces, or combine them into a single memory pool? Can an object be shared between multiple documents or simply copied? Can one document contain objects that live on multiple machines? How do you handle failures, of networks, machines and users? Must a programmer write code having already decided which objects are to be distributed, or can those decisions be deferred?

With stakes like these, little wonder that the opposing sides are weighing in so heavily and so early. This is at once the next generations of object-oriented software technology, graphical user interfaces, and network operating systems. Waiting on the sidelines are the anxious software developers whose products must interoperate using one standard or the other; corporate developers hungry to lower their costs; retailers wondering what this means to the ways we package, price and distribute software; and programmers everywhere trembling at the thought that everything they once knew about writing applications might have to be relearned and knowing that they’d better pick the right horse or be left behind.

This issue contains more detailed descriptions of OpenDoc and OLE elsewhere, so I won’t go into a great deal of technical detail here. Instead, I’ll focus on comparing and evaluating the two products.

Back To School

Of all that has been said and written about OLE and OpenDoc, the only uncontestable claim is that both are complicated pieces of technology.

The OLE documentation numbers in the hundreds of pages, but if you really want to know what’s going on you have to read Kraig Brockschmidt’s 900+ page book Inside OLE 2.0, along with numerous more recent papers from Microsoft. Come to think of it, Brockschmidt’s book doesn’t even begin to talk about OLE Automation, the substrate upon which scripting is built in OLE. It wouldn’t hurt to attend a few conferences to get a sneak preview of the upcoming Chicago and Cairo operating systems. Since OLE is best used from within the Microsoft Foundation Classes (MFC) class library, allocate a few mental cycles to become familiar with that set of documents as well.

As to OpenDoc, I haven’t even been able to estimate the page count of the documentation yet. You can get a quick tour by pouring through a couple hundred pages of OpenDoc preliminary - very preliminary - documentation, then wade into the 500+ page System Object Model (SOM) documentation from IBM. Oh, and if you haven’t yet had the pleasure, take the opportunity to read up on the Drag and Drop Manager and the Code Fragment Manager (CFM), because they, too, are connected.

If you are trying to compare the two, be prepared to be inundated with all manner of position papers from both sides. The level of sniping in these “black papers” is exceeded only by the casual attention to the real facts.

Egad! Don’t these people know that we all have jobs and families? Learning either of these technologies well enough to write a simple component takes a few days. Learning enough to design a major project looks to me more like weeks of reading and training backed by a at least a few months of hacking. Learning enough to intelligently make a long-term strategic choice is nearly impossible, especially since one of them doesn’t even exist yet.

Ghosts Of Objects Past, Present and Future

One of the surreal aspects of the debate is that it isn’t always clear what’s been there for a long time, what’s new and what is still vaporware. The OpenDoc consortium has done a brilliant job of positioning OpenDoc,tomorrow’s concept, against OLE 2.0, today’s product. Microsoft plays into their hands by claiming OLE is better because it’s already shipping, then in the same breath turning around and claiming that their vaporware will be better than OpenDoc’s vaporware. This stuff gives me a headache. Let’s start the comparison by defining our terms. The following is the situation as of this writing in mid-June.

The Past

On the Microsoft side, OLE 1.0 shipped in 1990 and OLE 2.0, a major enhancement, shipped early last year. It is available today in final form on all Windows and Windows NT platforms and on the Macintosh. Microsoft, lover of big numbers, claims there are about 1.5 million customers using OLE-aware applications and that most of the major software vendors are already committed to supporting OLE, most of them already pretty far down along the learning curve.

On the OpenDoc side, there are a couple of precursor technologies such as the Drag and Drop Manager but that’s about it. In short, Microsoft is right, as of today OLE is shipping and OpenDoc isn’t.

The Present

As you read this, there are various alpha and beta versions zooming around both worlds, making it horribly confusing to keep score. OpenDoc is in release A6 on the Macintosh as of the Apple World Wide Developers Conference in May. This includes a very early release of the IBM System Object Model implementation for the Macintosh. OpenDoc has not made its way to any other platforms as of this writing, although one piece of the underlying technology, SOM, is supported by IBM on AIX and OS/2.

Meanwhile, Microsoft has seeded a “select” group of 5000 developers with a prerelease version of the distributed version of OLE. They have also begun beta testing of OLE Controls, a small-component extension to OLE that is both a close analogy to OpenDoc parts and to Windows VBX controls.

The Future

Microsoft is busy implementing their “Windows API Everywhere” strategy. They have licensed the source code to the Windows operating system to at least two other vendors who have demonstrated partial ports to Unix. They have committed to licensing the Windows and OLE source more broadly, although the terms have not been made public. Where Windows goes, so goes OLE. In addition, Microsoft is working with DEC to refine COM somewhat to go from the “Component Object Model” to the “Common Object Model.” In addition to driving everyone nuts keeping the two identical acronyms straight, this will enable DEC to port OLE to other operating systems including OSF/1. Microsoft claims that a port to MVS on IBM mainframes is also under way. It should be made clear what this means: The OLE standard, including the objects themselves, will interoperate across both Windows and non-Windows operating systems.

One of the worst-kept secrets in the computing world is the work Microsoft is doing on their next generations of the Windows and Windows NT operating systems. Daytona, the successor to Windows NT and already in beta testing, will include native support for OLE 2.0 built into the operating system. Chicago, the successor to the Windows operating system and a major enhancement of the user interface, blurs whatever line is left between where the operating system ends and OLE begins. Cairo is an object-oriented operating system built on OLE from the ground up. Another technology Microsoft already has well into the pipeline is an integration of OLE with ODBC database support. All, with the possible exception of Cairo, are scheduled for release within roughly the same time frame as a final release of OpenDoc, so it is fair game to project them into the equation. (I consider anything within six months to be “roughly the same time frame.”)

The OpenDoc consortium has committed publicly to making the source code to OpenDoc available and to port it to a variety of platforms. They point to “support” from systems software vendors Apple, IBM, SunSoft, Taligent, Novell, and Xerox “with many, many others expected to follow” according to one white paper. (italics added) It is unclear from their white papers and documentation exactly what is being supported by all these vendors, OpenDoc or only pieces such as SOM. The phraseology cryptically lapses into support for “CILabs technologies,” CILabs being Component Integration Laboratories, the formal entity behind OpenDoc.

Recently the OpenDoc crowd has announced that OpenDoc will be written in such a way that it can interoperate with OLE. Details to follow.

Binary Object Models

Both products allow you to create objects in such a way that they can be accessed from programs written in other languages. This is the stuff of so-called “binary standards” for objects, a subject that has kept lots of folks busy over at the Object Management Group for quite a while. The two approaches are Microsoft’s COM and IBM’s SOM and they couldn’t be more different.

The Component Object Model

COM embodies a sharp separation between binary objects and the object-oriented programming languages used to create them. The COM model is based on delegation, not inheritance. Rather than have one class derive from another, in COM you would likely form what is known as an aggregate of two objects. The container object would delegate to the component object in much the same way that a derived class would pass along messages to its base class. Aggregation can also be used to implement the equivalent of complex data members.

COM represents a C++ way of thinking of objects at run time. Methods are dispatched using something that you would swear was a vtable if you didn’t know better. In fact, you can use C++ derivation to create your aggregates, as long as you add the interfaces to direct a client to the right base class or data member. Yet, the model is not exactly C++. While servers may be written in terms of derivation, clients see a flat object space. There is more mathematical consistency than in straight C++; if you do the equivalent of typecast from derived to base, you are guaranteed to be able to typecast back to the derived class in a typesafe way. This reflexivity is one of several rules of consistency imposed by COM to maintain type safety.

Speaking of type safety, COM is safe very. The way you access an object is through an interface. Each interface is uniquely identified by an interface ID (IID). The way you obtain an interface from an object is, you ask it. The specific way one would typecast from “derived” to “base,” for example, would be something like this:

Derived* d = (Derived*)object->QueryInterface(kDerivedID);

This returns either a valid pointer or NULL to signify that the object doesn’t support that interface. That is, in order to typecast any object you have to go through the object, which can verify that the interface is valid for that object. This is all done in either C or C++. It is easy to wrap existing classes with COM-style interface wrappers.

Normally, you will use interfaces that represent standard C++ class interfaces. COM also provides a special interface called IDispatch that essentially allows you to query an object about its capabilities at run time and dynamically dispatch to it using symbols, not precompiled function calls. This provide low-level support for scripting and high-level development tools that need a dynamic environment. Because of the delegation model, an object can change its interface and implementation at run time, within certain limitations to insure consistency.

What is involved in using COM? Basically, you just write the interfaces and the code. If you are using C++, it’s no more complicated than writing any other C++ program, but with a couple of extra member functions to write such as QueryInterface and perhaps support for OLE Automation. If you are a C programmer, there are some nifty macros that reduce the problem to about the same level of complexity as using C++, although I have a hard time understanding why anyone would want to use C rather than C++ for this particular purpose. If you are using other languages, you have to check with the language vendor. Many are being adapted to work with COM but it’s not a universal standard yet. As they are adapted, however, the objects you create using, say, C++ code will be accessible and manipulable from, say, Smalltalk and vice versa. Learning to use COM involves perhaps a few dozen pages of manuals and less than a day of effort.

The System Object Model

SOM frightens me. It really does. It’s as if a bunch of people with the software equivalent of bazookas sat around one day and said, “Hey, let’s see if we can solve a problem nobody cares about by taking all the major object-oriented languages and blasting them into smithereens.” Brrrr. What’s even more frightening is that no one seems to be taking IBM to task for this unbelievable act of engineering hubris - except Microsoft, and that is written off as competitive snarling. The rest of OpenDoc is a reasonably good piece of engineering, but if SOM comes within a hundred yards, duck.

The stated objective behind SOM is to provide “language-neutral” access to objects. It is based for the most part on the Object Management Group’s (OMG) efforts to standardize the binary representation of objects. The run-time model is dynamic, supporting run-time inheritance and polymorphism. Thus, you could use all the features of a Smalltalk or Common Lisp Object System (CLOS), such as creating classes or changing their derivation on the fly.

However laudable these aims, the result is an over-engineered mess. To use it, you have to first describe your objects using yet another language, the SOM Interface Description Language (IDL). This isn’t just a minor detail, but another programming language to master. The syntax is sort-of-Smalltalk, sort-of-C, sort-of-C++ but in the end it is IDL. It is chock full of macros, modifiers, conventions, and isms.

SOM isn’t language-neutral; I would call it C++ -hostile and sort of Smalltalk-friendly. The role of C++ is reduced to implementing and calling specific member functions. The C++ notions of class, derivation, constructors, destructors, operators, overloading, etc., are all subsumed by the IDL model. SOM does not support private, protected and public member specifications, unless you consider surrounding a member function with “#ifdef __PRIVATE__/#endif” to be “support.” In fact, it is hardly worth bothering to use C++ with SOM, since C does just about as well with what’s left. SOM is a slap in the face of the C++ community.

OK, so you don’t like C++. If what you really want is dynamic binding and such, Smalltalk or CLOS will do better than IDL at those tasks. Garbage collection is based primarily on reference counting. While this is also true of COM, COM doesn’t pretend to provide language neutrality. It is difficult to envision true Smalltalk- or CLOS-style programming using a reference-counted garbage collection scheme. SOM doesn’t have anything resembling code blocks from Smalltalk and Objective-C or macros from CLOS. There obviously aren’t anywhere near the diversity of libraries for SOM that there are for Smalltalk, Object-C or CLOS. It isn’t worth going further, because I wouldn’t stop for several days. To paraphrase the former candidate for Vice President and current Treasury Secretary, “I worked with Smalltalk. Smalltalk was my friend. Sir, you are no Smalltalk.”

Interoperability? COM is living proof that you can provide an equivalent level of language interoperability without first tossing conventional languages into a Cuisinart. Whether you like COM or not is not the issue; the issue is that SOM is to interoperating what a flamethrower is to mowing the lawn. Yes, the lawn won’t grow back for a while, but haven’t we forgotten something important about lawns?

The development lifecycle is bizarre. First, decide which objects should be treated as SOM objects and which as normal, local objects. If you later change your mind, it’s bad news unless you’re paid by the hour because you’ll have to pretty much start over. Next, write a description of your objects in IDL, then compile them using an IDL compiler. This funnels your description through an “emitter” which spits out template C or C++ or Smalltalk or whatever code. Fill in the blanks. Compile again, this time using your language compiler. On the client side, emit stuff to include with your program. Write your program. Compile. In short, take all your worst nightmares about software management and play them at fast forward with the color knob twisted all the way to green.

The funny thing is, there isn’t even a significant problem solved by SOM. Other than the vendors who forked over big bucks to participate in the Object Management Group, I haven’t noticed a particular groundswell of opinion that language interoperability is a big issue. Language vendors have managed to do a creditable job at that without having to toss everyone’s salad. A binary standard for how C++ objects are represented in memory would achieve 90% of what SOM does with about 5% of the hassle, because then C++ itself could also serve as the object interface description language. Extensions to C++ to handle intermixed dynamic and static binding of methods - permitted by the language spec, by the way, which does not require the use of vtables - would solve the rest. Even without that, as COM demonstrates, idioms for using C++ more intelligently can solve just about any problem a developer will encounter in real life. That’s what James Coplien’s book Advanced C++ Programming Styles and Idioms is all about.

It is telling that in OpenDoc one is only expected to use SOM for a small subset of objects that have to be accessed externally. The rest is conventional C++, including almost all of the OpenDoc Parts Framework class library.

Documents and Storage

Both products allow you to store arbitrary, nested, versioned objects in compound document files. Both use streaming as the primary means of turning an object into bits and back. Both provide a way to browse through a compound document file and extract only those pieces you care about. Both provide a way to maintain multiple versions of a document or specific objects within it. There are differences in the area of document storage, but not compelling ones in either direction.

The one major difference coming up is that the Structured Storage model of OLE is being built into the file system of upcoming Microsoft file systems. This will provide a higher level of protection and performance than one can expect from Bento, the underlying storage technology of OpenDoc. This will also fix a nagging problem with the current version, which uses file path names to locate things. If the user moves the file, OLE might not be able to find it when needed. SOM also uses path names, but OpenDoc apparently will stick to the Macintosh model of file references, which stay the same as a file is moved around.

Structured Storage has slightly better features for recovering from a crash, including a very good feature called “Transacted” mode, in which all changes are considered temporary until all are committed together. Under OpenDoc, changes are written incrementally, possibly leaving a file in an inconsistent state following a crash, at least as far as the user might be concerned. Structured Storage also makes it a little easier to implement a “Revert” command without having to duplicate the entire file, only the parts that change; OLE keeps track of which is which.

The User’s Perspective

OpenDoc looks very good at the user level compared to OLE. Apple has done a very good job of thinking through how a user will interact with parts that have different commands and different user interfaces for those commands; i.e., menus. In OpenDoc, the integration of the parts into a single user-visible document is pretty close to seamless. Click on an object and the user interface changes state as appropriate to that object. New menus show up, old ones go away, windoids come and go as they should. The standards for borders and other user feedback seem thoughtfully designed.

Comparing OpenDoc’s user interface directly to OLE’s is difficult because there are several levels of OLE support possible. Where OpenDoc has essentially one interface and architecture, OLE can potentially support several. This is not so much because the user will appreciate this as that OLE presents a graduated scale of options to the developer, depending on the degree of integration you want or how OLE-capable you are.

A minimal treatment is to turn a normal Windows application into an OLE-aware application, capable of being an OLE “server.” That is, the application’s document becomes a component in some other application’s - a container’s - document. When the user “activates” the embedded object, the originating application is fired up and the actual editing takes place in a separate, temporary window. This is a slight extension to the user’s concept of using two separate applications: There are still two apps, but now the data is combined. It is not necessary to have the original app around if all you want to do is draw and print, since the PICT or metafile needed to do those is er, embedded into each OLE server object. (By the way, OLE automatically turns PICTs into Windows metafiles and back when objects migrate between Macs and Windows machines.) Only if the user needs to edit or you want to do something fancy like animation or sound do you need the original application installed. Most of the work involved in becoming an OLE server is in supporting a stream interface for the clipboard and drag and drop to use in moving the data around.

On the receiving end is the OLE container, an application whose documents can contain OLE components. Like becoming a server, turning an arbitrary application into an OLE container is relatively easy although more involved than writing servers. A container-server is simply a combination of the two techniques. At this level one gets the use of drag and drop and OLE-style clipboard use.

It is also possible to create OLE servers that edit in place, rather than opening a separate window. These modify the user interface environment around them in much the same way that OpenDoc parts do: menus and toolbars swap to reflect the currently active server. This is known as in-place activation.

The next and most common treatment is an in-process server. This is a dynamic linked library (DLL) that runs as part of the client application, though it is loaded at run time. This allows lightweight display or editing code to go with the object without having to bring along the entire originating application.

Finally, there is the future: OLE Controls. These are direct analogies to OpenDoc parts. They are scriptable, they activate with single clicks, not double, even on deeply nested objects, and they run as part of the client process. In the long run they look a good bet to replace VBX controls in the Windows world.

So, which user interface is better? Unfortunately, the comparisons by the combatants aren’t always fair or accurate. More than in any other arena, one has to ask, “What version are you comparing to, today’s or tomorrow’s?” If you compare the user experience with OpenDoc parts to OLE 2.0’s out-of-place servers in OLE, OpenDoc wins, hands down, from the user experience. Users will appreciate not having to remember when to use a single click and when to use a double click to get the job done and not having to understand much about the underlying nested structure of their objects.

If you compare OpenDoc parts to OLE Controls, the differences become pretty minor both ways. Both activate in exactly the same way, with a single click, even on nested components. Both activate in-place. Both run in the same task space as the client application, although an OLEcontrol can run in its own space if written that way. OpenDoc still maintains a small edge because of its support for arbitrarily shaped borders around objects; as of today, OLE is limited to rectangular borders. However, Microsoft has stated publicly that they, too, are considering supporting funny-shaped embedded objects, so that distinction may not last for long.

So which user interface is better? Today, OLE 2.0 because OpenDoc isn’t here yet; this is one of those issues that has gotten clouded by comparing OLE today to OpenDoc tomorrow. Tomorrow, when OLE Controls and Cairo have come along, I suspect OpenDoc will still have the edge, though both will be excellent treatments.

Distributing the Objects

Finally, let’s talk about what happens when we stop looking at individual applications and start looking at entire networks. COM becomes a network-wide distributed object system by simply installing the Networked OLE DLL in place of the OLE 2.0 DLL. That’s it. No recompilation, no changes to source. OpenDoc requires the use of the Distributed System Object Model, or DSOM, the distributed extensions to SOM. DSOM is certainly a creditable implementation of distributed objects and the techniques used are very well-established. However, under straight DSOM, you have to decide up front what you want to distribute and what you don’t. The interfaces for creating, destroying and in some cases using objects under DSOM are quite different from those under straight SOM. The programming needed to implement DSOM servers adds even more complexity. This is not a trivial retrofit to existing code. Plus, it’s SOM++ and all that it implies.

Apple states that OPF will hide the differences between SOMand DSOM, and between local and remote objects. The A6 release code and documentation do not indicate how that will be done. [By the time you see this, Apple may have made this clear through a subsequent release - Ed stb]

As Microsoft is apt to point out, this relative simplicity of COM does not mean that the COM implementation is less robust; the opposite seems to be the case. For example, DSOM does not have any built-in capability to detect and resolve deadlocks while COM handles them properly. Programmers must manually design for support of multithreaded object servers under DSOM; not so under COM because of the operating system-level infrastructure underneath.

White Noise

There is a recurring problem with comparisons of the two technologies emanating from both camps. Both approaches are fundamentally sound and many of the “problems” can be fixed with little or no effort. For example, one of the most vocal criticisms of OpenDoc from Microsoft is that OpenDoc parts must be accompanied by the corresponding part handler code or they can’t be displayed or printed. In other words, if you don’t have the right part handlers installed on the user’s machine, she will just see big gray boxes where her objects should be. However, the OpenDoc folks could solve that at a stroke by requiring that all OpenDoc parts store a PICT image as a property, thereby allowing displaying and printing with or without the right part handler code.

There are outright distortions that show just how sloppy the discussion has become. For example, one supposed knock against OLE is that to activate a nested object one has to double click on the outermost object, then keep activating inward one object at a time. There are two problems with this: It is within the OLE spec to implement activation outside-in or inside-out a la OpenDoc and it is also within the OLE spec to use a single click rather than a double click to activate an object. OLE Controls illustrate these two facts by implementing an OpenDoc-style activation model - inside-out, single-click - while staying within the OLE specification. Apple also claims that the way that OLE finds server code, using file pathnames that a user can easily screw up, is a serious flaw in OLE. What Apple doesn’t point out is that Cairo, which is supposed to ship not too far from OpenDoc final, fixes all that. Granting equal time, Microsoft has decried OpenDoc’s requirement for part viewers on all the machines in a network that might need to access parts. This, it is claimed, requires engineering the software separately for each platform, neglecting to mention that the OpenDoc Parts Framework (OPF) class library is designed to provide source code compatibility across platforms and that only on machines supporting the Windows API is similar source compatibility assured for OLE. Both sides cite the supposedly extraordinary difficulty in using the other guy’s product, forgetting to mention that each has class libraries, OPF on the one hand and the Microsoft Foundation Classes on the other, that dramatically simplify things. Excuse me, my headache is back.

To Market, To Market, Lickety Split

Both OpenDoc and OLE present acceptable user interfaces. There are differences, and as of today’s specifications OpenDoc gets the nod. How much of a nod depends on your particular priorities, for while OpenDoc does, in fact, do a better job of improving the user experience, OLE is likely to get you into the component software market faster and is certainly not a bad user interface even today. In the time since Apple first started working on Taligent and later OpenDoc, Microsoft has released a steady stream of new, incrementally improving technologies, bringing the market and the industry along with them slowly but very surely. There are now hundreds of vendors of everything from DLLs to VBX controls to OLE objects that have experience in the marketplace for component software. It will take years for the Apple world to gain that level of experience with how to develop, package, price and support the sort of components needed to make OpenDoc a success.

Microsoft provides as seamless a growth path into OLE as OpenDoc presents a seamless user interface across parts. Companies that already know how to write Windows applications can start shipping OLE-aware applications with an absolute minimum of effort and with little or no rewriting or redesign of existing code. The ramp from there is gently sloped until one is shipping OLE Controls, the full-fledged equivalent of OpenDoc parts. Going from single-process COM to fully distributed objects involves no more than installing a new DLL. There are no new programming languages to learn, as there are with SOM. It is foolish to claim that all of these dollars and cents issues should be swallowed up by a single-minded quest for better user interfaces.

There are two different business models at work, Apple’s quantum leap approach to software and Microsoft’s continuous incremental refinement. I suspect there is room in the market for both, as long as the technological differences are narrow enough. Eventually, I expect this intense competition will reveal a case study in the way market economies are supposed to work: Both the OpenDoc consortium and Microsoft are being pushed to adopt the best features of each and this is improving the lot of consumers on both sides of the aisle.

The Mid-Term Report Card

OK, let’s cut through all this and reduce it to letter grades. To do this, I’m going to use a reasonable extrapolation of what will be available within twelve months; anything else gets too messy to do a direct comparison.

Subject OLE OpenDoc

Binary object standard B F

Compound document storage B B-

User interface A(1) A+

Distributed object support A+ B(2)

Ease of learning B C

Ease of adoption (3) A C

Overall grade A- B-

(1)Assumes OLE Controls and Cairo. Without these the grade would be C+ (no pun intended).

(2)This is purely for the DSOM extensions, since SOM itself is graded separately.

(3)Principally the smoothness of the growth path from existing development practices.

Overall grades are weighted evenly across all subjects. I won’t quibble with anyone that wanted to assign the user interface more weight, which would improve OpenDoc’s grade but not lower OLE’s. Were OpenDoc to use COM rather than SOM, the overall OpenDoc score would climb dramatically. Well, I can dream, can’t I?

 

Community Search:
MacTech Search:

Software Updates via MacUpdate

Latest Forum Discussions

See All

Make the passage of time your plaything...
While some of us are still waiting for a chance to get our hands on Ash Prime - yes, don’t remind me I could currently buy him this month I’m barely hanging on - Digital Extremes has announced its next anticipated Prime Form for Warframe. Starting... | Read more »
If you can find it and fit through the d...
The holy trinity of amazing company names have come together, to release their equally amazing and adorable mobile game, Hamster Inn. Published by HyperBeard Games, and co-developed by Mum Not Proud and Little Sasquatch Studios, it's time to... | Read more »
Amikin Survival opens for pre-orders on...
Join me on the wonderful trip down the inspiration rabbit hole; much as Palworld seemingly “borrowed” many aspects from the hit Pokemon franchise, it is time for the heavily armed animal survival to also spawn some illegitimate children as Helio... | Read more »
PUBG Mobile teams up with global phenome...
Since launching in 2019, SpyxFamily has exploded to damn near catastrophic popularity, so it was only a matter of time before a mobile game snapped up a collaboration. Enter PUBG Mobile. Until May 12th, players will be able to collect a host of... | Read more »
Embark into the frozen tundra of certain...
Chucklefish, developers of hit action-adventure sandbox game Starbound and owner of one of the cutest logos in gaming, has released their roguelike deck-builder Wildfrost. Created alongside developers Gaziter and Deadpan Games, Wildfrost will... | Read more »
MoreFun Studios has announced Season 4,...
Tension has escalated in the ever-volatile world of Arena Breakout, as your old pal Randall Fisher and bosses Fred and Perrero continue to lob insults and explosives at each other, bringing us to a new phase of warfare. Season 4, Into The Fog of... | Read more »
Top Mobile Game Discounts
Every day, we pick out a curated list of the best mobile discounts on the App Store and post them here. This list won't be comprehensive, but it every game on it is recommended. Feel free to check out the coverage we did on them in the links below... | Read more »
Marvel Future Fight celebrates nine year...
Announced alongside an advertising image I can only assume was aimed squarely at myself with the prominent Deadpool and Odin featured on it, Netmarble has revealed their celebrations for the 9th anniversary of Marvel Future Fight. The Countdown... | Read more »
HoYoFair 2024 prepares to showcase over...
To say Genshin Impact took the world by storm when it was released would be an understatement. However, I think the most surprising part of the launch was just how much further it went than gaming. There have been concerts, art shows, massive... | Read more »
Explore some of BBCs' most iconic s...
Despite your personal opinion on the BBC at a managerial level, it is undeniable that it has overseen some fantastic British shows in the past, and now thanks to a partnership with Roblox, players will be able to interact with some of these... | Read more »

Price Scanner via MacPrices.net

You can save $300-$480 on a 14-inch M3 Pro/Ma...
Apple has 14″ M3 Pro and M3 Max MacBook Pros in stock today and available, Certified Refurbished, starting at $1699 and ranging up to $480 off MSRP. Each model features a new outer case, shipping is... Read more
24-inch M1 iMacs available at Apple starting...
Apple has clearance M1 iMacs available in their Certified Refurbished store starting at $1049 and ranging up to $300 off original MSRP. Each iMac is in like-new condition and comes with Apple’s... Read more
Walmart continues to offer $699 13-inch M1 Ma...
Walmart continues to offer new Apple 13″ M1 MacBook Airs (8GB RAM, 256GB SSD) online for $699, $300 off original MSRP, in Space Gray, Silver, and Gold colors. These are new MacBook for sale by... Read more
B&H has 13-inch M2 MacBook Airs with 16GB...
B&H Photo has 13″ MacBook Airs with M2 CPUs, 16GB of memory, and 256GB of storage in stock and on sale for $1099, $100 off Apple’s MSRP for this configuration. Free 1-2 day delivery is available... Read more
14-inch M3 MacBook Pro with 16GB of RAM avail...
Apple has the 14″ M3 MacBook Pro with 16GB of RAM and 1TB of storage, Certified Refurbished, available for $300 off MSRP. Each MacBook Pro features a new outer case, shipping is free, and an Apple 1-... Read more
Apple M2 Mac minis on sale for up to $150 off...
Amazon has Apple’s M2-powered Mac minis in stock and on sale for $100-$150 off MSRP, each including free delivery: – Mac mini M2/256GB SSD: $499, save $100 – Mac mini M2/512GB SSD: $699, save $100 –... Read more
Amazon is offering a $200 discount on 14-inch...
Amazon has 14-inch M3 MacBook Pros in stock and on sale for $200 off MSRP. Shipping is free. Note that Amazon’s stock tends to come and go: – 14″ M3 MacBook Pro (8GB RAM/512GB SSD): $1399.99, $200... Read more
Sunday Sale: 13-inch M3 MacBook Air for $999,...
Several Apple retailers have the new 13″ MacBook Air with an M3 CPU in stock and on sale today for only $999 in Midnight. These are the lowest prices currently available for new 13″ M3 MacBook Airs... Read more
Multiple Apple retailers are offering 13-inch...
Several Apple retailers have 13″ MacBook Airs with M2 CPUs in stock and on sale this weekend starting at only $849 in Space Gray, Silver, Starlight, and Midnight colors. These are the lowest prices... Read more
Roundup of Verizon’s April Apple iPhone Promo...
Verizon is offering a number of iPhone deals for the month of April. Switch, and open a new of service, and you can qualify for a free iPhone 15 or heavy monthly discounts on other models: – 128GB... Read more

Jobs Board

IN6728 Optometrist- *Apple* Valley, CA- Tar...
Date: Apr 9, 2024 Brand: Target Optical Location: Apple Valley, CA, US, 92308 **Requisition ID:** 824398 At Target Optical, we help people see and look great - and Read more
Medical Assistant - Orthopedics *Apple* Hil...
Medical Assistant - Orthopedics Apple Hill York Location: WellSpan Medical Group, York, PA Schedule: Full Time Sign-On Bonus Eligible Remote/Hybrid Regular Apply Now Read more
*Apple* Systems Administrator - JAMF - Activ...
…**Public Trust/Other Required:** None **Job Family:** Systems Administration **Skills:** Apple Platforms,Computer Servers,Jamf Pro **Experience:** 3 + years of Read more
Liquor Stock Clerk - S. *Apple* St. - Idaho...
Liquor Stock Clerk - S. Apple St. Boise Posting Begin Date: 2023/10/10 Posting End Date: 2024/10/14 Category: Retail Sub Category: Customer Service Work Type: Part Read more
Top Secret *Apple* System Admin - Insight G...
Job Description Day to Day: * Configure and maintain the client's Apple Device Management (ADM) solution. The current solution is JAMF supporting 250-500 end points, Read more
All contents are Copyright 1984-2011 by Xplain Corporation. All rights reserved. Theme designed by Icreon.