The Gilbane Advisor

Curated for content, computing, data, information, and digital experience professionals

Page 273 of 935

Structured Editing & Wikis

If you know me you will realize that I tend to revisit XML authoring tools and processes frequently. It is one of my favorite topics. The intersection of structured tools and messy human thinking and behavior is an area fraught with usability issues, development challenges, and careful business case thinking. And therefore, a topic ripe for discussion.

I had an interesting conversation with a friend about word processors and XML editors the other day. His argument was that the word processing product model may not be the best, and certainly isn’t the only, way to prepare and manage structured content.

A word processor is software that has evolved to support the creation of documents. The word processing software model was developed when people needed to create documents, and then later added formatting and other features. This model is more than 25 years old (I remember using a word processor for the first time in college in 1980).

Of course it was logical to emulate how typewriters worked since the vast majority of information at the time was destined for paper documents. Now word processors include features for writing, editing, reviewing, formatting, and limited structural elements like links, indexes, etc. Again, all very document oriented. The content produced may be reused for other purposes if transformed in a post process (e.g., it could output HTML & PDF for Web, breaking into chunks for a repository or secondary use, etc.), but there are limits and other constraints, especially if your information is primarily designed to be consumed in print or document form.

It is easy to think of XML-structured editors, and the word processor software model they are based upon, as the most likely way to create structured content. But in my opinion, structured editors pay too much homage to word processing features and processes. I also think too many  project teams assume that the only way to edit XML content is in an XML document editor. Don’t get me wrong, many people have successfully deployed XML editors and achieved targeted business goals, myself included, but I can point out many instances where an alternative approach to editing content might be more efficient.

Database tools that organize the information logically and efficiently are not likely to store that data as documents. For instance, you may have an financial system with a lot of info in relational fields that is extracted to produce printable documents like monthly statements, invoices, etc.

Or software manuals that are customized for specific configurations using reusable data objects and related document maps instead maintaining the information as static, hierarchically-organized documents.

Or aircraft information that needs to match the configuration of a specific plane or tail number, selected from a complete library of data objects stored centrally.

Or statutes that start formatted as bills, then later appear as enacted laws, then later yet again as published, codified statutes, each with their own formatting and structural peccadilloes.

Or consider a travel guide publisher that collects information on thousands of hotels, restaurants, attractions, and services in dozens of countries and cities. Sure, the content is prepared with the intent of publishing it in a book, but it is easy to see how it can be useful for other uses, including providing hotel data to travel-related Web sites, or building specialized, custom booklets for special needs (e.g., a local guide for a conference, guides to historical neighborhoods, etc.).

In these examples of what some might call database publishing, system designers need to ask them selves what would be the best tool for creating and maintaining the information. They are great candidates for a database, some application dialogs and wizards, and some extraction and transformation applications to feed Web and other platforms for consumption by users. They may not even involve an editor per se, but might rely entirely a Wiki or other dialog for content creation and editing.

Word processors require a mix of skills, including domain expertise on the subject being written about, grammar and editing, and some formatting & design, use of the software itself, etc. While I personally believe everyone, not just teachers and writers, should be skilled in writing well and making documents look legible and appealing, I realize many folks are best suited for other roles. That is why we divide labor into roles. Domain experts (e.g., lawyers, aircraft engineers, scientists and doctors, etc.) are usually responsible for accuracy and quality of the ideas and information, while editorial and product support people clean up the writing and formatting and make it presentable. So, for domain experts, it may be more efficient to provide a tool that only manages the content creation, structuring, linking, organization, etc. with limited word processing capabilities, and leave the formatting and organization to the system or another department or automated style sheets.

In my mind, a Wiki is a combination of text functionality and database management features that allow content to be created and managed in a broader Web content platform (which also may include static pages, search interfaces, pictures, PDFs, etc.). In this model, the Web is the primary use and printing is secondary. Domain experts are not bothered with concepts like page layout, running heads, tables of content generation, justification & hyphenation, etc., much to the delight of the domain experts!

I am bullish on Wikis as content creation and management tools, even when the content is destined for print. I have seen some that hide much of the structure and technical “connective tissue” from the author, but produce well formatted, integrated information. The blogging tool I am using to create this article is one example of a Wiki-like interface that has a few bells and whistles for adding structure (e.g., keywords) dedicated to a specific content creation purpose. It only emulates word processing slightly with limited formatting tools, but is loaded with other features designed to improve my blog entries. For instance, I can pick a keyword from a controlled taxonomy from a pull-down list. And all within a Web browser, not a fat client editor package. This tool is optimized for making blog content, but not for, let’s say, scientific papers or repair manuals. It is targeted for a specific class of users, bloggers. Similarly, XML-editors as we have come to know them, are more adept at creating documents and document chunks than other interfaces.

Honestly, on more than one occasion I have pounded a nail with a wrench, or tightened a bolt with the wrong kind of pliers. Usually I get the same results, but sometimes it takes longer or has a less desirable result than if I had used a more appropriate tool. The same is true for editing tools.

On a final note, forgive me if I make a gratuitous plug, but authoring approaches and tools will be the subject of a panel I am chairing at the Gilbane San Francisco conference in early June if you want to hear more. </>

DPCI Joins Acquia Program to Deliver Drupal Publishing Solutions

DPCI announced it has joined the Acquia Partner program at the Platinum level. Through the program, DPCI will expand its open source content management offerings by developing and delivering custom publishing solutions utilizing Acquia’s value-added products and services for the Drupal social publishing system. Additionally the program allows DPCI to leverage the Acquia Network for support, site management, and remote network services. http://drupal.org, http://acquia.com, http://www.databasepublish.com

JustSystems Updates Content Contribution Capabilities in XMetaL Reviewer

JustSystems announced XMetaL Reviewer 5.5, the latest version of the company’s collaborative XML document reviewing software. Reinforcing the central role of document review and collaboration in the overall authoring process, XMetaL Reviewer lets subject matter experts and others contribute content directly to the documents under review. The latest version of XMetaL Reviewer also gains usability, connectivity, integration, and administration enhancements. XMetaL Reviewer moves the document review process beyond a read-only exercise that prevents reviewers from suggesting detailed edits to the documents being reviewed. Now, reviewers can actively contribute new content and rewrite entire sections of a document. Using the addendum editor, XMetaL Reviewer users with no knowledge of XML can contribute structured content, insert images, and create lists, tables, and whole sections for consideration as document modifications. In addition to harnessing the expertise of document reviewers, XMetaL Reviewer offers users a consolidated view of all their suggested changes which are displayed to all reviewers in real time, as the changes are being made. An audit trail tracks all document changes, and an inline chat feature lets reviewers discuss content changes in real time. XMetaL Reviewer 5.5 is available today with prices starting at $15,000 for the server and five concurrent user licenses. Additional concurrent licenses are $2,500 for five users. http://www.justsystems.com

Gilbane SF conference program posted

The conference session descriptions have now been posted at including bi-directional links with the schedule. The program (and the page) are about 90% complete – there are a couple of session descriptions to be added and not all speakers are listed yet. The speaker bio page is also posted at with many of the speakers listed. Bi-directional inks between the speaker bios are conference sessions are being worked on now. And attendee, speaker, and press registration are open at: .

Gilbane Group’s use of Twitter

We’ve started using Twitter in multiple ways, and while our use of it will certainly evolve, I thought it would be helpful to point how what we are doing with it. (I’ll save thoughts about how it could or should be used for another time.)

Any suggestions on how else we might use Twitter?

Federated Search: Setting User Expectations

In the past few months, it is rare that I am briefed on an enterprise search product without a claim to provide “federated search.” Having worked with the original ANSI standard, Z39.50, and on one of the many review committees for it back in the early 1990s, it is a topic that always catches my attention.

Some of the history of search federation is described in this rather sketchy article at Wikipedia. However, I want clarify the original call for such a standard. It comes from the days when public access to search technologies was available primarily through library on-line catalogs in pubic and academic institutional libraries. A demand for the ability to search not only one’s local library system and network (e.g. a university often standardized on one library system to include all the holdings of a number of its own libraries), but also the holdings of other universities or major public libraries. The problem was that the data structures and protocols from one library system product to the next varied in way that made it difficult for the search engine of the first system to penetrate the database of records in another system. Records might have been meta-tagged similarly, but the way the metadata were indexed and accessible to retrieval algorithms was not possible with a translating layer between systems. Thus, the Z39.50 standard was established, originally to let one library system’s user search from that library system into the contents of other libraries with different systems.

Ideally, results were presented to the searcher in a uniform citation format, organized to help the user easily recognize duplicated records, each marked with location and availability. Usually there was a very esoteric results presentation that could only be readily interpreted by librarians and research scholars.

Now we live in a digitized content environment in which the dissimilarities across content management systems, content repositories, publishers’ databases, and library catalogs have increased a hundred fold. The need for federating or translation layers to bring order to this metadata or metadata-less chaos has only become stronger. The ANSI standard is largely ignored by content platform vendors, thus leaving the federating solution to non-embedded search products. A buyer of search must do deep testing to determine if the enterprise search engine you have acquired actually stands up well under a load of retrieving across numerous disparate repositories. And you need a very astute and experienced searcher with expert familiarity of content in all the repositories to make an evaluation as to suitability for the circumstance in which the engine will be used.

So, let’s just recap what you need to know before you select and license a product claiming to support what you expect from search federation:

  • Federated search is a process for retrieving content either serially or concurrently from multiple targeted sources that are indexed separately, and then presenting results in a unified display. You can imagine that there will be a huge variation in how well those claims might be satisfied.
  • Federation is an expansion of the concept of content aggregation. It has play in a multi-domain environment of only internal sites OR a mix of internal and external sites that might include the deep (hidden) web. Across multiple domains complete federation supports at least four distinct functions:
    • Integration of the results from a number of targeted searchable domains, each with its own search engine
    • Disambiguation of content results when similar but non-identical pieces of content might be included
    • Normalization of search results so that content from different domains is presented similarly
    • Consolidation of the search operation (standardizing a query to each of the target search engines) and standardizing the results so they appear to be coming from a single search operation

In order to do this effectively and cleanly, the federating layer of software, which probably comes from a third-party like MuseGlobal, must have “connectors” that recognize the structures of all the repositories that will be targeted from the “home” search engine.

Why is this relevant? In short, because it is expected by users that when they search, all the results they are looking at represent all the content from all the repositories they believed they were searching in a format that makes sense to them. It is a very tall order for any search system to do this but when enterprise information managers are trying to meet a business manager’s or executive’s lofty expectations, anything less is viewed as the failure of enterprise search. Or else, they better set expectations lower.

Ingres Launches Open Source Enterprise Content Management Offering

Ingres Corporation announced the availability of the Ingres Icebreaker Enterprise Content Management (ECM) Appliance. Powered by Alfresco’s open source alternative software for enterprise content management, the Ingres Icebreaker ECM Appliance gives businesses a way to manage business content growth. Like other commercial open source solutions, the Ingres Icebreaker ECM Appliance lets IT purchasers pay only for the software and support they actually need. For the Ingres Icebreaker ECM Appliance, Ingres provides the open source database for a company’s advanced data repository needs, Alfresco provides the content management expertise, and their technology runs on the Ingres Database. It is an appliance that allows developers to bring two open technologies together on the open source Linux operating system. The Ingres Icebreaker ECM Appliance integrates the operating system, the database, and the ECM technology and is installed as a unit, managed as a unit, and maintained as a unit. The Ingres Icebreaker ECM appliance allows users to capture, store, and preserve data, assist in the management of data, and deliver data to customers and partners. To download the Ingres Icebreaker ECM appliance today, please go to http://esd.ingres.com, http://www.ingres.com

 

2009 Gilbane San Francisco Conference to Focus on Business Impact of Global Content Management and Social Media

One of the most closely watched technology trends for 2009 is the need for enterprises to leverage new web platforms while holding down the costs of managing their ever-proliferating content. Even in the midst of the current economic downturn, critical areas related to global business content will see double-digit growth over the next 12 months – including spending on search technologies which is projected to represent almost half of all digital spending by businesses in 2009, along with rising corporate investment in social media tools (Source: Winterberry Group). Reflecting these fundamental shifts in the way enterprises engage with customers and disseminate information, the sixth annual Gilbane San Francisco June 2-4, 2009 at the Westin Hotel in San Francisco – will focus on the timely theme “Where Content Management Meets Social Media.” Produced by The Gilbane Group and Lighthouse Seminars, the 2009 Gilbane San Francisco conference will offer enhanced programs tied to the business issues surrounding a company’s marketing, technical and enterprise content. The conference tracks have been organized around the four major areas of how enterprises use Web and content technologies and where they are most likely to invest, including: Web Business & Engagement; Managing Collaboration & Social Media; Enterprise Content: Searching, Integrating & Publishing; and Content Infrastructure. Gilbane San Francisco brings together industry experts from leading technology, enterprise IT, analyst, and consulting firms who provide attendees with the latest successful content management and new media strategies, technologies and techniques. In addition to the latest best practices, technology coverage within these four tracks will include enterprise and site search; content globalization; semantic technologies; publishing; XML; and social media tools and platforms from Twitter to business blogs, project wikis and microformats. The just-published schedule of conference sessions can be viewed at http://gilbanesf.com/conference-schedule.html.

Pre-conference workshops will feature industry thought leaders covering core topics in web content, new media, Sharepoint and more — the full schedule of workshops can be found at http://gilbanesf.com/workshops.html. IT and business professionals involved in content creation, management, delivery or analytics wishing to attend the conference may register at: http://gilbanesf.com/registration_information.html. Technology solution providers wishing to exhibit or sponsor should visit: http://gilbanesf.com/exhibitors_sponsors.html. Follow the conference on Twitter:
http://twitter.com/gilbanesf

 

« Older posts Newer posts »

© 2025 The Gilbane Advisor

Theme by Anders NorenUp ↑