Curated for content, computing, and digital experience professionals

Category: Web technologies & information standards (Page 23 of 58)

Here we include topics related to information exchange standards, markup languages, supporting technologies, and industry applications.

The Accountant Who Knew Too Much

It was a dark and rainy night. She toiled way past normal quitting time for all but accountants with Securities and Exchange Commission (SEC) filings deadlines looming. Cranking away on her first XBRL SEC filing, Debbie became quite frustrated. "I know what is supposed to go into all of these "other" accounts, but XBRL just doesn’t care," she lamented. You see, Debbie is the accountant who knew too much.

Debbie is not alone. In a recent conversation with Louis Matherne, former XBRL International President and Director, XBRL Services for Clarity Systems, the situation described above is actually a common occurrence for accountants tagging XBRL filings for the first time. He suggested a simple example:

The company balance sheet says "prepaid expenses and other". "Other" is there because it represents several accounts that aggregated to the balance sheet become immaterial as separate items. The registrant, however, knows what it is and thinks they should create a new taxonomy concept that better captures the details of "other". No where in the financial statements or footnotes to the financial statements do they describe what that ‘other’ is."

The object of using XBRL for compliance with the SEC mandate is to present the company’s required financial statements and footnote disclosures, not to expose the preliminary accounts and internal decisions the led to the final, top level reports. XBRL is not meant to extend, expand or further explain legal filings. It simple puts your disclosures into a machine readable form. The last word comes from Matherne: "XBRL for the SEC is primarily about the disclosure of the accounting".

The US GAAP XBRL taxonomies can be found here.

The Accountant Who Knew Too Much

Tweeting XBRL

Over the last few months, I have become acquainted with the wonders of the 140 character “tweet”. For those of you who are not “tweets”, I am referring to the combination of instant message and social networking that has converged at www.twitter.com. In essence, twitter asks “peeps” a simple question, What are you doing right now? In 140 spaces, you can communicate what you are thinking or what you have just read on the web.  Long URL’s are easily truncated leaving enough space to communicate simple messages.  If you find people that are doing or reading about interesting things, you can follow their “tweets”.

In efforts to keep up on XBRL, I use Yahoo and Google key word alerts as well as selected RSS feeds from the SEC and others. I have found, however, that this system falls short of the daily updates the people using www.twitter.com are providing for me. Its amazing how effective a 140 character message can be in sending you directly to fresh web content relevant to your interests. To improve my “hit” rate, I’ve added a few Tweet favorite tools such as tweetdeck (www.tweetdeck.com), TwitScoop, WeFollow, and MrTweet. Join up and send me a note. I’d love to follow you!

XML and Belly Buttons: How to “Sell” XML

Anyone who works with XML has probably had to “sell” the idea of using the standard instead of alternative approaches, whether as an internal evangelist of XML or in a formal sales role. We have developed some pretty convincing arguments, such as automating redundant processes, quality checking and validation of content, reuse of content using a single source publishing approach, and so on. These types of benefits are easily understood by the technical documentation department or developers and administrators in the IT group. And they are easy arguments to make.

Even so, that leaves a lot of people who can benefit from the technology but may never need know that XML is part of the solution. The rest of the enterprise may not be in tune with the challenges faced by the documentation department, and instead focus on other aspects of running a business, like customer support, manufacturing, fulfillment, or finance, etc.. If you tell them the software solution you want to buy has “XML Inside” they may stare off into space and let their eyes glaze over, even fall asleep. But if you tell them you have a way to reduce expensive customer support phone calls by making improvements to their public-facing Web content and capabilities, you might get more of their attention.

I have been around the XML community for a very long time, and we tend to look into our belly buttons for the meaning of XML. This is often doen at the expense of looking around us and seeing what problems are out there before we start talking about solutions to apply to them. Everything looks like a nail because we have this really nifty hammer called XML. But when CD-ROMs were introduced, people didn’t run around talking about the benefits of ISO 9660 (the standard that dictates how data is written to a CD). Okay they did at first to other technologists and executives in big companies adopting the standard, but rarely did the end consumer hear about the standard. Instead, we talked about the massive increase in data storage, and the flexibility of a consistent data storage format across operating systems. So we need to remember that XML is not what we want to accomplish, but rather how we may get things done to meet our goals. Therefore, we need to understand and describe our requirements in terms of these business drivers, not the tools we use to address them.

Part of the problem is that there are several potential audiences for the XML evangelism message, each with their own set of concerns and domain-specific challenges. End users want the ability to get the work out the door in a timely manner, at the right quality level, and that the tools are easy to use. Line Managers may add sensitivity to pricing, performance, maintenance and deployment costs, etc. These types of concerns I would classify as tactical departmental concerns focusing on operational efficiency (bottom line).

Meanwhile Product Managers, Sales, Customer Service, Fulfillment, Finance, etc. are more geared toward enterprise goals and strategies such as reducing product support costs, and increasing revenue, in addition to operational efficiency. Even stated goals like synchronizing releases of software and documentation, making data more flexible and robust to enable new Web and mobile delivery options, are really only supporting the efforts to achieve the first two objectives of better customer service and increased sales, which I would classify as strategic enterprise concerns.

The deft XML evangelist, to succeed in the enterprise discussion, needs to know about a lot more than the technology and processes in the documentation department, or he or she will be limited to tactical, incremental improvements. The boss may want, instead, to focus on how the data can be improved to make robust Web content that can be dynamically assembled according to the viewer’s profile. Or how critical updates can be delivered electronically and as fast as possible, while the complete collection of information is prepared for more time consuming, but equally valuable printed delivery in a multi-volume set of books. Or how content can be queried, rearranged, reformatted and delivered in a completely new way to increase revenue. Or how a business system can automatically generate financial reporting information in a form accurate and suitable enough for submission to the government, but without the army of documentation labor used previously.

At Gilbane we often talk about the maturity of XML approaches, not unlike the maturity model for software. We haven’t finalized a spectrum of maturity levels yet, but I think of XML applications as ad hoc, departmental, and enterprise in nature. Ad hoc is where someone decides to use an XML format for a simple process, maybe configuration files driving printers or other applications. Often XML is adopted with no formal training and little knowledge outside of the domain in which it is being applied.

Departmental applications tend to focus on operational efficiency, especially as it relates to creating and distributing textual content. Departmental applications are governed by a single department head but may interact with other groups and delivery feeds, but can standalone in their own environment.  An enterprise application of XML would need governance from several departments or information partners, and would focus on customer or compliance facing issues and possibly growth of the business. They tend to have to work within a broader framework of applications and standards.

Each of these three application types requires different planning and justification. For ad hoc use of XML it is usually up to the individual developer to decide if XML is the right format, if a schema will be needed, and what the markup and data model are, etc. Very little “selling” is needed here except as friendly debate between developers, architects and line managers. Usually these applications can be tweaked and changed easily with little impact beyond local considerations.

Departmental application of XML usually requires a team representing all stakeholders involved in the process, from users to consumers of the info. There may be some departmental architectural standards, but exceptions to these are easier to accommodate than with enterprise applications. A careful leader of a departmental application will look upstream and down stream in the information flow to include some of their needs. Also, they need to realize that the editing process in their department may become more complex and require additional skills and resources, but that these drawbacks are more than offset but savings in other areas, such as page layout, or conversion to Web formats which can be highly automated. Don’t forget to explain these benefits to the users whose work just got a little more complicated!

An Enterprise solution is by definition tied to the business drivers of the enterprise, even if that means some decisions may seem like they come at the expense of one department over another. This is where an evangelist could be useful, but not if they only focus on XML instead of the benefits it provides. Executives need to know how much revenue can be increased, how many problem reports can be avoided in customer service, and whether they can meet regulatory compliance guidelines, etc. This is a much more complicated set of issues with dependencies on and agreement with other departments needed to be successful. If you can’t provide these types of answers, you may be stuck in departmental thinking.

XML may be the center of my universe (my belly button so to speak), but it is usually not the center of my project’s sponsor’s universe. I have to have the right message to covince them to make signifiaccnt investment in the way their enterprise operates.  </>

Structured Editing & Wikis

If you know me you will realize that I tend to revisit XML authoring tools and processes frequently. It is one of my favorite topics. The intersection of structured tools and messy human thinking and behavior is an area fraught with usability issues, development challenges, and careful business case thinking. And therefore, a topic ripe for discussion.

I had an interesting conversation with a friend about word processors and XML editors the other day. His argument was that the word processing product model may not be the best, and certainly isn’t the only, way to prepare and manage structured content.

A word processor is software that has evolved to support the creation of documents. The word processing software model was developed when people needed to create documents, and then later added formatting and other features. This model is more than 25 years old (I remember using a word processor for the first time in college in 1980).

Of course it was logical to emulate how typewriters worked since the vast majority of information at the time was destined for paper documents. Now word processors include features for writing, editing, reviewing, formatting, and limited structural elements like links, indexes, etc. Again, all very document oriented. The content produced may be reused for other purposes if transformed in a post process (e.g., it could output HTML & PDF for Web, breaking into chunks for a repository or secondary use, etc.), but there are limits and other constraints, especially if your information is primarily designed to be consumed in print or document form.

It is easy to think of XML-structured editors, and the word processor software model they are based upon, as the most likely way to create structured content. But in my opinion, structured editors pay too much homage to word processing features and processes. I also think too many  project teams assume that the only way to edit XML content is in an XML document editor. Don’t get me wrong, many people have successfully deployed XML editors and achieved targeted business goals, myself included, but I can point out many instances where an alternative approach to editing content might be more efficient.

Database tools that organize the information logically and efficiently are not likely to store that data as documents. For instance, you may have an financial system with a lot of info in relational fields that is extracted to produce printable documents like monthly statements, invoices, etc.

Or software manuals that are customized for specific configurations using reusable data objects and related document maps instead maintaining the information as static, hierarchically-organized documents.

Or aircraft information that needs to match the configuration of a specific plane or tail number, selected from a complete library of data objects stored centrally.

Or statutes that start formatted as bills, then later appear as enacted laws, then later yet again as published, codified statutes, each with their own formatting and structural peccadilloes.

Or consider a travel guide publisher that collects information on thousands of hotels, restaurants, attractions, and services in dozens of countries and cities. Sure, the content is prepared with the intent of publishing it in a book, but it is easy to see how it can be useful for other uses, including providing hotel data to travel-related Web sites, or building specialized, custom booklets for special needs (e.g., a local guide for a conference, guides to historical neighborhoods, etc.).

In these examples of what some might call database publishing, system designers need to ask them selves what would be the best tool for creating and maintaining the information. They are great candidates for a database, some application dialogs and wizards, and some extraction and transformation applications to feed Web and other platforms for consumption by users. They may not even involve an editor per se, but might rely entirely a Wiki or other dialog for content creation and editing.

Word processors require a mix of skills, including domain expertise on the subject being written about, grammar and editing, and some formatting & design, use of the software itself, etc. While I personally believe everyone, not just teachers and writers, should be skilled in writing well and making documents look legible and appealing, I realize many folks are best suited for other roles. That is why we divide labor into roles. Domain experts (e.g., lawyers, aircraft engineers, scientists and doctors, etc.) are usually responsible for accuracy and quality of the ideas and information, while editorial and product support people clean up the writing and formatting and make it presentable. So, for domain experts, it may be more efficient to provide a tool that only manages the content creation, structuring, linking, organization, etc. with limited word processing capabilities, and leave the formatting and organization to the system or another department or automated style sheets.

In my mind, a Wiki is a combination of text functionality and database management features that allow content to be created and managed in a broader Web content platform (which also may include static pages, search interfaces, pictures, PDFs, etc.). In this model, the Web is the primary use and printing is secondary. Domain experts are not bothered with concepts like page layout, running heads, tables of content generation, justification & hyphenation, etc., much to the delight of the domain experts!

I am bullish on Wikis as content creation and management tools, even when the content is destined for print. I have seen some that hide much of the structure and technical “connective tissue” from the author, but produce well formatted, integrated information. The blogging tool I am using to create this article is one example of a Wiki-like interface that has a few bells and whistles for adding structure (e.g., keywords) dedicated to a specific content creation purpose. It only emulates word processing slightly with limited formatting tools, but is loaded with other features designed to improve my blog entries. For instance, I can pick a keyword from a controlled taxonomy from a pull-down list. And all within a Web browser, not a fat client editor package. This tool is optimized for making blog content, but not for, let’s say, scientific papers or repair manuals. It is targeted for a specific class of users, bloggers. Similarly, XML-editors as we have come to know them, are more adept at creating documents and document chunks than other interfaces.

Honestly, on more than one occasion I have pounded a nail with a wrench, or tightened a bolt with the wrong kind of pliers. Usually I get the same results, but sometimes it takes longer or has a less desirable result than if I had used a more appropriate tool. The same is true for editing tools.

On a final note, forgive me if I make a gratuitous plug, but authoring approaches and tools will be the subject of a panel I am chairing at the Gilbane San Francisco conference in early June if you want to hear more. </>

JustSystems Updates Content Contribution Capabilities in XMetaL Reviewer

JustSystems announced XMetaL Reviewer 5.5, the latest version of the company’s collaborative XML document reviewing software. Reinforcing the central role of document review and collaboration in the overall authoring process, XMetaL Reviewer lets subject matter experts and others contribute content directly to the documents under review. The latest version of XMetaL Reviewer also gains usability, connectivity, integration, and administration enhancements. XMetaL Reviewer moves the document review process beyond a read-only exercise that prevents reviewers from suggesting detailed edits to the documents being reviewed. Now, reviewers can actively contribute new content and rewrite entire sections of a document. Using the addendum editor, XMetaL Reviewer users with no knowledge of XML can contribute structured content, insert images, and create lists, tables, and whole sections for consideration as document modifications. In addition to harnessing the expertise of document reviewers, XMetaL Reviewer offers users a consolidated view of all their suggested changes which are displayed to all reviewers in real time, as the changes are being made. An audit trail tracks all document changes, and an inline chat feature lets reviewers discuss content changes in real time. XMetaL Reviewer 5.5 is available today with prices starting at $15,000 for the server and five concurrent user licenses. Additional concurrent licenses are $2,500 for five users. http://www.justsystems.com

XML Communities on LinkedIn

Just spent an excessive number of hours perusing the XML and related community groups on various Web community sites.  There are several social community tools and sites, but even just LinkedIn (http://www.linkedin.com/). seems to have dozens of community groups that at least mention XML in their descriptions, and several have it prominent in their names and logos. I joined several. Let’s see what happens. Stay tuned… </>

New Workshop on Implementing DITA

As part of our Gilbane Onsite Technology Strategy Workshop Series, we are happy to announce a new workshop, Implementing DITA.

Course Description

DITA, the Darwin Information Typing Architecture is an emerging standard for content creation, management, and distribution. How does DITA differ from other XML applications? Will it work for my vertical industry’s content? From technical documentation, to training manuals, from scientific papers to statutory publishing. DITA addresses one of the most challenging aspects of XML implementation, developing a data model that can be user and shared with information partners. Even so, DITA implementation requires effective process, software, and content management strategies to achieve the benefits promised by the DITA business case, cost-effective, reusable content. This seminar will familiarize you with DITA concepts and terminology, describe business benefits, implementation challenges, and best practices for adopting DITA. How DITA enables key business processes will be explored, including content management, formatting & publishing, multi-lingual localization, and reusable open content. Attendees will be able to participate in developing an effective DITA content management strategy.

Audience

This is an introductory course suitable for anyone looking to better understand DITA standard, terminology, processes, benefits, and best practices. A basic understanding of computer processing applications and production processes is helpful. Familiarity with XML concepts and publishing helpful, but not required. No programming experience required.

Topics Covered

  • The Business Drivers for DITA Adoption

  • DITA Concepts and Terminology

  • The DITA Content Model

  • Organizing Content with DITA Maps

  • Processing, Storing & Publishing DITA Content

  • DITA Creation, Management & Processing Tools

  • Multi-lingual Publishing with DITA

  • Extending DITA to work with Other Data Standards

  • Best Practices & Pitfalls for DITA Implementation

For more information and to customize a workshop just for your organization, please contact Ralph Marto by email or at +617.497.9443 x117

The Marvels of Project Guttenberg

p146.jpg

If you don’t know Project Guttenberg–and you should–it’s well worth spending your time over there to familiarize yourself with its contents and the way it has gone about creating the collection.

I keep track of it through the RSS feed of recently added books, which is updated nightly. That’s where I find out about new books like, The Pecan and its Culture, published in 1906, which includes the photo shown at left.

On their own, the one image and the one title are perhaps not so interesting or so significant (though I for one love these little snapshots of Americana, especially such primary material). What is significant of course is the mass nature of the digitization, and the care in which it is undertaken.  I compare this care with the sometimes abysmal scanning work being done by Google (and with much more fanfare). The fruits of Project Guttenberg are much more openly available, much easier to access, and much easier to migrate to reading devices like the Kindle.

So as we look at all the eBook and digitization efforts underway today, let’s not forget Project Guttenberg.

« Older posts Newer posts »

© 2024 The Gilbane Advisor

Theme by Anders NorenUp ↑