The Gilbane Advisor

Curated for content, computing, data, information, and digital experience professionals

Page 217 of 936

Classifying Searchers – What Really Counts?

I continue to be impressed by the new ways in which enterprise search companies differentiate and package their software for specialized uses. This is a good thing because it underscores their understanding of different search audiences. Just as important is recognition that search happens in a context, for example:

  • Personal interest (enlightenment or entertainment)
  • Product selection (evaluations by independent analysts vs. direct purchasing information)
  • Work enhancement (finding data or learning a new system, process or product)
  • High-level professional activities (e-discovery to strategic planning)

Vendors understand that there is a limited market for a product or suite of products that will satisfy every budget, search context and the enterprise’s hierarchy of search requirements. Those who are the best focus on the technological strengths of their search tools to deliver products packaged for a niche in which they can excel.

However, for any market niche excellence begins with six basics:

  • Customer relationship cultivation, including good listening
  • Professional customer support and services
  • Ease of system installation, implementation, tuning and administration
  • Out-of-the box integration with complementary technologies that will improve search
  • Simple pricing for licensing and support packages
  • Ease of doing business, contracting and licensing, deliveries and upgrades

While any mature and worthy company will have continually improved on these attributes, there are contextual differentiators that you should seek in your vertical market:

  • Vendor subject matter expertise
  • Vendor industry expertise
  • Vendor knowledge of how professional specialists perform their work functions
  • Vendor understanding of retrieval and content types that contribute the highest value

At a recent client discussion the application of a highly specialized taxonomy was the topic. Their target content will be made available on a public facing web site and also to internal staff. We began by discussing the various categories of terminology already extracted from a pre-existing system.

As we differentiated how internal staff needed to access content for research purposes and how the public is expected to search, patterns emerged for how differently content needs to be packaged for each constituency. For you who have specialized collections to be used by highly diverse audiences, this is no surprise. Before proceeding with decisions about term curation and determining the granularity of their metadata vocabulary, what has become a high priority is how the search mechanisms will work for different audiences.

For this institution, internal users must have pinpoint precision in retrieval on multiple facets of content to get to exactly the right record. They will be coming to search with knowledge of the collection and more certainty about what they can expect to find. They will also want to find their target(s) quickly. On the other hand, the public facing audience needs to be guided in a way that leads them on a path of discovery, navigating through a map of terms that takes them from their “key term” query through related possibilities without demanding arcane Boolean operations or lengthy explanations for advanced searching.

There is a clear lesson here for seeking enterprise search solutions. Systems that favor one audience over another will always be problematic. Therefore, establishing who needs what and how each goes about searching needs to be answered, and then matched to the product that can provide for all target groups.

We are in the season for conferences; there are a few next month that will be featuring various search and content technologies. After many years of walking exhibit halls and formulating strategies for systematic research and avoiding a swamp of technology overload, I try now to have specific questions formulated that will discover the “must have” functions and features for any particular client requirement. If you do the same, describing a search user scenario to each candidate vendor, you can then proceed to ask: Is this a search problem your product will handle? What other technologies (e.g. CMS, vocabulary management) need to be in place to ensure quality search results? Can you demonstrate something similar? What would you estimate the implementation schedule to look like? What integration services are recommended?

These are starting points for a discussion and will enable you to begin to know whether this vendor meets the fundamental criteria laid out earlier in this post. It will also give you a sense of whether the vendor views all searchers and their searches as generic equivalents or knows that different functions and features are needed for special groups.

Look for vendors for enterprise search and search related technologies to interview at the following upcoming meetings:

Enterprise Search Summit, New York, May 10 – 11 […where you will learn strategies and build the skill sets you need to make your organization’s content not only searchable but “findable” and actionable so that it delivers value to the bottom line.] This is the largest seasonal conference dedicated to enterprise search. The sessions are preceded by separate workshops with in-depth tutorials related to search. During the conference, focus on case studies of enterprises similar to yours for better understanding of issues, which you may need to address.

Text Analytics Summit, Boston, May 18 – 19 I spoke with Seth Grimes, who kicks off the meeting with a keynote, asking whether he sees a change in emphasis this year from straight text mining and text analytics. You’ll have to attend to get his full speech but Seth shared that he see a newfound recognition that “Big Data” is coming to grips with text source information as an asset that has special requirements (and value). He also noted that unstructured document complexities can benefit from text analytics to create semantic understanding that improves search, and that text analytics products are rising to challenge for providing dynamic semantic analysis, particularly around massive amounts of social textual content.

Lucene Revolution, San Francisco, May 23 – 24 […hear from … the foremost experts on open source search technology to a broad cross-section of users that have implemented Lucene, Solr, or LucidWorks Enterprise to improve search application performance, scalability, flexibility, and relevance, while lowering their costs.] I attended this new meeting last year when it was in Boston. For any enterprise considering or leaning toward implementing open source search, particularly Lucene or Solr, this meeting will set you on a path for understanding what that journey entails.

Adobe Introduces Creative Suite 5.5; Subscription Pricing

Adobe has introduced Creative Suite 5.5, a bundled package that offers all of Adobe’s designing and editing software in a single package. Adobe also planning to offer software rental options, allowing individuals to rent Adobe software, as well as plans for enterprise and student users. The Flash maker is considering offering its software to rent both on an annual basis, paid monthly, or a (more expensive) month-to-month payment plan. Rented software will check into Adobe servers every 30 days in order to check the credit card being used by the user is still valid and to make an automatic payment. If the card is not valid or has been removed by the user, the user gets a 5-day grace period for making the payments after which the software will cease to operate. http://www.adobe.com/

iBuildApp for iPad: Free Mobile Digital Publishing Interface

The iBuildApp solution has templates, is automated, and totally free to create and update, and will be integrated with blogging and CMS platforms for easy publishing. The fully functional publishing app takes just 2-3 hours to create and publish content. Just copy/paste content into pre-made templates for the iPad or add media from other CMS platforms such as WordPress, Drupal and Joomla. To integrate it requires snippets of code from iBuildApp to be inserted into the CMS code. By using this service, publishers can focus on their content and leave the formatting, publishing and distribution to iBuildapp. The finished product is built automatically by iBuildApp’s platform and all of the content, is hosted on our platform and can be updated quickly – Free of charge and no human intervention required. http://ibuildapp.com/

Nuxeo Releases BIRT Integration Package for Nuxeo ECM Applications

Nuxeo, the Open Source Enterprise Content Management (ECM) platform company, announced the availability of a new package in the Nuxeo Marketplace that connects Nuxeo Enterprise Platform (EP)-based applications and the open source Business Intelligence tool, Eclipse BIRT. The Nuxeo – BIRT Integration offers report design and access from within a Nuxeo ECM application. The Nuxeo – BIRT Integration offers integrated reporting capabilities on top of Nuxeo EP-based applications, enabling the design, rendering, and access of reports and analytics. The BIRT Engine has been integrated in Nuxeo EP, so that reporting capabilities are available without leaving the Nuxeo application, and security controls can be managed in one place. The Nuxeo – BIRT Integration package is available on the Nuxeo Marketplace. http://www.nuxeo.com/ http://www.eclipse.org/birt/phoenix/

Just Published: Outsell Gilbane Study on Multilingual Marketing Content

Our 2011 report describing the current state of practice for globalizing multilingual marketing content is available now through March 31 exclusively through study sponsors  Across Systems, ADAM Software, Lionbridge, and SDL.

Multilingual Marketing Content: Growing International Business With Global Content Value Chains features a major update of the global content value chain, Gilbane’s framework for helping companies plan and manage their globalization practices. The new value chain adds core competencies to the existing functional view of multilingual content processes, and it clearly ties the value chain to business outcomes.

Study data includes top business goals and objectives and the investments that marketing and localization managers are making in programs and initiatives that support those goals. The analysis covers what marketing organizations can learn from product content groups, who are generally further along the content globalization maturity curve.

The report will be available directly from the Gilbane website starting April 1. In the meantime, please visit a sponsor site to access the study, and check this blog for research highlights and insights.

e-Spirit Integrates FirstSpirit CMS Into Liferay Portal

e-Spirit AG has added the open source Liferay Portal to their range of possibilities for integration of the FirstSpirit content management system into enterprise portals. The module was developed in cooperation with e-Spirit’s technology partner USU. Integrating FirstSpirit in Liferay will allow organizations to create employee portals that combine Enterprise 2.0 functionality, IT applications, content and collaboration. Organizations will also be able to provide their employees with access to Web 2.0 functions such as forums, blogs and wikis and offering them a platform to efficiently organize collaboration and share information beyond individual departments. The new module will be available in May. http://www.e-spirit.com http://www.liferay.com/

ETL and Building Intelligence Behind Semantic Search

A recent inquiry about a position requiring ETL (Extraction/Transformation/Loading) experience prompted me to survey the job market in this area. It was quite a surprise to see that there are many technical positions seeking this expertise, plus experience with SQL databases, and XML, mostly in healthcare, finance or with data warehouses. I am also observing an uptick in contract positions for metadata and taxonomy development.

My research on Semantic Software Technologies placed me on a path for reporters and bloggers to seek my thoughts on the Watson-Jeopardy story. Much has been written on the story but I wanted to try a fresh take on the meaning of it all. There is a connection to be made between the ETL field and building a knowledgebase with the smarts of Watson. Inspiration for innovation can be drawn from the Watson technology but there is a caveat; it involves the expenditure of serious mental and computing perspiration.

Besides baked-in intelligence for answering human questions using natural language processing (NLP) to search, an answer-platform like Watson requires tons of data. Also, data must be assembled in conceptually and contextually relevant databases for good answers to occur. When documents and other forms of electronic content are fed to a knowledgebase for semantic retrieval, finely crafted metadata (data describing the content) and excellent vocabulary control add enormous value. These two content enhancers, metadata and controlled vocabularies, can transform good search into excellent search.

The irony of current enterprise search is that information is in such abundance that it overwhelms rather than helps findability. Content and knowledge managers can’t possibly contribute the human resources needed to generate high quality metadata for everything in sight. But there are numerous techniques and technologies to supplement their work by explicitly exploiting the mountain of information.

Good content and knowledge managers know where to find top quality content but may not know that, for all common content formats, there are tools to extract key metadata embedded (but hidden) in it. Some of these tools can also text mine and analyze the content for additional intelligent descriptive data. When content collections are very large but too small to justify (under a million documents) the most sophisticated and complex semantic search engines, ETL tools can relieve pressure on metadata managers by automating a lot of mining, extracting entities and concepts needed for good categorization.

The ETL tool array is large and varied. Platform tools from Microsoft (SSIS) and IBM (DataStage) may be employed to extract, transform and load existing metadata. Other independent products such as those from Pervasive and SEAL may contribute value across a variety of platforms or functional areas from which content can be dramatically enhanced for better tagging and indexing. The call for ETL experts is usually expressed in terms of engineering functions who would be selecting, installing and implementing these products. However, it has to be stressed that subject and content experts are required to work with engineers. The role of the latter is to help tune and validate the extraction and transformation outcomes, making sure terminology fits function.

Entity extraction is one major outcome of text mining to support business analytics, but tools can do a lot more to put intelligence into play for semantic applications. Tools that act as filters and statistical analyzers of text data warehouses will help reveal terminology for use in building specialized controlled vocabularies for use in auto-categorization. A few vendors that are currently on my radar to help enterprises understand and leverage their content landscape include EntropySoft Content ETL, Information Extraction Systems, Intelligenx, ISYS Document Filters, RAMP, and XBS, something here for everyone.

The diversity of emerging applications is a leading indicator that there is a lot of innovation to come with all aspects of ETL. While RAMP is making headway with video, another firm with a local connection is Inforbix. I spoke with a co-founder, Oleg Shilovitsky for my semantic technology research last year before they launched. As he then asserted, it is critical to preserve, mine and leverage the data associated with design and manufacturing operations. This area has huge growth potential and Inforbix is now ready to address that market.

Readers who seek to leverage ETL and text mining will gain know-how from the cases presented at the 2011 Text Analytics Summit, May 18-19 in Boston. As well, the exhibits will feature products to consider for making piles of data a valuable knowledge asset. I’ll be interviewing experts who are speaking and exhibiting at that conference for a future piece. I hope readers will attend and seek me out to talk about your metadata management and text mining challenges. This will feed ideas for future posts.

Finally, I’m not the only one thinking along these lines. You will find other ideas and a nudge to action in these articles.

Boeri, Bob. Improving Findability Behind the Firewall, 28 slides. Enterprise Search Summit 2010, NY, 05/2010.
Farrell, Vickie. The Need for Active Metadata Integration: The Hard Boiled Truth. DM Direct Newsletter, 09/09/2005, 3p
McCreary, Dan. Entity Extraction and the Semantic Web, Semantic Universe, 01/01/2009
White, David. BI or bust? KMWorld, 10/28/2009, 3p.

Enterprise Edition of Adobe Digital Publishing Suite Now Available

Adobe Systems Incorporated announced the immediate availability of the Enterprise Edition of Adobe Digital Publishing Suite, a set of hosted software services and viewer technology to create, distribute, monetize and analyze digital magazines, newspapers and publications. With output aimed at Android tablets, RIM PlayBook, and iOS tablet devices, the Enterprise Edition is designed for large publishers to implement a custom tablet publishing solution without disrupting existing processes and infrastructure. Today’s news follows the announcement that Adobe Digital Publishing Suite will support both Apple App Store Subscriptions and Google One Pass for magazine and newspaper publishers. http://www.adobe.com/

« Older posts Newer posts »

© 2026 The Gilbane Advisor

Theme by Anders NorenUp ↑