Federated Media Publishing, a “next-generation” media company, announced the acquisition of a platform for semantic and linguistic profiling of web-based content from TextDigger, a San Jose-based semantic search startup. FM provides a full suite of media and marketing services for brand advertisers that depends heavily on a proprietary media and marketing technology platform. TextDigger’s technology complements FM’s platform with a set of semantic solutions for content tagging, filtering and clustering, as well as related tools that enhance the user experience, ad targeting, and semantic search engine optimization for a site. TextDigger will continue its search business, all of TextDigger’s customers will continue to be supported by either FM or TextDigger, depending on the type of project or service. www.federatedmedia.net www.textdigger.com
Page 222 of 931
Mining content for facts and information relationships is a focal point of many semantic technologies. Among the text analytics tools are those for mining content in order to process it for further analysis and understanding, and indexing for semantic search. This will move enterprise search to a new level of research possibilities.
Research for a forthcoming Gilbane report on semantic software technologies turned up numerous applications used in the life sciences and publishing. Neither semantic technologies nor text mining are mentioned in this recent article Rare Sharing of Data Leads to Progress on Alzheimer’s in the New York Times but I am pretty certain that these technologies had some role in enabling scientists to discover new data relationships and synthesize new ideas about Alzheimer’s biomarkers. The sheer volume of data from all the referenced data sources demands computational methods to distill and analyze.
One vertical industry poised for potential growth of semantic technologies is the energy field. It is a special interest of mine because it is a topical area in which I worked as a subject indexer and searcher early in my career. Beginning with the 1st energy crisis, oil embargo of the mid-1970s, I worked in research organizations that involved both fossil fuel exploration and production, and alternative energy development.
A hallmark of technical exploratory and discovery work is the time gaps between breakthroughs; there are often significant plateaus between major developments. This happens if research reaches a point that an enabling technology is not available or commercially viable to move to the next milestone of development. I observed that the starting point in the quest for innovative energy technologies often began with decades-old research that stopped before commercialization.
Building on what we have already discovered, invented or learned is one key to success for many “new” breakthroughs. Looking at old research from a new perspective to lower costs or improve efficiency for such things as photovoltaic materials or electrochemical cells (batteries) is what excellent companies do.
How does this relate to semantic software technologies and data mining? We need to begin with content that was generated by research in the last century; much of this is just now being made electronic. Even so, most of the conversion from paper, or micro formats like fîche, is to image formats. In order to make the full transition to enable data mining, content must be further enhanced through optical character recognition (OCR). This will put it into a form that can be semantically parsed, analyzed and explored for facts and new relationships among data elements.
Processing of old materials is neither easy nor inexpensive. There are government agencies, consortia, associations, and partnerships of various types of institutions that often serve as a springboard for making legacy knowledge assets electronically available. A great first step would be having DOE and some energy industry leaders collaborating on this activity.
A future of potential man-made disasters, even when knowledge exists to prevent them, is not a foregone conclusion. Intellectually, we know that energy independence is prudent, economically and socially mandatory for all types of stability. We have decades of information and knowledge assets in energy related fields (e.g. chemistry, materials science, geology, and engineering) that semantic technologies can leverage to move us toward a future of energy independence. Finding nuggets of old information in unexpected relationships to content from previously disconnected sources is a role for semantic search that can stimulate new ideas and technical research.
A beginning is a serious program of content conversion capped off with use of semantic search tools to aid the process of discovery and development. It is high time to put our knowledge to work with state-of-the-art semantic software tools and by committing human and collaborative resources to the effort. Coupling our knowledge assets of the past with the ingenuity of the present we can achieve energy advances using semantic technologies already embraced by the life sciences.
IBM announced the acquisition of Datacap Inc., a privately-held company based in Tarrytown, NY. Datacap is a provider of software that enables organizations to transform the way they capture, manage and automate the flow of business information to improve business processes, reduce paper costs or manual errors and meet compliance mandates. Financial terms were not disclosed. The acquisition strengthens IBM’s ability to help organizations digitize, manage and automate their information assets, particularly in paper-intensive industries such as healthcare, insurance, government and finance. Additionally, regulations such as HIPAA and Sarbanes-Oxley have demanded new standards and now legislation is encouraging the adoption of new records management solutions, including scanning and capture to increase accuracy, lower costs and speed business processes to meet these regulations. http://ibm.com http://www.datacap.com/
Yesterday, it was announced that another CMS poster child of the late 90’s is to be acquired as Adobe Systems Incorporated and Day Software Holding AG announced the two companies have entered into a definitive agreement for Adobe to acquire all of the publicly held registered shares of Day Software in a transaction worth approximates US$240 million.
This follows Adobe’s acquisition of Omniture late last year and clearly demonstrates their intent in entering the web experience management (WEM) market place that we cover with interest here at Gilbane – as we anticipate they bring together the audience insight gained through the web analytics of Omniture and Day’s CRX content platform.
This will presumably add momentum to Day’s own move into the WEM space with their recent product marketing strategy, as they have reinvented themselves to be closer to the marketer with recent attention paid to functionality such as personalization, analytics, variant testing and messaging around using their repository for marketing campaigns and asset management. We await with interest firm integration plans.
In addition Day are a longtime advocate of CMS repository standards (JCR and CMIS), something that is also close to our heart at Gilbane. This announcement has also sent tremors through the Open Source community, as they wonder about Adobe’s commitment to the Apache projects like Sling and Jackrabbit that Day have been so supportive of.
Whilst Adobe and Day have been very quick to state that they will maintain Day’s commitment to these community projects, it’s hard not think that this commitment inside Day is cultural and we wonder whether this can realistically be maintained as the acquisition matures and Day is brought into the fold.
The acquisition also raises questions about what this means for Alfresco’s two year relationship with Adobe that runs pretty deep with OEM integration to Adobe LiveCycle – and Erik Larson (Senior Director of Product Management at Adobe) has publically stated the intention to integrate Day and LifeCycle to create a ‘full suite of enterprise technologies’. It will be important for the Adobe customers that have adopted the Alfresco based integration, to understand how this will affect them going forward.
One other area that I am sure my colleagues here at Gilbane in the Publishing Technologies practice will be watching with interest is the impact this will have on Adobe’s digital publishing offering.
As we’ve seen with previous acquisitions, it’s best to be cautious over what the future might hold. From a WEM product strategy perspective bringing Ominture and Day together makes a great deal of sense to us. The commitment to standards and open source projects is probably safe for now, it has been a part of the Day identity and value proposition for as long as I can remember and one of the most exciting things could be what this acquisition means for digital publishing.
Let’s wait and see…
Suggested further reading:
- Adobe Press Release
- Day Software FAQ
- CMSWire – Perspectives: What the Adobe + Day Software Deal Means
- Real Story Group – Adobe To Acquire Day – First Take ECM Perspective
- Adobe Buys ECM Vendor (and Apache Contributor) Day Software
Adobe Systems Incorporated and Day Software Holding AG announced the two companies have entered into a definitive agreement for Adobe to launch a public tender offer to acquire all of the publicly held registered shares of Day Software, for about $240 million. Adobe’s acquisition of Day will strengthen the company’s enterprise software solutions with Web Content Management (WCM), Digital Asset Management and Social Collaboration offerings. Day’s lweb solutions combined with Adobe’s existing enterprise portfolio will enable customers to better integrate their global web presence and business applications, unlocking value across their marketing, sales and service processes. In addition, Day customers will be able to leverage more interactive application and document capabilities from Adobe AIR, Adobe Flash, Flex, Adobe LiveCycle and PDF. http://www.adobe.com/ http://www.day.com
Version 6.0 of the Syncro SVN Client brings improved performance, a new application interface and a complete redesign of the Working Copy layout. New features in version 6.0: Redesigned application layout; Configurable Working Copy view; Performance improvements; and Synchronized history view. http://www.syncrosvnclient.com
It is not news that enterprise search has been relegated to the long list of failed technologies by some. We are at the point where many analysts and business writers have called for a moratorium on the use of the term. Having worked in a number of markets and functional areas (knowledge management/KM, special libraries, and integrated library software systems) that suffered the death knell, even while continuing to exist, I take these pronouncements as a game of sorts.
Yes, we have seen the demise of vinyl phonograph records, cassette tapes and probably soon musical CD albums, but those are explicit devices and formats. When you can’t buy or play them any longer, except in a museum or collector’s garage, they are pretty dead in the marketplace. This is not true of search in the enterprise, behind the firewall, or wherever it needs to function for business purposes. People have always needed to find “stuff” to do their work. KM methods and processes, special libraries and integrated library systems still exist, even as they were re-labeled for PR and marketing purposes.
What is happening to search in the enterprise is that it is finding its purpose, or more precisely its hundreds of purposes. It is not a monolithic software product, a one-size-fits-all. It comes in dozens of packages, models, and price ranges. It may be embedded in other software or standalone. It may be procured for a point solution to support retrieval of content for one business unit operating in a very narrow topical range, or it may be selected to give access to a broad range of documents that exist in numerous enterprise domains on many subjects.
Large enterprises typically have numerous search solutions in operation, implementation, and testing, all at the same time. They are discovering how to deploy and leverage search systems and they are refining their use cases based on what they learn incrementally through their many implementations. Teams of search experts are typically involved in selecting, deploying and maintaining these applications based on their subject expertise and growing understanding of what various search engines can do and how they operate.
After years of hearing about “the semantic Web,” the long sought after “holy grail” of Web search, there is a serious ramping of technology solutions. Most of these applications can also make search more semantically relevant behind the firewall. These technologies have been evolving for decades beginning with so-called artificial intelligence, and now supported by some categories of computational linguistics such as specific algorithms for parsing content and disambiguating terms. A soon to-be released study featuring some of noteworthy applications reveals just how much is being done in enterprises for specific business purposes.
With this “teaser” on what is about to be published, I leave you with one important thought, meaningful search technologies depend on rich linguistically-based technologies. Without a cornucopia of software tools to build terminology maps and dictionaries, analyze content linguistically in context to elicit meaning, parse and evaluate unstructured text data sources, and manage vocabularies of ever more complex topical domains, semantic search could not exist.
Language complexities are challenging and even vexing. Enterprises will be finding solutions to leverage what they know only when they put human resources into play to work with the lingo of their most valuable domains.
memoQ 4.2, by Kilgray, was integrated into XTRF 2.0, the management system for translation companies and corporate translation departments. XTRF supports the work of translation departments in three fields: management and administration of all company activities, management of workflow and of the production process, as well as management of the translation process. The integration of memoQ enables managing complete translation projects including from the very beginning until the translation is finished. http://www.xtrf.eu/ http://www.kilgray.com