The Gilbane Advisor

Curated for content, computing, data, information, and digital experience professionals

Page 312 of 934

The Social Language

Although it is already mid-January, I would still like to wish everyone a very good 2008! It definitely looks to be an interesting year.

Back to blogging, after a very long pause. The reason was my major geographical transition: after 8 very nice years in Boston, we returned to the bi-lingual Finland and the very multi-lingual European union last autumn. The time required for a trans-Atlantic move is not to be underestimated!

Leonor’s interview with Director General Lonnroth about the languages in the EU is an excellent description of the world on this side of the Atlantic. On a very personal note, I love tuning to YLE Mondo radio every time I am driving; a local station broadcasting news from several different countries. I even get the NPR! I listen to German, French, Spanish, and Italian news, and at the same time notice the differences there are not just in the language, but also in the content. Even more fun is to listen to news from Australia and South Africa, which really change the world perspective. A good reminder that from Africa or Australia, many things do look different than from the US or from Europe. How lovely it would be to understand what they say in Chinese, Japanese or Arabic, to name just a few languages!

Anyways, things are finally starting to find their places in their new home, so I am back to blogging. We had a wonderful Gilbane conference in Boston at the end of November; it got so many ideas going in my head, especially about the social aspects of content, search, collaboration – and of course language. The question “Where are languages in social media” was asked in the conference, and the first answer was on the lines of: gee, that is a tough thing to solve. True – and yet I am convinced that we will begin to see very new types of tools and solutions. It was interesting to note that several examples were given on how in corporations social media enabled people find a language speaker inside the organization. “Through our collaboration tool, we found someone who speaks Japanes and can check our translations.” “We realized someone in our German office could translate the materials we needed.” Language skills become yet another skill to be shared in communities.

Another interesting point was that MT and its usefulness came up. With the amount of user-generated information exploding, there is no chance to human-translate everything. Could this be the real coming of age of MT?

I spoke with one multilingual service provider who said that they have started receiving requests for checking user-generated content in corporate community sites. Interesting. I would guess that need for automated checking of “bad words” increases as more content on corporate sites comes not from employees but from anyone in the web. Enterprise searches have to be multilingual, but there is always room to improve.

As Leonor pointed out: collaboration yields knowledge. That knowledge is multilingual.

Oracle to Acquire BEA Systems

Oracle Corporation (NASDAQ: ORCL) and BEA Systems (NASDAQ: BEAS) announced they have entered into a definitive agreement under which Oracle will acquire all outstanding shares of BEA for $19.375 per share in cash. The offer is valued at approximately $8.5 billion, or $7.2 billion net of BEA’s cash on hand of $1.3 billion. The Board of Directors of BEA Systems has unanimously approved the transaction. It is anticipated to close by mid-2008, subject to BEA stockholder approval, certain regulatory approvals and customary closing conditions. www.oracle.com www.bea.com

Coveo Updates Enterprise Search Technology

Coveo Solutions Inc. announced that it has added new capabilities to its Coveo Enterprise Search technology. The following new capabilities are now available: Greater scalability with support for Windows 64-bit operating systems that improves search indexing and query performance and expands access to more powerful servers; New connectors for Symantec Enterprise Vault v2 offers more flexibility and controls; Out-of-the-box integration with Microsoft Exchange and Symantec Enterprise Vault email archives allows for integrated search across all corporate email content; Re-factored connector enables indexing and search of massive email archives; enhanced connector for salesforce.com improves performance and overall user experience; easier to deploy and delivers a CRM search interface out-of-the-box; and improved performance and precision of indexing and searching. http://www.coveo.com

Nothing Like a Move by Microsoft to Stir up Analysis and Expectations

Since I weighed in last week on the Microsoft acquisition of FAST Search & Transfer, I have probably read 50+ blog entries and articles on the event. I have also talked to other analysts, received emails from numerous search vendors summarizing their thoughts and expectations about the enterprise search market and had a fair number of phone calls asking questions about what it means. The questions range from “Did Microsoft pay too much?, to “Please define enterprise search,” to “What are the next acquisitions in this market going to be going to be?” My short and flippant answers would be “No,” “Do you have a few hours?” and “Everyone and no one.”

I have seen some excellent analysis contributing relevant commentary to this discussion, some misinterpretation of what the distinction’s are between enterprise search and Web search, and some conclusions that I would seriously debate. You’ll forgive me if I don’t include links to the pieces that influenced the following comments. But one by Curt Monash in his piece on January 14 summarized the state of this industry and its already long history. It is noteworthy that while the popular technology press has only recently begun to write about enterprise search, it has been around for decades in different forms and in a short piece he manages to capture the highlights and current state.

Other commentary seems to imply that Microsoft is not really positioning itself to compete with Google because Google is really about Web (Internet) searching and Microsoft is not. This implies that FAST has no understanding of Web searching. Several points must be made:

  1. FAST Search & Transfer has been involved in many aspects of search technologies for a decade. Soon after landing on our shores it was the search engine of choice for the U.S. government’s unifying search engine to support Internet-based searching of agency Web sites by the public. Since then it has helped countless enterprises (e.g. governments, manufacturers, e-commerce companies) expose their content, products and services via the Web. FAST knows a lot about how to make Web search better for all kinds of applications and they will bring that expertise to Microsoft.
  2. Google is exploiting the Web to deliver free business software tools that directly challenge Microsoft stronghold ( e.g. email, word processing). This will not go unanswered by the largest supplier of office automation software.
  3. Google has several thousand Google Enterprise Search Appliances installed in all types of enterprises around the world, so it is already as widely deployed in enterprises in terms of numbers as FAST, albeit at much lower prices and for simpler application. That doesn’t mean that they are not satisfying a very practical need for a lot of organizations where it is “good enough.”

For more on the competition between the two check this article out.

Enterprise search has been implied to mean only search across all content for an entire enterprise. This raises another fundamental problem of perception. Basically, there are few to no instances of a single enterprise search engine being the only search solution for any major enterprise. Even when an organization “standardizes” on one product for its enterprise search, there will be dozens of other instances of search deployed for groups, divisions, and embedded within applications. Just two examples are the use of Vivisimo now used for USA.gov to give the public access to government agency public content, even as each agency uses different search engines for internal use. Also, there is IBM, which offers the OmniFind suite of enterprise search products, but uses Endeca internally for its Global Services Business enterprise.

Finally, on the issue of expectations, most of the vendors I have heard from are excited that the Microsoft announcement confirms the existence of an enterprise search market. They know that revenues for enterprise search, compared to Web search, have been miniscule. But now that Microsoft is investing heavily in it, they hope that top management across all industries will see it as a software solution to procure. Many analysts are expecting other major acquisitions, perhaps soon. Frequently mentioned buyers are Oracle and IBM but both have already made major acquisitions of search and content products, and both already offer enterprise search solutions. It is going to be quite some time before Microsoft sorts out all the pieces of FAST IP and decides how to package them. Other market acquisitions will surely come. The question is whether the next to be acquired will be large search companies with complex and expensive offerings bought by major software corporations. Or will search products targeting specific enterprise search markets be a better buy to make an immediate impact for companies seeking broader presence in enterprise search as a complementary offering to other tools. There are a lot of enterprise search problems to be solved and a lot of players to divvy up the evolving business for a while to come.

W3C Opens Data on the Web with SPARQL

W3C (The World Wide Web Consortium) announced the publication of SPARQL, the key standard for opening up data on the Semantic Web. With SPARQL query technology, pronounced “sparkle,” people can focus on what they want to know rather than on the database technology or data format used behind the scenes to store the data. Because SPARQL queries express high-level goals, it is easier to extend them to unanticipated data sources, or even to port them to new applications. Many successful query languages exist, including standards such as SQL and XQuery. These were primarily designed for queries limited to a single product, format, type of information, or local data store. Traditionally, it has been necessary to formulate the same high-level query differently depending on application or the specific arrangement chosen for the relational database. And when querying multiple data sources it has been necessary to write logic to merge the results. These limitations have imposed higher developer costs and created barriers to incorporating new data sources. The goal of the Semantic Web is to enable people to share, merge, and reuse data globally. SPARQL is designed for use at the scale of the Web, and thus enables queries over distributed data sources, independent of format. Because SPARQL has no tie to a specific database format, it can be used to take advantage of “Web 2.0” data and mash it up with other Semantic Web resources. Furthermore, because disparate data sources may not have the same ‘shape’ or share the same properties, SPARQL is designed to query non-uniform data. The SPARQL specification defines a query language and a protocol and works with the other core Semantic Web technologies from W3C: Resource Description Framework (RDF) for representing data; RDF Schema; Web Ontology Language (OWL) for building vocabularies; and Gleaning Resource Descriptions from Dialects of Languages (GRDDL), for automatically extracting Semantic Web data from documents. SPARQL also makes use of other W3C standards found in Web services implementations, such as Web Services Description Language (WSDL). http://www.w3.org/

Atlassian Partners with EditGrid to Expand Features of Hosted Wiki

EditGrid, the online spreadsheet, is now available to customers that use the hosted versions of Confluence, the enterprise wiki from Atlassian. The EditGrid Plugin for Confluence allows for real-time collaboration of spreadsheets within Confluence Hosted or Confluence Enterprise Hosting. With EditGrid, users can create or insert spreadsheets into Confluence pages. They can also edit the spreadsheet collaboratively. The resulting spreadsheet is saved as an attachment within the Confluence pages in Microsoft Excel format, allowing Confluence to manage the revision history. Some of the features of the EditGrid plugin include: Real-time updates – allows multiple users to see dynamic changes to a spreadsheet; Remote data update – retrieves live financial data on the Web and stores it in a spreadsheet; Import and export: accepts file formats such as Microsoft Excel, CSV, HTML, Gnumeric, Lotus, OpenOffice and assigns fine-grained access control; and Live chat – enables multiple users to discuss changes from within EditGrid, no need to switch to another chat application. EditGrid is free for Confluence Hosted and Confluence Enterprise Hosting customers, and it is available starting today. For more information please visit, http://www.atlassian.com, http://www.editgrid.com

MadCap Software Debuts MadCap Lingo & MadCap Analyzer

MadCap Software announced MadCap Lingo, an XML-based, integrated translation memory system and authoring tool, aimed at eliminating the need for file transfers in order to complete translation. Document components, such as tables of content, topics, index keywords, concepts, glossaries, and variables all remain intact throughout the translation and localization process, so there is never a need to recreate them. MadCap Lingo also is integrated with MadCap Flare and MadCap Blaze, and it is Unicode enabled to help documentation professionals deliver a consistent user experience in print, online, and in any language. MadCap Lingo is being announced in conjunction with the new MadCap Analyzer, software that proactively recommends documentation content and design improvements. MadCap Lingo works with MadCap Flare, the company’s native-XML authoring product, and MadCap Blaze, the native-XML tool for publishing long print documents, which will be generally available in early 2008. A user creates a MadCap Lingo project to access the source content in a Flare or Blaze project via a shared file structure. Working through Lingo’s interface, the user accesses and translates the content. Because the content never actually leaves the structure of the original Flare or Blaze project, all the content and formatting is preserved in the translated version. Once a project is translated, it is opened in either Flare or Blaze, which generates the output and facilitates publishing. At the front end of the process, Flare and Blaze can import a range of document types to create the source content. Following translation, the products provide single-source delivery to multiple formats online and off, including the Internet, intranets, CDs, and print. MadCap Lingo is available and is priced at $2,199 per license, but is available at an introductory price of $899 for a limited time. MadCap Lingo also is available on a subscription basis for $649 per year. Fees for support start at $449 per year. http://www.madcapsoftware.com/

« Older posts Newer posts »

© 2025 The Gilbane Advisor

Theme by Anders NorenUp ↑