The Gilbane Advisor

Curated for content, computing, data, information, and digital experience professionals

Page 326 of 935

Sharepoint and Search

Sharepoint repositories are a prime content target for most search engines in the enterprise search arena, judging from the number of announcements I’ve previewed from search vendors in the last six month. This list is long and growing (Names link to press releases or product pages for Sharepoint search enabling):

Almost a year ago I began using a pre-MOSS version of Sharepoint to collect documents for a team activity. Ironically, the project was the selection, acquisition, implementation of a (non-Sharepoint) content management system to manage a corporate intranet, extranet, and hosted public Web site. The version of Sharepoint that was “set up” for me was strictly out of the box. Not being a development, I was still able to muddle my way through setting up the site, established users, posting announcements and categories of content to which I uploaded about fifty or sixty documents.

The most annoying discovery was the lack of a default search option. Later updating to MOSS solved the problem but at the time it was a huge aggravation. Because I could not guarantee a search option would appear soon enough, I had to painstakingly create titles with dates in order to give team members a contextual description as they would browse the site. Some of the documents I wanted to share were published papers and reviews of products. Dates were not too relevant for those, so I “enhanced” the titles with my own notations to help the finders select what they needed.

These silly “homemade” solutions are not uncommon when a tool does not anticipate how we would want to be able to use it. They persist as ways to handle our information storage and retrieval challenges. Since the beginning of time humans have devised ways to store things that they might want to re-use at some point in the future. Organizing for findability is an art as much at it is science. Information science only takes one so far in establishing the organizing criteria and assigning those criteria to content. Search engines that rely strictly on the author’s language will leave a lot of relevant content on the shelf for the same reasons as using Dewey Decimal classification without the complementary card catalog of subject topics. The better search engines exploit every structured piece of data or tagged content associated with a document, and that includes all the surrounding metadata assigned by “categorizers.” Categorizers might be artful human indexers or automated processes. Search engines with highly refined, intelligent categorizers to enable semantically rich finding experiences bring even more sophistication to the search experience.

But back to Sharepoint, which does have an embedded search option now, I’ve heard more than one expert comment on the likelihood that it will not be the “search” of choice for Sharepoint. That is why we have so many search options scrambling to promote their own Sharepoint search. This is probably because the organizing framework around contributing content to Sharepoint is so loosey goosey that an aggregation of many Sharepoint sites across the organization will be just what we’ve experienced with all these other homegrown systems – a dump full of idiosyncratic organizing tricks.

What you want to do, thoughtfully, is assess whether the search engine you need will share only Sharepoint repositories OR both structured and unstructured repositories across a much larger domain of types of content and applications. It will be interesting to evaluate the options that are out there for searching Sharepoint gold mines. Key questions: Is a product targeting only Sharepoint sites or diverse content? How will content across many types of repositories be aggregated and reflected organized results displays? How will the security models of the various repositories interact with the search engine? Answering these three questions first will quickly narrow your list of candidates for enterprise search.

Mashing-up the Wikipedia Code

Now here’s an interesting tidbit from the BBC, courtesy of my daughter (who’s a graduate student in London): Wikipedia ‘shows CIA page edits.’ It seems that staffers at the CIA, the Democratic National Campaign Committee, the Vatican, and many other well known institutions (who may be trying to remain nameless) have been ‘caught’ sprucing up various wikipedia articles. (Well of course this is a tarty British take on the matter!)
And the secret sauce that pulls back the curtain? Revealed at the end of the article, a simple mashup that links the IP addresses of contributors to an article (obtained through the “history” page) with a directory of organizations owning IP addresses. Both are publicly available. The results are hardly surprising.
The point is that when information is so widely and freely available, we have to begin to worry about the sources of information and how it is presented. There’s not a lot of anonymity on the public web — and quite possibly this is a good thing. But building community also includes notions of trust, expertise, and terms of reference. For example, when starting eBay, Pierre Omidyar came up with the notion of “rate the buyer” and “rate the seller” as a way of organically building trust within the community of eBayers . . . and the rest is history.
I hate to admit it but perhaps Ronald Regan said it the best. “Trust but verify.” What’s interesting is that mashing-up sources and IP addresses provides a whole new dimension to verification. I wonder what else is possible? Let’s start a discussion — comments?

Movable Type Releases Movable Type 4.0

Movable Type announced the release of Movable Type 4.0. “This is the biggest release of MT ever, a complete redesign of both the front end information architecture and the back end scaling infrastructure.” Movable Type 4 has a broad set of new capabilities, including: a redesigned user interface, more and better plugins, built in support for OpenID, community features, ability to aggregate content from multiple blogs, new support for standalone pages in addition to blog entries, “content management” features, smarter archiving (e.g., by author), more robust templates, and more. They also announced an upcoming open source version. http://www.movabletype.com/blog/2007/08/presenting-movable-type-40.html, http://www.movabletype.com/

2007 Gilbane Survey on the WCM User Experience

This is just a quick reminder to our Analyst on Demand subscribers that the results of our survey on the usability of commercially available content management solutions will be available in early September. The data will come directly from the feedback of the solution providers’ customers. Vendors covered will include Interwoven, Tridion, Vignette, FatWire, Percussion, RedDot, EMC/Documentum, CrownPeak, Mediasurface, PaperThin, Oracle, Day, Hot Banana, Clickability, Acumium, and others.

Poll of the Week: Got Process Bottlenecks?

We have not heard of an organization that doesn’t.

Content management and translation management each have their own set of process bottlenecks. Put them together and what do you get? An endless migraine, a major headache, a dull pain, and for the very few, a nuisance. Here’s some of the phrases we hear when we talk to our clients about the content and translation lifecycle:

  • “Undesired repetition and unpredictable outcomes.”
  • “A cost we don’t really have a handle on.”
  • “We’d have to survey each workgroup to figure it out.”
  • “Redundant, cumbersome, and expensive.”

Hence, the poll of the week. We’re gearing up for the Global Content Management track at Gilbane Boston, November 27-29. Our goal is to spend more time discussing the elimination of process bottlenecks rather than bemoaning their existence.

Help us shape the list for our sessions and discussions in Boston by taking our poll of the week. Got process bottlenecks? We want to know about them.

I’ve Got Infrastructure on My Mind

It’s been a rough few weeks for infrastructure.
Of course the collapse of the I-35 bridge over the Mississippi in Minneapolis last Wednesday is on all of our minds – how could the inspections fail and the road fall down? Significantly, there’s some videos that capture the moment, which hopefully will provide clues for determining the cause.
Closer to home, we had an eight-hour traffic jam on the I-93 loop in Braintree (a major highway south of Boston) a week ago Monday. A storm grate was thrown loose by a passing truck at the start of the morning rush, and landed on a near-by car. (Fortunately the driver survived.) Reportedly, the Massachusetts State Highway Department spent the rest of the day checking and welding shut all the grates on that highway. The next day, the same loose storm drain problem cropped up on a major road in Newton, near where I live. This time motorists were asked to dial a special code from their cell phones to report problems.
And then this morning New Yorkers awoke to a monsoon and a flooded mass transit system. The official M.T.A. web site could not keep up with the requests for information, and crashed when it was needed most.
You’ve gotta hand it to those hardy folks (and the New York Times) for that snarky, Big Apple attitude. Here’re a few priceless ones that I gleaned from NYTimes.Com during the day.

“Our transit system is not only frail but if it’s this vulnerable to rain attacks, then how vulnerable is it to terrorist attacks?”

“I walked from the Upper West Side all the way to work in Midtown, but thanks to Starbucks was able to stop in and cool myself every four blocks or so.”

(Now there’s somebody with brand loyalty!)

“And only the rats had no transportation problems.”

All this content about our physical infrastructure (user generated and otherwise) has the potential to bring social computing to a whole new level.
This got me thinking about the role of collaboration technologies for supporting our physical infrastructure. It’s great to be able to talk back — and let off a little steam. It’s even better to be able to call-in, and tell the authorities about the problem before there’s another horrible accident. But what else is possible? Could the bridge inspectors in Minneapolis have shared their observation reports, measurements, and perhaps photographs of the bridge’s structure over the past few inspection cycles, and had some semi-automated ways to detect the problems before the disaster? Unfortunately we’ll never know.
While I certainly don’t have it all figured out, I can begin to see some bread crumbs towards the workable solution we all want and need. We can no longer rely on human intelligence alone. Our worlds are much too complex and interdependent. We need to augment our understandings, and our abilities to take actions, by a variety of automated, concent-centric tools, such as semantic technologies. (With my colleagues Lynda Moulton and Frank Gilbane, we’re picking up coverage of this area, the ability to inject “meaning” and “context” into an enterprise environment. Be sure to check out the semantic technologies track at our upcoming conference in November.)
I’ve seen a couple of promising developments this month. SchemaLogic is finally reporting some progress in the publishing space, enabling publishers such as Associated Press to automatically repurpose content by synchronizing tags and managing metadata schemas. While pretty geeky, this is very neat! Now we need to see how this approach to managing semantics within the enterprise will impact collaboration and social computing.
Then project and portfolio management (PPM) systems — heretofore heavyweight (often mainframe) applications that are used to track resources for complex, engineering-driven projects — are being redeployed as Web 2.0 environments. In particular,eProject is now transforming its Web-based PPM environment into a broader collaborative tools suite. Seeking to capitalize on it’s expanded mission of bringing a PPM model to the Web, eProject’s also renaming itself in the process.
Where do these bread-crumbs lead? As a first step, we need to focus on how our collaboration infrastructure (fueled by our information architecture) can augment the work of people responsible for our physical infrastructure (ourselves included). At the end of the day, we need to be able to rely on this collaboration infrastructure to help us sense and respond to the challenges of simply getting from one place to another.

Recommendation to IT Directors: Constantly Track WCM Applications and Their Feature Sets

In recent conversations with several of Gilbane’s Analyst On Demand and Technology Acquisition Advisory clients, I have observed two careless practices that have prevented enterprises from being able to assess both the feature-functionality of their existing WCM applications and their requirements for selecting solutions to replace those applications. Both relate to a lack of documentation.
In the first case, it’s the absence of a master list of the WCM-related applications that have been developed in-house over the years. One company has “about 50” such applications, and geographically-dispersed individuals throughout the enterprise can tell me what some of them are, but no one can refer me to anyone or any system that has the complete listing. Discrete ongoing development projects exist for many of these applications, a few of which live buried deep in departmental silos. Needless to say, the functionality of applications within these silos is known only to a few people, is never re-used in other initiatives, and in fact often gets duplicated by newer siloed projects.
The second shortcoming is the non-documentation of feature-functions within the applications themselves. Even when applications are well known throughout the organization, their complete functionality sets are known to no one. This results in duplicate development, redundant purchases, and negative ROI — although no one knows just how negative.
At a minimum, enterprises should maintain master lists of both their WCM-related applications and the functionality within each one. To make effective use of such documentation, companies should establish effective dissemination processes. Examples range from the inclusion of key individuals in change control board meetings (for companies with predictive-style development methods) to informal cross-functional communication, especially between disparate technology groups, but also between IT and the business units whose requirements drive application development.

« Older posts Newer posts »

© 2025 The Gilbane Advisor

Theme by Anders NorenUp ↑