.NET Tutorials, Forums, Interview Questions And Answers
Welcome :Guest
Sign In
Win Surprise Gifts!!!

Top 5 Contributors of the Month
Post New Web Links

Crawl of TWiki site will not complete and produces Warnings

Posted By:      Posted Date: September 16, 2010    Points: 0   Category :SharePoint
My clients engineering team has an internal Twiki site (running on a Linux/Apache configuration) that's been up and running for years.   We recently installed SharePoint 2007 Enterprise for the client and are trying to configure the SharePoint search to crawl the internal Twiki site.   FYI... Twiki, for those of you that don't know, is an open source Wiki solution.  Read about the product here --> http://twiki.org/ Anyway..... We setup a new website content source via the SharePoint CA that points at the Twiki site and kicked off the crawl.   No problem there.  The crawl fired up and went to work. Here's the stumper.  After over 24 hours, the crawl was still running.  That's right, it ran for over 24 hours and still did not complete.  Assuming that something was wrong, we went ahead and stopped the crawl.   While the crawl was running, we were seeing the following behavior in the crawl log. Successfully crawls about 1,600 plus pieces of content on the Twiki site Produces over 250,000 Warnings with this message --> "Content for this URL is excluded by the server because a no-index attribute".  This is stunning.   I'm not a Twiki expert, but I find it hard to believe there are really 250,000 pieces of c

View Complete Post

More Related Resource Links

Only crawl one site collection

Hi We have an intranet with about 100 site collections. How can I set up one of those to be in a separate content source that can be crawled more often? Do I need to make two content sources with one containing the other 99 site collections with the setting "Crawl only the SharePoint Site of each start address" and the other one containg my prioritized site collection with the same setting? I also would like to ask if the crawl rules have any effect on in which order the content is crawled. If I put a certain site to be included with order 1 will that site always be crawled first? //Niclas

Cannot crawl sharepoint site and mysite after database attach upgrade form sharepoint 2007 to 2010.

After database attach upgrade site and mysite from sharepoint 2007 to 2010 , I have full crawl and get "The crawler could not communicate with the server. Check that the server is available and that the firewall access is configured correctly. If the repository was temporarily unavailable, an incremental crawl will fix this error. ( Error from SharePoint site: HttpStatusCode ServiceUnavailable The request failed with HTTP status 503: Service Unavailable. )" for mysite and get "Access is denied. Verify that either the Default Content Access Account has access to this repository, or add a crawl rule to crawl this repository. If the repository being crawled is a SharePoint repository, verify that the account you are using has "Full Read" permissions on the SharePoint Web Application being crawled." for sharepoint site. The content access account for search is "db_owner" of both of site and mysite. How do I solved this problem ?

Cannot crawl complex URL's without setting a site-wide rule to 'crawl as http content'. Help!

I have pages within a site that use a query string to provide dynamic data to the user (http://<site>/pages/example.aspx?id=1). I can get the content source to index these dynamic pages only if I create a rule which sets the root site (http://<site>/*) to 'include complex urls' and 'crawl sharepoint content as http content'. This is NOT acceptable as changing the crawling protocol from SharePoint's to HTTP will prevent any metadata from being collected on the indexed items. The managed metadata feature is a critical component to our SharePoint applications. To dispel any wondering of whether or not this is simply a configuration error on my part refer to http://social.technet.microsoft.com/Forums/en-US/sharepointsearch/thread/4ff26b26-84ab-4f5f-a14a-48ab7ec121d5 . The issue mentioned is my exact problem but the solution is unusable as I mentioned before. Keep in mind this is for an external publishing site and my search scope is being trimmed using content classes to only include documents/pages (STS_List_850 and STS_ListItem_DocumentLibrary). Creating a new web site content source and adding it to my scope presents 2 problems: duplicate content in scope and no content class defining it that I know of. What options do I have?

"The Server could not complete your request" on trying to connect to a site or create a new site wit

I am trying out SharePoint Designer 2010 on windows 2008 with SharePoint 2010 beta installed. When I try to connect to a site or create a new site I get the error: "The server could not complete your request.  For more specific information, click the Details button." When I clicked the Details button it says "Object moved" on one line then on the next it says "Object moved to here". I wanted to see the new workflow stuff but I cannot connect to a site.

How to copy a complete list from a sharepoint site to another when both are on different servers and

Hi All,
I have a sharepoint site which has a custom list. This list keeps getting updated with new items probably on a daily basis.
I need a copy of this list on my sharepoint site which is on a different server altogether. And i need the updates too on a daily basis. How can i achieve this programmatically or otherwise?How can i copy the entire list?
Some timer job to get my list updated?
Rgds, Ash

SP2010 SSL Crawl Fails - Accessing this site requires a client certificate.


Full error Message:  Accessing this site requires a client certificate. Specify a client certificate in the crawl rules.

I added the Crawl Rule and Content Source for the test site https://sp.xxxx.com/TestSite and tried specifying each available client certificates in the rules one by one, but the crawl still fails.





What kind of certificate does it needs and how to enable it? I know that the *.xxxx.com certificate is used for our SSL.


Search issues for with a php site not being able Crawl


Hi everyone,

I have setup a  content source of a php site but i am not able to crawl it. I have also added a php file type in the file tyes section

my crawl rule has two rules

1. http://*/*asdf*

2 http://*.*


i am able to crawl other sites but just not the php ones


can anyone please point me in the right direction





Sharepoint Services 3.0 - Cannot complete action. Please try again. No access to any site or Centr


I had a nice working sharepoint services 3.0 environment, which I added some webparts to.  these webparts worked fine, although it said I had no permission to view the data it was trying to get to (which is hosted on another server). 

I changed the authntication to windows integrated in the Central Administration part, created a new account which is a domain admin and added to the WSS and IIS groups, which would have the permissions to view the other servers data, and did a setspn to register the HTTP and to put is against that user.

Now I cannot access Central Administration or the site at all and although I have changed it all back, I get Cannot complete this action please try again.  I have already spoken to microsoft concierge and followed a barrage of pages, and they say to post here.

Any helpers out there today?

Do I need a full crawl if I add a new web in a site collection?

Hello, Do I need a full crawl if I add a new web in a site collection? Is an incremantal crawl enough? Thanks

How to save a COMPLETE site and create it on another server


I have a small site, but it has some custom web parts, with safe control listings in the web.confg and I need to save it as a whole (as a solution, or whatever), so I can move it from our test server to production, with all the web.config settings, and all the web parts on all the pages I've created.

I've tried saving as a template and including content, but that didn't work. Not only were the web parts not on the page, but the web.config didn't have all the safe control listings.

I don't know if this is even possible - - can anyone guide me in the correct direction on how to get this accomplished?

Crawl Error - Error in the Site Data Web Service. (Exception of type 'System.OutOfMemoryException'


I have a Sharepoint site that includes a WIki library.  Recently searches have stopped returning any hits from the Wiki pages.  (Search hits from documents and discussions are correctly returned.)  When I check the crawl log from Shared Services Admin site, I see the following error every time a crawl is performed:



Error in the Site Data Web Service. (Exception of type 'System.OutOfMemoryException' was thrown.)


The URL above is part of the Wiki library.  When I look at the raw Sharepoint logs, I see the following entries every time a crawl is performed:


11/26/2007 05:01:50.93  mssdmn.exe (0x1F98)                      0x1368 Search Server Common           MS Search Indexing             7hp2 Monitorable EnumerateListFolder fail. error 2147755542, strWebUrl http://cao.tfn.com/Caopedia, strListName {e9acf8e9-f23f-4afd-b8e9-6eaf6b0101db}, strFolder  
11/26/2007 05:01:50.93  mssdmn.exe (0x1F98)                      0x1368 Search Server Com

i need to crawl the data from particular site's database.

i need to crawl the data from particular site's database. Is it possible? they may protect their database. am i correct?. please reply me.

Changing the graphics on the home page of a SharePoint site

When you create a site in Microsoft Windows SharePoint Services by using the default Team Site template, the home page includes two default graphics. A Windows SharePoint Services graphic appears in the body of the page, and a picture of a house in a circle appears near the site title. You can change these graphics to customize your site. Changing the smaller graphic requires a Windows SharePoint Services compatible Web page editor

Take a tour of your SharePoint site

What is a SharePoint site for?
A SharePoint site is a Web site that provides a central storage and collaboration space for documents, information, and ideas. A SharePoint site is a tool for collaboration, just like a telephone is a tool for communication, or a meeting is a tool for decision making. A SharePoint site helps groups of people (whether work teams or social groups) share information and work together. For example, a SharePoint site can help you:

Coordinate projects, calendars, and schedules.
Discuss ideas and review documents or proposals.
Share information and keep in touch with other people.

Six easy ways to customize your site using Web Parts in SharePoint

Becoming familiar with Web Part customization techniques gives you a range of page-customization choices. For example, you can customize your site's home page so that it displays the specific information that your team needs. You can rearrange the list items on the home page, add lists, and even change the images on the home page. As you become more familiar with Web Part Page customization techniques, you can build interactive Web Part Pages that makes the data in lists more dynamic and useful, and you can add new Web Parts that provide access to external data and Web services.

Complete SQL Solution

SQL Server Solutions: SQL Server 2000/2005 Related Free Snippets.

Tips to Improve Your ASP.NET Web site performance

ASP.NET has been developed with lots of features which makes developer life easy. Start from ViewState, PostBack, Server Controls , User Controls and etc. All of them can make your web development experience fun and easy, but improper use of them can cause your site performance slowdown. Most of the controls generate overhead to the client and server. Therefore in this article, I I will list out about things you can do and don't to improve your ASP.NET website performance.
ASP.NetWindows Application  .NET Framework  C#  VB.Net  ADO.Net  
Sql Server  SharePoint  Silverlight  Others  All   

Hall of Fame    Twitter   Terms of Service    Privacy Policy    Contact Us    Archives   Tell A Friend