manifoldcf-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jorge Alonso Garcia <jalon...@gmail.com>
Subject Re: sharepoint crawler documents limit
Date Thu, 19 Dec 2019 16:42:16 GMT
Hi,
On UI shows 20,002 documents (on a firts phase show 10,001,and after
sometime of process raise to 20,002) .
It looks like a hard limit, there is more files on sharepoint with the used
criteria


Jorge Alonso Garcia



El jue., 19 dic. 2019 a las 16:05, Karl Wright (<daddywri@gmail.com>)
escribió:

> Hi Jorge,
>
> When you run the job, do you see more than 20,000 documents as part of it?
>
> Do you see *exactly* 20,000 documents as part of it?
>
> Unless you are seeing a hard number like that in the UI for that job on
> the job status page, I doubt very much that the problem is a numerical
> limitation in the number of documents.  I would suspect that the inclusion
> criteria, e.g. the mime type or maximum length, is excluding documents.
>
> Karl
>
>
> On Thu, Dec 19, 2019 at 8:51 AM Jorge Alonso Garcia <jalongar@gmail.com>
> wrote:
>
>> Hi Karl,
>> We had installed the shaterpoint plugin, and access properly http:/server/
>> _vti_bin/MCPermissions.asmx
>>
>> [image: image.png]
>>
>> Sharepoint has more than 20,000 documents, but when execute the jon only
>> extract these 20,000. How Can I check where is the issue?
>>
>> Regards
>>
>>
>> Jorge Alonso Garcia
>>
>>
>>
>> El jue., 19 dic. 2019 a las 12:52, Karl Wright (<daddywri@gmail.com>)
>> escribió:
>>
>>> By "stop at 20,000" do you mean that it finds more than 20,000 but stops
>>> crawling at that time?  Or what exactly do you mean here?
>>>
>>> FWIW, the behavior you describe sounds like you may not have installed
>>> the SharePoint plugin and may have selected a version of SharePoint that is
>>> inappropriate.  All SharePoint versions after 2008 limit the number of
>>> documents returned using the standard web services methods.  The plugin
>>> allows us to bypass that hard limit.
>>>
>>> Karl
>>>
>>>
>>> On Thu, Dec 19, 2019 at 6:37 AM Jorge Alonso Garcia <jalongar@gmail.com>
>>> wrote:
>>>
>>>> Hi,
>>>> We have an isuse with sharepoint connector.
>>>> There is a job that crawl a sharepoint 2016, but it is not recovering
>>>> all files, it stop at 20.000 documents without any error.
>>>> Is there any parameter that should be change to avoid this limitation?
>>>>
>>>> Regards
>>>> Jorge Alonso Garcia
>>>>
>>>>

Mime
View raw message