Setting up a folder crawler is an easy way to import data that is accessible to the server instance. This could e.g. be from a local hard drive or a network drive. If you want to upload local data to e.g. a cloud instance, using the cli is a good way to do it.
Setting up a Folder Crawler using the Admin UI
- Navigate to The Data Sources Hub
- Open the
Folder
Data Source
- Add a
Folder
Data Source
- Add a name for this source in the
Source
field and the absolut path the the data folder in thePath
field. TheName
field identifies a Task that runs the given schedule, whereas theSource
will show up as data source in the search system.
If you want to restrict the filetypes, add the MIME Types you want to ingest into the system.
- Press
Save
to save the scheduled task. - Press the Run Once button to run the task immediately. The file ingestion will then run on the given schedule.