Setting up a folder crawler is an easy way to import data that is accessible to the server instance. This could e.g. be from a local hard drive or a network drive. If you want to upload local data to e.g. a cloud instance, using the cli is a good way to do it.
Setting up a Folder Crawler using the Admin UI
Navigate to The Data Sources Hub
Add a name for this source in the
Sourcefield and the absolut path the the data folder in the
Namefield identifies a Task that runs the given schedule, whereas the
Sourcewill show up as data source in the search system.
If you want to restrict the filetypes, add the MIME Types you want to ingest into the system.
Saveto save the scheduled task.
Press the Run Once button to run the task immediately. The file ingestion will then run on the given schedule.