Concurrent queries in NLP
This MR adds a worker.nlp_conduit_chunk_size
setting which defines what is the chunk size when importing from API. This same setting affects how many concurrent threads will be started for NLP queries. Usually it's best to set this value to the number of CPUs on a machine, this way we use all corenlp threads.
This should speed up the import.
Related to #511