# this is the configuration file for the GROBID instance grobid: # where all the Grobid resources are stored (models, lexicon, native libraries, etc.), normally no need to change grobidHome: "grobid-home" # path relative to the grobid-home path (e.g. tmp for grobid-home/tmp) or absolute path (/tmp) temp: "tmp" # normally nothing to change here, path relative to the grobid-home path (e.g. grobid-home/lib) nativelibrary: "lib" pdf: pdfalto: # path relative to the grobid-home path (e.g. grobid-home/pdfalto), you don't want to change this normally path: "pdfalto" # security for PDF parsing memoryLimitMb: 6096 timeoutSec: 120 # security relative to the PDF parsing result blocksMax: 200000 tokensMax: 1000000 consolidation: # define the bibliographical data consolidation service to be used, either "crossref" for CrossRef REST API or # "glutton" for https://github.com/kermitt2/biblio-glutton #service: "crossref" service: "glutton" glutton: url: "http://sciencialab.ddns.net:8080" #url: "http://localhost:8080" crossref: mailto: "luca@sciencialab.com" # to use crossref web API, you need normally to use it politely and to indicate an email address here, e.g. #mailto: "toto@titi.tutu" token: # to use Crossref metadata plus service (available by subscription) #token: "yourmysteriouscrossrefmetadataplusauthorizationtokentobeputhere" proxy: # proxy to be used when doing external call to the consolidation service host: port: # CORS configuration for the GROBID web API service corsAllowedOrigins: "*" corsAllowedMethods: "OPTIONS,GET,PUT,POST,DELETE,HEAD" corsAllowedHeaders: "X-Requested-With,Content-Type,Accept,Origin" # the actual implementation for language recognition to be used languageDetectorFactory: "org.grobid.core.lang.impl.CybozuLanguageDetectorFactory" # the actual implementation for optional sentence segmentation to be used (PragmaticSegmenter or OpenNLP) sentenceDetectorFactory: "org.grobid.core.lang.impl.PragmaticSentenceDetectorFactory" # sentenceDetectorFactory: "org.grobid.core.lang.impl.OpenNLPSentenceDetectorFactory" # maximum concurrency allowed to GROBID server for processing parallel requests - change it according to your CPU/GPU capacities # for a production server running only GROBID, set the value slightly above the available number of threads of the server # to get best performance and security concurrency: 10 # when the pool is full, for queries waiting for the availability of a Grobid engine, this is the maximum time wait to try # to get an engine (in seconds) - normally never change it poolMaxWait: 1 delft: # DeLFT global parameters # delft installation path if Deep Learning architectures are used to implement one of the sequence labeling model, # embeddings are usually compiled as lmdb under delft/data (this parameter is ignored if only featured-engineered CRF are used) install: "../delft" pythonVirtualEnv: wapiti: # Wapiti global parameters # number of threads for training the wapiti models (0 to use all available processors) nbThreads: 0 models: # we configure here how each sequence labeling model should be implemented # for feature-engineered CRF, use "wapiti" and possible training parameters are window, epsilon and nbMaxIterations # for Deep Learning, use "delft" and select the target DL architecture (see DeLFT library), the training # parameters then depends on this selected DL architecture - name: "segmentation" # at this time, must always be CRF wapiti, the input sequence size is too large for a Deep Learning implementation engine: "wapiti" #engine: "delft" wapiti: # wapiti training parameters, they will be used at training time only epsilon: 0.0000001 window: 50 nbMaxIterations: 2000 delft: # deep learning parameters architecture: "BidLSTM_CRF_FEATURES" useELMo: false runtime: # parameters used at runtime/prediction max_sequence_length: 3000 batch_size: 1 training: # parameters used for training max_sequence_length: 3000 batch_size: 10 - name: "fulltext" # at this time, must always be CRF wapiti, the input sequence size is too large for a Deep Learning implementation engine: "wapiti" wapiti: # wapiti training parameters, they will be used at training time only epsilon: 0.0001 window: 20 nbMaxIterations: 1500 - name: "header" #engine: "wapiti" engine: "delft" wapiti: # wapiti training parameters, they will be used at training time only epsilon: 0.000001 window: 30 nbMaxIterations: 1500 delft: # deep learning parameters architecture: "BidLSTM_ChainCRF_FEATURES" #transformer: "allenai/scibert_scivocab_cased" useELMo: false runtime: # parameters used at runtime/prediction #max_sequence_length: 510 max_sequence_length: 3000 batch_size: 1 training: # parameters used for training #max_sequence_length: 510 #batch_size: 6 max_sequence_length: 3000 batch_size: 9 - name: "reference-segmenter" #engine: "wapiti" engine: "delft" wapiti: # wapiti training parameters, they will be used at training time only epsilon: 0.00001 window: 20 delft: # deep learning parameters architecture: "BidLSTM_ChainCRF_FEATURES" useELMo: false runtime: # parameters used at runtime/prediction (for this model, use same max_sequence_length as training) max_sequence_length: 3000 batch_size: 2 training: # parameters used for training max_sequence_length: 3000 batch_size: 10 - name: "name-header" engine: "wapiti" #engine: "delft" delft: # deep learning parameters architecture: "BidLSTM_CRF_FEATURES" - name: "name-citation" engine: "wapiti" #engine: "delft" delft: # deep learning parameters architecture: "BidLSTM_CRF_FEATURES" - name: "date" engine: "wapiti" #engine: "delft" delft: # deep learning parameters architecture: "BidLSTM_CRF_FEATURES" - name: "figure" engine: "wapiti" #engine: "delft" wapiti: # wapiti training parameters, they will be used at training time only epsilon: 0.00001 window: 20 delft: # deep learning parameters architecture: "BidLSTM_CRF" - name: "table" engine: "wapiti" #engine: "delft" wapiti: # wapiti training parameters, they will be used at training time only epsilon: 0.00001 window: 20 delft: # deep learning parameters architecture: "BidLSTM_CRF" - name: "affiliation-address" #engine: "wapiti" engine: "delft" delft: # deep learning parameters architecture: "BidLSTM_CRF_FEATURES" - name: "citation" #engine: "wapiti" engine: "delft" wapiti: # wapiti training parameters, they will be used at training time only epsilon: 0.00001 window: 50 nbMaxIterations: 3000 delft: # deep learning parameters architecture: "BidLSTM_CRF_FEATURES" #architecture: "BERT_CRF" #transformer: "michiyasunaga/LinkBERT-base" useELMo: false runtime: # parameters used at runtime/prediction max_sequence_length: 500 batch_size: 30 training: # parameters used for training max_sequence_length: 500 batch_size: 50 - name: "patent-citation" engine: "wapiti" #engine: "delft" wapiti: # wapiti training parameters, they will be used at training time only epsilon: 0.0001 window: 20 delft: # deep learning parameters architecture: "BidLSTM_CRF_FEATURES" #architecture: "BERT_CRF" runtime: # parameters used at runtime/prediction max_sequence_length: 800 batch_size: 20 training: # parameters used for training max_sequence_length: 1000 batch_size: 40 - name: "funding-acknowledgement" engine: "wapiti" #engine: "delft" wapiti: # wapiti training parameters, they will be used at training time only epsilon: 0.00001 window: 50 nbMaxIterations: 2000 delft: # deep learning parameters architecture: "BidLSTM_CRF_FEATURES" #architecture: "BERT_CRF" #transformer: "michiyasunaga/LinkBERT-base" useELMo: false runtime: # parameters used at runtime/prediction max_sequence_length: 800 batch_size: 20 training: # parameters used for training max_sequence_length: 500 batch_size: 40 - name: "copyright" # at this time, we only have a DeLFT implementation, # use "wapiti" if the deep learning library JNI is not available and model will then be ignored #engine: "delft" engine: "wapiti" delft: # deep learning parameters architecture: "gru" #architecture: "bert" #transformer: "allenai/scibert_scivocab_cased" - name: "license" # at this time, for being active, it must be DeLFT, no other implementation is available # use "wapiti" if the deep learning library JNI is not available and model will then be ignored #engine: "delft" engine: "wapiti" delft: # deep learning parameters architecture: "gru" #architecture: "bert" #transformer: "allenai/scibert_scivocab_cased" # for **service only**: how to load the models, # false -> models are loaded when needed, avoiding putting in memory useless models (only in case of CRF) but slow down # significantly the service at first call # true -> all the models are loaded into memory at the server startup (default), slow the start of the services # and models not used will take some more memory (only in case of CRF), but server is immediatly warm and ready modelPreload: true server: type: custom applicationConnectors: - type: http port: 8070 adminConnectors: - type: http port: 8071 registerDefaultExceptionMappers: false # change the following for having all http requests logged requestLog: appenders: [] # these logging settings apply to the Grobid service usage mode logging: level: INFO loggers: org.apache.pdfbox.pdmodel.font.PDSimpleFont: "OFF" org.glassfish.jersey.internal: "OFF" com.squarespace.jersey2.guice.JerseyGuiceUtils: "OFF" appenders: - type: console threshold: INFO timeZone: UTC # uncomment to have the logs in json format # layout: # type: json