Skip to content

Latest commit

 

History

History
61 lines (51 loc) · 8.27 KB

README.md

File metadata and controls

61 lines (51 loc) · 8.27 KB

mod-data-export-worker

Copyright (C) 2021-2023 The Open Library Foundation

This software is distributed under the terms of the Apache License, Version 2.0. See the file LICENSE for more information.

Introduction

API for Data Export Worker module.

Additional information

More detail can be found on Data Export Worker wiki-page: WIKI Data Export Worker.

Issue tracker

See project MODEXPW at the FOLIO issue tracker.

Other documentation

Other modules are described, with further FOLIO Developer documentation at dev.folio.org

Bulk edit

In case of no matched records found when uploading CSV file with items or users, link to download matched records is not available for user. The maximum value of size for uploading file is 15MB. It could be changed with spring.servlet.multipart.max-file-size application argument.

Memory configuration

To stable module operating the following mod-data-export-worker configuration is required: Java args -XX:MetaspaceSize=384m -XX:MaxMetaspaceSize=512m -Xmx2048m, AWS container: memory - 3072, memory (soft limit) - 2600, cpu - 1024.

Environment variables

This module uses separate storage of temporary (local) files for its work. These files are necessary for processing bulk-edit business flows. Any S3-compatible storage (AWS S3, Minio Server) supported by the Minio Client can be used as such storage. Thus, in addition to the AWS configuration (AWS_URL, AWS_REGION, AWS_BUCKET, AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY) of the permanent storage, one need to configure the environment settings for s3 subpathes (S3_SUB_PATH, S3_LOCAL_SUB_PATH). Typically, these options must specify a separate pathes. It is also necessary to specify variable S3_IS_AWS to determine if AWS S3 is used as files storage. By default this variable is false and means that MinIO server is used as files storage. This value should be true if AWS S3 is used as storage.

Name Default value Description
KAFKA_HOST localhost Kafka broker hostname
KAFKA_PORT 9092 Kafka broker port
KAFKA_CONSUMER_POLL_INTERVAL 3600000 Max interval before next poll. If long record processing is in place and interval exceeded then consumer will be kicked out of the group and another consumer will start processing the same message.
ENV folio Environment name
S3_URL http://127.0.0.1:9000/ AWS url
S3_REGION - AWS region
S3_BUCKET - AWS bucket
S3_ACCESS_KEY_ID - AWS access key
S3_SECRET_ACCESS_KEY - AWS secret key
S3_SUB_PATH mod-data-export-worker/remote S3 subpath for files storage
S3_LOCAL_SUB_PATH mod-data-export-worker/local S3 subpath for local files storage
S3_IS_AWS false Specify if AWS S3 is used as files storage
URL_EXPIRATION_TIME 604800 Presigned url expiration time (in seconds)
DATA_EXPORT_JOB_UPDATE_TOPIC_PARTITIONS 50 Number of partitions for topic
KAFKA_CONCURRENCY_LEVEL 30 Concurrency level of kafka listener
E_HOLDINGS_BATCH_JOB_CHUNK_SIZE 100 Specify chunk size for eHoldings export job which will be used to query data from kb-ebsco, write to database, read from database and write to file
E_HOLDINGS_BATCH_KB_EBSCO_CHUNK_SIZE 100 Amount to retrieve per request to mod-kb-ebsco-java (100 is max acceptable value)
AUTHORITY_CONTROL_BATCH_JOB_CHUNK_SIZE 100 Specify chunk size for authority control export job which will be used to query data from entities-links, and write to file
AUTHORITY_CONTROL_BATCH_ENTITIES_LINKS_CHUNK_SIZE 100 Amount to retrieve per request to mod-entities-links
MAX_UPLOADED_FILE_SIZE 40MB Specifies multipart upload file size