Update Frequency of JSONL Data File

Hi I am developing a simple-ish PWA app to enable me to create a food diary to keep on track with diet and health targets.

I have downloaded the big JSONL file and imported the relevant bits of the JSON into my own database (I am only interested in images, product names and brands, ingredients and nutritonal info).

I noticed that a product that I added to my diary had incorrect spelling of the product name and incorrect calorific values (the original poster had input the kj into the kcal field) so the product was reporting lots of calories for a small portion.

I logged into OpenFoodFacts and corrected the spelling and nutritional info and thought I’d just wait fo the correct data to be reflected in my database when my overnight script draws down and extracts the data from the JSONL file. I also added two new products that were not in the JSONL file.

The Open Food Facts website says the data downloads are updated nightly - but the revisions and additions haven’t appeared in last nights update.

Is there any merit in downloading the data every day or should I do it weekly or on some other schedule?

Many thanks :slight_smile:

Hello Steves,
The JSONL is indeed exported every night (daily maintenance scripts start around midnight CET), but we have many exports, which takes a lot of time.
Currently, the JSONL is generated around 4PM CET, so if you want the latest dump, you should download it after this time.
You can check the datetime of the last upload using curl, looking at the Last-Modified header:

curl --head https://openfoodfacts-ds.s3.eu-west-3.amazonaws.com/openfoodfacts-products.jsonl.gz

Which returns:

x-amz-id-2: 0+GVUzIUCRk7543wnsCJNvRPx38MH68LrVN7kpWyUWxXI8aG1uWr/syDxb6jY/uugD/Xj2W/lSA=
x-amz-request-id: FC7T5NZZ6J2AYVVS
Date: Mon, 21 Jul 2025 07:31:13 GMT
Last-Modified: Sun, 20 Jul 2025 13:51:13 GMT
ETag: "78dfb5505d8de3ae6a82214045592fe6-579"
x-amz-server-side-encryption: AES256
x-amz-version-id: THLtAR9Tgurcs2ulF5IXxvk1qIeT_iTu
Accept-Ranges: bytes
Content-Type: application/gzip
Content-Length: 9712456568
Server: AmazonS3

In the future, we would like to perform less exports on the main server (ideally only MongoDB and JSONL), and delegate the generation of the rest of the exports on a different server, which would allow reusers to get a dump much sooner in the day.

If you have any additional question, feel free to ask!

Hi Raphael,

Thanks for the response. I’ll leave my scheduled task as it is as although the JSONL seems to be ready early afternoon, I’d prefer my server to process it in the small hours.

Kind regards,
S

1 Like

Hi have noticed that the modified date on latest JSONL download (and indeed the MongoDB) is 12 days older than the date I requested it:

curl --head https://openfoodfacts-ds.s3.eu-west-3.amazonaws.com/openfoodfacts-products.jsonl.gz
HTTP/1.1 200 OK
x-amz-id-2: OKzs4g/1GleIuNLsZEuWfdniCxeFEh7duYhKvsLqOBHmhoF4wXXE2XuQpXAAWqJPDrUoxxrvhpvThVQ0/S9HcHw2ywVJtNnK
x-amz-request-id: ENPGJ2NCSTK4BRAF
Date: Wed, 18 Mar 2026 11:55:08 GMT
Last-Modified: Fri, 06 Mar 2026 20:51:58 GMT
ETag: “2990eb691c88aa18068a72c9bac5168e-650”
x-amz-server-side-encryption: AES256
x-amz-version-id: OHIGB7Jy9CiJO6FWThlsldVbdJV3.i_o
Accept-Ranges: bytes
Content-Type: application/gzip
Content-Length: 10896037142
Server: AmazonS3

Has the update frequency of this file changed?

No it’s still a daily export!

I tried again the `curl HEAD` command, the JSONL was successfully exported last night:

raphael@raphael-NS5x-NS7xPU ~> curl --head https://openfoodfacts-ds.s3.eu-west-3.amazonaws.com/openfoodfacts-products.jsonl.gz
HTTP/1.1 200 OK
x-amz-id-2: 6JZPs6st/vCC8ttm6QH382A2I9Q1ywEp7h3n6RsjRL4L3V0Ioo8aqXq/S5V0vmaBBvG5hA2LhkY=
x-amz-request-id: G17R6NR1A5AYX0AD
Date: Thu, 19 Mar 2026 08:40:16 GMT
Last-Modified: Wed, 18 Mar 2026 15:46:30 GMT
ETag: “14679ee4ded5d1ac0891046339d1d515-665”
x-amz-server-side-encryption: AES256
x-amz-version-id: m0DuFihlT8MRbDxXzPIC_Dq__BXbuP5P
Accept-Ranges: bytes
Content-Type: application/gzip
Content-Length: 11145360502
Server: AmazonS3

I will have a look tomorrow to see if there was an issue with the export. Thanks for bringing this up!

Hi Raphael0202,

Thanks for the reply. I’ve tried today and am getting this response from:

curl --head https://openfoodfacts-ds.s3.eu-west-3.amazonaws.com/openfoodfacts-products.jsonl.gz

HTTP/1.1 200 OK
x-amz-id-2: bINHIJ5VuWWg9ifhEdhDP+pxNACVpNB20hlaLODbA4ih6UDix6lHvgmuKy5GnULy/Qi0xlDY0wdELNANZ1TnnzjvMaUbOQvr
x-amz-request-id: CKJ037CBKSH4AAKG
Date: Tue, 24 Mar 2026 13:57:55 GMT
Last-Modified: Wed, 18 Mar 2026 15:46:30 GMT
ETag: “14679ee4ded5d1ac0891046339d1d515-665”
x-amz-server-side-encryption: AES256
x-amz-version-id: m0DuFihlT8MRbDxXzPIC_Dq__BXbuP5P
Accept-Ranges: bytes
Content-Type: application/gzip
Content-Length: 11145360502
Server: AmazonS3

So getting the same result as you pasted above - but several days later.

Hello Steve,
I’ve looked into it yesterday, and we have frequent failure due to some bug in the CSV export. I’ve put JSONL and MongoDB export higher in the export order, and it successfully exported last night. I will check in the coming days as well to see if the JSONL export is successful.
Sorry for the inconvenience!

Thanks for the update!

Just to let you know the download I did yesterday was considerably smaller than normal (1M entries) possibly caused by the issue you identified - but the download I’ve just done seems to be back to normal.

Will keep an eye on it over coming days.

Steve