Skip to content

Validating large data file (150GB) is trying to write the entire file to local temp space - error upon not enough local space #1241

@scholes-ds

Description

@scholes-ds

Checked for duplicates

Yes - I've already checked

🐛 Describe the bug

When our data archivist runs validate on a staged copy of very large data files in preparation for a PDS3 to PDS4 data conversion, validate throws a “not enough space on the disk” error. This is correct because we don't have that much disk space on the processing VM. Does it really need to copy all of the data file bytes locally?
The directory with contents that will be converted to PDS4: https://pds-geosciences.wustl.edu/mro/mro-m-sharad-5-3d-v1/mrosh_3001/data/

🕵️ Expected behavior

We expected validate to complete successfully.

📜 To Reproduce

  1. Make a temp xml label or we can provide one for data files in here: https://pds-geosciences.wustl.edu/mro/mro-m-sharad-5-3d-v1/mrosh_3001/data/

  2. There are also very large data files in the following archive location. Apparently a different error was thrown on these large files at the time, and it was not reported.

  3. https://pds-geosciences.wustl.edu/grail/grail-l-lgrs-5-rdr-v1/grail_1001/shbdr/

...

🖥 Environment Info

  • Version of this software: Validate 3.7
  • Operating System: Windows Server 2012R2
    ...

📚 Version of Software Used

Validate 3.7

🩺 Test Data / Additional context

No response

🦄 Related requirements

🦄 #xyz

⚙️ Engineering Details

No response

🎉 Integration & Test

No response

Metadata

Metadata

Assignees

Type

Projects

Status

ToDo

Status

Blocked

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions