Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Increased Volume Found after --continue #2257

Open
aberenguel opened this issue Jul 1, 2024 · 10 comments
Open

Increased Volume Found after --continue #2257

aberenguel opened this issue Jul 1, 2024 · 10 comments

Comments

@aberenguel
Copy link
Contributor

aberenguel commented Jul 1, 2024

I'm processing a dd file with master branch (profile pedo, enableOCR and enableFaceRecognition).
During the process, the process aborted, so I re-exectuted with --continue.
Then I noticed that the Volume Found (Volume Descoberto) increased.

image
TimePhoto_20240701_125326

@lfcnassif
Copy link
Member

lfcnassif commented Jul 1, 2024

Thanks @aberenguel for reporting. It shouldn't happen, possibly it is related to #telegram

A few weeks ago I noticed some items were receiving a different trackId when resuming (--continue), when it should be the same, so already processed items could be identified and ignored. I still didn't investigate the root cause, but that would explain the issue you reported.

Are you able to share a small image and provide the aborting point to resume to reproduce the issue?

@lfcnassif lfcnassif added the bug label Jul 1, 2024
@lfcnassif
Copy link
Member

Do you have Telegram databases in this case? @hauck-jvsh does Telegram parser extract subitems always in the same order? This is needed for --continue to work properly, since the subitem number/position is used in trackID computation. If a HashSet or HashMap, for example, is used somewhere into the parser to store subitems to be extracted, that may be the cause, just a blind guess...

@lfcnassif
Copy link
Member

Or maybe other parsers are affected by the hypothesis above...

@hauck-jvsh
Copy link
Member

I never thought about this, and never made any test to check if the items are being extracted in the same order.

@lfcnassif
Copy link
Member

Don't worry, I never gave this recommendation to contributors and just realized that might be the issue today.

@aberenguel
Copy link
Contributor Author

I think it is not related to Telegram. This is the case result:
image

@aberenguel
Copy link
Contributor Author

Are you able to share a small image and provide the aborting point to resume to reproduce the issue?

The trigger image has 1TB. I'll try to reproduce with another image.
Another test I'm going to do is to close the case when sleuth.db is complete e to resume with --continue.

@aberenguel
Copy link
Contributor Author

I was able to reproduce killing the java processes with kill -9 and resuming the processing with --continue.

@lfcnassif
Copy link
Member

Are you able to identify which files/artifacts were duplicated in the case? Look at the trackId property in Metadata filter panel, each value should occur just for 1 file. If it is not duplicated, maybe looking for files with same hashes and same paths may help to find the duplicated artifacts. And actually there is a very little chance nothing was duplicated, but it was just a minor bug in the Total Volume counter.

@lfcnassif
Copy link
Member

I took a look at --continue related code and seems it is taking into account the size of items that are containers AND subitems/carved items in the total volume count to process, while standard processing doesn't. So maybe this is just a minor issue with the volume count...

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants