Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

S3 cleanup needs a full redesign #768

Open
ivukotic opened this issue May 9, 2024 · 2 comments
Open

S3 cleanup needs a full redesign #768

ivukotic opened this issue May 9, 2024 · 2 comments
Assignees

Comments

@ivukotic
Copy link
Member

ivukotic commented May 9, 2024

Minio cleanup did not work as minio API always return only first 1000 objects in the list.
I have created a new cleanup code that relies on AWS boto3 API.
It now lists all the objects in the bucket but listing buckets returns only first 1000.
This is obviously a problem...
I have made a code that iteratively deletes empty buckets and buckets over the time limit.
Then it deletes oldest 500 buckets in the list of 1000, and repeats the process until we are under 1000 basket limit.

Once we have tens of users and some of them create hundreds or requests this all becomes a huge problem.

We have three solutions:

  • use a single bucket and have requestId as a prefix to the object name
  • use a single bucket but put object in "subdirectories"
  • we use multiple buckets as is the case now, but cleanup connects to postgresql DB to find what are all the buckets.
@ivukotic ivukotic self-assigned this May 9, 2024
@ponyisi
Copy link
Collaborator

ponyisi commented Jul 18, 2024

Is this addressed in #811 ?

@ivukotic
Copy link
Member Author

yes

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants