Skip to content

GitLab

  • Menu
Projects Groups Snippets
    • Loading...
  • Help
    • Help
    • Support
    • Community forum
    • Submit feedback
    • Contribute to GitLab
  • Sign in
  • G gitlabhq1
  • Project information
    • Project information
    • Activity
    • Labels
    • Members
  • Repository
    • Repository
    • Files
    • Commits
    • Branches
    • Tags
    • Contributors
    • Graph
    • Compare
  • Issues 21
    • Issues 21
    • List
    • Boards
    • Service Desk
    • Milestones
  • Merge requests 12
    • Merge requests 12
  • CI/CD
    • CI/CD
    • Pipelines
    • Jobs
    • Schedules
  • Deployments
    • Deployments
    • Environments
    • Releases
  • Monitor
    • Monitor
    • Incidents
  • Packages & Registries
    • Packages & Registries
    • Package Registry
    • Infrastructure Registry
  • Analytics
    • Analytics
    • CI/CD
    • Repository
    • Value stream
  • Wiki
    • Wiki
  • Snippets
    • Snippets
  • Activity
  • Graph
  • Create a new issue
  • Jobs
  • Commits
  • Issue Boards
Collapse sidebar
  • gpt
  • large_projects
  • gitlabhq1
  • Issues
  • #6316

Closed
Open
Created Feb 13, 2014 by Administrator@rootOwner

Abusive use of archive functionality by robots

Created by: zorun

Since our upgrade to the latest Gitlab version, we very frequently run out of disk space (i.e. almost every day).

After investigation, it turns out that many archives files (.tar.gz, .zip, etc) are created in /home/git/gitlab/tmp/repositories/. This is due to GET requests to URL like /johndoe/examplerepo/repository/archive.tar.gz?ref=cafecafecafecafecafecafecafecafecafecafe by robots, most notably GoogleBot.

Is there a way to tell robots not to index these kind of links? A workaround could be to delete these archives automatically after a short period of time.

We're running GitLab 6.5.1 2ffa03ab. Thanks!

Assignee
Assign to
Time tracking