1. 06 Apr, 2017 1 commit
  2. 05 Apr, 2017 1 commit
  3. 04 Apr, 2017 1 commit
  4. 03 Apr, 2017 1 commit
    • Sean McGivern's avatar
      Quiet pipeline emails · a1805cbc
      Sean McGivern authored
      1. Never send a pipeline email to anyone other than the user who created
         the pipeline.
      2. Only send pipeline success emails to people with the custom
         notification setting for enabled. Watchers and participants will
         never receive this.
      3. When custom settings are unset (for new settings and legacy ones),
         act as if failed_pipeline is set.
      a1805cbc
  5. 28 Mar, 2017 1 commit
  6. 18 Mar, 2017 1 commit
  7. 10 Mar, 2017 1 commit
  8. 07 Mar, 2017 1 commit
  9. 06 Mar, 2017 1 commit
  10. 03 Mar, 2017 2 commits
  11. 01 Mar, 2017 4 commits
  12. 24 Feb, 2017 1 commit
  13. 23 Feb, 2017 7 commits
  14. 20 Feb, 2017 1 commit
    • Z.J. van de Weg's avatar
      Transactional mattermost team creation · 444d71e0
      Z.J. van de Weg authored
      Before this commit, but still on this feature branch, the creation of
      mattermost teams where a background job. However, it was decided it was
      better that these happened as transaction so feedback could be displayed
      to the user.
      444d71e0
  15. 16 Feb, 2017 2 commits
  16. 08 Feb, 2017 1 commit
  17. 05 Feb, 2017 1 commit
  18. 03 Feb, 2017 4 commits
  19. 02 Feb, 2017 1 commit
  20. 31 Jan, 2017 4 commits
    • Kamil Trzcinski's avatar
      6e70870a
    • Kamil Trzcinski's avatar
      Asynchronously remove pages · e9e8a2f6
      Kamil Trzcinski authored
      e9e8a2f6
    • Kamil Trzcinski's avatar
      Add GitLab Pages · 120f9aba
      Kamil Trzcinski authored
      - The pages are created when build artifacts for `pages` job are uploaded
      - Pages serve the content under: http://group.pages.domain.com/project
      - Pages can be used to serve the group page, special project named as host: group.pages.domain.com
      - User can provide own 403 and 404 error pages by creating 403.html and 404.html in group page project
      - Pages can be explicitly removed from the project by clicking Remove Pages in Project Settings
      - The size of pages is limited by Application Setting: max pages size, which limits the maximum size of unpacked archive (default: 100MB)
      - The public/ is extracted from artifacts and content is served as static pages
      - Pages asynchronous worker use `dd` to limit the unpacked tar size
      - Pages needs to be explicitly enabled and domain needs to be specified in gitlab.yml
      - Pages are part of backups
      - Pages notify the deployment status using Commit Status API
      - Pages use a new sidekiq queue: pages
      - Pages use a separate nginx config which needs to be explicitly added
      120f9aba
    • Adam Pahlevi's avatar
      resolve deprecation warnings · 538d1bff
      Adam Pahlevi authored
      don’t pass AR object, use the ID to avoid depr warning
      
      pass in the id instead of AR object to specs for `ProjectDestroyWorker`
      538d1bff
  21. 25 Jan, 2017 1 commit
    • Yorick Peterse's avatar
      Fix race conditions for AuthorizedProjectsWorker · 88e627cf
      Yorick Peterse authored
      There were two cases that could be problematic:
      
      1. Because sometimes AuthorizedProjectsWorker would be scheduled in a
         transaction it was possible for a job to run/complete before a
         COMMIT; resulting in it either producing an error, or producing no
         new data.
      
      2. When scheduling jobs the code would not wait until completion. This
         could lead to a user creating a project and then immediately trying
         to push to it. Usually this will work fine, but given enough load it
         might take a few seconds before a user has access.
      
      The first one is problematic, the second one is mostly just annoying
      (but annoying enough to warrant a solution).
      
      This commit changes two things to deal with this:
      
      1. Sidekiq scheduling now takes places after a COMMIT, this is ensured
         by scheduling using Rails' after_commit hook instead of doing so in
         an arbitrary method.
      
      2. When scheduling jobs the calling thread now waits for all jobs to
         complete.
      
      Solution 2 requires tracking of job completions. Sidekiq provides a way
      to find a job by its ID, but this involves scanning over the entire
      queue; something that is very in-efficient for large queues. As such a
      more efficient solution is necessary. There are two main Gems that can
      do this in a more efficient manner:
      
      * sidekiq-status
      * sidekiq_status
      
      No, this is not a joke. Both Gems do a similar thing (but slightly
      different), and the only difference in their name is a dash vs an
      underscore. Both Gems however provide far more than just checking if a
      job has been completed, and both have their problems. sidekiq-status
      does not appear to be actively maintained, with the last release being
      in 2015. It also has some issues during testing as API calls are not
      stubbed in any way. sidekiq_status on the other hand does not appear to
      be very popular, and introduces a similar amount of code.
      
      Because of this I opted to write a simple home grown solution. After
      all, all we need is storing a job ID somewhere so we can efficiently
      look it up; we don't need extra web UIs (as provided by sidekiq-status)
      or complex APIs to update progress, etc.
      
      This is where Gitlab::SidekiqStatus comes in handy. This namespace
      contains some code used for tracking, removing, and looking up job IDs;
      all without having to scan over an entire queue. Data is removed
      explicitly, but also expires automatically just in case.
      
      Using this API we can now schedule jobs in a fork-join like manner: we
      schedule the jobs in Sidekiq, process them in parallel, then wait for
      completion. By using Sidekiq we can leverage all the benefits such as
      being able to scale across multiple cores and hosts, retrying failed
      jobs, etc.
      
      The one downside is that we need to make sure we can deal with
      unexpected increases in job processing timings. To deal with this the
      class Gitlab::JobWaiter (used for waiting for jobs to complete) will
      only wait a number of seconds (30 by default). Once this timeout is
      reached it will simply return.
      
      For GitLab.com almost all AuthorizedProjectWorker jobs complete in
      seconds, only very rarely do we spike to job timings of around a minute.
      These in turn seem to be the result of external factors (e.g. deploys),
      in which case a user is most likely not able to use the system anyway.
      
      In short, this new solution should ensure that jobs are processed
      properly and that in almost all cases a user has access to their
      resources whenever they need to have access.
      88e627cf
  22. 08 Jan, 2017 1 commit
    • Vincent Wong's avatar
      Record and show last used date of SSH Keys · b6df93a5
      Vincent Wong authored
      Addresses: Issue #13810
      
      1. Adds a last_used_at attribute to the Key table/model
      2. Update a key's last_used_at whenever it gets used
      3. Display how long ago an ssh key was last used
      b6df93a5
  23. 21 Dec, 2016 1 commit