Calendar Year 2018 Q4 OKRs

View GitLabs Objective-Key Results for quarter 4 2018. Learn more here!

CEO: Grow Incremental ACV according to plan. Sales prospects. Marketing achieves the SCLAU plan. Measure pipe-to-spend for all marketing activities.

  • VPE
    • Support: Improve Support expectations and documentation for Proof of Concept and Trial prospects, including temporary expanded Support focus for top 10 prospects (defined by Sales). 75% completed with segments redefined and workflows updated.
      • Amer East: 95% SLA for Premium/Ultimate/GitLab.com customers and 95% CSAT for all customers
      • Amer West: 95% SLA for Premium/Ultimate/GitLab.com customers and 95% CSAT for all customers
      • APAC: 95% SLA for Premium/Ultimate/GitLab.com customers and 95% CSAT for all customers
      • EMEA: 95% SLA for Premium/Ultimate/GitLab.com customers and 95% CSAT for all customers
      • Global team achieved 89% (93% achievement to the 95% goal), +5% increase over CYQ3 and trending up
  • CMO: Marketing achieves the SCLAU plan. 8 SCLAU per SDR per month, content 2x QoQ, website information architecture helps 50% QoQ.
    • Product Marketing: Complete Information Architecture for customer facing web site pages and add 50 product comparison pages.
    • Product Marketing: Deliver 15 enablement sessions each to sales and xDR teams - 30 total sessions.
    • Product Marketing: Target 6 industry analyst reports to include GitLab coverage.
    • MOps: Implement top three personas into our behavioural and demographic scoring model to get more qualified leads to SDRs.
    • Online Growth: Use data-driven approach and SEO/PPC/ABM/Paid social channels to drive traffic to high intent assets and increase signups for gated assets trials, webinars, and demos by 25% compared to last quarter.n
    • MPM: Support all field events with pre-event targeting, confirmations, sales notifications, and reminders to drive a minimum of 90% utilization per event.
    • MPM: Triple our marketable (aka opted-in for communications) database in Salesforce and Marketo by improving opt-in practices, revamping email communication positioning, and implementing strategic opt-in campaign to engage the database.
    • MPM: Launch an on-demand product demo and a weekly LIVE product Q&A session, generating 5x higher net new leads than previous weekly Enterprise Demo.
    • Content Marketing: Increase sessions on content marketing blog posts 5% MoM.
      • October: 55,583 sessions
      • November: 58,362 sessions
      • December: 61,280 sessions
  • CMO: Measure pipe-to-spend for all marketing activities. Content team based on it, link meaningful conversations to pipe, double down on what works.
    • Online Growth: Improve measurement of pipe-to-spend with more granular data. Work to expand reporting beyond initial pipe-to-spend calculation and ensure online actions are measured and reported.
    • Online Growth: Ensure site is properly tagged for attribution and optimized for crawlers and content and assets are aligned to Information Architecture.
    • MOps: Restructure Marketo/Bizible/SFDC to align the campaign & progression statuses across the system to allow for better multi-touch attribution tracking.
    • MOps: Streamline record management processes. Refresh SFDC layouts with unnecessary fields removed, create record views
  • VP Alliances: Establish differentiated relationships with Major Clouds. GitLab as a Service LoI on a major cloud, 500+ leads through joint marketing, GTM and Partner agreement in place tihe a large private cloud provider.
  • VP Alliances: establish GitLab as DevOps tool for CNCF. Three keynote/speaking engagements at CNCF events and 1000+ leads from Kubernetes partners
  • CFO: Use data to drive business decisions
    • Director of Business Operations: Create scope for User Journey that is documented and mapped to key data sources. 50%
    • Director of Business Operations: Top of funnel process and metrics defined and aligned with bottom of funnel. 70%
      • Data & Analytics: Able to define and calculate Customer Count, MRR and ARR by Customer, Churn by Cohort, Reason for Churn *75%- - Some metrics still require review
    • Director of Business Operations: Looker Explores generated for Customer Support, PeopleOps, GitLab.com Event Data, Marketing Data 100%
      • Data & Analytics: Single Data Lake for all raw and transformed company data - migration to Snowflake complete with DevOps workflow in place, GitLab.com production data extracted and modelled 75%
      • Data & Analytics: Configuration and Business processes documented with integrity tests that sync back with upstream source (SFDC), dbt docs deployed 100%
  • Product: Listen to and prioritize top feature requests from customers through sales / customer success. 10 features prioritized. => *130% Complete- 13 items prioritized

CEO: Popular next generation product. Finish 2018 vision of entire DevOps lifecycle. GitLab.com is ready for mission critical applications. All installations have a graph of DevOps score vs. releases per year.

CEO: Great team. ELO score per interviewer, executive dashboards for all key results, train director group.

  • CMO: Hire missing directors. Operations, Corporate, maybe Field.
  • VPE: Craft an impactful iteration to improve our career development framework and train the team => 70%, shipped a new BE eng benchmark and experience factor worksheets (also for SREs) to hopefully address career development and comp for a meaningful proportion of the engineering function.
  • VPE: Source 20 candidates by Oct 31 and hire 2 Engineering Directors: 100% (20/20) sourced, hired 1/2 directors => 60%, completed sourcing and hired a Sr Dir.
    • Frontend:
      • Scale process to handle incoming amount ofapplications and hire 2 engineering managers and 10 engineers: 1 engineering manager and 10 engineers hired (92%)
    • Dev Backend:
      • Gitaly:
        • Source 10 candidates by October 15 and hire 1 developer: 10 sourced (100%), 2 hired (200%) => Hired one extra developer to compensate for Alejandro’s departure
      • Gitter:
      • Plan:
        • Source 25 candidates (at least 5 by direct manager contact) by October 15 and hire 2 developers: 25 sourced (100%), 2 hired (100%)
      • Create:
        • Source 25 candidates (at least 5 by direct manager contact) by October 15 and hire 2 developers: 25 sourced (100%), 2 hired (100%)
        • Get 10 MRs merged into Go projects by team members without pre-existing Go experience: 6 merged, 1 in review (65%)
      • Manage:
        • Source 25 candidates (at least 5 by direct manager contact) by October 15 and hire 2 developers: 25 sourced (100%), 1 hired (50%)
        • Team to deliver 10 topic-specific knowledge sharing sessions (“201s”) by the end of Q4 (4 completed)
      • Geo:
        • At least one Geo team member to advance to Maintainer (0%) : Two engineers are in the group being trained to be maintainers.
      • Distribution:
        • Source 25 candidates (at least 5 by direct manager contact) by October 15 and hire 2 developers: 25 sourced (100%), 1 hired (50%)
      • Stan:
    • Infrastructure: Source 10 candidates by Oct 15 and hire 1 SRE (APAC) manager: X sourced (X%) X hired (X%)
      • SAE: Source 10 candidates by Oct 15 and hire 1 DBRE: 5 sourced (50%) 2 hired (200%)
      • SRE: Source 25 candidates by Oct 15 and hire 2 SREs: 0 sourced (0%) 2 hired (100%)
    • Ops Backend:
      • Hire 2 engineering managers (1 manager for Release, 1 manager for Secure): Hired 1 (50%)
      • Configure: Complete 10 introductory calls with prospective candidates by Nov 15 and hire 3 developers: 6 prospects (60%), hired 3 (150%)
      • Monitor: Source 30 candidates (at least 5 by direct manager contact) by November 15 and hire 3 developers: >30 sourced (100%), hired 2 (67%)
      • Secure: Source 75 candidates and hire 5 developers: 215 sourced (100%), 3 hired (60%)
      • CI/CD: Complete 10 introductory calls with prospective candidates by Nov 15 and hire 2 developers: 5 prospects (50%), hired 1 (50%)
    • Quality:
      • Source 100 candidates (at least 5 by direct manager contact) by October 15 and hire 2 test automation engineers: 52 sourced (52%), 2 hired (100%)
    • Security:
      • Source 100 candidates (at least 5 by direct manager contact) by October 15 and hire 3 security engineers: 100 sourced (100%), 6 hired (100%)
    • Support: Source 30 candidates by October 15 and hire 1 APAC Manager 100% - hired new manager in November
      • Amer East: Source 30 candidates by October 15 and hire 1 Support Agent
      • Amer West: Source 30 candidates by October 15 and hire 1 Support Engineer: hired 1 Support Engineer (100%)
      • APAC: Source 30 candidates by October 15 and hire 1 Support Engineer
    • UX:
      • Source 50 candidates by October 15 (at least 10 by direct manager contact) and hire 2 UX Designers and 1 UX Manager: 70 sourced (100%), 2 hired (66.6%)
    • VP Alliances: Scale team to interact with our key partners. Hire Content Manager and Alliance Manager
    • CFO: Improve processes and compliance
      • Legal: Implement a new contracting process to roll out with the vendor management process.
      • Legal: Update global stock option program to ensure global compliance.
    • CFO: Create plan for preparing the Company for IPO
    • CFO: Improve financial performance
      • FinOps: Annual planning process complete with company wide benchmarks and 100% of functions modeled to enable scenario analysis.
      • FinOps: Headcount planning linked to requisition process that provides for recruiting and hiring manager visibility against plan assumptions.
    • CCO: Improve Recruiting and People Ops Metrics, including relevant benchmarketing data
      • Recruiting: Hiring and Diversity stats and goals defined by functional group
      • Recruiting: Rent Index Evaluation by functional group
      • Recruting: Single Source of truth for Recruiting data, ownership and progress. If we can get this down to 2 sources in Q4, that will be a win.
      • PeopleOps: Added clarity to the turnover data and industry benchmarketing. Attempt to get benchmark data for remote companies.
    • CCO: Continue on the next iteration of Compensation
      • PeopleOps: Review and revise benchmarks to meet our current Compensation Philosophy
      • PeopleOps: Explore and decide on revising the compensation philosophy that supports hiring and retaining top talent, and have the plan in place to move forward.
      • PeopleOps: Review and revise Rent Indexes to align with the compensation philosophy.
    • CCO: Continue to improve Career Development and Feedback
      • Redesign the experience factor model to focus on supporting development and conduct trainings.
      • Define Cross-Functional and Functional Group specifc VP and Director Criteria
      • Launch an employee engagement survey that results in at least 85% participation and 3 themes to work on as a leadership team.
    • CCO: Improve Recruiting efficiency and effectiveness to build a solid foundation to scale:
      • Clearly define and standardize process, roles, responsibilities, and SLA’s
      • Improve continuity and accountability through recruiter/sourcer/coordinator alignment
      • Launch candidate survey in Greenhouse
      • Implement and integrate new background check provider
      • Enhance candidate communication and templates (invites, updates, decline notices)
      • Optimize Greenhouse and provide refresher training to hiring managers
      • Support Headcount planning process driven by Finance to ensure one source of truth and process for hiring plan and recruiter capacity
      • Continue work on ELO score for interviews. First interation complete by mid November.
    • Product: Hire 2 Directors of Product => *100% Complete- - hired Kenny and Eric

Engineering (VPE)

  • GOOD
    • Shipped several new impactful salary benchmarks
    • Shipped meaningful career development asset for BE engineers in the form of experience factor worksheets
    • GitLab.com is now meeting business goals
    • Error budgets allowed us to manage risk and solve some problems
  • BAD
    • Did not hire the second director of engineering we planned to
    • Did not address comp for every engineering role
    • Did not address career development for every engineering role
  • TRY
    • Delegate OKR process for managers to my directors in Q1
    • End error budget experiment in OKRs
    • Only do 2019 Q1 hiring KRs for teams that are not regularly hiring to plan
    • Delegate Director of Engineering hires to Sr Director of Development in Q1
    • Address comp expectations and career development for the remaining roles in engineering in Q1
    • Assume .com continues to meet business goals and not take on the mission critical goal in favor of more granular infra KRs in Q1
    • Arrange training in Q1 for career development facets (like writing secure code) for engineers

Frontend

  • GOOD
    • Met our engineering hiring goal and very close to closing also second manager. Ramped up the team for Hiring fast and the whole pipeline is now a team effort
    • Able to scale gitlab-ui, have the full workflow and the SSOT with design.gitlab.com while most of the work on documentation is automated
    • Onboarding new members with the frontend buddy system and constant prioritization of new hires worked well
    • We had a good analytical process of selecting a future proof charting library and drove it to integration a release cycle later
  • BAD
    • Scaling our CSS efforts is not effective enough and especially doesn’t scale
    • Work on performance and technical base work slowed down
    • Our assets grow sometimes suddenly due to changes and without looking constantly at Webpack reports no one is aware of it
  • TRY
    • Replan, Restructure and Rethink our CSS refactoring
    • Focus on performance improvements and distribute that wider in the team
    • Even as our tooling improved, there is way more that we can do to improve workflows, automation, setup time and consistency by tailoring solutions 100% to our needs
    • Bring more tooling for JS development to GitLab itself so we have it integrated in our workflow but also others can profit (Webpack reports, etc.)

Dev Backend

  • GOOD
    • Changes to retrospectives seem well-received: feedback is that they feel more productive and flow more cleanly
    • Clarified some terminology confusion with Product around prioritization
  • BAD
    • Cross-functional collaboration was hard to push due to absences.
    • Hard to have teams prioritize as we’d like due to gaps in the product.
    • Had a ‘regression’ in retrospective due to changing several things at once
  • TRY
    • Pushing more aggressively to prioritize what we need in the product
    • Keep in mind cross-functional collaboration may be difficult in Q4 due to holidays
    • Keep policy iterations (e.g. changes to retrospective format) atomic to better track success and avoid regressions

Plan

  • GOOD
    • Met the hiring target
    • No production incidents
    • One more person trained to lead technical interviews, another close to finishing their training
    • Single prioritised list works well for the team
  • BAD
    • Despite a significant amount of effort towards the Elasticsearch goal, we have not made any significant reductions in index disk usage
    • Elasticsearch issues have a tendency to get lost or deprioritised
    • One person left the team
    • Although we aren’t the most-impacted team, security issues have taken a lot of the team’s time over the past few months
  • TRY
    • Reframing Elasticsearch goal as improving admin controls, reduction in downtime during upgrades, progressive enablement, etc. These issues allow more incremental progress and don’t block improvement in index size
    • Reviewing security issues to ensure that we are clearing a backlog, rather than introducing new issues at the same rate as we fix old ones

Distribution

  • GOOD
    • The whole team is participating in all team projects (both Helm Charts and omnibus-gitlab)
    • Started with team trainings again which helped the team share knowledge faster
    • Keeping up with the incoming issues, all projects with open issues are trending down
    • Exposure of Helm Charts is growing and with that the feedback received
  • BAD
    • With the expanded work scope, context switching spiked
    • Helm Charts are taking majority of the teams time, which impacts the omnibus-gitlab package feature development
    • Higher complexity of the Helm Charts and the relative immaturity of tooling is making the team less efficient
    • Due to the new nature of Helm Charts, documentation is still lacking in spite of teams best effort to keep it updated frequently causing confusion with users
  • TRY
    • Find a way to remove duplicate work between the omnibus-gitlab package and Helm Charts in regards to introducing new configuration
    • Focus more on initial installation experience for the new users
    • Train more teams to be self sufficient with the Helm Charts
    • Ensure that the omnibus-gitlab package keeps receiving enough attention

Geo

  • GOOD
    • Reinstated the monthly team call which was received well by the team.
    • Cycle time on reviews (done on Geo work by the Geo team) has been reduced by concentrating on making sure that others are unblocked.
    • The team felt that they accomplished a lot on items like bug fixing, refactoring, performance improvements and documentation.
  • BAD
    • Some issues were not as prepared as they could have been and as a result the issues included in the milestone changed a lot during a month.
    • For some issues we were delayed in waiting for specialized reviewers. We know that there are improvements to the DB review process coming soon, and we look forward to these changes being in place.
    • We struggled with an erroneous revert as part of the automatic CE->EE merging. It wasn’t immediately clear that the revert had happened.
    • It was hard to know if some of the QA failures were due to our work or due to work on the QA framework itself. It is difficult and slow to figure out the difference
  • TRY
    • We need to spend more time in planning. We will introduce a process in 11.8 and see how this enables us to have more predictable deliverables in a milestone

Manage

  • GOOD
    • Stayed within the error budget, something we were nervous about after having an incident so early in the quarter.
    • We kept focus on deliverables for much of the quarter (e.g SmartCard, Group SAML, Billing improvements), which feels like an achievement given the size and scope of the team.
    • A decision has been made to create a new team (‘Sync’), which will help bring more focus to areas such as Billing & Licensing, while reducing the scope of the Manage team.
    • Made a conscious decision to prioritize reducing the security backlog in the coming milestones.
    • The impact of the 201 sessions (now known as Deep Dives) was really positive and has encouraged other teams to do similar.
  • BAD
    • Only hired 1 out of 2 positions.
    • Prioritization of security Issues will impact Deliverables until we clear down the backlog.
    • Number of Issues delivered was inconsistent for the 3 milestones in the quarter, making it hard to predict future delivery.
  • TRY
    • Work closer with the recruitment team, and other engineering teams, to understand what we need to do to speed up hiring.
    • Experiment with milestone planning to increase both velocity and predictability.

Create

  • GOOD
    • Met the hiring target!
    • No error budget points used
    • 2 people without significant pre-existing Golang experience are now comfortable contributing to our Golang projects
    • The pressure on and around the feature freeze was reduced by the new policy that allows MRs to be picked into stable after the 7th as long as they have a feature flag.
  • BAD
    • 1 team member left
    • 3 out of 5 people without significant pre-existing Golang experience did not end up contributing to Golang projects
    • We continue to overestimate the amount of work we can get done in a month, through underestimating the size of individual issues, resulting in higher pressure around the feature freeze and more slipping to the next release than desirable
    • Security issues have taken up a significant portion of each of the last few releases, resulting in less time to work on direction features
  • TRY
    • Continue to have people without Golang experience work on issues that require them to get familiar with our Golang components (Gitaly, Workhorse)
    • Break issues down more before scheduling them so that we have a better idea of the amount of work required and we can release them one part at a time, instead of the whole big MR slipping multiple releases in a row

Gitaly

  • GOOD
    • Delivered on 100% of OKRs
    • Hiring finally picked up for team
  • BAD
    • Delivered on 100% of OKRs :) Goals probably weren’t aggressive enough.
  • TRY
    • Setting more aggressive goals, particularly around delivering more rapid architectural improvements

Gitter

Stan

  • GOOD
    • Identified and fix an NFS bug for a number of customers
    • Reduced ~80 MB baseline memory usage by eliminating Linguist dependency
    • Reduced memory usage of the top 2 worst performing Sidekiq jobs by several gigabytes (UpdateMergeRequestsWorker and ReactiveCachingWorker)
    • Formalized maintainer role and started reviewing much more code
  • BAD
    • OKRs didn’t reflect a significant amount of the day-to-day work
    • Got pulled into scoping and improving audit logging for customer
    • Implemented Bitbucket Server importer for a customer
    • Got sidetracked helping many other teams (Geo, Infrastructure, etc.) and fixing production incidents (e.g. https://gitlab.com/gitlab-org/gitlab-ce/issues/53153)
  • TRY
    • Making OKRs actually reflect work done

Infrastructure

  • GOOD
    • We did well with hiring and onboarding
    • We improved our abilikty to manage workload (even as it still needs more work)
    • We cleaned up a significant backlog of team issues, which will allow us to fully focus on executing
  • BAD
    • OKRs were iffy at best: we treated them like homework rather than a proper management tool
    • We underestimated our workflow management at scale, which created confusiotn in the team
    • We failed to communicate new error budget frameork widely
  • TRY
    • Finish adopting OKRs and develop the framework of OKRs through GitLab issues
    • Ensure the OKR -> Epic -> Milestone -> Issue is crystal clear

SAE

  • GOOD
    • Hired 3 DBRE’s.
    • Delivered a good percentual of the OKR’s and another expressive project that was Patroni.
  • BAD
    • We missed an OKR to represent the project of Patroni.
  • TRY
    • Improve production best practices, work on better documentation and processes, to avoid self-inflicted outages.

SRE

  • GOOD
    • Got measurements in team velocity spreadsheet
    • Continuing to bring new team members up to speed
    • Hired 2 more SREs 1 EMEA / 1 Americas
  • BAD
    • DR not fully completed in quarter
    • 1 queue of issues for all team created information/tracking overload
  • TRY
    • Breaking apart team milestones to smaller groups tied to team level OKRs
    • Small sprint to finish DR work

Andrew

  • GOOD
    • Slack alerts to service issues with embedded grafana graphs to help visualise problems and reduce MTTD.
    • Delivered Correlation IDs. These are already proving very useful in diagnosing productions problems and also in self managed instances. This also formed the basis for distributed tracing, which will be delivered soon.
    • Abuse detection alerting is being used by the abuse team to detect and stop abuse.
  • BAD
    • Delivered two large, unplanned work pieces: Puma/Multithreaded App Server and Google Capacity Planning which were not accounted for in OKR planning or scoring.
    • More could be done to train team members on how to use new tools
  • TRY
    • Adjust OKRs during the quarter if priorities change
    • Focus on emitting more documentation through blueprints, videos and user documentation.

Ops Backend

  • GOOD
    • Completed goals for throughput OKR thanks to the support of the team to try a new model and seeing success with it.
    • Hired an engineering manager for release. Very happy to welcome Darby Frey to GitLab!
  • BAD
    • Did not hire a manager for secure. This has been a tough role to hire for despite our increased efforts in sourcing and filling the pipeline. We are seeing an uptick in the number of interviews in Q1 likely due to the efforts from Q4 but it’s disappointing that we are still in search.
  • TRY
    • Work with recruiting to diversify and improve our sourcing for the Secure team manager role.
    • Improve our presence at events and help widen our brand awareness on LinkedIn and other recruiting sites

Monitoring

  • GOOD
    • Sourced more candidates than our goal
    • No error budget points spent
  • BAD
    • Only hired 2 out of 3 positions
    • Didn’t meet throughput goal (only 79%)
    • Didn’t spend any error budget points. This is mostly due to our surface area still being so small.
  • TRY
    • Encourage smaller MRs merged earlier. Right now we have a number of small MRs, but they are all getting merged late in the cycle.

CI/CD

  • GOOD
    • Average throughput was significantly higher than our goal
  • BAD
    • Spent double our goal error budget
    • Only hired one engineer
  • TRY
    • More active sourcing and referral seeking to fill roles
    • Add some structure to how we prioritise and review community MRs to increase throughput

Secure

  • GOOD
    • Sourced more candidates then expected (215)
    • No error budget points spent
    • Met throughput goal
  • BAD
    • Only hired 3 out of 5 positions
  • TRY
    • Regarding our rejection ratio, source even more candidates

Configure

  • GOOD
    • Hired more than target, thanks to referrals and introductory calls
  • BAD
    • Didn’t meet the throughput goal (only 87%)
    • Reluctance to split GitLab issues due to the challenge of splitting our communication across multiple issues (significant tradeoff with the extensive async commonunication)
    • Not a single hire came from an application (all referral and sourced for introductory calls). Not clear if this is a problem since we exceeded our hiring target.
  • TRY
    • Stress the importance of smaller MRs more also try to make smaller issues
    • Ask for referrals more, encourage team members to share job ads on Linkedin
    • Continue our focus on Introductory calls since 6 calls led to 1 hire which is a good return on investment

Kamil

  • GOOD
    • Managed to finish major improvements related to CI scalability: database footprint and fixing Object Storage bugs,
  • BAD
    • The amount of OKR tasks didn’t really fit into spare time. I was constantly struggling with trying to find time for OKRs while doing regular development / reviews.
  • TRY
    • Focus on smaller tasks or define explicitly that OKR is not done in spare time, but rather in 1/3 of time to have better focus on.
    • Stop (disconnect) other interruptions while working on OKRs.

Philippe

  • GOOD
    • Several CFPs were posted, I should have at least one accepted.
  • BAD
    • Didn’t have time to benchmark any tool. It requires to be focused and available, which I’m not because of my interim EM position.
  • TRY
    • Find a new Secure Engineering Manager to allocate time for my new role OKRs.

Quality

  • GOOD
    • Hired two test automation engineers.
    • Completed Patroni Replication Manager testing with GitLab QA.
    • Completed TLS Depreciation testing with GitLab QA.
    • Completed Rails 5 migration testing with GitLab QA.
  • BAD
    • An update to the clone / fork button cause GitLab QA tests to break for 4 days.
    • Review Apps are broken (no DNS record) due to an increasing number of DNS records and too many API calls to AWS.
    • We had a few high severity regressions were as a result of Rails 5.
    • We are still lagging behind on test gaps.
  • TRY
    • We need to ensure that broken QA tests has the same urgency as a broken master.
    • Incentivize and encourage engineers to write E2E tests.

Security

  • GOOD
    • Public HackerOne bug bounty program rollout in December, with 1 hour first response time to new reports.
    • Security on-call rotation and automation successfully deployed with SLA of 15 minutes.
    • Completed initiative of discontinued support of TLS 1.0 and 1.1 on GitLab.com.
    • MTTM for S1/P1 security vulnerabilities consistently meets 30 day timeframe.
    • Biweekly application security office hours have been helpful so far.
  • BAD
    • Security releases, both critical and non-critical continue to be a challenge, with regression issues for previously hotpatched issues on GitLab.com.
    • MTTM of S2/P2 security issues are not consistently meeting 60 day timeframe.
  • TRY
    • Work with release delivery and management team to address security release issues.
    • Deploy mandatory onboarding secure coding training for all new (and existing) developers, to lower overall vulnerable code.

Support

  • GOOD
    • Positive team reaction to workflow changes to improve SLA response. Impressive KPI results with FRT trending close to 95% by EOQ.
    • Good cross-team coordination with Recruiting and hiring to plan.
    • Completed upgrade to Zendesk Enterprise and initial update to Zendesk workflows.
  • BAD
    • Experience factor review process had unnecessary friction suggesting iteration for a smoother Compensation process.
    • Although hiring goals were met we need to partner with Recruiting to lessen time to hire new roles.
  • TRY

Americas East

  • GOOD
    • First Response Hawk Process is improving SLAs
    • Team adopted more async workflows (Issue boards/Week-in-revew doc)
    • Offboarded APAC Support team to new manager successfully.
    • Will C. Developed Strace-parser which has helped tremendously with customer performance debugging.
  • BAD
    • We need more Geo Expertise.
    • Support Fix and Documentation processes were not as clearly defined as they should be which lead to delays and friction in using them.
  • TRY
    • Better communication and collaboration with docs to get changes merged in faster, more accurately.
    • Develop FRT Hawk Tooling to improve efficiency
    • Get other regions to use strace-parser through better training and documentation.

Americas West

  • GOOD
    • Hired to target
    • Have a solid first iteration on .com Engineering onboarding and clear next steps
    • .com performance in December was exceptional (after our BAD day)
    • Adhoc pairing sessions have reduced breaches and increased team knowledge
  • BAD
    • Weekend merge of a 4-hr SLA on Premium left the .com team scrambling
    • Coverage gap left .com with a 0% SLA day in December (only 1 ticket actually breached)
  • TRY
    • Have .com team members play the role of CMOC in the incident management process
    • Rotate through “FRT Hawk” role throughout team

APAC

  • GOOD
    • Manager hired for the region
  • BAD
    • Length of time to recruit manager for APAC
  • TRY
    • Define regional hours of coverage with the team and cover appropriately
    • Participate in the “FRT Hawk” role once completed in Americas
    • Define framework for multi-language support

EMEA

  • GOOD

    • Hired 2 Support Agents per target
    • Focused learning objectives for team with quarterly learning goals
    • Groundwork for knwledge sharing, documentation updates, search and tracking
    • Streamline Zendesk configuration to improve agent efficienc
  • BAD

    • Tracking of knowledge capture and documentation updates not yet in place
    • Frontend search of knowledge to deflect support tickets not yet in place
  • TRY

    • Track documentation updates and knowledge sharing by Support
    • Implement search of knowledge to deflect support tickets and track effectiveness
    • Continue to streamline and improve Zendesk configuration with help of Support Operations specialist when hired.

UX

  • GOOD
    • We developed and finalized an MVC for both group and instance level clusters with collaboration between UX, Frontend, Backend, and Product.
    • We also created another user journey towards creating a project cluster by introducing the Serverless tab. This navigates users towards the cluster page if they do not already have one. This was a joint effort between the Configure team and Triggermesh (a contracted group).
    • Completed a comprehensive competitor analysis of version control tools for designers.
    • Hired 2 UX Designers.
  • BAD
    • While we pushed for group and instance level clusters to be implemented, instance level clusters did not make it in Q4. The team did not have capacity.
    • Didn’t hire a UX Manager.
  • TRY
    • Better defining what we want and need from UX Managers.

Research

  • GOOD
    • We launched GitLab First Look, a new comprehensive research program for users of GitLab. Users are now able to sign-up to beta testing with GitLab.
    • Great collaboration with Product and Marketing.
  • BAD
    • Due to the holidays, some of the features we planned to beta test missed RCs/11.6 release therefore we’ve not had the opportunity to test out and re-iterate on our beta testing process.
  • TRY
    • Test the beta testing process for features in 11.7