<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Home on Arian Svirsky | DevOps Engineering</title><link>/</link><description>Recent content in Home on Arian Svirsky | DevOps Engineering</description><generator>Hugo</generator><language>en</language><atom:link href="/index.xml" rel="self" type="application/rss+xml"/><item><title>A Multi-Cloud Identity Platform</title><link>/experience/descope/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/experience/descope/</guid><description>&lt;p>&lt;strong>Descope&lt;/strong> — identity and authentication infrastructure&lt;/p>
&lt;p>Part of the platform team behind an auth product that processes millions of requests per day across 4 global regions. My remit is the boring but load-bearing part: the infra stays up, the deploys stay boring, and production never surprises us.&lt;/p>
&lt;h3 id="scale--surface-area">Scale &amp;amp; surface area&lt;/h3>
&lt;ul>
&lt;li>&lt;strong>25+ microservices&lt;/strong> in one platform, running simultaneously on &lt;strong>AWS (primary) and GCP&lt;/strong>&lt;/li>
&lt;li>&lt;strong>4 regions&lt;/strong>, each a fully independent production environment&lt;/li>
&lt;li>&lt;strong>Millions of auth requests/day&lt;/strong> across customer workloads&lt;/li>
&lt;li>&lt;strong>TypeScript + Pulumi&lt;/strong> as the IaC substrate — every cluster, every service, every env&lt;/li>
&lt;/ul>
&lt;h3 id="what-i-own-and-drive">What I own and drive&lt;/h3>
&lt;p>&lt;strong>Reliability &amp;amp; observability&lt;/strong> — Full Datadog footprint (APM, structured logs, dashboards, SLOs) across every region. I build and maintain the dashboards on-call actually uses, set the SLO targets for services I own, and drive the runbook standard for infra-level incidents. When latency creeps in Singapore, we see it before customers do.&lt;/p></description></item><item><title>Migrating a Security Product to Kubernetes</title><link>/experience/paloalto/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/experience/paloalto/</guid><description>&lt;p>&lt;strong>Palo Alto Networks&lt;/strong> (Cortex XSOAR, formerly Demisto) — 6 years&lt;/p>
&lt;p>Joined Demisto as a DevOps Engineer, stayed through the Palo Alto Networks acquisition, left as &lt;strong>Principal DevOps Engineer&lt;/strong>. Led infrastructure architecture and mentored a growing DevOps team through a startup-to-enterprise transition.&lt;/p>
&lt;h3 id="the-hard-problem-docker-in-docker-on-kubernetes">The hard problem: Docker-in-Docker on Kubernetes&lt;/h3>
&lt;p>Cortex XSOAR is a SOAR product — it runs customer automation playbooks inside Docker containers. When we moved from EC2 to Kubernetes, I led the design of how to securely run Docker-in-Docker inside Kubernetes pods.&lt;/p></description></item><item><title>The Linux Foundation</title><link>/experience/previous/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/experience/previous/</guid><description>&lt;p>&lt;strong>2014–2018&lt;/strong> — inManage, Interhost Networks, Calanit&lt;/p>
&lt;p>Before Kubernetes existed in my vocabulary, I was managing Linux servers, configuring DNS zones, hardening firewalls, and writing Bash scripts that are probably still running somewhere.&lt;/p>
&lt;p>&lt;strong>Interhost Networks&lt;/strong> was web hosting infrastructure — hundreds of customer sites, shared servers, the kind of environment where you learn DNS, SSL, and Apache configuration by necessity. When something broke at 2am, there was no Kubernetes self-healing. You SSH&amp;rsquo;d in and fixed it.&lt;/p></description></item><item><title>About</title><link>/about/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/about/</guid><description>&lt;p>&lt;img alt="Arian Svirsky" src="/images/arian.png">&lt;/p>
&lt;p>Staff-level DevOps / Infrastructure Engineer with 10+ years of hands-on experience. Specialized in Kubernetes, multi-cloud platforms, and production systems at scale. I&amp;rsquo;ve led infrastructure design at two very different companies — a unicorn identity startup and an enterprise security product — and taken both from &amp;ldquo;it works&amp;rdquo; to &amp;ldquo;it scales, it&amp;rsquo;s observable, and it doesn&amp;rsquo;t wake anyone up.&amp;rdquo;&lt;/p>
&lt;p>Started in tech support in 2014. Moved to Linux administration, then hosting infrastructure, then eCommerce platforms, then Docker, then Kubernetes, then owning multi-cloud platforms end-to-end. Each step was a deliberate climb toward harder problems and more ownership.&lt;/p></description></item><item><title>ariansvi.com — Cloud-Native on a Budget</title><link>/projects/this-site/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/projects/this-site/</guid><description>&lt;p>A personal website that runs on the same infrastructure patterns I use in production — but tuned to cost less than a coffee per month instead of a small car.&lt;/p>
&lt;h3 id="how-it-works">How it works&lt;/h3>
&lt;pre tabindex="0">&lt;code> git push → GitHub Actions → Artifact Registry → Cloud Run
 (lint, test, build) (Docker images) (serverless, scales to zero)
&lt;/code>&lt;/pre>&lt;p>Push code → GitHub Actions lints + tests + builds Docker images → pushes to Google Artifact Registry → deploys both services to Cloud Run via &lt;code>gcloud run deploy&lt;/code>. End-to-end in a few minutes. Auth to GCP is keyless (Workload Identity Federation).&lt;/p></description></item><item><title>Live Stats</title><link>/stats/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/stats/</guid><description/></item><item><title>Tools &amp; Technologies</title><link>/skills/overview/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/skills/overview/</guid><description>&lt;p>Not a checklist — these are things I actually use in production and have opinions about.&lt;/p>
&lt;h2 id="kubernetes">Kubernetes&lt;/h2>
&lt;p>EKS, GKE — I&amp;rsquo;ve run clusters on both. Helm for packages, Kustomize for overlays, ArgoCD for GitOps. Opinions: Autopilot is great until you need node-level access. Managed node groups are the sweet spot for most teams.&lt;/p>
&lt;h2 id="infrastructure-as-code">Infrastructure as Code&lt;/h2>
&lt;p>Terraform for cloud resources (modules, remote state, workspaces). Pulumi when the team prefers real programming languages over HCL. I&amp;rsquo;ve used both in production and can argue for either depending on the context.&lt;/p></description></item></channel></rss>