\nWe are looking for a skillful DevOps Engineer to join one of our teams working with US clients - referrals from the high-end American IT market.\n\n\nThis position offers a contract lasting for at least 1 year, with the potential for extension as new projects come in.\n\n\nOur tech stack is built around React, React Native, Node.js, and TypeScript. We are working with startups in their early stages (MVPs), which makes the job full of diverse and exciting challenges while crafting innovative solutions.\n\n\nEnjoy the flexibility of working within the Polish time zone and collaborate with clients from diverse industries such as FinTech, SportsTech, MedTech, and many others.\n\n\nSalary: PLN 9.000 - 13.000 + VAT (B2B)\n\n\n\n\n* Required skills: 1+ year of experience in DevOps or similar roles, basic understanding and hands-on experience with AWS, experience with CI/CD pipelines, knowledge of containerization, familiarity with infrastructure as code, scripting skills (i.e., Bash, Python), understanding of monitoring tools, English B2, Polish C1\n* Offer: B2B/UZL, min. 1 year contract, paid days off, 100% remote (or hybrid/onsite - as you prefer), Polish time zone, flexible hours, full-time position, Scrum Teams\n* Perks and Benefits: +6 000 PLN / year training budget, budget for health insurance and sports, Mentoring sessions, individual Career Path, and more\n* Tools you'll use: Apple MacBook Pro, Slack, GSuite, GitHub, Jira, Coda, Miro\n\n\n\nYour future tasks and duties:\n* Assist in developing and maintaining CI/CD pipelines\n* Help manage cloud infrastructure\n* Support implementation and oversight of monitoring and alerting systems\n* Collaborate with development teams to ensure seamless integration and deployment\n* Troubleshoot and resolve basic infrastructure issues\n* Participate in project discussions and provide technical insights\n* * Participate in on-call duty rotations to ensure 24/7 system availability, with additional compensation for on-call duties\n\n\n\nAbout you - Tech skills & Experience:\n* 1+ year of experience in DevOps or similar roles\n* Basic understanding and hands-on experience with AWS, including deployment and management of applications.\n* Experience with setting up and maintaining Continuous Integration and Continuous Deployment (CI/CD) pipelines using tools like Jenkins or GitHub Actions.\n* Understanding of container technologies such as Docker and Kubernetes for application packaging and deployment.\n* Familiarity with infrastructure as code (i.e., Terraform, Ansible)\n* Proficiency in scripting languages like Bash and Python for automation and configuration management.\n* Knowledge of monitoring and alerting tools such as Prometheus and Grafana to ensure system performance and availability.\n\n\n\nAbout you - Soft Skills:\n* Excellent communication skills in English (B2) and Polish (C1)\n* Open-minded and proactive in solving challenges\n* Strong team collaboration abilities\n* Eager to stay updated with industry trends\n* Problem-solving abilities to diagnose issues, suggest solutions, and make decisions based on project requirements\n\n\n\nNice to have:\n* Experience in an Agile/Scrum environment\n* Familiarity with security best practices\n* Previous work on early-stage product development\n\n\n\n\n\n\n\n\n\n\nMeaningful projects, top-notch experts and value-driven workplace\nSounds good?\nApply now! \n\n#Salary and compensation\n
No salary data published by company so we estimated salary based on similar jobs related to React, Python, Docker, DevOps, Cloud, Junior and Engineer jobs that are similar:\n\n
$60,000 — $100,000/year\n
\n\n#Benefits\n
๐ฐ 401(k)\n\n๐ Distributed team\n\nโฐ Async\n\n๐ค Vision insurance\n\n๐ฆท Dental insurance\n\n๐ Medical insurance\n\n๐ Unlimited vacation\n\n๐ Paid time off\n\n๐ 4 day workweek\n\n๐ฐ 401k matching\n\n๐ Company retreats\n\n๐ฌ Coworking budget\n\n๐ Learning budget\n\n๐ช Free gym membership\n\n๐ง Mental wellness budget\n\n๐ฅ Home office budget\n\n๐ฅง Pay in crypto\n\n๐ฅธ Pseudonymous\n\n๐ฐ Profit sharing\n\n๐ฐ Equity compensation\n\nโฌ๏ธ No whiteboard interview\n\n๐ No monitoring system\n\n๐ซ No politics at work\n\n๐ We hire old (and young)\n\n
\n\n#Location\nWrocลaw
๐ Please reference you found the job on Remote OK, this helps us get more companies to post here, thanks!
When applying for jobs, you should NEVER have to pay to apply. You should also NEVER have to pay to buy equipment which they then pay you back for later. Also never pay for trainings you have to do. Those are scams! NEVER PAY FOR ANYTHING! Posts that link to pages with "how to work online" are also scams. Don't use them or pay for them. Also always verify you're actually talking to the company in the job post and not an imposter. A good idea is to check the domain name for the site/email and see if it's the actual company's main domain name. Scams in remote work are rampant, be careful! Read more to avoid scams. When clicking on the button to apply above, you will leave Remote OK and go to the job application page for that company outside this site. Remote OK accepts no liability or responsibility as a consequence of any reliance upon information on there (external sites) or here.
The Role:\nAztec is gearing up to launch a sophisticated privacy-preserving blockchain on Ethereum. To achieve this, our engineers need support making our code run through first continuous integration, then test networks, and finally peer-to-peer software that anyone in the world can run. We need you! We seek engineers versed in devops and system architecture that are enthusiastic about building production systems with unique and challenging constraints\n\n\n\n* Design and build the foundation for monitoring a decentralized private network.\n\n* Develop metrics and monitoring of our hosted services.\n\n* Automate infrastructure provisioning and configuration.\n\n* Manage AWS services and resources ensuring cost effectiveness.\n\n* Collaborate with development teams to ensure they have the best development/CI experience.\n\n* Maintain a CI/CD pipeline for 30 engineers working across C++, Rust, Noir, Solidity, and Typescript.\n\n* Monitor system issues and create strategies for their detection.\n\n\n\nResponsibilities: \n\n\n* CI platforms such Github Actions, CircleCI.\n\n* AWS cloud services, EC2, ECS, ECR, etc.\n\n* Expert in Docker and creating fast efficient builds and containers. Earthly is a nice-to-have.\n\n* Terraform for defining IAC.\n\n* Prometheus and Grafana for metrics and monitoring, and working with codebases to extract required information.\n\n* Toolchain experience: C++ (clang, gcc, cmake). Rust. Solidity. NPM/Yarn.\n\n* Demonstrable experience in enabling an engineering team of our size have a robust and efficient development experience.\n\n* Self starter. Able to clearly identify areas to prioritise to deliver most value.\n\n\n\nQualifications: \n\n\n* Bachelor's or Master's degree in Computer Science, Information Technology, or a related field.\n\n* 4+ years of experience in site reliability engineering or DevOps roles, preferably in the cryptocurrency or financial services industry.\n\n* Strong communication and collaboration skills, with the ability to work effectively in a fast-paced, dynamic environment.\n\n* Can operate in Greenwich Mean Time Zone\n\n\n\nWhat we offer:\n\n\n* Flexible and remote work environment\n\n* 25 days holiday + bank holidays annually\n\n* Additional benefits include health insurance, retirement plans, and opportunities for professional development.\n\n* Quarterly offsite travel for collaboration \n\n* Events and conference budget\n\n* An opportunity to work at the cutting edge of blockchain and FinTech with a world class cryptography and engineering team\n\n\n\n\nCompensation Range: $165,000 - $209,000 + equity + additional benefits. The salary for this position will be commensurate with experience and qualifications. \n\n#Salary and compensation\n
No salary data published by company so we estimated salary based on similar jobs related to Docker, Travel, DevOps, Cloud and Engineer jobs that are similar:\n\n
$70,000 — $110,000/year\n
\n\n#Benefits\n
๐ฐ 401(k)\n\n๐ Distributed team\n\nโฐ Async\n\n๐ค Vision insurance\n\n๐ฆท Dental insurance\n\n๐ Medical insurance\n\n๐ Unlimited vacation\n\n๐ Paid time off\n\n๐ 4 day workweek\n\n๐ฐ 401k matching\n\n๐ Company retreats\n\n๐ฌ Coworking budget\n\n๐ Learning budget\n\n๐ช Free gym membership\n\n๐ง Mental wellness budget\n\n๐ฅ Home office budget\n\n๐ฅง Pay in crypto\n\n๐ฅธ Pseudonymous\n\n๐ฐ Profit sharing\n\n๐ฐ Equity compensation\n\nโฌ๏ธ No whiteboard interview\n\n๐ No monitoring system\n\n๐ซ No politics at work\n\n๐ We hire old (and young)\n\n
๐ Please reference you found the job on Remote OK, this helps us get more companies to post here, thanks!
When applying for jobs, you should NEVER have to pay to apply. You should also NEVER have to pay to buy equipment which they then pay you back for later. Also never pay for trainings you have to do. Those are scams! NEVER PAY FOR ANYTHING! Posts that link to pages with "how to work online" are also scams. Don't use them or pay for them. Also always verify you're actually talking to the company in the job post and not an imposter. A good idea is to check the domain name for the site/email and see if it's the actual company's main domain name. Scams in remote work are rampant, be careful! Read more to avoid scams. When clicking on the button to apply above, you will leave Remote OK and go to the job application page for that company outside this site. Remote OK accepts no liability or responsibility as a consequence of any reliance upon information on there (external sites) or here.
This job post is closed and the position is probably filled. Please do not apply. Work for Triplelift and want to re-open this job? Use the edit link in the email when you posted the job!
About TripleLift\n\nTripleLift, one of the fastest-growing ad tech companies in the world, is rooted at the intersection of creative and media. Its mission is to make advertising better for everyoneโ content owners, advertisers and consumersโby reinventing ad placement one medium at a time. With direct inventory sources, diverse product lines, and creative designed for scale using our Computer Vision technology, TripleLift is driving the next generation of programmatic advertising from desktop to television.\n\nAs of January 2021, TripleLift has recorded five years of consecutive growth of greater than 70 percent. TripleLift is a Business Insider Hottest Ad Tech Company, Inc. Magazine 5000, Crain's New York Fast 50, Deloitte Technology Fast 500 and among Incโs Best Workplaces. Find more information about how TripleLift is shaping the future of advertising at triplelift.com.\n\nThe Role\n\nTripleLift is seeking an experienced DevOps engineer to join our team full time. We are a fast-growing startup in the advertising technology sector, trying to tackle some of the most challenging problems facing the industry. As a DevOps engineer, you will be responsible for providing leverage to the engineering team to do the best possible work. This includes managing the infrastructure, working with them to improve their deployment and release process, as well as constantly searching for ways to improve our infrastructure.\n\nCore Technologies\n\nWe employ a wide variety of technologies here at TripleLift to accomplish our goals. From our early days, weโve always believed in using the right tools for the right job, and continue to explore new technology options as we grow. The DevOps team uses the following technologies at TripleLift:\n\nTools: Chef, Ansible, Terraform, Docker, Kubernetes, CircleCI, Spinnaker, Prometheus, Grafana, Vault, Consul, Snowflake, Airflow, Databricks \nDatabases: AeroSpike, RDS MySQL, Redshift, MongoDB, and more\nLanguages: Java, Python, Node.js, TypeScript, Scala, and more\nAmazon Web Services and Google Cloud (GCP) to keep everything humming\nResponsibilities\n\nCollaborate with the rest of the engineering team to come up with best practices for writing and scaling good code;\nImprove our infrastructure and deployment processes;\nBuild tools that make every engineer more productive;\nWork with each team to optimize their application performance;\nDevelop a unified system for monitor, logging and error handling;\nSearch for industry best practices and use them to drive our team forward.\nWork with teams to optimize and reduce cloud costs;\nDesired Skills and Attributes\n\nSignificant experience in a DevOps or SRE role;\nUnderstanding of container technologies, like Docker and what it takes to containerize applications. \nLoves automation and automating repetitive work;\nUnderstands best practices of application, data, and cloud security;\nUnderstands best practices around building scalable, reliable, and highly available secure infrastructure;\nStrong understanding of cloud networking and network architecture, especially in the context of multi-region applications. \nSkilled in software provisioning, configuration management, and infrastructure automation tools;\nAbility to code well in at least one programming language;\nComfortable taking ownership of projects and showcasing key accomplishments;\nStrives for continued learning opportunities to build upon craft;\nExcellent organizational skills and attention to detail;\nAbility to work quickly and independently with minimal oversight;\nAbility to work under pressure and multitask in a fast-paced start-up environment;\nDesire to accept feedback and constructive criticism;\nExtremely strong and demonstrable work ethic;\nProven academic and/or professional achievement.\nEducation Requirement\n\nA Bachelorโs degree in a technical subject is preferred, although candidates with relevant experience who hold other degrees will be considered.\n\nExperience Requirement\n\nAt least five years of working experience in a professional, collaborative environment.\n\nLocation\n\nNew York or Kitchener-Waterloo preferred, but open to remote candidates\n\nBenefits and Company Perks\n\n100% Medical, Dental & Vision Plans\nUnlimited PTO\n401k, FSA, Commuter Benefits\nWeekly Yoga & Bootcamp\nMembership to Headspace (Meditation)\nOngoing professional development\nAmazing company culture\nNote: The Fair Labor Standards Act (FLSA) is a federal labor law of general and nationwide application, including Overtime, Minimum Wages, Child Labor Protections, and the Equal Pay Act. This role is an FLSA exempt role.\n\nAwards\n\nWe love celebrating our achievements. They remind us of our contributions making advertising work for everyone, and the TripleLifters who make it all possible. TripleLift is proud to be recognized by Inc. as a Best Workplace for our culture and benefits, and among Incโs Best in Business for our innovations and positive impact on the industry. \n\nTo check out more of our awards and distinctions, please visit https://triplelift.com/ideas/#distinctions\n\nDiversity, Equity, Inclusion and Accessibility at TripleLift \n\nAt TripleLift, we believe in the power of diversity, equity, inclusion and accessibility. Our culture enables individuals to share their uniqueness and contribute as part of a team. With our DEIA initiatives, TripleLift is a place that works for you, and where you can feel a sense of belonging. At TripleLift, we will consider and champion all qualified applicants for employment without regard to race, creed, color, religion, national origin, sex, age, disability, sexual orientation, gender identity, gender expression, genetic predisposition, veteran, marital, or any other status protected by law. TripleLift is proud to be an equal opportunity employer.\n\nTripleLift does not accept unsolicited resumes from any type of recruitment search firm. Any resume submitted in the absence of a signed agreement will become the property of TripleLift and no fee shall be due. \n\nPlease mention the words **KID TWENTY SOLVE** when applying to show you read the job post completely (#RMy4xNDMuMjU1Ljkw). This is a feature to avoid spam applicants. Companies can search these words to find applicants that read this and see they're human.\n\n \n\n#Salary and compensation\n
$120,000 — $200,000/year\n
\n\n#Location\nUnited States, Eastern Standard Time Zone
# How do you apply?\n\nThis job post has been closed by the poster, which means they probably have enough applicants now. Please do not apply.
This job post is closed and the position is probably filled. Please do not apply. Work for Splitgraph and want to re-open this job? Use the edit link in the email when you posted the job!
# We're building the Data Platform of the Future\nJoin us if you want to rethink the way organizations interact with data. We are a **developer-first company**, committed to building around open protocols and delivering the best experience possible for data consumers and publishers.\n\nSplitgraph is a **seed-stage, venture-funded startup hiring its initial team**. The two co-founders are looking to grow the team to five or six people. This is an opportunity to make a big impact on an agile team while working closely with the\nfounders.\n\nSplitgraph is a **remote-first organization**. The founders are based in the UK, and the company is incorporated in both USA and UK. Candidates are welcome to apply from any geography. We want to work with the most talented, thoughtful and productive engineers in the world.\n# Open Positions\n**Data Engineers welcome!** The job titles have "Software Engineer" in them, but at Splitgraph there's a lot of overlap \nbetween data and software engineering. We welcome candidates from all engineering backgrounds.\n\n[Senior Software Engineer - Backend (mainly Python)](https://www.notion.so/splitgraph/Senior-Software-Engineer-Backend-2a2f9e278ba347069bf2566950857250)\n\n[Senior Software Engineer - Frontend (mainly TypeScript)](https://www.notion.so/splitgraph/Senior-Software-Engineer-Frontend-6342cd76b0df483a9fd2ab6818070456)\n\nโ [**Apply to Job**](https://4o99daw6ffu.typeform.com/to/ePkNQiDp) โ (same form for both positions)\n\n# What is Splitgraph?\n## **Open Source Toolkit**\n\n[Our open-source product, sgr,](https://www.github.com/splitgraph/splitgraph) is a tool for building, versioning and querying reproducible datasets. It's inspired by Docker and Git, so it feels familiar. And it's powered by PostgreSQL, so it works seamlessly with existing tools in the Postgres ecosystem. Use Splitgraph to package your data into self-contained\ndata images that you can share with other Splitgraph instances.\n\n## **Splitgraph Cloud**\n\nSplitgraph Cloud is a platform for data cataloging, integration and governance. The user can upload data, connect live databases, or "push" versioned snapshots to it. We give them a unified SQL interface to query that data, a catalog to discover and share it, and tools to build/push/pull it.\n\n# Learn More About Us\n\n- Listen to our interview on the [Software Engineering Daily podcast](https://softwareengineeringdaily.com/2020/11/06/splitgraph-data-catalog-and-proxy-with-miles-richardson/)\n\n- Watch our co-founder Artjoms present [Splitgraph at the Bay Area ClickHouse meetup](https://www.youtube.com/watch?v=44CDs7hJTho)\n\n- Read our HN/Reddit posts ([one](https://news.ycombinator.com/item?id=24233948) [two](https://news.ycombinator.com/item?id=23769420) [three](https://news.ycombinator.com/item?id=23627066) [four](https://old.reddit.com/r/datasets/comments/icty0r/we_made_40k_open_government_datasets_queryable/))\n\n- [Read our blog](https://www.splitgraph.com/blog)\n\n- Read the slides from our early (2018) presentations: ["Docker for Data"](https://www.slideshare.net/splitgraph/splitgraph-docker-for-data-119112722), [AHL Meetup](https://www.slideshare.net/splitgraph/splitgraph-ahl-talk)\n\n- [Follow us on Twitter](https://ww.twitter.com/splitgraph)\n\n- [Find us on GitHub](https://www.github.com/splitgraph)\n\n- [Chat with us in our community Discord](https://discord.gg/eFEFRKm)\n\n- Explore the [public data catalog](https://www.splitgraph.com/explore) where we index 40k+ datasets\n\n# How We Work: What's our stack look like?\n\nWe prioritize developer experience and productivity. We resent repetition and inefficiency, and we never hesitate to automate the things that cause us friction. Here's a sampling of the languages and tools we work with:\n\n- **[Python](https://www.python.org/) for the backend.** Our [core open source](https://www.github.com/splitgraph/splitgraph) tech is written in Python (with [a bit of C](https://github.com/splitgraph/Multicorn) to make it more interesting), as well as most of our backend code. The Python code powers everything from authentication routines to database migrations. We use the latest version and tools like [pytest](https://docs.pytest.org/en/stable/), [mypy](https://github.com/python/mypy) and [Poetry](https://python-poetry.org/) to help us write quality software.\n\n- **[TypeScript](https://www.typescriptlang.org/) for the web stack.** We use TypeScript throughout our web stack. On the frontend we use [React](https://reactjs.org/) with [next.js](https://nextjs.org/). For data fetching we use [apollo-client](https://www.apollographql.com/docs/react/) with fully-typed GraphQL queries auto-generated by [graphql-codegen](https://graphql-code-generator.com/) based on the schema that [Postgraphile](https://www.graphile.org/postgraphile) creates by introspecting the database.\n\n- [**PostgreSQL](https://www.postgresql.org/) for the database, because of course.** Splitgraph is a company built around Postgres, so of course we are going to use it for our own database. In fact, we actually have three databases. We have `auth-db` for storing sensitive data, `registry-db` which acts as a [Splitgraph peer](https://www.splitgraph.com/docs/publishing-data/push-data) so users can push Splitgraph images to it using [sgr](https://www.github.com/splitgraph/splitgraph), and `cloud-db` where we store the schemata that Postgraphile uses to autogenerate the GraphQL server.\n\n- [**PL/pgSQL](https://www.postgresql.org/docs/current/plpgsql.html) and [PL/Python](https://www.postgresql.org/docs/current/plpython.html) for stored procedures.** We define a lot of core business logic directly in the database as stored procedures, which are ultimately [exposed by Postgraphile as GraphQL endpoints](https://www.graphile.org/postgraphile/functions/). We find this to be a surprisingly productive way of developing, as it eliminates the need for manually maintaining an API layer between data and code. It presents challenges for testing and maintainability, but we've built tools to help with database migrations and rollbacks, and an end-to-end testing framework that exercises the database routines.\n\n- [**PostgREST](https://postgrest.org/en/v7.0.0/) for auto-generating a REST API for every repository.** We use this excellent library (written in [Haskell](https://www.haskell.org/)) to expose an [OpenAPI](https://github.com/OAI/OpenAPI-Specification)-compatible REST API for every repository on Splitgraph ([example](http://splitgraph.com/mildbyte/complex_dataset/latest/-/api-schema)).\n\n- **Lua ([luajit](https://luajit.org/luajit.html) 5.x), C, and [embedded Python](https://docs.python.org/3/extending/embedding.html) for scripting [PgBouncer](https://www.pgbouncer.org/).** Our main product, the "data delivery network", is a single SQL endpoint where users can query any data on Splitgraph. Really it's a layer of PgBouncer instances orchestrating temporary Postgres databases and proxying queries to them, where we load and cache the data necessary to respond to a query. We've added scripting capabilities to enable things like query rewriting, column masking, authentication, ACL, orchestration, firewalling, etc.\n\n- **[Docker](https://www.docker.com/) for packaging services.** Our CI pipeline builds every commit into about a dozen different Docker images, one for each of our services. A production instance of Splitgraph can be running over 60 different containers (including replicas).\n\n- **[Makefile](https://www.gnu.org/software/make/manual/make.html) and** [docker-compose](https://docs.docker.com/compose/) **for development.** We use [a highly optimized Makefile](https://www.splitgraph.com/blog/makefile) and `docker-compose` so that developers can easily spin-up a stack that mimics production in every way, while keeping it easy to hot reload, run tests, or add new services or configuration.\n\n- **[Nomad](https://www.nomadproject.io/) for deployment and [Terraform](https://www.terraform.io/) for provisioning.** We use Nomad to manage deployments and background tasks. Along with Terraform, we're able to spin up a Splitgraph cluster on AWS, GCP, Scaleway or Azure in just a few minutes.\n\n- **[Airflow](https://airflow.apache.org/) for job orchestration.** We use it to run and monitor jobs that maintain our catalog of [40,000 public datasets](https://www.splitgraph.com/blog/40k-sql-datasets), or ingest other public data into Splitgraph.\n\n- **[Grafana](https://grafana.com/), [Prometheus](https://prometheus.io/), [ElasticSearch](https://www.elastic.co/), and [Kibana](https://www.elastic.co/kibana) for monitoring and metrics.** We believe it's important to self-host fundamental infrastructure like our monitoring stack. We use this to keep tabs on important metrics and the health of all Splitgraph deployments.\n\n- **[Mattermost](https://mattermost.com/) for company chat.** We think it's absolutely bonkers to pay a company like Slack to hold your company communication hostage. That's why we self-host an instance of Mattermost for our internal chat. And of course, we can deploy it and update it with Terraform.\n\n- **[Matomo](https://matomo.org/) for web analytics.** We take privacy seriously, and we try to avoid including any third party scripts on our web pages (currently we include zero). We self-host our analytics because we don't want to share our user data with third parties.\n\n- **[Metabase](https://www.metabase.com/) and [Splitgraph](https://www.splitgraph.com) for BI and [dogfooding](https://en.wikipedia.org/wiki/Eating_your_own_dog_food)**. We use Metabase as a frontend to a Splitgraph instance that connects to Postgres (our internal databases), MySQL (Matomo's database), and ElasticSearch (where we store logs and DDN analytics). We use this as a chance to dogfood our software and produce fancy charts.\n\n- **The occasional best-of-breed SaaS services** **for organization.** As a privacy-conscious, independent-minded company, we try to avoid SaaS services as much as we can. But we still find ourselves unable to resist some of the better products out there. For organization we use tools like [Zoom](https://www.zoom.us) for video calls, [Miro](https://miro.com/) for brainstorming, [Notion](https://www.notion.so) for documentation (you're on it!), [Airtable for workflow management](https://airtable.com/), [PivotalTracker](https://www.pivotaltracker.com/) for ticketing, and [GitLab for dev-ops and CI](https://about.gitlab.com/).\n\n- **Other fun technologies** including [HAProxy](http://www.haproxy.org/), [OpenResty](https://openresty.org/en/), [Varnish](https://varnish-cache.org/), and bash. We don't touch them much because they do their job well and rarely break.\n\n# Life at Splitgraph\n**We are a young company building the initial team.** As an early contributor, you'll have a chance to shape our initial mission, growth and company values.\n\n**We think that remote work is the future**, and that's why we're building a remote-first organization. We chat on [Mattermost](https://mattermost.com/) and have video calls on Zoom. We brainstorm with [Miro](https://miro.com/) and organize with [Notion](https://www.notion.so).\n\n**We try not to take ourselves too seriously**, but we are goal-oriented with an ambitious mission.\n\n**We believe that as a small company, we can out-compete incumbents** by thinking from first principles about how organizations interact with data. We are very competitive.\n\n# Benefits\n- Fully remote\n\n- Flexible working hours\n\n- Generous compensation and equity package\n\n- Opportunity to make high-impact contributions to an agile team\n\n# How to Apply? Questions?\n[**Complete the job application**](https://4o99daw6ffu.typeform.com/to/ePkNQiDp)\n\nIf you have any questions or concerns, feel free to email us at [[email protected]](mailto:[email protected]) \n\nPlease mention the words **DESERT SPELL GOWN** when applying to show you read the job post completely (#RMy4xNDMuMjU1Ljkw). This is a feature to avoid spam applicants. Companies can search these words to find applicants that read this and see they're human.\n\n \n\n#Location\nWorldwide
# How do you apply?\n\nThis job post has been closed by the poster, which means they probably have enough applicants now. Please do not apply.