find a remote job
work from anywhere

Get new remote Apache + Cloud + Engineer + Security + Technical jobs sent to

Subscribe
×
๐Ÿ‘ฉโ€๐Ÿ’ป Join Remote OK ๐Ÿ‘‹  Log in
General
Remote OK Frontpage ๐Ÿ Remote jobs ๐ŸŒ—  Dark mode ๐Ÿ‘ฉโ€๐Ÿ’ป Hire remote workers ๐Ÿšจ Post a remote job ๐Ÿฑ Compact mode โœ๏ธ Remote work blog new
Top jobs
๐Ÿฆพ  AI Jobs
โฐ Async jobs ๐ŸŒŽ Distributed team ๐Ÿค“ Engineer jobs ๐Ÿ’ผ Executive jobs ๐Ÿ‘ต Senior jobs ๐Ÿค“ Developer jobs ๐Ÿ’ฐ Finance jobs โ™พ๏ธ Sys Admin jobs โ˜•๏ธ JavaScript jobs ๐Ÿ‘ Backend jobs
Companies
๐Ÿšจ Post a remote job ๐Ÿ“ฆ Buy a job bundle ๐Ÿท Ask for a discount Safetywing Health insurance for teams Safetywing Health insurance for nomads
Feeds
๐Ÿ›  Remote Jobs API ๐Ÿชš  RSS feed ๐Ÿช“  JSON feed

Hacker News mode  Hacker News mode

Safe for work mode  Safe for work mode

Help
๐Ÿ’ก  Ideas + bugs ๐Ÿš€  Changelog ๐Ÿ›๏ธ  Merch ๐Ÿ›Ÿ  FAQ & Help
Other projects
๐Ÿ“Š Remote work stats new ๐Ÿ‘ท Top remote companies ๐Ÿ’ฐ Highest paying remote jobs ๐Ÿงช State of remote work new
๐ŸŒ  Become a digital nomad
๐Ÿ”ฎ  Web3 Jobs
๐Ÿ“ธ  Photo AI
๐Ÿก  Interior AI
Post a remote job โ†’ Log in

๐Ÿ‘‰ Hiring for a Remote Apache + Cloud + Engineer + Security + Technical position?

Post a job
on the ๐Ÿ† #1 Remote Jobs board.
Minimum
$0k/year
๐Ÿ’ฐ 401(k)
๐ŸŒŽ Distributed team
โฐ Async
๐Ÿค“ Vision insurance
๐Ÿฆท Dental insurance
๐Ÿš‘ Medical insurance
๐Ÿ– Unlimited vacation
๐Ÿ– Paid time off
๐Ÿ“† 4 day workweek
๐Ÿ’ฐ 401k matching
๐Ÿ” Company retreats
๐Ÿฌ Coworking budget
๐Ÿ“š Learning budget
๐Ÿ’ช Free gym membership
๐Ÿง˜ Mental wellness budget
๐Ÿ–ฅ Home office budget
๐Ÿฅง Pay in crypto
๐Ÿฅธ Pseudonymous
๐Ÿ’ฐ Profit sharing
๐Ÿ’ฐ Equity compensation
โฌœ๏ธ No whiteboard interview
๐Ÿ‘€ No monitoring system
๐Ÿšซ No politics at work
๐ŸŽ… We hire old (and young)
Regions
๐ŸŒ Worldwide
โ›ฐ๏ธ North America
๐Ÿ’ƒ Latin America
๐Ÿ‡ช๐Ÿ‡บ Europe
๐Ÿฆ Africa
๐Ÿ•Œ Middle East
โ›ฉ Asia
๐ŸŒŠ Oceania
Countries
๐Ÿ‡บ๐Ÿ‡ธ United States
๐Ÿ‡จ๐Ÿ‡ฆ Canada
๐Ÿ‡ฌ๐Ÿ‡ง United Kingdom
๐Ÿ‡ฆ๐Ÿ‡บ Australia
๐Ÿ‡ณ๐Ÿ‡ฟ New Zealand
๐Ÿ‡ฎ๐Ÿ‡ณ India
๐Ÿ‡ต๐Ÿ‡น Portugal
๐Ÿ‡ฉ๐Ÿ‡ช Germany
๐Ÿ‡ณ๐Ÿ‡ฑ Netherlands
๐Ÿ‡ธ๐Ÿ‡ฌ Singapore
๐Ÿ‡ซ๐Ÿ‡ท France
๐Ÿ‡ญ๐Ÿ‡ฐ Hong Kong
๐Ÿ‡ง๐Ÿ‡ท Brazil
๐Ÿ‡ฌ๐Ÿ‡ท Greece
๐Ÿ‡ฆ๐Ÿ‡ช United Arab Emirates
๐Ÿ‡ธ๐Ÿ‡ช Sweden
๐Ÿ‡ต๐Ÿ‡ฑ Poland
๐Ÿ‡ช๐Ÿ‡ธ Spain
๐Ÿ‡ฒ๐Ÿ‡ฝ Mexico
๐Ÿ‡บ๐Ÿ‡ฆ Ukraine
๐Ÿ‡ฏ๐Ÿ‡ต Japan
๐Ÿ‡น๐Ÿ‡ญ Thailand
๐Ÿ‡จ๐Ÿ‡ฟ Czechia
๐Ÿ‡ท๐Ÿ‡บ Russia
๐Ÿ‡ฎ๐Ÿ‡ฑ Israel
๐Ÿ‡ซ๐Ÿ‡ฎ Finland
๐Ÿ‡จ๐Ÿ‡ณ China
๐Ÿ‡ฎ๐Ÿ‡ฉ Indonesia
๐Ÿ‡ฆ๐Ÿ‡ซ Afghanistan
๐Ÿ‡ฆ๐Ÿ‡ฑ Albania
๐Ÿ‡ฉ๐Ÿ‡ฟ Algeria
๐Ÿ‡ฆ๐Ÿ‡ธ American Samoa
๐Ÿ‡ฆ๐Ÿ‡ฉ Andorra
๐Ÿ‡ฆ๐Ÿ‡ด Angola
๐Ÿ‡ฆ๐Ÿ‡ฎ Anguilla
๐Ÿ‡ฆ๐Ÿ‡ถ Antarctica
๐Ÿ‡ฆ๐Ÿ‡ฌ Antigua and Barbuda
๐Ÿ‡ฆ๐Ÿ‡ท Argentina
๐Ÿ‡ฆ๐Ÿ‡ฒ Armenia
๐Ÿ‡ฆ๐Ÿ‡ผ Aruba
๐Ÿ‡ฆ๐Ÿ‡น Austria
๐Ÿ‡ฆ๐Ÿ‡ฟ Azerbaijan
๐Ÿ‡ง๐Ÿ‡ธ The Bahamas
๐Ÿ‡ง๐Ÿ‡ญ Bahrain
๐Ÿ‡ง๐Ÿ‡ฉ Bangladesh
๐Ÿ‡ง๐Ÿ‡ง Barbados
๐Ÿ‡ง๐Ÿ‡พ Belarus
๐Ÿ‡ง๐Ÿ‡ช Belgium
๐Ÿ‡ง๐Ÿ‡ฟ Belize
๐Ÿ‡ง๐Ÿ‡ฏ Benin
๐Ÿ‡ง๐Ÿ‡ฒ Bermuda
๐Ÿ‡ง๐Ÿ‡น Bhutan
๐Ÿ‡ง๐Ÿ‡ด Bolivia
๐Ÿ‡ง๐Ÿ‡ฆ Bosnia
๐Ÿ‡ง๐Ÿ‡ผ Botswana
๐Ÿ‡ง๐Ÿ‡ป Bouvet Island
๐Ÿ‡ฎ๐Ÿ‡ด British Indian Ocean Territory
๐Ÿ‡ง๐Ÿ‡ณ Brunei
๐Ÿ‡ง๐Ÿ‡ฌ Bulgaria
๐Ÿ‡ง๐Ÿ‡ซ Burkina Faso
๐Ÿ‡ง๐Ÿ‡ฎ Burundi
๐Ÿ‡ฐ๐Ÿ‡ญ Cambodia
๐Ÿ‡จ๐Ÿ‡ฒ Cameroon
๐Ÿ‡จ๐Ÿ‡ป Cape Verde
๐Ÿ‡ฐ๐Ÿ‡พ Cayman Islands
๐Ÿ‡จ๐Ÿ‡ซ Central African Republic
๐Ÿ‡น๐Ÿ‡ฉ Chad
๐Ÿ‡จ๐Ÿ‡ฑ Chile
๐Ÿ‡จ๐Ÿ‡ฝ Christmas Island
๐Ÿ‡จ๐Ÿ‡จ Cocos Islands
๐Ÿ‡จ๐Ÿ‡ด Colombia
๐Ÿ‡ฐ๐Ÿ‡ฒ Comoros
๐Ÿ‡จ๐Ÿ‡ฌ Congo
๐Ÿ‡จ๐Ÿ‡ฉ DR Congo
๐Ÿ‡จ๐Ÿ‡ฐ Cook Islands
๐Ÿ‡จ๐Ÿ‡ท Costa Rica
๐Ÿ‡ญ๐Ÿ‡ท Croatia
๐Ÿ‡จ๐Ÿ‡บ Cuba
๐Ÿ‡จ๐Ÿ‡ผ Curaรงao
๐Ÿ‡จ๐Ÿ‡พ Cyprus
๐Ÿ‡ฉ๐Ÿ‡ฐ Denmark
๐Ÿ‡ฉ๐Ÿ‡ฏ Djibouti
๐Ÿ‡ฉ๐Ÿ‡ฒ Dominica
๐Ÿ‡ฉ๐Ÿ‡ด Dominican Republic
๐Ÿ‡ช๐Ÿ‡จ Ecuador
๐Ÿ‡ช๐Ÿ‡ฌ Egypt
๐Ÿ‡ธ๐Ÿ‡ป El Salvador
๐Ÿ‡ฌ๐Ÿ‡ถ Equatorial Guinea
๐Ÿ‡ช๐Ÿ‡ท Eritrea
๐Ÿ‡ช๐Ÿ‡ช Estonia
๐Ÿ‡ช๐Ÿ‡น Ethiopia
๐Ÿ‡ซ๐Ÿ‡ฐ Falkland Islands
๐Ÿ‡ซ๐Ÿ‡ด Faroe Islands
๐Ÿ‡ซ๐Ÿ‡ฏ Fiji
๐Ÿ‡ฌ๐Ÿ‡ซ French Guiana
๐Ÿ‡น๐Ÿ‡ฑ East Timor
๐Ÿ‡น๐Ÿ‡ซ French Southern Territories
๐Ÿ‡ฌ๐Ÿ‡ฆ Gabon
๐Ÿ‡ฌ๐Ÿ‡ฒ Gambia
๐Ÿ‡ฌ๐Ÿ‡ช Georgia
๐Ÿ‡ฌ๐Ÿ‡ญ Ghana
๐Ÿ‡ฌ๐Ÿ‡ฎ Gibraltar
๐Ÿ‡ฌ๐Ÿ‡ฑ Greenland
๐Ÿ‡ฌ๐Ÿ‡ฉ Grenada
๐Ÿ‡ฌ๐Ÿ‡ต Guadeloupe
๐Ÿ‡ฌ๐Ÿ‡บ Guam
๐Ÿ‡ฌ๐Ÿ‡น Guatemala
๐Ÿ‡ฌ๐Ÿ‡ฌ Guernsey
๐Ÿ‡ฌ๐Ÿ‡ณ Guinea
๐Ÿ‡ฌ๐Ÿ‡ผ Guinea Bissau
๐Ÿ‡ฌ๐Ÿ‡พ Guyana
๐Ÿ‡ญ๐Ÿ‡น Haiti
๐Ÿ‡ญ๐Ÿ‡ฒ Heard Island and McDonald Islands
๐Ÿ‡ญ๐Ÿ‡ณ Honduras
๐Ÿ‡ญ๐Ÿ‡บ Hungary
๐Ÿ‡ฎ๐Ÿ‡ธ Iceland
๐Ÿ‡ฎ๐Ÿ‡ท Iran
๐Ÿ‡ฎ๐Ÿ‡ถ Iraq
๐Ÿ‡ฎ๐Ÿ‡ช Ireland
๐Ÿ‡ฎ๐Ÿ‡ฒ Isle of Man
๐Ÿ‡ฎ๐Ÿ‡น Italy
๐Ÿ‡จ๐Ÿ‡ฎ Cote d'Ivoire
๐Ÿ‡ฏ๐Ÿ‡ฒ Jamaica
๐Ÿ‡ฏ๐Ÿ‡ช Jersey
๐Ÿ‡ฏ๐Ÿ‡ด Jordan
๐Ÿ‡ฝ๐Ÿ‡ฐ Kosovo
๐Ÿ‡ฝ๐Ÿ‡ฐ Kosovo
๐Ÿ‡ฐ๐Ÿ‡ฟ Kazakhstan
๐Ÿ‡ฐ๐Ÿ‡ช Kenya
๐Ÿ‡ฐ๐Ÿ‡ฎ Kiribati
๐Ÿ‡ฐ๐Ÿ‡ต North Korea
๐Ÿ‡ฐ๐Ÿ‡ท South Korea
๐Ÿด Kurdistan
๐Ÿ‡ฐ๐Ÿ‡ผ Kuwait
๐Ÿ‡ฐ๐Ÿ‡ฌ Kyrgyzstan
๐Ÿ‡ฑ๐Ÿ‡ฆ Laos
๐Ÿ‡ฑ๐Ÿ‡ป Latvia
๐Ÿ‡ฑ๐Ÿ‡ง Lebanon
๐Ÿ‡ฑ๐Ÿ‡ธ Lesotho
๐Ÿ‡ฑ๐Ÿ‡ท Liberia
๐Ÿ‡ฑ๐Ÿ‡พ Libya
๐Ÿ‡ฑ๐Ÿ‡ฎ Liechtenstein
๐Ÿ‡ฑ๐Ÿ‡น Lithuania
๐Ÿ‡ฑ๐Ÿ‡บ Luxembourg
๐Ÿ‡ฒ๐Ÿ‡ด Macau
๐Ÿ‡ฒ๐Ÿ‡ฐ North Macedonia
๐Ÿ‡ฒ๐Ÿ‡ฌ Madagascar
๐Ÿ‡ฒ๐Ÿ‡ผ Malawi
๐Ÿ‡ฒ๐Ÿ‡พ Malaysia
๐Ÿ‡ฒ๐Ÿ‡ป Maldives
๐Ÿ‡ฒ๐Ÿ‡ฑ Mali
๐Ÿ‡ฒ๐Ÿ‡น Malta
๐Ÿ‡ฒ๐Ÿ‡ญ Marshall Islands
๐Ÿ‡ฒ๐Ÿ‡ถ Martinique
๐Ÿ‡ฒ๐Ÿ‡ท Mauritania
๐Ÿ‡ฒ๐Ÿ‡บ Mauritius
๐Ÿ‡พ๐Ÿ‡น Mayotte
๐Ÿ‡ซ๐Ÿ‡ฒ Micronesia
๐Ÿ‡ฒ๐Ÿ‡ฉ Moldova
๐Ÿ‡ฒ๐Ÿ‡จ Monaco
๐Ÿ‡ฒ๐Ÿ‡ณ Mongolia
๐Ÿ‡ฒ๐Ÿ‡ช Montenegro
๐Ÿ‡ฒ๐Ÿ‡ธ Montserrat
๐Ÿ‡ฒ๐Ÿ‡ฆ Morocco
๐Ÿ‡ฒ๐Ÿ‡ฟ Mozambique
๐Ÿ‡ฒ๐Ÿ‡ฒ Myanmar
๐Ÿ‡ณ๐Ÿ‡ฆ Namibia
๐Ÿ‡ณ๐Ÿ‡ท Nauru
๐Ÿ‡ณ๐Ÿ‡ต Nepal
๐Ÿ‡ง๐Ÿ‡ถ Caribbean Netherlands
๐Ÿ‡ณ๐Ÿ‡จ New Caledonia
๐Ÿ‡ณ๐Ÿ‡ฎ Nicaragua
๐Ÿ‡ณ๐Ÿ‡ช Niger
๐Ÿ‡ณ๐Ÿ‡ฌ Nigeria
๐Ÿ‡ณ๐Ÿ‡บ Niue
๐Ÿ‡ณ๐Ÿ‡ซ Norfolk Island
๐Ÿ‡ฒ๐Ÿ‡ต Northern Mariana Islands
๐Ÿ‡ณ๐Ÿ‡ด Norway
๐Ÿ‡ด๐Ÿ‡ฒ Oman
๐Ÿ‡ต๐Ÿ‡ธ Palestine
๐Ÿ‡ต๐Ÿ‡ฐ Pakistan
๐Ÿ‡ต๐Ÿ‡ผ Palau
๐Ÿ‡ต๐Ÿ‡ฆ Panama
๐Ÿ‡ต๐Ÿ‡ฌ Papua New Guinea
๐Ÿ‡ต๐Ÿ‡พ Paraguay
๐Ÿ‡ต๐Ÿ‡ช Peru
๐Ÿ‡ต๐Ÿ‡ญ Philippines
๐Ÿ‡ต๐Ÿ‡ณ Pitcairn Island
๐Ÿ‡ต๐Ÿ‡ซ Polynesia
๐Ÿ‡ต๐Ÿ‡ท Puerto Rico
๐Ÿ‡ถ๐Ÿ‡ฆ Qatar
๐Ÿ‡ท๐Ÿ‡ช Reunion
๐Ÿ‡ท๐Ÿ‡ด Romania
๐Ÿ‡ท๐Ÿ‡ผ Rwanda
๐Ÿ‡ธ๐Ÿ‡ญ Saint Helena
๐Ÿ‡ฐ๐Ÿ‡ณ Saint Kitts and Nevis
๐Ÿ‡ฑ๐Ÿ‡จ Saint Lucia
๐Ÿ‡ต๐Ÿ‡ฒ Saint Pierre and Miquelon
๐Ÿ‡ป๐Ÿ‡จ Saint Vincent and the Grenadines
๐Ÿ‡ผ๐Ÿ‡ธ Samoa
๐Ÿ‡ธ๐Ÿ‡ฒ San Marino
๐Ÿ‡ธ๐Ÿ‡น Sao Tome and Principe
๐Ÿ‡ธ๐Ÿ‡ฆ Saudi Arabia
๐Ÿ‡ธ๐Ÿ‡ณ Senegal
๐Ÿ‡ท๐Ÿ‡ธ Serbia
๐Ÿ‡ธ๐Ÿ‡จ Seychelles
๐Ÿ‡ธ๐Ÿ‡ฑ Sierra Leone
๐Ÿ‡ฒ๐Ÿ‡ซ Saint-Martin
๐Ÿ‡ธ๐Ÿ‡ฝ Sint Maarten
๐Ÿ‡ธ๐Ÿ‡ฐ Slovakia
๐Ÿ‡ธ๐Ÿ‡ฎ Slovenia
๐Ÿ‡ธ๐Ÿ‡ง Solomon Islands
๐Ÿ‡ธ๐Ÿ‡ด Somalia
๐Ÿ‡ฟ๐Ÿ‡ฆ South Africa
๐Ÿ‡ฌ๐Ÿ‡ธ South Georgia and the South Sandwich Islands
๐Ÿ‡ธ๐Ÿ‡ธ South Sudan
๐Ÿ‡ฑ๐Ÿ‡ฐ Sri Lanka
๐Ÿ‡ธ๐Ÿ‡ฉ Sudan
๐Ÿ‡ธ๐Ÿ‡ท Suriname
๐Ÿ‡ธ๐Ÿ‡ฏ Svalbard and Jan Mayen Islands
๐Ÿ‡ธ๐Ÿ‡ฟ Swaziland
๐Ÿ‡จ๐Ÿ‡ญ Switzerland
๐Ÿ‡ธ๐Ÿ‡พ Syria
๐Ÿ‡น๐Ÿ‡ผ Taiwan
๐Ÿ‡น๐Ÿ‡ฏ Tajikistan
๐Ÿ‡น๐Ÿ‡ฟ Tanzania
๐Ÿ‡น๐Ÿ‡ฌ Togo
๐Ÿ‡น๐Ÿ‡ฐ Tokelau
๐Ÿ‡น๐Ÿ‡ด Tonga
๐Ÿ‡น๐Ÿ‡น Trinidad and Tobago
๐Ÿ‡น๐Ÿ‡ณ Tunisia
๐Ÿ‡น๐Ÿ‡ท Turkey
๐Ÿ‡น๐Ÿ‡ฒ Turkmenistan
๐Ÿ‡น๐Ÿ‡จ Turks and Caicos Islands
๐Ÿ‡น๐Ÿ‡ป Tuvalu
๐Ÿ‡บ๐Ÿ‡ฌ Uganda
๐Ÿ‡บ๐Ÿ‡พ Uruguay
๐Ÿ Hawaii
๐Ÿ‡บ๐Ÿ‡ฒ USA Minor Outlying Islands
๐Ÿ‡บ๐Ÿ‡ฟ Uzbekistan
๐Ÿ‡ป๐Ÿ‡บ Vanuatu
๐Ÿ‡ป๐Ÿ‡ฆ Vatican City
๐Ÿ‡ป๐Ÿ‡ช Venezuela
๐Ÿ‡ป๐Ÿ‡ณ Vietnam
๐Ÿ‡ป๐Ÿ‡ฌ British Virgin Islands
๐Ÿ‡ป๐Ÿ‡ฎ United States Virgin Islands
๐Ÿ‡ผ๐Ÿ‡ซ Wallis and Futuna Islands
๐Ÿ‡ช๐Ÿ‡ญ Western Sahara
๐Ÿ‡พ๐Ÿ‡ช Yemen
๐Ÿ‡ฟ๐Ÿ‡ฒ Zambia
๐Ÿ‡ฟ๐Ÿ‡ผ Zimbabwe
Apply for this job
๐Ÿ’ต Salary
๐ŸŽช Benefits
๐Ÿš Apache Remove this filter
โ˜๏ธ Cloud Remove this filter
๐Ÿค“ Engineer Remove this filter
๐Ÿ”‘ Security Remove this filter
๐Ÿ›  Technical Remove this filter
โŒ Clear 9 results
Nomad Insurance by SafetyWing
Global health coverage for remote workers and nomads

Ad


Hawk

 This job is getting a high amount of applications right now (11% of viewers clicked Apply)

๐ŸŒ Probably worldwide
๐Ÿ’ฐ $55k - $90k*

Support

Software

Code

DevOps

Financial

Java

Reliability

Engineering

Hawk

Apply now

๐Ÿ‘€ 6,606 views

โœ… 744 applied (11%)

Share this job:
Get a rok.co short link

Hawk is hiring a

Remote DevOps Engineer

About Usย Hawk is the leading provider of AI-supported anti-money laundering and fraud detection technology. Banks and payment providers globally are using Hawkโ€™s powerful combination of traditional rules and explainable AI to improve the effectiveness of their AML compliance and fraud prevention by identifying more crime while maximizing efficiency by reducing false positives. With our solution, we are playing a vital role in the global fight against Money Laundering, Fraud, or the financing of terrorism. We offer a culture of mutual trust, support and passion โ€“ while providing individuals with opportunities to grow professionally and make a difference in the world.ย Your MissionAs a DevOps Engineer, you will play a crucial role in ensuring the scalability, security, and reliability of our AI-driven financial crime prevention platform. You will automate cloud infrastructure, implement monitoring and observability solutions, and build secure, scalable CI/CD pipelines. Your work will directly contribute to maintaining high availability for a platform that fights financial crime 24/7. This role is based on the East Coast, U.S. and requires expertise in cloud infrastructure, automation, security best practices, and continuous integration/deployment (CI/CD).Your Responsibilities* Provision, manage, and scale multi-cloud environments using Infrastructure as Code (IaC) (e.g., Terraform).\n* Maintain high availability (HA), fault tolerance, and least-privilege security practices, while optimizing cloud costs.\n* Design and maintain developer-friendly CI/CD workflows, container templates, and reusable artifacts for seamless software delivery.\n* Implement real-time monitoring, alerting, and observability solutions (e.g., Elastic Stack, Prometheus, Grafana, CloudWatch) to proactively detect and resolve issues.\n* Implement and enforce cloud security best practices, identify and mitigate vulnerabilities, and ensure compliance with data protection regulations.\n* Provide technical guidance to clients running Hawkโ€™s platform in their own VPC environments, supporting onboarding and integration.\n* Develop structured documentation for cloud architectures, best practices, and deployment processes, ensuring seamless team collaboration.\n\nYour Profile* 5+ years of experience in DevOps, Site Reliability Engineering (SRE), or Cloud Engineering roles.\n* Bachelorโ€™s degree in Computer Science, Engineering, or a related field (or equivalent experience).\n* Strong expertise in Kubernetes, containerized applications, and cloud-native technologies.\n* Hands-on experience with AWS or GCP and their core services.\n* Proficiency with Terraform and Infrastructure as Code (IaC) methodologies.\n* Experience with CI/CD tools such as GitLab CI, GitHub Actions, or similar.\n* Strong knowledge of observability and monitoring tools (e.g., Elastic Stack, Prometheus, Grafana, CloudWatch).\n* Solid understanding of cloud security principles, least-privilege access, and automated security policies.\n* Ability to diagnose complex technical challenges and provide scalable, secure solutions.\n* Strong communication and collaboration skills; able to work effectively in a remote, cross-functional environment.\n* Comfortable in a fast-paced, hands-on role, with a willingness to get your hands dirty and embrace feedback for continuous improvement.\n\nPreferred Qualifications* Experience in cybersecurity, penetration testing, and cloud compliance.\n* Familiarity with Java Spring Boot & Apache Kafka.\n* Experience in 24/7 uptime environments with on-call rotations.\n* Knowledge of big data systems (PostgreSQL, S3/Azure Blob Storage, Elasticsearch).\n\n \n\n#Salary and compensation\n No salary data published by company so we estimated salary based on similar jobs related to DevOps, Java, Cloud and Engineer jobs that are similar:\n\n $55,000 — $90,000/year\n
\n\n#Benefits\n ๐Ÿ’ฐ 401(k)\n\n๐ŸŒŽ Distributed team\n\nโฐ Async\n\n๐Ÿค“ Vision insurance\n\n๐Ÿฆท Dental insurance\n\n๐Ÿš‘ Medical insurance\n\n๐Ÿ– Unlimited vacation\n\n๐Ÿ– Paid time off\n\n๐Ÿ“† 4 day workweek\n\n๐Ÿ’ฐ 401k matching\n\n๐Ÿ” Company retreats\n\n๐Ÿฌ Coworking budget\n\n๐Ÿ“š Learning budget\n\n๐Ÿ’ช Free gym membership\n\n๐Ÿง˜ Mental wellness budget\n\n๐Ÿ–ฅ Home office budget\n\n๐Ÿฅง Pay in crypto\n\n๐Ÿฅธ Pseudonymous\n\n๐Ÿ’ฐ Profit sharing\n\n๐Ÿ’ฐ Equity compensation\n\nโฌœ๏ธ No whiteboard interview\n\n๐Ÿ‘€ No monitoring system\n\n๐Ÿšซ No politics at work\n\n๐ŸŽ… We hire old (and young)\n\n
Apply for this job

๐Ÿ‘‰ Please reference you found the job on Remote OK, this helps us get more companies to post here, thanks!

When applying for jobs, you should NEVER have to pay to apply. You should also NEVER have to pay to buy equipment which they then pay you back for later. Also never pay for trainings you have to do. Those are scams! NEVER PAY FOR ANYTHING! Posts that link to pages with "how to work online" are also scams. Don't use them or pay for them. Also always verify you're actually talking to the company in the job post and not an imposter. A good idea is to check the domain name for the site/email and see if it's the actual company's main domain name. Scams in remote work are rampant, be careful! Read more to avoid scams. When clicking on the button to apply above, you will leave Remote OK and go to the job application page for that company outside this site. Remote OK accepts no liability or responsibility as a consequence of any reliance upon information on there (external sites) or here.


Government Employees Insurance Company


MD Chevy Chase (Office) - JPS
๐Ÿ’ฐ $48k - $98k*

Design

 

SaaS

 

Python

Students

Training

Full-Stack

Docker

Support

Developer

Software

DevOps

Financial

Edu

NoSQL

API

Leader

Strategy

Lead

Senior

Operational

Excel

Health

Engineering

Government Employees Insurance Company is hiring a

Remote Staff Engineer

Our Senior Staff Engineer works with our Staff and Sr. Engineers to innovate and build new systems, improve and enhance existing systems and identify new opportunities to apply your knowledge to solve critical problems. You will lead the strategy and execution of a technical roadmap that will increase the velocity of delivering products and unlock new engineering capabilities. The ideal candidate is a self-starter that has deep technical expertise in their domain. Position Responsibilities As a Senior Staff Engineer, you will: Provide technical leadership to multiple areas and provide technical and thought leadership to the enterprise Collaborate across team members and across the tech organization to solve our toughest problems Develop and execute technical software development strategy for a variety of domains Accountable for the quality, usability, and performance of the solutions Utilize programming languages like C#, Java, Python or other object-oriented languages, SQL, and NoSQL databases, Container Orchestration services including Docker and Kubernetes, and a variety of Azure tools and services Be a role model and mentor, helping to coach and strengthen the technical expertise and know-how of our engineering and product community. Influence and educate executives Consistently share best practices and improve processes within and across teams Analyze cost and forecast, incorporating them into business plans Determine and support resource requirements, evaluate operational processes, measure outcomes to ensure desired results, and demonstrate adaptability and sponsoring continuous learning Qualifications Exemplary ability to design, perform experiments, and influence engineering direction and product roadmap Experience partnering with engineering teams and transferring research to production Extensive experience in leading and building full-stack application and service development, with a strong focus on SAAS products / platforms. Proven expertise in designing and developing microservices using C#, gRPC, Python, Django, Kafka, and Apache Spark, with a deep understanding of both API and event-driven architectures. Proven experience designing and delivering highly-resilient event-driven and messaging based solutions at scale with minimal latency. Deep hands-on experience in building complex SAAS systems in large scale business focused systems, with great knowledge on Docker and Kubernetes Fluency and Specialization with at least two modern OOP languages such as C#, Java, C++, or Python including object-oriented design Great understanding of open-source databases like MySQL, PostgreSQL, etc. And strong foundation with No-SQL databases like Cosmos, Cassandra. Apache Trino etc. In-depth knowledge of CS data structures and algorithms Ability to excel in a fast-paced, startup-like environment Knowledge of developer tooling across the software development life cycle (task management, source code, building, deployment, operations, real-time communication) Experience with Micro-services oriented architecture and extensible REST APIs Experience building the architecture and design (architecture, design patterns, reliability, and scaling) of new and current systems Experience in implementing security protocols across services and products: Understanding of Active Directory, Windows Authentication, SAML, OAuth Fluency in DevOps Concepts, Cloud Architecture, and Azure DevOps Operational Framework Experience in leveraging PowerShell scripting Experience in existing Operational Portals such as Azure Portal Experience with application monitoring tools and performance assessments Experience in Azure Network (Subscription, Security zoning, etc.) Experience 10+ years full-stack development experience (C#/Java/Python/GO), with expertise in client-side and server-side frameworks. 8+ years of experience with architecture and design 6+ years of experience in open-source frameworks 4+ years of experience with AWS, GCP, Azure, or another cloud service Education Bachelorโ€™s degree in Computer Science, Information Systems, or equivalent education or work experience Annual Salary $115,000.00 - $260,000.00 The above annual salary range is a general guideline. Multiple factors are taken into consideration to arrive at the final hourly rate/ annual salary to be offered to the selected candidate. Factors include, but are not limited to, the scope and responsibilities of the role, the selected candidateโ€™s work experience, education and training, the work location as well as market and business considerations. At this time, GEICO will not sponsor a new applicant for employment authorization for this position. Benefits: As an Associate, youโ€™ll enjoy our Total Rewards Program* to help secure your financial future and preserve your health and well-being, including: Premier Medical, Dental and Vision Insurance with no waiting period** Paid Vacation, Sick and Parental Leave 401(k) Plan Tuition Reimbursement Paid Training and Licensures *Benefits may be different by location. Benefit eligibility requirements vary and may include length of service. **Coverage begins on the date of hire. Must enroll in New Hire Benefits within 30 days of the date of hire for coverage to take effect. The equal employment opportunity policy of the GEICO Companies provides for a fair and equal employment opportunity for all associates and job applicants regardless of race, color, religious creed, national origin, ancestry, age, gender, pregnancy, sexual orientation, gender identity, marital status, familial status, disability or genetic information, in compliance with applicable federal, state and local law. GEICO hires and promotes individuals solely on the basis of their qualifications for the job to be filled. GEICO reasonably accommodates qualified individuals with disabilities to enable them to receive equal employment opportunity and/or perform the essential functions of the job, unless the accommodation would impose an undue hardship to the Company. This applies to all applicants and associates. GEICO also provides a work environment in which each associate is able to be productive and work to the best of their ability. We do not condone or tolerate an atmosphere of intimidation or harassment. We expect and require the cooperation of all associates in maintaining an atmosphere free from discrimination and harassment with mutual respect by and for all associates and applicants. For more than 75 years, GEICO has stood out from the rest of the insurance industry! We are one of the nation's largest and fastest-growing auto insurers thanks to our low rates, outstanding service and clever marketing. We're an industry leader employing thousands of dedicated and hard-working associates. As a wholly owned subsidiary of Berkshire Hathaway, we offer associates training and career advancement in a financially stable and rewarding workplace. Opportunities for Students & Grads Learn more about GEICO Learn more about GEICO Diversity and Inclusion Learn more about GEICO Benefits \n\n#Salary and compensation\n No salary data published by company so we estimated salary based on similar jobs related to Design, SaaS, Python, Docker, DevOps, Education, Cloud, API, Senior and Engineer jobs that are similar:\n\n $47,500 — $97,500/year\n
\n\n#Benefits\n ๐Ÿ’ฐ 401(k)\n\n๐ŸŒŽ Distributed team\n\nโฐ Async\n\n๐Ÿค“ Vision insurance\n\n๐Ÿฆท Dental insurance\n\n๐Ÿš‘ Medical insurance\n\n๐Ÿ– Unlimited vacation\n\n๐Ÿ– Paid time off\n\n๐Ÿ“† 4 day workweek\n\n๐Ÿ’ฐ 401k matching\n\n๐Ÿ” Company retreats\n\n๐Ÿฌ Coworking budget\n\n๐Ÿ“š Learning budget\n\n๐Ÿ’ช Free gym membership\n\n๐Ÿง˜ Mental wellness budget\n\n๐Ÿ–ฅ Home office budget\n\n๐Ÿฅง Pay in crypto\n\n๐Ÿฅธ Pseudonymous\n\n๐Ÿ’ฐ Profit sharing\n\n๐Ÿ’ฐ Equity compensation\n\nโฌœ๏ธ No whiteboard interview\n\n๐Ÿ‘€ No monitoring system\n\n๐Ÿšซ No politics at work\n\n๐ŸŽ… We hire old (and young)\n\n
\n\n#Location\nMD Chevy Chase (Office) - JPS
Apply for this job

๐Ÿ‘‰ Please reference you found the job on Remote OK, this helps us get more companies to post here, thanks!

When applying for jobs, you should NEVER have to pay to apply. You should also NEVER have to pay to buy equipment which they then pay you back for later. Also never pay for trainings you have to do. Those are scams! NEVER PAY FOR ANYTHING! Posts that link to pages with "how to work online" are also scams. Don't use them or pay for them. Also always verify you're actually talking to the company in the job post and not an imposter. A good idea is to check the domain name for the site/email and see if it's the actual company's main domain name. Scams in remote work are rampant, be careful! Read more to avoid scams. When clicking on the button to apply above, you will leave Remote OK and go to the job application page for that company outside this site. Remote OK accepts no liability or responsibility as a consequence of any reliance upon information on there (external sites) or here.


SADA India


๐Ÿ’ฐ $60k - $100k*

Cryptocurrency

 

Teach

Support

Software

Growth

Management

Senior

Operational

Engineering

SADA India is hiring a

Remote Senior Data Engineer

\nJoin SADA India as a Senior Data Engineer, Enterprise Support service!\n\nYour Mission \n\nAs a Sr. Data Engineer on the Enterprise Support service team at SADA, you will reduce customer anxiety about running production workloads in the cloud by implementing and iteratively improving observability and reliability. You will have the opportunity to engage with our customers in a meaningful way by defining, measuring, and improving key business metrics; eliminating toil through automation; inspecting code, design, implementation, and operational procedures; enabling experimentation by helping create a culture of ownership; and winning customer trust through education, skill sharing, and implementing recommendations. Your efforts will accelerate our customersโ€™ cloud adoption journey and we will be with them through the transformation of their applications, infrastructure, and internal processes. You will be part of a new social contract between customers and service providers that demands shared responsibility and accountability: our partnership with our customers will ensure we are working towards a common goal and share a common fate.\n\nThis is primarily a customer-facing role. You will also work closely with SADAโ€™s Customer Experience team to execute their recommendations to our customers, and with Professional Services on large projects that require PMO support.\n\nPathway to Success \n\n#MakeThemRave is at the foundation of all our engineering. Our motivation is to provide customers with an exceptional experience in migrating, developing, modernizing, and operationalizing their systems in the Google Cloud Platform.\n\nYour success starts by positively impacting the direction of a fast-growing practice with vision and passion. You will be measured bi-yearly by the breadth, magnitude, and quality of your contributions, your ability to estimate accurately, customer feedback at the close of projects, how well you collaborate with your peers, and the consultative polish you bring to customer interactions.\n\nAs you continue to execute successfully, we will build a customized development plan together that leads you through the engineering or management growth tracks.\n\nExpectations\n\nCustomer Facing - You will interact with customers on a regular basis, sometimes daily, other times weekly/bi-weekly. Common touchpoints occur when qualifying potential opportunities, at project kickoff, throughout the engagement as progress is communicated, and at project close. You can expect to interact with a range of customer stakeholders, including engineers, technical project managers, and executives.\n\nOnboarding/Training - The first several weeks of onboarding are dedicated to learning and will encompass learning materials/assignments and compliance training, as well as meetings with relevant individuals.\n\nJob Requirements\n\nRequired Credentials:\n\n\n* Google Professional Data Engineer Certified or able to complete within the first 45 days of employment \n\n* A secondary Google Cloud certification in any other specialization\n\n\n\n\nRequired Qualifications: \n\n\n* 4+ years of experience in Cloud support\n\n* Experience in supporting customers preferably in 24/7 environments\n\n* Experience working with Google Cloud data products (CloudSQL, Spanner, Cloud Storage, Pub/Sub, Dataflow, Dataproc, Bigtable, BigQuery, Dataprep, Composer, etc)\n\n* Experience writing software in at least two or more languages such as Python, Java, Scala, or Go\n\n* Experience in building production-grade data solutions (relational and NoSQL)\n\n* Experience with systems monitoring/alerting, capacity planning, and performance tuning\n\n* Experience with BI tools like Tableau, Looker, etc will be an advantage\n\n* Consultative mindset that delights the customer by building good rapport with them to fully understand their requirements and provide accurate solutions\n\n\n\n\nUseful Qualifications:\n\n\n* \n\n\n* Mastery in at least one of the following domain areas:\n\n\n\n\n\n* \n\n\n* Google Cloud DataFlow: building batch/streaming ETL pipelines with frameworks such as Apache Beam or Google Cloud DataFlow and working with messaging systems like Pub/Sub, Kafka, and RabbitMQ; Auto scaling DataFlow clusters, troubleshooting cluster operation issues\n\n* Data Integration Tools: building data pipelines using modern data integration tools such as Fivetran, Striim, Data Fusion, etc. Must have hands-on experience configuring and integrating with multiple Data Sources within and outside of Google Cloud\n\n* Large Enterprise Migration: migrating entire cloud or on-prem assets to Google Cloud including Data Lakes, Data Warehouses, Databases, Business Intelligence, Jobs, etc. Provide consultations for optimizing cost, defining methodology, and coming up with a plan to execute the migration.\n\n\n\n\n\n\n\n\n\n* Experience with IoT architectures and building real-time data streaming pipelines\n\n* Experience operationalizing machine learning models on large datasets\n\n* Demonstrated leadership and self-direction -- a willingness to teach others and learn new techniques\n\n* Demonstrated skills in selecting the right statistical tools given a data analysis problem\n\n* Understanding of Chaos Engineering\n\n* Understanding of PCI, SOC2, and HIPAA compliance standards\n\n* Understanding of the principle of least privilege and security best practices\n\n* Understanding of cryptocurrency and blockchain technology\n\n\n\n\n  \n\n#Salary and compensation\n No salary data published by company so we estimated salary based on similar jobs related to Cloud, Senior and Engineer jobs that are similar:\n\n $60,000 — $100,000/year\n
\n\n#Benefits\n ๐Ÿ’ฐ 401(k)\n\n๐ŸŒŽ Distributed team\n\nโฐ Async\n\n๐Ÿค“ Vision insurance\n\n๐Ÿฆท Dental insurance\n\n๐Ÿš‘ Medical insurance\n\n๐Ÿ– Unlimited vacation\n\n๐Ÿ– Paid time off\n\n๐Ÿ“† 4 day workweek\n\n๐Ÿ’ฐ 401k matching\n\n๐Ÿ” Company retreats\n\n๐Ÿฌ Coworking budget\n\n๐Ÿ“š Learning budget\n\n๐Ÿ’ช Free gym membership\n\n๐Ÿง˜ Mental wellness budget\n\n๐Ÿ–ฅ Home office budget\n\n๐Ÿฅง Pay in crypto\n\n๐Ÿฅธ Pseudonymous\n\n๐Ÿ’ฐ Profit sharing\n\n๐Ÿ’ฐ Equity compensation\n\nโฌœ๏ธ No whiteboard interview\n\n๐Ÿ‘€ No monitoring system\n\n๐Ÿšซ No politics at work\n\n๐ŸŽ… We hire old (and young)\n\n
\n\n#Location\nThiruvananthapuram, Kerala, India
Apply for this job

๐Ÿ‘‰ Please reference you found the job on Remote OK, this helps us get more companies to post here, thanks!

When applying for jobs, you should NEVER have to pay to apply. You should also NEVER have to pay to buy equipment which they then pay you back for later. Also never pay for trainings you have to do. Those are scams! NEVER PAY FOR ANYTHING! Posts that link to pages with "how to work online" are also scams. Don't use them or pay for them. Also always verify you're actually talking to the company in the job post and not an imposter. A good idea is to check the domain name for the site/email and see if it's the actual company's main domain name. Scams in remote work are rampant, be careful! Read more to avoid scams. When clicking on the button to apply above, you will leave Remote OK and go to the job application page for that company outside this site. Remote OK accepts no liability or responsibility as a consequence of any reliance upon information on there (external sites) or here.


Chan Zuckerberg Initiative


๐Ÿ’ฐ $43k - $83k*

Design

 

Amazon

 

Recruiter

Support

Software

Growth

Web

Edu

Microsoft

Management

Senior

Operations

Operational

Engineering

Linux

Chan Zuckerberg Initiative

Apply now

๐Ÿ‘€ 696 views

โœ… 30 applied (4%)

Share this job:
Get a rok.co short link

Chan Zuckerberg Initiative is hiring a

Remote Senior AI Infra Engineer AI ML and Data Infrastructure

The Team\n\nAcross our work in Science, Education, and within our communities, we pair technology with grantmaking, impact investing, and collaboration to help accelerate the pace of progress toward our mission. Our Central team provides the support needed to push this work forward. \n\nThe Central team at CZI consists of our Finance, People & DEI, Real Estate, Events, Workplace, Facilities, Security, Brand & Communications, Business Systems, Central Operations, Strategic Initiatives, and Ventures teams. These teams provide strategic support and operational excellence across the board at CZI.\nThe Opportunity\n\nBy pairing engineers with leaders in our science and education teams, we can bring AI/ML technology to the table in new ways to help drive solutions. We are uniquely positioned to design, build, and scale software systems to help educators, scientists, and policy experts better address the myriad challenges they face. Our technology team is already helping schools bring personalized learning tools to teachers and schools across the country. We are also supporting scientists around the world as they develop a comprehensive reference atlas of all cells in the human body, and are developing the capacity to apply state-of-the-art methods in artificial intelligence and machine learning to solve important problems in the biomedical sciences. \n\nThe AI/ML and Data Engineering Infrastructure organization works on building shared tools and platforms to be used across all of the Chan Zuckerberg Initiative, partnering and supporting the work of a wide range of Research Scientists, Data Scientists, AI Research Scientists, as well as a broad range of Engineers focusing on Education and Science domain problems. Members of the shared infrastructure engineering team have an impact on all of CZI's initiatives by enabling the technology solutions used by other engineering teams at CZI to scale. A person in this role will build these technology solutions and help to cultivate a culture of shared best practices and knowledge around core engineering.\nWhat You'll Do\n\n\n* Participate in the  technical design and building of efficient, stable, performant, scalable and secure AI/ML and Data infrastructure engineering solutions.\n\n* Active hands-on coding working on our Deep Learning and Machine Learning models\n\n* Design and implement complex systems integrating with our large scale AI/ML GPU compute infrastructure and platform, making working across multiple clouds easier and convenient for our Research Engineers, ML Engineers, and Data Scientists. \n\n* Use your solid experience and skills in building containerized applications and infrastructure using Kubernetes in support of our large scale GPU Research cluster as well as working on our various heterogeneous and distributed AI/ML environments.  \n\n* Collaborate with other team members in the design and build of our Cloud based AI/ML platform solutions, which includes Databricks Spark, Weaviate Vector Databases, and supporting our hosted Cloud GPU Compute services running containerized PyTorch on large scale Kubernetes.\n\n* Collaborate with our partners on data management solutions in our heterogeneous collection of complex datasets.\n\n* Help build tooling that makes optimal use of our shared infrastructure in empowering  our AI/ML efforts with world class GPU Compute Cluster and other compute environments such as our AWS based services.\n\n\n\nWhat You'll Bring\n\n\n* BS or MS degree in Computer Science or a related technical discipline or equivalent experience\n\n* 5+ years of relevant coding experience\n\n* 3+ years of systems Architecture and Design experience, with a broad range of experience across Data, AI/ML, Core Infrastructure, and Security Engineering\n\n* Scaling containerized applications  on Kubernetes or Mesos, including expertise  with creating custom containers using secure AMIs and continuous deployment systems that integrate with Kubernetes or Mesos. (Kubernetes preferred)\n\n* Proficiency with Amazon Web Services (AWS), Google Cloud Platform (GCP), or Microsoft Azure, and experience with On-Prem and Colocation Service hosting environments\n\n* Proven coding ability with a systems language such as Rust,C/ C++, C#, Go, Java, or Scala\n\n* Shown ability with a scripting language such as Python, PHP, or Ruby\n\n* AI/ML Platform Operations experience in an environment with challenging data and systems platform challenges - including large scale Kafka and Spark deployments (or their coralaries such as Pulsar, Flink, and/or Ray) as well as Workflow scheduling tools such as Apache Airflow, Dagster, or Apache Beam  \n\n* MLOps experience working with medium to large scale GPU clusters in Kubernetes (Kubeflow),  HPC environments, or large scale Cloud based ML deployments\n\n* Working knowledge of Nvidia CUDA and AI/ML custom libraries.  \n\n* Knowledge of Linux systems optimization and administration\n\n* Understanding of Data Engineering, Data Governance, Data Infrastructure, and AI/ML execution platforms.\n\n* PyTorch, Karas, or Tensorflow  experience a strong nice to have\n\n* HPC with and Slurm experience a strong nice to have\n\n\n\nCompensation\n\nThe Redwood City, CA base pay range for this role is $190,000 - $285,000. New hires are typically hired into the lower portion of the range, enabling employee growth in the range over time. Actual placement in range is based on job-related skills and experience, as evaluated throughout the interview process. Pay ranges outside Redwood City are adjusted based on cost of labor in each respective geographical market. Your recruiter can share more about the specific pay range for your location during the hiring process.\nBenefits for the Whole You \n\nWeโ€™re thankful to have an incredible team behind our work. To honor their commitment, we offer a wide range of benefits to support the people who make all we do possible. \n\n\n* CZI provides a generous employer match on employee 401(k) contributions to support planning for the future.\n\n* Annual benefit for employees that can be used most meaningfully for them and their families, such as housing, student loan repayment, childcare, commuter costs, or other life needs.\n\n* CZI Life of Service Gifts are awarded to employees to โ€œlive the missionโ€ and support the causes closest to them.\n\n* Paid time off to volunteer at an organization of your choice. \n\n* Funding for select family-forming benefits. \n\n* Relocation support for employees who need assistance moving to the Bay Area\n\n* And more!\n\n\n\nCommitment to Diversity\n\nWe believe that the strongest teams and best thinking are defined by the diversity of voices at the table. We are committed to fair treatment and equal access to opportunity for all CZI team members and to maintaining a workplace where everyone feels welcomed, respected, supported, and valued. Learn about our diversity, equity, and inclusion efforts. \n\nIf youโ€™re interested in a role but your previous experience doesnโ€™t perfectly align with each qualification in the job description, we still encourage you to apply as you may be the perfect fit for this or another role.\n\nExplore our work modes, benefits, and interview process at www.chanzuckerberg.com/careers.\n\n#LI-Remote  \n\n#Salary and compensation\n No salary data published by company so we estimated salary based on similar jobs related to Design, Amazon, Recruiter, Cloud, Senior and Engineer jobs that are similar:\n\n $42,500 — $82,500/year\n
\n\n#Benefits\n ๐Ÿ’ฐ 401(k)\n\n๐ŸŒŽ Distributed team\n\nโฐ Async\n\n๐Ÿค“ Vision insurance\n\n๐Ÿฆท Dental insurance\n\n๐Ÿš‘ Medical insurance\n\n๐Ÿ– Unlimited vacation\n\n๐Ÿ– Paid time off\n\n๐Ÿ“† 4 day workweek\n\n๐Ÿ’ฐ 401k matching\n\n๐Ÿ” Company retreats\n\n๐Ÿฌ Coworking budget\n\n๐Ÿ“š Learning budget\n\n๐Ÿ’ช Free gym membership\n\n๐Ÿง˜ Mental wellness budget\n\n๐Ÿ–ฅ Home office budget\n\n๐Ÿฅง Pay in crypto\n\n๐Ÿฅธ Pseudonymous\n\n๐Ÿ’ฐ Profit sharing\n\n๐Ÿ’ฐ Equity compensation\n\nโฌœ๏ธ No whiteboard interview\n\n๐Ÿ‘€ No monitoring system\n\n๐Ÿšซ No politics at work\n\n๐ŸŽ… We hire old (and young)\n\n
\n\n#Location\nRedwood City, California, United States
Apply for this job

๐Ÿ‘‰ Please reference you found the job on Remote OK, this helps us get more companies to post here, thanks!

When applying for jobs, you should NEVER have to pay to apply. You should also NEVER have to pay to buy equipment which they then pay you back for later. Also never pay for trainings you have to do. Those are scams! NEVER PAY FOR ANYTHING! Posts that link to pages with "how to work online" are also scams. Don't use them or pay for them. Also always verify you're actually talking to the company in the job post and not an imposter. A good idea is to check the domain name for the site/email and see if it's the actual company's main domain name. Scams in remote work are rampant, be careful! Read more to avoid scams. When clicking on the button to apply above, you will leave Remote OK and go to the job application page for that company outside this site. Remote OK accepts no liability or responsibility as a consequence of any reliance upon information on there (external sites) or here.


Capco


๐Ÿ’ฐ $58k - $93k*

Design

 

Python

Consultancy

Recruiter

Growth

Financial

Scala

Management

Senior

Operational

Recruitment

Capco is hiring a

Remote Mid Big Data Engineer

\nCAPCO POLAND \n\n*We are looking for Poland based candidate. The job is remote but may require some business trips.\n\nJoining Capco means joining an organisation that is committed to an inclusive working environment where youโ€™re encouraged to #BeYourselfAtWork. We celebrate individuality and recognize that diversity and inclusion, in all forms, is critical to success. Itโ€™s important to us that we recruit and develop as diverse a range of talent as we can and we believe that everyone brings something different to the table โ€“ so weโ€™d love to know what makes you different. Such differences may mean we need to make changes to our process to allow you the best possible platform to succeed, and we are happy to cater to any reasonable adjustments you may require. You will find the section to let us know of these at the bottom of your application form or you can mention it directly to your recruiter at any stage and they will be happy to help.\n\nCapco Poland is a global technology and management consultancy specializing in driving digital transformation across the financial services industry. We are passionate about helping our clients succeed in an ever-changing industry.\n\nWe also are experts in      focused on development, automation, innovation, and long-term projects in financial services. In Capco, you can code, write, create, and live at your maximum capabilities without getting dull, tired, or foggy.\n\nWe're seeking a skilled Mid Big Data Engineer to join our Team. The ideal candidate will be responsible for designing, implementing and maintaining scalable data pipelines and solutions on on-prem / migration / cloud projects for large scale data processing and analytics.\n\n \n\nTHINGS YOU WILL DO\n\n\n* Design, develop and maintain robust data pipelines using Scala or Python, Spark, Hadoop, SQL for batch and streaming data processing\n\n* Collaborate with cross-functional teams to understand data requirements and design efficient solutions that meet business needs \n\n* Optimize Spark jobs and data processing workflows for performance, scalability and reliability\n\n* Ensure data quality, integrity and security throughout the data lifecycle\n\n* Troubleshoot and resolve data pipeline issues in a timely manner to minimize downtime and impact on business operations\n\n* Stay updated on industry best practices, emerging technologies, and trends in big data processing and analytics\n\n* Document, design specifications, deployment procedures and operational guidelines for data pipelines and systems\n\n* Provide technical guidance and mentorship for new joiners\n\n\n\n\n \n\nTECH STACK: Python or Scala, OOP, Spark, SQL, Hadoop\n\nNice to have: GCP, Pub/Sub, Big Query, Kafka, Juniper, Apache NiFi, Hive, Impala, Cloudera, CI/CD\n\n \n\nSKILLS & EXPERIENCES YOU NEED TO GET THE JOB DONE\n\n\n* min. 3-4 years of experience as a Data Engineer/Big Data Engineer\n\n* University degree in computer science, mathematics, natural sciences, or similar field and relevant working experience\n\n* Excellent SQL skills, including advanced concepts\n\n* Very good programming skills in Python or Scala\n\n* Experience in Spark and Hadoop\n\n* Experience in OOP\n\n* Experience using agile frameworks like Scrum\n\n* Interest in financial services and markets\n\n* Nice to have: experience or knowledge with GCP \n\n* Fluent English communication and presentation skills\n\n* Sense of humor and positive attitude\n\n\n\n\n \n\nWHY JOIN CAPCO?\n\n\n* Employment contract and/or Business to Business - whichever you prefer\n\n* Possibility to work remotely\n\n* Speaking English on daily basis, mainly in contact with foreign stakeholders and peers\n\n* Multiple employee benefits packages (MyBenefit Cafeteria, private medical care, life-insurance)\n\n* Access to 3.000+ Business Courses Platform (Udemy)\n\n* Access to required IT equipment\n\n* Paid Referral Program\n\n* Participation in charity events e.g. Szlachetna Paczka\n\n* Ongoing learning opportunities to help you acquire new skills or deepen existing expertise\n\n* Being part of the core squad focused on the growth of the Polish business unit\n\n* A flat, non-hierarchical structure that will enable you to work with senior partners and directly with clients\n\n* A work culture focused on innovation and creating lasting value for our clients and employees\n\n\n\n\n \n\nONLINE RECRUITMENT PROCESS STEPS*\n\n\n* Screening call with the Recruiter\n\n* Technical interview: first stage\n\n* Client Interview\n\n* Feedback/Offer\n\n\n\n\n*The recruitment process may be modified \n\n#Salary and compensation\n No salary data published by company so we estimated salary based on similar jobs related to Design, Python, Recruiter, Cloud, Senior and Engineer jobs that are similar:\n\n $57,500 — $92,500/year\n
\n\n#Benefits\n ๐Ÿ’ฐ 401(k)\n\n๐ŸŒŽ Distributed team\n\nโฐ Async\n\n๐Ÿค“ Vision insurance\n\n๐Ÿฆท Dental insurance\n\n๐Ÿš‘ Medical insurance\n\n๐Ÿ– Unlimited vacation\n\n๐Ÿ– Paid time off\n\n๐Ÿ“† 4 day workweek\n\n๐Ÿ’ฐ 401k matching\n\n๐Ÿ” Company retreats\n\n๐Ÿฌ Coworking budget\n\n๐Ÿ“š Learning budget\n\n๐Ÿ’ช Free gym membership\n\n๐Ÿง˜ Mental wellness budget\n\n๐Ÿ–ฅ Home office budget\n\n๐Ÿฅง Pay in crypto\n\n๐Ÿฅธ Pseudonymous\n\n๐Ÿ’ฐ Profit sharing\n\n๐Ÿ’ฐ Equity compensation\n\nโฌœ๏ธ No whiteboard interview\n\n๐Ÿ‘€ No monitoring system\n\n๐Ÿšซ No politics at work\n\n๐ŸŽ… We hire old (and young)\n\n
\n\n#Location\nKrakรณw, Lesser Poland Voivodeship, Poland
Apply for this job

๐Ÿ‘‰ Please reference you found the job on Remote OK, this helps us get more companies to post here, thanks!

When applying for jobs, you should NEVER have to pay to apply. You should also NEVER have to pay to buy equipment which they then pay you back for later. Also never pay for trainings you have to do. Those are scams! NEVER PAY FOR ANYTHING! Posts that link to pages with "how to work online" are also scams. Don't use them or pay for them. Also always verify you're actually talking to the company in the job post and not an imposter. A good idea is to check the domain name for the site/email and see if it's the actual company's main domain name. Scams in remote work are rampant, be careful! Read more to avoid scams. When clicking on the button to apply above, you will leave Remote OK and go to the job application page for that company outside this site. Remote OK accepts no liability or responsibility as a consequence of any reliance upon information on there (external sites) or here.


Ocrolus Inc.

 This job is getting a high amount of applications right now (10% of viewers clicked Apply)

๐Ÿ’ฐ $65k - $115k*

Software

 

Redis

 

System

Front End

Support

Growth

Code

Fintech

Java

Git

Senior

Junior

Engineering

Backend

Ocrolus Inc.

Apply now

๐Ÿ‘€ 2,021 views

โœ… 212 applied (10%)

Share this job:
Get a rok.co short link

Ocrolus Inc. is hiring a

Remote Senior Software Engineer

\nThe Mortgage Engineering team is seeking a highly skilled and experienced Senior Backend Engineer with a strong focus on microservices architecture to join our team. The ideal candidate will be proficient in Java, and possess in-depth knowledge of Kafka, SQS, Redis, Postgres, Grafana, and Kubernetes. You are an expert in working with and scaling event-driven systems, webhooks, RESTful APIs and solving challenges with concurrency and distributed systems. As a Senior Backend Engineer at Ocrolus, you will be responsible for designing, developing, and maintaining highly scalable and reliable backend systems. You will work closely with product managers, designers, and other engineers to ensure our services meet the highest standards of performance and reliability, specifically tailored to the needs of the mortgage industry.\n\nKey Responsibilities:\n\n\n* Design, develop, and maintain backend services and microservices architecture using Java.\n\n* Implement event-driven systems utilizing Kafka and AWS SQS for real-time data processing and messaging.\n\n* Optimize and manage in-memory data stores with Redis for high-speed caching and data retrieval.\n\n* Develop and maintain robust database solutions with Postgres, ensuring data integrity and performance with PgAnalyze.\n\n* Deploy, monitor, and manage containerized applications using Kubernetes and Terraform and ensure its scalability and resilience and our manage cloud infrastructure.\n\n* Collaborate closely with product managers and designers to understand requirements and deliver technical solutions that meet business needs.\n\n* Develop and maintain RESTful APIs and gRPC services to support seamless integration with frontend applications and third-party services.\n\n* Ensure secure and efficient authentication and authorization processes using OAuth.\n\n* Manage codebases in a monorepo environment using Bazel for build automation.\n\n* Troubleshoot and resolve client support issues in a timely manner, ensuring minimal disruption to service.\n\n* Continuously explore and implement new technologies and frameworks to improve system performance and efficiency.\n\n* Write and maintain technical documentation on Confluence to document technical plans and processes, and facilitate knowledge sharing across the team.\n\n* Mentor junior engineers and contribute to the overall growth and development of the engineering team.\n\n\n\n\nRequired Qualifications:\n\n\n* Bachelorโ€™s or Masterโ€™s degree in Computer Science, Engineering, or a related field.\n\n* 5+ years of professional experience in backend development with a focus on microservices.\n\n* Proficiency in Java, with a strong preference for expertise in Java and the Spring framework.\n\n* Strong experience with Apache Kafka for building event-driven architectures.\n\n* Hands-on experience with AWS SQS for message queuing and processing.\n\n* Expertise in Redis for caching and in-memory data management.\n\n* Solid understanding of Postgres or other relational databases, including performance tuning, migrations, and optimization.\n\n* Proven experience with Kubernetes for container orchestration and management.\n\n* Proficiency in developing and consuming RESTful APIs and gRPC services.\n\n* Proficiency with command line and Git for version control and Github for code reviews.\n\n* Familiarity with OAuth for secure authentication and authorization.\n\n* Strong understanding of software development best practices, including version control, testing, and CI/CD automation.\n\n* Excellent problem-solving skills and the ability to work independently and as part of a team.\n\n* Strong communication skills and the ability to articulate complex technical concepts to non-technical stakeholders.\n\n\n\n\nPreferred Qualifications:\n\n\n* Experience working in the mortgage and fintech industries, with a deep understanding of domain-specific challenges and B2B SaSS requirements.\n\n* Experience managing codebases in a monorepo environment with Bazel for build automation.\n\n* Understanding of security best practices and implementation in microservices.\n\n* Experience with performance monitoring and logging tools such as Grafana, Sentry, PgAnalyze, Prometheus, and New Relic.\n\n* Familiarity with cloud platforms such as AWS.\n\n* Familiarity with Python.\n\n\n \n\n#Salary and compensation\n No salary data published by company so we estimated salary based on similar jobs related to Redis, Java, Cloud, Git, Senior, Junior, Engineer and Backend jobs that are similar:\n\n $65,000 — $115,000/year\n
\n\n#Benefits\n ๐Ÿ’ฐ 401(k)\n\n๐ŸŒŽ Distributed team\n\nโฐ Async\n\n๐Ÿค“ Vision insurance\n\n๐Ÿฆท Dental insurance\n\n๐Ÿš‘ Medical insurance\n\n๐Ÿ– Unlimited vacation\n\n๐Ÿ– Paid time off\n\n๐Ÿ“† 4 day workweek\n\n๐Ÿ’ฐ 401k matching\n\n๐Ÿ” Company retreats\n\n๐Ÿฌ Coworking budget\n\n๐Ÿ“š Learning budget\n\n๐Ÿ’ช Free gym membership\n\n๐Ÿง˜ Mental wellness budget\n\n๐Ÿ–ฅ Home office budget\n\n๐Ÿฅง Pay in crypto\n\n๐Ÿฅธ Pseudonymous\n\n๐Ÿ’ฐ Profit sharing\n\n๐Ÿ’ฐ Equity compensation\n\nโฌœ๏ธ No whiteboard interview\n\n๐Ÿ‘€ No monitoring system\n\n๐Ÿšซ No politics at work\n\n๐ŸŽ… We hire old (and young)\n\n
\n\n#Location\nGurgaon, Haryana, India
Apply for this job

๐Ÿ‘‰ Please reference you found the job on Remote OK, this helps us get more companies to post here, thanks!

When applying for jobs, you should NEVER have to pay to apply. You should also NEVER have to pay to buy equipment which they then pay you back for later. Also never pay for trainings you have to do. Those are scams! NEVER PAY FOR ANYTHING! Posts that link to pages with "how to work online" are also scams. Don't use them or pay for them. Also always verify you're actually talking to the company in the job post and not an imposter. A good idea is to check the domain name for the site/email and see if it's the actual company's main domain name. Scams in remote work are rampant, be careful! Read more to avoid scams. When clicking on the button to apply above, you will leave Remote OK and go to the job application page for that company outside this site. Remote OK accepts no liability or responsibility as a consequence of any reliance upon information on there (external sites) or here.


Daniel J Edelman Holdings


๐ŸŒ Probably worldwide
๐Ÿ’ฐ $70k - $100k*

Design

 

Amazon

 

System

Consultancy

Support

Software

Strategy

Management

Analytics

Engineering

Daniel J Edelman Holdings

Apply now

๐Ÿ‘€ 781 views

โœ… 69 applied (9%)

Share this job:
Get a rok.co short link

Daniel J Edelman Holdings is hiring a

Remote Platform Engineer

\nWe are seeking a skilled Platform Engineer to join our team. They will be instrumental in developing and optimizing our cloud architecture as well as managing our development and analytics tooling. The ideal candidate has a proven track record of maintaining and developing cloud-based architecture which supports large-scale data initiatives for data science and software development teams. \n \nThe Platform Engineer will be responsible for designing, implementing, and managing cloud architectures to meet business requirements. The ideal candidate possesses expertise in translating business needs into scalable cloud solutions while ensuring security, reliability, and cost-effectiveness.\n\n\n\nResponsibilities:\n* Collaborate with business stakeholders to understand the product requirements and translate them into scalable and resilient cloud architectures.\n* Collaborate closely with Data Engineering, Data Science and Software Development teams to contribute to the design of the cloud solutions for our products.\n* End-to-end implementation of the optimized and secure cloud-based/cloud-native architecture. \n* Develop and document cloud architecture designs, ensuring alignment with industry best practices.\n* Provide expertise in cloud and platform engineering to the Product Data team, ensuring alignment with the company's strategic goals.\n* Contribution to the selection and integration of cloud-based  vendors, tools and frameworks.\n* Keep up with emerging trends in cloud engineering  and introduce new technologies or practices that can benefit the organization.\n* Implement security measures to safeguard cloud environments, including identity and access management, encryption, and compliance controls.\n* Conduct regular security assessments and address vulnerabilities promptly.\n* Monitor and optimize cloud infrastructure for performance, cost, and reliability.\n* Implement performance tuning strategies to enhance overall system efficiency.\n* Continuous improvement and innovation for cloud and platform engineering.\n* Implement and manage the provisioning of cloud resources based on project requirements.\n* Responsible for the maintenance and support of cloud-based and cloud-native architecture including access controls, security and networking.\n* Configure and fine-tune cloud infrastructure components for optimal performance.\n* Perform audits and assessments of cloud environments to ensure compliance with security and regulatory standards.\n* Provide recommendations for continuous improvement and adherence to best practices.\n* Lead the deployment of applications onto our cloud platform, ensuring seamless integration and functionality.\n* Manage and monitor cloud applications to maintain performance, availability, and scalability.\n\n\n\nQualifications:\n* 3-5 years of proven experience as a Platform Engineer or similar role in designing, implementing, and managing cloud architectures.\n* Expertise in constructing, installing, and maintaining large-scale cloud-native and cloud-based architecture.\n* Database management expertise: Postgres, Snowflake, Lucene-based search engines (Apache Solr/AWS OpenSearch/Elastic Search)\n\n\n* Cloud-Native tooling expertise: Amazon S3, AWS EMR, Amazon EC2 Amazon RDS, Amazon Sagemaker, Amazon ECS, Amazon ECR, Amazon VPC, Amazon IAM (*alternatives from other cloud providers are acceptable) \n* Cloud-based Application tooling: Databricks administration\n* Strong communication in English. Ability to communicate technical concepts to non-technical audience.\n* Cloud Certifications from Cloud providers (AWS. GCP, Azure) \n* Experience with Streaming Technologies such as Apache Kafka and AWS Kinesis\n* Experience Productionizing ML based cloud solutions.\n\n\n\n\n\n#LI-RT9\n\n\nEdelman Data & Intelligence (DXI) is a global, multidisciplinary research, analytics and data consultancy with a distinctly human mission.\n\n\nWe use data and intelligence to help businesses and organizations build trusting relationships with people: making communications more authentic, engagement more exciting and connections more meaningful.\n\n\nDXI brings together and integrates the necessary people-based PR, communications, social, research and exogenous data, as well as the technology infrastructure to create, collect, store and manage first-party data and identity resolution. DXI is comprised of over 350 research specialists, business scientists, data engineers, behavioral and machine-learning experts, and data strategy consultants based in 15 markets around the world.\n\n\nTo learn more, visit: https://www.edelmandxi.com \n\n#Salary and compensation\n No salary data published by company so we estimated salary based on similar jobs related to Design, Cloud and Engineer jobs that are similar:\n\n $70,000 — $100,000/year\n
\n\n#Benefits\n ๐Ÿ’ฐ 401(k)\n\n๐ŸŒŽ Distributed team\n\nโฐ Async\n\n๐Ÿค“ Vision insurance\n\n๐Ÿฆท Dental insurance\n\n๐Ÿš‘ Medical insurance\n\n๐Ÿ– Unlimited vacation\n\n๐Ÿ– Paid time off\n\n๐Ÿ“† 4 day workweek\n\n๐Ÿ’ฐ 401k matching\n\n๐Ÿ” Company retreats\n\n๐Ÿฌ Coworking budget\n\n๐Ÿ“š Learning budget\n\n๐Ÿ’ช Free gym membership\n\n๐Ÿง˜ Mental wellness budget\n\n๐Ÿ–ฅ Home office budget\n\n๐Ÿฅง Pay in crypto\n\n๐Ÿฅธ Pseudonymous\n\n๐Ÿ’ฐ Profit sharing\n\n๐Ÿ’ฐ Equity compensation\n\nโฌœ๏ธ No whiteboard interview\n\n๐Ÿ‘€ No monitoring system\n\n๐Ÿšซ No politics at work\n\n๐ŸŽ… We hire old (and young)\n\n
Apply for this job

๐Ÿ‘‰ Please reference you found the job on Remote OK, this helps us get more companies to post here, thanks!

When applying for jobs, you should NEVER have to pay to apply. You should also NEVER have to pay to buy equipment which they then pay you back for later. Also never pay for trainings you have to do. Those are scams! NEVER PAY FOR ANYTHING! Posts that link to pages with "how to work online" are also scams. Don't use them or pay for them. Also always verify you're actually talking to the company in the job post and not an imposter. A good idea is to check the domain name for the site/email and see if it's the actual company's main domain name. Scams in remote work are rampant, be careful! Read more to avoid scams. When clicking on the button to apply above, you will leave Remote OK and go to the job application page for that company outside this site. Remote OK accepts no liability or responsibility as a consequence of any reliance upon information on there (external sites) or here.


SADA India


๐Ÿ’ฐ $60k - $110k*

Cryptocurrency

 

Teach

Support

Software

Growth

Management

Senior

Operational

Engineering

SADA India is hiring a

Remote Senior Data Engineer

\nJoin SADA India as a Senior Data Engineer!\n\nYour Mission \n\nAs a Sr. Data Engineer on the Enterprise Support service team at SADA, you will reduce customer anxiety about running production workloads in the cloud by implementing and iteratively improving observability and reliability. You will have the opportunity to engage with our customers in a meaningful way by defining, measuring, and improving key business metrics; eliminating toil through automation; inspecting code, design, implementation, and operational procedures; enabling experimentation by helping create a culture of ownership; and winning customer trust through education, skill sharing, and implementing recommendations. Your efforts will accelerate our customersโ€™ cloud adoption journey and we will be with them through the transformation of their applications, infrastructure, and internal processes. You will be part of a new social contract between customers and service providers that demands shared responsibility and accountability: our partnership with our customers will ensure we are working towards a common goal and share a common fate.\n\nThis is primarily a customer-facing role. You will also work closely with SADAโ€™s Customer Experience team to execute their recommendations to our customers, and with Professional Services on large projects that require PMO support.\n\nPathway to Success \n\n#MakeThemRave is at the foundation of all our engineering. Our motivation is to provide customers with an exceptional experience in migrating, developing, modernizing, and operationalizing their systems in the Google Cloud Platform.\n\nYour success starts by positively impacting the direction of a fast-growing practice with vision and passion. You will be measured bi-yearly by the breadth, magnitude, and quality of your contributions, your ability to estimate accurately, customer feedback at the close of projects, how well you collaborate with your peers, and the consultative polish you bring to customer interactions.\n\nAs you continue to execute successfully, we will build a customized development plan together that leads you through the engineering or management growth tracks.\n\nExpectations\n\nCustomer Facing - You will interact with customers on a regular basis, sometimes daily, other times weekly/bi-weekly. Common touchpoints occur when qualifying potential opportunities, at project kickoff, throughout the engagement as progress is communicated, and at project close. You can expect to interact with a range of customer stakeholders, including engineers, technical project managers, and executives.\n\nOnboarding/Training - The first several weeks of onboarding are dedicated to learning and will encompass learning materials/assignments and compliance training, as well as meetings with relevant individuals.\n\nJob Requirements\n\nRequired Credentials:\n\n\n* Google Professional Data Engineer Certified or able to complete within the first 45 days of employment \n\n* A secondary Google Cloud certification in any other specialization\n\n\n\n\nRequired Qualifications: \n\n\n* 5+ years of experience in Cloud support\n\n* Experience in supporting customers preferably in 24/7 environments\n\n* Experience working with Google Cloud data products (CloudSQL, Spanner, Cloud Storage, Pub/Sub, Dataflow, Dataproc, Bigtable, BigQuery, Dataprep, Composer, etc)\n\n* Experience writing software in at least two or more languages such as Python, Java, Scala, or Go\n\n* Experience in building production-grade data solutions (relational and NoSQL)\n\n* Experience with systems monitoring/alerting, capacity planning, and performance tuning\n\n* Experience with BI tools like Tableau, Looker, etc will be an advantage\n\n* Consultative mindset that delights the customer by building good rapport with them to fully understand their requirements and provide accurate solutions\n\n\n\n\nUseful Qualifications:\n\n\n* \n\n\n* Mastery in at least one of the following domain areas:\n\n\n\n\n\n* \n\n\n* Google Cloud DataFlow: building batch/streaming ETL pipelines with frameworks such as Apache Beam or Google Cloud DataFlow and working with messaging systems like Pub/Sub, Kafka, and RabbitMQ; Auto scaling DataFlow clusters, troubleshooting cluster operation issues\n\n* Data Integration Tools: building data pipelines using modern data integration tools such as Fivetran, Striim, Data Fusion, etc. Must have hands-on experience configuring and integrating with multiple Data Sources within and outside of Google Cloud\n\n* Large Enterprise Migration: migrating entire cloud or on-prem assets to Google Cloud including Data Lakes, Data Warehouses, Databases, Business Intelligence, Jobs, etc. Provide consultations for optimizing cost, defining methodology, and coming up with a plan to execute the migration.\n\n\n\n\n\n\n\n\n\n* Experience with IoT architectures and building real-time data streaming pipelines\n\n* Experience operationalizing machine learning models on large datasets\n\n* Demonstrated leadership and self-direction -- a willingness to teach others and learn new techniques\n\n* Demonstrated skills in selecting the right statistical tools given a data analysis problem\n\n* Understanding of Chaos Engineering\n\n* Understanding of PCI, SOC2, and HIPAA compliance standards\n\n* Understanding of the principle of least privilege and security best practices\n\n* Understanding of cryptocurrency and blockchain technology\n\n\n\n\n  \n\n#Salary and compensation\n No salary data published by company so we estimated salary based on similar jobs related to Cloud, Senior and Engineer jobs that are similar:\n\n $60,000 — $110,000/year\n
\n\n#Benefits\n ๐Ÿ’ฐ 401(k)\n\n๐ŸŒŽ Distributed team\n\nโฐ Async\n\n๐Ÿค“ Vision insurance\n\n๐Ÿฆท Dental insurance\n\n๐Ÿš‘ Medical insurance\n\n๐Ÿ– Unlimited vacation\n\n๐Ÿ– Paid time off\n\n๐Ÿ“† 4 day workweek\n\n๐Ÿ’ฐ 401k matching\n\n๐Ÿ” Company retreats\n\n๐Ÿฌ Coworking budget\n\n๐Ÿ“š Learning budget\n\n๐Ÿ’ช Free gym membership\n\n๐Ÿง˜ Mental wellness budget\n\n๐Ÿ–ฅ Home office budget\n\n๐Ÿฅง Pay in crypto\n\n๐Ÿฅธ Pseudonymous\n\n๐Ÿ’ฐ Profit sharing\n\n๐Ÿ’ฐ Equity compensation\n\nโฌœ๏ธ No whiteboard interview\n\n๐Ÿ‘€ No monitoring system\n\n๐Ÿšซ No politics at work\n\n๐ŸŽ… We hire old (and young)\n\n
\n\n#Location\nThiruvananthapuram, Kerala, India
Apply for this job

๐Ÿ‘‰ Please reference you found the job on Remote OK, this helps us get more companies to post here, thanks!

When applying for jobs, you should NEVER have to pay to apply. You should also NEVER have to pay to buy equipment which they then pay you back for later. Also never pay for trainings you have to do. Those are scams! NEVER PAY FOR ANYTHING! Posts that link to pages with "how to work online" are also scams. Don't use them or pay for them. Also always verify you're actually talking to the company in the job post and not an imposter. A good idea is to check the domain name for the site/email and see if it's the actual company's main domain name. Scams in remote work are rampant, be careful! Read more to avoid scams. When clicking on the button to apply above, you will leave Remote OK and go to the job application page for that company outside this site. Remote OK accepts no liability or responsibility as a consequence of any reliance upon information on there (external sites) or here.

closed

ZOE


๐Ÿ’ฐ $60k - $105k*

Python

 

Ceo

Software

Growth

Code

Senior

Reliability

Health

Healthcare

Engineering

ZOE

๐Ÿ‘€ 832 views

โœ… 72 applied (9%)

Share this job:
Get a rok.co short link

This job post is closed and the position is probably filled. Please do not apply. Work for ZOE and want to re-open this job? Use the edit link in the email when you posted the job! ๐Ÿค– Closed by robot after apply link errored w/ code 404 1 year ago


\nWe are redefining how people approach their health\n\n\nZOE is combining scientific research at a scale never before imagined and cutting-edge AI to improve the health of millions.  \n\n\nCreated by the worldโ€™s top scientists, our personalised nutrition program is reimagining a fundamental human need โ€“ eating well for your own body. Currently available in the US and the UK, ZOE is already helping > 100k ZOE members to adopt healthier habits and live better. Our work and expertise in biology, engineering, data science, and nutrition science have led to multiple breakthrough papers in leading scientific journals such as Nature Medicine, Science, The Lancet, and more.\n\n\nTo learn more, head to Spotify, Apple Podcasts, or Audible to listen to our Science & Nutrition Podcast (with 3 million listens!) \n\n\nA remote-first, high-growth startup, we are backed by founders, investors, and entrepreneurs who have built multi-billion dollar technology companies. We are always looking for innovative thinkers and builders to join our team on a thrilling mission to tackle epic health problems. Together, we can improve human health and touch millions of lives. \n\n\nWe value inclusivity, transparency, ownership, open-mindedness and diversity. We are passionate about delivering great results and learning in the open. We want our teams to have the freedom to make long-term, high-impact decisions, and the well-being of our teammates and the people around us is a top priority.\n\n\nCheck out what life is like for our tech team on ZOE Tech. \n\n\nWeโ€™re looking for a Senior Data Engineer to take ZOE even further.\n\n\nAbout the team\n\n\nThe mission of the Core Science team is to transform research trials and data into personalised actionable recommendations that reach our members. We are currently developing a feedback loop to measure the efficacy of ZOE's nutrition and health advice, which will drive the evolution of our recommendations. In addition, the team is conducting supplementary studies in key areas, such as the microbiome, and constructing a platform to facilitate these trials alongside the main product. The team also maintains close collaboration with other stream-aligned teams to deliver scientific discoveries directly to the app.\n\n\nWe operate in a very dynamic and rewarding environment, where we work closely with all sorts of stakeholders to find the best solutions for both the business and our potential customers. Our agile, cross-functional teams use continuous delivery and regular feedback to ensure we deliver value to our customers on a daily basis. Our systems make use of Python, dbt, Apache Airflow, Kotlin, Typescript, React, and FastAPI. We deploy and operate our software using Kubernetes and ML models using VertexAI in GCP. \n\n\nAbout the role\n\n\nAs a Senior data engineer in the Core Science team, you will be working with scientists, data scientists and other engineers to build a platform that empowers our team to conduct scientific research trials and improve the efficacy of ZOEโ€™s nutrition and health advice. Every line of code you write will be a catalyst for groundbreaking discoveries.\n\n\nIn this role, you will also have the opportunity to make a significant impact on the personal and professional development of our team by providing guidance, support, and expertise. You will play a crucial role in helping individuals achieve their goals, overcome challenges, and maximise their potential.\n\n\n\nYou'll be\n* Defining the data requirements from the research trials that the core Science team will run alongside data coming from the main product experience.\n* Automating data collection from a variety of sources (e.g. labs, questionnaires, study coordination tools).orchestrating integration of data derived from these trials into our data warehouse. \n* Coordinating with different product teams to ensure a seamless App experience for both study participants and paid customers. \n* Ensuring consistency and accuracy of all study data used for research and product development.  \n* Conducting exploratory data analysis to understand data patterns and trends. \n* Creating algorithms and ML models when necessary.  \n* Ensuring data security and compliance with regulatory standards.\n* Ensuring data accessibility to internal and external stakeholders with up-to-date documentation on data sources and schemes. \n\n\n\nWe think youโ€™ll be great if you...\n* +6 years of experience in data engineering roles, with a proven track record of working on data integration, ETL processes, and data warehousing \n* Are proficient in Python and SQL and have experience with Data Warehouses (ex: BigQuery, SnowFlake) and interactive computing environments like Jupyter Notebooks.\n* Have knowledge of data governance principles and best practices for ensuring data quality, security, and compliance with regulatory standards.\n* Are detail-oriented and data-savvy to ensure the accuracy and reliability of the data. \n* Are someone who strives to keep their code clean, tests complete and maintained, and their releases frequent.\n* Have experience with cloud platforms like Google Cloud Platform (GCP) and platforms to schedule and monitor data workflows like Apache Airflow.\n* Have a solid understanding of best practices around CI/CD, containers and what a great release process looks like. \n* Have the ability to collaborate effectively with cross-functional teams and communicate technical concepts to non-technical stakeholders.\n* Have a mindset of collaboration, innovation, and a passion for contributing to groundbreaking scientific discoveries.\n\n\n\nNice to have\n* Have experience with dbt and Apache Airflow. \n* Have experience with ML modelling and ML Ops. \n* Have experience with privacy-preserving technologies such as federated learning and data synthesis. \n\n\n\n\n\nThese are the ideal skills, attributes, and experience weโ€™re looking for in this role. Donโ€™t worry if you donโ€™t tick all the boxes, especially on the skills and experience front, weโ€™re happy to upskill for the right candidate. \n\n\nLife as a ZOEntist โ€“ what you can expect from us:\nAs well as industry-benchmarked compensation and all the hardware and software you need, we offer a thoughtfully-curated list of benefits. We expect this list to evolve as we continue supporting our team membersโ€™ long-term personal and professional growth, and their wellbeing. \n\n\nRemote-first: Work flexibly โ€“ from home, our London office, or anywhere within the EU \nStock options: So you can share in our growth \nPaid time off: 28 days paid leave (25 holiday days, plus 2 company-wide reset days, and 1 โ€œlife eventโ€ day) \nEnhanced Parental Leave: On top of the statutory offering\nFlexible private healthcare and life assurance options\nPension contribution: Pay monthly or top up  โ€“ your choice. \nHealth and wellbeing: Like our Employee Assistance Program and Cycle to Work Scheme\nSocial, WFH, and Growth (L&D) budgets. Plus, multiple opportunities to connect, grow, and socialise \n\n\nWeโ€™re all about equal opportunities \nWe know that a successful team is made up of diverse people, able to be their authentic selves. To continue growing our team in the best way, we believe that equal opportunities matter, so we encourage candidates from any underrepresented backgrounds to apply for this role. You can view our Equal Opportunities statement in full here. \n\n\nA closer look at ZOE \nThink youโ€™ve heard our name somewhere before? We were the team behind the COVID Symptom Study, which has since become the ZOE Health Study (ZHS). We use the power of community science to conduct large-scale research from the comfort of contributorsโ€™ own homes. Our collective work and expertise in biology, engineering, and data/nutrition science have led to multiple breakthrough papers in leading scientific journals such as Nature Medicine, Science, The Lancet, and more.\n\n\nSeen ZOE in the media recently? Catch our co-founder Professor Tim Spector (one of the worldโ€™s most cited scientists) and our Chief Scientist Dr Sarah Berry on this BBC Panorama, and listen to CEO Jonathan Wolf unpack the latest in science and nutrition on our ZOE podcast. \n\n\nOh, and if youโ€™re wondering why ZOE? It translates to โ€œLifeโ€ in Greek, which weโ€™re helping ZOE members enjoy to the fullest. \n\n#Salary and compensation\n No salary data published by company so we estimated salary based on similar jobs related to Python, Cloud, Senior and Engineer jobs that are similar:\n\n $60,000 — $105,000/year\n
\n\n#Benefits\n ๐Ÿ’ฐ 401(k)\n\n๐ŸŒŽ Distributed team\n\nโฐ Async\n\n๐Ÿค“ Vision insurance\n\n๐Ÿฆท Dental insurance\n\n๐Ÿš‘ Medical insurance\n\n๐Ÿ– Unlimited vacation\n\n๐Ÿ– Paid time off\n\n๐Ÿ“† 4 day workweek\n\n๐Ÿ’ฐ 401k matching\n\n๐Ÿ” Company retreats\n\n๐Ÿฌ Coworking budget\n\n๐Ÿ“š Learning budget\n\n๐Ÿ’ช Free gym membership\n\n๐Ÿง˜ Mental wellness budget\n\n๐Ÿ–ฅ Home office budget\n\n๐Ÿฅง Pay in crypto\n\n๐Ÿฅธ Pseudonymous\n\n๐Ÿ’ฐ Profit sharing\n\n๐Ÿ’ฐ Equity compensation\n\nโฌœ๏ธ No whiteboard interview\n\n๐Ÿ‘€ No monitoring system\n\n๐Ÿšซ No politics at work\n\n๐ŸŽ… We hire old (and young)\n\n
\n\n#Location\nUK/EU or compatible timezone (Remote)
# How do you apply?\n\nThis job post has been closed by the poster, which means they probably have enough applicants now. Please do not apply.
419ms