Pyspark Production Code Example Jobs Remote Jobs in Usa
5 positions found
“Let goodness, fairness, and most importantly, love prevails in business; profits will inevitably follow.” – NK Chaudhary, founder
What we do for our team members:
- Comprehensive Benefits: Company Paid Holidays, PTO, Parental Involvement Leave, Maternity/Paternity Leave, EAP, No Cost Employee Medical Plan, Vision, Dental, and Company Paid Life Insurance. We also include a match on retirement (401K/Roth).
- Career Development: We're committed to providing growth for career development within the company, supporting our team members' aspirations with a well-defined succession plan that includes a variety of training and development opportunities.
- Pet-Friendly Workplace: We welcome your furry friends! Our 'Bring Your Dogs to Work' policy creates a pet-friendly atmosphere, allowing our team members to enjoy the companionship of their dogs during the workday.
- Wellness Support: Not only do we support an active lifestyle with our on-site basketball court and yoga studio, but we host quarterly mental health events to assist in creating a well-rounded work-life harmony for our team members.
- Sustainability Efforts: Reuse, Renew, and Refresh by joining our Green Team! Responsible for harvesting from the organic community garden, donating goods to local pet shelters and schools, creating educational workshops, leading nature walks, and much more, they promote well-being through sustainable practices.
Our Values
Empowerment • Inclusiveness • Responsibility • Progressive
Learn more about our company story here: Jaipur Rugs Foundation
Since 2004, the Jaipur Rugs Foundation has worked to improve the lives of rug-weaving artisans in India. This is done through training, skills development, and social interventions. By focusing on the ideas and solutions that create social value, the Foundation supports the dignity and heritage of these traditional artisans, believing that healthy and sustainable communities are key to the survival of traditional rug weaving. Jaipur Living has made ethical and socially conscious global citizenship the foundation of its business. Through social initiatives and the Jaipur Rugs Foundation, the company supports a supplier ecosystem without a middleman of more than 40,000 artisans in 700 villages across India by providing them with a livable wage, access to health care, leadership education, and opportunities for personal growth and development. Combining time-honored techniques and of-the-moment trends, every Jaipur Living product is as ethically and responsibly made as it is beautiful.
Learn more about the Jaipur Rugs Foundation here: are a fast-growing, design-led B2B home décor and textiles brand with big ambitions. Over the last 12 months, we have revolutionized our technical foundation, investing in Microsoft Dynamics 365 (F&O) and a Microsoft Fabric ecosystem. We are now looking for a seasoned leader to refine our existing infrastructure, optimize our end-to-end data workflows, and bridge the gap between "raw data" and "reliable business intelligence."
This role demands a strong balance of technical depth and operational management. While you must possess expert-level proficiency in data engineering, specifically within the Microsoft Fabric ecosystem and modern data platforms, we also need a leader who is experienced in analytics, data visualization, BI, and translating business needs into analytical solutions. You will be responsible for defining and executing an outcome-based Data & Analytics strategy, building and developing a global team of data engineers, BI developers, and data analysts, and ensuring the company has trusted, scalable, and decision-ready data at every level of the organization. The ideal candidate is a Fabric-certified or Fabric-trained leader, an exceptional communicator, and a proven people manager who can balance hands-on technical depth with strategic leadership.
Key Responsibilities:
Strategic Management & Outcome-Based Delivery
- Tactical Roadmap: Develop and execute a multi-year roadmap that aligns data engineering, BI, and advanced insights with business priorities (e.g., inventory efficiency, margin protection, and growth).
- Process Standardization: Define what “good” looks like for data reliability, documentation, insight quality, and business impact
- Baseline Maturity: Shift the organization from ad-hoc reporting to repeatable, trusted, decision-ready data products
- Advance Automation: Assess the current-state landscape and define a clear path from foundational reporting to automated, predictive analytics.
- Executive Communication: Serve as the single point of accountability for all data and analytics capabilities, translating technical progress into business-relevant implications across the organization
Infrastructure Optimization & Fabric Engineering
- Systemic Optimization: Lead the audit and refinement of the existing Fabric environment (Lakehouse, Pipelines, Notebooks) to improve overall performance, stability, and refresh reliability
- Engineering Standards: Set the "gold standard" for architecture, data modeling, testing, and deployment (CI/CD), ensuring the stack is hardened for enterprise-scale growth
- Reduce Manual Effort: Minimize operational risk by standardizing pipelines, refresh processes, and metric calculations
- Automation & Reliability: Systematically identify and eliminate manual reporting and spreadsheet-based workflows through robust automation in PySpark and Fabric
- Proactive Governance: Establish monitoring, alerting, and exception-handling processes to manage data quality and refresh failures before they impact the business
Analytics & Decision Enablement
- High-Quality BI Delivery: Oversee the design and delivery of visually appealing Power BI dashboards that simplify complexity and adhere to our design-led brand standards
- Metric Governance: Ensure KPI definitions and reporting logic are consistent across the company, acting as the arbiter of "the truth" for business metrics
- Advanced Analytics: Identify and operationalize high-value use cases for predictive analytics (e.g., demand forecasting, product lifecycle analysis) as platform maturity increases
- Business Translation: Partner with business leaders to translate business requirements into scalable, intuitive, impactful analytics solutions
- Business Evolution: Lead the transition from descriptive and diagnostic reporting to forward-looking insights that support planning and decision-making
Global Team Leadership & Talent Development
- People Leadership: Directly lead and develop a 3–5 person global team (primarily based in India), establishing clear roles, accountability, and a high-performance culture
- Skill Development: Create career paths and skill-development plans for engineers and analysts to ensure consistent, high-quality delivery
- Operating Model: Build a scalable offshore capability that delivers at speed while maintaining rigorous standards for code quality and documentation
Skills & Minimum Qualifications:
To perform this job successfully, an individual must be able to perform each essential duty satisfactorily. The requirements listed below are representative of knowledge, skill, and/or ability required. Reasonable accommodation may be made to enable individuals with disabilities to perform essential functions.
- 10+ years of experience in data engineering, analytics, or BI, with director-level scope or equivalent ownership
- Deep hands-on experience with Microsoft Fabric (Lakehouse, Pipelines, Notebooks, semantic models)
- Fabric certification or formal Fabric training strongly preferred
- Strong experience with PySpark and Spark-based transformations
- Strong understanding of Azure data services and modern data architectures
- Exceptional dashboard-development skills using Power BI; portfolio-quality experience preferred
- Strong understanding of data storytelling, executive-ready visualization, and intuitive UI/UX design
- Experience gathering business requirements and translating them into analytical products
- Proven experience leading and developing global / offshore teams
- Strong communicator with the ability to influence at senior levels
- Experience supporting ERP-driven environments; Dynamics 365 preferred
- Ability to juggle strategy, execution, and stakeholder communication simultaneously
Success Measures (First 12–18 Months)
- Strategy Execution: An outcome-based Data & Analytics strategy that is fully operational and tied to business outcomes
- Optimized Infrastructure: A trusted, scalable Fabric platform with significantly reduced manual reporting and 99%+ data availability
- Dashboard Adoption: A suite of high-quality dashboards used daily and weekly by business leaders to drive decision-making
- Team Growth: A high-performing global team with a track record of delivering complex analytics products with speed and precision
Physical Requirements:
- Remaining in a seated position for long periods of time
- Standing is to remain on one’s feet in an upright position without moving about
- The ability to alternate between sitting and standing is present when a worker has the flexibility to choose between sitting or standing as needed when this need cannot be accommodated by schedules breaks and/or lunch period
- Lifting and transporting items that could weight up to 25 pounds
- Entering text or data into a computer by means of a traditional keyboard
- Expressing or exchanging ideas by means of the spoken work to impart oral information to clients and talent and convey detailed spoken instructions to other workers accurately and quickly
- The ability to hear, understand, and distinguish speech and/or other sounds such as in person and telephone
- Clarity of vision to see computer screens and workspace
They are required to have experience modernizing legacy Microsoft BI environments (including SSIS).
This is not an SSIS-only role.
The consultant will design, modernize, and enhance enterprise data and analytics solutions supporting Cyber Security, Physical Security, Electronic Security and Police operations.
This role includes evolving legacy SQL Server/SSIS-based processes into modern Azure data architectures while designing scalable new ETL/ELT pipelines and delivering executive-level analytics solutions.
The consultant will work directly with stakeholders to deliver production-grade reporting and analytics capabilities across multiple enterprise systems.
This requires architectural thinking and hands-on technical execution.
Core Responsibilities: Candidates must have direct experience building enterprise-grade ETL pipelines and executive Power BI dashboards.
Design and implement modern ETL/ELT pipelines in Azure Assess and refactor existing SSIS packages as part of broader modernization efforts Architect Lakehouse / Medallion data models Develop optimized dimensional data models (star schema) Integrate data from SQL Server, Oracle, APIs, and security platforms Design and deploy enterprise Power BI dashboards Build paginated reports using Power BI Report Builder Optimize DAX and dataset performance Implement Row-Level Security (RLS) Support CI/CD and DevOps deployment processes Produce technical documentation and data lineage artifacts Engage directly with executive stakeholders Required Technical Skills: (Must-Have) Data Engineering & Architecture: Strong ETL/ELT design and optimization experience Advanced SQL (expert-level required) Python / PySpark Dimensional data modeling (star schema required) REST API integrations Azure Data Stack: • Azure Data Factory • Azure Databricks • Azure Synapse Analytics • Azure Data Lake Storage Microsoft Data Platform: • Experience with SQL Server data warehouse environments • Working knowledge of SSIS and experience modernizing or migrating SSIS workflows to Azure-based solutions Power BI: Power BI Desktop (expert-level) Advanced DAX Executive dashboard development Paginated reports (Power BI Report Builder) Data Gateway configuration Incremental refresh Row-Level Security (RLS) Nice to Have: Microsoft Purview Terraform (Infrastructure-as-Code) Orchestration tools (Airflow or equivalent) Security systems data integration experience Experience with C# / .NET web application development (for integration with internal systems or APIs) Experience Requirements: 7+ years of hands-on data engineering / analytics delivery Demonstrated experience building production data pipelines in Azure Proven experience delivering executive-facing Power BI solutions Experience working in complex enterprise environments Software Skills: 4–6 years of experience in Azure for building, deploying, and managing cloud-based data and application services.
Technical Skills: 2–4 years of experience in .NET code development for developing and maintaining enterprise applications and data processing components.
6+ years of experience in Data Modeling including designing logical and physical data models for enterprise data warehouses and analytics systems.
6+ years of experience in Python scripting for data processing, automation, ETL development, and data transformation tasks.
6+ years of experience in Structured Query Language (SQL) for writing complex queries, stored procedures, performance tuning, and data manipulation.
Project description
The customer is looking for a JavaFS / Spring / C# / CSS
Responsibilities
Collaborates with leaders, business analysts, project managers, IT architects, technical leads and other developers, along with internal customers, to understand requirements and develop needs according to business requirements
Maintains and enhances existing enterprise services, applications, and platforms using domain driven design and test-driven development
Troubleshoots and debugs complex issues; identifies and implements solutions
Creates detailed project specifications, requirements, and estimates
Researches and implements new technologies to enhance current processes, security, and performance
Supports the development of coding standards and adheres to best practices and security guidelines
Works closely with software architects and technical leads to ensure decisions meet long-term enterprise growth needs"
Skills
Must have
Bachelor's degree in Computer Science, Computer Engineering, Technology, Information Systems (CIS/MIS), Engineering or related technical discipline, or equivalent experience/training
5+ years of experience designing, developing, and implementing large-scale solutions in production environments TOP Skills Required: Java FullStack / focus on Front End
REACT .NET/C# SQL DataFactory UI Skills, Licenses & Certifications
Proficiency in Full Stack Development
Proficiency with the following technologies:
Programming Languages: C# required, PySpark and Python is a plus.
Frameworks: Spring/SpringBoot, FastAPI, .Net Core, .Net Framework
Front End Technologies: CSS, jQuery, AJAX, Hands on experience in Single Page Architectures such as React is required
Source Control: GitHub, Azure DevOps
CICD: GitHub Actions, Azure DevOps
Data management: PostgreSQL, MS SQL
Integration/APIs Technologies: Kafka, REST, GraphQL
Cloud Providers such as Azure
Other Dev Ops Toolchain: Selenium, SonarQube, GitHub, Kibana, New Relic 10. Proficiency in Networking and security concepts: OAuth, VNet Injection, VNet Integration, Private Endpoints, Azure Firewall 1
Proficiency in object-oriented design techniques and principles 1
Proficiency in Development Tools: Visual Studio, Visual Studio Code, SQL Server Management Studio, Azure Data Studio
Ability to concisely convey ideas verbally, in writing, in code, and in diagrams
Ability to work through others by cooperatively sharing responsibilities
Proficiency in Agile methodologies, such as SCRUM
Principal Data Engineer, you will be the primary architect of Digital Remedy’s next-generation data platform.
You won't just move data; you will design the blueprints for how we handle massive-scale AdTech telemetry.
You will bridge the gap between business priorities and technical execution, ensuring our 'Partner' integrations are scalable, cost-efficient, and resilient.
Key Responsibilities Architectural Leadership: Design and oversee the migration/optimization of our data lakehouse architecture using Databricks.
Production Excellence: Serve as the escalation point for the platform, leveraging a proven record of handling production issues to ensure 24/7 reliability of mission-critical pipelines.
Strategic Capacity Planning: Work directly with the SVP of Tech to translate product roadmaps into technical requirements.
This requires the ability to multi-task across various high-priority workstreams without losing sight of architectural integrity.
Infrastructure as Code: Drive the adoption of Databricks Asset Bundles (DABs) to standardize deployments across dev, staging, and production.
Mentorship & Governance: Provide high-level guidance to engineering squads in India and the US, defining schemas and governance models for BigQuery and Databricks.
Work with engineers across multiple geographies and time zones.
Mandatory Requirements Expert-Level Databricks: Minimum 4+ years of hands-on experience specifically within the Databricks ecosystem (Delta Lake, Unity Catalog, Photon).
The Stack: Deep proficiency in Spark (PySpark/Scala), Python, and SQL.
Experience with GCP (BigQuery) is a major plus.
Scale: Proven experience managing petabyte-scale datasets and high-concurrency pipelines (AdTech experience preferred).
Independent Execution: Must be a self-starter capable of taking a high-level concept from the SVP and driving it to completion with minimal supervision.
Experience: 10+ years in Data Engineering, with at least 3 years in a Principal or Architect capacity.
Preferred Qualifications Incident Management: A history of successfully diagnosing and resolving complex, large-scale production bottlenecks and data outages.
Context Switching: A demonstrated ability to multi-task, balancing long-term architectural R&D with immediate business requests and partner integrations.
FinOps & Communication: Strong understanding of optimizing Databricks/BigQuery costs and the ability to communicate technical trade-offs to non-technical stakeholders.
Please note that this position is not eligible for visa sponsorship, including employer sponsorship for an H-1B visa, OPT-STEM employment, or related programs.
Why Digital Remedy? We are in a high-growth phase, integrating innovative tech from recent acquisitions.
You will have a seat at the table to influence our long-term technical strategy and help us 'build and scale' for the future.
Sinclair is proud to be an equal opportunity employer and a drug free workplace.
Employment practices will not be influenced or affected by virtue of an applicant's or employee's race, color, religion, sex (including pregnancy, gender identity, and sexual orientation), national origin, age, disability, genetic information, military or veteran status or any other characteristic protected by law.
About Sinclair Sinclair, Inc.
(Nasdaq: SBGI) is a diversified media company and a leading provider of local news and sports.
The Company owns, operates and/or provides services to 178 television stations in 81 markets affiliated with all major broadcast networks; owns Tennis Channel, the premium destination for tennis enthusiasts; multicast networks CHARGE, Comet, ROAR and The Nest.
Sinclair’s AMP Media produces a growing portfolio of digital content and original podcasts.
Additional information about Sinclair can be found at .
About Digital Remedy Digital Remedy is a performance media partner for agencies, brands, and media companies.
Through proprietary technology, services, and partnerships, Digital Remedy offers outcome-based media solutions for marketers empowering them to exceed beyond standard media goals.
Whether brand or outcome focused, marketers can access insights, media planning, activation, measurement, reporting, and optimization capabilities to deliver the highest performance tied to ROI.
The effectiveness of Digital Remedy’s solutions is supported by a tenured client roster of leading brands, agencies, and media companies.
In addition, Digital Remedy is a proud recipient of several major award wins including the Digiday Technology Award and Crain’s Best Places to Work several years in a row including a rank of #1 in 2024.
About the Team The life-blood of our organization is our people.
We have a compelling story, a goal-oriented culture, and we take really good care of people.
How good? Here is a glimpse: great benefits, open door policy, upward mobility and a strong desire to see you succeed.
Ready to be part of a winning team? Let’s talk.
The base salary compensation range for this role is $131,250 to $175,000.
Final compensation for this role will be determined by various factors such as a candidate’s relevant work experience, skills, certifications, and geographic location.
Full time positions are eligible for benefits that include participation in a retirement plan, quarterly and annual incentive plan bonuses, life and disability insurance, health, dental and vision plans, flexible spending accounts, 15 paid vacation days, 2 paid personal days, 9 paid holidays, 40 hours of paid sick leave, parental leave, and employee stock purchase plan.
- $78.00 Negotiable based upon years of experience Global Financial Firm located in RUTHERFORD, NJ has an immediate contract opportunity for an experienced Data Scientist Summary: Immediate contract opportunity for an experienced professional Location: Rutherford, NJ Role involves establishing and implementing new or revised applications and programs in the Technology team Focus on data extraction and analysis from structured and unstructured sources Develop systems to clean results and build predictive and prescriptive models Implement models in a production environment Address complex problems involving financial data with a focus on credit risk management Open to learning new and advanced models in LLM and GenAI Responsibilities: Architect, design, and implement robust Python-based solutions Lead global technology teams across different time zones, enabling a hybrid work model Drive Object-Oriented Programming (OOPS) principle-based development Mentor team members to leverage advanced Python coding techniques and features Partner with multiple management teams for integration of functions and system enhancements Resolve high impact problems/projects through in-depth evaluation of complex business processes Provide expertise in applications programming and ensure design adheres to architecture blueprint Develop standards for coding, testing, debugging, and implementation Provide in-depth analysis with interpretive thinking for defining issues and developing solutions Serve as advisor or coach to mid-level developers and analysts Implement Unit Testing and TDD to ensure software quality and maintainability Assess risk when making business decisions, demonstrating consideration for the company's reputation Requirements: PhD or master’s degree in computer science, data science, statistics, mathematics, engineering, or related field 5 years of industry experience specializing in ML modeling, ranking, recommendations, or personalization systems 5 years of experience designing and developing scalable and reliable machine learning systems Strong background in ML/DL/LLM algorithms, model architectures, and training techniques Proficiency in Python, SQL, Spark, PySpark, TensorFlow or other analytical/model-building programming languages Ability to work independently and collaboratively within a team Preferred Skills: Experience in GenAI/LLMs projects Familiarity with distributed data/computing tools (e.g., Hadoop, Hive, Spark, MySQL) Background in financial business like banking, risk management Familiarity with capital markets and financial instruments and modeling