Pre-screened and vetted in New Jersey.
Senior Data Engineer specializing in cloud data platforms and large-scale ETL
“Data engineer focused on large-scale ETL/ELT pipelines across cloud stacks (GCP and AWS), including Spark-based transformations and orchestration with Airflow. Has experience loading up to ~2TB per BigQuery target table and designing atomic loads to multiple downstream systems (Elasticsearch + Kafka), with Kubernetes deployment and Jenkins CI/CD.”
Mid-level Business Analyst specializing in BI, reporting, and data analytics
“Finance data and reporting professional with PwC experience who bridges accounting and technology, especially around GL-related reconciliations, reporting accuracy, and close support. While not a direct PeopleSoft GL owner, they bring strong SQL-driven troubleshooting, ETL/data mapping remediation, and process automation experience that helped shorten close cycles and improve audit readiness.”
Mid-level Data Analyst specializing in BI, ETL, and operational analytics
Mid-level Data Analyst specializing in healthcare and financial risk analytics
Mid-level Data Engineer specializing in real-time analytics and FinTech data platforms
Mid-level AI/ML Engineer specializing in Generative AI and healthcare data
“Built and deployed a production RAG-based document Q&A system on Azure OpenAI to help business teams search thousands of PDFs/Word files, using Qdrant vector search, MongoDB, and a Flask API. Demonstrates strong production engineering (streaming large-file ingestion, parallel preprocessing, monitoring/retries) plus systematic prompt/embedding/chunking experimentation to improve accuracy and reduce hallucinations, and has hands-on orchestration experience with ADF/Airflow/Databricks/Synapse.”
Mid-level Data Analyst specializing in healthcare and finance analytics
“Built an end-to-end Alexa smart-home IoT application controlling a Wi-Fi bulb, including ESP32 firmware (MQTT) and an AWS serverless backend (IoT Core/Device Shadow, Lambda, DynamoDB) with a REST API. Demonstrates strong real-time scalability patterns (streaming ingestion, stateless processing, partition-key design) and full-stack delivery with Spring Boot + React (JWT auth, CORS, data-heavy dashboards).”
Mid-level Data Analyst specializing in banking and healthcare analytics
Mid-level AI/ML Engineer specializing in GenAI and financial risk & compliance analytics
“Built and deployed a production LLM-powered financial risk and compliance platform to reduce manual trade exception handling and speed up insights from regulatory documents. Implemented a LangChain multi-agent workflow with structured/unstructured data integration (Redshift + vector DB) and emphasized hallucination reduction for regulatory safety using Amazon Bedrock. Strong MLOps/orchestration background across Kubernetes, Airflow, Jenkins, and monitoring/testing with MLflow, Evidently AI, and PyTest.”
Mid-level Data Analyst specializing in analytics, ETL, and cloud data platforms
“Data analyst with 4 years of experience spanning banking and retail/marketing analytics. Has hands-on experience building churn analytics pipelines in SQL and Python, optimizing large-query performance, and turning stakeholder-aligned metrics into recurring dashboards and business actions.”
Mid-level Business Data Analyst specializing in financial risk and payments analytics
Mid-level Data Scientist specializing in LLMs and applied machine learning
Senior Data Scientist specializing in healthcare ML, LLMs, and responsible AI
“Clinical data scientist who has built an agentic LLM-powered literature review assistant (with RAG-style storage/retrieval) to identify predictors for downstream predictive modeling. Also delivered a patient-focused progression analysis model using Databricks + Airflow orchestration, partnering closely with clinicians to define targets and validate that model insights aligned with clinical expectations.”
Mid-level Data Analyst specializing in CRM and business intelligence analytics
Mid-level Data Scientist specializing in GenAI, RAG, and forecasting
“ML/NLP engineer focused on large-scale data linking for e-commerce-style catalogs and customer records, combining transformer embeddings (BERT/Sentence-BERT), NER, and FAISS-based vector search. Has delivered measurable lifts (e.g., +30% matching accuracy, Precision@10 62%→84%) and built production-grade, scalable pipelines in Airflow/PySpark with strong data quality and schema-drift handling.”
Mid-level AI/ML Engineer specializing in NLP, computer vision, and MLOps
“Built and deployed a production LLM/RAG intelligent document understanding platform for healthcare clinical documents (notes, discharge summaries, diagnostic reports), integrating spaCy entity extraction, Pinecone vector search, and a Spring Boot API on AWS with monitoring and guardrails. Demonstrates strong MLOps/orchestration (LangChain, Airflow, Kubeflow/Kubernetes) and a metrics-driven evaluation approach, and partnered with a healthcare operations manager to cut manual review time by 80%.”
Mid-level Healthcare Data Analyst specializing in clinical and claims analytics
Junior Data Analyst specializing in BI, SQL, and business analytics
“Analytics professional with experience across Dreamline AI, Ultron Technologies, and Infolabz, building SQL/Python data pipelines and BI dashboards for incentive, FMCG, and retail use cases. Stands out for turning messy multi-source data into trusted reporting, automating recurring analytics, and tying dashboard adoption to measurable business outcomes like 50% faster reporting and 30% ROI improvement.”
Mid-level Data Analyst specializing in analytics, AI, and business intelligence
Mid-level Data Analyst specializing in BI, reporting automation, and operational analytics
Junior AI/ML Engineer specializing in applied machine learning and data pipelines
“Built and deployed an LLM-powered automation pipeline that ingests voice and documents, transcribes/extracts key information into structured data, and routes it through backend workflows using Python/FastAPI. Uses n8n to orchestrate multi-step AI processes with validation, retries, and monitoring, and iterates with stakeholders via rapid demos to refine changing requirements.”