Hacker Read top | best | new | newcomments | leaders | about | bookmarklet login

SEEKING WORK - Bucharest or REMOTE

Experienced Data Scientist.

Keywords: Apache Spark, scaling algorithms.

Well-versed in digging through data to find key insights and curating a compelling story from complex analyses, passionate about delving into data from different systems, at different timescales, and in complex formats to uncover hidden relationships.

Machine Learning knowledge acquired from personal experimentation with Spark: Linear / Logistic Regression, Decision Trees, NaiveBayes, Alternating Least Squares (Recommender Systems), TF-IDF

Professional Background (formerly): ETL Developer / Traditional DWHs / Kimball's Methodology

Computer Science Skills / Core: Data Structures, Algorithms, Functional Programming Paradigm, Relational Databases

Big Data Framework / Core: Spark

Big Data / Other: Apache Kafka => Spark Streaming from Kafka topics

Source Control: GitHub

Source Control / Other: BitBucket

DevOps / Other: Docker / DockerHub

Programming Languages / Core: Python, Scala

Programming Language / Other: Haskell

Keen interest in experimenting with open-source Big Data technologies.

E-mail address in the profile.



sort by: page size:

SEEKING WORK - REMOTE ONLY

Experienced Data Scientist.

Keywords: Apache Spark, scaling algorithms.

Well-versed in digging through data to find key insights and curating a compelling story from complex analyses, passionate about delving into data from different systems, at different timescales, and in complex formats to uncover hidden relationships.

Machine Learning with Spark: Linear / Logistic Regression, Decision Trees, NaiveBayes, Alternating Least Squares (Recommender Systems), TF-IDF, Frequent Pattern Mining

Professional Background (formerly): ETL Developer / Traditional DWHs / Kimball's Methodology

Computer Science Skills / Core: Data Structures, Algorithms, Functional Programming Paradigm, Relational Databases

Big Data Framework / Core: Spark

Big Data / Other: Apache Kafka => Spark Streaming from Kafka topics

Big Data / Other: Apache Cassandra => Data Modeling

Source Control: GitHub

Source Control / Other: BitBucket

DevOps / Other: Docker / DockerHub

Programming Languages / Core: Scala, Python

Programming Language / Other: Haskell

Keen interest in experimenting with open-source Big Data technologies.

E-mail address in the profile.

I'm currently doing work on the freelancing site upwork.com, if you feel more comfortable "trying before you buy" we can start with a limited-scope, fixed-budget project there.


Location: Bucharest

Remote: Yes

Willing to relocate: No

Technologies: Spark, Kafka, Python, Scala, Machine Learning, Data Science

Résumé: https://www.linkedin.com/in/apachesparkdataengineer

Email: in the profile

Experienced data scientist.

Well-versed in digging through data to find key insights and curating a compelling story from complex analyses, passionate about delving into data from different systems, at different timescales, and in complex formats to uncover hidden relationships.


SEEKING WORK - REMOTE ONLY

Keywords: Spark Streaming, Cassandra data modeling.

Well-versed in digging through data to find key insights and curating a compelling story from complex analyses, passionate about delving into data from different systems, at different timescales, and in complex formats to uncover hidden relationships.

Machine Learning with Spark: Linear / Logistic Regression, Decision Trees, NaiveBayes, Alternating Least Squares (Recommender Systems), TF-IDF, Frequent Pattern Mining

Professional Background (formerly): ETL Developer / Traditional DWHs / Kimball's Methodology

Computer Science Skills / Core: Data Structures, Algorithms, Functional Programming Paradigm, Relational Databases

Big Data Framework / Core: Spark

Big Data / Other: Apache Kafka => Spark Streaming from Kafka topics

Big Data / Other: Apache Cassandra => Data Modeling

Source Control: GitHub

Source Control / Other: BitBucket

DevOps / Other: Docker / DockerHub

Programming Languages / Core: Scala, Python

Programming Language / Other: Haskell

Keen interest in experimenting with open-source Big Data technologies.

E-mail address in the profile.


SEEKING WORK - REMOTE ONLY

Keywords: Spark Streaming, Cassandra data modeling.

Well-versed in digging through data to find key insights and curating a compelling story from complex analyses, passionate about delving into data from different systems, at different timescales, and in complex formats to uncover hidden relationships.

Machine Learning with Spark: Linear / Logistic Regression, Decision Trees, NaiveBayes, Alternating Least Squares (Recommender Systems), TF-IDF, Frequent Pattern Mining

Professional Background (formerly): ETL Developer / Traditional DWHs / Kimball's Methodology

Computer Science Skills: Data Structures, Algorithms, Functional Programming Paradigm, Relational Databases

Big Data / Core Skill: Spark

Big Data / Core Skill: Apache Cassandra => Data Modeling

Big Data / Other: Apache Kafka => Spark Streaming from Kafka topics

Programming Languages: Scala, Python

Keen interest in experimenting with open-source Big Data technologies.

E-mail address in the profile.


SEEKING WORK - REMOTE ONLY

Keywords: Spark Streaming, Cassandra data modeling, Spark GraphFrames.

Well-versed in digging through data to find key insights and curating a compelling story from complex analyses, passionate about delving into data from different systems, at different timescales, and in complex formats to uncover hidden relationships. Machine Learning with Spark: Linear / Logistic Regression, Decision Trees, NaiveBayes, Alternating Least Squares (Recommender Systems), TF-IDF, Frequent Pattern Mining

Big Data / Core Skill: Apache Spark

Big Data / Core Skill: Apache Cassandra (Data Modeling)

Big Data / Core Skill: Graph Modeling / Algorithms / Queries (with Spark GraphFrames and Neo4J)

Big Data / Other: Apache Kafka (incl. KafkaConnect), ElasticSearch, RedShift

Programming Languages: Scala, Cypher (Neo4J's query language)

Secondary Skills: Git, Docker, AWS, Avro, Parquet, Zookeeper, HDFS, Yarn, Akka, Mesos, Linux.

Keen interest in experimenting with open-source Big Data technologies.

E-mail address in the profile.


Hi,

I have 5+ years of Big Data and Data Science Experience and I'm a DataBricks Certified Apache Spark Developer, MapR Certified Hadoop Developer, Cloudera Certified Hadoop and Spark Developer, Cloudera Certified Hadoop Administrate, DataStax Certified Apache Cassandra Developer and I have very good experience in working with USA clients. My profile here. https://in.linkedin.com/in/sandishkumar

https://streamsets.com/blog/visualizing-netflow-data-streams...

https://www.phdata.io/visualizing-netflow-data-with-apache-k...

--

Thanks, Regards, SandishKumar HN


Hi,

I have 5+ years of Big Data and Data Science Experience and I'm a DataBricks Certified Apache Spark Developer, MapR Certified Hadoop Developer, Cloudera Certified Hadoop and Spark Developer, Cloudera Certified Hadoop Administrate, DataStax Certified Apache Cassandra Developer and I have very good experience in working with USA clients. My profile here. https://in.linkedin.com/in/sandishkumar

https://streamsets.com/blog/visualizing-netflow-data-streams...

https://www.phdata.io/visualizing-netflow-data-with-apache-k...

--

Thanks, Regards, SandishKumar HN


Hi,

I have 5+ years of Big Data and Data Science Experience and I'm a DataBricks Certified Apache Spark Developer, MapR Certified Hadoop Developer, Cloudera Certified Hadoop and Spark Developer, Cloudera Certified Hadoop Administrate, DataStax Certified Apache Cassandra Developer and I have very good experience in working with USA clients. My profile here. https://in.linkedin.com/in/sandishkumar

https://streamsets.com/blog/visualizing-netflow-data-streams...

https://www.phdata.io/visualizing-netflow-data-with-apache-k...

--

Thanks, Regards, SandishKumar HN


Hi,

I have 5+ years of Big Data and Data Science Experience and I'm a DataBricks Certified Apache Spark Developer, MapR Certified Hadoop Developer, Cloudera Certified Hadoop and Spark Developer, Cloudera Certified Hadoop Administrate, DataStax Certified Apache Cassandra Developer and I have very good experience in working with USA clients. My profile here. https://in.linkedin.com/in/sandishkumar

https://streamsets.com/blog/visualizing-netflow-data-streams...

https://www.phdata.io/visualizing-netflow-data-with-apache-k...

--

Thanks, Regards, SandishKumar HN


Hi,

I have 5+ years of Big Data and Data Science Experience and I'm a DataBricks Certified Apache Spark Developer, MapR Certified Hadoop Developer, Cloudera Certified Hadoop and Spark Developer, Cloudera Certified Hadoop Administrate, DataStax Certified Apache Cassandra Developer and I have very good experience in working with USA clients. My profile here. https://in.linkedin.com/in/sandishkumar

https://streamsets.com/blog/visualizing-netflow-data-streams...

https://www.phdata.io/visualizing-netflow-data-with-apache-k...

--

Thanks, Regards, SandishKumar HN


SEEKING WORK, Primarily Remote (based in Eindhoven, NL)

I am a strong data engineer who is passionate about large-scale distributed systems and streaming pipelines, and cares about producing clean, elegant, maintainable, robust, well-tested Scala / Spark code.

Core Skills:

? Kafka, Spark Streaming, Avro

? Cassandra (DevOps, Data Modeling)

? Graph Modeling / Algorithms / Queries (with Spark GraphFrames and Neo4J)

? Programming Languages: Scala (highly proficient, 8 years exp.), Python (proficient)

Other Skills: Git, Docker, Akka Streams, Apache Ignite, Parquet, HBase, Zookeeper, HDFS, ElasticSearch, AWS (EC2/AutoScG/S3/RedShift/EMR).

Educational Background: Computer Science

Solid experience working remotely.

All of my recent work history (8 years) is exclusively with startups.

Profile: https://angel.co/dan-serban

My rate is $125/hour.

E-mail address in the profile.


SEEKING WORK, Data Engineer, Primarily Remote

Well-rounded Scala data engineer with deep knowledge of the internals of distributed datastores. Solid experience working remotely and working with teams that are distributed geographically. I typically work Pacific Time hours.

Core Skills:

? Cassandra (Data Modeling, Troubleshooting Performance And Operational Issues)

? Stream Processing At Scale: Kafka, Flink, Spark Streaming

? Custom-Crafted TF-IDF / Embeddings, Vector-Based Semantic Search, Deep Intent Recognition In Search Engine Queries

? Languages: Scala, Python, SQL (proficient), Rust, Golang (ramping up)

Other Skills: Airflow, DBT, Snowflake, Qdrant, Databricks/DeltaLake, BigQuery, Redshift, Kinesis, PrestoSQL/Trino, Airbyte, Reverse-Engineering Search Engine Technologies.

Educational Background: Computer Science.

More details: https://drive.google.com/file/d/1fYGY5QBpkE1OmTNV7fhJPQyxv8f...

E-mail address in the profile.


SEEKING WORK, Primarily Remote (based in Eindhoven, NL)

I am a strong data engineer who is passionate about large-scale distributed systems and streaming pipelines, and cares about producing clean, elegant, maintainable, robust, well-tested Scala / Spark code.

Core Skills:

? Kafka, Spark Streaming, Avro

? Cassandra (DevOps, Data Modeling)

? Graph Modeling / Algorithms / Queries (with Spark GraphFrames and Neo4J)

? Programming Languages: Scala (highly proficient, 8 years exp.), Python (proficient)

Other Skills: Git, Docker, Akka Streams, HBase, Zookeeper, HDFS, ElasticSearch, AWS (EC2/S3/EMR).

Educational Background: Computer Science

Solid experience working remotely.

All of my recent work history (8 years) is exclusively with startups.

Profile: https://angel.co/dan-serban

Rate: $125/hour.

E-mail address in the profile.


SEEKING WORK, Data Engineer, Primarily Remote

Well-rounded Scala data engineer with deep knowledge of the internals of distributed datastores. Solid experience working remotely and working with teams that are distributed geographically. I typically work Pacific Time hours.

Core Skills:

? Cassandra (Data Modeling, Troubleshooting Performance And Operational Issues)

? Stream Processing At Scale: Kafka, Flink, Spark Streaming

? Custom-Crafted TF-IDF / Embeddings, Vector-Based Semantic Search, Deep Intent Recognition In Search Engine Queries

? Languages: Scala, Python, SQL (proficient), Rust, Golang (ramping up)

Other Skills: Airflow, DBT, Snowflake, Qdrant, Databricks/DeltaLake, BigQuery, Redshift, Kinesis, PrestoSQL/Trino, Airbyte, Reverse-Engineering Search Engine Technologies.

Educational Background: Computer Science.

More details: https://drive.google.com/file/d/1yGBkmrW-9iwDgfJW2G1El6btwDU...

E-mail address in the profile.


SEEKING WORK, Data Engineer, Primarily Remote

Well-rounded Scala data engineer with deep knowledge of the internals of distributed datastores. Solid experience working remotely and working with teams that are distributed geographically. I typically work Pacific Time hours.

Core Skills:

? Cassandra (Data Modeling, Troubleshooting Performance And Operational Issues)

? Stream Processing At Scale: Kafka, Flink, Spark Streaming

? Custom-Crafted TF-IDF / Embeddings, Vector-Based Semantic Search, Deep Intent Recognition In Search Engine Queries

? Languages: Scala, Python, SQL (proficient), Rust, Golang (ramping up)

Other Skills: Airflow, DBT, Snowflake, Qdrant, Databricks/DeltaLake, BigQuery, Redshift, Kinesis, PrestoSQL/Trino, Airbyte, Reverse-Engineering Search Engine Technologies.

Educational Background: Computer Science.

More details: https://drive.google.com/file/d/1c22IdKLc09TQSGiQnoVpFl4XVnT...

E-mail address in the profile.


SEEKING WORK, Primarily Remote (based in Eindhoven, NL)

I am a strong data engineer who is passionate about large-scale distributed systems and streaming pipelines, and cares about producing clean, elegant, maintainable, robust, well-tested Scala / Spark code.

Core Skills:

? Kafka, Spark Streaming, Avro

? Cassandra (DevOps, Data Modeling)

? Graph Modeling / Algorithms / Queries (with Spark GraphFrames and Neo4J)

? Programming Languages: Scala (highly proficient, 8 years exp.), Python (proficient)

Other Skills: Git, Docker, Akka Streams, Apache Ignite, Parquet, HBase, Zookeeper, HDFS, ElasticSearch, AWS (EC2/AutoScG/S3/RedShift/EMR).

Professional Background (formerly): ETL Developer / Traditional DWHs / Kimball's and Data Vault Methodologies

Educational Background: Computer Science

Solid experience working remotely.

All of my recent work history (8 years) is exclusively with startups.

Profile: https://angel.co/dan-serban

My hourly rate is EUR 80.

E-mail address in the profile.


SEEKING WORK, Data Engineer, Primarily Remote

Well-rounded Scala data engineer with deep knowledge of the internals of distributed datastores. Solid experience working remotely and working with teams that are distributed geographically. I typically work Pacific Time hours.

Core Skills:

? Cassandra (Data Modeling, Troubleshooting Performance And Operational Issues)

? Stream Processing At Scale: Kafka, Flink, Spark Streaming

? Custom-Crafted TF-IDF / Embeddings, Vector-Based Semantic Search, Deep Intent Recognition In Search Engine Queries

? Languages: Scala, Python, SQL (proficient), Rust, Golang (ramping up)

Other Skills: Airflow, DBT, Snowflake, Qdrant, Databricks/DeltaLake, BigQuery, Redshift, Kinesis, PrestoSQL/Trino, Airbyte, Reverse-Engineering Search Engine Technologies.

Educational Background: Computer Science.

More details: https://drive.google.com/file/d/1fYGY5QBpkE1OmTNV7fhJPQyxv8f...

E-mail address in the profile.


SEEKING WORK | Berlin or Remote

Data Engineer

I have experience designing and implementing data processing systems in the Hadoop ecosystem as well as developing algorithms in a telecommunication environment for distributed systems.

Technologies: PySpark, HDFS, Parquet, YARN, SQL, Postgres, Tensorflow, Jupyter, Pandas, gensim, Flask, Docker, AWS, Linux

Programming languages: Python(proficient), Java(familiar), C++(familiar) Languages: English, German

Recent machine learning side project: https://github.com/grbtm/rssBriefing

CV on request

Email in profile

Website: https://grbtm.github.io/


Location: EU-based, 5 hours overlap with PST

Remote: Yes

Willing to relocate: Yes

Technologies: Kafka, Flink, Spark, Cassandra, Druid, Scala, Python, Zookeeper, BigQuery, Redshift, Kinesis, Airflow, DBT, Snowflake, Stitch, FiveTran, Big Data Engineering

Résumé/CV: https://is.gd/dserbanresume202203

Email: profile

I'm pursuing a role in the high-scalability distributed systems space.

I'm a well-rounded Scala data engineer with deep knowledge of the internals of distributed datastores.

Core Skills:

? Cassandra (Data Modeling, Troubleshooting Performance And Operational Issues)

? Druid (Stream Ingestion, Cluster Ops, Data Modeling, Scaling Complex Queries)

? Stream Processing At Scale: Kafka, Flink, Spark Streaming

? Programming Languages: Scala (highly proficient, 8 years exp.), Python (proficient)

Other Skills: Zookeeper, BigQuery, Redshift, Kinesis, Airflow, DBT, Snowflake, Stitch, FiveTran.

Educational Background: Computer Science.

Solid experience working remotely.

next

Legal | privacy