Nutch hbase jobs

Filter

My recent searches
Filter by:
Budget
to
to
to
Type
Skills
Languages
    Job State
    755 nutch hbase jobs found, pricing in INR

    #Your code goes here import '' import '' def jbytes(*args) { |arg| arg.to_s.to_java_bytes } end def put_many(table_name, row, column_values) table = (@, table_name) p = (*jbytes(row)) do |column, value| family, qualifier = (':') (jbytes(family, qualifier), jbytes(value)) end (p) end # Call put_many function with sample data put_many 'wiki', 'DevOps', { "text:" => "What DevOps IaC do you use?", "revision:author" => "Frayad Gebrehana", "revision:comment" => "Terraform" } # Get data from the 'wiki' table get 'wiki', 'DevOps' #Do not remove the exit call below exit

    ₹5004 (Avg Bid)
    ₹5004 Avg Bid
    7 bids

    ...Visualization of JanusGraph with Elasticsearch Integration for Relationship Analysis in Banking" Requirements Analysis: a. Conduct stakeholder interviews to gather system requirements b. Document use cases and user stories c. Define data schema and relationship mapping for JanusGraph d. Assess technical constraints and system integrations Planning and Design: a. Select the datastore (HBase or Cassandra) after analysing performance and scalability b. Define the JanusGraph schema, data model, and query patterns c. Plan data migration strategy and sequence from Elasticsearch to JanusGraph d. Design the algorithm for relationship creation between Main party and Other party e. Evaluate visualization libraries and choose the most appropriate for the Link Analysis cha...

    ₹25019 (Avg Bid)
    ₹25019 Avg Bid
    1 bids

    ...Visualization of JanusGraph with Elasticsearch Integration for Relationship Analysis in Banking" Requirements Analysis: a. Conduct stakeholder interviews to gather system requirements b. Document use cases and user stories c. Define data schema and relationship mapping for JanusGraph d. Assess technical constraints and system integrations Planning and Design: a. Select the datastore (HBase or Cassandra) after analysing performance and scalability b. Define the JanusGraph schema, data model, and query patterns c. Plan data migration strategy and sequence from Elasticsearch to JanusGraph d. Design the algorithm for relationship creation between Main party and Other party e. Evaluate visualization libraries and choose the most appropriate for the Link Analysis cha...

    ₹7005 (Avg Bid)
    ₹7005 Avg Bid
    1 bids

    ...looking for an advanced Hadoop trainer for an online training program. I have some specific topics to be covered as part of the program, and it is essential that the trainer can provide in-depth knowledge and expertise in Hadoop. The topics to be discussed include Big Data technologies, Hadoop administration, Data warehousing, MapReduce, HDFS Architecture, Cluster Management, Real Time Processing, HBase, Apache Sqoop, and Flume. Of course, the trainer should also have good working knowledge about other Big Data topics and techniques. In addition to the topics mentioned, the successful candidate must also demonstrate the ability to tailor the course to meet the learner’s individual needs, making sure that the classes are engaging and fun. The trainer must also possess out...

    ₹1168 / hr (Avg Bid)
    ₹1168 / hr Avg Bid
    1 bids

    I am looking for a freelancer who c...through WebEx meetings. Here are the project requirements: Specific Azure topics: - Azure Networking Assistance type: - Virtual Assistance Preferred meeting type: - WebEx Meeting and AZURE Azure Data Factory (ADE), Azure DataBricks, Azure Data Lake Services (ADLS), Azure Blob Services, Azure SQL DB, Azure Active Directory (AAD), Azure Dev Ops. Languages: Scala, Core Java, Python Databases Hive, Hbase Data Ingestion: Sqoop, Kafka, Spark Streaming Data Visualization:Table and AZURE:ADF Databricks Azure Skills and experience: - Strong understanding of Azure Networking - Experience in providing virtual assistance - Proficiency in conducting WebEx meetings If you have the required skills and experience, please bid on this summary

    ₹40197 (Avg Bid)
    ₹40197 Avg Bid
    4 bids

    I am looking for a freelancer who can develop an Search engine (Apache Nutch 1.0) Crawler system with an integrated AI backend. The project requires the following functionalities: - Optimized crawling and indexing - Advanced crawling and indexing with custom plugins - Crawling, indexing, and AI-driven data analysis - single page search engine The data analysis should be integrated with an existing database system. Additionally, the ideal candidate should have experience in working with Neural Networks as the AI algorithm.

    ₹21985 (Avg Bid)
    ₹21985 Avg Bid
    7 bids

    ...SupportLocation: Austin, TXDuration: 12 Months Job Description: We are looking someone having strong experience in production support, administration and Development experience with Hadoop technologies.• Minimum Experience 8 Years• Must have Hands-on experience on managing Multiple Hortonworks Clusters. Troubleshooting, Maintaining and Monitoring is the key responsibility here.• Must be conversant in HBase, OpenBSD & Grafana related issues handling in order to ensure the data flow is smooth and consistent.• Experience on Kafka for stream processing of Data• Experience in Deployment of new services, patching of hosts etc.• Good hands-on experience on Linux (preferably Red hat) server platform• Should have knowledge in at least one o...

    ₹23518 (Avg Bid)
    ₹23518 Avg Bid
    17 bids

    Hi Maste...needed 4. Ability to bring a vision to life 5. Honesty and realism when it comes to agreed project deadlines 6. Reasonably accessible when needed 7. Available to provide continuous feedback as appropriate Plugins and Algorithms: • WP Web Scraper, Web Scraper Shortcode, Web Scraper, Web Scraper and SEO Tool for web scraping • Scrapy (Python), Beautiful Soup (Python) • Cheerio (JavaScript), Apache Nutch • Heritrix • Application Programming Interfaces (APIs) • Parsehub, Scrapinghub, Octoparse for data extraction • Tableau, Power BI, Looker • AI Chatbot for AI plugin enhancements. • Google Maps API, Google Search API for Application Programming Interfaces (APIs) Note: The above plugins and algorithms are not limited and may or ma...

    ₹4170 (Avg Bid)
    ₹4170 Avg Bid
    1 bids

    ...topics like AWS Azure GCP DigitalOcean Heroku Alibaba Linux Unix Windows Server (Active Directory) MySQL PostgreSQL SQL Server Oracle MongoDB Apache Cassandra Couchbase Neo4J DynamoDB Amazon Redshift Azure Synapse Google BigQuery Snowflake SQL Data Modelling ETL tools (Informatica, SSIS, Talend, Azure Data Factory, etc.) Data Pipelines Hadoop framework services (e.g. HDFS, Sqoop, Pig, Hive, Impala, Hbase, Flume, Zookeeper, etc.) Spark (EMR, Databricks etc.) Tableau PowerBI Artificial Intelligence Machine Learning Natural Language Processing Python C++ C# Java Ruby Golang Node.js JavaScript .NET Swift Android Shell scripting Powershell HTML5 AngularJS ReactJS VueJS Django Flask Git CI/CD (Jenkins, Bamboo, TeamCity, Octopus Deploy) Puppet/Ansible/Chef Docker Kubernetes ECS/EKS Test...

    ₹2835 (Avg Bid)
    ₹2835 Avg Bid
    23 bids

    .../ Define the problem. Create Tables with constraints Design a Schema based on tables and explain the schema. Create primary keys, foreign keys. Create Procedures. Create functions. Create Views Create Index Use of the following Clauses: Example : order by, between, group by, having, order by, AND, OR, with Use Aggregate Functions Use of nested queries, Scalar Subquery. Part 2 has to be done in HBASE Create Tables – 4 tables with Column family and columns Column family - 5 column families: Make sure have different parameter. Ex: versions Minimum 4 Columns in each Column family Insert records Delete records Perform basic queries like your assignment1 Try to extract data using timestamp Insert partial data in a row Describe table. Check table status – enabled or disable...

    ₹12092 (Avg Bid)
    ₹12092 Avg Bid
    33 bids

    .../ Define the problem. Create Tables with constraints Design a Schema based on tables and explain the schema. Create primary keys, foreign keys. Create Procedures. Create functions. Create Views Create Index Use of the following Clauses: Example : order by, between, group by, having, order by, AND, OR, with Use Aggregate Functions Use of nested queries, Scalar Subquery. Part 2 has to be done in HBASE Create Tables – 4 tables with Column family and columns Column family - 5 column families: Make sure have different parameter. Ex: versions Minimum 4 Columns in each Column family Insert records Delete records Perform basic queries like your assignment1 Try to extract data using timestamp Insert partial data in a row Describe table. Check table status – enabled or disable...

    ₹3753 (Avg Bid)
    ₹3753 Avg Bid
    10 bids

    Looking for Flutter ( Dart) developer work for long term project with fixed monthly payment . Required skills : 1: At least 2 app published 2: Dart 3: API 4: Sqlite ,hbase 5: In app purchase integration experience 6: Bluetooth experience

    ₹29522 (Avg Bid)
    ₹29522 Avg Bid
    75 bids

    ...oriented discussion. Must Have: ● At least 6+ years of total IT experience ● At least 4+ years of experience in design and development using Hadoop technology stack and programming languages ● Hands-on experience in 2 or more areas: o Hadoop, HDFS, MR o Spark Streaming, Spark SQL, Spark ML o Kafka/Flume. o Apache NiFi o Worked with Hortonworks Data Platform o Hive / Pig / Sqoop o NoSQL Databases HBase/Cassandra/Neo4j/MongoDB o Visualisation & Reporting frameworks like D3.js, Zeppellin, Grafana, Kibana Tableau, Pentaho o Scrapy for crawling websites o Good to have knowledge of Elastic Search o Good to have understanding of Google Analytics data streaming. o Data security (Kerberos/Open LDAP/Knox/Ranger) ● Should have a very good overview of the current landscape and ability to...

    ₹2335 / hr (Avg Bid)
    ₹2335 / hr Avg Bid
    6 bids

    Data Engineers 6+ yrs : At least 6+ years of total IT experience ● At least 4+ years of experience in design and development using Hadoop technology stack and programming languages ● Hands-on experience in 2 or more areas: o Hadoop, HDFS, MR o Spark Streaming, Spark SQL, Spark ML o Kafka/Flume. o Apache NiFi Worked with Hortonworks Data Platform o Hive / Pig / Sqoop o NoSQL Databases HBase/Cassandra/Neo4j/MongoDB o Visualisation & Reporting frameworks like D3.js, Zeppellin, Grafana, Kibana Tableau, Pentaho o Scrapy for crawling websites o Good to have knowledge of Elastic Search o Good to have understanding of Google Analytics data streaming. o Data security (Kerberos/Open LDAP/Knox/Ranger) ● Should have a very good overview of the current landscape and ability t...

    ₹159787 (Avg Bid)
    ₹159787 Avg Bid
    2 bids

    LDAP service on Oracle Linux 7.3 with rpm packages. LDAP for HDP-2.5.3.0 3. Install and configure Ranger HDP service security on Hbase and Solr tables and collection and validate the security with 5 users

    ₹12760 (Avg Bid)
    ₹12760 Avg Bid
    1 bids

    ...taking advantage of the CI/CD pipelines when possible - Help with troubleshooting and configuration fine-tuning on several platforms (apache, haddoop, hbase etc) - Build and maintain a local testing environment replica for developers. - Help plan for "non hyper cloud" deployments. OpenStack, ProxMox, Kubernetes. All are on the table but the most "appropriate" one must be selected considering the architecture and CI/CD capabilities. - Build and maintain "on prem" alternatives of the AWS structure. This will include hardware planing (server) but also deployment of several VMs (or containers at some point) with techs including php+nginx, hadoop with hbase (and phoenix), sql database (probably mysql) and CEPH object storage. - Be the technical cha...

    ₹1418 / hr (Avg Bid)
    ₹1418 / hr Avg Bid
    17 bids

    The purpose of this project is to develop a working prototype of a network monitoring and reporting Platform that receives network health and status, traffic data from several network infrastructure monitoring sources, and produces an aggr...Platform that receives network health and status, traffic data from several network infrastructure monitoring sources, and produces an aggregate of network status data for processing by a data analytics engine. This prototype will be known as NetWatch. The NetWatch solution will utilize data processing and analytics services via the Hadoop infrastructure, and data reporting features of the Hbase or MYSQL/Datameer tool. The prototype will be used by the Network A&E team to determine its viability as a working engine for network status ...

    ₹614 - ₹1535
    ₹614 - ₹1535
    0 bids

    Please have a look at the below stack. 1. Bash Scripting. 2. Hive 3. Scala Spark 4. HBase and other regular big data technologies.

    ₹42532 (Avg Bid)
    Local
    ₹42532 Avg Bid
    16 bids

    - Backup HBase database on internal infrastructure

    ₹1334 / hr (Avg Bid)
    ₹1334 / hr Avg Bid
    3 bids

    We are looking for a machine learning engineer who must have the following experience: 1. python coding: +7 years of experience 2. Machine Leaning: +5 years of experience (Scikit-Learn, TensorFlow, Caffe, MXNet, Keras, XGBoost) 3. AI/Deep Learning: +5 years of experience 4. Cloud computing: AWS, S3, EC2, EMR, SageMaker, ECS, Lambda, IAM 5. distributed computing technology: Hadoop, Spark, HBase, Hive / Impala, or any similar technology Should be an independent developer, NO CONSULTING COMPANY There will be series of technical interview about python coding, machine learning, AI , cloud computing. Candidate must have an excellent skill in python coding and be able to answer challenging python questions during the interview

    ₹4587 / hr (Avg Bid)
    ₹4587 / hr Avg Bid
    13 bids

    Design, code, test Hive, Sqoop, HBase, Yarn, UNIX Shell scripting Spark and Scala mandatory You should have working experience in previous projects not a beginner level projects so please be ready to design develop and fix the bugs. Working hours and all We can decide over the chat.

    ₹4587 / hr (Avg Bid)
    ₹4587 / hr Avg Bid
    4 bids

    am trying to run hbase backup command and got below error root@machine:~/hbase-2.4.12# hbase backup Error: Could not find or load main class backup Caused by: : backup need to fix that some tips below : Hbase install below Just enable the configuration on xml file and start the hbase and confirm is working well run Hbase on linux Ubuntu some helps below:

    ₹1084 / hr (Avg Bid)
    ₹1084 / hr Avg Bid
    3 bids

    moving data from wkc to atlas. There is an issue in one of the category relationship mapping

    ₹7089 (Avg Bid)
    ₹7089 Avg Bid
    3 bids

    Roles And R...high-performance web services for data tracking. High-speed querying. Managing and deploying HBase. Being a part of a POC effort to help build new Hadoop clusters. Test prototypes and oversee handover to operational teams. Propose best practices/standards. Skills Required: Good knowledge in back-end programming, specifically java, JS, Node.js and OOAD Good knowledge of database structures, theories, principles, and practices. Ability to write Pig Latin scripts. Hands on experience in HiveQL. Familiarity with data loading tools like Flume, Sqoop. Knowledge of workflow/schedulers like Oozie. Analytical and problem solving skills, applied to Big Data domain Proven understanding with Hadoop, HBase, Hive, Pig, and HBase. Good aptitude in multi-threading and...

    ₹917 / hr (Avg Bid)
    ₹917 / hr Avg Bid
    1 bids

    Hi Tapasi K., I noticed your profile and would like to offer you my project. Write a spark submit job that accesses a data in a hive table in one hadoop/spark cluster , accesses data in an hbase table in another hadoop cluster , combine( do some aggregation) this data and save result in both hive and hbase. P.S. Hive is in a different hadoop cluster than hbase ( both in same network / VPC subnet ) .

    ₹8340 (Avg Bid)
    ₹8340 Avg Bid
    1 bids

    We need to hire a Hadoop and Spark expert. Tasks to be done: - Configure properly Hadoop cluster in HA mode - Configure properly Spark cluster in HA mode - Install and configure HBase - Install and configure Oozie - Install and configure SSL for all the tools mentioned above. - Configure authentication for all the tools mentioned above. Installation will be done in an on-premise environment. Linux based OS (centos 9) will be used. All the Hadoop and Spark software will be the full open source version. We are not using Cloudera, Hortonworks, MapR or similars. Project will be payed by an hourly rate for the amount of time it takes to finish the tasks mentioned above. Only tech folks with experience will be considered! :)

    ₹4754 / hr (Avg Bid)
    ₹4754 / hr Avg Bid
    2 bids

    - Existing infrastructure needs to be backed up with Ansible - Should have knowledge of the following technologies - Ansible - Terraform - Docker - Kubernetes - Postgres - HBase - Gitlab

    ₹917 / hr (Avg Bid)
    ₹917 / hr Avg Bid
    4 bids

    ...full product life-cycles • Coding skills in JavaScript with a strong base in object-oriented design and functional programming • Strong Experience in Node.Js and React.Js web framework • Understanding of basic data structures & algorithms • Experienced with relational databases(MySQL, Postgres, etc) good working knowledge of SQL Experience with non-relational databases (MongoDb, Cassandra, Hbase, DynamoDb), designing schemas • Experience in API design and best practices • Experience in building microservices-based architectures • Strong experience on any of frameworks such as Express, Koa, Sails, StrongLoop etc. • Web fundamentals like HTML5 and CSS3 • Good design and prototyping skills • Ability to technically l...

    ₹126929 (Avg Bid)
    ₹126929 Avg Bid
    9 bids

    Design, code, test Hive, Sqoop, HBase, Yarn, UNIX Shell scripting Spark and Scala mandatory You should have experience in previous projects not a beginner level projects so please be ready to design develop and fix the bugs. Working hours and all We can decide over the chat.

    ₹1084 / hr (Avg Bid)
    ₹1084 / hr Avg Bid
    3 bids

    ...Docker, Kubernetes, CI/CD GitLab, GitHub, JFrog Artifactory, Docker, Kubernetes, RESTful API, HEAT, Tosca, YAML Validation: Jenkins, Gherkin, Cucumber, Ruby, HP Quality Center Cloud, Python, Docker Developer: (Must have): Python, Pandas, Pytest, CI/CD, Jira, Confluence, GitHub, (Good to have): Python Flask, Dash, DevOps, Big Data architecture, SCRUM, SonarQube (Good to have): Docker, Hadoop, HBase, Kafka, NiFi, Camunda, Databricks, Cloudera, Kubernetes. Engineer capable of designing solutions, writing code, testing code, automating testing and deployment Proven skills, knowledge and experience with statistical computer languages. (Python, etc.) and associated ecosystem (jupyterlab, jupyter notebook, ...), Good knowledge of CI/CD and automated testing using Python Cloud, P...

    ₹128179 (Avg Bid)
    ₹128179 Avg Bid
    11 bids

    Need someone who have experience with Big Data Technology 1. Spark 2. Hadoop 3. HBase 4. Kafka 5. Zookeeper

    ₹1418 (Avg Bid)
    ₹1418 Avg Bid
    7 bids

    Need someone who have experience with Big Data Technology 1. Spark 2. Hadoop 3. HBase 4. Kafka 5. Zookeeper

    ₹1835 (Avg Bid)
    ₹1835 Avg Bid
    3 bids

    Column family databases are best known because of Google’s BigTable implementation. They are very similar on the surface to relational databases, but they have critical conceptual differences. You will not be able to apply the same sort of solutions that you used in a relational database to a column database.

    ₹834 - ₹2502
    ₹834 - ₹2502
    0 bids

    ...Cross Region Replication 8. DynamoDB Performance and Partition Key Selection 9. Snowball and AWS Big Data 10. AWS DMS 11. AWS Aurora in Big Data 12. Key Takeaways Module 5 - AWS Big Data Processing Services 1. Learning Objective 2. Amazon EMR 3. Apache Hadoop 4. EMR Architecture 5. EMR Releases and Cluster 6. Choosing Instance and Monitoring 7. Demo - Advance EMR Setting Options 8. Hive on EMR 9. HBase with EMR 10. Presto with EMR 11. Spark with EMR 12. EMR File Storage 13. AWS Lambda 14. Key Takeaways Module 6 - Analysis 1. Learning Objective 2. Redshift Intro and Use cases 3. Redshift Architecture 4. MPP and Redshift in AWS Eco-System 5. Columnar Databases 6. Redshift Table Design - Part 2 7. Demo - Redshift Maintenance and Operations 8. Machine Learning Introduction 9. Machine...

    ₹17847 (Avg Bid)
    ₹17847 Avg Bid
    3 bids

    ...Cross Region Replication 8. DynamoDB Performance and Partition Key Selection 9. Snowball and AWS Big Data 10. AWS DMS 11. AWS Aurora in Big Data 12. Key Takeaways Module 5 - AWS Big Data Processing Services 1. Learning Objective 2. Amazon EMR 3. Apache Hadoop 4. EMR Architecture 5. EMR Releases and Cluster 6. Choosing Instance and Monitoring 7. Demo - Advance EMR Setting Options 8. Hive on EMR 9. HBase with EMR 10. Presto with EMR 11. Spark with EMR 12. EMR File Storage 13. AWS Lambda 14. Key Takeaways Module 6 - Analysis 1. Learning Objective 2. Redshift Intro and Use cases 3. Redshift Architecture 4. MPP and Redshift in AWS Eco-System 5. Columnar Databases 6. Redshift Table Design - Part 2 7. Demo - Redshift Maintenance and Operations 8. Machine Learning Introduction 9. Machine...

    ₹11759 (Avg Bid)
    ₹11759 Avg Bid
    6 bids

    ...like GCP, AWS. • Experience developing/deploying ML solutions in a public cloud (AWS/Azure/Google Cloud) platforms • • Has a Bachelor's degree or equivalent real-world experience, preferably in a related field (Engineering, Computer Science, Statistics, Applied Math) • Is always willing to learn and apply new techniques where appropriate • Distributed computing platforms, such as Hadoop (Hive, HBase, Pig), Spark, GraphLab • Databases (traditional and noSQL) • Bonus Points if you have: • Experience with a mass-market consumer-facing product • Familiarity with auto-differentiation libraries (e.g., TensorFlow, PyTorch, etc.) • Proficiency with the pydata stack (NumPy, scipy, pandas) • Familiarity with scheduling and orch...

    ₹192061 (Avg Bid)
    ₹192061 Avg Bid
    26 bids

    • At least 5 years of consu...Catalog, Cosmo Db, ML Studio, AI/ML, Azure Functions, ARM Templates, Azure DevOps, CI/CD etc. • Cloud migration methodologies and processes including tools like Azure Data Factory, Event Hub, etc. • Familiarity with the Technology stack available in the industry for data management, data ingestion, capture, processing, and curation: Kafka, StreamSets, Attunity, GoldenGate, Map Reduce, Hadoop, Hive, Hbase, Cassandra, Spark, Flume, Hive, Impala, mongoDB, PostgreSQL etc. • Bachelors or higher degree in Computer Science or a related discipline. • Experience in clinical domain is preferred. Budget is 550 to 750$ / a Month . Work is 10 Hours a day and 6 days a week . 1 day will be off contract will be 3 Months and then it will get r...

    ₹50705 (Avg Bid)
    ₹50705 Avg Bid
    3 bids

    • At least 5 years of consu...Catalog, Cosmo Db, ML Studio, AI/ML, Azure Functions, ARM Templates, Azure DevOps, CI/CD etc. • Cloud migration methodologies and processes including tools like Azure Data Factory, Event Hub, etc. • Familiarity with the Technology stack available in the industry for data management, data ingestion, capture, processing, and curation: Kafka, StreamSets, Attunity, GoldenGate, Map Reduce, Hadoop, Hive, Hbase, Cassandra, Spark, Flume, Hive, Impala, mongoDB, PostgreSQL etc. • Bachelors or higher degree in Computer Science or a related discipline. • Experience in clinical domain is preferred. Budget is 550 to 750$ / a Month . Work is 10 Hours a day and 6 days a week . 1 day will be off contract will be 3 Months and then it will get r...

    ₹20849 - ₹62547
    Sealed NDA
    ₹20849 - ₹62547
    3 bids

    Hi all, Looking for support on below skillset Spark with Java Spark core, spark sql, spark streaming Hive tables Hbase Kafka Shell scripts Oozie workflows Language: Java Let me know if you have experience in it

    ₹12760 (Avg Bid)
    ₹12760 Avg Bid
    1 bids

    Responsibilities • Lead projects end-to-end. Meet with business users to determine requirements, analyze the data lake for relevant datasets, collaborate with other developers to design a technical solution, and see the project through to completion. • Design and build analytical workflows that take da...new technologies, innovate and deliver solution for business critical applications Preferred Skills • SDLC Methodology - Agile / Scrum / Iterative Development • Job Scheduling Tools (Autosys) • Version Control System (Git, Bitbucket) • Continuous Integration / Continuous Delivery (CI/CD) pipelines (Jenkins) • Real Time Streaming (Kafka) • Visual Analytics Tools (Tableau) • No SQL Technologies (Hbase) Timings- 2-3 hours/weekdays Experienc...

    ₹25519 (Avg Bid)
    ₹25519 Avg Bid
    2 bids

    Responsibilities • Lead projects end-to-end. Meet with business users to determine requirements, analyze the data lake for relevant datasets, collaborate with other developers to design a technical solution, and see the project through to completion. • Design and build analytical workflows that take dat...new technologies, innovate and deliver solution for business critical applications Preferred Skills • SDLC Methodology - Agile / Scrum / Iterative Development • Job Scheduling Tools (Autosys) • Version Control System (Git, Bitbucket) • Continuous Integration / Continuous Delivery (CI/CD) pipelines (Jenkins) • Real Time Streaming (Kafka) • Visual Analytics Tools (Tableau) • No SQL Technologies (Hbase Timings- 2-3 hours/weekdays Experienc...

    ₹41698 (Avg Bid)
    ₹41698 Avg Bid
    4 bids

    I need crawled results from (APACHE NUTCH) connected to website search software (MANTICORE SEARCH) Manticore will index those results to be searchable by a user search box. The link provided will guide on how to use and connect folder for Manticore to use. Coder will need to download and work with the two softwares in order to fully understand how everything works in his own home server. If coder not able to connect by using tutorial videos he can create his own plugin that gives the same results im looking for. If coder creates plugin the code must be provided for other coders to use on my server. Any lines of coder used from Manticore to connect softwares must be provided to be used by other coders and future projects Resources Apache nutch- website scraper

    ₹38529 (Avg Bid)
    ₹38529 Avg Bid
    12 bids

    implementation a red black tree algorithm to hbase

    ₹1251 (Avg Bid)
    ₹1251 Avg Bid
    2 bids

    implementation algorithm to hbase to test the search process and calculate the time complexity

    ₹834 - ₹2502
    ₹834 - ₹2502
    0 bids

    Implementation a red black tree algorithm to hbase to test the search process and calculate the time complexity

    ₹1835 (Avg Bid)
    ₹1835 Avg Bid
    2 bids

    implementation a red black tree algorithm to hbase

    ₹2085 (Avg Bid)
    ₹2085 Avg Bid
    1 bids

    implementing a Red Black Tree algorithm on hbase to test the search process

    ₹417 / hr (Avg Bid)
    ₹417 / hr Avg Bid
    1 bids

    Responsibilities • Design, code, test and debug new features and enhancement over Big Data platform following best practices of agile software development Skills Required: • 6+ years in Java /J2EE skills • 3+ years in AWS • Min 3+ Yrs experience in using Eclipse, Spring Framework, STRUTS, Web services • Working experience in DevOps Environment • Working knowledge of AWS cloud, Voltdb, Hbase • Working knowledge of XML, HTTP, REST, SOAP, JSON, AJAX • Working knowledge in Hibernate, Hadoop, XST, MySQL • Basic Knowledge in Linux Shell Commands and Scripting

    ₹29856 (Avg Bid)
    ₹29856 Avg Bid
    2 bids

    SKILL SET REQUIRED:- Spark, Strong datawarehousing Skills, Python, Hive, Hbase EXPERIENCE REQUIRED:- 4-6 years CANDIDATES MUST DEDICATE 8-9 HOURS PER DAY. Interested people can Bid. Further details will be provided through chat,

    ₹58544 (Avg Bid)
    ₹58544 Avg Bid
    6 bids

    Expertise in designing and deployment of Hadoop Cluster and different analytical tools including Pig, Hive, HBase, Sqoop, Kafka Spark with Cloudera distribution. Working on a live 20 nodes Hadoop cluster running on CDH4.4. Working with highly unstructured and semi structured data of 40 TB in size (120 TB with replication factor of 3) Managing external tables in Hive for optimized performance. Very good understanding of Partitions and Bucketing in Hive Developed Spark scripts using Scala as per the requirement using Spark 1.5 framework. Using Spark API’s over Cloudera Hadoop Yarn to perform analytics on data used for Hive stored at HDFS. Developed Scala Scripts, UDFs using both Data frames/SQL and RDD in Spark for data aggregation, queries and writing data back onto HDFS. Explo...

    ₹1084 / hr (Avg Bid)
    ₹1084 / hr Avg Bid
    6 bids