Amaltas
  • Consulting
  • Training
  • About

bigdata

HDFS Architecture

HDFS Architecture

HDFS Architecture Diagram Data Replication How does HDFS manage to provide data reliability and high fault tolerance when it uses commodity (and therefore cheap and prone to failure) hardware ? HDFS does it using a very simple idea. 1. Break a large file into series of chunks or blocks (each block
Anay Tamhankar Nov 14, 2023
Hadoop Distributed File System (HDFS)

Hadoop Distributed File System (HDFS)

What is HDFS ? HDFS is a distributed file system designed to run on commodity hardware. HDFS is highly fault tolerant and provides high throughput access to "application data". It is suitable for applications which have large datasets. Distributed File Systems are file systems that manage storage across a network of
Anay Tamhankar Nov 13, 2023
Introduction to Hadoop

Introduction to Hadoop

What is Hadoop ? Hadoop is a collection of open source software utilities that facilitate solving big data problems using a network of computers. Hadoop comprises of a software framework for distributed storage (HDFS) and a software framework for query and analysis (Map Reduce). Design Principal 💡Hardware failure are common occurrences
Anay Tamhankar Nov 10, 2023
Big Data - Challenges and Opportunities

Big Data - Challenges and Opportunities

From a technical standpoint, Big Data can be defined as data that is so large, complex and fast changing that none of the traditional data management tools are able to store and process it efficiently. Therefore the term "big data" or "big data technologies" refers to the technologies that can
Anay Tamhankar Sep 8, 2023
Amaltas Technologies LLP © 2025.