Message Queue

 

What is Message Passing?
Message passing is a technique to enable inter-process communication (IPC), or for inter-thread communication within the same process communication between two distributed or non-distributed parallel processes in synchronous or asynchronous mode, The communications are completed by the sending of messages (functions, signals and data packets) to recipients.
Continue reading Message Queue

Algorithm

What is an Algorithm?

Since there is no universally agreed-on wording to describe this notion, there is general agreement about what the concept means:

An algorithm is a sequence of unambiguous instructions for obtaining a required output for any legitimate input in a finite amount of time
Continue reading Algorithm

How to dump data from Redshift to JSON

Extracting data from redshift is nearly as common as getting data in. Sometimes, the results of hard computations done in Redshift are necessary for serving systems. Other times, a large export is needed for analysis in Excel or other tools. UNLOAD command can be used to extract data from redshift to s3 in various formates like Delimited or fixed-width formate. Continue reading How to dump data from Redshift to JSON

What is Machine Learning?

What is Machine Learning?

Machine learning is a type of artificial intelligence (AI) that provides computers with the ability to learn without being explicitly programmed. Machine learning is a part of computer science which focuses on the development of computer programs that can teach themselves to grow and change based on the data it is exposed to. Continue reading What is Machine Learning?

Normalisation

Normalisation is the process of eliminating the redundancy, minimising the use of null values and prevention of the loss of information by establishing relations and ensuring data integrity.

Data should only be stored once and avoid storing data that can be calculated from other data already held in the database. During the process of normalisation redundancy must be removed, but not at the expense of breaking data integrity rules.

Continue reading Normalisation

MapReduce

MapReduce is a framework for processing large amount of data residing on hundreds of computers, its an extraordinarily powerful paradigm. MapReduce was first introduced by Google in 2004 MapReduce: Simplified Data Processing on Large Clusters.

In this article we’ll see how MapReduce processes the data, I am considering the Word Count program as a example, yeah!! this is the worlds most famous MapReduce program!!

Continue reading MapReduce