About A PhD Candidate student who is doing research in the field of Data-Intensive Scalable Computing within the EU BigFoot project (http://bigfootproject.eu). Primarily work with OpenSource Frameworks such as:
- Apache Hadoop: a framework that allows for the distributed processing of large data sets across clusters of computers (several to thousands of machines) using MapReduce programming models.
- Apache Pig: a platform for analyzing large data sets that consists of Pig-Latin high-level language for expressing data analysis programs, coupled with infrastructure for executing these programs on framework like Apache Hadoop. (http://pig.apache.org)
- Apache HBase: the Hadoop database, a distri...