International Journal of Advance Research In Science And Engineering http://www.ijarse.com IJARSE, Vol. No.4, Issue 03, March 2015 ISSN-2319-8354(E) 61 | Page COMPREHENSIVE AND COORDINATED SECURITY OF KNOX GATEWAY IN BIG DATA M.Chithik Raja 1 , M. Munir Ahamed Rabbani 2 1 Research Scholar , AMET University , (India) 2 Supervisor, B. S. Abdur Rahman University , ( India) ABSTRACT Adding security to Hadoop is challenging because all the interactions do not follow the classic client-server pattern: the file system is partitioned and distributed requiring authorization checks at multiple points; a submitted batch job is executed at a later time on nodes different from the node on which the client authenticated and submitted the job; job tasks from different users are executed on the same compute node; secondary services such as a workflow system access Hadoop on behalf of users; and the system scales to thousands of servers and tens of thousands of concurrent tasks. To address these challenges, the base Kerberos authentication mechanism is supplemented by delegation and capability-like access tokens and the notion of trust for secondary services.In order to secure a Hadoop cluster all you needed was a firewall that restricted network access to only authorized users. This eventually evolved into a more robust security layer in Hadoop layer that could augment firewall access with strong authentication. The enterprise has placed requirements on the platform to not only provide perimeter security, but to also integrate with all types of authentication mechanisms and all the while, be easy to manage and to integrate with the rest of the secured corporate infrastructure. Kerberos can still be a great provider of the core security technology but with all the touch-points that a user will have with Hadoop, something more is needed. Keywords: Git Hub EDW, GFS, HBASE, GCE, HDFS I INTRODUCTION Hadoop comes from open source, and there’s no shortage of open source Apache projects aimed at building security functionality into the core Hadoop stack. A quick list would include Apache Knox for authentication, Apache Falcon for data governance, and Apache Sentry, for role-based authorization. We take Kerberos for authentication, while hart on network gave us open source Project Rhino, a GitHub-hosted effort to develop a Hadoop framework for encryption, key management, and authorization.But stitching these components together in a comprehensive and cohesive way is not so easy. Apache Sentry, for example, requires working with XML files, which is not something everybody will feel comfortable with. Many of these open source projects are quite green still and lack polish.This