CSIC 2014 ( November ) - Computer Society Of India

` 50/ISSN 0970-647X | Volume No. 38 | Issue No. 8 | November 2014
www.csi-india.org
Cover Story
Visualization-Techniques,
Methods and Tools 7
Article
Information Technology to
Curb Piracy in Bollywood 23
Technical Trends
Big Data Visualization using
Cassandra and R 15
IT Industry Perspective
Future of Medical Transcription Jobs in
India - Need to Extend it beyond Record
Generating Process 29
Research Front
A Walkthrough & Pathfinder for Research
Novitiates: Google Scholar Vs Microsoft
Academic Search 18
Security Corner
A Case Study of Netrapur Police 39
CSI Communications | November 2014 | 1
CSI Job Specific Certifications
Computer Society of India enters into Knowledge Partnership with
InfoComm International, USA
In its drive towards empowering IT students beyond regular diplomas and degrees, CSI has launched professional certification programme
to certify them for direct employment in niche domains. As the members are aware, CSI has already entered into collaboration with the Open
Group for TOGAF Foundations Certificate to students in the graduate engineering stream. InfoComm certification is the second in line.
InfoComm International of USA is a nonprofit association serving the professional audio video communication industry worldwide.
InfoComm University education supported by CSI would offer those Indian students from the IT stream, pursuing diplomas as well as
degree, the opportunity to qualify as “AV Essential Technology Specialist”.
Initially, a pilot programme that would open around 200 vacancies for successful certified students will be made available by InfoComm.
The course would consist of:
•
Quick start to the AV Industry,
•
The essentials of audio visual technology,
•
Certification preparation, and,
•
Recognized Audio Visual Technologist Certificate.
The delivery of the pilot programme will be via the InfoComm University online learning platform supported by CSI and their partner
Institutions.
Each year, InfoComm International will consolidate the vacancies arising among its member organizations in India and will recruit competent
certified students through the CSI Educational Directorate.
All CSI member educational institutions are invited to be part of the partnership programme, which offers placement opportunities
to the young professionals. Interested institutions may please contact the Director (Education), Computer Society of India, Education
Directorate, CIT Campus, 4th Cross Road, Taramani, Chennai - 600 113.
Phone : +91-44-2254 1102/1103/2874
Email : director.edu@csi-india.org
Institutions in the Mumbai Region may also contact Computer Society of India, Headquarters, Samruddhi Venture Park, Unit No.3, 4th Floor,
MIDC, Andheri (E). Mumbai-400093 Maharashtra,
India. Phone: 91-22-29261700, Fax: 91-22-28302133
Email: hq@csi-india.org
www.csi-india.org
CSI Communications
Contents
Volume No. 38 • Issue No. 8 • November 2014
Editorial Board
Chief Editor
Dr. R M Sonar
Editors
Dr. Debasish Jana
Dr. Achuthsankar Nair
Resident Editor
Mrs. Jayshree Dhere
Published by
Executive Secretary
Mr. Suchit Gogwekar
For Computer Society of India
Design, Print and
Dispatch by
CyberMedia Services Limited
7
10
Cover Story
Visualization-Techniques, Methods and
Tools
Dr. KVSN Rama Rao, Mr. Surya Putchala and
Midhun Thaduru
Visualization for STEM Subjects
A.B. Karthick Anand Babu, D. Maghesh Kumar
and G. RajaRaja Cholan
Technical Trends
11
12
Visualization Tool for Data Mining
15
Big Data Visualization using Cassandra
and R
16
Visualization Methods for Vector Fields:
An Insight
18
A Walkthrough & Pathfinder for Research
Novitiates: Google Scholar Vs Microsoft
Academic Search
Dr. B. Eswara Reddy and Mr. K. Rajendra Prasad
Information Visualization in Gene
Expression Data
Sreeja Ashok, Dr. M. V. Judy and N. Thushara
Vijayakumar
Rishav Singh and Dr. Sanjay Kumar Singh
22
27
Dilip Kumar Dalei, B. V. Hari Krishna Nanda
and N. Venkataramanan
28
29
Amol Dhumane and Dr. Rajesh Prasad
PLUS
Protection of Software as Intellectual
Property
Dr. M Hanumanthappa, Mrs. S Regina
Lourdhu Suganthi and Mrs. Rashmi S
Practitioner Workbench
Programming.Tips() »
Fun with ‘C’ Programs – Reversing a
String using a Bitwise Operator
Programming.Learn(“R”) »
RStudio- Studio of R
Umesh P and Silpa Bhaskaran
IT Industry Perspective
Future of Medical Transcription Jobs
in India - Need to Extend it beyond
Record Generating Process
Prof (Dr.) D G Jha
36
37
Articles
Active Queue Management
Sumith Kumar Puri and Dr. H K Anasuya Devi
Wallace Jacob
Research Front
Anchal Garg, Madhurima, Madhulika and Saru Dhir
Please note:
CSI Communications is published by Computer
Society of India, a non-profit organization.
Views and opinions expressed in the CSI
Communications are those of individual authors,
contributors and advertisers and they may
differ from policies and official statements of
CSI. These should not be construed as legal or
professional advice. The CSI, the publisher, the
editors and the contributors are not responsible
for any decisions taken by readers on the basis of
these views and opinions.
Although every care is being taken to ensure
genuineness of the writings in this publication,
CSI Communications does not attest to the
originality of the respective authors’ content.
© 2012 CSI. All rights reserved.
Instructors are permitted to photocopy isolated
articles for non-commercial classroom use
without fee. For any other copying, reprint or
republication, permission must be obtained
in writing from the Society. Copying for other
than personal use or internal reference, or of
articles or columns not owned by the Society
without explicit permission of the Society or the
copyright owner is strictly prohibited.
23
25
Information Technology to Curb
Piracy in Bollywood
Innovations in India
Software User Experience Maturity
Model
Rajiv Thanawala and Prachi Sakhardande
Security Corner
Information Security »
A Quick Look at Virtual Private
Database Security
Jignesh Doshi and Bhushan Trivedi
39
Case Studies in IT Governance, IT Risk
and Information Security »
A Case Study of Netrapur Police
Dr. Vishnu Kanhere
Brain Teaser
42
Dr. Debasish Jana
Ask an Expert
43
Dr. Debasish Jana
Happenings@ICT
44
H R Mohan
On the Shelf!
Peeyush Chomal
45
CSI Report
46
CSI News
48
Published by Suchit Gogwekar for Computer Society of India at Unit No. 3, 4th Floor, Samruddhi Venture Park, MIDC, Andheri (E), Mumbai-400 093.
Tel. : 022-2926 1700 • Fax : 022-2830 2133 • Email : hq@csi-india.org Printed at GP Offset Pvt. Ltd., Mumbai 400 059.
CSI Communications | November 2014 | 3
Know Your CSI
Executive Committee (2013-14/15)
President
Mr. H R Mohan
president@csi-india.org
»
Vice-President
Prof. Bipin V Mehta
vp@csi-india.org
Hon. Secretary
Mr. Sanjay Mohapatra
secretary@csi-india.org
Hon. Treasurer
Mr. Ranga Rajagopal
treasurer@csi-india.org
Immd. Past President
Prof. S V Raghavan
ipp@csi-india.org
Nomination Committee (2014-2015)
Prof. P. Kalyanaraman
Mr. Sanjeev Kumar
Mr. Subimal Kundu
Region - I
Mr. R K Vyas
Delhi, Punjab, Haryana, Himachal
Pradesh, Jammu & Kashmir,
Uttar Pradesh, Uttaranchal and
other areas in Northern India.
rvp1@csi-india.org
Region - II
Mr. Devaprasanna Sinha
Assam, Bihar, West Bengal,
North Eastern States
and other areas in
East & North East India
rvp2@csi-india.org
Region - III
Prof. R P Soni
Gujarat, Madhya Pradesh,
Rajasthan and other areas
in Western India
rvp3@csi-india.org
Region - V
Mr. Raju L kanchibhotla
Karnataka and Andhra Pradesh
rvp5@csi-india.org
Region - VI
Dr. Shirish S Sane
Maharashtra and Goa
rvp6@csi-india.org
Region - VII
Mr. S P Soman
Tamil Nadu, Pondicherry,
Andaman and Nicobar,
Kerala, Lakshadweep
rvp7@csi-india.org
Division-I : Hardware (2013-15)
Prof. M N Hoda
div1@csi-india.org
Division-II : Software (2014-16)
Dr. R Nadarajan
div2@csi-india.org
Division-III : Applications (2013-15)
Dr. A K Nayak
div3@csi-india.org
Division-IV : Communications
(2014-16)
Dr. Durgesh Kumar Mishra
div4@csi-india.org
Division-V : Education and Research
(2013-15)
Dr. Anirban Basu
div5@csi-india.org
Regional Vice-Presidents
Division Chairpersons
Region - IV
Mr. Hari Shankar Mishra
Jharkhand, Chattisgarh,
Orissa and other areas in
Central & South
Eastern India
rvp4@csi-india.org
Publication Committee (2014-15)
Dr. S S Agrawal
Prof. R K Shyamasundar
Prof. R M Sonar
Dr. Debasish Jana
Dr. Achuthsankar Nair
Dr. Anirban Basu
Prof. A K Saini
Prof. M N Hoda
Dr. R Nadarajan
Dr. A K Nayak
Dr. Durgesh Kumar Mishra
Mrs. Jayshree Dhere
Chairman
Member
Member
Member
Member
Member
Member
Member
Member
Member
Member
Member
Important links on CSI website »
About CSI
Structure and Orgnisation
Executive Committee
Nomination Committee
Statutory Committees
Who's Who
CSI Fellows
National, Regional & State
Student Coordinators
Collaborations
Distinguished Speakers
Divisions
Regions
Chapters
Policy Guidelines
Student Branches
Membership Services
Upcoming Events
Publications
Student's Corner
CSI Awards
CSI Certification
Upcoming Webinars
About Membership
Why Join CSI
Membership Benefits
BABA Scheme
Special Interest Groups
http://www.csi-india.org/about-csi
http://www.csi-india.org/web/guest/structureandorganisation
http://www.csi-india.org/executive-committee
http://www.csi-india.org/web/guest/nominations-committee
http://www.csi-india.org/web/guest/statutory-committees
http://www.csi-india.org/web/guest/who-s-who
http://www.csi-india.org/web/guest/csi-fellows
http://www.csi-india.org/web/guest/104
http://www.csi-india.org/web/guest/collaborations
http://www.csi-india.org/distinguished-speakers
http://www.csi-india.org/web/guest/divisions
http://www.csi-india.org/web/guest/regions1
http://www.csi-india.org/web/guest/chapters
http://www.csi-india.org/web/guest/policy-guidelines
http://www.csi-india.org/web/guest/student-branches
http://www.csi-india.org/web/guest/membership-service
http://www.csi-india.org/web/guest/upcoming-events
http://www.csi-india.org/web/guest/publications
http://www.csi-india.org/web/education-directorate/student-s-corner
http://www.csi-india.org/web/guest/csi-awards
http://www.csi-india.org/web/guest/csi-certification
http://www.csi-india.org/web/guest/upcoming-webinars
http://www.csi-india.org/web/guest/about-membership
http://www.csi-india.org/why-join-csi
http://www.csi-india.org/membership-benefits
http://www.csi-india.org/membership-schemes-baba-scheme
http://www.csi-india.org/special-interest-groups
Membership Subscription Fees
Membership and Grades
Institutional Membership
Become a member
Upgrading and Renewing Membership
Download Forms
Membership Eligibility
Code of Ethics
From the President Desk
CSI Communications (PDF Version)
CSI Communications (HTML Version)
CSI Journal of Computing
CSI eNewsletter
CSIC Chapters SBs News
Education Directorate
National Students Coordinator
Awards and Honors
eGovernance Awards
IT Excellence Awards
YITP Awards
CSI Service Awards
Academic Excellence Awards
Contact us
http://www.csi-india.org/fee-structure
http://www.csi-india.org/web/guest/174
http://www.csi-india.org /web/guest/institiutionalmembership
http://www.csi-india.org/web/guest/become-a-member
http://www.csi-india.org/web/guest/183
http://www.csi-india.org/web/guest/downloadforms
http://www.csi-india.org/web/guest/membership-eligibility
http://www.csi-india.org/web/guest/code-of-ethics
http://www.csi-india.org/web/guest/president-s-desk
http://www.csi-india.org/web/guest/csi-communications
http://www.csi-india.org/web/guest/csi-communicationshtml-version
http://www.csi-india.org/web/guest/journal
http://www.csi-india.org/web/guest/enewsletter
http://www.csi-india.org/csic-chapters-sbs-news
http://www.csi-india.org/web/education-directorate/home
http://www.csi-india.org /web/national-studentscoordinators/home
http://www.csi-india.org/web/guest/251
http://www.csi-india.org/web/guest/e-governanceawards
http://www.csi-india.org/web/guest/csiitexcellenceawards
http://www.csi-india.org/web/guest/csiyitp-awards
http://www.csi-india.org/web/guest/csi-service-awards
http://www.csi-india.org/web/guest/academic-excellenceawards
http://www.csi-india.org/web/guest/contact-us
Important Contact Details »
For queries, correspondence regarding Membership, contact helpdesk@csi-india.org
CSI Communications | November 2014 | 4
www.csi-india.org
President’s Message
H R Mohan
From
: President’s Desk:: president@csi-india.org
Subject : President's Message
Date
: 1st November, 2014
Dear Members
On behalf of the CSI Execom 2014-15, I am pleased to invite you to
CSI-2014, the 49th Annual Convention on the theme “Emerging ICT for
Bridging Future”. It will be held as a part of CSI@50 during 12-14, Dec 2014
at Jawaharlal Nehru Technological University (JNTU), Hyderabad. CSI2014 is hosted by CSI Hyderabad chapter in association with JNTU and
Defense Research Development Organization (DRDO). The Convention
is spearheaded by Mr. JA Chowdary (OC), Dr. A Govardhan (PC) and
Mr. Gautam Mahapatra (FC) and a dedicated team of CSI members,
academic and industry professionals from Hyderabad. The arrangements for
CSI-2014 are in full swing. The team is putting up an excellent programme
which is under its final stage and will soon be available at the convention
website www.csi-2014.org. I am sure that the sessions by a wide range of
eminent speakers on a variety of topics will be a feast to all participants. Prior
to the CSI-2014, the Annual Students Convention will be held on 10-11, Dec
2014 at GNIT campus. Under the theme “Campus to Corporate and Beyond”,
the programme will focus on career opportunities and enhancements. It is
hosted by Guru Nanak Institutions under the chairmanship of Dr. HS Saini.
I look forward to active participation by the members in CSI-2014 and
student members in the student convention. As a heritage city, Hyderabad
has a lot to offer first-time visitors. Please plan your trip well in advance and
have a great time at CSI-2014. We will be delighted to see you at Hyderabad.
I was happy to be among the CSI veterans at Pune who had organized
a meet as part of CSI@50 events along with the International Conference
on Advances in Cloud Computing (ACC-2014) on 10th Oct 2014. Launched
in 1972, CSI Pune has gained prominence by organizing a number of
national and international events including two annual conventions.
Incidentally, CSI-77 at Pune was my first convention as a regular member
after I graduated in 1976. In the CSI@50 meet, after a briefing on chapter
activities by Mr. Anand Joglekar, I shared my views on CSI and the way
to move forward in the current context of capacity building and creating
intellectual property. In a special talk, Dr. Mathai Joseph, a seasoned
computer scientist, academic, researcher and one of the most respected
members of the Pune’s software industry took us down the memory lane
on computing in India from 1960s till now. He highlighted the growth as
well the opportunities we had missed. Mr. Shekhar Sahasrabudhe, a former
RVP-VI and past chairman of CSI Pune, made a lively presentation on
past chairpersons of CSI Pune and their contributions to the growth of the
chapter. A few other former chair persons and Dr. Deepak Shikarpur, former
chairman of CSI Pune, and fellow of CSI, shared their experiences. The conf.
ACC-2014, held under the Div V on “Education & Research” headed by
Dr. Anirban Basu, was well attended with over 100 enthusiastic participants
who had interacted actively with the speakers. The panel on “IoT – the
present and the future” was the highpoint of ACC-2014. Dr. Rajesh Ingle
and his team brought out an excellent proceeding with 13 papers, which
will be available in the CSI Digital Library soon. The CSI Pune deserves a
hearty congratulation for showcasing their organizing capacity once again.
Formal Methods (FM), a discipline in theoretical computing,
plays an important role by applying mathematical techniques in the
specification, development and verification of software and hardware
systems. NCFM - the National workshop-cum-Conference on Formal
Methods - recently held during 15-17, Oct 2014 in Bangalore was the
first of its kind that brought together scientists and engineers who
are active in the area of FM and are interested in exchanging their
experiences in the industrial usage of these methods. The NCFM was
organized by the new CSI SIG on Formal Methods whose convener is
Ms. Bhanumathi Shekhar, along with CSI Bangalore and hosted by IISc.
Dr. Ramaswamy Srinivasan, Global R&D Project Manager of ABB Global
Industries, Bangalore in his inaugural address highlighted the growing
needs in FM. In view of the fact that 400 out of 600 talent-starved R&D
centers of international companies are now located in India and the role
of CSI SIG-FM in creating awareness and motivating students to work
in FM, the need to educate our students and working professionals in
FM were the highlights of the special address by the President. The CSI
BC Chairman Chander Mannar and his team have spared no efforts in
making this conf. a technically rich one with a lot of interactions and
learning by the participants. While speaking to Prof. N. Balakrishnan,
of the Supercomputer Education and Research Centre at IISc, it was
given to understand that funding is available for research in the Formal
Methods in Cyber Security and related areas, and that SIG-FM can work
with them in popularizing Formal Methods among researchers.
While we at CSI are taking advantage of our association with IEEE
& IEEE Computer Society in organizing a number of Distinguished Visitor
Programme talks and tutorials at various CSI Student Branches, our own
home grown Distinguished Speaker Programme (DSP) requires attention
for its effective implementation. The growth in the number of engineering
institutions has created a shortage of qualified teachers; DSP was initiated to
rope in both experienced academic and industry professionals to supply an
alternative stream of teachers who will help our students become industry
ready. While it is quite easy to organize programmes with speakers from
abroad, my personal experience in getting our senior members of academia
and industry for talks is quite a mixed one. Some have been overwhelmingly
responsive to our request but a few do not even acknowledge the request,
left alone coming forward to speak at conferences organized by CSI.
I am happy to share some of the activities of our Education
Directorate in their continued effort in service to our members which
include: Signing of MoU with Telecom Sector Skill Council (TSSC)
for skill development initiatives in the Telecommunications domain;
Conducting the SEARCC International School Software Competition
ISSC-2014 at Rajalakshmi Engineering College, Chennai in which ROC
Taiwan came first, followed by India (Team-B) and Sri Lanka (Team-A) in
the second and third places respectively. The prizes were presented by
Mr. HR Mohan, President, CSI, Dr. Thangam Meganathan, Chairperson,
Rajalakshmi Group of Institutions, Prof. P Thrimurthy, Past President,
CSI, Mr. S Ramanathan, Past Hony. Secretary and Mr. Bhaskaran, Vice
Chairman, CSI Chennai; and Continuing its campaign on Open Source
Technology, by organizing a free workshop for faculty members on BOSS
MOOL at JNTU, Anantapur, Andhra, jointly with CDAC, IIT Madras. The
workshop, attended by over 100 was inaugurated by Prof. Lal Kishore,
Vice Chancellor, JNTU-A and facilitated by Prof. Ananda Rao.
I am happy to note that CSI Chennai in association with IEEE Computer
Society and IEEE Professional Communication Society is organizing an Essay
Contest for the school and college students on the topics “ICT for Digital
India”, “ICT for Make in India” and “ICT for Clean India” which are the
important initiatives of the Government of India. I wish all the best for their
efforts and request them to share the views of the young minds to the PMO
and DeitY.
Due to space constraints, I will discuss some of the recent
developments in the country such as eCommerce, ICT in manufacturing
and other initiatives in the next month. While I conclude my message, let
me once again remind you to enroll new members in CSI by letting them
know about the 15% discount on Life Membership being offered for a limited
period. This will end by Dec 2014.
Once again I extend my personal invitation to all for your participation
in CSI-2014, the flagship event of CSI at Hyderabad. I eagerly look forward
to seeing you at Hyderabad in Dec 2014.
With best regards
H R Mohan
President
Computer Society of India
CSI Communications | November 2014 | 5
Editorial
Rajendra M Sonar, Achuthsankar S Nair, Debasish Jana and Jayshree Dhere
Editors
Dear Fellow CSI Members,
We have come a long way since the time when data was displayed
only in text form. With increasing computing power over the years,
and deluge of data from a variety of fields from e-business to gene
sequencing, the visual representation of data has assumed greater
and greater importance. Techniques and technologies for visualisation
have emerged over the years making it possible and feasible to make
sense out of the overwhelming amount of data. Information today can
be represented in various formats which makes computing more and
more interesting with a variety of applications in different fields. The
advent of Google glass which displays information in a smartphonelike hands-free format, is a pointer to the future. The theme of CSI
Communications this month is related to the vast field of Visualization
Technologies. The articles of course are not covering the breadth of
the field, but as with other such fields, the theme would be taken up
again in future to do justice to the coverage.
We start our cover story section with an article titled “Visualization
– Techniques, Methods and Tools” by Dr. KVSN Rama Rao, Mr.
Surya Putchala and Midhun Thaduru. The article talks about how
visualization adds value to the data that needs to be analyzed. After
explaining the process of visualization the article provides information
about various open source tools available for visualization. The
second article is by A.B. Karthick Anand Babu, D. Maghesh Kumar
and G. RajaRaja Cholan and is titled “Visualization for STEM Subjects”.
The article discusses how visualization enhances learning experience
for students of subjects such as Science, Technology, Engineering and
Mathematics. It also gives information about visualization tools that
can be used and concludes saying that domain knowledge is required
for customizing and designing effective visualization using readily
available open source tools.
With increasing computing power over the years, and deluge of
data from a variety of fields from e-business to gene sequencing,
the visual representation of data has assumed greater and greater
importance. Techniques and technologies for visualisation have
emerged over the years making it possible and feasible to make
sense out of the overwhelming amount of data.
Our Technical Trends section is enriched with three articles. The first
article is written by Dr. B. Eswara Reddy and Mr. K. Rajendra Prasad on
“Visualization Tool for Data Mining”. This article provides information
about a tool called Visual Access Tendency (VAT) which is useful
for detecting information of number of data clusters (or classes) in
visual form. The second article is by Sreeja Ashok, Dr. M. V. Judy
and N. Thushara Vijayakumar n on “Information Visualization in Gene
Expression Data”. The article talks about significance of visualization
in the context of gene expression data and then goes about explaining
the process of visualization. The third article is by Rishav Singh and
Dr. Sanjay Kumar Singh and is titled “Big Data Visualization Using
Cassandra and R” which discusses the use of distributed database
system Cassandra and compares it with SQL database.
Research Front section has two articles. First one is by Dilip Kumar
Dalei, B. V. Hari Krishna Nanda and N. Venkataramanan titled
“Visualization Methods for Vector Fields: An Insight”. The article
provides introduction to scientific data visualization and provides
overview of fundamental flow visualization technique. The second
article is by Anchal Garg, Madhurima, Madhulika and Saru Dhir titled
“A Walkthrough & Pathfinder for Research Novitiates: Google Scholar Vs
Microsoft Academic Search”. The article explains how searching can be
done while researching.
In regular article section, we have three articles on different topics.
The first article titled “Active Queue Management” is written by
Amol Dhumane and Dr. Rajesh Prasad who write about new queue
management technique used for handling congestion in networks.
The second article is on an interesting topic of “Information Technology
CSI Communications | November 2014 | 6
to Curb Piracy in Bollywood” written by Sumith Kumar Puri and Dr. H K
Anasuya Devi. This article makes us aware about CineCat, a software
application idea. The third article titled “Protection of Software as
Intellectual Property” is sent by Dr. M Hanumanthappa, Mrs. S Regina
Lourdhu Suganthi and Mrs. Rashmi S. The article explains the meaning
of intellectual property and discusses the protection provided by law
for such intangible property in the form of patents and copyrights. It
compares the two and concludes saying that for software adequate
protection in the form of patent right is highly desirable and dual IP
protection in the form of Patents and Copyrights can also be explored
to cover both functional and non-functional aspects of software.
... for software adequate protection in the form of patent right is
highly desirable and dual IP protection in the form of Patents and
Copyrights can also be explored to cover both functional and nonfunctional aspects of software.
In Practitioner workbench section we have two articles from our
regular contributors. The first one: “Fun with ‘C’ Programs: Reversing
a String using a Bitwise Operator” under Programming.Tips() section
is written by Prof. Wallace Jacob of Tolani Maritime Institute and
second one: “RStudio – Studio of R” under Programming.Learn(“R”)
section is written by Umesh P and Silpa Bhaskaran, Department of
Computational Biology and Bioinformatics.
Under IT industry perspective we have an article titled “Future of
Medical Transcription Jobs in India – Need to Extend it Beyond Record
Generating Process?” by Prof (Dr.) DG Jha. This article discusses
various points such as importance of medical transcriptions (MT),
why the MT work is outsourced, quality standards applicable to MT,
stages in the process of MT, technological innovations which are
creating a doubt whether it is dead end for smaller Indian companies
doing MT jobs, factors obstructing outsourcing MT work to India and
finally talks about revival strategy that can make India a preferred
destination once again for MT work.
Under Innovations in India column, which we recently started from
July 2014 issue, we have a brief article titled “Software User Experience
Maturity Model” by Rajiv Thanawala and Prachi Sakhardande of TCS.
In this article they explain how metrics can be used for measuring
quality of software products in terms of user experience.
Under information security section of Security Corner column, this
time we have an article on “A Quick Look at Virtual Private Database
Security” by Jignesh Doshi and Bhushan Trivedi, where they explain
how Oracle’s virtual private database technology can be used for
prevention of theft of sensitive data. In the second section of Security
Corner column, we have a case study in IT Governance, IT Risk and
Information Security by Dr. Vishnu Kanhere, who writes about case
of Netrapur Police who plan to implement new surveillance system
called San-nirikshan and the kind of disruption it is expected to create.
In our regular section called Brain Teaser we have a cross-word by Dr.
Debasish Jana, Editor on Visualization Technologies. He also answers
readers’ questions under the column Ask an Expert: Your question,
Our answer. Briefs of various ICT news of October 2014 are compiled
and brought to us by H R Mohan, ICT Consultant, President, CSI and
former AVP (Systems), The Hindu, Chennai under Happenings@ICT.
We have other regular features like CSI Announcements, CSI Reports,
Chapter and Student Branch news etc. Please remember we welcome
your suggestions and feedback at csic@csi-india.org. Please do write
and help us serve you better. Wish you happy reading and learning.
With warm regards,
Rajendra M Sonar, Achuthsankar S Nair,
Debasish Jana and Jayshree Dhere
Editors
www.csi-india.org
Cover
Story
Dr. KVSN Rama Rao*, Mr. Surya Putchala** and Midhun Thaduru***
*Prof., Dept. of CSE, MLR Institute of Technology,Hyderabad
**CEO and Chairman, Zettamine Technologies
***Associate Consultant, Zettamine Technologies
Visualization-Techniques, Methods and Tools
Why Visualization?
It is a well known fact that “A picture
is worth a thousand words.“ To extract
and analyze the massive amount of
generated data, visualization plays an
incredible role. Further it amplifies the
cognition by helping in pattern detection
and enhancing visual insight of a large
quantity of data. It helps us to see data in
context, analyze and discover knowledge.
For companies across different industries
– retail, logistics, banking and Finance,
Insurance, energy etc., data visualization
offer terrific opportunities to identify
new products or uncover customer
propensities that can provide insights
of tremendous value to Businesses. For
example, in a retail industry, increased use
of geo-spatial visualization and analysis,
the location of store, the difference in
market size according to region, price and
compensation studies in regard to specific
regions, etc. reflect more clearly their
potential advantage.
Data Visualization techniques often
comes handy while representing large
quantities of data and help making sense
of big data and thus provide an exploratory
platform for gaining deeper and clear
insights. Some of the key functions of
visualization are highlighted below.
•
To present large volumes of data
(structured
or
unstructured)
effectively and elegantly.
•
Provide a platform for exploring
various facets of information
aesthetically and interactively.
•
Promote a deeper level of
understanding of the data under
investigation and assists us in
drawing conclusions.
•
Share information effectively to
persuade, collaborate and emphasize
important aspects of data.
A Simple Process for Visualization
As the data and the number of sources of
information keeps growing, extraction of
suitable information and presenting in a
human consumable form becomes a great
challenge. However, a data analyst can
systemically increase the value of data
through two major processes – one, Data
Cleansing/Pre Processing and the second,
Visualization process. Visualization is a six
step process.
•
•
•
•
•
•
Mapping : is encoding of data into
visual form. It is used for achieving
accurate relationship between data
points and visual objects that are to
be described.
Selection : of attributes from the
data which aims for right pictorial
representation.
Presentation : is effective management
and organization of information in
the available screen space.
Interactivity : is providing facilities to
organize explore and rearrange the
visualization.
Human factors : are easy readability
and accessibility of information for
end user.
Evaluation: is finding out effectiveness
in the created visual, if we have
succeeded in reaching our goal in
creating lucid and easy to understand
graphics.
A Brief Survey of Graphical Techniques
Charting or Graphing is small subset of
visualization where the data in question is
explained with the help of bar charts, line
charts or pie charts.
Information graphics or infographics
are graphic visual representations of
information, data or knowledge intended
to present complex information quickly
and clearly. They can improve cognition
by utilizing graphics to enhance the
human visual system’s ability to see
patterns and trends. Infographics are
used to communicate a message, to
simplify presentation of large amount of
data, often different facets arranged in a
thematic way.
A Scorecard is a tabular visualization
of measures and their respective targets
with visual indicators to see how each
measure is performing against their
targets at a glance. Scorecards may
contain columns that show trends in spark
lines. It measures performance against
goals. It displays the graphic indicators
that visually convey the overall success or
failure of an organization in its efforts to
achieve a particular goal.
A report is the presentation of
data transformed into formatted and
organized according to specific business
requirement. Reports contain detailed
data in a tabular format and typically
display numbers and text only, but
they can use visualization to highlight
key data.
Dashboarding takes visualization
a step further by aggregating several
different pieces of visual information in
a single location. As quoted by Stephen
Few, “dashboarding is a visual display of
the most information needed to achieve
one or more objectives which fits entirely
on a single computer screen so that it
can be monitored at a glance.” A typical
dashboard might contain a scorecard,
an analytical report and an analytical
chart. Digital dashboards are laid out to
track the flow inherent in the business
process that they monitor. Dashboard is
a user interface that is used to organize
and present information in a way that is
easy to read. A good dashboard presents
information about important data, with
fewer graphs and time overview.
To generate visualization, data plays
a major role. In order to visualize, data
need to be in one of the format.
Following are various data formats
that are available.
•
Spreadsheets
are
electronic
document in which the attributes are
stored in columns and the objects are
stored in rows.
•
JSON (Java Script Object Notation) is
a simple human readable file format
with data objects consisting of key
value pairs.
•
XML (extensible Markup language)
is a flexible way to create common
information formats and share both
the format and the data on the world
wide web
•
Delimited Separated Values is format
of data that is used to store various
two dimensional arrays of data by
separating the values in each row
with specific delimiter characters.
Various types of delimiters are
comma, space, tab, and semi-colon.
•
RDF
(Resource
Description
Framework) is a general framework
for describing website metadata, or
information about information.
•
HTML (Hyper Text Mark-up Language)
is a language used for describing webpages using ordinary text.
In many cases, there will be a need
to use more than one data format (variety
CSI Communications | November 2014 | 7
of data). In such situations, we need to
integrate different varieties of data. To
create visualizations for such complex
data, there are several popular approaches
and methods which are discussed below.
Popular Approaches/Methods for Data
Visualizations
There have been some conventional
ways to visualize data in the form of
tables, histograms, pie charts and bar
graphs. However to convey a message
effectively there are some exciting
visual techniques that are available.
In addition to the above methods, there
are certain popular open source tools.
Technique
Description
Choropleth
Thematic map where each spatial unit is filled with pattern or color which are scaled and normalized.
Choropleth used to show spatial variation of one or two variables at time by using color, shades
and/or patterns. E.g. population density of each state in a country.
Contour Heat Map
Chord Diagram
Collapsible Tree
Tree Map
Calendar Heat Map
used to display density from the vector point of data. Contour heat maps are used to plot when there are
large number of clustered and continuous data points which can also take categorical variables. These
maps do not actually plot the data but designs a surface fit to the data. E.g. density of population.
used to display inter-relationship between group of entities. The data is arranged around the circle
and the relationships are displayed by arcs.
Used for hierarchical data of long nested lists on the web pages can be difficult to understand.
The tree view, a user interface widget that displays hierarchical data as nested lists, solves this
problem by making lists collapsible and expandable; a list can be opened by or closed by clicking
on its parent list item. where the parent node and children nodes are are joined by arcs or line. E.g.
government divisions and sub-divisions.
Used for hierarchical data that shows attributes of leaf nodes using size and color coding. Tree
maps enable users to compare nodes and sub-trees even at varying depth in the tree and help them
spot the pattern. E.g. visualization of continents, countries, population and area
Chart time series onto vector of dates. E.g. Twitter activity of a celebrity on each day.
Method
Word cloud
Description
is a method of visualizing unstructured text data. Word cloud expresses the occurrence of words in
the text form with size of word or phrase directly related to the frequency of occurrence with which
the word has occurred in the text document. Word cloud helps quickly analyze the main focus or
topics of discussion and can also help us in sentiment analysis.
Association Trees
are used for understanding word association in large quantities of text. Association is most
commonly used in social media text, News analysis or customer feedback. Latent Semantic
Analysis (LSA) is a statistical computation used to identify relationships between a set of
documents and the terms contain by producing a set of concepts related to the documents and
terms. The main idea behind LSA is that all the word combinations in which a given word and
word does not appear can determine the similarity of word meanings. Association trees are a
way defining such similarities.
Cubism Horizon graphs
is used to analyze time series, or streaming content. Cubism Horizon graphs is an intuitive way of
project real-time time series plot. Horizon charts reduce vertical space without losing resolution.
is used for relationships and gaining insights from multidimensional data from multiple sources. A
topological network represents the data by grouping similar data points into nodes and connecting
Self-Organizing maps or Topological those nodes by an edge if the corresponding collections have data points in common. The
visualization techniques comes under the heading of scatter plot methods, where the data points
Analysis
are projected on to 2D or 3D dimensional space, then plotting projections on the coordinates in
usual way.
Network Graphs
is used to study meaning and relationships between large contextual data. These graphs are used
to quantify relationships between different vertices of data. These graphs can be directional or
non-directional based on requirement. A network is a collection of points; called vertices with lines
connection these points are called arcs.
CSI Communications | November 2014 | 8
www.csi-india.org
Open Source Visualization Tools
R is a programming language used for
Statistical Analysis, Data Visualization and
Predictive Modelling. R is an implementation
of S language combined with lexical scoping
semantics. R is a scripting and an interpreted
language i.e. a programming language for
which most of the implementations execute
instructions directly without compiling the
program into machine language. R is an open
source software with great contribution
from R-community towards R-programming
in the form of packages which are available
on Comprehensive R Archive Network
(CRAN).
The base graphs in R are used most
commonly and are a very powerful system
for creating 2-D graphics. The main
function for base graphic is plot(). The
base graphs are loaded by default into R.
Grid is an alternative graphics system
added to R that allows for the creation of
multiple regions on a single graphics page.
The grid package needs to be loaded before
it can be used by using library function.
lattice graphics is a powerful
Implementation
and elegant highlevel data visualization system with an
emphasis on multivariate data. The lattice
package is an implementation of Trellis
graphics for R originally developed for the
S-Language. The lattice consists of highlevel generic functions each designed
to create a particular type of display by
default. Lattice gives advantage of high
user controllable settings.
ggplot2 is a plotting system which
takes the best from the base and lattice
graphics. The plot can be split into
scales and layers which gives the added
advantage over base plot.
rShiny is an interactive web application
framework for r which helps us do our
analysis in dynamic fashion. Shiny combines
the computational power of R with
interactivity of modern web. rShiny has
its own capabilities which doesn’t require
HTML, CSS or JavaScript Knowledge.
D3.js is a Java Script Library which helps
build data visualization framework. D3
stands for Data Driven Documentations.
D3.js is a powerful tool for creating dynamic
and interactive data visualizations. D3.js
uses Scalable Vector Graphics, JavaScript,
HTML5, and Cascading Style Sheets
(CSS3) standards.
google charts is a simple and powerful
open source which can be used to visualize
simple line charts to complex hierarchical
tree maps. Google charts are a specialist
for geocharts. Google charts can easily
connect charts and controls into interactive
dashboards. It can also be used to connect
to data in real time using variety of data
connection tools and protocols.
Gephi is an interactive visualization
and exploration platform for all kinds of
network and complex systems, dynamic
and hierarchical graphs. Gephi is used for
exploratory data analysis, link analysis,
social network analysis, and biological
network analysis and poster creation.
Lumify is a open source big data analysis
and visualization platform. Its intuitive
web-based interface helps users discover
connections and explore relationships in
their data via a suite of analytic options,
including 2D and 3D graph visualizations,
full-text
faceted
search,
dynamic
histograms, interactive geographic maps,
and collaborative workspaces shared in
real-time.
Conclusion
Visualization provides great value addition
for the data that is to be analyzed. There
are several techniques, methods and open
source tools for visualization. Once the data
is ready in a particular format, visualization
can be generated by using these techniques.
References
[1] M Khan and S S Khan, (2011). “Data
and
Information
Visualization
Methods,
and
Interactive
Mechanisms:
A
Survey”,
International Journal of Computer
Application (0975-8887), vol. 34 –
No.1, November 2011
[2] https://www.dashingd3js.com/whybuild-with-d3js
[3] S Card, J MacKinlay, and
B Shneiderman, (1998). “Readings
in
Information
Visualization:
Using Vision to Think”. Morgan
Kaufmann.
[4] Alfredo R Teyseyre and Marcelo
R Campo, (2009). “An Overview
of 3D Software Visualization”, IEEE
Transactions on Visualization and
Computer Graphics, vol.15, No.1.
[5] L Chittaro, (2006). “Visualizing
Information on Mobile Devices”,
ACM Computer, v.39 n.3, p.40-45.
[6] Edward R Tufte, (2007). “The Visual
Display of Quantitative Inforamtion”,
Second Edition, Graphics Press.
[7] h t t p : //w h a t i s .t e c h t a rge t .co m /
definition/infographics
[8] h t t p : // t h e n e x t w e b . c o m /
d d /2 0 1 3/ 1 0/ 1 6 / 1 0 - ways - u s e infographics/
[9] h t t p : / / b l o g s . w s j . c o m /
cmo/2014/06/24/outside-voicesthe-visual-web-is-changingeverything-in-media-andadvertising/
[10] h t t p : //s p o t f i r e . t i b c o . c o m /
blog/?cat=34
[11] http://www.computer.org/portal/
w e b /c o m p u t i n g n o w/a r c h i v e /
january2014
[12] http://smartdatacollective.com/
jgptec/140486/3-big-trends-datavisualization
[13] http://code.stephenmorley.org /
javascript/collapsible-lists/
n
About the Authors
Dr. KVSN Rama Rao is a Professor,Dept of CSE at MLR Institute of Technology,Hyderabad.He is a Doctrate in Computer Science
with over two decades of academic experience. He has published several papers in reputed International and National journals and
conferences. His research interests are Cyber Security and Big data.
Surya Putchala is CEO and Chairman for ZettaMine Technologies, a firm focuses on providing high end educational and Management
consulting services to Business around the world. It also aims to become the first think tank in “Big Data” space in India. Over the last 2
decades, he provided thought leading consulting solutions in the areas of Business Intelligence, Data Warehousing, Data Management
and Analytics to Fortune 500 Clients. He has architected commercial Analytical Applications such as Product MDM and Procurement
Optimization. He has held senior leadership roles with firms such as GE capital, Cognizant, Accenture and HCL.
Midhun Thaduru is an Associate Consultant at ZettaMine Technologies. He is with the Data Science team and focusses on Statistical
Analysis, Exploratory Data Analysis, machine learning and Visualization. He extensively uses R, python, D3.js for his day to day
programming. His area of focus is in Life Sciences (Pharma) and Insurance, particularly, Payer and Provider analytics. His interests are
in developing high performance algorithms for Predictive modelling. Midhun graduated from BITS, Pilani.
CSI Communications | November 2014 | 9
Cover
Story
A.B. Karthick Anand Babu*, D. Maghesh Kumar** and G. RajaRaja Cholan***
*Managing Director, KK Infotech, Thanjavur
**Assistant Professor, Department of Software Engineering, Periyar Maniammai University, Vallam, Thanjavur
***PRIST Univerity, Thanjavur
Visualization for STEM Subjects
The process of learning in Science,
Technology, Engineering, and Mathematics
(STEM) fields using visualization will be
more interesting and can surely improve
the effectiveness in the learning and
teaching process. Besides making it easier
for students to understand the STEM
subjects, facilitates teachers in teaching
STEM lessons material in the classroom.
Visualization
Visual means of processed information
is referred as Visualization. The foremost
objectives
of
any
visualization
techniques
are
data
exploration
and
communicating
information
effectively to the intended audience.
Data Exploration is the practice of
using visualization techniques to find
unforeseen
relationships
between
data points or sets of points from huge
databases. Once a relationship has been
found, a similar visualisation will be used
to communicate that will be a reference
to others. Visualization techniques can
also be applied to information that is
already known and it has the potential
to organize large amounts of data in
meaningful ways. Visualizations are
often static, dynamic and interactive,
allowing users to manipulate the data
that they observe. Static visualizations
are the one that do not change with
time. Dynamic visualizations are
those in which the graphical elements
being displayed can change with time.
Interactive
visualisation
involves
humans’ interaction with computers
to create graphic illustrations of
information.
Visualization in Education
The application areas of Visualization
are many, one such area is Education.
Thomas Carpenter and James Hiebert
in[1] address the role of visualization in
education and also represent a framework
for knowledge representation in the
domain of education way back in early
1990s. The purpose of any visualization
to be used in an educational context
is to facilitate the learning of some
knowledge in the form of algorithm,
concept, idea, relationship, fact, decision
making and application. In order to
accomplish this visualization must make
connections between knowledge of the
learner has and the knowledge being
taught. Knowledge of a learner might
be of Fragmented and knowledge being
taught might be Coherent. Fragmented
knowledge results in domains in
which the learner has had little or no
experience. Coherent knowledge refers
to a wealth of information to draw upon
from a domain.
STEM Education
Education in any domain can be viewed
as the externally facilitated development
and representation of knowledge.
Education particularly in the field of
Science, Technology, Engineering, and
Mathematics (STEM) subjects is quite
difficult for the students to understand.
This is because the STEM subjects are
not simple and it requires high levels
of thinking and reasoning from the
students, especially in the areas of
Concepts and Logics. In addition, the
subject material is also quite difficult to
teach. Teachers, as educators, need a
S.No Tools
Science Tools
1.
ChemSketch
2.
Step
Technology Tools
3.
gEDA
4.
Player Project
specific way to teach in order to make the
material easily understood by students.
To make teaching more attractive and
effective, it needs a learning model or
a medium which can provide a concrete
conceptual, illustration, and definitive,
in order to improve the effectiveness of
the learning process and to achieve the
learning objectives.
Visualization tools in STEM Education
Teachers and students have long been
using charts and graphs to analyze and
make sense of data. The effective use
of any technology in teaching requires
thoughtful consideration and planning.
Technology has brought these tools to
a new level. In order to design effective
visualizations in STEM subjects it is
necessary to know the level of audience.
Table 1, list some of the open source
visualization tools used by educators to
provide STEM lessons’ material in the
classroom. This visualization tools acts
either as a primary medium or as an
alternative tool in the teaching learning
environment.
Features
•
•
•
•
Tool for learning Chemistry
Includes comprehensive chemical drawing package,
Offers drawing of polymers, organic elements and structures
h t t p : //w w w. a c d l a b s . c o m /r e s o u r c e s /f r e e w a r e /
chemsketch/
• Tool for learning physics
• Simulator supports classical mechanics, particles, springs,
gravitational and coulomb forces, collisions, molecular
dynamics, and much more.
• https://edu.kde.org/applications/all/step
• Refers to GPL'd Electronic Design Automation tools,
• Supports the teaching of electronics to the students of
technical institutions.
• offers various analog, digital simulation and printed circuit
board (PCB) layout capabilities.
• http://git.geda-project.org/
• Supports teaching of robotics in technical institutions
• Provides a simulated network interface for sensors and to
control robots
• Simulates the 2D and 3D robotic interaction in environment.
• http://playerstage.sourceforge.net/index.php?src=index
Continued on Page 41
CSI Communications | November 2014 | 10
www.csi-india.org
Engineering Tools
5.
OpenModelica • Tool for modelling and simulating industrial applications
such as control system design, embedded system modelling
and numerical algorithms
• https://www.openmodelica.org/download/download-linux
6.
OpenSCAD
• Provides a software for creating solid 3D Computer Aided
Design models
• offers two modeling techniques such as constructive solid
geometry and extrusion of 2D outlines
• http://www.openscad.org/about.html
Mathematical Tools
7.
About the Authors
8.
Maxima
GeoGebra
• Tool for Algebra and calculus teaching
• Manipulates and provides good numerical results
• http://www.gnu.org/software/maxima/
Conclusion
Visualization technology will provide
a detailed representation and creates
interesting definitions of Concepts,
Theories, Formulas, and the Principles
that are contained in the STEM materials.
Visualization tools can also be useful for
bridging the gap between domains of
knowledge and applications. Knowledge
base of the audience is much needed
for customizing and designing effective
visualizations. There are visualization tools
that are available as open source products,
which allow students and educators to
customize the tool and to learn the domain
of education.
References
• The award-winning math app combines tools for, arithmetic
algebra, calculus and geometry.
• Supports instructors to create worksheets and suitable
teaching aids for teaching maths to the students of schools,
colleges and university.
• http://www.geogebra.org/cms/
[1] Hiebert, James & Carpenter, Thomas
P. "Learning and Teaching with
Understanding"
Handbook
on
Mathematics Teaching and Learning1992.
[2] h t t p : //e n . w i k i p e d i a . o r g /w i k i /
Interactive_visualization
[3] h ttp://w w w. dpi . s t at e . nc . us /c t e /
n
program-areas/technology/
A.B. Karthick Anand Babu is member of CSI and Managing Director, KK Infotech, Thanjavur, Tamilnadu. He has rich experience in teaching and
software development. He can be contacted through karthickanandbabu.ab@gmail.com.
D. Maghesh Kumar is devoted Assistant Professor of Faculty of Software Engineering, Periyar Maniammai University,Tamilnadu. He has more
than 20 years of teaching experience. His research area includes “Big Data” , “Cloud Computing” and “Internet of Things”. Reach him through
maghesh.d@gmail.com
G. Raja Raja Cholan, Assistant Professor,Department of Computer Science and Engineering, PRIST University. He has also served as a Coordinator
for Centre for Knowledge Management. His primary research area includes Network Security, Internet of Things and Data Science. He can be
reached at gsrajarajacholan@gmail.com
CSI Communications | November 2014 | 11
Technical
Trends
Dr. B. Eswara Reddy* and Mr. K. Rajendra Prasad**
*Professor, Dept. of CSE, JNTUA, Ananthapur
**Associate Professor, Dept. of IT, RGM College of Engg. & Tech.,Nandyal
Visualization Tool for Data Mining
Introduction
Characterizing and classifying the data
is an emerging need for data mining
functionality. The aim of data mining is
to extract and classify the data based on
the data characteristics. We may have
pre-requisite to prior knowledge about
the number of distinct data classes for
any data mining functionality. For this
valuable determination, we introduce the
visualization tool called as Visual Access
Tendency (VAT), which is used to detect the
information of number of data clusters (or
classes) in visual form. The VAT has been
introduced by author Bezdek. The visual
pattern apparent will presents with the more
clarity of visual results for data classes (or
clusters). A very beginner of data clustering
can makes use of this tool for accessing of
data clusters from the organized data.
Motivation
Many algorithms of data clustering such
as k-means, hierarchical, density based
clustering algorithm etc have to produce
the clustering results without knowing
any prior knowledge about number of
clusters. Suppose the user has trying the
clustering results with incorrect k value in
the k-means, then the k-means procedure
may also produce the inaccurate clustering
results. This is the key motivation for
choosing of visualization tool (VAT) and
it has capable for detecting the number of
clusters (k value) as correctly. Thus, this
visualization tool helps for achieving the
best clustering results.
History
Initially, the author Bezdek has developed
the algorithm of VAT tool for classifying
of clusters. Later he has modified with the
spectral approach called as SpecVAT which
works better than VAT where in the cases of
tough data such as it may be huge, higher
dimensionality etc. He has also developed
another version iVAT (Improved VAT) for
the cases of path based data. Therefore,
we can use any one of the visualization tool
among the three based on our degree of the
complexity of the data.
Visual Access Tendency (VAT)
The following key diagram introduces
the processing steps of VAT tool.
Data Objects
The organization of data consists of a set
of data objects which is further clustered
based on the similarity features for
Fig. 2a: Dissimilarity matrix ‘R’ and its dissimilarity image ‘I’ (before applying the VAT tool)
Fig. 1: Processing steps of VAT
every two-element subset data objects.
Initially, the data is organized as data
matrix (n x m) form which can be shown
as follows with n number of objects and
m number of properties
O\P
P1
P2
...............
Pm
O1
X11
X12
..................
X1m
O2
.
.
.
On
Xn1
4.
Xn2
..................
Xnm
These n data objects are compared
using the distance measures such as
Euclidean, Mhanttan, and Minkowski etc
for obtaining the dissimilarity matrix R.
The following matrix is represents the
dissimilarity matrix R (with size n x n)
O\P
O1
O2
...............
On
O1
0
D12
..................
D1n
O2
.
0
.
.
On
Dn1
..................
0
Dn2
Dissimilarity Matrix and its VAT Image
The dissimilarity matrix R is the input of VAT,
and the VAT tool outputs the VAT Image
with finite number of visual squared shaped
dark blocks. Thus, we detect the number of
clusters by counting of square shaped dark
blocks along the diagonal of VAT Image. The
accessing of number of data clusters by VAT
tool is depicted by the following results
Advantages of VAT Tool
1. It presents the information in visual
so that even normal user also can
able the number of clusters
2. The VAT tool is uses only the Prim’s
logic, so there is no complexity about
understanding of how it works
3. We use this visualization to any kind
of data such as text, image, audio,
CSI Communications | November 2014 | 12
Fig. 2b: After applying VAT tool
and video etc for defining of clusters
or classes
It is possible to extract the correct
k value; so that it is useful for the
k-means
Limitations
1. It takes much time for solving of k
value where in the cases of big data
2. We have less clarity of visual results
if the data clusters are overlapped
3. The visualization results will depends
on distance measures
Designing of VAT Tool
The VAT uses the logic of Prim’s
algorithm for the major purpose of
changing the current indices of the data
objects. In this way, the indices of data
objects are reordered by the VAT tool.
Reordering the indices of data objects
would shown the number of clusters by
squared shaped dark blocks along the
diagonal in the VAT Image, this is clearly
shown in Fig. 2a. The main aspect of VAT
tool is to display the hidden clustering
structure for a set of data objects. The
logical steps of VAT tool are described
as follows
1. Choose the longest edge weight
e(vi,vj) from the dissimilarity matrix R
Index=1;P(index)=vj; Set I = { };
J = {1,2,…,n};
I = I ∪ {j} and J = J – {j}
2. Use Prim’s Logic
For the interactions t = 2,…, n
Select (i,j)∈ arg minp∈I,q∈J {Rpq} ;
P(t) = j;
I = I ∪ {j} , and J = J – {j}
3. Reordered dissimilarity matrix
[RRi,j] =[RP(i),P(j)]
www.csi-india.org
4.
Display the VAT Image of RR
Let R=[0.00 0.73 0.19 0.71 0.16;
0.73 0.00 0.59 0.12 0.78;
0.19 0.59 0.00 0.55 0.19;
0.71 0.12 0.55 0.00 0.74;
0.16 0.78 0.19 0.74 0.00];
Tracing steps for VAT Tool Algorithm for
the Input R(Dissimilarity Matrix)
Finally, we display the VAT image
in MATLAB by executing the command
of imshow(RR), then it displays image of
Fig. 2b. Hence, we can extract the number
of clusters information from this visualized
image.
Conclusion
We make use of visualization tool as VAT
Step No.
Iter. No.
especially in the data mining. The VAT
tool visualizes and reveals the similarity
structure of data objects. Hence we have
a prior knowledge about the number of
data classes (or clusters) which helps
the further processing of data mining
functionality. Specifically this visualization
tool has been used for the purpose of
unsupervised learning. In the k-means,
we can attempt the good clustering
results by this prior k value. Other
clustering algorithms are also required the
knowledge about number of clusters.
[3]
[4]
[5]
References
[1] Liang Wang, James Bezdek, Enhanced
Visual Analysis for cluster tendency
Choose the Max edge P
from R (In Step1) or
Choosing of Min Edge
from the vertices I to J
(In Step 2)
I
[6]
J
[7]
1
-
(2,5)-Max Edge
P(1)=5
{5}
{1,2,3,4}
2
3
t=2
t=3
(5,1)-Min Edge
(1,3)-Min Edge
P(2)=1
P(3)=3
{1,5}
{1,3,5}
{2,3,4}
{2,4}
4
t=4
(3,4)-Min Edge
P(4)=4
{1,3,4,5}
{2}
5
t=5
(4,2)-Min Edge
P(5)=2
{1,2,3,4,5}
{ }
About the Authors
[2]
[8]
assessment and data partitioning,
EEE Trans. on Knowledge and Data
Engineering, 22(10), 1401-1413(2010).
http://www.ece.mtu.edu/~thavens/
pubs.html
http://www.ece.mtu.edu/~thavens/
code/VAT.m
http://www.ece.mtu.edu/~thavens/
code/iVAT.m
Timothy C Havens ,and James
C Bezdek, An Efficient Formulation
of the Improved Visual Assessment
of Cluster Tendency (iVAT) Algorthm,
IEEE Trans. on Knowledge and Data
Engineering. 22( 10), 1401-1413(2010).
Liang
Wang,
James
Bezdek,
Automatically
Determining
the
Number of Clusters in Unlabeled
Datasets, IEEE Trans. on Knowledge
and Data Engineering, 21(3), 335349.(2009).
B Eswara Reddy, K Rajendra Prasad,
Reducing runtime values in minimum
spanning tree based clustering by
visual access tendency, International
Journal of Data Mining & Knowledge
Management Process, 2(3), 11-22
(2012).
James Bezdek, VAT: A Tool for Visual
Assessment of Cluster Tendency, Proc.
Int’l Joint Conf. Neural Networks. 22252230 (2002).
n
Dr. B. Eswara Reddy Graduated in B.Tech.(CSE) from Sri Krishna Devaraya University in 1995. He received Masters Degree in M.Tech.
(Software Engineering), from JNT University, Hyderabad, in 1999. He received Ph.D in Computer Science & Engineering from JNT University,
Hyderabad, in 2008. Currently he is working as professor in the Dept. of CSE, JNTUACE, Anantapur. His research interests include Pattern
Recognition & Image Analysis, Data Warehousing & Mining and Software Engineering. He is a life member of CSI, ISTE, IE, ISCA and
member IEEE.
Mr. K. Rajendra Prasad has a research scholar at JNTUA, Anantapur. He has completed B.Tech(CSE) at RGM Engg College, Nandyal and
M.Tech(CSE) at KBN College of Engg, Gulbarga. Currently, he is pursuing PhD (CSE) at JNTUACE, Ananthapur. He is life member of CSI and
a member of IEEE.
CSI Communications | November 2014 | 13
Technical
Trends
Sreeja Ashok*, Dr. M. V. Judy** and N. Thushara Vijayakumar***
*JRF, DST Funded Research Project, Amrita School of Arts and Sciences, Amrita Vishwa Vidyapeetham University, Kochi
**Associate Professor, HOD, CS and IT, Amrita School of Arts and Sciences, Amrita Vishwa Vidyapeetham University, Kochi
***JRF, DST Funded Research Project, Amrita School of Arts and Sciences, Amrita Vishwa Vidyapeetham University, Kochi
Information Visualization in Gene
Expression Data
Significance of Visualization in Gene
Expression Data
Bioinformatics is an exploration area
that manages huge amount of data.
Gene expression analysis is a significant
instance of Bioinformatics which tries to
identify expression levels of genes through
microarray experiments. Gene expression
data generated through this process
typically represents thousands of gene
expressions across multiple experiments.
Mostly, gene expression data are noisy
and of high dimensionality which makes
the analysis process difficult. They
have to be transformed into a reduced
set of genes for subsequent analysis.
Transforming large-scale gene expression
data into a meaningful set of data can be
done through preprocessing techniques
and through feature reduction methods.
The primary objective of analysis could
be to find functional groupings of genes
by discovering similarity or dissimilarity
among gene expression profiles, or
predicting the pathways of previously
uncharacterized genes. This can be
done through different supervised and
unsupervised methods. The challenge
is to interpret and to make sense out of
this processed information. Representing
the analysis output through a suitable
visualization framework simplifies the
reasoning process and allows users to
easily explore the relationships among
genes and conditions.
Visualization Process
Data mining and Visualization techniques
work hand in hand to enable complete
elucidation and user interpretation of
large datasets. The sandwich technology
that combines both techniques is
useful in molecular biology where large
volumes of sequences and gene arrays
can be efficiently mined and represented
in graphs, trees and chains to extract
meaningful information. Visualization
process of gene expression data focuses
on data preprocessing, data reduction,
clustering/classification,
visualization
technique and Analysis & Knowledge
discovery. The visualization process flow
is shown in below Fig. 1.
Genes that contribute noise will
lead to wrong analysis and results.
Preprocessing techniques helps in
converting raw data to meaningful
biological data by removing noise, low
intensity, bad quality and empty spots from
gene expression data using Normalization,
Filtration, Sampling, Extraction, Labeling,
Scanning etc. Feature reduction procedure
is an effective approach to downsize the
data. For example, when the dataset has
thousands of genes and few samples and
the objective is to classify novel samples
into known disease type, dimensionality
Fig. 1: Visualization process flow
reduction methods helps in finding
a subset of informative genes which
can be processed for further analysis.
Classification helps in understanding the
complex relationship/interaction among
the various conditions and features of a
biological object. For example, a training
dataset has diseased and normal cells and
when a new cell is obtained, classification
process has to automatically determine
whether it is normal or a diseased
cell. Cluster analysis helps in grouping
genes with similar function or grouping
samples with similar expression profiles.
Visualization techniques help in visually
inspecting and interacting with two/
three dimensional view of processed data
set. Following table shows the available
visualization techniques.
Available Visualization Techniques
Visualization
techniques
Objective
Methods
Geometric
techniques
Visualization of geometric transformations and
projections of the data
Scatterplots Matrices , Hyperslice, Parallel Coordinates, Projection
Pursuit Techniques, Prosection Views, Landscapes
Icon- based
techniques
Visualization of the data values as features of icons
Chernoff-Faces , Stick Figures ,Shape Coding, Color icons, Tile Bars
Pixel-oriented
techniques
The basic idea of pixel-oriented techniques is to map
each data value to a colored pixel. Each attribute value
is represented by a pixel with a color tone proportional
to a relevance factor in a separate window
Query-Independent Techniques
-Simple Techniques
-Recursive Pattern Technique
-Space-Filling Curves
Query-Dependent Techniques
-Spiral Technique
-Circle Techniques
-Axes Technique
Hierarchical
techniques
Visualization of the data using a hierarchical
partitioning into two- or three-dimensional subspaces
Treemap , Cone trees , Dimensional Stacking , InfoCube ,Worldswithin-Worlds , Box plots
CSI Communications | November 2014 | 14
www.csi-india.org
Graph-based
visualization
Graphs (edges + nodes) with labels and attributes
are used where emphasis is on data relationship
(databases, telecom)
Useful for discovering patterns
2D and 3D graph
Basic Graphs (e.g., Straight-Line, Polyline, Curved-Line, Orthogonal
Graphs)
Enrichment Map
Distortion
techniques
Global context/view of the information content.
Hyperbolic tree , Graphical Fisheye view, Perspective wall , Polyfocal
Display , 2D Bifocal Display
Dynamic/
Interaction
techniques
Providing interaction mechanism that make it possible
to manipulate visualization effectively and effortlessly
Brushing , Linking , Zooming & Panning,
Detail on Demand , Filtering (Selection, Querying), Data-toVisualization Mapping, Projections
Matrix
Techniques
To simultaneously explore the associations of up to
thousands of subjects, variables, and their interactions,
without first reducing dimension
Heat map
Hybrid
Techniques
Integrated use of multiple techniques in one or
multiple windows to enhance the expressiveness of
the visualizations
Scatter plot of icons with dynamic zooming and mapping
Dynamically link and brush scatterplot matrices, star icons, parallel
coordinates.
Graph Matrix Visualization.
Brief summary of some of the prominently
used visualization techniques for gene
expression data are given below
•
Heat map: Provides an overview of
the data. A colored matrix display
will represent the matrix of values
as a grid, number of rows equal to
the number of genes being analyzed,
and the number of columns equal
to expression levels (Fig. 2a). A big
challenge for interpreting patterns in
a colored matrix is that the rows and
columns need to be re-ordered in a
meaningful way.
•
Scatter plots : Scatter plots are
useful for pair wise comparisons,
finding which genes are excessively
expressed, they provide an overview
of the multivariate distribution of
expression values (Fig. 2b)
•
Parallel
coordinates:
Parallel
coordinates are a common information
visualization technique for highdimensional data sets. Expression
data for each gene corresponds to a
dimension in the data set, with data for
each gene being represented by one of
a series of parallel vertical axes. Colorcoding of expression profiles are very
efficient (Fig. 2c)
•
Dendrogram:
The
dendrogram
method plots the hierarchical
tree information obtained as a
graph from the cluster pane. The
numbers along the horizontal axis
represent genes and the height of
the diagram indicates the distance
between the genes. Different color
codes can be assigned to represent
null values or zero values. Shades
represent intensity or magnitude of
expression(Fig. 2d)
Conclusion
Every day new discoveries are made in the
field of molecular biology & genetics and
sheer volume of data comes out from studies.
These vast quantities of heterogeneous,
2008.
[2] Tangirala Venkateswara Prasad and
Syed Ismail Ahson , “Visualization
of microarray gene expression
data”, Bioinformation by Biomedical
Fig. 2: Visualization techniques for gene expression data
dynamic
and
largely
unprocessed
information have to be transformed into
a coherent and user friendly format easily
accessible to all. Analytical methods
together with visualization techniques play
a major role in exploring, analyzing and
presenting meaningful inferences.
References
[1] Ashraf S Hussein “Analysis and
Visualization of Gene Expressions
and Protein Structures”, Journal of
Software , Vol. 3, No. 7, October
Informatics Publishing Group, May
03, 2006.
[3] Wegman, E “ Hyperdimensional data
analysis using parallel coordinates.”
Journal of American Statistics
Association 85, 664-675.
[4] Pavlopoulos, G A, Wegener, A L,
& Schneider, R “A survey of
visualization tools for biological
network analysis. Biodata mining”,
1(1), 1-11.
[5] Chun Tang, Li Zhang and Aidong
CSI Communications | November 2014 | 15
About the Authors
Zhang , “Interactive Visualization and
Analysis for Gene Expression Data” ,
HICSS-35: 35th Hawaii International
Conference on System Sciences.
[6] Poornima. S, Dr. J Jeba Emilyn,”
A
Comparative
Analysis
on
Visualization of Microarray Gene
Expression Data”, ISSN: 1694-2108 |
Vol. 12, No. 1. APRIL 2014.
[7] Purvi Saraiya, Chris North, Karen
Duca, “An Evaluation of Microarray
Visualization Tools for Biological
Insight”, IEEE Symposium on
Information Visualization 2004,
October 10-12, Austin, Texas, USA.
[8] Janne Nikkila, Petri Toronenb,
Samuel Kaskia, Jarkko Vennaa, Eero
Castrenb, Garry Wongb,” Analysis and
visualization of gene expression data
using Self-Organizing Maps”, Neural
Networks 15 (2002) 953–966.
n
Sreeja Ashok is currently working as JRF for a DST Funded Research project at Amrita School of Arts and Sciences, Amrita Vishwa Vidyapeetham
University, Kochi. She has completed MCA, has 14 years of experience in IT industry (Avenir & Wipro Technologies) with expertise in Software
Engineering, Project Management, Quality Management & Data Analysis, 2 years of teaching experience.
Dr. M. V. Judy is Associate Professor, Head of the Department of CS and IT at Amrita School of Arts and Sciences, Amrita Vishwa Vidyapeetham
University, Kochi and also principle Investigator for DST Funded Research Project under Cognitive Science Research Initiative. She has
completed MCA, Mphil and Ph.D in Computer Science from reputed universities/institutes and has 12 years of teaching experience.
N.Thushara Vijayakumar is a former lecturer at NIT Calicut, completed M-Tech in Genetic Engineering from SRM University, Chennai and
B-tech from Anna university, Chennai. Currently working as JRF for a DST Funded Research project at Amrita School of Arts and Sciences,
Amrita Vishwa Vidyapeetham University, Kochi.
CSI Communications | November 2014 | 16
www.csi-india.org
Technical
Trends
Rishav Singh* and Dr. Sanjay Kumar Singh**
*Working in Infosys
**Associate professor, IIT-BHU, Department of Computer Science and Engineering
Big Data Visualization using Cassandra and R
In the current world each organization
is moving towards big data. Big data
basically consist of data storage and data
analysis. Earlier people used to store
only important data, since there was
no cost effective way of storing all the
data. But most of the organisations and
researchers felt the need of huge volume
of data for building up mathematical
models, so that complex problems can
be solved. However, in the last few years
many frameworks and databases have
been designed to store huge volume
of data using commodity hardware.
Whenever we talk about big data there
are 3 V’s that come to our mind Velocity
(the rate at which the data is growing),
Variety (structured, semi-structured and
unstructured) and Volume (Size of the
data in terms of petabytes or xetabytes).
There are many solutions which have
been proposed for the same, but most of
the time people are not aware of which
framework or database to choose. While
creating a database for any particular
application, the architect should
consider three features: consistency,
availability and partition tolerance.
Consistency means even if concurrent
updates are happening, users will get
the updated result irrespective of the
place from where they are accessing the
data. Availability means data should be
available to the users 24*7 irrespective
of the load on the server. Partition
tolerance means even in the case of
the partial failure the system should
be functional. Brewer suggested a CAP
theorem, which is applicable for all the
databases. According to him a database
cannot have all the three features, i.e
consistency, availability and partition
tolerance. They can have only two
features out of the three, for Example:
Consistent & Available: postgres
My SQL, SQL server.
Available & Partition Tolerance:
Cassandra, RIAK, Voldemort, CouchDB.
Consistent & partition tolerance: Neo4J,
MongoDB, HBase, Hypertable, Redis.
Cassandra
There are many distributed database
systems available but they have only
two features out of the three desired
features. Cassandra is a distributed and
decentralized database, which has high
availability and partition tolerance, but at
the same time it is eventually consistent
which means that its consistency level
can be increased by sacrificing the
availability. It is capable of storing all the
forms of digital data i.e structured, semi –
structured and unstructured data.
Cassandra[1] was designed with
a motive to have a database that
should be able to write at a very high
speed irrespective of the load on the
server or even if most of the nodes in
the cluster are dead or unavailable. In
Cassandra[2] the write operation is very
fast as compared to the read operation.
During the write operation, data is first
written to the commit log along with the
instructions for the durability purpose
and the flag bit is set to one,
then the data is moved to the
memtable which resides in RAM,
when the memtable exceeds a
given threshold value the data
is flushed into the SStable which
is present in the secondary
memory. Once the data is written
in the SStable, flag bit is again set
to zero or if anything goes wrong
the data will again be sent to the
memtable from the commit log.
But the user gets the confirmation that
the data is written successfully, when the
data is first written inside the commit log
file, and he does not have to wait for all
this process to complete.
Cassandra is a distributed system
and all of us know that the distributed
system works on network, which is one
of the highly unreliable resource in this
world, we don’t know when the traffic on
the network is going to increase/decrease
or it will get disconnected. Keeping this
thing in mind Cassandra was provided
with a mechanism known as HintedHandoff, which means that if a user wants
to write any data on a particular node(A)
which is unavailable, then a hint will be
written on its neighbor node(B), stating
that whenever node(A) comes back
node(B) will pass this hint.
R
Many organizations are moving towards
big data analysis and visualization. R
is an open source tool which is being
used by most of the analysts for building
up a mathematical model such as
recommendation system and predictive
system. To build up such models many
optimized algorithms packages[3] are
present in R such as SVM, kNN, Decision
Trees, outliers, Naïve Bayes, adaboost ,
JRip and many more.
R has a Cassandra package
(RCassandra[4]) to access the functionality
of apache cassandra cluster such as login,
updates, queries and read/write operation
without using java.
Following plot shows the iris data in 4
dimensions which has been first inserted
into Cassandra database using R and then
read the same data for analysis. Plot()
function is used for the representation of
data.
Cassandra Vs SQL:
The SQL was designed with a motive to
have database that can store only the
normalized form of data because at that
time the hardware cost was very high, but
due to normalization the query needs to
refer multiple tables for fetching a particular
record, due to this the latency time of the
query is very high and which affects the
overall performance of the application.
SQL was designed to handle data
in terms of Gigabytes and up to some
extent Terabytes. But the Cassandra can
handle the data in terms of petabytes and
xetabytes.
In SQL the table schemas are defined
upfront and the users are forced to insert
data for each column or else they have to
use the NULL value.
Scaling up and scaling down depending
upon the load is very expensive in SQL, but
in Cassandra you just have to plug in a new
node or remove the node. No configuration
needs to be done on the cluster side.
CSI Communications | November 2014 | 17
About the Authors
In SQL first the database is designed,
then the queries are identified to solve the
problem. But in Cassandra first the queries
are identified and then the data is organized
around the queries. There are no unions and
joins. Each query will refer to a single table.
Since SQL is a 100% consistent
database they need to use locks on
each transaction and if in a distributed
system the locks are not used properly
it will either lead it to starvation or
deadlock. Most of the time people say
that my data is important and it needs
to be consistent, but there are many
organizations such as Google, Facebook,
Amazon which have an inconsistent
database although there data is very
important.
References:
[1] http://cassandra.apache.org/
[2] http://www.datastax.com/
[3] http://en.wikibooks.org/wiki/Data_
Mining_Algorithms_In_R
n
[4] http://cran.r-project.org/
Rishav Singh is currently working in Infosys and also pursuing his Ph.D (part time) from Indian School of Mines, Dhanbad.
Dr. Sanjay Kumar Singh is an Associate professor in IIT-BHU in the department of Computer Science and Engineering.
His main area of interest are Biometrics, Computer Vision, Image Processing, Video Processing, Pattern Recognition and
Artificial Intelligence.
CSI Communications | November 2014 | 18
www.csi-india.org
Research
Front
Dilip Kumar Dalei*, B. V. Hari Krishna Nanda** and N. Venkataramanan***
*Scientist (ANURAG), Defence R & D Organization (DRDO), Hyderabad
** Scientist (ANURAG), Defence R & D Organization (DRDO), Hyderabad
*** Scientist (ANURAG), Defence R & D Organization (DRDO), Hyderabad
Visualization Methods for Vector Fields:
An Insight
Introduction
Scientific Data Visualization is the branch
of science which deals with transformation
of scientific data into meaningful visuals
to comprehend and gain insight into the
underlying scientific phenomena. The
output data from scientific experiments
and simulations are normally sampled
in 2D or 3D grids. The grids provide
an efficient data structure for data
storage and retrieval. A variety of grids
with different topology and geometry
are constantly used for numerous
applications. The choice of a right kind
of grid depends solely on the application
domain. These grid-based data are then
processed by different visualization
methods to generate visuals suitable for
human interpretation.
Over the years, researchers have
devised various methods, which form the
core part of any scientific visualization
system. There are various schemes of
classification for visualization methods.
Based on type of input data, these
can be broadly put into three major
categories, viz. Scalar, Vector and Tensor
visualization methods.
Scalar methods operate on the
data that represents scalar quantity viz.
temperature, pressure, density etc. In
vector field visualization, we deal with
vector quantity based data (like velocities).
The Tensor data type is more complex
than other ones. They generally arise
from applications such as engineering
analysis (Stress-Strain tensors), molecular
diffusion measurements etc.
What are Vector Fields
Vector fields are defined as the collection
of vectors over a region of space in
the domain. Mathematically, they are
represented as a vector-valued function
that assigns a vector at each point of the
field. The examples of vector fields are
gravitational field, electrical field that
surrounds charges, magnetic fields and
velocity fields (Fig. 1).
Fig. 1: Vector field with equation
Vx = x, Vy = -y in Matlab
In a vector field the vector
quantities can vary over space and time.
A static vector field represents vectorvalues that change over space, while a
time-dependent one changes over both
space and time. One of the important
applications of vector field visualization
is study of flow (velocity). In case of
flow, the field is termed as velocity
field, which is frequently encountered
in vector visualization arena. We have
assumed a flow vector field for the study
of visualization methods. The concepts
are still applicable to other vector fields.
The flow field is further divided into two
categories: steady (time-independent)
and unsteady (time-dependent).
Visualization Methods
The vector field needs to be visualized
effectively
by
considering
both
magnitude and direction simultaneously.
The problem gets challenging if the data
size is large and time-variant. In recent
times, vector visualization techniques
have received widespread attention.
This is due to extensive research in
areas such as Metrological simulation,
Medical blood flow, Computational
simulation of air flow around aircrafts,
ships and differential equation systems.
The vector fields representing fluids
are transparent in nature, thus making
the flow pattern invisible to us. The
visualization discerns the flow patterns
to convey its qualitative and quantitative
information. To get an insight of vector
field it is necessary to explore the flow
both locally and globally. Based on the
density of representation[1], the vector
based methods can be classified as local
and global techniques. A local technique
visualizes only a portion of domain, while
the complete flow is considered in global
techniques.
The local techniques create a
representation based on flow lines –
streamlines, streaklines and pathlines
etc. The flow lines are trajectory of
mass-less particles through the vector
field. They describe flow path with start
(come from) and end (will go). On the
other hand, global techniques visualize
the entire flow in single picture with
minimal user intervention. Examples
are line plots (hedgehogs), arrow plots,
glyphs, texture based methods (LIC
& Texture Splats). They give a sense
of vector magnitude and direction at
a point. E.g. at any point what is the
strength of flow and direction.
Lines, Arrows & Glyphs
These are the elementary visualization
techniques to make an overall picture of
the vector field. They put icons such as
lines, arrows or 3D objects to highlight
local features (magnitude, direction,
time dependency) of the field. The icons
are drawn directly at each data point.
Their scaling is controlled according to
magnitude and orientation of vectors.
Fig. 2: Arrowplot for Air flow over a city model
These methods are quite suitable
for 2D vector fields. In 3D, they
pose the problem of cluttering and
overlapping. As the number increases,
it is difficult to understand the
position and orientation of the field.
CSI Communications | November 2014 | 19
There is also no satisfactory scaling
mechanism exists. With large scaling,
they occlude each other. If the scaling
is fixed to study directions only, then
the magnitude information is lost.
Spot Noise & Line Integral Convolution
Spot noise and Line Integral
Convolutions (LIC) are texture synthesis
methods to visualize flow. Both employ
same
mathematical
principle
of
convolution to create the visual effect.
Spot Noise distributes a large number of
small intensity functions – called spots
– over the vector field. The shape of the
intensity functions is deformed by the
vector field. In LIC, a random noise is
convoluted with a piece of streamline. The
final result is a smeared image, which can
give the idea of fluid motion.
Flowlines: Streamline, Streakline,
Pathline
The flow lines visualize the flow starting
from a particular location. They are
calculated as integral lines that simply
assume the path traversed by a particle if
dropped in the vector field. The examples
of flow lines are Streamlines, Streaklines
and Pathlines. The lines are identical in
a steady flow, where the velocity field
remains constant over time. But, in
unsteady flow they are different. They
are displayed using standard graphical
primitives (points, lines, arrows).
Mathematically, the flowlines are the
solution of simple ordinary differential
equations (ODE) as given below for a 2D
vector field.
dx = vx(x(t), yt) dy = vy(x(t), y(t),t)
dt
dt
Intial Conditon: At t=0 x(0),
Fig. 3: 2D Line Integral convolution texture
applied to space shuttle[4]
Textured Splats
It is an extension of a popular volume
rendering technique, called splatting to
visualize vector data. It also uses texture
as a medium of visualization. Spaltting
is the projection of data volumes on a
2D viewing plane. The splats are finally
composited on top of each other to
produce the final image. Texture splatting
combines multiple textures and blend
them accordingly.
y(0)
The clear advantage with flow line
methods is the ease of implementation
and rendering. But the technique heavily
relies on seed point (start point of flow)
placement, which becomes a major
drawback. The efficient seed generation
is still an active area of research. A
significant amount of user’s involvement
and prior knowledge of vector field is
required to generate effective flow lines.
The other issue is related to number of
flowlines to be generated. With large
number, the computation becomes
expensive.
Streamlines represents a family
of curves that are tangential to fluid
flow everywhere. At any instant of time
streamlines provide a snapshot of a region
of the field. A streamline start from a given
initial position (seed point) and move
along the lines tangential to the vectors.
In an unsteady flow, streamlines keep
changing as the vector at each point varies
with time.
The streamlines mostly don’t
intersect each other, as there cannot
be two velocities at the same point.
The streamlines are easier to compute
compared to others, because they have
no temporal Interpolation. The technique
is better suited for static visualization.
The streamlines are further extended to
encode more spatial information in 3D
through the use of geometrical objects.
Some of the extensions are streamtubes,
streamribbons, streamsurfaces, streamballs and streamarrows.
The path lines are the curves
describing the traversals of all the
particles. They are identical to streamlines
in a steady vector field. The path lines are
drawn only in forward direction. But an
unsteady field has different pathlines and
streamlines. At every moment the vector
field changes, so the streamlines. This
makes the paths traversed by particles
different. The paths are determined by the
streamline at that time stamp.
The streak line is the locus of all the
particles passed through a certain position.
It is also similar to streamlines when the
flow is steady. In unsteady flow, when
particles are released from a fixed point
they travel along different paths because
of continuous change of vector field with
time. The collection of such points at later
instant of time gives a streakline.
Conclusion
Vector field visualization is an active topic
for research and analysis. An overview of
fundamental flow visualization techniques
are discussed in the current paper. The
aforementioned techniques are applicable
to both 2D and 3D vector fields. In 3D
Vector fields they generally pose issues
such as occlusion, Visual data density,
depth perception. People have come
up with suitable modifications - sparse
representations, color differences, semitransparency and many more to overcome
these issues.
Acknowledgement
We would like to thank members of
visualization wing and Director, ANURAG
for constant support and encouragement.
Fig. 4: Test tornado data rendered using the
textured splat[5]
Fig. 5: Streamlines of air flow over city model
CSI Communications | November 2014 | 20
References
[1] Charles D Hansen, Chris R Johnson
“The
Visualization
Handbook”,
Elsevier Academic Press, pp. 261-277,
2005.
[2] J
van
Wijk,
Spot
Noise,
www.csi-india.org
TextureSynthesis
for
Data
Visualization, ACM SIGGRAPH ‘91.
[3] B Cabral, C Leedom, Imaging Vector
Fields Using Line Integral Convolution
ACM SIGGRAPH ‘93.
[4] Roger A Crawfis, Han-Wei Shen, Nelson
Max, Flow visualization techniques
for CFD using Volume rendering, 9th
International symposium on flow
visualization, 2000.
[5] Roger A Crawfis, Nelson Max, Texture
Splats for 3D Vector and Scalar
Field Visualization, Proceedings
Visualization ‘93 IEEE CS Press.
n
About the Authors
Shri Dilip Kumar Dalei is Scientist at ANURAG, Defence Research & Development Organization (DRDO), Hyderabad. He received his Bachelor
degree in 2004 from NIT, Rourkela, India. He obtained his master degree in the field of Computer Science and Engineering in 2010 from
IIT, Kharagpur, India. His areas of interest are Scientific Data Visualization, Computer Graphics, 3D Displays and GPU programming. He has
publications in International and National Conferences.
Shri B. V. Hari Krishna Nanda received his M.Sc degree in Computer Science from Andhra University in 1992 and M.Tech in Computer Science
& Engineering from Acharya Nagarjuna University in 2010. He is presently working as scientist at ANURAG, Defence Research & Development
Organization (DRDO), Hyderabad. He has more than 20 years of experience in the area of Scientific Data Visualization, Computer Graphics
and Database Management.
Shri N. Venkataramanan received his M. Sc. in Computer Science from University of Mysore. He is presently working as Scientist at ANURAG,
Defence Research & Development Organization (DRDO), Hyderabad. He has more than 20 years of experience in the area of Scientific Data
Visualization and Computer Graphics. He has published research papers in various International and National Conferences.
CSI Communications | November 2014 | 21
Research
Front
Anchal Garg*, Madhurima**, Madhulika*** and Saru Dhir ****
*Assistant Professor, Department of CSE, Amity University Uttar Pradesh, Noida, India
**Assistant Professor, Department of Information Technology, Amity University Uttar Pradesh, Noida, India
***Assistant Professor, Department of CSE, Amity University Uttar Pradesh, Noida, India
****Assistant Professor, Department of Information Technology, Amity University Uttar Pradesh, Noida, India
A Walkthrough & Pathfinder for Research
Novitiates: Google Scholar Vs Microsoft
Academic Search
What Exactly Research is?
In our day to day life we are surrounded
by data, information, knowledge, opinions,
ideas etc. Initially we study, learn and use
them for our own advantage and later,
this information is disseminated for the
benefit of the society and the mankind.
Research is search for new knowledge.
When we extend the knowledge by
digging the unveiled information, we are
doing research. We often come across
this term through peers and while working
on it. Beginners in formal research often
encounter problems searching for relevant
scholarly papers[1]. They generally tend to
waste their enormous time in searching
and very often get baffled and frustrated.
2.
3.
Fig. 1: Showing the scholar.google.com page in the Google
Chrome Web Browser
Tutorial Insight
This short tutorial explains very precisely
and clearly how to begin and understand
the
searching
mechanism
while
researching in very simple steps. We
have explained the steps using Google
Scholar and Microsoft Academic Search.
Searching results are explained below
with the help of snapshots.
The steps below are showing the
searching tricks in Google Scholar.
Step 1- Go to Google Scholar by typing www.
scholar.google.com on the Web Browser
The Google Scholar is an excellent aid
for searching when we have to search for
different articles, thesis, books, abstracts
and research papers across different
disciplines[2]. The Fig. 1 below shows a
snapshot when you type www.scholar.
google.com in the Web Browser’s address
bar.
Step 2- Enter the topic to be searched (For
example artificial intelligence)
Then in the search field provided,
enter the topic for which you want to
fetch information. For example, in the
Fig. 2 below we have typed “artificial
intelligence” to get all the scholarly
articles related to it.
Fig. 2: Showing the topic “artificial intelligence” for searching
Step 3- Click on the
Search button. You will
get the list of scholarly
articles.
Further click on the
search button. This
will give access to a
list of scholarly articles
related to “artificial
intelligence” as shown
in the Fig. 3 below.
The scholarly articles
may include research
papers, thesis, books,
abstracts, etc.
Fig. 3: Showing the results or list of scholarly articles generated after
clicking on the search button.
The boxes represented in the figure are
explained below:
1.
CSI Communications | November 2014 | 22
4.
You can view the
paper by clicking on
this link.
Citations
for
a
particular paper can
be
seen
through
the “Cited by” link.
Here
“Cited
by
1903”indicates that
this paper has been
referred by 1903
paper [3].
People generally have
problems
writing
references. Click on
“Cite” in order to cite/
refer this paper in
your research paper.
You can find MLA,
APA and Chicago
referencing
styles
here. You can change
the referencing style
as required. Each
journal has its own
referencing style.
Clicking on “Since 2010”, you can
view papers published from the year
2010.
5.
6.
7.
It shows the Academic Publisher.
Here the Publisher is Elsevier.
It shows the name of the Journal.
Here the name of the Journal is
“Artificial Intelligence”.
It shows the year of publication. Here
www.csi-india.org
Fig. 7: Showing the topic “artificial intelligence” for searching
Fig. 4: Showing the results after clicking on “Cited by”
year in which the research paper was
published is “1987”.
Step 4- Click on “Cited by” to view papers
that have referred this paper in their research
papers.
For instance, if you click on “Cited by
1903” of the article titled “On agent based
software engineering”, then you will be
navigated to the webpage containing all
Step 5-Click on
“Cite” to refer this
article in your paper.
It
has
several
formats; you can use
any one of them to
add it as a reference
in your paper. Fig.
5 below shows the
Fig. 8: Showing various options for selection after clicking the search
button
Fig. 5: Snapshot of the web page when the user clicks on “cite”.
the scholarly articles that have referred
the above article. This new webpage will
help you find further work on the research
paper titled “On agent based software
engineering”.
Fig. 9: Search Result showing list of publications
snapshot when we
click on “Cite”.
Searching through Microsoft Academic
Search
Similarly, we can search the scholarly
articles on Microsoft Academic Search. The
Microsoft
Academic
Search
searches
research
papers
from various sources
based on the selected
conferences
and
journals.
Fig. 6: Showing the academic.research.microsoft.com page in the
Google Chrome Web Browser
The steps below are
showing the searching
tricks in Microsoft
Academic Search.
Step 1: To search
a
research
paper
on Microsoft Academic search, type
“academic.research.microsoft.com” on the
address bar.
Step 2: Next enter the topic to be
searched. As in previous case, type
“artificial intelligence”.
Step 3: Next, select the “Field of
Study”. For instance, select “Computer
Science” and click on search button as
shown in Fig. 8.
On clicking the search button, you
will be navigated to a new webpage
containing the following information as
shown in Fig. 9.
1. List of renowned authors in your
research area as shown in Fig. 9.
2. List of best rated conferences in your
research area as shown in Fig. 9.
3. List of Publications as shown in Fig. 9.
CSI Communications | November 2014 | 23
information on “International Joint
Conference on Artificial Intelligence”.
The famous authors, keywords and the
list of research papers published in IJCAI
(shown in Fig. 12).
Closing remarks
Fig. 10: Search Result showing list of publications
Fig. 11: Showing the information related to “Journal on Artificial Intelligence”. The most cited
authors, the keywords and list of research papers.
4.
List of best journals in your field of
study (Shown in Fig. 10.)
Scrolling down, we can see the list of
journals relevant to our research field.
Step 4: Clicking on the “AI” shown
in Fig. 10 (extreme left), we will be
navigated to new webpage. The Fig.
11 shows the information related to
“Journal on Artificial Intelligence”. The
most cited authors, the keywords and
list of research papers.
Similarly, clicking on “IJCAI” shown
in Fig. 9 (extreme left), we will be
navigated to the new webpage displaying
This article gives an overview for
effectively searching scholarly articles
using Google Scholar and Microsoft
Academic Search. This tutorial has
provided a step by step procedure for
efficiently searching the relevant articles
related to your research. Google scholar
offers a more user friendly interface as
compared to its counterpart Microsoft
Academic Search. Also, it is more
popular and up to date. Google scholar
has rich repository of academic materials
and citations than Microsoft Academic
Search. The main advantage of Microsoft
Academic Search over Google Scholar
is that more parochial proportionate
service. Also, the data processing in
Google Scholar is more time expensive.
Furthermore, Microsoft Academic Search
is more suitable search for retrieving
results from multidisciplinary fields[4] [5].
We sincerely hope the above
information will prove to be a useful aid to
your research navigation. We expect that
the information shared above facilitates
enough explanation to your unanswered
questions. We also hope that this concise
article will make your search for research
much trivial.
Acknowledgements
Authors are very thankful to Dept. of
Information Technology and Computer
Science & Engineering, Amity School
of Engineering and Technology, Amity
University Uttar Pradesh, Noida, India for
their support to carry out this study.
End Notes
[1]
[2]
[3]
[4]
Fig. 12: Showing information on “International Joint Conference on Artificial Intelligence”
CSI Communications | November 2014 | 24
http://scholar.google.co. in
www.researchgate.net/
http://en.wikipedia.org/wiki/Citation
Orduna-Malea, E, Ayllon, J M,
Martin-Martin, A, & López-Cózar,
E D (2014). Empirical Evidences
in Citation-Based Search Engines:
Is Microsoft Academic Search
dead?. arXiv preprint arXiv:1404.7045.
[5] Ortega, JL, & Aguillo, I F (in
press). Microsoft academic search
and Google scholar citations:
Comparative analysis of author
profiles. Journal of the Association for
Information Science and Technology.
n
www.csi-india.org
Anchal Garg has more than twelve years of teaching experience. She is working as Assistant Professor in Department of Computer Science
and Engineering at Amity University, Noida. Her major research areas are information systems, business process management, data mining,
process mining and higher education and research. She is a member of ACM and IET (UK). She has number of international and national
publications to her credit.
About the Authors
Madhurima did her Master’s Degree in Information Technology from Guru Gobind Singh Indraprastha University, New Delhi. She is at
presently doing her Ph.D in Information Technology from Sri Venkateshwara University. She has 7.6 years of working experience in academic
field. She has number of international and national publications to her credit. She has published one book with the title “Computer Networks”
with Laxmi Publications. Her M.Tech work has been published as a book titled “Video Object Tracking” by LAP LAMBERT Academic
Publishing GmbH & Co. KG, Germany. She is a member of CSI and ACM. Her primary research area includes image segmentation, object
tracking and object oriented testing, ajax and databases.
Madhulika is working as Assistant Professor in Department of Computer Science and Engineering at Amity University, Noida. She holds
diploma in Computer Science Engineering, B.E in Computer Science Engineering, MBA in Information Technology, M.Tech in Computer
Science & Pursuing Ph.D from Amity University, Noida. She has total 8 years of teaching experience. She published almost 15 Research
Papers in National, International conferences and journals. Her primary research area includes video object tracking and soft computing
techniques.
Saru Dhir did her Master's Degree in Computer Science and Technology from Amity University Uttar Pradesh, Noida. She is at presently
doing her Ph.D in Engineering. She has 5 years of working experience in academic field. She has number of international and national
publications to her credit. She is a member of IEEE. Her primary research area includes Software Engineering, Software Testing, Agile
Methodology and databases.
CSI Communications | November 2014 | 25
Article
Amol Dhumane* and Dr. Rajesh Prasad**
*Assistant Professor, Department of Computer Engineering, NBN Sinhgad School of Engineering, Ambegaon (Bk), Pune
**Professor & Head, Department of Computer Engineering, NBN Sinhgad School of Engineering, Ambegaon (Bk), Pune
Active Queue Management
Congestion is a complex process. It
is really a hard task to define congestion
but it can be easily sensed through the
tainted performance of the network.
Sometimes, user experiences long delays
in the delivery of messages, perhaps
significant losses due to buffer overflows.
One of the reasons of high packet loss rate
is the disappointment from the network
to provide early congestion notification
to the sources of traffic. It is essential
to avoid high packet loss rates in the
internet. Packet loss means the wastage
of the resources that have been used by
the packet. Normally network congestion
occurs when a link or node is carries
huge amount of data which deteriorates
networks quality of service.
In the traditional Droptail queuing
algorithm a mechanism of simple queue is
used but the problem with the algorithm
was, it was not able to handle the bursty
traffic.
To sort out the issue of bursty traffic
large buffers at the routers were used, but
it introduced the problem of high queuing
delays in the network especially during
the congestion. Small buffers reduced the
problem of large queuing delays but added
a problem of high packet loss, which will
further result into underuse of the link in
Droptail queues.
So we want a fair and sensible
tradeoff between high link utilization and
low queuing delays.
These limitations of endpoint
congestion are highlighted by peer to peer,
multimedia and short lived web traffic.
These applications mainly required less
delay and jitter, high bandwidth and less
packet drop. This created a need of new
queue management technique against
traditional Droptail queues, which becomes
the point of origin for the development of
new queue management technique known
as “Active Queue Management”.
Since the introduction to IP networks
in 1993 there has been a steady stream
of research output with respect to Active
Queue Management (AQM).
Congestion occurs when the total
requirement of a resource, (e.g. link
bandwidth) go beyond the capacity of the
available resource. Congestion results
into: high latencies in data transfers;
wasted resources due to packet losses;
and network failure in extreme case for
which there is essentially no data transfer
through the network: the throughput
drops to zero and the response time
goes to infinity. The aim would then to
be control congestion or more preferably
avoid congestion. Initial work is done in
the area of congestion control primarily
focused around the responsive flow
(TCP) traffic.
Endpoint
congestion
control
mechanism was employed at end points of
the network by keeping the core network
simple. In contrast AQM mechanism had
taken the advantage of powerful hardware
technology and tried to implement it
in the core network components (such
as routers).
Unlike Droptail, AQM is a proactive
congestion control scheme by which the
network informs the sources when the
incipient congestion is detected. The
network can inform the sources by means
of packet drops or with the help of Explicit
Congestion Notification (ECN) marks
inside the packet headers.
AQM marks the packets according
to the severity of the congestion. When
congestion increases the probability of
packet dropping (or marking) will also
increase.
Increase in the packet marking
reduces the transmission rate of the
sources which results into decreasing
the queue length and reducing the packet
losses by preventing the queue overflow.
As AQM algorithm used to drop
the packets randomly before buffer
overflow, it avoids the problem of
“global synchronization” that was
present in the Droptail algorithms
AQM Components are shown in the
diagram below:
Congestion indicator gives signal
of the occurrence of congestion, control
CSI Communications | November 2014 | 26
function proposes the strategies to
deal with the congestion and feedback
mechanism makes the sources aware
of it and asks them to adjust their pace
of transmission to bring the congestion
under control.
At earlier stages AQM algorithms
were entirely based on the queue length
as a congestion indicator, but as the
research progressed it has been found
that there are harms in using only queue
length as a congestion indicator. So for
detecting link congestion, AQM may
utilize any combination of the following
parameters: queue length (average
or instantaneous), input rate (packet
arrival rate or queue drain rate), queuing
delay and events of buffer overflow and
emptiness.
Congestion
control
function
calculates the probability of packet
marking (or dropping) based on the
level of congestion shown by the various
congestion indicating parameters and
decides which packets to drop (or mark).
Congestion control function can use
different parameters or blend of different
parameters. For getting better results, it
is essential to tune the congestion control
parameters according to the current traffic
condition in the network.
One important point here to note that,
in current AQM mechanisms, different
control function types are suited for
different traffic environments.
There is not a unique control function
that is suitable in all network conditions.
There are two major approaches to
in designing the control function: heuristic
approach and control theoretic approach.
Heuristic approach is completely
based on intuition of the algorithm
developer. There is always simplicity in
design and implementation to certain
extent. Mathematical modeling is generally
not present in this approach and the control
parameters are tuned manually depending
on the network traffic conditions.
Sometimes this approach may lead the
system into unstable state. As heuristic
approach is not always trustworthy, control
theoretic approach came into. It is based
on mathematical modeling where there
is always a explanation exists for the
www.csi-india.org
About the Authors
selection of the parameter values. Most
of the control theoretic approaches are
responsive flows centric.
Finally Feedback Mechanism informs
the sources about the incipient or transient
congestion in the network with the help of
ECN notification or packet drops (using
TCP’s triple duplicate acknowledgement
policy or retransmission policy).
Issues upsetting AQM performance:
a) Buffer size and link capacity
b) Traffic load and RTT
c) The manner in which flows enter
into the queue i.e. in-synch or out-ofphase way.
d) Routing matrix gain (it is used
to
represent
the
topologies
mathematically).
Network
robustness is inversely proportional
to this routing matrix gain.
e) The presence of short lived flows and
unresponsive flows.
f) Reverse path asymmetry.
Some of the existing AQM algorithms
are listed below:
Heuristic
Approach
Control-Theoretic
Approach
Random Early
Detection
Proportional Integral
(PI) controller
Adaptive
Random Early
Detection
Proportional
Derivative (PD)
controller
BLUE
Proportional
Integral-Derivative
controller
BLACK
Gain adaptive smith
predictor
Hyperbola RED
PI-PD
Load/Delay
Controller
Various fuzzy logic
controllers.
LUBA
And many more
And many more
Acknowledgements
We are thankful to our Principal Dr. S
D Markande, Prof. S P Patil, Head, IT
department and entire NBN Sinhgad
School of Engineering for their support in
our work.
References
[1] Richelle Adams, “Active Queue
Management: A Survey”, IEEE
Communications Surveys & Tutorials,
Vol. 15, No. 3, Third Quarter 2013, pp
1425-1476.
[2] S Floyd and V Jacobson, “Random
Early Detection Gate-ways for
Congestion
Avoidance”,
IEEE/
ACM Transactions on Networking,
1(4):397-413, Aug. 1993.
[3] W Feng, D Kandlur, D Saha and
K Shin, “A Self Configuring RED
Gateway,” in Proc.IEEEINFOCOM’99,
March (1999), pp.1320-1328.
n
Amol Dhumane has received his masters (M.E. Computer Engg.) from Bharati Vidyapeeth University, Pune. He is having 10 years of teaching
experience. His area of interest includes congestion control and IoT. He has published more than 10 papers in national and international
conferences. He is a life member of ISTE.
Dr. Rajesh Prasad has received masters (M.E. Computer Engg.) from College of Engineering, Pune and his doctorate from Swami Ramanand
Tirth Marathwada University, Nanded. He is having 18 years of experience. His area of interest is Soft computing, Text Analytic and
Information management. He has published more than 40 papers in national and international journals. He is a life member of IAENG, CSI
and ISTE.
CSI Communications | November 2014 | 27
Article
Sumith Kumar Puri* and Dr. H K Anasuya Devi**
*Bachelor of Engineering [Information Science and Engineering], Sri Revana Siddeshwara Institute of Technology, Bengaluru
**Researcher & Professor, Artificial Intelligence, CCE, Indian Institute of Science, Bengaluru
Information Technology to Curb Piracy
in Bollywood
This was it! The huge 1.5 GB download
was done. He observed that he was the
fastest seed, as he had close to a 60Mbps
connection. Albert now opened his
facebook friends' list and shot across a
message, announcing the link to all. By the
way, this was the latest in the series of scifi superhero movies that he and majority
of his peer group were awaiting. This
being the third day of the official release.
All set to watch the movie now, Albert
opened up his favorite video player on the
computer. But his joy was shortlived... The
movie quality did not live up to what was
mentioned on the site. It had promised
'HD 1080p video and DVD 384kbps
audio' as part of its title and description.
In spite of wanting to tear his hair apart at
the jumping of scenes, he plans to watch
the entire movie. The audio being hardly
audible, the video jittery and sometimes
tilted, the only pleasure he seemed to
have was his favorite cheese popcorn
and a glass of cola. Interestingly, the last
time Albert had ever paid for a movie or
music clip online was about three years
ago. That too, was actually an accidentally
downloaded clip, through auto-pay, as he
had his credit card registered on one of the
sites.
This is a normal daily or weekly
scenario of almost all of the netizens
who are within the teenage, graduate or
starting professional levels. Given the
same situation a few years earlier, we
would have seen the likes of Albert along
with his friends queuing up at a movie
theatre or waiting eagerly for the video
tape or compact disc release of the movie.
As per one of the reports the total
loss to Bollywood alone, in the year 2010,
was close to a billion dollars. This loss
was also attributed to camcorders' being
used in theatres to create pirated compact
discs and digital video discs. So, Hasn't
any one taken any steps to get back what
should be rightfully theirs? Though there
are many options including enhancing
the already available digital watermarking
techniques, policing and governance. But
in the reality of things, How much of this is
actually enforced and How much does the
film creator actually bother about? As per
a study by the Harvard Business School in
2011, the implementation of some of the
steps could actually outdo the financial
benefits associated with it. Hence,
sometimes even the creators overlook
these measures.
So, How about implementing a pure
information technology solution or system
that is not only helpful for the creators, but
also a fun and legal way for the users. My
proposal is not only to create a software
mechanism that promotes a legal way, but
one that also that is cost effective, cheaper
to implement, generates user interest,
direct business to customer, promotes
legality and is enjoyable. What's more,
it promises to even re-employ pirates as
legal vendors of the same media, albeit at
a lesser profit to them.
CineCat is what I call this software
idea, which will allow people of all ages
to easily and legally access any media,
primarily related to Bollywood. CineCat
acts not only as a directory or database
of movie information but as an online
distributor of, primarily, movies and music.
Unlike the YouTube's of our era, it works in
a mode that is even more closely tied to
Commerce, Anti-Piracy and Bollywood.
The simple driving process is that when
the movie is ready to release and it hits off
a calculated cooling period, the software
allows the creator to provide an online
release.
With
proprietary
security
mechanisms that are tied to the hardware,
it aims to distribute the release at equal or
lesser the cost to download or purchase
a pirated movie disc. The delivery can be
either digital or on an actual disc, through
post. Since the security mechanisms
are tied to hardware or the registered
player and to a specific user, they
cannot be duplicated or cracked easily.
The calculation of a cooling period is
done based on history, at times being a
calculated risk. It also aims to generate
further interest in the users who await
that ultimate digital experience and
would outright deny any pirated media. It
will create further avenues for revenues,
CSI Communications | November 2014 | 28
a legal approach which is attractive to
educated users and curb piracy effectively.
All of these can be combined with many
other digital and traditional marketing
techniques, to maintain users' interest.
The difference in the solution that we
are proposing is that an online release
would not only use CineCat as a form
of distribution but also as a player and
security control mechanism. It is not only
in India, but in even bigger countries such
as the United States, tying movies with
ordinary post or digital delivery is a tried
mechanism that has got good response.
So, the loose ends that remain in
such a proposal is what is that appropriate
time when the online release would outdo
or curb piracy to abysmal levels. Also, it
should allow the creator to gain revenues
that is rightfully theirs. Then there is the
issue of periodic or perpetual rights on
the creation itself. These are only minor
issues or calculated risks, that would
never take away the normal theatre,
multiplex or big screen watchers. Since
CineCat would allow a digital database,
which is equivalent of a movie wikipedia
and a lifelong availability of the original
or high quality creation itself – the issue
of rights is a decision to make that would
only eventually give benefits, now starting
from an earlier period.
But why would one chose a tedious
process as an user? What would appeal to
him, like our own Albert?
Albert is one of the normal graduate
or undergraduate that you would come
across, who has a group of geeky friends.
Albert is not the only one who prefers the
highest quality of video, audio or even
a trailer – his entire peer group does. At
the same time, they all vie to create the
best repository or collection of music or
movies. They are also the ones who don't
bother to spend a very small amount
of money, not only because it provides
quality – but also because it is now cool to
pay for this form of viewing. Also, being an
avid fan of sci-fi superhero movies, he is
even more enthusiastic of using CineCat.
He expects an autographed compact
disc to be delivered to him, and doesn't
www.csi-india.org
lose hope that he may win the contest on
CineCat to meet the superhero (actor)
himself. Even other buyers off the road
of a pirated compact disc, have now an
option – even if they do not use a personal
computer. This will mostly be tied to
the serial number of the playing device.
Since most household's may own a single
device (or at the most two) – a suitable
serial number can be provided to encode
it for playing with CineCat loaded onto the
movie disc itself.
What about the other side, that of
the creator? Does he see anything that
is different or lucrative from what he was
doing till now?
This could lead to a cheaper form of
distribution, that may infact itself pay for
itself in many ways, because of including
advertising to one extent – with CineCat
taking care of almost all processes which
include online security, distribution,
proprietary mechanisms, it would keep a
constant check (and update) on crackers
and any form of threats to its reliability.
We, in this article, are not covering
or studying any other application or
software system that is already in place
that is close to what CineCat does or is
intended to do.
But then there is a common sense
that may prevail in many minds. Why
curtail some piracy, When it is actually
good? There maybe positive vibes due to
the piracy. Well, there are steps to actually
take that too into account, as is done to one
extent even today. CineCat can be used as
an extended trailer, an extended promo to
ascertain not only the content, interest of
the user but also the quality, security and
reliability processes. Most of the times for
an user a very short wait is mostly worth
the watch (beyond the trailers).
All of the mentioned policies and
compared to the prevalent usage, and
of that do exist a small percentage are
enforced. One aspect of CineCat is also a
stronger enforcement of cyber laws related
to media and information technology. It
could lead to a combination of banning of
certain sites, internet service providers,
torrent sites, peer-to-peer sharing, file
sharing sites and even sale of pirated cd's.
CineCat can curb or provide alerts for
many of the online piracy methods and
also provide a way to tackle the offline
menace (out of this discussion).
Albert is back to watching the movie,
which is almost at the end. No, He isn't
the one who has lost hope... All this while,
with the movie quality not capturing his
interest - he spent some time reading
this article about CineCat. He is the one
eagerly awaiting the first time he gets to
see an online movie release on CineCat!
Glossary of Terms
CineCat was a Software Application
Idea based on the concepts mentioned
in this article. This was to be built by the
now closed startup, named TechArmy,
of one of the authors.
About the Authors
advertising, marketing, delivery and even
policing. Some of these ideas can be difficult
in implementing because of curtailed
internet speeds, but then there always is the
option of delivery of discs by post.
Also, since CineCat relies on
ideas are also applicable to music,
events, short films and any other form
of distribution that may need exclusive
rights, or are subject to piracy.
We are evolving into a society
where the cyber laws are too loose when
Interesting Read (Google Search Terms)
Hollywood and Bollywood Join Arms...
Protecting against the Pirates of Bollywood
Bollywood & Rights, Piracy Is Mostly Via
Torrents
Bollywood No Longer Worrying About
Piracy...
ETC Fights against Movie Piracy...
Film Piracy Funding Terror in India...
Piracy cost Bollywood $959m: Report
YRF digitizes Dhoom 3 to combat piracy
n
Sumith Kumar Puri holds a Bachelor of Engineering [Information Science and Engineering] from Sri Revana Siddeshwara Institute
of Technology, Bengaluru. He has also completed his Proficience [Cryptography & Network Security] and Proficience [Intelligent
Agents] from the Indian Institute of Science, Bengaluru. Sumith has recently completed a Part-Time Acting Course from Actor
Prepares, Mumbai. He has more than 9 years of experience in various facets of Software Development. You can reach him at sumith.
puri@sumithpuri.me
Dr. H K Anasuya Devi received her Ph D degree from the Indian Institute of Science, Bangalore in 1985. Her research interests include
– Computational Linguistics and Language Techonology, Artificial Intelligence and Expert Systems, Remote Sensing Applications
and Geographic Information Systems, Archaeology and Epigraphy, Sports. She possesses extensive research experience with over
25 years in inter-disciplinary areas, including 15 years of teaching experience at graduate and post-graduate levels. Over the last
decade and half, she has guided over 150 students in their research projects, published over 298 papers in all forms. You can reach
her at hkadevi@yahoo.com
CSI Communications | November 2014 | 29
Article
Dr. M Hanumanthappa*, Mrs. S Regina Lourdhu Suganthi** and Mrs. Rashmi S***
*Associate Professor, Department of Computer Science and Applications, Bangalore University, Bangalore
**Research Scholar, Department of Computer Science and Applications, Bangalore University, Bangalore
***Research Scholar, Department of Computer Science and Applications, Bangalore University, Bangalore
Protection of Software as Intellectual Property
Each system of Law has its own
definition of what can be bought and
sold, i.e., what is tangible property?
Therefore, by virtue of implication
of law, the ownership of property
such as tangible goods is defined and
apportioned. Whereas, the ownership
in Intellectual Property (IP), represents
a proprietary right in intangible products
of human minds, which are generally
referred to as “knowledge goods” and
the ownership of IP, is also defined by
law. However, as against the existence
of laws that define and protect the
ownership of tangible goods, a separate
form of legal protection is recognized
for the protection of knowledge goods,
since knowledge goods are intangible,
non-excludable
and
inexhaustible.
Accordingly, the laws that govern IP
define ownership rights in IP. Therefore,
the term “Intellectual Property Rights
(IPRs)”, is bundle of legal rights,
granted for knowledge goods, such
as inventions (Patent Right), brand
names (Trademark), artistic and literary
creations (Copyright), aesthetic designs
(Designs) and industrial secrets (Trade
secret).
Traditionally, computer software is
recognized as a literary work and hence
its legal protection is provided under
the law of Copyright (albeit statutory
registration is not mandatory). Prior to the
unbundling of computer software from
hardware and where there was no distinct
value for software, copyright protection
was deemed sufficient. It is important to
understand that Copyright Law provides
protection only to expression of ideas
(non-functional ideas) and not to ideas
(functional ideas) per se.
However, with the unbundling of
computer software from hardware in 1970s
and software assuming an independent
work of creation, the protection of
software, only under the law of Copyright
was found grossly inadequate.
In Lotus Development Corporation v.
Borland International, U.S. Supreme Court,
while deciding a copyright infringement
case, in a dispute between Borland and
Quattro Pro (Lotus 1-2-3), exposed the
limitation of legal protection of copyright
for software. In this case none of the source
code or machine code that generated the
menus was copied, but the names of the
commands and the organization of those
commands into a hierarchy were virtually
identical. The Court, while denying the
Copyright protection for the software,
where the functional aspects of software
are claimed, pronounced that copyright
was limited only to the protection of
non-functional aspects of software, as a
literary work.
Copyright law provides protection
for literary works, which are original
and non-functional in nature. With the
separation of software from hardware
and the greater value of realization for
software as an intellectual asset, a need
was felt to provide a comprehensive IP
protection for the functional aspects
of software and an effort was made to
seek protection of software as invention
(functional idea).
Accordingly, protection of software
in the form of Patents gained momentum,
in countries like US and Europe, by
considering them as inventions rather
than literary works.
Legal protection in the form of Patents
is granted by a State (Registration is
mandatory for exercising the legal rights),
for inventions, that are New, Inventive
and having Industrial Application. In
addition to the satisfaction of these legal
conditions, it is also imperative to ensure
that such inventions do not fall under the
category of ‘non-statutory inventions”. In
other words, patent statutes of countries,
do have negative list, with which they bar,
explicitly, certain categories of inventions,
irrespective of such inventions being
Novel and Inventive.
One such category of inventions,
which are often being considered as
“non-statutory” is software. For instance,
under US Patent Law (35 U.S.C. §101), the
subject matter of inventions relating to
laws of nature, natural phenomena, and
abstract ideas, are specifically excluded
from patentability. The rejection of
software as a patentable subject matter
is based on the premise that they are
abstract ideas and hence not patentable.
US Supreme Court in Gottschalk v.
Benson(409 U.S. 63 (1972)) ruled that a
process claim directed to a numerical
CSI Communications | November 2014 | 30
algorithm, as such, was not patentable
because the patent would wholly preempt the mathematical formula and in
practical effect would be a patent on
the algorithm itself and hence it would
amount to allowing a patent on an
abstract idea.
However, the US Supreme Court, in
the case Diamond v. Diehr, 450 U.S. 175
(1981), held that controlling the execution of
a physical process, by running a computer
program did not preclude patentability of
the invention as a whole. In other words,
the mere presence of a software element
did not make an otherwise patenteligible machine or process un-patentable.
This decision of US Supreme Court opened
up vistas for the protection of software as
patents in the USA.
Similarly, in Europe, in terms of
European Patent Convention (EPC), (Art
52 of EPC), "programs for computers" are
not regarded as inventions for the purpose
of granting European patents, but this
exclusion from patentability only applies
to the extent to which a European patent
application relates to a computer program
“as such”. In other words, inventions
relating to claiming of “mere algorithms”
fall foul of this provision of law. In
Europe, patents for software inventions
are granted, as long as they have new
“technical effect” and they do not relate to
the implementation of normal functioning
of computer hardware.
The term “technical effect” is
generally interpreted to include, solution
to a technical problem, higher speed,
reduced hard-disk access time, more
economical use of memory, more efficient
data base search strategy, more effective
data compression techniques, improved
user interface, better control of robotic
arm, improved reception/transmission of
a radio signal etc.
In India, the subject matter of the
inventions relating to “mathematical or
business methods or a computer program per
se or algorithms”, fall outside the purview of
patent protection. Nevertheless, software
can be protected as literary work under
Indian Copyright Law. In view, Indian
Patent Law governing software inventions
are similar to European Patent Law, Indian
Patent Law, also reckons “new technical
www.csi-india.org
effect” in considering the patentability of
inventions claiming computer software.
To conclude, Software as an IP, needs
to be framed as an invention and adequate
protection in the form of patent right is highly
desirable, since patents offer comprehensive
protection for functional ideas. If necessary,
dual IP protection for software in the form of
Patents and Copyrights can also be explored
to cover both functional and non-functional
aspects of software.
References
[1]
Manish Arora, “Guide to Patents Law”,
Universal Law Publishing Co. Pvt. Ltd, 2002.
[2] P Narayanan, “Law of Copyright and
Industrial Designs”, Eastern Law House,
2002.
[3] Frederick W. Mostert& Lawrence E
Aploson, “ From Edison to iPod”, Dorling
Kindersley Limited, 2007.
Web References
[1]
http://en.wikipedia.org/wiki/Lotus_Dev._
Corp._v._Borland_Int'l,_Inc.
[2] http://supreme.justia.com/cases/federal/
us/409/63/case.html
[3] http://www.invispress.com/law/patents/
diehr.html
[4] h t t p : //e n . w i k i p e d i a . o r g / w i k i /
Gottschalk_v._Benson
[5] h t t p : //e n . w i k i p e d i a . o r g / w i k i /
Diamond_v._Diehr
[6] http://en.wikipedia.org/wiki/Software_
patents_under_the_European_Patent_
Convention
n
About the Authors
Dr. M Hanumanthappa is currently working as Associate Professor in the Department of Computer Science and Applications, Bangalore
University, Bangalore, India. He has over 15 years of teaching (Post Graduate) as well as Industry experience. He is member of Board of
Studies/Board of Examiners for various Universities in Karnataka, India. He is actively involved in the funded research project and guiding
research scholars in the field of Data Mining and Network Security.
Mrs. S Regina Lourdhu Suganthi is a Research Scholar in the Department of Computer Science and Applications, Bangalore University,
Bangalore. She has two decades of rich teaching experience in the field of Computer Science. Areas of interest include Data Mining, Image
Processing, Algorithms and Problem solving. She also assists in computer related patent drafting work.
Mrs. Rashmi S is a Research Scholar in the Department of Computer Science and Applications, Bangalore University, Bangalore, India. She
also has over 3 years of teaching as well as Industry experience. Her specialization in research is Data Mining. She has published several
papers in various National and International conference and Journals.
CSI Communications | November 2014 | 31
Practitioner
Workbench
Wallace Jacob
Senior Assistant Professor, Tolani Maritime Institute
Programming.Tips() »
Fun with ‘C’ Programs – Reversing a String using a
Bitwise Operator
The program and its output below exemplify a novel method of
reversing a string using one of the bitwise operators:
Program listing one
#include<stdio.h>
#include<string.h>
main()
{
char str[80];
int i, len;
printf(“Enter a string (less than 80
characters): “);
scanf(“%[^\n]s”, str);
printf(“\nBefore reversal, str= %s”, str);
len=strlen(str)-1;
for(i=0;i<len;i++,len--)
{
str[i] ^= str[len];
str[len] ^= str[i];
str[i] ^= str[len];
}
printf(“\nAfter reversal, str= %s”, str);
return 0;
}
The underlying logic is explored below:
c
\0
str[0] str[1] str[2] str[3] . . .
str[79]
If we convert each character of str, into its corresponding ASCII
About the Author
98
99
0
str[0] str[1] str[2] str[3]
...
str[79]
The binary representation of the ASCII codes would be as follows:
01000001 01100010 01100011
str[0]
str[1]
str[2]
00000000
str[3]
...
str[79]
Now let us dissect the statements:
str[i] ^= str[len];
str[len] ^= str[i];
str[i] ^= str[len];
In the first run of the for loop, it would imply:
i.e. the contents of str[0] and str[2] have been interchanged.
The rest is simple iteration using the for loop.
The string (Abc) would be stored as follows:
b
65
str[0] ^= str[2]; /* str[0] = 01000001 ^
01100011; i.e. str[0] = 00100010 */
str[2] ^= str[0];
/* str[2] =
01100011 ^ 00100010; i.e. str[2] =
01000001 */
str[0] ^= str[2]; /* str[0] = 00100010 ^
01000001; i.e. str[0] = 01100011 */
Sample output:
Enter a string (less than 80 characters):
Before reversal, str= Abc
After reversal, str= cbA
A
code, then the storage can be visualised as follows:
[Why should we use bitwise operators? For the simple reason
that bitwise operations are faster and use less power because of
reduced resource usage.]
n
Wallace Jacob is a Senior Assistant Professor at Tolani Maritime Institute, Induri, Talegaon-Chakan Road,
Talegaon Dabhade, Pune, Maharashtra. He has contributed articles to CSI Communications especially in the
Programming.Tips section under Practitioner Workbench.
E-mail: wallace_jacob@yahoo.co.in
Office Contact No: 02114 242121
CSI Communications | November 2014 | 32
www.csi-india.org
Practitioner
Workbench
Umesh P and Silpa Bhaskaran
Department of Computational Biology and Bioinformatics, University of Kerala
Programming.Learn("R") »
RStudio- Studio of R
RStudio is one of the most popular open source Integrated
Development Environment (IDE) for R. Both Desktop version
for Windows, Linux and Mac and Server version of RStudio are
available. RStudio has both Open Source Edition and Commercial
License. The open source edition is supported by the Community
forums only, whereas commercial users will get email support for
eight hour response during business hours. RStudio is written in
the C++ programming language and the Qt framework was used
to develop its graphical user interface. RStudio makes the life of
data scientist easy by offering an open source and enterpriseready professional software.
Let us have a look at the RStudio IDE.
RStudio has mainly four windows –
1. Source Window (Data objects in Spreadsheet form).
2. Console Window, where you can type commands and see
output (as in the R environment )
3. Environment and History Window- Environment Tab shows
all the active objects and history tab shows a list of commands
used so far. By using the environment tab, you can
import/ load data into R.
4. File, Plot, Package, Help and Viewer area.
You can also make changes of the look and feel of
the environment by editing the Global Options (go to
tools> Global Options) of RStudio.
R studio provides syntax highlighting facility and
line numbering which help you in writing code in R.
Importing data from various resources including web
goes interactively in RStudio. You can add packages by
a single click on the IDE and load the package into the
environment by checking the box. Also you can view
plots and export figure into the format you wish to.
A new R script can be created by using
RStudio by either clicking the icon or from the file
menu. You can also create a new text file from the
Rstudio interface itself. RStudio also supports C++
programming.
One of the significant features of RStudio is that
you can make presentations and html documentation
using it. A particular block of code in R can be inserted in the html
code and you can run the block. In the documentation, you can
show the code and gives its plot/result.
In a similar way, you can also use RStudio as an editor to
produce documentation of your R experiment or tutorial and save
it in PDF.
RStudio has a number of packages like Packrat, Shiny, ggvis,
dplyr, knitr, R Markdown, which are actually assets of statisticians.
Packrat is one of such package which helps you to handle the
package dependencies. This makes the package installed in your
machine portable and thus you can move your project easily and
reproduce result in some other machine. Shiny helps to create
interactive web applications on R. We will discuss more on Shiny
in the next tutorial.
n
CSI Communications | November 2014 | 33
IT Industry
Perspective
Prof (Dr.) D G Jha
Professor & Area Chairperson – IT; Programme Coordinator-MCA
Future of Medical Transcription Jobs in
India - Need to Extend it beyond Record
Generating Process
The real India still lives in villages
and farming continues to be the main
occupation of the people of this nationa country among the largest agriculture
producers in the world, a nation that
survives necessarily because of the agro
sector; INDIA is and will continue to
be viewed globally as a giant granary.
However, this perception that India is
still only an agro-based nation is fast
changing and going by the recent policy
statements of Government of India, it
definitely reflects the determination that
it is ready to recognise that the services
sectors are slowly but surely occupying
the drivers’ seat in the economy. Indian
government official plan for 2020 says,
“Our vision of India in 2020 is of a nation
busting with energy, entrepreneurship and
innovation” (Yojana Aayog;-Planning
Commission.2000, p21). This is an
opportunity.
Given the kind of support the
government
provided
to
service
sector specially the ITES (Information
Technology Enabled Services) sector,
it certainly has managed to accelerate
India’s transformation from an agrarian
economy to a service economy to a
certain extent. For the sector to thrive
further, it should be understood that ITES
holds enormous promise for India but at
the same time it has to be handled with
utmost care. Having, a powerful revenue
generating model and support policy of the
government alone will not be the primary
success determining factor; the ability to
manage human resources and maintain
quality will be paramount. ITES is a global
industry – sourced locally – and India can
succeed only if it is globally competitive
(Dhar-Patel & Vishwanath eds. 2002, p1).
IT Enabled Services (ITES) is defined
as all those services which can be
delivered to the end user from a remote
location with the aid of information
technology (Dhar-Patel & Vishwanath
eds. 2002, p7). In simple terms, the set
of processes that can be outsourced
and (can be) enabled with the help of
Information Technology can be defined as
Information Technology Enabled Services
or more simply as ITES. The sectors
(indicative) that can be classified as the
part of ITES are: Call Centers; Back Office
Processing; Medical Transcription (MT);
Geographic Information System (GIS);
Knowledge Process Outsourcing (KPO).
The Economic survey of 2012-13 indicates
four major sub components of IT/ITeS
industry as: IT services, business process
outsourcing (BPO), engineering services
and R & D and software products. (http://
indiabudget.nic.in. Economic Survey 201213, p223)
Importance of Medical Transcription
“Health is one of the most informationintensive businesses you will find, and
that information can have a direct impact
on the quality of patient care” notes Prof
Michael Smith – editor of Healthcare
Informatics Journal (cited in Brady et al.
2001, p118).
Since, the entire healthcare industry
in the US revolves around insurance; the
detailed, comprehensive, accurate and
complete patient records are needed
for processing insurance claims making
medical transcription one of the most
challenging and fastest growing aspects
within healthcare.
Medical Transcription assumes
greater importance as it is concerned with
preserving each and every patient-related
document, including the doctor’s note.
•
A patient visiting a clinic may require
investigation by physicians with
different medical specialties, nurses,
therapists, and technicians, all of
whom will record observation or test
data in separate files
•
During examination the patients may
be asked about his/her condition or
lifestyle by all of the above involved
in diagnosis
•
A patient upon investigation may
be prescribed several drugs and
CSI Communications | November 2014 | 34
treatments by different physicians
Each of these responses can get recorded
in different ways and in different files.
(Nair (Dr) cited in www.chillibreeze.com).
Reasons for Outsourcing of Medical
Transcription Process
The processing of insurance claim
requires comprehensive documentation of
the patient’s medical consultation history
and records of the every interaction with
physicians. Most of such documents that
recorded the interaction of the patients
and the physicians were handwritten in
the past and that made at times difficult
for insurance auditors and lawyers to
decipher and interpret. They started
insisting on the typed medical records.
This gave birth to the concept of Medical
Transcription. An efficient record-keeping
system thus becomes important and
indispensable (especially in the US) since
insurance claims form an integral part of
the medical industry.
Transcribe (across + write in Latin)
means writing something out in full
from notes or shorthand. It also means
transferring of information from one way
of storing it on a computer to another,
or from computer to an external storage
device.
Medical records noted or dictated
by the doctor (or his assistants or nurses
or a clinician), physical therapists, health
professionals such as dietician or social
psychologist are loaded into the tape or
onto the digital voice processing system
precisely and aptly transcribed i.e.,
converted into the word document by MT
or MLS (Medical Language Specialist). All
the clinic notes, office notes, operative or
consultation notes, patient’s history and
physical reports, psychiatric evaluations,
laboratory reports, x-ray reports and
discharge summaries after transcription
are proofread up to 98% accuracy before
being uploaded back to the doctor’s office
or clinic (Dhar-Patel & Vishwanath eds.,
2002, p.100) .
www.csi-india.org
Since the US is the only country in
the world that has such stringent policies
on maintaining medical records, medical
transcription is largely the US based
industry. The need to reduce the cost of
administering these records pushed many
hospitals towards adopting electronic
formats for documentation. India, for its
excellence in interpretation and large
English-speaking population base, became
the preferred destination (till recently) for
the US medical practitioners.
The documents are expected to be
returned within the stipulated time frame
of 24 hours and time zone difference that
India has with US makes it quite easy.
Apart from need to cut cost, the other
apparent reason is increased reliance on
core competence that prompted hospital
administration to outsource this job. The
benefits of ordering out for the medical
practitioners are:
•
Concentration on core activity
•
Production based compensation
•
Minimized expenses on perks and
other employee benefits
•
Elimination of recruitment and
training expenses
•
Flexibility in choosing quality
manpower at reduced cost
•
Greater
accountability
and
transparency
in
production
standards.
(Dhar-Patel & Vishwanath eds.,
2002, p.101)
Quality Standards for Medical
Transcription
The healthcare industry in US needs
to adhere to stringent quality standards.
One such standard adopted by software
companies focused on healthcare domain
is HL7 (Health Level Seven) standards.
The HL7 standards (Shet 2005,
p.326-328) addresses the following
key processes as outlined in capability
Maturity Model or more simply CMM:
•
Requirements management
•
Software quality assurance
•
Organisation process focus
•
Training programmes
•
Software quality management
•
Technology change management
•
Defect prevention
It is one of the several ANSI
(American National Standard Institute)
-accredited
Standards
Developing
Organisation (SDO) operating in the
Healthcare domain. Members of the HL7
are collectively known as working group,
which is mainly organized into technical
committees (TC) and Special Interest
Groups (SIG).
The role of the Technical Committee
(TC) is to:
•
Identify the scope and range of data
elements
•
Work with other SIG, TCs, or any
other related organisation to identify
appropriate controlled vocabulary for
encoding those data elements
•
Identify or define messages (or
objects) required to support the
specific information exchange needs
of applications, both as input to the
application as well as output from the
application.
The Special Interest Groups (SIGs)
provides/certifies the standard for
electronic data interchange pertaining
to healthcare. It keeps on enhancing
the current standards and works in
conjunctions with technical committees
to advance the quality standards and
improve HL7 standards.
The goals of the HL7 standards are to
create flexible, cost effective approaches,
standards, guidelines, methodologies
and related services that help in the
interoperability of healthcare information
system with other dependent system
within the domain.
Another important and specific
standard that all healthcare organisations
Fig. 1: The entire medical transcription process
Fig. 2: The entire medical transcription process
[source: http://www. outsource2india.com]
[source: http:// www.slideshare.net/rohitpate l203/medical-transcription-industry]
CSI Communications | November 2014 | 35
are required to adhere to is the HIPAA
compliance. Health Insurance Portability
and Accountability Act of 1996 -or more
simple referred to as HIPAA complianceis the federal law amended to the
Internal Revenue Code of 1996. The
healthcare insurance industry to improve
portability and continuity for the groups
as well individuals rely heavily on the
HIPAA compliance standards. The main
objectives of HIPAA are:
•
To increase the efficiency and
effectiveness of health information
systems through improvements in
electronic health care transactions
recording and maintenance process
•
To maintain security and privacy
of individual identifiable health
information.
The processes and information
impacted by HIPAA are (but not limited
to):
•
Health claims and equivalent
encounter information
•
Enrolment in and disenrollment from
a health plan
•
Eligibility for health plan
•
Healthcare payment and remittance
advice
•
Health plan premium payments
•
Referral
certification
and
authorization
•
Coordination of benefits
•
Prescriptions
(www.call-centers-india.com/ites.
html)
Stages in Medical Transcription Process
The parameters that define a successful
medical transcription unit are efficient
receiving of voice files, allotting,
transcribing and sending of transcribed
files in cost-efficient manner and without
any error. Accuracy and quick delivery
is paramount in the field of medical
transcription. Figure 1 and 2 describes
the Standardized Medical Transcription
Process as stipulated by HIPAA.
Traditionally, Indian MT sector
followed:
• Dictation
The first step in MT begins with
physician dictating the observation
and diagnosis during examination
of the patient. The dictation is done
onto a device such as tape recorder,
Dictaphone, digital Dictaphone, a PC
or a normal telephone after dialling a
toll free telephone number provided by
the outsourced MT service provider.
First of such dictating machine was
engineered in the late 1880s and since
then it has progressed to several phases
of development from grooves cut on
a cylinder, to flexible belts of plastic,
magnetic media, digital dictation and
speech recognition.
Dictating through toll-free number
is currently the most used technique by
the physician. Since it is possible that a
physician may meet the same patient
at more than one hospital, the doctors
are allotted phone numbers where the
last digit indicates the hospital at which
the patient has been/is being examined.
The innovation like this helps the MT
service provider to locate the hospital
and accordingly decide upon the format
for recording, as different hospitals may
follow different formats.
To make the task of dictation easier
and efficient, voice recognition devices
that can record the message directly into
the PC are now being employed by the MT
service providers. Also, Personal Digital
Assistants or more popularly known
as PDA’s are now made available with
embedded voice recognition capacity or
dictation capture system.
• Storage and retrieval of audio files
The audio files are directed towards the
MT service provider’s voice capturing
server using the intelligent network.
The MT Company’s server host
application capable of digitizing the
voice on real time basis pushes these
voice data (files) to the production unit
through Internet.
• Encryption and transmission
The digitized data is then compressed
and properly encrypted for onward
transmission through satellite link
to MT service provider’s destination
(such as India) where the transcription
gets carried out. The files are then
allotted by the local unit of MT
Company or consultant to various
medical transcriptionist or franchisee.
• Allotment and actual transmission
Depending upon the familiarity with the
physician or based on specific stream
of medicine, the files are classified and
allotted to various transcriptionists
matching the required skill set and
needed turn-around time. The files
are then transcribed in the formats
as prescribed by the physicians or
hospitals. The completed files are
uploaded into the quality controllers
CSI Communications | November 2014 | 36
work area where these are checked for
quality before uploading it back. The
various types of error that may occur
while transcribing are:
• Medical terminology misunderstood (wrong disease attributed to
the patient)
• Omitted dictation (omission of
the laboratory finding as the value
dictated could not be heard or was
misunderstood)
• Medical terminology wrongly spelt
• English word misinterpreted and
hence wrongly spelt (entered
‘elicit’ instead of ‘illicit’ or ‘dissent’
instead of ‘descent’
• Grammatical errors
• Punctuation errors
• Inappropriate blanks
• Typographical errors (typos)
• Formatting errors
• Returning Completed Transcription
The quality checked transcripts
are generally uploaded to the MT’s
company server and are then sent
back to the individual doctors or the
hospitals where they are filled for
storage. Ideally, the medical records
are returned to the hospital servers
in real-time, ready to be accessed.
The encrypted transcribed files are
commonly known as EMR (Electronic
Medical Records).Transcribing/editing
all the hospital reports dictated by
hospital medical staff (physicians,
nurses and physician assistants)
is the responsibility of the Medical
Information
Transcriptionist/editor
also at times referred to as Medical
Language Specialists. The other tasks
include but not restricted to:
• Utilisation of sourcing hospital
information system in order to
place the transcribed report into
the patients account
• Based on sense of the report
and the patient types, selection
of apt report format and print
destinations
• Complete the required number
of transcription tasks within the
stipulated time frame without any
error as per the guidelines set by
the sourcing hospital
(Dhar-Patel & Vishwanath eds., 2002,
p.105-107)
www.csi-india.org
ITeS/MT Industry in India: Over the Years
“India controls 44 percent of the global
offshore outsourcing market for software
and back-office services, with revenues
of US $17.2 billion (euro14.07 billion) in
the year ended March 2005…” (Source:
Associated Press, June 2005. http://
www.medicaltranscription-olutions.com/
medical_transcription_statistics.html).
Over the years, the slowdown across
the globe has impacted the revenues of
this sector and this gets reflected in a
study by NASSCOM indicating that the
growth reduced from 15 percent in 201112 to an estimated 8.4 percent in 2012-13.
The estimated growth for 2013-14 are 1315 percent for total IT-BPM revenue (of
which MT is still a component perhaps to
reckon with), 12-14 percent for exports and
13-15 percent for domestic circuit. IT and
Business Process Management (however)
sector revenues have grown from
1.2 percent in 1997-98 to an estimated
8 percent in 2012-13. (http://indiabudget.
nic.in. Economic Survey 2012-13, p224)
When it comes to MT, starting from
1990, MT ranks amongst the first set of
ITeS-BPO activities to be sourced from
India. Currently MT segment in India
is estimated to earn annual revenue
aggregate of US$ 220-240 million.
(http://www.aiita.org /news/blogs/3322medical-transcription-the-way-to-getyour-bright-career.html)
In India, there are about 120-150 mid
to large size vendors offering medical
transcription services and 70% of the
revenues in the industry are generated
by the large players (mostly US medical
transcription service organizations MTSOs) with offshore centers in India.
Cbay is one the major vendors (www.
cbayscribe.com /cbay systems/ index.
htm) with aggressive growth plan having
headcount of over 3,500. Apart from
large-size and mid-size players there
are few small-size vendors offering MT
services with average employee strength
of about 50 competing for remaining 30%
market. (http://www .sourcingnotes.com/
content/ view/333/58/).
Technological Innovations in MT – Does
it mean Dead End for smaller Indian MT
Outsourcing Agencies…
To imagine that the industry will have
reliable speech recognition output without
human intervention was never envisaged,
and hence the elimination of medical
transcriptionist was completely ruled out
until the new disruptive voice recognition
(VR) technology got introduced. The
automated VR technology allows the data
to be fed directly by the doctors into the
system.
Nuance and Royal Philips Electronics
are the two key players in providing
technology that caters to speech-to-text
market for health care professionals.
The use of speech-to-text technology
by US military’s health system grew by
100% over the last year with about 6000
professional adopting Nuance’s software
across all branches of the military. (http://
w w w. s o u rc i n g n o t e s .co m /co n t e n t /
view/333/58/)
Role of Information as a Business
Resource (IaBR) in healthcare industry
and its impact on patients care, reputation
of the hospital, seamless archiving of
patients data, health tourism and ease
of retrieval has been well-established.
Some of the important terms (http://
w w w. c h i r o c a r e . c o m /w p - c o n t e n t /
Fig. 3: Electronic Medical Record [Source: Nuesoft XpressTM Electronic Medical Record accessed at
http://mtherald.com/how-emr-ehr-is-going-to-affect-medical-transcription-industry]
uploads/2013/03/HITTerms.pdf) related
to health information systems are:
•
EMR (Electronic Medical Record):
Capturing
patient
data
and
facilitating access of these details
to authorized clinical staff from any
specified location characterises a
computer-based medical recording
system popularly referred to as EMR.
Some of important deliverables
associated with EMR are: accurate
and complete claim processing as
required by insurance companies,
providing automated alerts for
drugs’ allergic reactions (and
interactions), generating clinical
notes, prescriptions and schedules,
allowing labs to communicate with
clinical staff etc. The advanced EMR
systems now include units which
keep track of all relevant (from
internal as well as external sources)
medical information and are now
often referred to PMS (practice
management system)
•
EHR (Electronic Health Record):
EHR comprises of processes such
as recording, storing, classifying,
retrieving, querying, generating
reports with respect to patients
physical
and
mental
health.
Past, progressive and predictive
information is then used for providing
primary-health care and healthrelated services. EHR also comprises
of decision-based tools that helps
clinical staff to access and take
decisions on the basis of evidences.
Apart, from clinical purpose EHR
also has modules that help collect
and manipulate data with regards
to billing, quality management,
outcome reporting and public-health
disease surveillance and reporting
modules. The primary contents of
a typical Electronic Health Record
would include: patient demographics,
progress notes, SOAP (Subjective,
Objective, Assessment and Plan)
notes, persisting problems, vital
signs, diagnosis details, medical
history, immunizations time-lines,
laboratory data and radiological data.
Also, using Discrete Reportable
Transcription (DRT) facility, clinical
documentation on EHRs can be further
improved. It not only directly populates
transcribed information into the EHR’s
fields and templates, but also facilitates
structured documentation (right format)
of information in EHR.
CSI Communications | November 2014 | 37
Fig. 4: Sample EHR [Source: http://www.assistmed.com/ blog/ bid/79204/Save-Physicians-90-Minutes-Per-Day-with-DRT-Transcription]
Fig. 5: Sample PHR [Source: http://mtherald.com/how-emr-ehr-is-going-to-affect-medicaltranscription-industry/]
•
PHR (Personal Health Record):
Electronically
maintaining
and
managing
individuals’
health
information requires a very secured
and
confidential
environment.
PHR provides an application that
provides access of the details only
to the authorized person. It controls
an keeps track of accesses made
(accountability), authorization (provide
access to right people in right format at
the right time), authentication (verify
the credentials before allowing the
access) and privacy.
Factors that Obstruct Outsourcing MT
Jobs to India
The healthcare industry in US brought in a
major change in the 1990’s by recognising
and adopting standards that would clearly
define medical terminologies which
was later referred to as standardized
transcription style.
In order to lay
emphasis on standardized documentation
and accuracy the American Association of
Medical Transcription (AAMT) published
Book of Style for Medical Transcription in
1995. The increased use of standards
made hospitals and clinics outsource
their work to medical transcription service
organizations (MTSOs) or home-based
US medical transcriptionists having their
own centers or at the most leased out
the task and services (sub-contract) to
smaller medical transcription companies
within the US. These MTSOs account
for 40% of the transcription job that are
outsourced in the US while only 5% of
these task are currently offshored to low
cost destinations such India, Philippines
and other Asian countries. (http://
w w w. s o u rc i n g n o t e s .co m /co n t e n t /
view/333/58/). This very clearly indicates
that several hospitals and physician
groups out in the US doesn’t favour the
off-shoring of medical records.
CSI Communications | November 2014 | 38
There have been widespread protests
by unions and employees of a hospital in
the east Midlands town of Leicester against
outsourcing of medical transcription work
to India as a part of plans to cut costs. For
the management, outsourcing of task such
as typing of letters by doctors would cut to
48 hours and save about 500,000 pounds
annually. The proposed changes when
implemented would cause loss of jobs for
the medical secretary’s - fears unions and
employees of the medical hospitals in UK.
(http://articles.economictimes.
indiatimes.com/2012-05-13/
news/31689733_1_outsourcing-movemedical-transcription-typing).
Use of technology has added another
dimension. The speech recognition
software is said to get upto 98%
accuracy in results. For the large player
in India adapting itself to the changes in
technology is seamless and in tandem
with the requirements of the industry. Also
these big players not only can afford to get
in new technology but also complement it
with apt training schedule which in turn can
tackle employee attrition/retention issues.
For the small India-based MT companies
bringing in the new technology is always
not affordable and hence it means lesser
flows of works to them which they could
bag either directly or via sub-contracting
earlier. Since smaller offshore vendors
are dependent on medical transcription
service organizations (MTSOs) for
subcontracted work, the non-compatible
obsolete technology makes it difficult to
change the task-giver perception that the
www.csi-india.org
capabilities of small offshore vendors are
found wanting on technology front. This
apparently means loss of opportunity
in MT offshoring job to smaller MT
companies.
For the small India based MT
industry facing stiff internal competition
and lack of stable and trained manpower;
advancements in speech recognition
and Electronic Health Records (EHR)
technology is adding new challenges
to their very existence. The technology
due to features such as easy to use is
gaining popularity and is expected to
be increasingly adopted by clinics in the
US.
(http://www.sourcingnotes.com/
content/ view/333/58/). This would
simply mean that a clinic on successful
implementation of EMR can stop
outsourcing transcription work and can
become self compliant to fulfil statutory
requirements of state/s altogether.
Availability of low cost technology
with low operational cost has definitely
boosted the confidence of physician to
acquire these applications in-house.
However, not all the EMR applications
follow the same methodology for creating
patient records. While some are strictly
non-customizable
but
standardized
template-driven point and click technique
others are DRT enabled, which allows
physicians to use traditional dictation
technique for feeding the data into the
system. Needless to say if the former
becomes more prevalent, need for
transcription will further decrease. (http://
mtherald.com/how-emr-ehr-is-going-toaffect-medical-transcription-industry/)
Currently one of the major concerns
for the medical practitioners is medical
accuracy of the patients’ data and this
view is getting stronger that Indian MT
organisations are not fully equipped to
handle this. But still India continues to
get the job outsourced to its shore. But,
what if the law makes it compulsory for
the physicians do it on their own to avoid
medical errors?
Revival Strategy that can Make India
Once Again the Preferred Destination
Though technology is a threat for Indian
MT industry but then it is and will not be
possible for all the doctors’ to switchover
instantly, physicians with less volume
of practice may not be able to afford to
automate and implement EMR solution
into their clinics. And these are the
targets for Indian MT industry. However,
once it becomes affordable - the word of
caution is, EMR in not too tough a task to
perform on a PC or a tablet, the easy to
use interface and training (as part of their
academics) is enough for the doctors to
fill in the prescription with few clicks. It’s
time clicking fast for Indian MT industry
to change and think fast to take advantage
of brand that India has created in the
outsourcing business over the years.
According to the Bureau of Labor
Statistics (US):
"Employment
of
medical
transcriptionists is projected to grow
faster than the average; job opportunities
should be good, especially for those who
are certified. Employment of medical
transcriptionists is projected to grow 14
percent from 2006 to 2016, faster than
the average for all occupations." (Source:
http://www.bls.gov/oco/ocos271.htm and
http:// mteducationonline.com)
About 80 medical transcriptionist
(MT) firms in Philippines have formed
the Medical Transcription Industry
Association of the Philippines Inc.
(MTIAPI), in order to provide standardized
training in various functions. It aims
at providing standardised training and
increasing the workforce three times as
that of present. The vision states “The
Organization envisions the Philippines to be
the offshore destination of choice for Medical
Transcription services.”(http:// www.ncc.
gov.ph/default.php).
In India too, efforts are being
made towards launching the training
programmes in medical transcription
especially for the rural youth. The 4-6
months program covering the subjects
like Science, US accent training, English
typing and grammar aims at improving
the quality of MT jobs and deliverables in
India. This would then help India spruce
up MT industry and make it ready to
tackle the threat from the countries such
as Philippines which are churning out
adequate numbers of trained manpower.
(http://www.sourcingnotes.com/content/
view/333/58/)
In
August
2004,
Medical
Transcription firms in the country formed
the Indian Medical Transcription Industry
Association with the aim of boosting the
growth to US$ 100 million (www.imtia.
net). Accessing the website, however,
gives the feeling that the association is
defunct with web links of some of the
founder members leading to load page
error. In the news and event section it
still shows the latest news pertaining
to August 6, 2004. One of the revival
strategies could be to have IMTIA as more
visible association representing India in
the field of MT globally. Consortiums
always help. Another key factor is the
wage. Currently, the wages for the medical
transcriptionist in India is:
National Salary Data
Salary
Rs 72,650 - Rs 348,068
per year ($1207.51 $5785.22 per year)
Bonus
Rs 0.00 – Rs 39,732 per
year
Certified
Professional
Coder
Employees
Rs 180,000 – Rs 3,60,000
per year ($2991.71
-$5983.55 per year)
(Country: India | Currency: INR | Updated:
9 Jun 2014 | Individuals Reporting: 336;
http://www.payscale.com/research/IN/
Job=Medical_Transcriptionist/Salary)
The above figures don’t inspire the
current generation youth to join the MT
industry as they perceive this to be the
dead end to their careers. Not only have
wages remained stagnant since the new
millennium, but the introduction of voice
recognition has actually reduced wages
by nearly one-half. Moreover even after
acquiring the certified profession coder
credential the salary ranges to about Rs
15,000 – Rs 30,000 per month which does
not meet the GenNext aspiration. With
perception that MT industry in India is not
doing enough to stand the competition,
the sector is moving towards the dead
end opines several of the MT professional
interacting on social network? Wage
correction and defining the standards can
help regain the past glory of MT industry
in India. Compare this to the wages in US:
US Data
$34,020 per year
$16.36 per hour
Entry-Level
Postsecondary
Education
non-degree award
Work experience in None
a related occupation
On-the-Job-Training None
Number of Jobs
84,100
2012
Job Outlook, 2012-22 8% (As fast as
average)
Employment
6,400
Change 2012-22
Quick Facts: Medical Transcription in
US (Source: http://www.bls.gov/ooh/
healthcare/ medical- transcriptionists.
htm)
2012 Median Pay
CSI Communications | November 2014 | 39
Inaccuracies in medical records
can put patients’ lives to risk. Once the
records are generated by transcriptionist,
the outsourced agency should make their
process of checking more rigorous and
ensure that records are accurate. This
would help rebuild the confidence of the
hospitals and doctors outsourcing their
task to India as they don’t have to recheck
the records before replacing the patient
information into the shelf.
Hacking, breach of security and
leakage of sensitive patient data are cited
as other reasons by hospitals and clinics
for not sourcing the MT tasks to India. It is
important for MT industry in India to build
the robust and protective security system
that will prevent the hacking as technology
advances, merely stating that hacking
happens day in and day out and even
military or banking data are not immune
to these kind of attempt will not suffice
and further alienate the MT outsourcing
task from India. (http://mtherald.com/
how-emr-ehr-is-going-to-affect-medicaltranscription-industry/)
or palm tops, connected to internet at ever
increasing speed, the internet of things
(IOT) is increasingly becoming a reality. The
practising doctor in U.S. is not an exception
to the bevy of users getting smarter in use
of these devices and technologies. In such a
condition the absence of even a functional
industry association may seem like the last
nail on the coffin- specially when a country
like Philippines is trying to catch up to give
a tough competition to established player
like India.
Will the Indian MT industry sit up and
notice? Will it use the same technology
disruption to its advantage? Quality is the
mantra. Can we take it to six sigma level
which can instil confidence in the hearts of
the users? Can we make it secure enough to
pass laws like HIPPA and standards like HL7.
Can we make it more productive so that we
can attract better talent who are in search of
better salaries? That is the need of the hour.
It remains to be seen how the Indian MT
industry copes with this challenge brought
about by new wave of technology.
References
Conclusion
For whom the bell tolls? The answer may
seem obvious to many given the rate of
change of technology. The same technology
which brought the MT work from the
faraway shores of U.S.A many years back
today seems to be a hindrance to the health
of the Medical Transcription job in general.
Technologies like speech to text and
handwriting recognition are terms which
we are becoming aware more and more in
everyday life. With the proliferation of hand
held devices, be it smart phones or tablets
[1]
[2]
[3]
[4]
[5]
[6]
[7]
Associated Press, June 2005. http://
www.medicaltranscription-solutions.
com/medical_
transcription_statistics.
html
Brady, JA Monk, EF & Wagner, BJ, 2001.
Concepts in Enterprise Resource Planning.
Bangalore: Thomson Course Technology.
Dhar-Patel, M & Vishwanath, CV eds.,
2002. The Economic Times: IT enabled
services – An ETIG report. Mumbai: The ET
knowledge series.
en.wikipedia.org
http:// mteducationonline.com
http:// www.ncc.gov.ph/default.php
http://articles.economictimes.indiatimes.
com/2012-05-13/news/31689733_1_
outsourcing-move-medical-transcriptiontyping
[8] http://indiabudget.nic.in.
Economic
Survey 2012-13
[9] http://mtherald.com/how-emr-ehr-isgoing-to-affect-medical-transcriptionindustry/
[10] http://www .sourcingnotes.com/content/
view/333/58/
[11] http://www. outsource2india.com
[12] http://www.aiita.org /news/blogs/3322medical-transcription-the-way-to-getyour-bright-career.html
[13] http://www.assistmed.com/
blog/
bid/79204/Save-Physicians-90Minutes-Per-Day-with-DRT-Transcription
[14] http://www.bls.gov/oco/ocos271.htm
[15] http://www.bls.gov/ooh/healthcare/
medical- transcriptionists.htm)
[16] http://www.chirocare.com/wp-content/
uploads/2013/03/HITTerms.pdf
[17] http://www.payscale.com/research/IN/
Job=Medical_Transcriptionist/Salary)
[18] http://www.slideshare.net/rohitpate
l203/medical-transcription-industry
[19] http://www.sourcingnotes.com/content/
view/333/58
[20] Nair, K (Dr.)., 2008. Medical Transcription
and India: The current scenario and the future.
Available at http://chillibreeze.com/
articles/MedicaltranscriptionandIndia.
asp [accessed 16 December 2008].
[21] Nuesoft Xpress TM Electronic Medical
Record accessed at http://mtherald.com/
how-emr-ehr-is going-to-effect-medicaltranscription-industry/
[22] www.call-centers-india.com/ites.html
[23] www.cbayscribe.com/cbaysystems/
index.htm
[24] www.imtia.net
[25] Yojana Aayog;-Planning Commission.
2000. India Vision 2020 [pl_vsn2020.
pdf], Government of India. Available at
http://www.planningcommission.gov.
in/plnre/pl_vsn2020.pdf; pg21 of 108.
n
[accessed on 5 April 2008]
Some of the facts are taken from blogs and other social networks, where the Medical transcriptionist and others associated with industry have expressed their opinions. These are
represented here for the academic purpose only. Also, websites are liberally referred - to get the overview of the industry and the prevailing conditions, the inferences are based purely
on the availability of these secondary data. The trademarks used here belong to the respective organisations and have been used here for the academic purpose as well.
CSI Communications | November 2014 | 40
www.csi-india.org
Kind Attention: Prospective Contributors of CSI Communications
Please note that Cover Themes for forthcoming issues are planned as follows:
•
December 2014 – Algorithmic Computing
•
January 2015 – IT Infrastructure
•
February 2015 – Quantum Computing
•
March 2015 – Machine Translation
Articles may be submitted in the categories such as: Cover Story, Research Front, Technical Trends and Article. Please send your contributions
before 20th of a month prior to the issue month for which you are contributing. The articles may be long (2500-3000 words maximum)
or short (1000-1500 words) and authored in as original text. Plagiarism is strictly prohibited.
Please note that CSI Communications is a magazine for membership at large and not a research journal for publishing full-fledged
research papers. Therefore, we expect articles written at the level of general audience of varied member categories. Equations and
mathematical expressions within articles are not recommended and, if absolutely necessary, should be minimum. Include a brief
biography of four to six lines for each author with high resolution author picture.
Please send your articles in MS-Word and/or PDF format to the CSI Communications Editorial Board via email id csic@csi-india.org.
(Issued on behalf of Editorial Board of CSI Communications)
CSI Communications | November 2014 | 41
Innovations
in India
Rajiv Thanawala* and Prachi Sakhardande**
Product Experience CoE, Component Engg. Group, TCS
Software User Experience Maturity Model
To improve software code quality, we use
quality metrics that are measurable. Defects
per Line of Code, Cyclometric Complexity,
and so on, are examples of measurable
metrics that enable this. User Experience
(UX) is a critical quality aspect for software.
How can we then measure UX quality?
We assessed the existing UX
assessment methods. They revealed
that though the current UX assessments
helped identify usability flaws, they
lacked quantifiable metrics for an
objective review of UX quality, and also
lacked increasing levels of UX maturity
using which one can strive for continuous
improvement of UX quality.
To objectively and quantitatively
assess UX quality, we then formulated
User Experience Maturity Model
(UXMM) – a system and framework built
based on established UX assessment
techniques. UXMM is applicable to
software products and applications,
irrespective of its domain or device.
Figure 1 shows diagrammatic representation of User Experience Maturity Model.
Main features of UXMM are as follows:
•
4 hierarchical levels of UX maturity—
Usable, Useful, Desirable and
Delightful—indicating progression in
user experience
•
14 Key UX Parameters (KUXP)] –
Ease of Use, Speed of Use and others
as shown in the figure
•
Each KUXP comprising of attributes.
E.g. “Ease of Use” KUXP has
attributes like Ease of access, Ease
of Data Input, Visibility of System
Status etc. (not shown in the figure)
•
Levels 1 and 2 based on first 10
KUXPs—from ‘Ease of Use’ to ‘Help’
•
Levels 3 and 4 based on additional
4 KUXPs—from ‘Brand Recall’ to
‘Greater Good’
‘Expert Review’ used as assessment
method for certifying software at Level 1
•
‘Usability Tests’, ‘Competitor Analysis’
and ‘Emotional Heuristics’ used for
certifying software at Levels 2, 3 and 4
•
Quantitative scores as result of
assessments
The approach for UXMM assessment is as
follows:
1. Capture context of use of a software
application or product through a preassessment questionnaire. Based
on the context, UXMM system
generates weightages for each KUXP.
2. Steps to Certify for each Level:
2.1 UXMM system generates a level
specific benchmark score.
2.2 A usability expert rates the
application across KUXPs with
a scorecard based on a specific
assessment technique (‘E.g.
Expert Review’ for Level 1).
2.3 If this score is lesser than
benchmark score, the expert
identifies improvement areas.
Else, the software is considered
certified for that specific Level.
2.4 Once a level is achieved, repeat
steps 2.1 to 2.3 for the next level,
all the way up to Level 4: Delightful
Over the last two and half years, more
than 100 UXMM assessments have been
used to certify UX quality of our software
products and applications. Use of a
standardised assessment mechanism has
helped improve our products’ UX quality
and institutionalised UX as a practice.
•
Reference
[1]
About the Authors
Fig. 1: User Experience Maturity Model
TCS Published Patent Application – US
20130254735 A1 – ‘User Experience
Maturity Level Assessment’ – Prachi
Sakhardande, Rajiv Thanawala
n
(© 2014, Tata Consultancy Services Limited, Printed
with permission)
Rajiv has 25 yrs of experience in IT consulting and product development. He leads Product Experience Center of
Excellence which provides User Experience, Information Visualization, User Assistance and Customer Experience
services to a products group at TCS.
Prachi Sakhardande has consulted with Fortune companies and leads User Experience and Customer Experience
within Product Experience Center of Excellence for a products group at TCS. Her team is responsible for UX design
and benchmarking.
Innovators interested in publishing about their work can send a brief write up in 150 words to Dr Anirban Basu, Chairman, CSI Div V, at div5@csi-india.org.
CSI Communications | November 2014 | 42
www.csi-india.org
Security Corner
Jignesh Doshi* and Bhushan Trivedi**
*Associate Professor, L J Institute of Management studies
**Ph.D.
Information Security »
A Quick Look at Virtual Private Database Security
Introduction
Internet users and usage have increased
drastically in past few years. Internet
is widely used for business. Millions of
transactions are taking place via web
applications. Cyber attacks are becoming
more sophisticated day by day with
evolution of technologies and increasing in
numbers. Security is a big challenge for IT.
As per Open World Application Security
Projects (2013), 5 out of 10 top attacks are
related to databases.
A new attack, “Sensitive Data
Exposure” is added into top 10 list.
After getting access, attacker can dump
sensitive data like credit card number,
social security number, financial data
etc. As a result it may impact data loss,
negative publicity and a loss of customer
confidence.
Oracle Virtual Private Database
(VPD) is one of the key technologies
powering sensitive data theft prevention.
In this article, some of the ways in which
sensitive data exposure can be prevented
using Oracle Virtual private Database is
discussed.
Limitations of Traditional Security Policies
There are two types of database
security available in oracle databases
are Discretionary access control using
grant and revoke (DAC) and Mandatory
Access Control (MAC).
Two level of access privileges can
be granted using DAC security policy, at
account level ( like create table, drop table
etc.) and table level (like insert, update,
select or delete). DAC is very flexible
policy and DBA must provide selective
access using Grant or Revoke statements.
In many applications, additional security
policy is needed for classifying data and
users based on security classes. This
approach is known as Mandatory Access
Control.
Most commercial DBMS provide
mechanisms only for Discretionary
Access Control.
Security Flaws in DAC and MAC is
summarized as below:
What is VPD?
Oracle Virtual Private Database (VPD)
enables you to create security policies to
control database access at the row and
column level. VPD can be enforced directly
on database tables, views, or synonyms.
We can apply VPD policies on SELECT,
INSERT, UPDATE, INDEX, and DELETE
statements. Oracle database engine
automatically applies security functions
whenever a user access data and there is no
way to bypass security i.e. adds a dynamic
WHERE clause to a SQL statement.
Security Policy Drawbacks
1) In this security
DAC
policy. users either
have or do not have
a certain privilege.
2) No method to Limit
propagation
e.g.
one can not restrict
GRANT
OPTION
privilege to at most
‘n’ other accounts.
3) Not suitable for high
security data.
MAC
1) Not flexible as a
result it can not
work with all types
of applications.
How Oracle Virtual Private Database
Works?
VPD execution flow is narrated in Fig. 1.
Whenever any statement is executed,
DBMS engine checks whether security
policy is configured on the objects found
in query or not. If security policy is
configured on the objects of query, it will
modify query and execute the query.
With DAC security policy, we can not
restrict data selection at horizontally
(row) or vertically (column) level. i.e. if
we want to restrict that AR user can see
only his department data or he can not see
salary of employees.
How to Configure?
Pre-requisites: Fine grained access
control must be enabled in database.
Check using query “select * from v$option
where parameter = 'Fine-grained access
control”, it should be TRUE.
ROW LEVEL VPD Security Policy
One can create multiple security policies
on different objects using same function.
We can apply row level security directly
on database tables, views, or synonyms.
Row Level VPD configuration is a two
step process.
i) In first step, we will create the stored
function.
ii) Second step comprise of creation
of security policy using DBMS_RLS
CREATE OR REPLACE FUNCTION
SEC_AUTH_ORDERS
(SCHEMA_VAR IN VARCHAR2,
TABLE_VAR IN VARCHAR2 )
RETURN VARCHAR2 IS RETURN_
VAL VARCHAR2 (400);
BEGIN
RETURN_VAL := 'SALESMAN_ID
= 101';
RETURN RETURN_VAL;
END AUTH_ORDERS;
Fig. 1: Oracle VPD processing flow
Table 1: Security Function
CSI Communications | November 2014 | 43
package for stored function created
in step 1.
Example: If you have configured policy
such that JOHN can view only his sales
orders and employee id of John is 101.
Step 1: Create Security Function
Create the stored function (refer to
Table 1), which will append the WHERE
SALESMAN_ID = 101 clause to any SELECT
statement on the OE.ORDERS table.
Step 2: Create the Oracle Virtual Private
Database Policy
BEGIN
DBMS_RLS.ADD_POLICY (
OBJECT_SCHEMA => ‘OE’,
OBJECT_NAME => ‘ORDERS’,
POLICY_NAME => ‘SEC_POLICY_
ORDERS’,
FUNCTION_SCHEMA => ‘SYS’,
POLICY_FUNCTION => ‘SEC_
AUTH_ORDERS’,
STATEMENT_TYPES => ‘SELECT,
INSERT, UPDATE, DELETE’ );
END;
Table 2: Security Function
using DBMS_RLS package.
Two ways to implement column level
security policies are as below:
•
Displaying Only the Column Rows
Relevant to the Query
•
Using Column Masking to Display
Sensitive Columns as NULL Values
We can apply column-level security
policies to tables and views, but not
to synonyms. Let us understand using
example.
Example: Security policy in which
Chicago sales users cannot see the salaries
a) Security Function
CREATE OR REPLACE FUNCTION
SEC_HIDE_SALES_DATA
(SCHEMA_VAR IN VARCHAR2,
TABLE_VAR IN VARCHAR2 )
RETURN VARCHAR2 IS RETURN_
VAL VARCHAR2 (400);
BEGIN
RETURN_VAL := 'LOCID=15';
RETURN RETURN_VAL;
END AUTH_ORDERS;
b) Configure security Policy
BEGIN
DBMS_RLS.ADD_POLICY (
OBJECT_SCHEMA => 'OE',
OBJECT_NAME => 'EMPLOYEE',
POLICY_NAME => 'SEC_POLICY_
HIDE_SAL',
POLICY_FUNCTION => 'SEC_
HIDE_SALES_DATA',
SEC_RELEVANT_COLS =>
'SAL,COMM');
END;
Next, create the policy (refer to Table 2)
using the ADD_POLICY procedure in the
DBMS_RLS package.
How security policy will work?
Query Executed
by JOHN
Converted Query
by VPD
SELECT* FROM
OE.ORDERS;
SELECT* FROM
OE.ORDERS
WHERE SALESMAN
_ID = 101;
About the Authors
COLUMN LEVEL VPD Security Policy
Column Level VPD configuration is a two
step process. In first step, we will create
the stored function. Using it we will define
row level access control Second step
comprise of creation of security policy
and commission details of people outside
the Chicago. (Location id is 15 for Chicago).
Above security policy will display salary
and commission details of Chicago sales
representatives only. However, super user
can view all details.
Conclusion
Key Benefits of Oracle Virtual Private
Database security policies are as below:
•
High Level of Security:
˚ As plemented at database level, it
provides consistent access control
on tables, views or synonyms, no
matter from where user access
data ( Front end, back end or third
party tools)
we
can mitigate Sensitive data
˚
exposure risk using VPD security
policies.
•
Simplicity;
˚ Adding security policy is very easy
and simple.
•
Flexibility.
˚ One can create different policies
for different operations (select,
insert, update or delete) on same
table.
Security policy functions run as it had
been declared with definer’s rights so
do not declare it with invoker’s rights.
We can extend user Role based access
control (RBAC) to fine grained row or
column level.
References
[1] OWASP:
https://www.owasp.org/
index.php/Top_10_2013-A1-Injection:
accessed 31st May 2014
[2] Internet hosting statistics : http://www.
netcraft.com/internet-data-mining:
accessed 31st May 2014
[3] Common Weakness Enumeration:
h t t p : //c w e . m i t r e . o r g /d a t a /
definitions/89.html : accessed 3rd June
2013
[4] Internet
users
:
http://www.
internetlivestats.com/internet-users:
accessed 14th June 2014
[5] Oracle documentation ; docs.oracle.
com
[6] DBMS_RLS :docs.oracle.com: visited on
25th july 2014.
n
Jignesh Doshi received the B.Sc (Maths) and M.C.A. degrees in 1989 and 1992, respectively. He is an Oracle Certified
Professional ( OCP 10g). He has nearly 22 years of experience. During 1992-2008, he worked in various IT firms like
Patni Computer Systems, Vodafone (Fascel), Gujarat Lease Financing Ltd., Erhardt + Leimer (I) Ltd.. Since 2008, he is
working as associate professor at L J Institute of Management studies. His areas of interest are Database, Database
security and Data warehousing and data mining.
Bhushan Trivedi received Ph.D. in 2008. He has nearly 25 years of teaching experience. His areas of interest are
Intrusion Detection with mobile agents, Sensor Networks, Using artificial intelligence techniques to solve real world
Problems. He had conducted workshops on Effective Teaching (20), "how to debug a network with TCPDUMP and
WireShark" (15) and “how to write an effective research paper” across India. He has written books are on ANSI
C++ and Computer Networks under Oxford publications. He is guiding 8 students in their doctorate degree.
CSI Communications | November 2014 | 44
www.csi-india.org
Security Corner
Dr. Vishnu Kanhere
Convener SIG – Humane Computing of CSI (Former Chairman of CSI Mumbai Chapter)
Case Studies in IT Governance, IT Risk and Information Security »
Visualization – to make visible: Visualization covers a wide range from mental and creative visualization to geo visualization, flow
visualization, computer graphics, data visualization and even network visualization. Visualization has been used by mankind since
its early days and is not a new concept. The early prehistoric cave paintings and the later Egyptian hieroglyphs are examples of the
use of visual symbols and pictures by humans to communicate, convey and interact meaningfully. Visualization has found growing
acceptance and extensive use in science, education, engineering, communication, multimedia and medicine to name a few. In the
field of computing the growth and emphasis on use of visualization was limited due to the lack of graphics power. However, since late
1990’s the field has seen an accelerated growth which has been further boosted with the use of computer animation. Moving forward
from the more familiar digital animations used for weather pictures and satellite photos and the complex scientific 2D & 3D imaging,
- visualization is now being used for educational animations, and dynamic representation of timelines.
Computer Visualization has two main aspects. The first is a major benefit in terms of providing a powerful tool for better and
smarter governance. Governance whether of nations, corporate, organizations, entities or individuals requires a robust measurement
and reporting system and efficient MIS for better control, monitoring, corrective action and delivery. Computer visualization has enabled
improved meaningful MIS which has provided a strong boost at a micro as well as macro level, for both IT and corporate governance.
In today’s information age an increasing amount of data is being created by the internet and social media and the various sensors
in and around us – be it the CCTV’s in offices, public places, housing societies or at traffic signals, which in turn does at times get
related or is relatable to data bases holding personally identifiable information like PAN data base, Aadhaar data base etc. This data is
often referred to as Big Data. Computer visualization includes Data visualization which enables and improves the processing, analyzing
and communicating of this Big Data. This is the second aspect. In this context visualization of Big Data presents a variety of ethical
and analytical challenges for visualization.
The potential threats to the data and from the use of such data are many – these primarily arise from the SMAC threats which
are ever present in mobile, social media, cloud environment and in analytics – since visualization does require and makes extensive
use of analytics.
Mobile devices are not always secured, are often mishandled, lost or stolen and used for almost all activities, both personal and
professional exposing data.
Social media has its own vulnerabilities with users, their friends and relations indiscriminately and accidentally posting and
exchanging sensitive information.
Analytics aggregates and analyzes such information making it even more critical, hence enlarging the threat.
Finally, the Cloud has removed the barriers between the entity dealing with the information and the outside world. Even if you are
not on the cloud your service providers, associates, users, are bound to be there and that is enough to increase the risk further.
In such situation the need has arisen for a specialized field in Governance – the field of data governance to improve data security
and to ensure accountability and ethics in use of data within a robust framework that promotes public confidence.
Given this background the current Case in Information Systems is being presented. The facts of the case are based on information
available in media reports, online information and some real life incidents. Although every case may cover multiple aspects it will have
a predominant focus on some aspect which it aims to highlight.
A case study cannot and does not have one right answer. In fact answer given with enough understanding and application of mind
can seldom be wrong. The case gives a situation, often a problem and seeks responses from the reader. The approach is to study the
case, develop the situation, fill in the facts and suggest a solution. Depending on the approach and perspective the solutions will differ
but they all lead to a likely feasible solution. Ideally a case study solution is left to the imagination of the reader, as the possibilities
are immense. Readers’ inputs and solutions on the case are invited and may be shared. A possible solution from the author’s personal
viewpoint is also presented.
A Case Study of Netrapur Police
Satyarthi the SP of Netrapur was newly
posted to this state. He is tech savvy and
wishes to go in for major initiatives to make
his policing better and more effective.
He has unveiled his policing plans and
initiatives. He has planned to introduce
state of the art surveillance system “Sannirikshan” initially in the urban and the
remote far flung areas followed up by
coverage of the semi urban and rural
parts of Netrapur. He is convinced that
adopting state of the art monitoring
technology with face recognition software
for CCTVs and cross referencing with
universal data base of Aadhaar and PAN
will enable quick identification of persons
from footage and help prevent, detect
and resolve crimes effectively. It will also
improve policing with rapid pinpointed
and effective action. Coupled with this
he also plans to introduce cyber-policing
of the internet, social media and mobile
messaging using network traffic analysis.
Combining these two in “san-nirikshan”
will enable compilation of Big Data and
computer visualization of all the activities
of the population helping in curbing of
criminal, anti social as well as socially and
morally reprehensible practices.
He calls a meeting of his top officials
and shares his viewpoint. He emphasizes
that adopting “san-nirikshan” and using
visualization will enable Netrapur police to •
Go beyond geo-political, cultural,
social and physical barriers
•
Deploy cost effective policing
CSI Communications | November 2014 | 45
solutions enabling comprehensive 24
by 7 coverage
•
Enhance operations and logistics
enabling rapid and effective response
•
Enable better data gathering and
analysis with visualization for crime
prevention and promoting moral and
ethical society
•
Promote inter-state and inter-agency
collaboration and exchange of data
•
Use real time data analytics and
advanced visualization techniques
thereby improving the policing
process
enabling
deterrence,
prediction and prevention of crimes
that are about to happen
•
Enable rapid scaling up and right
sizing of the Police force and support
internal and external security
His CP Varun Pal and Joint CP
Crime Vishal are apprehensive of the far
reaching changes and possible impact
especially given the fact that such a major
initiative is being adopted without taking
the people into confidence and virtually
being thrust on the public. Samar the DSP
is extremely bullish and is certain that this
would provide a definite edge to the police
force over criminals, terrorist networks,
anti-social elements and trouble makers
and keep the pseudo intellectuals who
generally foment trouble over human
rights as well as the media in check.
Anticipating some resistance to the
project – “san-nirikshan” from intellectuals
and vested interests Satyarthi emphasizes
that there are both privacy / human rights
concerns as well as deployment issues.
He selects a crack team consisting of
Samar, Varun and Vishal to overcome the
social resistance and nip it in the bud to
enable smooth deployment. They are
also tasked to put in place systems to
address the various issues and concerns
about deploying “san-nirikshan”. The
political set up is fully supportive of the
project and has already given a go ahead
using administrative powers. The project
involves sensitive internal security issues
and hence is outside the purview of
RTI, and is to be kept confidential with
complete media blackout.
The crack team develops a plan and
submits it. Varun is very uneasy with the
manner in which the whole project is
steamrolled. He seeks an appointment
with Satyarthi and voices his reservations
about privacy and human rights and
freedom of the press. He suggests that
the project be introduced in remote
areas first as a pilot and then extended
to metro cities in a transparent manner. If
implemented initially on a small scale and
then deployed in phases taking the people
into confidence it will be both acceptable
and successful.
Satyarthi rejects Varun’s approach,
points out the increasing criminal activities
and terrorist attacks and stresses the
need for a full deployment with utmost
secrecy. He advises Varun to go ahead and
implement the plan as it is.
After much persuasion Varun relents
and the project is implemented. The
project achieves quick and remarkable
success. A number of criminal gangs are
neutralized, terrorist activities are visibly
reduced. The backbone of the Naxalite
resistance in remote areas is broken and as
a result the law and order situation in the
state of Netrapur improves considerably.
The police are lauded for their good work
and Satyarthi is promoted. Varun is still
ill at ease and opts for a posting in the
police housing project where he remains
till retirement.
What are your views on project “san-nirikshan”?
Solution
The situation:
Visualization techniques using computer
graphics and imagery, analytics and
presentation using visualization techniques
relating to Big Data especially from Social,
Mobile, Analytics and Cloud technologies
have added a new dimension to surveillance.
The eyes and ears of the police – the Big
Brother are now omnipresent. Surveillance
and use of covert means and secret agents
including agencies like FBI, MI6 and KGB
are common in countries subscribing
to all shades of political, economic and
social viewpoints. The recent use of similar
techniques by the US, to tackle terror and
ensure ‘Homeland Security’, has been
accepted without question. It has had a
significant impact across the world and an
even more profound impact on privacy and
human rights.
There exists a case both for and
against the use of technology for such
means and the case both for and against
use of visualization techniques for internal
and external security is equally strong.
Information technology when combined
with technological advances like street
CSI Communications | November 2014 | 46
level view satellites and micro drones
carrying eavesdropping equipment and
remote CCTVs make the visualization and
analytical technology even more powerful.
The way forward therefore cannot
be either to pull the plug on use of Data
visualization for internal policing and / or
external defense or to wait and watch till
the technology is deployed, stabilizes and
is eventually safe and acceptable enough
to adopt.
The challenge:
Disruptive technology as has often been
said needs proper governance. In the same
fashion use of Data visualization has to be
understood, the risks appreciated and the
technology proactively adopted.
There are primarily four challenges
which need to be understood and
considered.
1. Preventing the potential abuse of the
technology and the power it provides
either for selective victimization,
oppression, political misuse, or moral
policing
2. Preventing the leakage, compromise
and misuse of the data resulting
in harm to the community or to
individuals
3. Preventing the loss of privacy and
fundamental human rights
4. Preventing the loss of freedom
of expression and preserving the
opportunity for reasonable dissent
and difference of opinion either
collective or individual.
The consequences:
Continuing in the existing fashion with
the project, given the lack of proper
governance and transparency may despite
initial encouraging results, ultimately lead
to a totalitarian situation where a handful
few control the multitude. It will pose a
clear threat to the democratic polity and
cultural ethos. Ethical values may not be
maintained and freedom of expression
and privacy may be lost.
The strategy:
The right strategy for the state of Netrapur
at this stage would be:
1. The four arms – the parliament, the
executive, the judiciary and the press
and public need to understand the
technology and the pressing need to
deploy it.
2. Understand and agree on the need
of using Big Data Visualization and
www.csi-india.org
3.
4.
About the Author
5.
analytics for internal and external
security and protection.
Understand potential threats of
abuse, leakage, and threats to privacy,
fundamental rights and freedom.
Develop a GRC (Governance, Risk,
and Compliance) framework with
proper safeguards and security built
in to protect integrity and ethical
values and prevent misuse.
Deploy framework and manage
Vulnerabilities and threats.
6.
Ensure appropriate controls to
protect the system from future abuse
or compromise.
It is evident from the reported
and unreported incidents till date that
aberrations are possible and misuse
cannot be entirely prevented. However
given the pressing need for adopting
emerging technologies the way forward is
to proactively adopt Data visualization and
analytics. The police in Netrapur cannot
afford to lag behind their counterparts in
other countries and states, nor fall short
of the criminal and terrorist networks.
The police have to be where the action
is and even where the action is going to
be in a proactive manner. Likewise the
governance and security framework and
its scope have to deal both with the threats
and the key issues identified above.
An effective solution is generally
expected to proceed on these lines.
n
Dr. Vishnu Kanhere Dr. Vishnu Kanhere is an expert in taxation, fraud examination, information systems security and system audit and has done
his PhD in Software Valuation. He is a practicing Chartered Accountant, a qualified Cost Accountant and a Certified Fraud Examiner. He has over
30 years of experience in consulting, assurance and taxation for listed companies, leading players from industry and authorities, multinational
and private organizations. A renowned faculty at several management institutes, government academies and corporate training programs, he has
been a key speaker at national and international conferences and seminars on a wide range of topics and has several books and publications to
his credit. He has also contributed to the National Standards Development on Software Systems as a member of the Sectional Committee LITD17
on Information Security and Biometrics of the Bureau of Indian Standards, GOI. He is former Chairman of CSI, Mumbai Chapter and has been a
member of Balanced Score Card focus group and CGEIT- QAT of ISACA, USA. He is currently Convener of SIG on Humane Computing of CSI and
Topic Leader – Cyber Crime of ISACA(USA). He can be contacted at email id vkanhere@gmail.com
CSI Communications | November 2014 | 47
Brain Teaser
Dr. Debasish Jana
Editor, CSI Communications
Crossword »
Test your Knowledge on Visualization Technologies
Solution to the crossword with name of first all correct solution provider(s) will appear in the next issue. Send your answers to CSI
Communications at email address csic@csi-india.org with subject: Crossword Solution - CSIC November 2014
CLUES
ACROSS
1.
5.
6.
7.
8.
9.
10.
11.
15.
16.
17.
19.
21.
22.
23.
24.
27.
28.
29.
Displays tree data structure in a way that expands outwards (9)
A property or characteristic of data (9)
A common data visualization method used in computational fluid
dynamics (10)
A data structure optimized to quickly answer multi-dimensional
analytical queries (4)
A schematic representation of a sequence of operations in a computer
program (9)
Used in project management to illustrate schedule of a project (5, 5)
Type of diagram used for data visualization (11)
Type of diagram showing the flow of data through an information system (4, 4)
A graphical representation of data (5)
Field lines resulting from this vector field description of the flow (11)
Type of fishbone shaped diagram that shows the causes of a specific event (8)
open source JavaScript charting library (8)
Type of dictionary that illustrates the meaning of words primarily
through pictures (6)
A sequence of related images viewed in rapid succession to show
apparent movement of objects (9)
Online analytical processing (4)
comma separated values (3)
Online transactional processing (4)
Displays a list of events in chronological order (8)
Microsoft Windows-based visual mapping software (10)
DOWN
2.
3.
4.
9.
12.
13.
14.
18.
20.
Did you know view of Hal Varian on how the Web challenges managers?
Hal Varian, Google’s chief economist commented in The
McKinsey Quarterly, Jan 2009:
The ability to take data—to be able to understand it, to process
it, to extract value from it, to visualize it, to communicate it—
that’s going to be a hugely important skill in the next decades,
… because now we really do have essentially free and ubiquitous
data. So the complimentary scarce factor is the ability to understand that data
and extract value from it.
25.
26.
Organic-looking n-dimensional objects used in computer graphics for
visualization (8)
An open-source Java based framework that allows the creation of a wide
variety of charts (10)
An assignment of a vector to each point in a subset of space used in vector
calculus (11)
Visualization technique supporting geospatial data analysis (16)
Describes the property that allows light to partially pass through and
partially reflect (12)
What you see is what you get (7)
A measure of spatial extent, especially width, height, or length (9)
an open-source network analysis and visualization software package
written in Java (5)
An effect that causes different signals to become indistinguishable when
sampled (8)
Type of diagram that shows all possible logical relations between a finite
collection of sets (4)
Symbol or icon used to represent data values (5)
Solution to October 2014 crossword
(More details can be found in
http://www.mckinsey.com/insights/innovation/hal_varian_on_how_the_web_challenges_managers)
We are overwhelmed by the responses and solutions received from our enthusiastic readers
Congratulations!
ALL correct answers to October 2014 month’s crossword received from the
following readers:
Prof. Suresh Kumar (Dept. of Computer Science and Engineering, Manav
Rachna International University, Faridabad)
CSI Communications | November 2014 | 48
www.csi-india.org
Ask an Expert
Dr. Debasish Jana
Editor, CSI Communications
Your Question, Our Answer
“There is a magic in graphs. The profile of a curve reveals in a flash a whole situation — the life history of an epidemic, a panic, or an era of
prosperity. The curve informs the mind, awakens the imagination, convinces.”
~ Henry D. Hubbard, 1939
On Object Oriented Programming
From: Himanshu Raghav, B.Sc. Final Year,
Dev Sanskriti Vishwavidyalaya, Haridwar,
Uttrakhand
I want to know the concepts of object oriented
programming with example that I can easily
understand them and not forget them in my whole
life time.
A
Object Oriented Programming (OOP) involves
a new paradigm of concepts to programmers
of traditional procedural languages such as Pascal,
C, FORTRAN, COBOL etc. These new ideas, such as,
data hiding, encapsulation and polymorphism lie at the
heart of OOP. In OO system, the system is based
on objects rather than on actions. Here, procedures
operate on abstract values called objects rather than
on stored representations. Objects interact with each
other through well-defined interfaces. Objects can be
created and destroyed dynamically. Commonality of
objects can be made explicit by using inheritance.
OO paradigm has two important philosophies:
data hiding and data abstraction. Data hiding philosophy
says to partition the program so that data is hidden
in modules such that users of the service shouldn't
know the underlying implementation. Internal
representation can be accessed from internal
implementation and not by the users of the modules.
Data abstraction philosophy says to decide which
types are needed, to provide full set of operations
for each type so that a new type of data if defined
can be used similar to built in type of data with all
sort of operations permissible. Data abstraction is
the decision to concentrate on properties, which are
shared by many objects or situations in the real world,
and to ignore the differences between them.
In contrast to procedural programming paradigm
to have a large single store where all procedures work,
in Object-Oriented paradigm, procedures operate
on abstract values called objects rather than on
stored representations. OO programming paradigm
is based on the idea of communicating between
objects to simulate the temporal evolution of a set
of real world phenomena. Data as well as operations
are encapsulated in objects. Information hiding is
used to protect internal properties of an object.
Objects interact by means of message passing or
function invocations. Objects are active entities that
communicate with each other. State of an object may
change in response to some interaction requested
from some other object. In object-oriented paradigm,
objects are grouped in classes. Objects in classes are
similar enough to allow programming of the classes,
as opposed to programming of the individual objects.
Let’s take an example of approach in
implementing say, a stack using procedural paradigm
through C and same implementation using object
oriented paradigm using C++, and let’s compare the
philosophy by discussing pro’s and con’s.
‘C’ implementaion of Stack
#include <stdio.h>
#define MAX_SZ 10
typedef struct {
int data[MAX_SZ];
int sp;
} STACK;
void Init(STACK *which) {
which->sp = -1; //initialize the
stack
}
bool Push(int d, STACK *which) {
if (which->sp == MAX_SZ-1) //
overflow
return false;
which->data[++which->sp]=d;
return true;
}
bool Pop(int *d, STACK *which) {
if (which->sp == 1) // underflow
return false;
*d = which->data[which->sp--];
return true;
}
int main() {
STACK s1, s2;
Init(&s1);
Init(&s2);
Push(3, &s1);
Push(4, &s2);
Push(5, &s2);
int d;
Pop(&d, &s1); // d becomes 3
Pop(&d, &s2); // d becomes 5
Pop(&d, &s2); // d becomes 4
return 0;
}
In the above approach, we find little strength and few
weaknesses, which are kind of risky.
int data[MAX_SZ];//by default,
private
int sp; // by default, private
public:
STACK() { this->sp = -1; }
bool Push(int d) {
if (this->sp == MAX_SZ-1) //
overflow
return false;
this->data[++this->sp]=d;
return true;
}
bool Pop(int &d) {
if (this->sp == 1) // underflow
return false;
d = this->data[this->sp--];
return true;
}
};
int main() {
STACK s1, s2; //constructor call
s1.Push(3);
s2.Push(4);
s2.Push(5);
int d;
s1.Pop(d); //d passed as reference,
//d becomes 3
s2.Pop(d); // d becomes 5
s2.Pop(d); // d becomes 4
return 0;
}
Strength, first
The effect?
1.
1.
2.
The program is structured, written in a
methodical way
Multiple stacks can be used, since stack pointer
is not kept as a global data. Keeping one global
data as stack pointer restricts the usage of
multiple stacks
2.
3.
Weaknesses
1.
2.
3.
4.
5.
#define has its own hazards, if we by mistake
put a semicolon after #define MAX_SZ 10; or a
line comment i.e. //, we fall into trap of macros.
#define does not ensure datatypes, which are
essence of any data
The implementation is well exposed, yet not
protected from wrong usage by the user. For
example, the person who writes main, may
start accessing s1.data[2] directly, absolutely no
control on that.
What happens if the user (who writes main)
forgets to initialize the stack by calling Init?
Push, Pop take d as a pointer, and the pointer is
passed by value (not reference) in C. Passing
pointer is kind of risky, as, we don’t know
whether the pointer is pointing to a valid content
or not, a pointer may be null or spurious having
uninitialized data, which is very risky.
Now, let’s see if we take a C++ route through
object-orientedness.
#include <stdio.h>
const int MAX_SZ = 10;//put ; at your
will
class STACK {
4.
5.
#define has its own hazards, we don’t fall into
trap of macros. Instead, we define an integer
constant.
const int ensures datatypes, which is the
essence of any data
The implementation is well exposed, yet
protected from wrong usage by the user.
For example, the person who writes main, is
prevented from accessing s1.data[2] directly, as
the data is private within a stack.
No chance of forgetting by the user (who writes
main) to initialize the stack. constructors are
called automatically when variables (objects)
are declared.
Push, Pop take d as a integer reference, saves
memory, also, no pointer is passed explicitly,
putting less risk and also looks elegant. No risk
of un-initialized data exists.
A class is a construct for implementing a userdefined type. Once defined, such types may be
conveniently used as the languages primitive types.
The instances of a class are called objects. A class
specifies the representation of objects and a set of
operations that are applicable to such objects. We
have already discussed two important philosophies
of object oriented programming as data hiding and
data abstraction. Ideally, a class is an implementation
of an abstract data type. This implies that the
implementation details of the class are private to
the class. The public interface of a class is exposed
externally.
n
Do you have something to ask? Send your questions to CSI Communications with subject line ‘Ask an Expert’ at email address csic@csi-india.org
CSI Communications | November 2014 | 49
Happenings@ICT
H R Mohan
ICT Consultant, Former AVP (Systems), The Hindu & President, CSI
Email: hrmohan.csi@gmail .com
ICT News Briefs in October 2014
The following are the ICT news and headlines
of interest in October 2014. They have been
compiled from various news & Internet sources
including the dailies - The Hindu, Business Line,
and Economic Times.
Voices & Views
•
Smartphones, data plans cheapest in
India, the Internet access remains beyond
the grasp of about 95 crore people. It is
predicted about 30 crore Internet users will
be added by 2018, taking the total number
to 50 crore users. Even the cheapest data
plans are simply too expensive, equal to
13% of the total spending of people in the
segment – McKinsey.
•
The demand for regular satellite capacity
in India has been growing at over 6%
between 2008 and 2013 and now
reached 214 transponder equivalents –
Euroconsult.
•
Better urban planning through telecom
networks – LIRNEasia.
•
The global semiconductor revenue is
on pace to reach $338 billion in 2014,
a 7.2% increase from 2013. In 2014,
unit production of smartphones and
ultramobiles to increase 27% and 18.9%
respectively – Gartner.
•
Smartphones, a tool for driving
productivity. Globally, one-third of
business smartphone users said their
devices save them more than 5 hours
during an average working week – survey.
•
The domestic demand of electronic goods
is projected to grow to $400 billion by
2020, of which domestic production can
cater to only $100 billion.
•
India is a test-bed to create tools, ideas
for the world. Internet growth in India will
have a big global impact – Zuckerberg.
•
India is destination of choice for
engineering R&D - BVR Mohan Reddy,
Vice-Chairman, Nasscom.
•
India today is a home to over 600 ER&D
companies, and over 400 global ER&D
organisations with 2,00,000 engineers
are employed by service providers and
engineering firms.
•
As e-commerce grows, protecting
consumers is a challenge – Paswan.
•
Cyclone Hudhud hits, 1000 crore revenue
generating IT industry at Vizag.
•
Online retailing - both direct and through
marketplaces - will become a Rs. 50,000
crore industry by 2016, growing at a
whopping 50-55% annually over the next
three years - Crisil Research.
•
Mobile messaging apps eating into telcos’
SMS revenue. Operators worldwide will
lose $14 billion to OTT services this year
- Juniper Research.
•
E-commerce sector, which is a little less
than one per cent (at $3.2 billion) of the
$700-billion retail industry, has attracted
several global investors in a short span of
time – Industry experts.
•
India is the third largest base for startups in the world with 3,100 of them after
the US (41,500) and the UK (4,000) Nasscom.
Govt, Policy, Telecom, Compliance
•
Vizag smart city to get US-funding.
•
DoT ropes in ISRO scientists to explore
use of satellite for broadband.
•
Four years after receiving 4G spectrum,
telecom operators - Airtel, Reliance Jio and
Aircel have approached the DoTseeking
another five years to complete their
rollout obligation.
•
Two electronic manufacturing clusters
(EMC) coming up in Madhya Pradesh
at Purva near Jabalpur and Badwai near
Bhopal. They will be the first greenfield
EMCs in the country.
•
Audit by CAG finds under-utilisation
(compared to a total collection of Rs.
58,579 crore, only Rs. 17,947 crore has
been disbursed), misuse of Universal
Services Obligation funds. In three circles
(Kerala, Tamil Nadu and Karnataka),
subsidy amount of Rs. 20.34 crore was
allowed for 32,759 multiple connections
in the name of one person/address and
for more than one member of a family at
the same address.
•
Centre to use Aadhaar-based system to
track staff attendance.
•
Zuckerberg meets PM and IT Minister to
discuss how Facebook can help with the
Digital India mission.
•
Govt to start reimbursing electronic firms
under special incentive scheme soon.
•
2G auction: TRAI for 10% hike in 1800
MHz base price at Rs. 2,138 crore per
MHz.
•
Karnataka drafts policy on rural BPOs;
Cabinet nod soon.
•
India may dilute stand on Net control.
Proposes backing for the popular view on
a multi-stakeholder approach to Internet
Governance.
•
US-based Iridium wants a piece of the
Indian satellite services play.
•
Green signal for entire SmartCity Kochi
project.
•
IT firms get domestic boost due to Govt.
IT spending.
•
Japan’s SoftBank to invest $10 billion in
India. Picks up stake in Snapdeal ($627m),
Ola ($210m)
IT Manpower, Staffing & Top Moves
•
EMC at Jabalpur will generate around
3,000 direct employment opportunities
and around 9,000 indirect employment
opportunities.
•
Kris Gopalakrishnan’s innings at Infosys
set to end on 10th Oct 2014. He recently
donated $35 million to IISc for brain
research and $1.8 million to CMU to fund
its research partnership with IISc.
•
Yahoo! India R&D Centre lays off 600 in
engineering & product development.
•
S. Ramadorai steps down from Tata
companies.
•
Out of Yahoo! into waiting arms of startups, e-commerce firms.
•
Pegasystems to tap varsities, Java, .NET
professionals.
•
TCS plans to hire 35,000 from campus
this year.
•
Software majors on hiring spree in Sept
quarter.
CSI Communications | November 2014 | 50
Company News: Tie-ups, Joint Ventures, New
Initiatives
•
The country’s, claimed to be the first
dedicated medical search engine - www.
braham.in to emerge as a one-stop-shop
information source for all medical and
healthcare-related information has been
lauched.
•
Online car rental companies - Meri Cabs,
Ola Cabs have launched digital wallets.
Others like Savaari, Taxiforsure, TabCab,
Uber and Mytaxiindia are also said to be
in talks with firms such as Citrus, PayU
and Paytm for creating a mobile or digital
wallet.
•
Microsoft to continue with the Nokia
brand continues in the entry and feature
phones for the next 10 years. In the
smartphones space, they will be branded
as Lumia.
•
Flipkart fumbles on the “Big Billion Day”
on 6th Oct 2014, as server fails. Flipkart
says sorry for mega snag. But claims to
a have received a billion hits with about
1.5 million people shopped and achieved
a 24-hour sales target of $100 million in
gross merchandise value in just 10 hours.
•
Hewlett-Packard spinoff likely to hit
Mphasis outsourcing biz.
•
Flipkart campaign leads retailers to
Amazon as www.bigbillionday.com were
redirected to rival online marketplace
www.amazon.in
•
Pre booking starts by Apple for its launch
of iPhone 6 and iPhone 6 Plus in India on
October 17.
•
Microsoft Devices will now import phones
from the Hanoi (Vietnam) factory instead
of the facility owned by Nokia in Chennai.
•
DigitSecure has launched online and
mobile social wallet platform – HotRemit
for money transfer and payments service.
•
Citi downgrades IT sector; shares of Infy,
TechM, Mindtree tank.
•
Facebook’s Zuckerberg in India to focus
on driving Internet access.
•
Infibeam.com’s model, ‘live online stores’
- buildabazaar.com, has gathered over
30,000 live stores, enabling merchants
or brand owners to set up their own
e-retailing portals and offer discounts as
they deem fit for their business.
•
Microsoft
Ventures
is
accepting
applications for the next batch of its
Accelerator programme aimed at
technology start-ups.
•
With Drones and 3D visuals, real-estate
players innovate to attract buyers.
•
Flipkart to unveil world’s first dual-screen
phone today .
•
Philips unveils handset, with large font
size keypad, a torch-cum-SOS button and
long lasting battery life targeting senior
citizens.
•
Google’s Inbox promises to act as your
personal secretary.
•
Zensar Tech sets up the first National
Digital Literacy Mission (NDLM) Centre
in Pune.
•
Dell offers cloud-based solution targeting
mid-size hospital chains.
n
www.csi-india.org
On the Shelf!
Peeyush Chomal
Sr. Technical Officer, C-DAC, Mumbai
Book Review »
Book Title
:
C++ and Object-Oriented Programming Paradigm, 3rd edition
Author
:
Dr. Debasish Jana
ISBN
:
978-81-203-5033-5
No. of Pages
:
551
Price
:
Rs. 495/-
Publisher
:
Prentice Hall India
About the Author
First I must congratulate Dr. Debasish Jana for successfully
bringing out 3rd edition of his book C++ and Object-Oriented
Programming
Paradigm,
which
reflects
upon
acceptance of his
textbook as a fine
academic material
amongst students,
professionals and
casual
readers’
community. I have
grown up reading
books
in
C++
dominated largely
by foreign authors
such as Stroustoup,
Lippman
with
few contributions
in that area by
Indian
Authors.
It
is
definitely
pleasant to see a quality material coming from one more author
with Indian origin who has painstakingly addressed almost each
aspect of C++ and OO paradigm together in roughly 500 pages.
In other words, the book is not voluminous which otherwise tends
to discourage students/casual readers in carrying a bulky load of
material to read through but at the same time is addressing all the
critical mass within the paradigm and C++ as a tool for attaining it.
What has further impressed me is that Dr. Jana has tried to
introduce all the relevant concepts surrounding a particular topic
while discussing it in his book. For example, most books limit their
talk to static
binding
and
dynamic binding
whereas Dr Jana
has introduced
his readers to
even
concept
of static typing
and dynamic typing and presented the matrix showcasing which
language/tool provides for all four in his very first chapter. But he
has done this without digressing too much into the nitty-gritties
which otherwise would tend to distract reader from main idea into
unnecessary detailing. He has basically given pointers to readers who
can at their will then decide to pursue the in-depth understanding
from alternative material subject to his/her interest in the area.
The book consists of 15 chapters with last two chapters
discussing on Object Oriented Design and Modelling, and
Unified Modelling Language. Each chapter is concluded with
questions challenging reader to assess his/her conceptual
understanding of the topic. The book also comprises of problems
(for laboratory workouts), glossary, bibliography and Index which
take up around 50 pages towards the end. I also appreciate him
for covering pre-processor directives, operator overloading and
advanced concepts in dedicated chapters of their own. He also
has placed emphasis on importance of program design through
chapters on Data abstraction through classes and user defined
data types, and Data Structures and Applications in C++. I have
found the textual material well-balanced and well-supported
with diagrams and numerous examples. I therefore recommend
this book to find place on your shelf, your collections and your
Institution’s library, if it has one.
n
Peeyush Chomal serves as Sr Technical Officer at C-DAC, Mumbai (Erstwhile NCST) in Research & Development
areas of Software Architectures, Middleware Computing and Sensor Technologies. Equipped with Postgraduation from
University of Madras, Graduation from University of Mumbai and MBA from JBIMS, he has 15 years of experience
spread across Turnkey Project deliveries for DeitY, IAEA, BARC, DGoV and NPCIL. In his free time, he enjoys fiddling
with open source tools, trying out new gadgets and messing with ROMs on Android/Window platform.
CSI Communications | November 2014 | 51
CSI Report
From CSI SIG and Divisions »
Please check detailed news at:
http://www.csi-india.org/web/guest/csic-reports
SPEAKER(S)
TOPIC AND GIST
Information Retrieval Society of India, IEEE Uttar Pradesh Section, IEEE Computer Society GLA University SB and CSI Region-1, Division-1 &
Mathura Chapter, Department of Computer Engineering & Applications, GLA University, Mathura, India
Prof. DS Chauhan, Prof. Naresh Chauhan, Dr. Sujoy Das, 27 September 2014: One day National Workshop on “Emerging Trends in
Prof. MM Sufyan Beg, Prof. Krishna Kant, Prof. Anand Singh Information Retrieval (ETIR – 2014)”
Jalal, Prof. Charul Bhatnagar, Dr. Dilip Kumar Sharma
L-R: Dr. Dilip Kr Sharma, Prof. Krishna Kant, Prof. DS Chauhan,
Dr. Sujoy Das, Prof. Naresh Chauhan & Prof. AS Jalal
CSI Communications | November 2014 | 52
Dr. Dilip introduced workshop theme & focused on need of Information Retrieval
in day to day life. He discussed about how other Computer Science domains
are contributing to Information Retrieval (IR) research area. Emerging trends of
Information Retrieval include cross linguistic IR systems, recommender systems,
social network analysis & temporal Information Retrieval. He explained how
music IR system can play song(s) based on our mood. Dr. Sujoy discussed basic
concepts of IR, IR models & evolution and explained how to search faster. He
gave emphasis on how to index data, phases of indexing & its applications while
assigning tokens. Prof Beg unveiled concept of Precisiating natural language for
question answer system. He explained how natural language processes English
language as well as functioning of computer system to understand English
language through Stenford POS tagger. Dr. Sujoy Das along with Ms. Aarti
Kumar, Ms. Anubha Jain, Avinash Samuel & Mohd. Amir Khan gave hands on
experience on Terrier IR tool to participants.
www.csi-india.org
A.V. Ramachandra Rao Memorial Award for Best Papers in Hardware Area
We are pleased to announce a new award - A.V. Ramachandra Rao Memorial Award for Best Papers in Hardware Area, named after
the late husband of one of our Fellows, Dr Swarnalatha Rao.
Two cash awards of Rs. 10,000/= and Rs. 5,000/= each will be presented to the Best Papers presented in the area of Hardware in the
conferences organized by the CSI Units – Chapters, Divisions, Regions during the year and ranked First & Second.
These awards will be presented for the first time in the Annual Convention CSI-2015 to be held at New Delhi in 2015.
The Programme Committee chairs of conferences being held from Apr 2014 till Mar 2015 are requested to nominate and forward
three shortlisted papers in the area of Hardware to awards@csi-india.org for review and final selection. For more details on the award,
pl. visit www.csi-india.org
Brief profile of
Late Shri. A. V. Ramachandra Rao
(1-2-1934 to 2-5-2014)
Late Shri. A. V. Ramachandra Rao was an eminent technologist and entrepreneur in electronic devices and equipment. Over a period
of two decades, in the 1960’s and 1970’s, in the formative years of Bharat Electronics Ltd (BEL), he lead teams in the development
and manufacturing of X-Ray tubes and systems, electronic tubes of various kinds for civilian and military communication and radar
applications, black and white CRTs for use in medical diagnostic systems and in televisions etc. He had aslo contributed to the
development of germanium semiconductor devices.
On leaving BEL, Shri Ramachandra Rao became an entrepreneur, and concentrated in developing and manufacturing components and
subsystems for import substitution in important electronic equipment, such as deflection components in TVs, pulse transformers
especially for high end power electronic equipment etc.
Shri Ramachandra Rao on invitation by the West Bengal Electronics Corporation, involved in the setting up their TV Picture Tube
manufacturing plant – Webel Video Devices in late 1970’s and served as its Managing Director.
He had also set up successful companies in Chennai and Bangalore, concentrating on import substitutions in areas which were
important at that time, thus enabling the country to save precious foreign exchange.
Shri. Ramachandra Rao was a keen follower of technological developments taking place across the world. He admired the culture
and achievements of western societies. He was deeply interested in education and training of the younger generation in India to do
innovative contributions at the level of the advanced industrialized countries.
CSI Communications | November 2014 | 53
DETAILS OF CSI GRANTS AVAILABLE FOR CHAPTERS/STUDENT
BRANCHES/MEMBERS
(Presented by Mr. Ranga Rajagopal. Hon. Treasurer – CSI)
CSI has made available several new categories of grants for the benefit of members during the last few years. Following are the
categories of grants available to CSI chapters, student branches and members (as applicable). Members/chapters are requested
to avail the grants provided for the fiscal year before 31st March 2015. Members may share any feedback/suggestions in this
regard with treasurer@csi-india.org.
- Tech Bridge - Grant of 5k per event of the chapter (upto 2 events per year) to be availed before 31st March 2015.
Refer http://www.csi-india.org/grant-to-csi-chapters for details
Total Grant budget for 2014-15 Rs. 5.0 Lacs
- Golden Jubilee Chapter Grant (50k for Cat A, 25K for Cat B and 15K for Cat C) - a one
time grant to chapters for a Golden Jubilee event to be conducted before March 31st 2015 to enable all chapter members/Past OBs/
Patrons/Fellows to participate. Refer http://www.csi-india.org/grant-to-csi-chapters for details
Total Grant budget for 2014-15 Rs. 10. Lacs
- Divisional Grant - 25 K available for any chapter level technical event conducted in association with any Division. Event should be
planned and announced well in advance. Refer http://www.csi-india.org/divisions for details
Total Grant budget for 2014-15 Rs. 5.0 Lacs
- Tech Bridge (for Student chapters) - 5k for 1 event during the year for the 100
proposals received during the year 2014-15.
Refer http://www.csi-india.org/web/education-directorate/grant-to-student-branch
Total Grant budget for 2014-15 Rs. 5.0 Lacs
- Travel grant for Research scholars/students for participation in International Conferences upto Rs. 25K. (Available for first 12
proposals received during the year).
Refer http://www.csi-india.org/c/document_library/get_file?uuid=944dd521-20c9-4541-aaf5-a2a6de694be0&groupId=10157
Total Grant budget available for 2014-15 Rs. 3.0 Lacs
- Student Convention grant for Student Branches (25k for State Student convention, 35k for Regional Student Convention and 70k
for National Student Convention). Claim format available with Education Directorate. Convention proposal to be submitted to ED/
NSC for approval as per guidelines in advance
Total Grant budget for 2014-15 Rs. 6.0 Lacs
-Support grant to chapters for conducting Chapter level/Regional level/National level rounds of Discover Thinking School Quiz,
Discover Thinking Project Contest. Alan Turing Programming Contest, Discover Thinking Online Quiz for Student members. Refer
notification in CSIC or contact csiprograms@csi-india.org
Total Grant budget for 2014-15 Rs. 5.0 Lacs
CSI Communications | November 2014 | 54
www.csi-india.org
Application for Travel Grants for Researchers
Research Committee of Computer Society of India has decided to partly fund CSI Life Members to the extent of Rs. 25000/ for
travelling abroad to present research papers at conferences.
CSI Life Members who have been invited to present papers abroad and have received partial or no funding are eligible to apply for the
same. They have to apply within December 31, 2014 to div5@csi-india.org and furnish:
1.
2.
3.
4.
5.
6.
7.
8.
9.
Name of the Applicant, Organization Details and Bio Data of Applicant
CSI Life Membership Number
Name of the International Conference with details of the organizers
Conference Venue and Date
Copy of the Research Paper
Copy of the Invitation Letter received from the organizers
Details of funding received from/applied to for funding to any other agency
Justification for requesting support (in 100 words).
Two References (including one from head of the organization)
Dr Anirban Basu
Chairman,
CSI Division V (Education and Research)
CSI Communications | November 2014 | 55
CSI News
From CSI Chapters »
Please check detailed news at:
http://www.csi-india.org/web/guest/csic-chapters-sbs-news
SPEAKER(S)
TOPIC AND GIST
DELHI (REGION I)
Dr. Roop N Bharadwaj, VK Gupta, Dr. AK Bansal
and SD Sharma
13 September 2014: Technical talk on “Innovation Means of IT &
Telecommunication of Education”
Mr. Sharma explained how new ideas and innovations in IT are impacting our
day to day life. Dr. Bhardwaj covered the topic on e-learning - education with
innovative means, how it was 30 years before with limited technical support
and how it is as on date with latest technologies. He spoke about how IT is
playing useful role in imparting education to masses living in remote areas.
Speaker and participants during technical talk
HARIDWAR (REGION I)
Dr. Satish K. Peddoju, Lt.Col. (Rtd.) PK Jain,
Prof. VK Sharma, Dr. Mayank Aggrawal and
Dr. Mahendra Singh Aswal
11 October 2014: Expert Session on “Cloud Computing”
Dr. Satish delivered lecture on very current and popular issue of “Cloud
Computing”. He covered all aspects of cloud computing including
Introduction, tools knowledge, practical view of cloud, simulators required
for research and how it works. The session was full of information for all
students, researchers, teachers etc.
Faculty members and participants
CHENNAI (REGION VII)
HR Mohan, Judges-Prof. P Thrimurthy, S Ramanathan
and Bhaskaran assisted by Prof. P Kumar and
Bhuvaneswaran
9 & 12 October 2014: SEARCC – International Schools’ Software Competition
(ISSC) 2014
Participating team had 3 students each. 2 teams from India, 2 from
Sri Lanka, 1 from ROC Taiwan and 1 from Papua New Guinea participated.
Trial competition was held on 11th Oct 2014 and all teams participated in
it for getting experience. Software displaying minute to minute position
was very handy. Result: First- ROC Taiwan, Second-India (Team-B) and
Third- Sri Lanka (Team-A). Mr. Mohan gave away prizes and addressed
the gathering. SEARCC Rolling Trophy was given away to the Taiwan team.
Prize winners with CSI President HR Mohan, Dr. Thangam Meganathan and
other dignitaries of CSI
TRIVANDRUM (REGION VII)
Mr. Vinod Purushothaman
6 August 2014: Technical talk on “Connecting with Agile Principles and
Practices”
CSI Trivandrum Chapter organized a technical talk on ‘Connecting with
Agile Principles and Practices’ by Mr. Vinod Purushothaman, Technical
Architect, Envestnet Inc. (NYSE: ENV) at Institution of Engineers Hall,
Thiruvananthapuram.
Mr. Vinod Purushothman delivering the lecture
CSI Communications | November 2014 | 56
www.csi-india.org
TRIVANDRUM (REGION VII)
24 September 2014: Technical talk on the topic “Time
Management Using A Tomato”
Mr. Ramnath Jayakumar
CSI Trivandrum Chapter organized a Technical talk on the
topic ‘Time Management Using A Tomato’ by Mr. Ramnath
Jayakumar, Lead Engineer, Envestnet Inc. (NYSE: ENV),
Thiruvananthapuram at The Institution of Engineers Hall,
Thiruvananthapuram.
Mr. Ramnath Jayakumar delivering the lecture
From Student Branches »
(REGION-V)
CMR TECHNICAL CAMPUS, HYDERABAD
16th Oct 14: Mr. Somagiri delivered a lecture on “Big Data Analytics"
(REGION-V)
(REGION-V)
ANURAG GROUP OF INSTITUTIONS, HYDERABAD
9th Oct, 2014 : Conducted an one day “TECHNICAL QUIZ CONTEST”
to bring out the talent of the students in Computer programming skills
and winners with certificates awarded by the Principal
(REGION-V)
SREE VIDYANIKETHAN INSTITUTE OF MANAGEMENT, TIRUPATI
LAKIREDDY BALI REDDY COLLEGE OF ENGINEERING, MYLAVARAM
14th Oct, 2014: New student branch was inaugurated by Consultant
Mr. S. Ramasamy and Mr. Y. Kathiresan, Dr. Mohan, Dean, Mr. Prabhakar,
Guest speaker are during the inauguration
20th and 21st of Sep, 2014: Two day workshop on “ASP.NET MVC4.0”
was conducted by Mr. V. Phani Sekhar, Senior Software Engineering from
HPS, Bangalore
(REGION-V)
(REGION-V)
KLE DR M S SHESHGIRI COLLEGE OF ENGINEERING AND TECHNOLOGY,
BELGAUM
SWARNANDHRA INSTITUTE OF ENGINEERING AND TECHNOLOGY (SIET),
NARASAPUR
27th Aug, 2014: Mr. Prasad Inchal, Senior Consultant, HP, Bangalore
inaugurated the CSI activities for the year 2014-15
15th Sep, 2014: Activities for 2014-15 inaugurated Prof, J V R Murthy,
JNTUK – Kakinada. Prof. Rajesh C Jampala, Chairman - CSI Vijayawada
chapter and Principal, Dr. T Madhu with the Student branch Plaque
CSI Communications | November 2014 | 57
(REGION-VI)
(REGION-VII)
P.E.S COLLEGE OF ENGINEERING, AURANGABAD
DE PAUL INSTITUTE OF SCIENCE AND TECHNOLOGY, ANGAMALY
12th Oct, 2014: Orphan children and Students of the college along with
Prof. V.A.Losarwar, HOD & Prof. A.U.Jadhav
7th Oct, 2014: Workshop on Ethical Hacking and Web Application
Security by Mr. Manu Zacharia, Information Security Evangelist
(REGION-VII)
(REGION-VII)
SRINIVASA RAMANUJAN CENTRE, SASTRA UNIVERSITY, KUMBAKONAM
SENGAMALA THAYAR EDUCATIONAL TRUST STUDENT BRANCH
9th Oct 14: Event INNOVATE-AD based on advertisement for enriching
student’s creativity and innovative ideas organized wherein 26 teams
from various streams of Engineering, Arts and Science participated.
26th Sep, 2014: National Seminar on Cloud computing was conducted.
Correspondent , Principal, Tamilnadu State student Coordinator
Dr. Srinath and Mr. S. Ramasamy, Consultant are on the dais
(REGION-VII)
(REGION-VII)
JYOTHI ENGINEERING COLLEGE, THRISSUR
MCA DEPARTMENT OF COMPUTING SCIENCE, VELS UNIVERSITY, CHENNAI
22nd Sep, 2014: Computer Awareness Programme for plus students was
conducted by CSI Student Volunteers Mr. Gijo Vargese, Mr. Joe Mathew
and Mr. Sanjo Simon under the guidance of Mr. Viju Shankar, SBC
1st Oct, 2014: Mr. Rajan T.Joseph Director-Education, CSI was the Chief
Guest and delivered the Inaugural address in the National level Technical
symposium CIPHERMIX 2014
(REGION-VII)
(REGION-VII)
JEPPIAAR ENGINEERING COLLEGE , CHENNAI
ST. PETER’S UNIVERSITY, CHENNAI
25th & 26th Sep, 14: Workshop On NS2 Simulations by Mr. Pradeep Kumar Inauguration of New student branch by Mr. Rajan Joseph, Director
along with Principal Dr. Sushil Lal Das , Director Mrs. M. Regeena Jeppiaar, Education with Vice Chancellor
CSE HOD Dr. V. L. Jyothi, IT HOD Dr. R. Sabeetha
Please send your student branch news to Education Director at director.edu@csi-india.org. News sent to any other email id will not be considered.
Low-resolution photos and news without gist will not be published. Please send only 1 photo per event, not more.
CSI Communications | November 2014 | 58
www.csi-india.org
CSI Calendar
2014
Date
Prof. Bipin V Mehta
Vice President, CSI & Chairman, Conf. Committee
Email: bvmehta@aesics.ac.in
Event Details & Organizers
Contact Information
November 2014 events
14–16 Nov 2014
International Conference on Emerging Computing Technologies-2014
(ICECT-2014) Organized by Dept. of Computer Science and Applications, M.
D. University, Rohtak in association with CSI Region – I and CS Division – I.
Prof. R S Chhillar
chhillar02@gmail.com
14–16 Nov 2014
International Conference on Information and Communication Technology
for Competitive strategies (ICTCS-2014) Organized by: Computer Society of
India, Udaipur Chapter, Division IV, I, SIG-WNs, Hosted by: Sunrise Group of
Institutions, Udaipur. http://www.csi-udaipur.org/ictcs-2014
Prof. Amit Joshi
Organizing Secretary
amitjoshiudr@gmail.com
28-30 Nov 2014
International Conference on Advance in Computing Communication and
Informatics at COER School of Management, Roorkee , Uttrakhand
http://coer.ac.in/ICACCI2014/index.html
Dr. Vishal Singhal, Convener
fdpcoersm@coer.ac.in
December 2014 events
10-11 Dec 2014
49th Annual Student Convention, Organized by Computer Society of India,
Hyderabad Chapter In association with GNIT, Hyderabad. Theme: “ Campus to
Corporate” Venue: GNIT, Ibrahimpatnam, Rangareddy District Telangana.
http://www.csihyderabad.org/csi-2014
Dr. DD Sarma,
Shri Raju Kanchibhotla
Shri Chandra Sekhar Dasaka.
http://www.csihyderabad.
org/csi-2014
12-14 Dec 2014
49th Annual Convention ,Organized by Computer Society of India,
Hyderabad Chapter In association with JNTU-Hyderabad & DRDO. Theme:
Emerging ICT for Bridging Future Venue: JNTUH, Kukatpally, Hyderabad
http://www.csihyderabad.org/csi-2014
Sri. J A Chowdary
Sri. GautamMahapatra
csi2014@csihyderabad.org
12-14 Dec 2014
Special session on “Cyber Security and Digital Forensics” during Computer Society
of India Annual Convention - 2014 by CSI Special Interest Group on Cyber Forensics,
JNTU Hyderabad
Dr. Vipin Tyagi
dr.vipin.tyagi@gmail.com
16-20 Dec 2014
ICISS-2014: International Conference on Information Systems Security. At
Institute for Development & Research in Banking Technology (IDRBT), Hyderabad,
India. Co-sponsored by CSI Division IV and CSI SIG-IS.
http://www.idrbt.ac.in/ICISS_2014/
iciss2014@idrbt.ac.in
19-21 Dec 2014
EAIT-2014: Fourth International Conference on Emerging Applications of
Information Technology at Kolkata. Organized by CSI Kolkata at Indian Statistical
Institute, Kolkata https://sites.google.com/site/csieait/ For paper ssubmission :
https://cmt.research.microsoft.com/EAIT2014
Prof. Aditya Bagchi
Dr. Debasish Jana
Prof. Pinakpani Pal
Prof. R T Goswami
csieait@gmail.com
22-24 Dec 2014
ICHPCA-2014: International Conference on High Performance Computing and
Applications Organized by: CV Raman College of Engg. in association with CSI
Div-V and IEEE Kolkata Section http://www.ichpca-2014.in/
Prof. (Dr.) Rachita Misra
ichpca2014@gmail.com
9th INDIACom; 2015 2nd International Conference on “Computing for Sustainable
Global Development” Organized by Bharati Vidyapeeth’s Institute of Computer
Applications and Management (BVICAM), New Delhi
Prof. MN Hoda
conference@bvicam.ac.in,
indiacom2015@gmail.com
March 2015
11–13 Mar 2015
CSI Communications | November 2014 | 59
Registered with Registrar of News Papers for India - RNI 31668/78
Regd. No. MH/MR/N/222/MBI/12-14
Posting Date: 10 & 11 every month. Posted at Patrika Channel Mumbai-I
Date of Publication: 10 & 11 every month
If undelivered return to :
Samruddhi Venture Park, Unit No.3,
4th floor, MIDC, Andheri (E). Mumbai-400 093
CSI-2014
Annual Convention and
International Conference on Emerging ICT for Bridging Future
Hosted by: CSI Hyderabad Chapter
In Association with JNTU Hyderabad & DRDO
Dates: 12th- 14th December 2014, Venue: JNTU Hyderabad
www.csihyderabad.org/csi-2014, www.csi-2014.org
Call for Sponsors/ Participation
Introduction: CSI-2014, the 49th Annual Convention of Computer Society of India (CSI) is being organized as a part of CSI@50, the Golden
Jubilee celebrations of CSI by CSI Hyderabad Chapter, in association with Jawaharlal Nehru Technological University, Hyderabad and DRDO.
The Golden Jubilee Celebration along with International Conference will be held at JNTU Hyderabad on 12th ,13th and 14th December 2014. The
theme of the convention is “Emerging ICT for Bridging Future”. The objective of this convention is to bring together researchers, engineers,
developers, practitioners, IT professionals from academia, industry, government establishments, SME, Public Sectors and multi-national
companies and share their experience, exchange ideas and update their knowledge on the latest developments in emerging areas. As part
of this convention Knowledge sharing sessions on e-governance have been organized where the implementers, policy makers, users and
developers from various agencies will be deliberating regarding successful implementation of E-Governance for achieving the vision of Digital
India. In this convention many IT luminaries, famous personalities from industries, Govt and Public sectors are participating and deliberating
from various aspects of ICT’s for IT enabling of India. Number of Keynote sessions, CIO’s panel discussions is also part of this convention. Large
scale exhibition from various IT firms is one of the main attractions. We have already received more than 160 high quality research papers on
all aspects of the ICT’s and same will be presented in this convention in various parallel tracks.
National Student convention is also being organized at Gurunanak Group of Institutions, Ibrahimpatnam, Hyderabad to make the student
community glide through “Campus to Corporate and Beyond” on 10th and 11th December 2014.
Invitation: Govt, Public sectors, Educational Institutes, Software firms, industries and business houses are invited to participate in the
convention and present and exhibit their products and services. Online registrations facilities are provided in the website: www.csi-2014.org.
We also invites proposals for workshops, pre-conference tutorials and doctoral consortium.
Registration fee and Sponsorship plan is given below:
Souvenir
Registrations
Delegate Type
CSI Ins(Rs)
Others(Rs)
Regular
3000
4000
150
Paper Presenter
Springer
5000
6000
250
Student For Stu.
Convention
400
500
25
Amount (Rs)
Full Page
50000
Half Page
25000
Qtr. Page
15000
Exhibition Stall 9x6ft:Rs 30000
Sponsorship Plans
Facilities Given
Amount
(Rs)
Registrations
Exhibition Stall
Ad Pages
Crown
700000
10
2
2
Platinum
500000
5
2
1
Diamond
300000
3
1
1
Gold
200000
2
1
1
Silver
100000
1
Shared
Half Page
Bronze
50000
1
None
Half Page
Plan
Page
Overseas($)
Delegates will be provided accommodation on first come
first serve basis.Transport to venue will be provided by the
convention team.
*12.36% service tax extra.
Address for Communication:
CSI-Hyderabad Chapter,
#302,Archana Aracde,10-3-190,
Opp: Railway Reservation Complex,
Secunderabad,Telangana-500025
Email:csi2014@csihyderabad.org
contact@csihyderabad.org
Payments to be made by DD/Cheque drawn in favour of:
“CSI Annual Convention” Payable at MUMBAI or by RTGS/NEFT at A/c no: 34242332507
IFSC Code: SBIN0007074, Service Tax Registration No: AAATC1710FSD001
Sri. J A Chowdary
Dr. A Govardhan, JNTU Hyderabad
Organizing Committee
Programme Committee
Sri. Gautam Mahapatra, RCI, DRDO
Finance Committee