Call for Abstract

3rd International Conference on Big Data, Data Structures and Data Applications , will be organized around the theme “Advance Technologies for Creative Inventions in Data Structures ”

Data Structures 2017 is comprised of keynote and speakers sessions on latest cutting edge research designed to offer comprehensive global discussions that address current issues in Data Structures 2017

Submit your abstract to any of the mentioned tracks.

Register now for the conference by choosing an appropriate package suitable to you.

Ransomware is a type of malicious software that carries out the cryptoviral extortion attack from cryptovirology that blocks access to data until a ransom is paid and displays a message requesting payment to unlock it. Simple ransomware may lock the system in a way which is not difficult for a knowledgeable person to reverse. More advanced malware encrypts the victim's files, making them inaccessible, and demands a ransom payment to decrypt them. The ransomware may also encrypt the computer's Master File Table (MFT) or the entire hard drive. Thus, ransomware is a denial-of-access attack that prevents computer users from accessing files since it is intractable to decrypt the files without the decryption key. Ransomware attacks are typically carried out using a Trojan that has a payload disguised as a legitimate file.

Cyber security includes controlling physical access to the hardware, as well as protecting against harm that may come via network access, data and code injection. Also, due to malpractice by operators, whether intentional, accidental, IT security is susceptible to being tricked into deviating from secure procedures through various methods.The field is of growing importance due to the increasing reliance on computer systems and the Internet in mostly developed (first-world) societies, wireless networks such as Bluetooth and Wi-Fi, and the growth of "smart" devices, including smartphones, televisions and tiny devices as part of the Internet of Things.

Information Mining Applications in Engineering and Medicine centres to offer data excavators who wish to apply unmistakable data some help with mining frameworks. These applications consolidate Data mining structures in cash related business segment examination, Application of data mining in preparing, Data mining and Web Application, Medical Data Mining, Data Mining in Healthcare, Engineering data mining, Data Mining in security, Social Data Mining, Neural Networks and Data Mining, these are a segment of the uses of data Mining.

  • Data mining systems in financial market analysis
  • Application of data mining in education
  • Data mining and processing in bioinformatics, genomics and biometrics
  • Advanced Database and Web Application
  • Medical Data Mining
  • Data Mining in Healthcare data
  • Engineering data mining
  • Data mining in security


Big data analytics examines huge amounts of data to uncover hidden patterns, correlations and other insights. With today’s technology, it’s possible to analyze our data and get answers from it almost instantly – an effort that’s slower and less efficient with more traditional intelligence solutions

  • Track 4-1Volume Growth of Analytic Big Data
  • Track 4-2Managing Analytic Big Data
  • Track 4-3Data Types for Big Data.
  • Track 4-4Refresh Rates for Analytic Data
  • Track 4-5Replacing Analytics Platforms
  • Track 4-6Big Data Analytics Adoption.
  • Track 4-7Benefits of Big Data Analytics
  • Track 4-8Barriers to Big Data Analytics.

Data Mining devices and program design projects incorporate Big Data Security and Privacy, Data Mining and Predictive Analytics will be in Machine Learning, Boundary to Database Systems and Software Systems.

  • Track 5-1Big Data Security and Privacy
  • Track 5-2E-commerce and Web services
  • Track 5-3Medical informatics
  • Track 5-4Visualization Analytics for Big Data
  • Track 5-5Predictive Analytics in Machine Learning and Data Mining
  • Track 5-6Interface to Database Systems and Software Systems
  • Track 5-7Potential Growth versus Commitment for Big Data Analytics Options.
  • Track 5-8 Trends for Big Data Analytics Options.

In assuming, a data conveyance focus, generally called an attempt data stockroom (EDW), is a structure used for recording and data examination. Data Warehousing are central chronicles of facilitated data from one or more different sources. This data warehousing combines Data Warehouse Architectures, Case contemplates: Data Warehousing Systems, Data warehousing in Business Intellect, Role of Hadoop in Business Intelligence and Data Warehousing, Commercial procedures of Data Warehousing, Computing EDA (Exploratory Data Analysis) Techniques, Machine Learning and Data Mining.

  • Track 6-1Data mining systems in financial market analysis
  • Track 6-2Application of data mining in education
  • Track 6-3Data mining and processing in bioinformatics, genomics and biometrics
  • Track 6-4Advanced Database and Web Application
  • Track 6-5Medical Data Mining
  • Track 6-6Data Mining in Healthcare data
  • Track 6-7Engineering data mining
  • Track 6-8Data mining in security

Big data is definitely one of the biggest tinkle phrases in IT today. Combined with virtualization and cloud computing, big data is a scientific capability that will force data centers to significantly transform and evolve within the next five years. Parallel to virtualization, big data infrastructure is unique and can create an architectural upheaval in the approach systems, storage, and software infrastructure are associated and managed.

  • Track 7-1Cloud/Grid/Stream Computing for Big Data
  • Track 7-2 High Performance/Parallel Computing Platforms for Big Data
  • Track 7-3Autonomic Computing and Cyber-infrastructure, System Architectures, Design and Deployment.
  • Track 7-4Energy-efficient Computing for Big Data
  • Track 7-5Programming Models and Environments for Cluster, Cloud, and Grid Computing to Support Big Data
  • Track 7-6Software Techniques and Architectures in Cloud/Grid/Stream Computing Big Data Open Platforms
  • Track 7-7New Programming Models for Big Data beyond Hadoop/MapReduce, STORM Software Systems to Support Big Data Computing

The objective of big data organization is to ensure a high level of data quality and accessibility for business intelligence and big data analytics uses. Corporations, government agencies and other organizations employ big data management strategies to help them struggle with fast-growing pools of data, typically involving many terabytes or even petabytes of information saved with different file formats. Effective big data management helps companies locate valuable information in big sets of unstructured data and semi-structured data from a selection of sources, including call detail records, system logs and social media sites.

  • Track 8-1Search and Mining of variety of data including scientific and engineering,social,sensor/IoT/IoE, and multimedia data
  • Track 8-2Mobility and Big Data
  • Track 8-3Semantic-based Data Mining and Data Pre-processing
  • Track 8-4Cloud/Grid/Stream Data Mining- Big Velocity Data Link and Graph Mining
  • Track 8-5Large-scale Recommendation Systems and Social Media Systems
  • Track 8-6Computational Modelling and Data Integration
  • Track 8-7Visualization Analytics for Big Data
  • Track 8-8Data Acquisition, Integration, Cleaning, and Best Practices
  • Track 8-9Big Data Search Architectures, Scalability and Efficiency
  • Track 8-10Distributed, and Peer-to-peer Search
  • Track 8-11Algorithms and Systems for Big Data Search
  • Track 8-12Multimedia and Multi-structured Data- Big Variety Data

During latest years, data production rate has been growing exponentially. Many organizations demand efficient solutions to store and analyse these big amount data that are preliminary generated from various sources such as high quantity instruments, sensors or connected devices. For this determination, big data technologies can utilize cloud computing to provide significant benefits, such as the availability of automatic tools to assemble, connect, configure and reconfigure virtualized properties on demand. These make it much easier to meet organizational goals as organizations can easily deploy cloud services.

  • Track 9-1Intrusion Detection for Gigabit Networks
  • Track 9-2Sociological Aspects of Big Data Privacy
  • Track 9-3HCI Challenges for Big Data Security & Privacy
  • Track 9-4Privacy Preserving Big Data Collection/Analytics
  • Track 9-5Privacy Threats of Big Data
  • Track 9-6Threat Detection using Big Data Analytics
  • Track 9-7Visualizing Large Scale Security Data
  • Track 9-8High Performance Cryptography
  • Track 9-9Anomaly and APT Detection in Very Large Scale Systems
  • Track 9-10Trust management in IoT and other Big Data Systems

Online Analytical Processing is an innovation that is utilized to make choice bolster programming. OLAP empowers application clients to rapidly divide data that has been outlined into multidimensional perspectives and chains of importance. Through editing anticipated inquiries into multidimensional perspectives preceding run time, OLAP apparatuses give the benefit of expanded execution over conventional database access devices. The vast majority of the asset serious count that is essential to compress the information is done before an inquiry is submitted.


  • Track 10-1Data Storage and Access
  • Track 10-2OLAP Types (WOLAP, DOLAP, RTOLAP)
  • Track 10-3OLAP Architecture
  • Track 10-4OLAP tools and internet
  • Track 10-5OLAP Functional requirements
  • Track 10-6Control of spread sheets and SQL

Enormous Data is a liberal wonder which is a standout amongst the most every now and again talked about subjects in the present age, and is relied upon to remain so within a reasonable time-frame. Aptitudes, equipment and programming, calculation design, accurate centrality, the sign to commotion proportion and the way of Big Data itself are distinguished as the significant problems which are tarnishing the way toward acquiring important gauges from Big Data.


  • Track 11-1Challenges for Forecasting with Big Data
  • Track 11-2Applications of Statistical and Data Mining Techniques for Big Data Forecasting
  • Track 11-3Forecasting the Michigan Confidence Index
  • Track 11-4Forecasting targets and characteristics

Disseminated computing is a kind of Internet-based figuring that gives shared handling assets and information to PCs and unlike devices on attention. It is a typical for authorizing pervasive, on-interest access to a common pool of configurable process assets which can be quickly provisioned and discharged with insignificant administration exertion. Distributed calculating and volume arrangements supply clients and ventures with different abilities to store and method their info in outsider info trots. It depends on distribution of assets to accomplish rationality and economy of scale, like a utility over a system.


  • Track 12-1Reference models for cloud computing
  • Track 12-2Cloud deployment models
  • Track 12-3Cloud Automation and Optimization
  • Track 12-4High Performance Computing (HPC)
  • Track 12-5Cloud-based systems
  • Track 12-6Overarching concerns

Information illustration or information observation is seen by numerous orders as a present likeness visual correspondence. It is not requested by any one field, yet rather discovers translation crosswise over numerous It envelops the arrangement and examination of the visual representation of information, signifying "data that has been dreamy in some schematic arrangement, including attributes or variables for the units of data".


  • Track 13-1Large data visualization
  • Track 13-2Scalable parallel rendering algorithms
  • Track 13-3Practical data visualization
  • Track 13-4Visual analytics techniques for large, complex network data
  • Track 13-5Future trends in scientific visualization

The way toward splitting information from source frameworks and bringing it into the information distribution center is ordinarily called ETL, which ruins for extraction, change, and stacking. Note that ETL eludes to a wide procedure, and not three very much categorized strides. The acronym ETL is maybe excessively short-sighted, on the grounds that it overlooks the passage stage and suggests that each of alternate periods of the procedure is particular. All things considered, the whole procedure is known as ETL.


  • Track 14-1ETL Basics in Data Warehousing
  • Track 14-2ETL Tools for Data Warehouses
  • Track 14-3Logical Extraction Methods
  • Track 14-4ETL data structures
  • Track 14-5Cleaning and conforming
  • Track 14-6Delivering dimension tables

Info mining frameworks and counts an interdisciplinary subfield of programming building is the computational system of discovering case in marvellous data sets including strategies like Big Data Search and Mining, Novel Theoretical Models for Big Data, High implementation data mining figuring’s, Methodologies on far reaching scale data mining, Methodologies on broad gauge data mining, Big Data and Analytics

  • Track 15-1Novel Theoretical Models for Big Data
  • Track 15-2New Computational Models for Big Data
  • Track 15-3High performance data mining algorithms
  • Track 15-4Methodologies on large-scale data mining
  • Track 15-5Empirical study of data mining algorithms

Bunching can be viewed as the most vital unsupervised learning issue; along these lines, as each other issue of this kind, it manages meaning a structure in a gathering of unlabelled information. A free meaning of bunching could be the way toward sorting out items into assemblies whose individuals are comparable somehow

  • Track 16-1Shared File System Accessibility
  • Track 16-2Density Based Clustering
  • Track 16-3Spectral and Graph Clustering
  • Track 16-4Clustering Validation