major project document
major project document
PROJECT REPORT
ON
BACHELOR OF TECHNOLOGY
IN
COMPUTER SCIENCE AND ENGINEERING
SUBMITTED BY
BACHELOR OF TECHNOLOGY
IN
COMPUTER SCIENCE AND ENGINEERING
SUBMITTED BY
CERTIFICATE
This is to certify that the project report titled “A HYBRID DEEP LEARNING
APPROACH FOR CYBERBULLYING DETECTION IN TWITTER SOCIAL
MEDIA PLATFORM” is submitted by B.UMA DEVI (196M1A0572),V.SRUTHI
(196M1A0593),N.ALEKY( 196M1A05B5) ,A .RUPA BINDU
(196M1A05B6),B.SRAVANI (206M5A0502), in B.Tech 4 2 semester COMPUTER
SCIENCE AND ENGINEERING is a record bonofide work carried out by them.
The results embodied in this report have not been submitted to any other university
for the award of any degree.or prize.
PRINCIPAL
EXTERNAL EXAMINER Dr.RAJEER SHRIVASTAVA
MTech,Ph.D
DECLARATION
The present project work in the several days study of the various aspects of the
project development. During this the effort in the present study. We have received a
great amount of help from out Chairman Dr. S. PRABAKAR RAO which we wish
to knowledge and thank from of depth our hearts.
We thankful to our Principle Dr. SRI VASTAV, Ph.D for permitting and
encouraging us in during this project.
We deeply intended to Dr. ARUL DALTON, Ph.D, and Head of the
Department, CSE whose motivation and constant encouragement has led to pursue a
project in the field of software development.
We very much obliged and thankful to our internal guide Ms. SANKU
VASAVI,MTech CSE for providing this opportunity and constant encouragement
given by her during the course. We grateful to her valuable guidance and suggestions
during our project work.
Our Parents have put ourself ahead of themselves. Because of their hard
work and dedication, we have had opportunity beyond our wildest dreams. Our heart
felt thanks to them for giving us all we ever needed to be successful student and
individual.
Finally we express our thanks to all our other professors, classmates,
friends, neighbours and our family members who helped us for the completion of our
project and without infinite love and patience this would never have been possible.
CONTENT
ABSTRACT
LIST OF SCREENS
1. INTRODUCTION 1-6
1.1 Motivation
2.1 Introduction
3. ANALYSIS 14-19
I
II
3.3.1 Algorithms
3.3.2 flowchart
4. DESIGN 20-25
5.1 Introduction
6.1 INTRODUCTION
II
III
7. VALIDAION 58-61
7.2 CONCLUSION
8. CONCLUSION 62
REFERENCES 63-64
III
IV
ABSTRACT
IV
V
LIST OF FIGURES
1 System Architecture 20
4 Class Diagram 22
5 Sequence Diagram 23
V
VI
LIST- SCREENS
VI
VII
VI
I
1
1.INTRODUCTION
1.1 MOTIVATION
Social media networks such as Face book, Twitter, Flickr, and Instagram have
become the preferred online platforms for interaction and socialization among people
of all ages. While these platforms enable people to communicate and interact in
previously unthinkable ways, they have also led to malevolent activities such as
cyber-bullying. Cyber bullying is a type of psychological abuse with a significant
impact on society. Cyber-bullying events have been increasing mostly among young
people spending most of their time navigating between different social media
platforms. Particularly, social media networks such as Twitter and Face book are
prone to CB because of their popularity and the anonymity that the Internet provides
to abusers. In India, for example, 14 percent of all harassment occurs on Face book
and Twitter, with 37 percent of these incidents involving youngsters [1]. Moreover,
cyber bullying might lead to serious mental issues and adverse mental health effects.
Most suicides are due to the anxiety, depression, stress, and social and emotional
difficulties from cyber-bullying events [2]_[4]. This motivates the need for an
approach to identify cyber bullying in social media messages (e.g., posts, tweets, and
comments).
In this article, we mainly focus on the problem of cyber bullying
detection on the Twitter platform. As cyber bullying is becoming a prevalent problem
in Twitter, the detection of cyber bullying events from tweets and provisioning
preventive measures are the primary tasks in battling cyber bullying threats [5].
Therefore, there is a greater need to increase the research on social networks-based
CB in order to get greater insights and aid in the development of effective tools and
approaches to effectively combat cyber bullying problem [6]. Manually monitoring
and controlling cyber bullying on Twitter platform is virtually impossible [7].
Furthermore, mining social media messages for cyber bullying detection is quite
difficult. For example, Twitter messages are often brief, full of slang, and may include
emojis, and gifs, which makes it impossible to deduce individuals' intentions and
meanings purely from social media messages. Moreover, bullying can be difficult to
detect if the bully uses strategies like sarcasm or passive-aggressiveness to conceal it.
1
2
Despite the challenges that social media messages bring, cyber bullying detection on
social media is an open and active research topic. Cyber bullying detection within the
Twitter platform has largely been pursued through tweet classification and to a certain
extent with topic modeling approaches. Text classification based on supervised
machine learning (ML) models are commonly used for classifying tweets into
bullying and non-bullying tweets [8]_[17]. Deep learning (DL) based classifiers have
also been used for classifying tweets into bullying and non-bullying tweets [7],
[18]_[22]. Supervised classifiers have low performance in case the class labels are
unchangeable and are not relevant to the new events [23]. Also, it may be suitable
only for a pre-determined collection of events, but it cannot successfully handle
tweets that change on the _y. Topic modeling approaches have long been utilized as
the medium to extract the vital topics from a set of data to form the patterns or classes
in the complete dataset. Although the concept is similar, the general unsupervised
topic models cannot be efficient for short texts, and hence specialized unsupervised
short text topic models were employed [24]. These models effectively identify the
trending topics from tweets and extract them for further processing. These models
help in leveraging the bidirectional processing to extract meaningful topics. However,
these unsupervised models require extensive training to obtain sufficient prior
knowledge, which is not adequate in all cases [25]. Considering these limitations, an
efficient tweet classification approach must be developed to bridge the gap between
the classifier and the topic model so that the adaptability is significantly proficient.
2
3
_ Propose DEA-RNN by combining the Elman type RNN and the improved DEA for
optimal classification of tweets;
_ A new Twitter dataset is collected based on cyber bullying keywords for evaluating
the performance of DEA-RNN and the existing methods; and
_ The efficiency of DEA-RNN in recognizing and classifying cyber bullying tweets is
assessed using Twitter datasets. The thorough experimental results reveal that DEA-
RNN outperforms other competing models in terms of recall, precision, accuracy, F1
score, and specificity.
The rest of this article is structured as the following: Recent related works
are reviewed and analyzed in Section II. The proposed DEA-RNN model is described
in Section III. Section IV discusses the experimental analysis, performance metrics,
and results analysis. The discussion is introduced
in Section V. Finally, Section VI offers the conclusion and possible future direction.
1. The Cyber-attacks are growing faster than usual rate, it became evident that
necessary steps should be taken in-order to get them under control. Among various
cyber-attacks, Phishing websites is one of the popular and commonly used attack
to steal users personal information and financial information by manipulating the
website URL and IP addresses.
2. The main focus in this project is to implement the better model for detecting
these phishing websites using ML algorithms.
This paper presents a phishing URL dataset using legitimate login websites to obtain
the URLs from such pages. Then, we evaluate machine and deep learning techniques
for recommending the method with higher accuracy. Next, we show how models
trained with legitimate homepages struggle to classify legitimate login URLs,
demonstrating our hypothesis about phishing detection and legitimate login URLs.
3
4
Additionally, we show how the accuracy decrease with the time on models trained
with datasets from 2016 and evaluated on data collected in 2020. Finally, we provide
an overview of current phishing encounters, explaining attacker tricks and
approaches.
there are limitations and risks that need to be addressed to ensure success. The three
primary constraints that project managers should be familiar with are time, scope, and
cost. These are frequently referred to as the triple constraints or the project
management triangle.
Assumptions are things that we believe to be true and which we therefore build into
the project plan. Constraints are things that we know to be true and which must be
accounted for in the plan so that we can work around them. And risks are factors that
Stretched resources.
Operational mishaps.
Low performance.
Lack of clarity.
Scope creep.
High costs.
Time crunch.
4
5
Objectives: A project objective describes the desired results of the work. We have
mentioned about the work we are trying to accomplish in this section.
limitations:
Chapter 2
Literature Survey: the purpose of a literature review is to gain an understanding of
the existing resources to a particular topic or area of study. We have referred to many
research papers relevant to our work in a better way.
Chapter 3
System Analysis: System Analysis is a document that describes about the existing
system and proposed system in the project. And also describes about advantages and
disadvantages in the project.
Chapter 4
System design: System design is a document that describes about the project
modules, Activity diagram, Use Case Diagram, Data Flow Diagram, and Sequence
Diagram detailed in the project.
Chapter 5
Implementation: Implementation is a document that describes about the detailed
concepts of the project. Also describes about the algorithm with their detailed steps.
And also, about the codes for implementation of the algorithm
5
6
Result: The result of project describes the output of the project and it gives
certain output for certain input
Chapter 6
Testing: Testing is a document that describes about the a. Methods of testing: This
contains the information about Unit testing, Validation testing, Functional testing,
Integration testing, User Acceptance testing. b. Test Cases: In Test Cases we contain
the detailed description about program
Chapter 7
Conclusion: Conclusion and Future Enhancement is a document that describes about
the brief summary of the project and undetermined events that will occur in that time.
6
7
2.LITERATURE SURVEY
2.1 INTRODUCTION
This study extracted features using Word2Vec and TF-IDF methods. Dalvi et al. [12]
[27] used SVM and Random Forests (RF) models with TF-IDF for feature extraction
for detecting cyberbullying in tweets. Although SVM in these models achieved high
performance, the model complexity increases when the class labels are increased. Al-
classifiers such as RF, Naïve Bayes (NB), and SVM based on various extracted
features from Twitter such as (tweet content, activity, network, and user). Huang et
al.
[29] suggested an approach for identifying CB from social media, which integrated
the social media features and textual content features. The features are ranked using
IG method. Well-known classifies such as NB, J48, and Bagging and Dagging are
utilized. The findings implied that social characteristics could aid in increasing the
(C4.5) classifier with a social network, personal and textual features to identify
MySpace. Balakrishnan et al. [31] utilized different ML algorithms such as RF, NB,
and J48 to detect cyberbullying events from tweets and classify tweets to different
cyberbullying classes such as aggressors, spammer, bully, and 25858 VOLUME 10,
normal. The study concluded that the emotional feature does not impact the detection
rate. Despite its efficiency, this model is limited to a small dataset with fewer class
labels. Alam et al. [32] proposed an ensemble-based classification approach using the
7
8
models utilized decision tree, LR, and Bagging ensemble model classifiers for the
feature extraction models. On analysis over the Twitter dataset, the Bagging ensemble
model provided the best precision but considered other parameters. Although, these
ensemble models reduced the training and execution time for classification, the major
limitation comes when utilized sarcasm tweets and multiple-meaning acronym terms.
Chia et al. [8] also utilized different ML and feature engineering-based approaches to
classify irony and sarcasm from cyber-bullying tweets. In this approach, many
classifiers and feature selection methods were tested; while this approach greatly
detects the sarcasm and irony terms among cyber-bullying tweets, the detection rate is
still very low [33]. Similarly, Rafiq et al. [17] utilized decision tree, AdaBoost, NB,
dataset. Authors collected the Vine media dataset and labeled it using Crowd-Sourced
information, and profile as the features. Nahar et al. [34] suggested a semi-supervised
learning method for detecting CB in social media in which training data samples are
augmented, and a fuzzy SVM method is applied. The augmented training approach
expands and extracts the training set from the unclassified streaming text
automatically. The learning is performed using a small limited training set given as an
initial input. The suggested method overcomes the dynamic and complex character of
streaming data. Xu et al. [35] provided many off-the-shelf methods, including LDA
and LSA-based modeling and Bagof-Words models for predicting bullying traces on
introduced by Cheng et al. [36] to detect cyberbullying from the Twitter dataset. PI-
Bully composes three elements: a global element that determines the characteristics
8
9
that all users have in common, a personalized element that captures the distinctive
features of each user, and a peer influence element capable of quantifying the various
influences of other users people. Deep learning (DL) based approaches for
Yuvaraj et al. [9] used Artificial Neural Network (ANN) and Deep Reinforcement
Learning (DRL) to classify cyberbullying tweets. However, this approach has higher
computational complexity. Chen et al. [37] used a text classification model based on
CNN and 2-D TF-IDF features to enhance the sentiment analysis task performance.
The experimental results showed that the CNN model obtained optimal results
compared to the baselines LR and SVM models. Agrawal [16] utilized LSTM with
Marginalized Denoising Autoencoder) was suggested by Zhao et al. [38] for detecting
Following that, The numerical representations that have been learned can be input
into SVM. Zhang [39] Suggested a new model which integrates the Gated Recurrent
unit Network GRU layers and CNN layers to detect hate speech. Al-Hassan and Al-
Dossari [19] utilized SVM as the baseline classifier and compared it against four DL
models, namely CNN + LTSM, LTSM, CNN + GRU, and GRU to detect
datasets. Natarajan Yuvaraj et al. [18] proposed a new classification model for CB
detection from Twitter data. It used deep decision-tree classification with multi-
feature based AI for tweet classification. The deep decision tree classifier has been
designed by integrating the hidden layers of deep neural networks with the decision
9
10
tree classifier. This approach also utilized three feature selection approaches: Chi-
data with such accuracy. Fang et al. [20] designed a classification model that
GRU) to detect cyberbullying in tweets. This model employed merit for learning the
underlying relationships between words using BI-GRU and used it together with a
Ilavarasan
tweets. This model utilizes Light Gradient Boosting Machine to classify the tweets
and RoBERTa to create word embedding. This approach overcomes the context-
independent limitations of traditional word embedding methods. Yet, this model has a
higher training time compared to the CNN models. Paul and Saha [40] proposed a
Iwendi et al. [21] introduced a model to detect cyberbullying based on Bi-LSTM and
RNN. This model showed that the RNN could achieve high performance, but still, the
Bi-LSTM has significantly high efficiency. In some cases, CNN also performs better.
Akhter et al. [41] performed many DL models such as LSTM, CLSTM, CNN, and
BLSTM, and other ML models to discover an abusive language from Urdu social
media text. Some other studies utilized CNN’s to enhance the cyberbullying detection
[42]–[46]. Tripathy et al. [47] proposed a fine-tuning approach for detecting CB based
on ALBER. Agarwal et al. [7] utilized RNN based on Under-Sampling and Class
Weighting. These modifications helped the RNN model to perform better than the
LSTM model
1
11
Purnamasari et al. [26] utilized the SVM and Information Gain(IG) based feature
selection method for detecting cyberbullying events in tweets. Muneer and
Fati [11] used various classifiers, namely AdaBoost(ADB), Light Gradient Boosting
Machine (LGBM), SVM, RF, Stochastic Gradient Descent (SGD), Logistic
Regression (LR), and MNB, and for cyberbullying events identification in tweets.
This study extracted features using Word2Vec and TF-IDF methods.
Dalvi et al. [12] [27] used SVM and Random Forests (RF) models with TF-IDF for
feature extraction for detecting cyberbullying in tweets. Although SVM in these
models achieved high performance, the model complexity increases when the class
labels are increased. Al-garadi et al. [28] investigated cyberbullying identification
using different ML classifiers such as RF, Naïve Bayes (NB), and SVM based on
various extracted features from Twitter such as (tweet content, activity, network, and
user). Huang et al. [29] suggested an approach for identifying CB from social media,
which integrated the social media features and textual content features.
The features are ranked using IG method. Well-known classifies such as NB, J48, and
Bagging and Dagging are utilized. The findings implied that social characteristics
could aid in increasing the accuracy of cyberbullying detection. Squicciarini et al.
[30] utilized a decision tree (C4.5) classifier with a social network,
not impact the detection rate. Despite its efficiency, this model is limited to a small
dataset with fewer class labels.
1
12
Alam et al. [32] proposed an ensemble-based classification approach using the single
and double ensemble-based voting model. These ensemble-based voting models
utilized decision tree, LR, and Bagging ensemble model classifiers for the
classification while utilizing mutual information bigrams and unigram TF-IDF as
feature extraction models. On analysis over the Twitter dataset, the Bagging ensemble
model provided the best precision but considered other parameters. Although, these
ensemble models reduced the training and execution time for classification, the major
limitation comes when utilized sarcasm tweets and multiple-meaning acronym terms.
Chia et al. [8] also utilized different ML and feature
1
13
_ Propose DEA-RNN by combining the Elman type RNN and the improved
DEA for optimal classification of tweets;
_ A new Twitter dataset is collected based on cyberbullying keywords for
evaluating the performance of DEA-RNN and the existing methods; and
_ The efficiency of DEA-RNN in recognizing and classifying cyberbullying
tweets is assessed using Twitter datasets. The thorough experimental results
reveal that DEA-RNN outperforms other competing models in terms of recall,
precision, accuracy, F1 score, and specificity
2.3.1 ADVANTAGES OF PROPOSED SYSTEM
The proposed system effectively identifies the trending topics from tweets and
extracts them for further processing. An effective models help in leveraging the
bidirectional processing to extract meaningful topics.
1
14
3.ANALYSIS
System Requirement Specifications is used for the programming contexture that are
basically being for the functionality of the system can do, and also for the work
behind the organization for describing and to understand the client’s needs. The
system. It generally provides for the capable clients needful conditions at a particular
complete that describes the behavior of system to be developed. Use case techniques
quality of standards
➢ RAM - 4 GB (min)
➢ Hard Disk - 20 GB
➢ Monitor - SVGA
1
15
6 Front-End : Python.
7 Back-End : Django-ORM
The Content diagram is an extension of UML notation. The purpose of the Content
diagram is to generate or represent a project structure (diagrams) and relations
between them. The Content table works as a table of contents for a project.
1
16
3.3.1 ALGORITHMS
Decision tree
Decision tree classifiers are used successfully in many diverse areas. Their most
important feature is the capability of capturing descriptive decision making
knowledge from the supplied data. Decision tree can be generated from training sets.
The procedure for such generation based on the set of objects (S), each belonging to
one of the classes C1, C2, …, Ck is as follows:\
Step 1. If all the objects in S belong to the same class, for example Ci, the decision
tree for S consists of a leaf labeled with this class
Step 2. Otherwise, let T be some test with possible outcomes O1, O2,…, On. Each
object in S has one outcome for T so the test partitions S into subsets S1, S2,… Sn
where each object in Si has outcome Oi for T. T becomes the root of the decision tree
and for each outcome Oi we build a subsidiary decision tree by invoking the same
procedure recursively on the set Si.
Gradient boosting
1
17
Whenever we have a new data to classify, we find its K-nearest neighbors from the
training data
Naïve Bayes
Yet, despite this, it appears robust and efficient. Its performance is comparable to
other supervised learning techniques. Various reasons have been advanced in the
literature. In this tutorial, we highlight an explanation based on the representation
bias. The naive bayes classifier is a linear classifier, as well as linear discriminant
analysis, logistic regression or linear SVM (support vector machine). The difference
lies on the method of estimating the parameters of the classifier (the learning bias)
Random Forest
1
18
outperform decision trees, but their accuracy is lower than gradient boosted trees.
However, data characteristics can affect their performance.
SVM
3.3.1 FLOWCHART
1
19
Service provider:
Remote user:
1
20
4. ANALYSIS
GOALS:
The Primary goals in the design of the UML are as follows:
1. Provide users a ready-to-use, expressive visual modeling Language so that
they can develop and exchange meaningful models.
2. Provide extendibility and specialization mechanisms to extend the core
concepts.
3. Be independent of particular programming languages and development
process.
4. Provide a formal basis for understanding the modeling language.
5. Encourage the growth of OO tools market.
6. Support higher level development concepts such as collaborations,
frameworks, patterns and components.
7. Integrate best practices
2
21
2
22
2
23
2
24
1. The DFD is also called as bubble chart. It is a simple graphical formalism that
can be used to represent a system in terms of input data to the system, various
processing carried out on this data, and the output data is generated by this
system.
2. The data flow diagram (DFD) is one of the most important modeling tools. It
is used to model the system components. These components are the system
process, the data used by the process, an external entity that interacts with the
system and the information flows in the system.
3. DFD shows how the information moves through the system and how it is
modified by a series of transformations. It is a graphical technique that depicts
information flow and the transformations that are applied as data moves from
input to output.
4. DFD is also known as bubble chart. A DFD may be used to represent a system
at any level of abstraction. DFD may be partitioned into levels that represent
increasing information flow and functional detail.
2
25
MODULES DESCRIPTION:
Service Provider
In this module, the Service Provider has to login by using valid user name and
password. After login successful he can do some operations such as
Login, Train & Test URL Data Sets, View URL Data Sets Trained and Tested
Accuracy in Bar Chart, View URL Data Sets Trained and Tested Accuracy Results,
View Prediction Of URL Type, View URL Type Ratio, Download Predicted Data
Sets, View URL Type Ratio Results, View All Remote Users.
In this module, the admin can view the list of users who all registered. In this, the
admin can view the user’s details such as, user name, email, address and admin
authorizes the users.
Remote User
In this module, there are n numbers of users are present. User should register before
doing any operations. Once user registers, their details will be stored to the database.
After registration successful, he has to login by using authorized user name and
password. Once Login is successful user will do some operations like REGISTER
AND LOGIN, PREDICT URL TYPE, VIEW YOUR PROFILE..
2
26
A. DATA COLLECTION
The input dataset is made up of tweets collected through Twitter API streaming with
the help of around 32 cyber-bullying keywords. Idiot, ni**er, LGBTQ (le***an, g*y,
bisexual, transgender, and queer), whore, pussy, faggot, shit, sucker, slut, donkey,
live, afraid, moron, poser, rape, fuck, fucking, ugly, bitch, ass, whale, etc. are some of
the keywords as recommended in psychology literature [30], [36], [57]. Whereas the
other keywords such as ban, kill, die, evil, hate, attack, terrorist, threat, racism, black,
Muslim, Islam, and Islamic were suggested in [39]
B. DATA ANNOTATION
This section mainly concentrates on annotating and labeling the selected tweets from
the original Twitter dataset. After selecting 10000 tweets randomly from the
collecting tweets, the selected tweets were labeled manually into two labels, either
‘‘0’’ non-cyber bullying or ‘‘1’’ cyberbullying, by a set of three human annotators
over a period of one and half months. In the labeling procedure, the human annotators
labeled the instances based on whether it was considered to involve cyberbullying and
also the guidelines described in detail in [57]. The making decision of the
cyberbullying instances depends on the following guidelines: character attacks,
insults, competence attacks, malediction, verbal abuse, teasing, name-calling,
mockery, threats, and physical appearance. Initially, each tweet was classified by two
annotators, and the level of agreement rate between the two annotators was 91%
approximately at this phase. Then, a third annotator was tasked with resolving the
discrepancies discovered during the initial annotation process. Finally, we obtained
the final dataset after resolving discrepancies and cleaned up the data, which
contained 10000 labeled tweets, among which 6,508 (0.65%) are non-cyberbullying,
and 3492 (0.35%) are cyberbullying tweets. By observing the number of
cyberbullying and non-cyberbullying tweets, the labeled Twitter dataset is
imbalanced. The number of tweets in classes is greatly variable
2
27
2
28
DATASETS
To test the model robustness against URLs collected in different periods, we used the five
phishing datasets shown in Table 3. These datasets are grouped into two different categories
depending on their recollection strategy: (i) category A: PWD2016, 1M-PD and PIU-60K
collected legitimate samples by inspecting the top-visited domains and (ii) category B:
Ebbu2017 and PLU-60K visited those websites and performed further actions: in the case of
Ebbu2017, its authors retrieved the inner URLs and, in the case of PLU60K, we looked for
the login form page. Therefore most of the URLs include a path. Table 4 shows the
distribution of sample structure within the datasets.
EXPERIMENTAL SETTINGS
Precision = TP TP + FP (1)
Recall = TP TP + FN (2)
Accuracy = TP + TN TP + TN + FN + FP (3)
TP denotes the true positives, i.e., how many phishing websites were correctly
classified. FP refers to the false positives and represents the number of legitimate
samples wrongly classified as phishing. TN (i.e., the true negatives) denotes the
number of legitimate samples correctly classified. Finally, FN represents the false
2
29
--
-- Database: `dearnn`
--
CREATE DATABASE IF NOT EXISTS `dearnn` DEFAULT CHARACTER SET
latin1 COLLATE latin1_swedish_ci;
USE `dearnn`;
--
-- Table structure for table `auth_group`
--
2
30
--
-- Table structure for table `auth_group_permissions`
--
--
-- Table structure for table `auth_permission`
--
--
-- Dumping data for table `auth_permission`
--
3
31
--
-- Table structure for table `auth_user`
--
--
-- Table structure for table `auth_user_groups`
--
3
32
--
-- Table structure for table `auth_user_user_permissions`
--
--
-- Table structure for table `django_admin_log`
--
--
-- Table structure for table `django_content_type`
--
3
33
--
-- Dumping data for table `django_content_type`
--
--
-- Table structure for table `django_migrations`
--
--
-- Dumping data for table `django_migrations`
--
3
34
--
-- Table structure for table `django_session`
--
--
-- Dumping data for table `django_session`
--
3
35
('0jpcgnd1gmwbp3e8tw54e6nxjylsogyo',
'YmM4NjE0MDQ2MzBmYWIxNzIzNTkxZjBiN2I5M2MxMzQyYTE0YmMxODp7InVzZ
XJpZCI6Mn0=', '2020-02-21 08:52:28.687500'),
('1avxwyhltuaclq2zfq40bjmwzxgup6hp',
'eyJ1c2VyaWQiOjJ9:1ml5JO:9_XIwCjkeG7Vu6-k169B1gbde6aRX-
mqiwhgCKNqrRA', '2021-11-25 08:19:58.276708'),
('49qo7iki5uxczhyymi8ka7dnh6a2wva5',
'MmE4N2EzZmM3NTI1ODc3MjUxYjUxNWM3OWM4ZGExNWViMzRkN2MzYTp7Im5hb
WUiOjF9', '2019-05-08 09:19:45.486328'),
('4df7s82pddaszour6twx23d86058ppjq',
'ZmNkODA5MmI1ZGQ0Yjk5MmZlNzEyNTcwNTcxNjk2ZWYxZTE3NThkMjp7InVzZ
XJpZCI6NX0=', '2020-11-23 11:49:21.396484'),
('4io28d085qjfib7a5s2qbhc8qp4wfiva',
'eyJ1c2VyaWQiOjE2fQ:1mAtmi:oIUbcN3WzJiaWnxMBZ6eIGMTo8NS2y701Jl
pwqvzBUk', '2021-08-17 12:44:40.453750'),
('4x6b78w9rfcn34v650kd2j7oij6atr8p',
'Zjk0Y2RlYjc4OTJhNWMyZjQyNmM4ZGRhYTVjNmVlNDFhZGE4ZmU3NTp7InVzZ
XJpZCI6Nn0=', '2019-12-27 12:07:42.082031'),
('6h7f2lif2b5xzckoqy0dzpuf2ssln6u6',
'eyJ1c2VyaWQiOjJ9:1oOxNy:rn921d2FdMCS42DJKaBJnAlT20byTRiY2P4X7
Wm76Qc', '2022-09-02 08:29:46.357161'),
('7ai80sml3hwq0jsvdfl1vm5m13x8d8k9',
'eyJ1c2VyaWQiOjF9:1mt5vR:3pScFzSqVCPEshZhoFMwZb-
rfCX09_pAUFEuMV8fNKA', '2021-12-17 10:36:21.166280'),
('7ixdamflp4fqyjecn17bd7xfbsi7eowq',
'eyJ1c2VyaWQiOjEwfQ:1mBzQr:5DfHs08xtygiklJxfW3kZFCrxnrA_igxR5g
bDcKt2e8', '2021-08-20 12:58:37.947486'),
('7ph664obz14m207786d3oubrzgjnisom',
'eyJ1c2VyaWQiOjN9:1ml5U2:RJ7eMbREY4fk71sBmCItxM6E3kDDc-R-clIp-
QGiiuI', '2021-11-25 08:30:58.769500'),
('9vom7mmn5muyoiy8nytc9mxown1q1g2b',
'eyJ1c2VyaWQiOjE5fQ:1mDRHr:QCoJ_gmMMx_cxknA5j_5NlcTLnENHFouosR
uxYZlYbI', '2021-08-24 12:55:19.038051'),
('au3tqhab9csr4r2g5p8wxgktebzxone0',
'eyJ1c2VyaWQiOjExfQ:1mexd7:BmPTZn93Z2602ApV03LTh7BmDypyoNMN2YR
KctHrGF8', '2021-11-08 10:55:01.820399'),
('b9cu6cjsfqfm5mame5dy1ikpiiy7yn3w',
'OTk3NTk2YTE0NjM5MWQ0OGQ0MjY3NzBjNzdhOTc0ZWJhM2ZkMzdkMjp7InVzZ
XJpZCI6MX0=', '2019-05-09 11:00:08.480453'),
('bhfid9lacfwlfi5yu3rgdg1uo5fp2bq8',
'eyJ1c2VyaWQiOjE4fQ:1mBH4F:2wUorkPET_MGY07bWd-Zp-
9HZUsjS3bGCHCu1j6BN-s', '2021-08-18 13:36:19.192796'),
('ct13q5fpn94zvnij8ekixwzcky2imc5e',
'YWUzM2IzMWJiYmQ3YmY2YzlkMGFlNTM1YmU5ZGM4YjQ0MmY1YTc0NTp7InVzZ
XJpZCI6NH0=', '2019-05-14 11:44:10.978515'),
('e07j4duysh402dedtomm8icctvs9ljgy',
'MmE4N2EzZmM3NTI1ODc3MjUxYjUxNWM3OWM4ZGExNWViMzRkN2MzYTp7Im5hb
WUiOjF9', '2019-05-09 06:08:12.306625'),
('fq0czwxzas1e5bjz5cx5pr6ytm8uhejy',
'eyJ1c2VyaWQiOjExfQ:1mfKoD:eQbeRUgZ8NFqCleEdS6fE0NAoRs3zn6_B82
CZb4YtiQ', '2021-11-09 11:40:01.472965'),
('gq3vdjxoy34hxkorw2d8nztdusdknegl',
'eyJ1c2VyaWQiOjF9:1mpUvs:Lry5yh51WzsY8judWu-
ApNb05fEC5oHytMQ9bZSqbGU', '2021-12-07 12:29:56.098464'),
3
36
('gsqk7v1ei7yhuvcbxw6r8vxhgbuzz7zx',
'eyJ1c2VyaWQiOjJ9:1mkkhL:_T8wyqi-MJi-K7_a-0EPz-
h6HRouyRfpeAhHqTS5N04', '2021-11-24 10:19:19.270816'),
('h2up0dvopjvwswxnvprj7id9lgrivhus',
'eyJ1c2VyaWQiOjIzfQ:1mfM3f:FAuAUdY-
ly6qun6t571yt1pYKVGhXfjbjhiruld5rNs', '2021-11-09
13:00:03.703809'),
('hbv74sg6w6e4wp89vq807vw0xhkh5s1h',
'MzU0ZWYzNTQ3MjM4MWZlOTVjM2M1MWQ4MmE5ODE0OTlkNDRkNDkwMDp7InVzZ
XJpZCI6MX0=', '2020-01-10 07:40:38.067382'),
('hhtt48je70l9nzw6dee4ocuxxm9blqej',
'NGRhY2JkNmQ4ZTM4OTU0Y2UzMzFlZmZmOTgzYmE0MWVkOThiNjc2NTp7Im5hb
WUiOjEsInVzZXJpZCI6MX0=', '2019-05-09 10:12:38.380843'),
('hsb5814on7cph0wvy0yls67ca94ngcq3',
'eyJ1c2VyaWQiOjE5fQ:1mBzgz:cug3sAkQKH-bQBkB9O5l0UsDJL-
37eDV8mR9Qau3elA', '2021-08-20 13:15:17.195464'),
('i530ldontosd9c37rlmr7i190cc8j54c',
'eyJ1c2VyaWQiOjExfQ:1mfGFZ:PSpZPmdPYnGzwCScqY4tYBkDj8BMVATwweZ
jjxmG5dk', '2021-11-09 06:47:57.629900'),
('i77fui9jgj9yk7ncx7u4ph5d6kg0nl6c',
'eyJ1c2VyaWQiOjE5fQ:1mDPJC:kqt800XGsVGRjHS3TmeLFrJbrpIK4-
GbH4ZirwIc7S4', '2021-08-24 10:48:34.860638'),
('ic3hqykgws5iy6fz5ns6h6f921jbjzmt',
'eyJ1c2VyaWQiOjExfQ:1kywHL:I_tahJ0VJb7myAbMbXpWZu9XrSaAMmduNxG
d2x5gtmY', '2021-01-25 12:26:35.043761'),
('iz6wcyx97x1w6mpfc51g1tj72z2xghfn',
'eyJ1c2VyaWQiOjl9:1kwlIp:YKOKMwJARe6w057AKTGY1-
GCuRcZAeAbJ0bdQao23wY', '2021-01-19 12:19:07.663490'),
('jgcbya9z2s6b6mmldfv28lm18imc73m8',
'eyJ1c2VyaWQiOjIxfQ:1mDnYP:GTRQ2I-UYLdsCCyA0-
WsFSAVBNno1wLo6lk4M8JS0OU', '2021-08-25 12:41:53.312085'),
('jpkxxiej4bdjin5tpdjm0xqhdooexz9o',
'eyJ1c2VyaWQiOjExfQ:1mBEdk:YOk6fHHfBMmtt5ZvSyzgy13Az8JS59iXbU4
LO1Ps1RI', '2021-08-18 11:00:48.423036'),
('k7dyn4irgrj5wb4jucb4po527iw724dp',
'eyJ1c2VyaWQiOjEzfQ:1l0JrY:2_TJ4L_XoHdOW51Zdp0MOdyBEZEzntk5pdX
ZFDmX9x4', '2021-01-29 07:49:40.202848'),
('ktjsa2dwmkzggkc8htfro0m1zf2kt78d',
'eyJ1c2VyaWQiOjI0fQ:1mfiev:rUgpc2VOr-
8MQnmWwZsSVM_IqXQAA3Bacheqmp_LQ2o', '2021-11-10
13:08:01.503971'),
('kxla8qbe1hs1tb4a04ewpmdmi5fx1evc',
'eyJ1c2VyaWQiOjJ9:1mpVxM:dUv2r33P7CFNLdnAjy8jCk-
OJOlisN07sHd5Rf7zCLg', '2021-12-07 13:35:32.342260'),
('o7x1vhluuypdfmgv7fmv6nohgfn5ub55',
'NzMyZjlhNzFhZjk2ZGUzZmFiMmIzYjMwNTJkYTg5MDUzNmNlMDk4Mjp7InVzZ
XJpZCI6MTZ9', '2020-01-02 12:51:55.659179'),
('oc4pzt7ijx1rzj09m2ve6b6y5uwc6wt5',
'eyJ1c2VyaWQiOjF9:1mkk7W:6GDuPOPoLHMwOYLys3lXIgfM79Fq36HHW31-
utzbeyk', '2021-11-24 09:42:18.649798'),
('owqt9fqa6pkheboduh6f4k5p4lkwj0yc',
'eyJ1c2VyaWQiOjExfQ:1mfiXk:Wzn12pygxu_2Z1TzCSC4bKDawuXj_i7_BFL
hJjKx-10', '2021-11-10 13:00:36.684741'),
('psdjoq42u7lfqwfodftic5x6z9ij34nk',
'eyJ1c2VyaWQiOjExfQ:1mAXDq:a8YYY1YJU3jPv03qo9-
VcrjRHnDWRSqGseiR93n0GVM', '2021-08-16 12:39:10.518259'),
3
37
('q6hp9a2l9dbrclvox0o02x1aamx1ukj7',
'eyJ1c2VyaWQiOjIyfQ:1mfGz3:wkq7ZgyB738cK3Jugrc0viqb3eI1C0gWhyy
pHF_DE-A', '2021-11-09 07:34:57.860407'),
('qnaolidvfx6bu9ra3uyqvkgva7bv92f1',
'OTk3NTk2YTE0NjM5MWQ0OGQ0MjY3NzBjNzdhOTc0ZWJhM2ZkMzdkMjp7InVzZ
XJpZCI6MX0=', '2019-05-14 05:34:50.069335'),
('r0lswf36qh5gt0i2afmqqeu550j9vohq',
'eyJ1c2VyaWQiOjF9:1oOvxv:-
o0Ngd4h_vqB4DUhIuathTt7_CkrfaHFDCAmdaRTbwU', '2022-09-02
06:58:47.164716'),
('r9qk0kd407g591hugz99fhps8zofh69s',
'eyJ1c2VyaWQiOjE4fQ:1mBxnG:qkd9MTM_FhhghUpV90qngEkwkoSKYdLbfwR
KBLhK7Qg', '2021-08-20 11:13:38.356475'),
('rfq3uvadj7qsqrz7qlcyie9wscsqz1nr',
'eyJ1c2VyaWQiOjI1fQ:1mk2vF:mTIne3EU3rECWccrUfyPmy7XxIwdAzxhig4
S5oOngho', '2021-11-22 11:34:45.897995'),
('rn48bwukkb2yv60kvkacr8nc0njr5xky',
'eyJ1c2VyaWQiOjExfQ:1mfggv:xsbmrrGzxtrEFgspA0Wp7oWTp9qDl0shDSl
GG8fHJo4', '2021-11-10 11:01:57.519295'),
('s7ui2zx2cslubpch6dm7iaxlz2wlsdgg',
'eyJ1c2VyaWQiOjExfQ:1mk1Xz:fpol-krFazPxkK0b4gEoocpOXHcd-
eaYoxg26CU420Q', '2021-11-22 10:06:39.923928'),
('sdcvtwp7s5yj8q1lb0mdvlg8nj5wujqo',
'eyJ1c2VyaWQiOjEyfQ:1kzJ3p:0g6nRuJv3TXWVpANqNgbJcrUv96ZU5UQwv3
bgqBbL1I', '2021-01-26 12:46:09.538596'),
('tejgl09oettnyva23kqdbns5nfz5g8ug',
'OTk3NTk2YTE0NjM5MWQ0OGQ0MjY3NzBjNzdhOTc0ZWJhM2ZkMzdkMjp7InVzZ
XJpZCI6MX0=', '2019-05-09 11:19:24.387679'),
('tx26u0gjaebi1m6miqvms83dw452rw7c',
'eyJ1c2VyaWQiOjExfQ:1mg01H:i0OHhHK7t2WdfySWwMXXs92TaT7Vn1UQf2i
0eZBO70s', '2021-11-11 07:40:15.053714'),
('u5icgvq3qt5nthdlv99go3r804ccghbo',
'MmE4N2EzZmM3NTI1ODc3MjUxYjUxNWM3OWM4ZGExNWViMzRkN2MzYTp7Im5hb
WUiOjF9', '2019-05-09 06:00:13.573226'),
('ws2o4cq1jbqepe0e9s9v7n4erxatq9ic',
'eyJ1c2VyaWQiOjE1fQ:1l2CgI:SmlpAnZzplZhPTFJ_rkEJstnZRl2CYWyTca
h7PHPv-M', '2021-02-03 12:33:50.352453'),
('x5tpr0r7bu57jws6fpdwhi841252o6o2',
'eyJ1c2VyaWQiOjJ9:1oFxxT:uvC1LIN0lwgSyiMgEmEl1yH9W_bC1GTe5QDFs
BdOQqM', '2022-08-08 13:17:15.650911'),
('xc6si4gdotxq06mslnwwjtewdhzuyh44',
'eyJ1c2VyaWQiOjR9:1mnfuL:VMKdQ2gcb4wQppelEf-
wAdfTSSkA6nWPCeQODqD3NOM', '2021-12-02 11:48:49.418566'),
('zega5sz46ivu1tb1o1mtmg3v2ysxog1w',
'eyJ1c2VyaWQiOjh9:1kuVm4:L7RizVvw4EC0IyYCYAIhGjC8lvZol_Z1abqVw
dkdKkY', '2021-01-13 07:20:00.767751'),
('zq2idu1e3gkyelhzpa24k5jixqhj74da',
'eyJ1c2VyaWQiOjJ9:1mne5B:sOSzXh5WGI5se1746XNpUgE1_UZZw07Xzb03K
IcZXfY', '2021-12-02 09:51:53.585889'),
('zqfzqdolomzyjb8lckwouaztv90i58gi',
'eyJ1c2VyaWQiOjJ9:1mt6QE:e6TQ4CyC7rlTpZA0HyGK2zrDnv_EbGQrGGZUz
gCcnCM', '2021-12-17 11:08:10.160908');
--
3
38
--
-- Dumping data for table `remote_user_clientregister_model`
--
--
-- Table structure for table
`remote_user_cyberbullying_detection_type`
--
--
-- Dumping data for table
`remote_user_cyberbullying_detection_type`
--
3
39
--
-- Table structure for table `remote_user_detection_accuracy`
--
--
-- Dumping data for table `remote_user_detection_accuracy`
--
--
-- Table structure for table `remote_user_detection_ratio`
--
3
40
--
-- Dumping data for table `remote_user_detection_ratio`
--
4
41
Register page
Register form
4
42
User Login
Preduction page
4
43
Line chart
Pie chart
4
44
Ratio
4
45
Bar graph
Information Stored
4
46
In this section, the evaluation of DEA-RNN is performed over datasets crawled from
Twitter utilizing these metrics: recall, precision, F-measure, accuracy, and specificity.
The input dataset and the data annotation are described in sections III-A and III-B.
Two baseline cyberbullying models based on deep learning, namely Bi-LSTM [21],
RNN [21], and three baseline cyberbullying models based on machine learning
models, namely, SVM [26], Multinomial Naive Bayes (MNB) [11], and R [11] are
used for the comparison with the proposed DEA-RNN model. These models have
been selected from state-of-the-art cyberbullying detection in social media.
1. A. EVALUATION METRICS
This sub-section briefly highlights the evaluation metrics utilized in this study to
evaluate the efficiency of DEA-RNN. The evaluation process is performed based on
the following metrics: accuracy, recall, precision, F-measure, specificity and
computing training time. However, each method is run (N = 20) times for all
experiments to obtain an average of obtained results for each evaluation metric. These
performance metrics are described in Table 2.
4
47
2. EXPERIMENTAL RESULTS
4
48
A. AVERAGE ACCURACY
4
49
The Training time of the proposed model was compared with baseline models.
Where, scenario 2 has been taken into consideration for computing the training time.
It can be observed that the proposed DEA-RNN model has less training time
compared to other deep learning Bi-LSTM, RNN baseline models. The training time
of Bi-LSTM is more than the proposed model, RNN as well as the machine learning
models, but the achievement of Bi-LSTM is better than the other baseline models and
less the proposed model.
4
50
6.TESTING
6.1 INTRODUCTION
Software system meets its requirements and user expectations and does not fail in an
unacceptable manner. There are various types of test. Each test type addresses a
specific testing requirement.
TYPES OF TESTS
The following are the Testing Methodologies:
o Unit Testing.
o Integration Testing.
o User Acceptance Testing.
o Output Testing.
o Validation Testing.
Unit Testing
Unit testing focuses verification effort on the smallest unit of Software design that is
the module. Unit testing exercises specific paths in a module’s control structure to
ensure complete coverage and maximum error detection. This test focuses on each
module individually, ensuring that it functions properly as a unit. Hence, the naming
is Unit Testing.
During this testing, each module is tested individually and the module
interfaces are verified for the consistency with design specification. All important
processing path are tested for the expected results. All error handling paths are also
tested.
Integration Testing
5
51
Integration testing addresses the issues associated with the dual problems of
verification and program construction. After the software has been integrated a set of
high order tests are conducted. The main objective in this testing process is to take
unit tested modules and builds a program structure that has been dictated by design.
3. Bottom-up Integration
This method begins the construction and testing with the modules at the lowest
level in the program structure. Since the modules are integrated from the bottom
up, processing required for modules subordinate to a given level is always
available and the need for stubs is eliminated. The bottom up integration strategy
may be implemented with the following steps:
The low-level modules are combined into clusters into clusters that
perform a specific Software sub-function.
A driver (i.e.) the control program for testing is written to coordinate
test case input and output.
The cluster is tested.
Drivers are removed and clusters are combined moving upward in the
program structure
The bottom up approaches tests each module individually and then each module is
module is integrated with a main module and tested for functionality.
5
52
User Acceptance of a system is the key factor for the success of any system. The
system under consideration is tested for user acceptance by constantly keeping in
touch with the prospective system users at the time of developing and making changes
wherever required. The system developed provides a friendly user interface that can
easily be understood even by a person who is new to the system.
Output Testing
After performing the validation testing, the next step is output testing of the proposed
system, since no system could be useful if it does not produce the required output in
the specified format. Asking the users about the format required by them tests the
outputs generated or displayed by the system under consideration. Hence the output
format is considered in 2 ways – one is on screen and another in printed format.
Validation Checking
Text Field:
The text field can contain only the number of characters lesser than or equal to
its size. The text fields are alphanumeric in some tables and alphabetic in other
tables. Incorrect entry always flashes and error message.
Numeric Field:
The numeric field can contain only numbers from 0 to 9. An entry of any character
flashes an error messages. The individual modules are checked for accuracy and what
it has to perform. Each module is subjected to test run along with sample data. The
individually tested modules are integrated into a single system. Testing involves
executing the real data information is used in the program the existence of any
program defect is inferred from the output. The testing should be planned so that all
the requirements are individually tested.
A successful test is one that gives out the defects for the inappropriate data and
produces and output revealing the errors in the system.
5
53
Taking various kinds of test data does the above testing. Preparation of test data
plays a vital role in the system testing. After preparing the test data the system under
study is tested using that test data. While testing the system by using test data errors
are again uncovered and corrected by using above testing steps and corrections are
also noted for future use.
Live test data are those that are actually extracted from organization files. After
a system is partially constructed, programmers or analysts often ask users to key in a
set of data from their normal activities. Then, the systems person uses this data as a
way to partially test the system. In other instances, programmers or analysts extract a
set of live data from the files and have them entered themselves.
Artificial test data are created solely for test purposes, since they can be
generated to test all combinations of formats and values. In other words, the artificial
data, which can quickly be prepared by a data generating utility program in the
information systems department, make possible the testing of all login and control
paths through the program.
The most effective test programs use artificial test data generated by persons
other than those who wrote the programs. Often, an independent team of testers
formulates a testing plan, using the systems specifications.
The package “Virtual Private Network” has satisfied all the requirements
specified as per software requirement specification and was accepted.
5
54
Whenever a new system is developed, user training is required to educate them about
the working of the system so that it can be put to efficient use by those for
whom the system has been primarily designed. For this purpose the normal
working of the project was demonstrated to the prospective users. Its working
is easily understandable and since the expected users are people who have
good knowledge of computers, the use of this system is very easy.
6.3 MAINTAINENCE
This covers a wide range of activities including correcting code and design
errors. To reduce the need for maintenance in the long run, we have more accurately
defined the user’s requirements during the process of system development. Depending
on the requirements, this system has been developed to satisfy the needs to the largest
possible extent. With development in technology, it may be possible to add many
more features based on the requirements in future. The coding and designing is simple
and easy to understand which will make maintenance easier.
TESTING STRATEGY:
A strategy for system testing integrates system test cases and design techniques
into a well planned series of steps that results in the successful construction of
software. The testing strategy must co-operate test planning, test case design, test
execution, and the resultant data collection and evaluation .A strategy for software
testing must accommodate low-level tests that are necessary to verify that a small
source code segment has been correctly implemented as well as high level tests
that validate major system functions against user requirements.
5
55
Software once validated must be combined with other system elements (e.g.
Hardware, people, database). System testing verifies that all the elements are proper
and that overall system function performance is
achieved. It also tests to find discrepancies between the system and its original
objective, current specifications and system documentation.
UNIT TESTING:
In unit testing different are modules are tested against the specifications
produced during the design for the modules. Unit testing is essential for verification of
the code produced during the coding phase, and hence the goals to test the internal
logic of the modules. Using the detailed design description as a guide, important
Conrail paths are tested to uncover errors within the boundary of the modules. This
testing is carried out during the programming stage itself. In this type of testing step,
each module was found to be working satisfactorily as regards to the expected output
from the module.
Integration testing
5
56
Functional test
Functional tests provide systematic demonstrations that functions tested are
available as specified by the business and technical requirements, system
documentation, and user manuals.
invoked.
System Test
System testing ensures that the entire integrated software system meets
requirements. It tests a configuration to ensure known and predictable results. An
example of system testing is the configuration oriented system integration test.
System testing is based on process descriptions and flows, emphasizing pre-driven
process links and integration points.
5
knowledge of the inner workings, structure and language of the software, or at least its
5
57
purpose. It is purpose. It is used to test areas that cannot be reached from a black box
level.
5
58
7.VALIDATIOn
7.1 INTRODUCTION
Text Field:
The text field can contain only the number of characters lesser than or equal to
its size. The text fields are alphanumeric in some tables and alphabetic in other
tables. Incorrect entry always flashes and error message.
Numeric Field:
The numeric field can contain only numbers from 0 to 9. An entry of any
character flashes an error messages. The individual modules are checked for accuracy
and what it has to perform. Each module is subjected to test run along with sample
data. The individually tested modules are integrated into a single system. Testing
involves executing the real data information is used in the program the existence of
any program defect is inferred from the output. The testing should be planned so
that all the requirements are individually tested.
A successful test is one that gives out the defects for the inappropriate data
and produces and output revealing the errors in the system.
Taking various kinds of test data does the above testing. Preparation of test
data plays a vital role in the system testing. After preparing the test data the system
under study is tested using that test data. While testing the system by using test data
errors are again uncovered and corrected by using above testing steps and corrections
are also noted for future use.
Live test data are those that are actually extracted from organization files.
After a system is partially constructed, programmers or analysts often ask users to key
in a set of data from their normal activities. Then, the systems person uses this data as
5
59
a way to partially test the system. In other instances, programmers or analysts extract
a set of live data from the files and have them entered themselves.
Artificial test data are created solely for test purposes, since they can be generated to
test all combinations of formats and values. In other words, the artificial data, which
can quickly be prepared by a data generating utility program in the information
systems department, make possible the testing of all login and control paths through
the program.
The most effective test programs use artificial test data generated by persons other
than those who wrote the programs. Often, an independent team of testers formulates
a testing plan, using the systems specifications.
The package “Virtual Private Network” has satisfied all the requirements specified as
per software requirement specification and was accepted.
USER TRAINING
Whenever a new system is developed, user training is required to educate them about
the working of the system so that it can be put to efficient use by those for whom the
system has been primarily designed. For this purpose the normal working of the
project was demonstrated to the prospective users. Its working is easily
understandable and since the expected users are people who have good knowledge of
computers, the use of this system is very easy.
MAINTAINENCE
This covers a wide range of activities including correcting code and design errors. To
reduce the need for maintenance in the long run, we have more accurately defined the
6
60
TESTING STRATEGY :
A strategy for system testing integrates system test cases and design techniques into a
well planned series of steps that results in the successful construction of software.
The testing strategy must co-operate test planning, test case design, test execution,
and the resultant data collection and evaluation .A strategy for software testing must
accommodate low-level tests that are necessary to verify that a small source code
segment has been correctly implemented as well as high level tests that validate
major system functions against user requirements.
Software testing is a critical element of software quality assurance and represents the
ultimate review of specification design and coding. Testing represents an interesting
anomaly for the software. Thus, a series of testing are performed for the proposed
system before the system is ready for user acceptance testing.
SYSTEM TESTING:
Software once validated must be combined with other system elements (e.g.
Hardware, people, database). System testing verifies that all the elements are proper
and that overall system function performance is achieved. It also tests to find
discrepancies between the system and its original objective, current specifications
and system documentation.
UNIT TESTING:
In unit testing different are modules are tested against the specifications produced
during the design for the modules. Unit testing is essential for verification of the code
produced during the coding phase, and hence the goals to test the internal logic of
the modules. Using the detailed design description as a guide, important Conrail
paths are tested to uncover errors within the boundary of the modules. This testing is
carried out during the programming stage itself. In this type of testing step, each
6
61
module was found to be working satisfactorily as regards to the expected output from
the module.
7.2 CONCLUSION
6
62
8.CONCLUSION
6
63
REFERENCE
6
64