Next Article in Journal
Terminal Sliding Mode Control with a Novel Reaching Law and Sliding Mode Disturbance Observer for Inertial Stabilization Imaging Sensor
Next Article in Special Issue
Blockchain-Based Data Sharing and Trading Model for the Connected Car
Previous Article in Journal
A Reliable Fault Diagnosis Method for a Gearbox System with Varying Rotational Speeds
Previous Article in Special Issue
Toward Efficient Image Recognition in Sensor-Based IoT: A Weight Initialization Optimizing Method for CNN Based on RGB Influence Proportion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Recurrent GANs Password Cracker For IoT Password Security Enhancement †

Graduate School of Information Security, Korea University, Seoul 02841, Korea
*
Author to whom correspondence should be addressed.
This paper is an extension version of conference paper published in: Nam, S.; Jeon, S.; Moon, J. A New Password Cracking Model with Generative Adversarial Networks. In Proceedings of the International Workshop on Information Security Applications, Jeju Island, Korea, 21–24 August 2019; pp. 247–258.
Sensors 2020, 20(11), 3106; https://doi.org/10.3390/s20113106
Submission received: 26 April 2020 / Revised: 25 May 2020 / Accepted: 28 May 2020 / Published: 31 May 2020
(This article belongs to the Special Issue Selected papers from WISA 2019)

Abstract

:
Text-based passwords are a fundamental and popular means of authentication. Password authentication can be simply implemented because it does not require any equipment, unlike biometric authentication, and it relies only on the users’ memory. This reliance on memory is a weakness of passwords, and people therefore usually use easy-to-remember passwords, such as “iloveyou1234”. However, these sample passwords are not difficult to crack. The default passwords of IoT also are text-based passwords and are easy to crack. This weakness enables free password cracking tools such as Hashcat and JtR to execute millions of cracking attempts per second. Finally, this weakness creates a security hole in networks by giving hackers access to an IoT device easily. Research has been conducted to better exploit weak passwords to improve password-cracking performance. The Markov model and probabilistic context-free-grammar (PCFG) are representative research results, and PassGAN, which uses generative adversarial networks (GANs), was recently introduced. These advanced password cracking techniques contribute to the development of better password strength checkers. We studied some methods of improving the performance of PassGAN, and developed two approaches for better password cracking: the first was changing the convolutional neural network (CNN)-based improved Wasserstein GAN (IWGAN) cost function to an RNN-based cost function; the second was employing the dual-discriminator GAN structure. In the password cracking performance experiments, our models showed 10–15% better performance than PassGAN. Through additional performance experiments with PCFG, we identified the cracking performance advantages of PassGAN and our models over PCFG. Finally, we prove that our models enhanced password strength estimation through a comparison with zxcvbn.

1. Introduction

As the computing power of IoT devices such as drones and smartwatches have improved, they have been utilized not only for entertainment purposes but also in various fields, such as military services and delivery services. Such enhanced computing ability allows the devices to operate on a modern operating system (OS) such as Linux that contains various applications, including a file transfer protocol (FTP), which is susceptible to massive network-based attacks such as a distributed denial of service (DDoS) attack [1]. Thus, proper security mechanisms for IoT devices are required to avoid the abuse of the devices. Among many security mechanisms, password-based authentication, especially text-based passwords, is a popular choice. The text-based password is a basic and fundamental authentication method, and it often plays a crucial role in system security. Passwords are both easy to understand and use and easy to implement [2]. Password authentication depends only on people’s ability to remember their passwords but does not need any additional equipment. This characteristic makes passwords the most popular authentication method, including for IoT devices, and passwords are often used as the first authentication method in multi-factor authentication. The fact that a password relies on a user’s ability to remember can also be a weakness. The dependence on memory induces people to use passwords in easy-to-remember word patterns [3,4,5]—for example, “iloveyou12”, and these are easily cracked by rule-governed dictionary-based password attacks. Using easy-to-remember passwords or known default passwords makes the device vulnerable to a security breach, as well as more sophisticated attacks involving IoT devices such as DDoS. To improve on the weakness of these passwords, password strength estimation studies have been conducted to recommend the usage of strong passwords, and organizations such as the National Institute of Standards and Technology (NIST) recommend the use of strong passwords at the policy level [6]. As a strong password usage policy, major sites such as gmail.com and microsoft.com enforce the use of a combination of uppercase and lowercase letters, digits, and special characters in passwords.
To crack high-complexity passwords, a password cracking method is required beyond a rule-based dictionary attack, and research into the development of such a technique is needed. The following is a detailed description of the cases in which password cracking technology needs to be developed. The first case is that people often forget their passwords; in particular, if a user sets a complicated password (which is unlike a pattern used in the past), the user will forget it more quickly and will need an effective password cracking method. Second, national agencies may need to crack passwords to obtain encrypted criminal evidence or intelligence information. Finally, valid password cracking methods are needed to secure passwords properly. Password cracking techniques can be used to estimate the strength of passwords practically. The zxcvbn method [7], which is used in DropBox, uses simple password crackers to estimate password strength. In this paper, we focus on improving password cracking performance rather than evaluating password strength.
Password cracking methods can be broadly divided into two types: the first type is the brute-force attack—i.e., an exhaustive attack—while the dictionary-based attack is the second type, which also has a hybrid attack variation [8]. The hybrid attack applies transformation or adds masking to the attack dictionary. Leaked passwords or an English dictionary are usually used as an attack dictionary. In some cases, a hybrid attack is distinguished from dictionary-based attacks, but in this study, we consider hybrid attacks to be dictionary-based attacks in a broad sense from the perspective of an expanded dictionary-based attack. The brute-force attack (or exhaustive attack) is a method of cracking a password by generating all the character combinations within a given password length range. The number of hash calculations per second executed in a brute-force attack has increased due to the development of GPU technology and the development of heterogeneous computing technology. In addition, if the time required for hash calculations such as MD5 or NTLM is short, the number of operations per second is high, meaning that the crack efficiency is good [9]. However, the recent use of slow hash algorithms such as bcrypt and md5crypt has resulted in poor cracking efficiency for the brute-force attack [9]. The dictionary-based cracking option involves the preparation of a word or combination of words that are likely to be used as a password, the calculation of the hash value of this item, and the comparison of the hash to the cracking target. Although cracking can be done at a constant rate regardless of password length, the disadvantage is that the password cracking search range is limited to the dictionary used for the attack. To overcome this, a method for defining and utilizing password transformation rules or adding masking rules to broaden the password search range is used. Hashcat [10] and John the Ripper (JtR) [11] are the most popular free tools for password cracking, and they support all of the password attack methods described above.
In this study, we focus on dictionary-based attacks. The dictionary-based password attack can be said to be a data-based attack, and the definition and use of dictionary transformation rules supported by Hashcat [10] and JtR [11] can be said to be the most basic data-driven expanded attack. This is because the password transformation rules are used by experts to analyze leaked passwords, identify frequently used transformation rules, and describe them with a specific format.
However, the methods used by security experts to analyze leaked passwords and create rules have limitations: the first is that the number of leaked passwords is too large for people to analyze. There are about 14 million plaintext passwords in RockYou [12,13], and there are about 60 million LinkedIn [14] plaintext passwords cracked at Hashes.org. Analyzing tens of millions of leaked passwords to create password transformation rules is time-consuming and tedious for people to do. Second, it is challenging to cover tens of millions of leaked password objects with thousands of rules. Because passwords reflect personal preferences and linguistic characteristics (Korean, Japanese, Chinese, etc.), it is challenging to represent all possibilities with specific rules, and if as many rules are generated as the number of leaked passwords, the efficiency of password cracking will vanish. Therefore, a study was conducted to automatically generate a password that would be likely to be used by people by analyzing a leaked password. The probability-based analysis of leaked passwords was conducted to generate new passwords, followed by template-based password generation studies using probabilistic context-free grammar (PCFG) [15,16]. With the development of machine learning, especially deep learning technology, studies on the application of deep learning technology to password cracking have recently been published. The most recently published study was PassGAN [17], and our team has introduced rPassGAN, which applied the recurrent neural network (RNN) to PassGAN. In this study, we present an additional RNN-based PassGAN model and confirm the superiority of the RNN-based deep learning model through a performance comparison between deep learning models. In addition, through a performance comparison analysis with PCFG, which showed excellent performance overall, we identified the complementary characteristics of PCFG and deep learning models.
The structure of this manuscript is as follows. In Section 2, we briefly explain the study of probability-based password cracking methods, including PCFG, as well as providing basic knowledge about deep learning and PassGAN. In Section 3, we first describe the model we presented in previous research, and then we investigate the disadvantages of the previous model. We explain how we tried to solve the problem with the previous model, and we propose a new model that does not suffer from the same problem. Section 4 describes the experimental environment and the process of deriving deep learning parameters. In Section 5, a performance comparison between deep learning models and a performance comparison that includes PCFG is presented; in this section, we further identify the causes of the performance gap between the PCFG and rPassGAN models in a performance comparison with PCFG. The final section presents the conclusions of this study.

2. Related Studies

In the following section, an assessment of password prediction methods will be introduced. A detailed analysis of the generative adversarial networks (GANs) will be provided in Section 3.

2.1. Markov and Context-Free Grammar Approaches

In using a probabilistic approach, a process using the Markov model was recommended. Narayanan and colleagues put forward a process that would create a pseudo password, working with the Markov model [18]. The central premise of this approach is that when people create a new password, they generally use simple passwords that they can remember easily; these passwords use a regular pattern of various alphanumeric sequences, which defines the probability of each sequence that is created. The research was subsequently extended by Ma et al. and Dürmuth et al. [4,5]. Ma attempted to find a better way of measuring password cracking performance and introduced the N-gram model for statistical language modeling to password modeling. Ma [5] showed that whole-string Markov models outperformed PCFG in their experiments. Dürmuth [4] also proposed an efficient password guessing program based on a Markov model, which was called OMEN. OMEN used an improved enumeration algorithm, which was called “enumPwd”, which allowed OMEN to produce the most likely passwords first. The application of the PCFG approach to the password prediction process, which was initially suggested by Weir and colleagues [15], was the most crucial aspect of these experiments. This PCFG-based analysis has continued to broaden and improve [19,20]. The current complex password grammatical structure is a combination of alphanumerical sequences, special characters, and keyboard-walks. This grammatical structure is used to analyze and calculate the distribution probability from leaked passwords. This information is then stored for later use. The PCFG generates a password using the grammatical structure with the highest probability using the information from the leaked passwords; this means that the PCFG improves its cracking capability over time. Methods using PCFG succeeded in cracking passwords at a higher rate when compared to attack methods that used a dictionary with Hashcat’s built-in rules as its source. In turn, the password cracking range has been effectively expanded by the use of this method. Houshmand and colleagues [20] focused on improving the cracking performance of keyboard walk structures and employed a smoothing technique. In their experiments, they showed better performance for cracking keyboard-walk patterned passwords than their previous approach.

2.2. Deep Learning Approaches

The RNN model of password prediction is a deep learning method that states that the characters that constitute the password are based on formerly used characters. This deep learning approach, which is comparable to the Markov method, performs remarkably well in the area of natural language processing and is often used in numerous functions, such as chat-bot applications, translation, and auto-completion. Melicher and colleagues [21] proposed a password prediction process that employs deep learning by applying an RNN [22]. This password prediction process creates a character unit password by using leaked passwords as training data. Furthermore, Hitaj and his colleagues developed the PassGAN [17] method, which also uses a deep learning-based password prediction process.
The relatively new and improved Wasserstein GAN (IWGAN) [23] was utilized in the creation of the PassGAN method. The GAN, which is used in PassGAN, is a deep learning model that has become the most popular generation model. Presented by Goodfellow and colleagues [24], the intent of the first GAN model was to create examples that they expected the population to include throughout the exercise. This was done to achieve a distribution that was indistinguishable from the high-dimension population. The GAN contains a unique structure compared to the current neural network model. The GAN applies two deep neural networks (DNNs) known as generative DNN (denoted as G) and the discriminative DNN (denoted as D): the role of G is to create perfect pseudo password samples that are indistinguishable from the authentic passwords, while the role of D is to differentiate between authentic passwords and the pseudo password samples created by G. This minimax problem, which was proven mathematically by Goodfellow, has a global optimum when the distribution of pseudo samples created by G is indistinguishable to the distribution of the factual data. The minimax problem can be expressed as follows:
m i n G   m a x D   V ( D , G ) = E x P d a t a ( x ) [ l o g D ( x ) ] + E z P z ( z ) [ l o g ( 1 D ( G ( z ) ) ) ]
Multiple GAN models with improved performance have been suggested since Goodfellow first proposed his GAN model. The Wasserstein GAN (WGAN) [25] and the IWGAN [23] are two of the proposed GAN methods that offer the stability that allows the GAN model to find the global optimum. IWGAN was proposed by Gulrajani and colleagues [23], who introduced the concept of a gradient penalty instead of gradient clipping in WGAN and achieved stable training of the GAN model without divergence. The IWGAN showed that a CNN-based IWGAN model could be utilized for text generation. PassGAN utilizes RockYou leaked passwords [12,13] as training data and originated from the results of this experiment. In PassGAN, the role of D is to differentiate between the authentic passwords and the pseudo password samples created by G, while G is programed to create passwords that are comparable to the leaked passwords, which ultimately deceive D. In experiments, PassGAN displayed the ability to create passwords that Hashcat is unable to create [17].
Liu and his colleagues proposed the GENPass method, which uses both the generator and classifier as deep learning models for password generation [26]. The generator uses LSTM and PCFG. But the classifier uses CNN and heuristic constant. The discriminator judges whether a candidate will be accepted. In this paper, we will focus on the improvement of a deep learning model without using the multi-source approach.

3. Proposed Model

In our previous research [27], we approached the issue of improving password cracking performance from two standpoints. The first standpoint was to alter the discriminator and generator of the deep learning model, while the second was to add the structure change to our first approach.

3.1. Transformation of Neural Network Type

Based on IWGAN, PassGAN was developed with the generator and the discriminator utilizing CNNs. The CNN, although not the ideal answer for text generation, can be used for this purpose. Passwords are created by characters that follow a certain pattern and obey specific rules. For example, consider the password “p@ssw0rd”: if we recognize the @ symbol and that letter s follows the letter p, then it is possible to infer that the letter s will be the next character. We arrive at this conjecture on account of people typically using characters in an established arrangement, with the preceding characters having a predictable result on the later characters. In this area of password cracking, the RNN [22] is a fitting deep-learning model that performs well in handling this sequential information, whereas long short-term memory (LSTM) [28] and gated recurrent unit (GRU) [29] are suitable RNN cell types. A few text generation studies have used the GAN model, which was based on LSTM or GRU [30,31,32]. Password cracking performance is likely to be improved when the model of D and G is shifted from a CNN model to an RNN model without modifying PassGAN’s design, as shown in Figure 1b. This new model is called rPassGAN.

3.2. Transformation of Architecture

We also proposed a change to the model of D and G from CNN to RNN, while simultaneously changing the structure of the GAN. When changing the GAN structure, we proposed the application of dual discriminators that utilize the IWGAN’s cost function. D1’s role is to identify the authentic passwords among the samples, whereas D2’s role is to identify the pseudo passwords among the samples. Both discriminators would execute the same role as in Nguyen and colleagues’ D2GAN method [33]. The objective of G in this dual-discriminator GAN model would be the deception of both discriminators, where the distribution of both the authentic and pseudo password samples created by G has to be indistinguishable. This creates a situation in which D1 and D2 are not able to differentiate between the authentic and pseudo passwords. Each discriminator’s cost is defined, in this first dual-discriminator model, with both discriminators being trained to minimize the cost. In contrast, the generator is trained to augment the total costs (Algorithm 1). This dual-discriminator combination PassGAN is known by its abbreviation rPassD2CGAN (Figure 2).
Algorithm 1 rPassD2CGAN calculates each discriminator’s gradient penalty. We use the default values λ = 10 , n c r i t i c = 10 , n g e n = 40 , α = 0.0001 , β 1 = 0.5 and β 2 = 0.9 .
Require: Gradient penalty coefficient λ , number of critic iterations per generator n c r i t i c , number of generator iterations per discriminator n g e n , batch size m, and Adam hyper-parameters α , β 1 , and β 2 .
Require: Initial D 1 and D 2 critic parameters w 0 and u 0 , and initial generator parameter θ 0 .
while θ has not converged do
  for  t = 1 , , n c r i t i c   do
   for  i = 1 , , m   do
    Sample real data x P r , latent variable z p ( z ) , and a random number ϵ U [ 0 , 1 ] .
     x ˜ G θ ( z )
     x ^ ϵ x + ( 1 ϵ ) x ˜
     x ¯ ϵ x ˜ + ( 1 ϵ ) x
     L D 1 i D w ( x ˜ ) D w ( x ) + λ ( x ^ D w ( x ^ ) 2 1 ) 2
     L D 2 i D u ( x ) D u ( x ˜ ) + λ ( x ¯ D w ( x ¯ ) 2 1 ) 2
     L D 2 c o m b i = L D 1 i + L D 2 i
   end for
    ( w , u ) Adam ( ( w , u ) 1 m i = 1 m L D 2 c o m b ( i ) , w , u , α , β 1 , β 2 )
  end for
  for  t = 1 , , n g e n   do
   Sample a batch of latent variable { z ( i ) } i = 1 m p ( z )
    θ Adam ( θ 1 m i = 1 m ( D w ( G θ ( z ) ) + D u ( G θ ( z ) ) , θ , α , β 1 , β 2 )
  end for
end while

3.3. Limitation of rPassD2CGAN

rPassD2CGAN exhibited some unstable training results. The data used in the default training data feeding mode are identical in all discriminators. We provided D1 and D2 with the same fake and real passwords; however, when class 4 (number, special characters, alphabet lower and upper) long password data were provided, the rPassD2CGAN training status was unstable. Usually, the Jensen–Shannon divergence (JSD) reduces and then reaches an almost constant value. In the case of rPassD2CGAN, the JSD value did not converge but exhibited instability, as shown in Figure 3a. To solve this problem, we changed some hyper-parameters such as the batch size, G/D ratio, layer dimension, and learning-rate, but these measures were not effective. After many attempts, we decided to change the training data feeding modes. We tested two cases: one used the same real training data and different pseudo data for two discriminators, and the second used different real and pseudo training data feeding modes for two discriminators. These two training data feeding modes in Figure 4 enable rPassD2CGAN to carry out stable training. The password cracking results were similar but were not exactly the same.

3.4. New Dual-Discriminator Model

Here, we introduce another dual-discriminator model that is slightly different from rPassD2CGAN. In rPassD2CGAN, two discriminators D1 and D2, were encapsulated; thus, the two discriminators operated as one discriminator, from the perspective of loss. In the second dual-discriminator model, D1 and D2 are not combined, and the hyper-parameters of D1 and D2 are trained sequentially. Algorithm  2 describes this operation. We use the same training parameter values of rPassD2CGAN for this model. We call this model rPassD2SGAN, and its block diagram is shown in Figure 5. The training sequence is D1, D2, and G, as shown in Figure 5b. rPassD2CGAN and rPassD2SGAN performed similarly in terms of their password cracking performance.
Algorithm 2 rPassD2SGAN calculates each discriminator’s gradient penalty. We use default λ = 10 , n c r i t i c = 10 , n g e n = 40 , α = 0.0001 , β 1 = 0.5 , β 2 = 0.9 .
Require: The gradient penalty coefficient λ , the number of critic iteration per generator n c r i t i c , the number of generator iteration per discriminator n g e n , the batch size m, Adam hyper-parameters α , β 1 , β 2 .
Require: initial D 1 , D 2 critic parameters w 0 and u 0 , initial generator parameter θ 0
while  θ has not converged do
  for  t = 1 , , n c r i t i c   do
   for  i = 1 , , m   do
    Sample real data x P r , latent variable z p ( z ) , a random number ϵ U [ 0 , 1 ] .
     x ˜ G θ ( z )
     x ^ ϵ x + ( 1 ϵ ) x ˜
     L D 1 i D w ( x ˜ ) D w ( x ) + λ ( x ^ D w ( x ^ ) 2 1 ) 2
   end for
    w Adam ( w 1 m i = 1 m L D 1 ( i ) , w , α , β 1 , β 2 )
   for  i = 1 , , m   do
    Sample real data x P r , latent variable z p ( z ) , a random number ϵ U [ 0 , 1 ] .
     x ˜ G θ ( z )
     x ¯ ϵ x ˜ + ( 1 ϵ ) x
     L D 2 i D u ( x ) D u ( x ˜ ) + λ ( x ¯ D w ( x ¯ ) 2 1 ) 2
   end for
    u Adam ( u 1 m i = 1 m L D 2 i , u , α , β 1 , β 2 )
  end for
  for  t = 1 , , n g e n   do
   Sample a batch of latent variable { z ( i ) } i = 1 m p ( z )
    θ Adam ( θ 1 m i = 1 m ( D w ( G θ ( z ) ) + D u ( G θ ( z ) ) , θ , α , β 1 , β 2 )
  end for
end while
In summary, we described two approaches in this section. The first is the neural network type transformation. PassGAN used CNN in the module (discriminator and generator); in contrast, our models used an RNN. This transformation changed the core architecture of PassGAN, and all hyper-parameters should be configured adequately for stable training. The second approach is the transformation of the architecture. D2GAN [33] used the integral probability metric (IPM)-based cost function, which is based on a cross entropy. GAN using the IPM-based cost function has a fundamental weakness; the probability of real data being real should decrease as the probability of fake data being real increases [34]. We applied D2GAN architecture using a non-IPM cost function (WGAN-GP) to rPassGAN. Furthermore, to solve the unstable training of the D2GAN architecture, a sequential training algorithm for dual discriminators was suggested.

4. Experiments

Experiments were devised to analyze if the prospective models have superior performance to that of PassGAN. Furthermore, the results were not only evaluated on their simple success rate at cracking the passwords, but also according to the extensibility of the cracked password. The exact experiment setup was as follows.

4.1. Training Configuration

4.1.1. Training Parameters

An essential training framework equal to that of PassGAN (training values are detailed in Algorithm 1) was used in both models. Our experiments were conducted to obtain the suitable G/D training ratio for our models, with only a single G/D training ratio applied differently from that of PassGAN. The arrangement of the optimized G/D ratio experiment was as follows: the subjects of this experiment were three models that use RNN, excluding PassGAN. Utilizing data acquired from RockYou, the ratio of G/D training was increased from 10:10 to 100:10 with 20,000 (20k) training epochs. N-gram (N = 3, 4) JSD values of RockYou were contrasted with the dictionary generated by each RNN-based model to evaluate the results of the experiments.
The lower the JSD value, the closer the value to the password distribution generated by people. In Figure 6, RNN-based models had the lowest JSD at 40:10 as G/D training iteration parameter. When we compared the cracking performances of rPassGAN, rPassD2CGAN, and rPassD2SGAN, we used the same hyper-parameters, such as the learning rate, G/D iteration ratio, and Adam optimizer parameters, to reduce other factors which may affect the performance.

4.1.2. Training Data

Data in the form of plaintext passwords from RockYou were used during the training of the three deep-learning models. In this experiment, the RockYou dictionary was defined before the models were trained, and the Unicode characters were removed. Next, the length of the passwords found in the RockYou dictionary was analyzed, with nearly 14 million passwords with a length of 1 to 32 characters being used. These passwords were then separated into groups depending on their length, with training dictionaries created for each group using RockYou. The groups consisted of passwords with a length of 1–8 characters, 9–15 characters, and 16–32 characters, respectively. This was done to identify which model showed the best performance for each password length. In addition, when compared with PCFG, passwords of length 1–10 from RockYou and LinkedIn were used. This experiment condition replicated the experimental settings of Hitaj’s PassGAN [17].

4.2. Password Cracking

Passwords leaked from the LinkedIn website were used as the targets for our password cracking experiments. In total, 20% of the leaked LinkedIn passwords with a character length of 1–32 characters were used as the target in the password cracking performance test for the RockYou dictionary-trained models. Just as in the RockYou training data case, the Unicode passwords were removed from the password cracking targets. The best64 rule was used when testing the cracking performance with Hashcat. This was done for two reasons: first, conditions that are comparable to field conditions, where best64 is commonly applied to the password dictionary, were needed to test the performance of the password prediction methods; second, the billions of passwords utilizing the deep-learning model required a considerable amount of time to generate. Thus, even if a limited amount of passwords were generated, we could successfully crack many passwords that were identified with a smooth transformation by employing the best64 rule. As already mentioned in the above section, when comparing with PCFG, we did not employ the best64 rule to all models to create an experimental environment similar to that of PassGAN.

5. Evaluation

5.1. Dictionary Quality Perspective

Two aspects were considered when we evaluated the quality of the password-prediction dictionary. The JSD was used as the evaluation criteria to measure the similarity of pseudo passwords generated by the model. First, the similarities between the dictionary generated by the model and the training dictionary created by people were considered. Second, the number of unique passwords predicted by the dictionary using the model was also considered. To avoid any degradation of the password cracking performance, the duplicated passwords generated by the model needed to be minimized as much as possible. The JSD value for each N-gram was compared with that of a human-generated password dictionary (the RockYou leaked password dictionary). As shown in Figure 7, the RNN-based models contained diminished JSD values compared to PassGAN for all N-grams. In terms of N-grams, the RNN-based models exhibited a higher resemblance to that human-generated password distribution than PassGAN.
The RNN-based models performed better than PassGAN when considering the rate of redundancy of the generated candidate passwords in Table 1. The space for generating the passwords widened exponentially when the length of the generated passwords increased. Therefore, the password group with the length of 16–32 characters saw its redundancy rate vanish. Even in a limited password space, the RNN-based models exhibited low password generation redundancy, causing the passwords to be cracked efficiently.

5.2. Cracking Performance Perspective

The large number of passwords successfully cracked and the contribution to the cracking of hashes that previously had not been cracked were the most appealing aspects of creating a password-guessing dictionary using models that were trained on a leaked-password dictionary. The improved password cracking performance and cracking area expansion capabilities of rPassGAN, rPassD2CGAN, and rPassD2SGAN compared with PassGAN will now be considered.

5.2.1. Total Password Cracking

First, the cracking performance for each training epoch will be presented in Figure 8. A modest contrast in password cracking performance after all models were trained for 100,000 (denoted as 100k) epochs was exhibited in the results of the RockYou password length groups. Acceptable password cracking performance for all password lengths and training epochs were exhibited in the RNN-based models.
The connection between the number of password predictions and the number of increased password cracks is displayed in Figure 9. A deviation appears between the early training (10k epoch) model and the final (200k epoch) model when the number of increased password cracks is analyzed. At the beginning and even until the latter half of the experiment, both rPassD2CGAN and rPassD2SGAN display the highest number of increased crackings for the early training (10k epoch) model. When analyzing the 200k epochs, the only difference that can be confirmed is between PassGAN and the three RNN models due to the slim difference between the RNN models. A negligible difference can be seen in the results for the password-length groups. However, rPassD2CGAN and rPassD2SGAN performed better than rPassGAN in the case of the password length groups 9–15 and 16–32.
The password cracking performances between deep learning-based models were compared. Furthermore, we conducted password cracking performance measurement experiments that included PCFG, similar to Hitaj’s PassGAN study [17]. In these experiments, unlike in the previous experiments, we used RockYou and LinkedIn leaked passwords with the length 1–10 characters as training data and cracking targets. We created two sets of training data and cracking targets by using the Linux command “sort-R”. We split the randomly sorted data into training data (80%) and cracking targets (20%). Table 2 shows the password cracking performance for each model. “Tr” means training data, and “Cr” means cracking targets. Postfix digit means the index of the data set. For rk0–rk0 and rk0–lk0, deep learning models, including PassGAN, achieved similar or better password cracking performance than PCFG; except for these two cases, PCFG achieved the best performance.
We decided to measure JSD to estimate the difference in the probabilistic distribution between the training data and the cracking targets. Table 3 shows the result of the measurement. All cases except rk0-rk0 and rk0-lk0 showed a very low JSD value; in particular, JSD1 was almost zero. For rk0-rk0 and rk0-lk0, the JSD values were relatively high; the higher JSD values mean that the probabilistic distribution of the training data is quite different from that of the cracking target. This means that the deep learning models, including PassGAN, have the ability to crack passwords with new or strange patterns that did not exist in the training data.

5.2.2. Cracking Extensibility

Below, the results of the password cracking space extension of each model’s password-predicting dictionary experiment are shown. Two perspectives were utilized when analyzing the experimental results of the password cracking space extensibility: first, the extension of the password dictionary generated by each model for the cracking space of the authentic passwords used in the training of each model was analyzed; second, the proportion of passwords that could be cracked by the model, and the number of passwords that the model did not crack for all proposed models, was compared.
Table 4 shows the disparity in the number of cracks generated by the password dictionary for each model. Regions in which all proposed models had exclusive password cracking success where the other models had none are also detailed. These data are significant, on account of the expansion of the password cracking space which the use of these dictionaries would create. rPassGAN was the model with the largest number of unique cracks, followed by rPassD2CGAN, rPassD2SGAN among our models. rPassD2SGAN had the largest number of unique cracks when compared with PassGAN. The RNN models exhibited a much smaller difference to each other than in comparison to PassGAN.

5.3. Password Strength Estimation

The models we selected for the password strength estimation experiment are zxcvbn [7], PCFG, and rPassGAN (with the minimum cracking performance model of the rPassGAN series). To estimate the entropy of passwords, the zxcvbn framework was basically used, because zxcvbn is practically used for password strength estimation for DropBox. Moreover, it has the function that external password candidates can be supplied to zxcvbn. With this function, the effectiveness of password candidates of both PCFG and rPassGAN could be analyzed as regards the estimation of the strength of passwords; if they were effective, the entropy of the password strength would be lower than that of zxcvbn’s basic mode [7]. For this experiment, we ran the password cracking experiment with the same configuration and training data. However, we decided to use a small amount of cracking target passwords for the estimation experiment in the case of the rk0 dataset. Since estimating the strength of 2 million passwords is time-consuming, we randomly made three 10k password sets from the rk0 dataset. In addition, we created some class 4 (uppercase letters, lowercase letters, digits, special symbols such as “!” and “@”) training and estimating datasets from RockYou, LinkedIn, iMesh, ZooSK, and Myspace leaked passwords. These cracked plaintext passwords can be downloaded from Hashes.org. In zxcvbn [7], neither KeePass nor NIST formalize the type of entropy they model; zxcvbn [7] assumes that n bits of strength means that guessing the password is equivalent to guessing a value of a random variable X according to
n = H ( X ) = i p ( x i ) l o g 2 p ( x i )
Assuming the guesser knows the distribution over X and attempts guesses x i in a decreasing order of probability p ( x i ) , a lower bound on the expected number of guesses E [ G ( X ) ] can be shown as
E [ G ( X ) ] 2 H ( X ) 2 + 1
provided that H ( X ) 2 . This conservative lower bound to convert bits into guesses in zxcvbn [7] was used. Table 5 shows the percentage of cases in which each model performs best, corresponding to the same passwords. A lower entropy and lower guessing number of a password mean that the strength of this password is not high; thus, if a model showed lower entropy than others, this model would represent a robust password strength estimator.
When the password candidates of rPassGAN and PCFG were supplied to zxcvbn, a better password strength estimating performance was observed. This means that rPassGAN and PCFG’s password candidates are helpful for stricter password strength estimation. Further, the entropy of rPassGAN was slightly lower than that of PCFG and zxcvbn with the rk0 dataset, where JSD is higher than the others (see Table 3). In the case of class 4 (password length 14–16), rPassGAN showed slightly lower entropy than PCFG except for passwords with a length of 16. In the case of cls4(16), rPassGAN showed weakness at estimating keyboard-walk passwords; in contrast, PCFG showed the advantages of estimating keyboard-walk passwords. In total, 20% passwords of PCFG’s best cases in cls4(16) were keyboard-walk passwords. rPassGAN should be trained with a collection of keyboard-walk passwords to make up for this weakness and generate the keyboard-walk password candidates, which would help zxcvbn in strictly estimating passwords. Table 6 shows some examples of password strength estimation results.

6. Conclusions

The three models introduced in this study performed better at cracking passwords than PassGAN—the first GAN-based password predicting model. We applied the RNN as the primary DNN model to our models. Although PassGAN performed better when changing from CNN to RNN, we achieved substantially improved performance with the adoption of a dual discriminator structure. From this experiment, we derived the G/D training ratio for the enhanced training in the RNN-based IWGAN. Through the use of password cracking performance tests, it has been shown that RNN-based models performed better than PassGAN. rPassGAN and PCFG both displayed improved password estimation results in the password strength estimation experiment. An extensive amount of time was needed to generate the password-guessing dictionary and train the exiting password dictionary using the deep-learning model; nonetheless, by performing these procedures, the time invested resulted in a secure password-predicting dictionary.
The selection of a password generation model for the password cracking extension of PCFG could be influenced by the results of this experiment. We advise the use of a single RNN-based model for password cracking when enough time is available, but the desired computing capability is not. The RNN-model chosen in this situation is not essential; however, if time is limited but adequate computer capability is available, the use of rPassD2CGAN or rPassD2SGAN is advised since both achieved sufficient password cracking performance with a small epoch, such as 10k. When a maximum number of cracked passwords is needed, all models should be used since each has its own set of cracked passwords. Although PCFG achieved a better cracking performance than the deep learning models, our results revealed that the deep learning models had advantages for certain password cracking cases; specifically, when the probabilistic distribution difference between the training and the cracking targets was significant, rPassGAN showed better cracking performance than PCFG. This means that the generality of the deep learning models was better than that of PCFG. Because PCFG as a model is suitable for cracking passwords that have patterns similar to those in training data, the first password-cracking step using PCFG is to crack passwords with similar patterns as the training data. In the second step, the deep learning model is used to implement password cracking with different learning data and patterns, which effectively maximizes the overall password cracking space. The maximized password candidates from both the RNN-based model and PCFG are helpful in setting default passwords for IoT devices. Furthermore, modern deep learning libraries, such as Google’s tensorflow [35], support the loading of the trained model with Javascript and for mobile devices. Thus, we can easily apply the deep learning password cracking methodology to web browsers and IoT devices.

Author Contributions

Conceptualization, S.N.; methodology, S.N.; software, S.J.; validation, S.J.; data curation, H.K.; writing—original draft preparation, S.N.; writing—review and editing, J.M.; supervision, J.M.; All authors have read and agreed to the submitted version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

This research was supported by a Korea University Grant.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
GANGenerative Adversarial Network
PCFGProbabilistic Context-Free Grammar
RNNRecurrent Neural Network
CNNConvolutional Neural Network
JtRJohn the Ripper
JSDJensen–Shannon Divergence

References

  1. Antonakakis, M.; April, T.; Bailey, M.; Bernhard, M.; Bursztein, E.; Cochran, J.; Durumeric, Z.; Halderman, J.A.; Invernizzi, L.; Kallitsis, M.; et al. Understanding the Mirai Botnet. In Proceedings of the 26th USENIX Security Symposium, Vancouver, BC, Canada, 16–18 August 2017; pp. 1093–1110. [Google Scholar]
  2. Herley, C.; Van Oorschot, P. A research agenda acknowledging the persistence of passwords. IEEE Secur. Priv. 2012, 10, 28–36. [Google Scholar] [CrossRef] [Green Version]
  3. Dell’ Amico, M.; Michiardi, P.; Roudier, Y. Password Strength: An Empirical Analysis. In Proceedings of the 2010 IEEE INFOCOM, San Diego, CA, USA, 15–19 March 2010; pp. 1–9. [Google Scholar] [CrossRef]
  4. Dürmuth, M.; Angelstorf, F.; Castelluccia, C.; Perito, D.; Chaabane, A. OMEN: Faster Password Guessing Using an Ordered Markov Enumerator. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2015; pp. 119–132. [Google Scholar] [CrossRef] [Green Version]
  5. Ma, J.; Yang, W.; Luo, M.; Li, N. A study of probabilistic password models. In Proceedings of the IEEE Symposium on Security and Privacy, San Jose, CA, USA, 18–21 May 2014; pp. 689–704. [Google Scholar] [CrossRef]
  6. National Institute of Standards and Technology. Digital Identity Guidelines. Available online: https://pages.nist.gov/800-63-3/ (accessed on 10 June 2017).
  7. Wheeler, D.L. ZxCVBN: Low-budget password strength estimation. In Proceedings of the 25th USENIX Security Symposium, Austin, TX, USA, 10–12 August 2016. [Google Scholar]
  8. Tasevski, P. Password Attacks and Generation Strategies. Master’s Thesis, Tartu University: Faculty of Mathematics and Computer Sciences, Tartu, Estonia, 2011. [Google Scholar]
  9. Gosney, J.M. Hashcat Benchmarks 8x Nvidia GTX 1080 Hashcat Benchmarks. Available online: https://gist.github.com/epixoip/a83d38f412b4737e99bbef804a270c40 (accessed on 12 July 2018).
  10. Hashcat team. Hashcat Advanced Password Recovery. Available online: https://hashcat.net/wiki/ (accessed on 12 July 2018).
  11. Openwall.com. John the Ripper password cracker. Available online: http://www.openwall.com/john/ (accessed on 12 July 2018).
  12. Skullsecurity. RockYou Online. Available online: https://wiki.skullsecurity.org/Passwords (accessed on 12 July 2018).
  13. Cubrllovic, N. Rockyou Hack: From Bad To Worse. Available online: https://techcrunch.com/2009/12/14/rockyou-hack-security-myspace-facebook-passwords (accessed on 12 July 2018).
  14. Hashes.org. LinkedIn Online. Available online: https://hashes.org/public.php (accessed on 5 October 2018).
  15. Weir, M.; Aggarwal, S.; De Medeiros, B.; Glodek, B. Password cracking using probabilistic context-free grammars. In Proceedings of the 2009 30th IEEE Symposium on Security and Privacy, Oakland, CA, USA, 17–20 May 2009. [Google Scholar] [CrossRef]
  16. Weir, M.; Aggarwal, S.; Collins, M.; Stern, H. Testing metrics for password creation policies by attacking large sets of revealed passwords. In Proceedings of the ACM Conference on Computer and Communications Security, Chicago, IL, USA, 4–8 October 2010; pp. 162–175. [Google Scholar] [CrossRef]
  17. Hitaj, B.; Gasti, P.; Ateniese, G.; Perez-Cruz, F. PassGAN: A Deep Learning Approach for Password Guessing. In Applied Cryptography and Network Security; Springer: Cham, Switzerland, 2019; pp. 217–237. [Google Scholar] [CrossRef] [Green Version]
  18. Narayanan, A.; Shmatikov, V. Fast dictionary attacks on passwords using time-space tradeoff. In Proceedings of the ACM Conference on Computer and Communications Security, Alexandria, VA, USA, 7–11 November 2005. [Google Scholar] [CrossRef]
  19. Yazdi, S.H. Probabilistic Context-Free Grammar Based Password Cracking: Attack, Defense and Applications. Ph.D. Thesis, Florida State University: Department of Computer Science, Tallahassee, FL, USA, 2015. [Google Scholar]
  20. Houshmand, S.; Aggarwal, S.; Flood, R. Next Gen PCFG Password Cracking. IEEE Trans. Inf. Forensics Secur. 2015, 10, 1776–1791. [Google Scholar] [CrossRef]
  21. Melicher, W.; Ur, B.; Segreti, S.M.; Komanduri, S.; Bauer, L.; Christin, N.; Cranor, L.F. Fast, lean, and accurate: Modeling password guessability using neural networks. In Proceedings of the 25th USENIX Security Symposium, Austin, TX, USA, 10–12 August 2016. [Google Scholar]
  22. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  23. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A. Improved training of wasserstein GANs. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  24. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  25. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein Generative Adversarial Networks. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; Volume 70, pp. 214–223. [Google Scholar]
  26. Liu, Y.; Xia, Z.; Yi, P.; Yao, Y.; Xie, T.; Wang, W.; Zhu, T. GENPass: A General Deep Learning Model for Password Guessing with PCFG Rules and Adversarial Generation. In Proceedings of the 2018 IEEE International Conference on Communications (ICC), Kansas City, MO, USA, 20–24 May 2018; pp. 1–6. [Google Scholar] [CrossRef]
  27. Nam, S.; Jeon, S.; Moon, J. A New Password Cracking Model with Generative Adversarial Networks. In Information Security Applications; Springer: Cham, Switzerland, 2020; pp. 247–258. [Google Scholar] [CrossRef]
  28. Gers, F.A.; Schmidhuber, J.A.; Cummins, F.A. Learning to Forget: Continual Prediction with LSTM. Neural Comput. 2000, 12, 2451–2471. [Google Scholar] [CrossRef] [PubMed]
  29. Cho, K.; van Merrienboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1724–1734. [Google Scholar] [CrossRef]
  30. Wang, H.; Qin, Z.; Wan, T. Text Generation Based on Generative Adversarial Nets with Latent Variable. In Advances in Knowledge Discovery and Data Mining; Springer: Cham, Switzerland, 2018. [Google Scholar]
  31. Press, O.; Bar, A.; Bogin, B.; Berant, J.; Wolf, L. Language Generation with Recurrent Generative Adversarial Networks without Pre-training. arXiv 2017, arXiv:1706.01399. [Google Scholar]
  32. Subramanian, S.; Rajeswar, S.; Dutil, F.; Pal, C.; Courville, A. Adversarial Generation of Natural Language. arXiv 2017, arXiv:1705.10929. [Google Scholar] [CrossRef]
  33. Nguyen, T.D.; Le, T.; Vu, H.; Phung, D. Dual Discriminator Generative Adversarial Nets. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 2667–2677. [Google Scholar]
  34. Jolicoeur-Martineau, A. The relativistic discriminator: A key element missing from standard GAN. arXiv 2018, arXiv:1807.00734. [Google Scholar]
  35. Google Inc. Tensorflow Lite. Available online: https://www.tensorflow.org/lite/ (accessed on 1 December 2019).
Figure 1. The model architecture of PassGAN & rPassGAN: PassGAN’s G and D consist of a five-layered residual convolutional neural network (CNN) with filter size 5. Otherwise, we replace the five-layered residual CNN with a one-layer RNN. In the rPassGAN, the red arrow stands for the recurrent flow. We mainly use gated recurrent unit (GRU) cells, because the training and sampling speed is better than long short-term memory (LSTM). FC stands for the fully-connected multilayer perceptron (MLP) in both Figure 1a,b. GAN stands for generative adversarial network.
Figure 1. The model architecture of PassGAN & rPassGAN: PassGAN’s G and D consist of a five-layered residual convolutional neural network (CNN) with filter size 5. Otherwise, we replace the five-layered residual CNN with a one-layer RNN. In the rPassGAN, the red arrow stands for the recurrent flow. We mainly use gated recurrent unit (GRU) cells, because the training and sampling speed is better than long short-term memory (LSTM). FC stands for the fully-connected multilayer perceptron (MLP) in both Figure 1a,b. GAN stands for generative adversarial network.
Sensors 20 03106 g001aSensors 20 03106 g001b
Figure 2. rPassD2CGAN block diagram.
Figure 2. rPassD2CGAN block diagram.
Sensors 20 03106 g002
Figure 3. rPassD2CGAN showed some unstable training results. The use of different training data feeding modes stabilized rPassD2CGAN during training.
Figure 3. rPassD2CGAN showed some unstable training results. The use of different training data feeding modes stabilized rPassD2CGAN during training.
Sensors 20 03106 g003
Figure 4. Different training data feeding modes.
Figure 4. Different training data feeding modes.
Sensors 20 03106 g004
Figure 5. rPassD2SGAN block diagram and training step sequence.
Figure 5. rPassD2SGAN block diagram and training step sequence.
Sensors 20 03106 g005
Figure 6. N-gram JSD (N = 3, 4) by G/D ratio (X-axis: G/D training ratio; Y-axis: JSD value). The data applied in training consisted of a RockYou password dictionary of length 1–8.
Figure 6. N-gram JSD (N = 3, 4) by G/D ratio (X-axis: G/D training ratio; Y-axis: JSD value). The data applied in training consisted of a RockYou password dictionary of length 1–8.
Sensors 20 03106 g006
Figure 7. N-gram JSD of generated password dictionary (X-axis: password length; Y-axis: JSD value).
Figure 7. N-gram JSD of generated password dictionary (X-axis: password length; Y-axis: JSD value).
Sensors 20 03106 g007
Figure 8. Total cumulative cracked passwords by training epoch.
Figure 8. Total cumulative cracked passwords by training epoch.
Sensors 20 03106 g008
Figure 9. Total cumulative cracked passwords by the number of guessed passwords.
Figure 9. Total cumulative cracked passwords by the number of guessed passwords.
Sensors 20 03106 g009
Table 1. Redundancy rate of password guessing candidate dictionary.
Table 1. Redundancy rate of password guessing candidate dictionary.
Models1–89–1516–321–32
PassGAN7.52%0.55%0.10%4.08%
rPassGAN4.42%0.15%0.10%2.02%
rPassD2CGAN4.34%0.15%0.09%2.05%
rPassD2SGAN4.32%0.15%0.09%2.05%
Table 2. Password cracking performance results.
Table 2. Password cracking performance results.
Tr/CrPassGANrPassGANrPassD2CGANrPassD2SGANPCFG
rk0-rk0251,592287,745289,753292,626141,278
lk0-lk0666,571735,264764,762764,0822,725,501
rk0-lk0677,542765,324772,956777,5862,186,212
lk0-rk0346,303391,565399,256402,696417,802
rk1-rk1396,851440,020449,401448,772845,539
lk1-lk1707,567732,795753,700763,4942,724,828
rk1-lk1738,499803,508824,561823,6392,505,716
lk1-rk1299,449314,142321,967324,575961,771
Password guessing number: 250 million, with no Hashcat rule such as best64 employed, Tr means training data index; Cr means cracking data index. PCFG: probabilistic context-free grammar.
Table 3. JSD between training data and cracking target data in Table 2.
Table 3. JSD between training data and cracking target data in Table 2.
Tr/CrJSD1JSD2JSD3JSD4
rk0-rk00.2794750.3477730.4330520.536094
lk0-lk00.0000000.0000310.0020260.030252
rk0-lk00.0106360.0245970.0467710.096937
lk0-rk00.1980820.2506800.3185820.415374
rk1-rk10.0000010.0001160.0046430.041829
lk1-lk10.0000000.0000300.0020320.030271
rk1-lk10.0022100.0133190.0320670.079866
lk1-rk10.0021880.0133750.0344950.095603
Table 4. Difference between each deep learning model’s cracking result set.
Table 4. Difference between each deep learning model’s cracking result set.
ModelsPassGANrPassGANrPassD2CGANrPassD2SGAN
PassGAN0107,103107,406113,048
rPassGAN235,9370143,182149,495
rPassD2CGAN235,127142,0690148,390
rPassD2SGAN241,456149,069149,0770
The number of each element in the table stands for the size of the difference of the element between the two models. A better model has more unique cracked passwords; thus, for a better model, the size of difference will be larger than that of others. The elements are calculated by the following rule: m1 = PassGAN, m2 = rPassGAN, m3 = rPassD2CGAN, m4 = rPassD2SGAN S m i = {x | x is cracked passwords of model mi} |D(i, j)| = | S m i S m j |, (i is row index, j is column index.) For example, |D(2, 1)| = |SrPassGANSPassGAN|.
Table 5. Percentage of the lowest entropy cases for each model.
Table 5. Percentage of the lowest entropy cases for each model.
DatasetrPassGANPCFGzxcvbnNone
rk0-015.6% (17.91)2.3% (18.13)0.5% (18.19)81.6%
rk0-18.4% (31.84)12.1% (31.79)0.0% (32.37)79.5%
rk0-212.3% (32.59)12.3% (32.74)0.0% (33.57)75.4%
cls4(14)4.7% (38.43)5.7% (38.48)0.1% (39.03)89.5%
cls4(15)8.4% (38.68)4.9% (39.40)0.1% (39.87)86.6%
cls4(16)1.4% (45.18)4.6% (44.61)0.8% (45.32)93.2%
The stronger the password estimator is, the lower the entropy and guessing number will be. The values in parentheses are the average of all target passwords’ entropy for each model. The “None” column described the percentage of cases when the three models resulted in the same entropy value; this means that no models contributed to enhancing password estimation. The Prefix cls4 datasets were created from RockYou, LinkedIn, iMesh, ZooSK, and Myspace from Hashes.org. And, the number in the parentheses indicates the length of estimating passwords.
Table 6. The example results of password strength estimation, where rPassGAN showed the lowest entropy.
Table 6. The example results of password strength estimation, where rPassGAN showed the lowest entropy.
PasswordsrPassGANPCFGzxcvbn
Charlotte.201018.83 ( 1.17 × 10 5 ) 21.852 ( 9.46 × 10 5 ) 23.916 ( 3.96 × 10 6 )
C0mm3m0r47!0ns18.83 ( 1.17 × 10 5 ) 41.11 ( 2.97 × 10 11 ) 40.11 ( 2.97 × 10 11 )
MyLinkedIn101!25.333 ( 1.06 × 10 7 ) 28.605 ( 1.02 × 10 8 ) 35.598 ( 1.30 × 10 10 )
Linkedin2011@@17.897 ( 6.10 × 10 4 ) 19.167 ( 1.47 × 10 5 ) 27.644 ( 5.24 × 10 7 )
Profe$$1ona11$m18.527 ( 9.44 × 10 4 ) 29.824 ( 2.38 × 10 8 ) 29.824 ( 2.38 × 10 8 )
Concep+ua1!z!ng21.27 ( 6.32 × 10 5 ) 41.066 ( 5.75 × 10 11 ) 47.09 ( 5.75 × 10 11 )
C0ns4nguini+i3s14.997 ( 8.18 × 10 3 ) 47.09 ( 3.74 × 10 13 ) 47.09 ( 3.74 × 10 13 )
September27,198722.682 ( 1.68 × 10 6 ) 40.457 ( 3.77 × 10 11 ) 40.457 ( 3.77 × 10 11 )
JLN@linkedin201128.022 ( 6.81 × 10 7 ) 51.308 ( 6.97 × 10 14 ) 51.308 ( 6.97 × 10 14 )
@WSX$RFV1qaz3edc46.003 ( 1.76 × 10 13 ) 28.724 ( 1.11 × 10 8 ) 46.003 ( 1.76 × 10 13 )
!QAZ1qaz@WSX2wsx31.184 ( 6.10 × 10 8 ) 30.032 ( 2.74 × 10 8 ) 46.003 ( 1.76 × 10 13 )
#$ERDFCV34erdfcv55.142 ( 9.94 × 10 15 ) 24.188 ( 4.78 × 10 6 ) 55.142 ( 9.94 × 10 15 )
1qaz)OKM2wsx(IJN46.003 ( 1.76 × 10 13 ) 30.741 ( 4.49 × 10 8 ) 46.003 ( 1.76 × 10 13 )
!QAZ#EDC%TGB7ujm48.253 ( 8.81 × 10 13 ) 28.843 ( 1.20 × 10 8 ) 48.325 ( 8.81 × 10 13 )
The number in parentheses is the guessing number from Equation (3). In keyboard-walk passwords, rPassGAN showed the weakness.

Share and Cite

MDPI and ACS Style

Nam, S.; Jeon, S.; Kim, H.; Moon, J. Recurrent GANs Password Cracker For IoT Password Security Enhancement. Sensors 2020, 20, 3106. https://doi.org/10.3390/s20113106

AMA Style

Nam S, Jeon S, Kim H, Moon J. Recurrent GANs Password Cracker For IoT Password Security Enhancement. Sensors. 2020; 20(11):3106. https://doi.org/10.3390/s20113106

Chicago/Turabian Style

Nam, Sungyup, Seungho Jeon, Hongkyo Kim, and Jongsub Moon. 2020. "Recurrent GANs Password Cracker For IoT Password Security Enhancement" Sensors 20, no. 11: 3106. https://doi.org/10.3390/s20113106

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop