Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Suppressor, wordlist, rules: bug fix and enhancements/cleanup #5716

Merged
merged 6 commits into from
Mar 26, 2025

Conversation

solardiz
Copy link
Member

Previously, we only supported a repeated initialization call after cracking mode transition, but as magnum found we may also call more than once from within wordlist mode.

Fixes #5714

Previously, we only supported a repeated initialization call after cracking
mode transition, but as magnum found we may also call more than once from
within wordlist mode.

Fixes openwall#5714
@solardiz solardiz requested a review from magnumripper March 25, 2025 00:27
Copy link
Member

@magnumripper magnumripper left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks good to me. I can test it tomorrow with the files from the original problem.

@solardiz
Copy link
Member Author

I sneaked a couple of other suppressor enhancements in here (separate commits).

@solardiz solardiz changed the title Suppressor: Handle repeated initialization from wordlist mode correctly Suppressor, wordlist, rules: bug fix and enhancements/cleanup Mar 25, 2025
@solardiz
Copy link
Member Author

Got yet another subtly related change in here:

Wordlist, rules: Drop the dupe check against the previous word

This check dates back to 1990s when our primary target was descrypt with
its length truncation, its computation was much slower, and we didn't have
the generic dupe suppressor.  In the suppressor, a word matching the
previous one would be an L1 cache hit, so we don't gain much by
pre-checking for that special case.  And if the suppressor is disabled,
that's probably because we don't want any overhead like that.

Besides, this check didn't behave correctly along with an external filter()
possibly modifying the word (a bug that also dates back to 1990s).

OTOH, we do not currently support the suppressor on top of stacked rules
(we do before stacked rules), so we do lose a bit of functionality with
this change.

This is something I had thought of doing years ago, and especially after I added the suppressor in 2022.

Testing with our default wordlist and rules, but without suppressor, this changes the output only slightly. In the first 100M candidates produced (including dupes), here are the numbers of unique, before:

Total lines read: 100000000, unique lines written: 90070771 (90%), no slow passes

After:

Total lines read: 100000000, unique lines written: 90070624 (90%), no slow passes

meaning that ~150 more dupes got into the first 100M. I think that difference is so tiny it wasn't worth checking for those dupes, even without suppressor, and now the suppressor would take care of all of them when enabled.

@solardiz
Copy link
Member Author

~150 more dupes got into the first 100M. I think that difference is so tiny it wasn't worth checking for those dupes, even without suppressor, and now the suppressor would take care of all of them when enabled.

Confirming with -dup=256, before:

Total lines read: 100000000, unique lines written: 98577332 (98%), no slow passes

After:

Total lines read: 100000000, unique lines written: 98577332 (98%), no slow passes

@magnumripper
Copy link
Member

Testing with our default wordlist and rules, but without suppressor, this changes the output only slightly

Is there any measurable performance improvement? Perhaps it's insignificant.

@solardiz
Copy link
Member Author

Testing with our default wordlist and rules, but without suppressor, this changes the output only slightly

Is there any measurable performance improvement? Perhaps it's insignificant.

On the system I'm testing on, there's no obvious performance change when running with rules (there's fluctuation of a few percent), but there's a ~10% speedup when running without rules. Perhaps that ARCH_WORD sized pre-check in rules.c was very efficient, and the strcpy / strcmp in wordlist.c not so efficient.

@solardiz
Copy link
Member Author

Looks good to me. I can test it tomorrow with the files from the original problem.

Did you? OK to merge?

We could have already disabled the suppressor before the current mode ends,
in which case a second update on mode end would be incorrect.
to make it friendly to suppressor's random writes.
This check dates back to 1990s when our primary target was descrypt with
its length truncation, its computation was much slower, and we didn't have
the generic dupe suppressor.  In the suppressor, a word matching the
previous one would be an L1 cache hit, so we don't gain much by
pre-checking for that special case.  And if the suppressor is disabled,
that's probably because we don't want any overhead like that.

Besides, this check didn't behave correctly along with an external filter()
possibly modifying the word (a bug that also dates back to 1990s).

OTOH, we do not currently support the suppressor on top of stacked rules
(we do before stacked rules), so we do lose a bit of functionality with
this change.
@solardiz solardiz force-pushed the suppressor-double-init branch from 5f0e5fe to 7e2b4ba Compare March 26, 2025 00:20
@solardiz solardiz merged commit d265b3c into openwall:bleeding-jumbo Mar 26, 2025
35 of 36 checks passed
@solardiz solardiz deleted the suppressor-double-init branch March 26, 2025 04:03
@magnumripper
Copy link
Member

Looks good to me. I can test it tomorrow with the files from the original problem.

Did you? OK to merge?

I was delayed by the LM test vector problem. I expect it to be fine now but I will verify it when possible.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants