are close to 0% of the total possible different patterns
Interestingly, the content of such numbers increases with the growth and height of the range.
Anyone who wishes can calculate independently in the 130 range, but it will be a very long time:
1. Sounds normal, as you increase bit length each existing constraint gets factorized, and new ones also appear. Do you want to exclude 99.97% of the search space because they have 2 consecutive hex digits out of, let's say, 32 hex digits? Because that is where this series will end up.
2. It's computed wrong anyways. Something like 0x7f80 has 2 consecutive hex characters. And I didn't make a typo, maybe think about it in bits. Or should we only exclude byte-aligned bit patterns? See, any seemingly random string can be made into a pattern depending on how you look at it.
Do you know how many numbers need to be generated per second to find Puzzle 65 in 10 minutes?
It's an mission impossible . Even in C++
It is very funny that you fell into the same trap as everyone else. Here's the reality: /dev/urandom is not a secure random number generator. It fall backs to a PRNG when too much data gets read from it.