| # Licensed under the Apache License, Version 2.0 (the "License"); |
| # you may not use this file except in compliance with the License. |
| # You may obtain a copy of the License at |
| # |
| # http://www.apache.org/licenses/LICENSE-2.0 |
| # |
| # Unless required by applicable law or agreed to in writing, software |
| # distributed under the License is distributed on an "AS IS" BASIS, |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or |
| # implied. |
| # See the License for the specific language governing permissions and |
| # limitations under the License. |
| |
| from __future__ import absolute_import, division, print_function |
| |
| import binascii |
| import collections |
| from contextlib import contextmanager |
| import re |
| |
| import pytest |
| |
| import six |
| |
| from cryptography.exceptions import UnsupportedAlgorithm |
| |
| import cryptography_vectors |
| |
| |
| HashVector = collections.namedtuple("HashVector", ["message", "digest"]) |
| KeyedHashVector = collections.namedtuple( |
| "KeyedHashVector", ["message", "digest", "key"] |
| ) |
| |
| |
| def select_backends(names, backend_list): |
| if names is None: |
| return backend_list |
| split_names = [x.strip() for x in names.split(',')] |
| # this must be duplicated and then removed to preserve the metadata |
| # pytest associates. Appending backends to a new list doesn't seem to work |
| selected_backends = [] |
| for backend in backend_list: |
| if backend.name in split_names: |
| selected_backends.append(backend) |
| |
| if len(selected_backends) > 0: |
| return selected_backends |
| else: |
| raise ValueError( |
| "No backend selected. Tried to select: {0}".format(split_names) |
| ) |
| |
| |
| def check_for_iface(name, iface, item): |
| if name in item.keywords and "backend" in item.funcargs: |
| if not isinstance(item.funcargs["backend"], iface): |
| pytest.skip("{0} backend does not support {1}".format( |
| item.funcargs["backend"], name |
| )) |
| |
| |
| def check_backend_support(item): |
| supported = item.keywords.get("supported") |
| if supported and "backend" in item.funcargs: |
| if not supported.kwargs["only_if"](item.funcargs["backend"]): |
| pytest.skip("{0} ({1})".format( |
| supported.kwargs["skip_message"], item.funcargs["backend"] |
| )) |
| elif supported: |
| raise ValueError("This mark is only available on methods that take a " |
| "backend") |
| |
| |
| @contextmanager |
| def raises_unsupported_algorithm(reason): |
| with pytest.raises(UnsupportedAlgorithm) as exc_info: |
| yield exc_info |
| |
| assert exc_info.value._reason is reason |
| |
| |
| def load_vectors_from_file(filename, loader): |
| with cryptography_vectors.open_vector_file(filename) as vector_file: |
| return loader(vector_file) |
| |
| |
| def load_nist_vectors(vector_data): |
| test_data = None |
| data = [] |
| |
| for line in vector_data: |
| line = line.strip() |
| |
| # Blank lines, comments, and section headers are ignored |
| if not line or line.startswith("#") or (line.startswith("[") |
| and line.endswith("]")): |
| continue |
| |
| if line.strip() == "FAIL": |
| test_data["fail"] = True |
| continue |
| |
| # Build our data using a simple Key = Value format |
| name, value = [c.strip() for c in line.split("=")] |
| |
| # Some tests (PBKDF2) contain \0, which should be interpreted as a |
| # null character rather than literal. |
| value = value.replace("\\0", "\0") |
| |
| # COUNT is a special token that indicates a new block of data |
| if name.upper() == "COUNT": |
| test_data = {} |
| data.append(test_data) |
| continue |
| # For all other tokens we simply want the name, value stored in |
| # the dictionary |
| else: |
| test_data[name.lower()] = value.encode("ascii") |
| |
| return data |
| |
| |
| def load_cryptrec_vectors(vector_data): |
| cryptrec_list = [] |
| |
| for line in vector_data: |
| line = line.strip() |
| |
| # Blank lines and comments are ignored |
| if not line or line.startswith("#"): |
| continue |
| |
| if line.startswith("K"): |
| key = line.split(" : ")[1].replace(" ", "").encode("ascii") |
| elif line.startswith("P"): |
| pt = line.split(" : ")[1].replace(" ", "").encode("ascii") |
| elif line.startswith("C"): |
| ct = line.split(" : ")[1].replace(" ", "").encode("ascii") |
| # after a C is found the K+P+C tuple is complete |
| # there are many P+C pairs for each K |
| cryptrec_list.append({ |
| "key": key, |
| "plaintext": pt, |
| "ciphertext": ct |
| }) |
| else: |
| raise ValueError("Invalid line in file '{}'".format(line)) |
| return cryptrec_list |
| |
| |
| def load_hash_vectors(vector_data): |
| vectors = [] |
| key = None |
| msg = None |
| md = None |
| |
| for line in vector_data: |
| line = line.strip() |
| |
| if not line or line.startswith("#") or line.startswith("["): |
| continue |
| |
| if line.startswith("Len"): |
| length = int(line.split(" = ")[1]) |
| elif line.startswith("Key"): |
| # HMAC vectors contain a key attribute. Hash vectors do not. |
| key = line.split(" = ")[1].encode("ascii") |
| elif line.startswith("Msg"): |
| # In the NIST vectors they have chosen to represent an empty |
| # string as hex 00, which is of course not actually an empty |
| # string. So we parse the provided length and catch this edge case. |
| msg = line.split(" = ")[1].encode("ascii") if length > 0 else b"" |
| elif line.startswith("MD"): |
| md = line.split(" = ")[1] |
| # after MD is found the Msg+MD (+ potential key) tuple is complete |
| if key is not None: |
| vectors.append(KeyedHashVector(msg, md, key)) |
| key = None |
| msg = None |
| md = None |
| else: |
| vectors.append(HashVector(msg, md)) |
| msg = None |
| md = None |
| else: |
| raise ValueError("Unknown line in hash vector") |
| return vectors |
| |
| |
| def load_pkcs1_vectors(vector_data): |
| """ |
| Loads data out of RSA PKCS #1 vector files. |
| """ |
| private_key_vector = None |
| public_key_vector = None |
| attr = None |
| key = None |
| example_vector = None |
| examples = [] |
| vectors = [] |
| for line in vector_data: |
| if ( |
| line.startswith("# PSS Example") or |
| line.startswith("# OAEP Example") or |
| line.startswith("# PKCS#1 v1.5") |
| ): |
| if example_vector: |
| for key, value in six.iteritems(example_vector): |
| hex_str = "".join(value).replace(" ", "").encode("ascii") |
| example_vector[key] = hex_str |
| examples.append(example_vector) |
| |
| attr = None |
| example_vector = collections.defaultdict(list) |
| |
| if line.startswith("# Message"): |
| attr = "message" |
| continue |
| elif line.startswith("# Salt"): |
| attr = "salt" |
| continue |
| elif line.startswith("# Seed"): |
| attr = "seed" |
| continue |
| elif line.startswith("# Signature"): |
| attr = "signature" |
| continue |
| elif line.startswith("# Encryption"): |
| attr = "encryption" |
| continue |
| elif ( |
| example_vector and |
| line.startswith("# =============================================") |
| ): |
| for key, value in six.iteritems(example_vector): |
| hex_str = "".join(value).replace(" ", "").encode("ascii") |
| example_vector[key] = hex_str |
| examples.append(example_vector) |
| example_vector = None |
| attr = None |
| elif example_vector and line.startswith("#"): |
| continue |
| else: |
| if attr is not None and example_vector is not None: |
| example_vector[attr].append(line.strip()) |
| continue |
| |
| if ( |
| line.startswith("# Example") or |
| line.startswith("# =============================================") |
| ): |
| if key: |
| assert private_key_vector |
| assert public_key_vector |
| |
| for key, value in six.iteritems(public_key_vector): |
| hex_str = "".join(value).replace(" ", "") |
| public_key_vector[key] = int(hex_str, 16) |
| |
| for key, value in six.iteritems(private_key_vector): |
| hex_str = "".join(value).replace(" ", "") |
| private_key_vector[key] = int(hex_str, 16) |
| |
| private_key_vector["examples"] = examples |
| examples = [] |
| |
| assert ( |
| private_key_vector['public_exponent'] == |
| public_key_vector['public_exponent'] |
| ) |
| |
| assert ( |
| private_key_vector['modulus'] == |
| public_key_vector['modulus'] |
| ) |
| |
| vectors.append( |
| (private_key_vector, public_key_vector) |
| ) |
| |
| public_key_vector = collections.defaultdict(list) |
| private_key_vector = collections.defaultdict(list) |
| key = None |
| attr = None |
| |
| if private_key_vector is None or public_key_vector is None: |
| continue |
| |
| if line.startswith("# Private key"): |
| key = private_key_vector |
| elif line.startswith("# Public key"): |
| key = public_key_vector |
| elif line.startswith("# Modulus:"): |
| attr = "modulus" |
| elif line.startswith("# Public exponent:"): |
| attr = "public_exponent" |
| elif line.startswith("# Exponent:"): |
| if key is public_key_vector: |
| attr = "public_exponent" |
| else: |
| assert key is private_key_vector |
| attr = "private_exponent" |
| elif line.startswith("# Prime 1:"): |
| attr = "p" |
| elif line.startswith("# Prime 2:"): |
| attr = "q" |
| elif line.startswith("# Prime exponent 1:"): |
| attr = "dmp1" |
| elif line.startswith("# Prime exponent 2:"): |
| attr = "dmq1" |
| elif line.startswith("# Coefficient:"): |
| attr = "iqmp" |
| elif line.startswith("#"): |
| attr = None |
| else: |
| if key is not None and attr is not None: |
| key[attr].append(line.strip()) |
| return vectors |
| |
| |
| def load_rsa_nist_vectors(vector_data): |
| test_data = None |
| p = None |
| salt_length = None |
| data = [] |
| |
| for line in vector_data: |
| line = line.strip() |
| |
| # Blank lines and section headers are ignored |
| if not line or line.startswith("["): |
| continue |
| |
| if line.startswith("# Salt len:"): |
| salt_length = int(line.split(":")[1].strip()) |
| continue |
| elif line.startswith("#"): |
| continue |
| |
| # Build our data using a simple Key = Value format |
| name, value = [c.strip() for c in line.split("=")] |
| |
| if name == "n": |
| n = int(value, 16) |
| elif name == "e" and p is None: |
| e = int(value, 16) |
| elif name == "p": |
| p = int(value, 16) |
| elif name == "q": |
| q = int(value, 16) |
| elif name == "SHAAlg": |
| if p is None: |
| test_data = { |
| "modulus": n, |
| "public_exponent": e, |
| "salt_length": salt_length, |
| "algorithm": value, |
| "fail": False |
| } |
| else: |
| test_data = { |
| "modulus": n, |
| "p": p, |
| "q": q, |
| "algorithm": value |
| } |
| if salt_length is not None: |
| test_data["salt_length"] = salt_length |
| data.append(test_data) |
| elif name == "e" and p is not None: |
| test_data["public_exponent"] = int(value, 16) |
| elif name == "d": |
| test_data["private_exponent"] = int(value, 16) |
| elif name == "Result": |
| test_data["fail"] = value.startswith("F") |
| # For all other tokens we simply want the name, value stored in |
| # the dictionary |
| else: |
| test_data[name.lower()] = value.encode("ascii") |
| |
| return data |
| |
| |
| def load_fips_dsa_key_pair_vectors(vector_data): |
| """ |
| Loads data out of the FIPS DSA KeyPair vector files. |
| """ |
| vectors = [] |
| # When reading_key_data is set to True it tells the loader to continue |
| # constructing dictionaries. We set reading_key_data to False during the |
| # blocks of the vectors of N=224 because we don't support it. |
| reading_key_data = True |
| for line in vector_data: |
| line = line.strip() |
| |
| if not line or line.startswith("#"): |
| continue |
| elif line.startswith("[mod = L=1024"): |
| continue |
| elif line.startswith("[mod = L=2048, N=224"): |
| reading_key_data = False |
| continue |
| elif line.startswith("[mod = L=2048, N=256"): |
| reading_key_data = True |
| continue |
| elif line.startswith("[mod = L=3072"): |
| continue |
| |
| if not reading_key_data: |
| continue |
| |
| elif reading_key_data: |
| if line.startswith("P"): |
| vectors.append({'p': int(line.split("=")[1], 16)}) |
| elif line.startswith("Q"): |
| vectors[-1]['q'] = int(line.split("=")[1], 16) |
| elif line.startswith("G"): |
| vectors[-1]['g'] = int(line.split("=")[1], 16) |
| elif line.startswith("X") and 'x' not in vectors[-1]: |
| vectors[-1]['x'] = int(line.split("=")[1], 16) |
| elif line.startswith("X") and 'x' in vectors[-1]: |
| vectors.append({'p': vectors[-1]['p'], |
| 'q': vectors[-1]['q'], |
| 'g': vectors[-1]['g'], |
| 'x': int(line.split("=")[1], 16) |
| }) |
| elif line.startswith("Y"): |
| vectors[-1]['y'] = int(line.split("=")[1], 16) |
| |
| return vectors |
| |
| |
| # http://tools.ietf.org/html/rfc4492#appendix-A |
| _ECDSA_CURVE_NAMES = { |
| "P-192": "secp192r1", |
| "P-224": "secp224r1", |
| "P-256": "secp256r1", |
| "P-384": "secp384r1", |
| "P-521": "secp521r1", |
| |
| "K-163": "sect163k1", |
| "K-233": "sect233k1", |
| "K-283": "sect283k1", |
| "K-409": "sect409k1", |
| "K-571": "sect571k1", |
| |
| "B-163": "sect163r2", |
| "B-233": "sect233r1", |
| "B-283": "sect283r1", |
| "B-409": "sect409r1", |
| "B-571": "sect571r1", |
| } |
| |
| |
| def load_fips_ecdsa_key_pair_vectors(vector_data): |
| """ |
| Loads data out of the FIPS ECDSA KeyPair vector files. |
| """ |
| vectors = [] |
| key_data = None |
| for line in vector_data: |
| line = line.strip() |
| |
| if not line or line.startswith("#"): |
| continue |
| |
| if line[1:-1] in _ECDSA_CURVE_NAMES: |
| curve_name = _ECDSA_CURVE_NAMES[line[1:-1]] |
| |
| elif line.startswith("d = "): |
| if key_data is not None: |
| vectors.append(key_data) |
| |
| key_data = { |
| "curve": curve_name, |
| "d": int(line.split("=")[1], 16) |
| } |
| |
| elif key_data is not None: |
| if line.startswith("Qx = "): |
| key_data["x"] = int(line.split("=")[1], 16) |
| elif line.startswith("Qy = "): |
| key_data["y"] = int(line.split("=")[1], 16) |
| |
| if key_data is not None: |
| vectors.append(key_data) |
| |
| return vectors |
| |
| |
| def load_fips_ecdsa_signing_vectors(vector_data): |
| """ |
| Loads data out of the FIPS ECDSA SigGen vector files. |
| """ |
| vectors = [] |
| |
| curve_rx = re.compile( |
| r"\[(?P<curve>[PKB]-[0-9]{3}),SHA-(?P<sha>1|224|256|384|512)\]" |
| ) |
| |
| data = None |
| for line in vector_data: |
| line = line.strip() |
| |
| if not line or line.startswith("#"): |
| continue |
| |
| curve_match = curve_rx.match(line) |
| if curve_match: |
| curve_name = _ECDSA_CURVE_NAMES[curve_match.group("curve")] |
| digest_name = "SHA-{0}".format(curve_match.group("sha")) |
| |
| elif line.startswith("Msg = "): |
| if data is not None: |
| vectors.append(data) |
| |
| hexmsg = line.split("=")[1].strip().encode("ascii") |
| |
| data = { |
| "curve": curve_name, |
| "digest_algorithm": digest_name, |
| "message": binascii.unhexlify(hexmsg) |
| } |
| |
| elif data is not None: |
| if line.startswith("Qx = "): |
| data["x"] = int(line.split("=")[1], 16) |
| elif line.startswith("Qy = "): |
| data["y"] = int(line.split("=")[1], 16) |
| elif line.startswith("R = "): |
| data["r"] = int(line.split("=")[1], 16) |
| elif line.startswith("S = "): |
| data["s"] = int(line.split("=")[1], 16) |
| elif line.startswith("d = "): |
| data["d"] = int(line.split("=")[1], 16) |
| |
| if data is not None: |
| vectors.append(data) |
| |
| return vectors |