| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929 |
- # This file is dual licensed under the terms of the Apache License, Version
- # 2.0, and the BSD License. See the LICENSE file in the root of this repository
- # for complete details.
- from __future__ import absolute_import, division, print_function
- import binascii
- import collections
- import json
- import math
- import os
- import re
- from contextlib import contextmanager
- import pytest
- import six
- from cryptography.exceptions import UnsupportedAlgorithm
- import cryptography_vectors
- HashVector = collections.namedtuple("HashVector", ["message", "digest"])
- KeyedHashVector = collections.namedtuple(
- "KeyedHashVector", ["message", "digest", "key"]
- )
- def check_backend_support(backend, item):
- for mark in item.node.iter_markers("supported"):
- if not mark.kwargs["only_if"](backend):
- pytest.skip("{} ({})".format(
- mark.kwargs["skip_message"], backend
- ))
- @contextmanager
- def raises_unsupported_algorithm(reason):
- with pytest.raises(UnsupportedAlgorithm) as exc_info:
- yield exc_info
- assert exc_info.value._reason is reason
- def load_vectors_from_file(filename, loader, mode="r"):
- with cryptography_vectors.open_vector_file(filename, mode) as vector_file:
- return loader(vector_file)
- def load_nist_vectors(vector_data):
- test_data = None
- data = []
- for line in vector_data:
- line = line.strip()
- # Blank lines, comments, and section headers are ignored
- if not line or line.startswith("#") or (line.startswith("[") and
- line.endswith("]")):
- continue
- if line.strip() == "FAIL":
- test_data["fail"] = True
- continue
- # Build our data using a simple Key = Value format
- name, value = [c.strip() for c in line.split("=")]
- # Some tests (PBKDF2) contain \0, which should be interpreted as a
- # null character rather than literal.
- value = value.replace("\\0", "\0")
- # COUNT is a special token that indicates a new block of data
- if name.upper() == "COUNT":
- test_data = {}
- data.append(test_data)
- continue
- # For all other tokens we simply want the name, value stored in
- # the dictionary
- else:
- test_data[name.lower()] = value.encode("ascii")
- return data
- def load_cryptrec_vectors(vector_data):
- cryptrec_list = []
- for line in vector_data:
- line = line.strip()
- # Blank lines and comments are ignored
- if not line or line.startswith("#"):
- continue
- if line.startswith("K"):
- key = line.split(" : ")[1].replace(" ", "").encode("ascii")
- elif line.startswith("P"):
- pt = line.split(" : ")[1].replace(" ", "").encode("ascii")
- elif line.startswith("C"):
- ct = line.split(" : ")[1].replace(" ", "").encode("ascii")
- # after a C is found the K+P+C tuple is complete
- # there are many P+C pairs for each K
- cryptrec_list.append({
- "key": key,
- "plaintext": pt,
- "ciphertext": ct
- })
- else:
- raise ValueError("Invalid line in file '{}'".format(line))
- return cryptrec_list
- def load_hash_vectors(vector_data):
- vectors = []
- key = None
- msg = None
- md = None
- for line in vector_data:
- line = line.strip()
- if not line or line.startswith("#") or line.startswith("["):
- continue
- if line.startswith("Len"):
- length = int(line.split(" = ")[1])
- elif line.startswith("Key"):
- # HMAC vectors contain a key attribute. Hash vectors do not.
- key = line.split(" = ")[1].encode("ascii")
- elif line.startswith("Msg"):
- # In the NIST vectors they have chosen to represent an empty
- # string as hex 00, which is of course not actually an empty
- # string. So we parse the provided length and catch this edge case.
- msg = line.split(" = ")[1].encode("ascii") if length > 0 else b""
- elif line.startswith("MD") or line.startswith("Output"):
- md = line.split(" = ")[1]
- # after MD is found the Msg+MD (+ potential key) tuple is complete
- if key is not None:
- vectors.append(KeyedHashVector(msg, md, key))
- key = None
- msg = None
- md = None
- else:
- vectors.append(HashVector(msg, md))
- msg = None
- md = None
- else:
- raise ValueError("Unknown line in hash vector")
- return vectors
- def load_pkcs1_vectors(vector_data):
- """
- Loads data out of RSA PKCS #1 vector files.
- """
- private_key_vector = None
- public_key_vector = None
- attr = None
- key = None
- example_vector = None
- examples = []
- vectors = []
- for line in vector_data:
- if (
- line.startswith("# PSS Example") or
- line.startswith("# OAEP Example") or
- line.startswith("# PKCS#1 v1.5")
- ):
- if example_vector:
- for key, value in six.iteritems(example_vector):
- hex_str = "".join(value).replace(" ", "").encode("ascii")
- example_vector[key] = hex_str
- examples.append(example_vector)
- attr = None
- example_vector = collections.defaultdict(list)
- if line.startswith("# Message"):
- attr = "message"
- continue
- elif line.startswith("# Salt"):
- attr = "salt"
- continue
- elif line.startswith("# Seed"):
- attr = "seed"
- continue
- elif line.startswith("# Signature"):
- attr = "signature"
- continue
- elif line.startswith("# Encryption"):
- attr = "encryption"
- continue
- elif (
- example_vector and
- line.startswith("# =============================================")
- ):
- for key, value in six.iteritems(example_vector):
- hex_str = "".join(value).replace(" ", "").encode("ascii")
- example_vector[key] = hex_str
- examples.append(example_vector)
- example_vector = None
- attr = None
- elif example_vector and line.startswith("#"):
- continue
- else:
- if attr is not None and example_vector is not None:
- example_vector[attr].append(line.strip())
- continue
- if (
- line.startswith("# Example") or
- line.startswith("# =============================================")
- ):
- if key:
- assert private_key_vector
- assert public_key_vector
- for key, value in six.iteritems(public_key_vector):
- hex_str = "".join(value).replace(" ", "")
- public_key_vector[key] = int(hex_str, 16)
- for key, value in six.iteritems(private_key_vector):
- hex_str = "".join(value).replace(" ", "")
- private_key_vector[key] = int(hex_str, 16)
- private_key_vector["examples"] = examples
- examples = []
- assert (
- private_key_vector['public_exponent'] ==
- public_key_vector['public_exponent']
- )
- assert (
- private_key_vector['modulus'] ==
- public_key_vector['modulus']
- )
- vectors.append(
- (private_key_vector, public_key_vector)
- )
- public_key_vector = collections.defaultdict(list)
- private_key_vector = collections.defaultdict(list)
- key = None
- attr = None
- if private_key_vector is None or public_key_vector is None:
- # Random garbage to defeat CPython's peephole optimizer so that
- # coverage records correctly: https://bugs.python.org/issue2506
- 1 + 1
- continue
- if line.startswith("# Private key"):
- key = private_key_vector
- elif line.startswith("# Public key"):
- key = public_key_vector
- elif line.startswith("# Modulus:"):
- attr = "modulus"
- elif line.startswith("# Public exponent:"):
- attr = "public_exponent"
- elif line.startswith("# Exponent:"):
- if key is public_key_vector:
- attr = "public_exponent"
- else:
- assert key is private_key_vector
- attr = "private_exponent"
- elif line.startswith("# Prime 1:"):
- attr = "p"
- elif line.startswith("# Prime 2:"):
- attr = "q"
- elif line.startswith("# Prime exponent 1:"):
- attr = "dmp1"
- elif line.startswith("# Prime exponent 2:"):
- attr = "dmq1"
- elif line.startswith("# Coefficient:"):
- attr = "iqmp"
- elif line.startswith("#"):
- attr = None
- else:
- if key is not None and attr is not None:
- key[attr].append(line.strip())
- return vectors
- def load_rsa_nist_vectors(vector_data):
- test_data = None
- p = None
- salt_length = None
- data = []
- for line in vector_data:
- line = line.strip()
- # Blank lines and section headers are ignored
- if not line or line.startswith("["):
- continue
- if line.startswith("# Salt len:"):
- salt_length = int(line.split(":")[1].strip())
- continue
- elif line.startswith("#"):
- continue
- # Build our data using a simple Key = Value format
- name, value = [c.strip() for c in line.split("=")]
- if name == "n":
- n = int(value, 16)
- elif name == "e" and p is None:
- e = int(value, 16)
- elif name == "p":
- p = int(value, 16)
- elif name == "q":
- q = int(value, 16)
- elif name == "SHAAlg":
- if p is None:
- test_data = {
- "modulus": n,
- "public_exponent": e,
- "salt_length": salt_length,
- "algorithm": value,
- "fail": False
- }
- else:
- test_data = {
- "modulus": n,
- "p": p,
- "q": q,
- "algorithm": value
- }
- if salt_length is not None:
- test_data["salt_length"] = salt_length
- data.append(test_data)
- elif name == "e" and p is not None:
- test_data["public_exponent"] = int(value, 16)
- elif name == "d":
- test_data["private_exponent"] = int(value, 16)
- elif name == "Result":
- test_data["fail"] = value.startswith("F")
- # For all other tokens we simply want the name, value stored in
- # the dictionary
- else:
- test_data[name.lower()] = value.encode("ascii")
- return data
- def load_fips_dsa_key_pair_vectors(vector_data):
- """
- Loads data out of the FIPS DSA KeyPair vector files.
- """
- vectors = []
- for line in vector_data:
- line = line.strip()
- if not line or line.startswith("#") or line.startswith("[mod"):
- continue
- if line.startswith("P"):
- vectors.append({'p': int(line.split("=")[1], 16)})
- elif line.startswith("Q"):
- vectors[-1]['q'] = int(line.split("=")[1], 16)
- elif line.startswith("G"):
- vectors[-1]['g'] = int(line.split("=")[1], 16)
- elif line.startswith("X") and 'x' not in vectors[-1]:
- vectors[-1]['x'] = int(line.split("=")[1], 16)
- elif line.startswith("X") and 'x' in vectors[-1]:
- vectors.append({'p': vectors[-1]['p'],
- 'q': vectors[-1]['q'],
- 'g': vectors[-1]['g'],
- 'x': int(line.split("=")[1], 16)
- })
- elif line.startswith("Y"):
- vectors[-1]['y'] = int(line.split("=")[1], 16)
- return vectors
- def load_fips_dsa_sig_vectors(vector_data):
- """
- Loads data out of the FIPS DSA SigVer vector files.
- """
- vectors = []
- sha_regex = re.compile(
- r"\[mod = L=...., N=..., SHA-(?P<sha>1|224|256|384|512)\]"
- )
- for line in vector_data:
- line = line.strip()
- if not line or line.startswith("#"):
- continue
- sha_match = sha_regex.match(line)
- if sha_match:
- digest_algorithm = "SHA-{}".format(sha_match.group("sha"))
- if line.startswith("[mod"):
- continue
- name, value = [c.strip() for c in line.split("=")]
- if name == "P":
- vectors.append({'p': int(value, 16),
- 'digest_algorithm': digest_algorithm})
- elif name == "Q":
- vectors[-1]['q'] = int(value, 16)
- elif name == "G":
- vectors[-1]['g'] = int(value, 16)
- elif name == "Msg" and 'msg' not in vectors[-1]:
- hexmsg = value.strip().encode("ascii")
- vectors[-1]['msg'] = binascii.unhexlify(hexmsg)
- elif name == "Msg" and 'msg' in vectors[-1]:
- hexmsg = value.strip().encode("ascii")
- vectors.append({'p': vectors[-1]['p'],
- 'q': vectors[-1]['q'],
- 'g': vectors[-1]['g'],
- 'digest_algorithm':
- vectors[-1]['digest_algorithm'],
- 'msg': binascii.unhexlify(hexmsg)})
- elif name == "X":
- vectors[-1]['x'] = int(value, 16)
- elif name == "Y":
- vectors[-1]['y'] = int(value, 16)
- elif name == "R":
- vectors[-1]['r'] = int(value, 16)
- elif name == "S":
- vectors[-1]['s'] = int(value, 16)
- elif name == "Result":
- vectors[-1]['result'] = value.split("(")[0].strip()
- return vectors
- # https://tools.ietf.org/html/rfc4492#appendix-A
- _ECDSA_CURVE_NAMES = {
- "P-192": "secp192r1",
- "P-224": "secp224r1",
- "P-256": "secp256r1",
- "P-384": "secp384r1",
- "P-521": "secp521r1",
- "K-163": "sect163k1",
- "K-233": "sect233k1",
- "K-256": "secp256k1",
- "K-283": "sect283k1",
- "K-409": "sect409k1",
- "K-571": "sect571k1",
- "B-163": "sect163r2",
- "B-233": "sect233r1",
- "B-283": "sect283r1",
- "B-409": "sect409r1",
- "B-571": "sect571r1",
- }
- def load_fips_ecdsa_key_pair_vectors(vector_data):
- """
- Loads data out of the FIPS ECDSA KeyPair vector files.
- """
- vectors = []
- key_data = None
- for line in vector_data:
- line = line.strip()
- if not line or line.startswith("#"):
- continue
- if line[1:-1] in _ECDSA_CURVE_NAMES:
- curve_name = _ECDSA_CURVE_NAMES[line[1:-1]]
- elif line.startswith("d = "):
- if key_data is not None:
- vectors.append(key_data)
- key_data = {
- "curve": curve_name,
- "d": int(line.split("=")[1], 16)
- }
- elif key_data is not None:
- if line.startswith("Qx = "):
- key_data["x"] = int(line.split("=")[1], 16)
- elif line.startswith("Qy = "):
- key_data["y"] = int(line.split("=")[1], 16)
- assert key_data is not None
- vectors.append(key_data)
- return vectors
- def load_fips_ecdsa_signing_vectors(vector_data):
- """
- Loads data out of the FIPS ECDSA SigGen vector files.
- """
- vectors = []
- curve_rx = re.compile(
- r"\[(?P<curve>[PKB]-[0-9]{3}),SHA-(?P<sha>1|224|256|384|512)\]"
- )
- data = None
- for line in vector_data:
- line = line.strip()
- curve_match = curve_rx.match(line)
- if curve_match:
- curve_name = _ECDSA_CURVE_NAMES[curve_match.group("curve")]
- digest_name = "SHA-{}".format(curve_match.group("sha"))
- elif line.startswith("Msg = "):
- if data is not None:
- vectors.append(data)
- hexmsg = line.split("=")[1].strip().encode("ascii")
- data = {
- "curve": curve_name,
- "digest_algorithm": digest_name,
- "message": binascii.unhexlify(hexmsg)
- }
- elif data is not None:
- if line.startswith("Qx = "):
- data["x"] = int(line.split("=")[1], 16)
- elif line.startswith("Qy = "):
- data["y"] = int(line.split("=")[1], 16)
- elif line.startswith("R = "):
- data["r"] = int(line.split("=")[1], 16)
- elif line.startswith("S = "):
- data["s"] = int(line.split("=")[1], 16)
- elif line.startswith("d = "):
- data["d"] = int(line.split("=")[1], 16)
- elif line.startswith("Result = "):
- data["fail"] = line.split("=")[1].strip()[0] == "F"
- assert data is not None
- vectors.append(data)
- return vectors
- def load_kasvs_dh_vectors(vector_data):
- """
- Loads data out of the KASVS key exchange vector data
- """
- result_rx = re.compile(r"([FP]) \(([0-9]+) -")
- vectors = []
- data = {
- "fail_z": False,
- "fail_agree": False
- }
- for line in vector_data:
- line = line.strip()
- if not line or line.startswith("#"):
- continue
- if line.startswith("P = "):
- data["p"] = int(line.split("=")[1], 16)
- elif line.startswith("Q = "):
- data["q"] = int(line.split("=")[1], 16)
- elif line.startswith("G = "):
- data["g"] = int(line.split("=")[1], 16)
- elif line.startswith("Z = "):
- z_hex = line.split("=")[1].strip().encode("ascii")
- data["z"] = binascii.unhexlify(z_hex)
- elif line.startswith("XstatCAVS = "):
- data["x1"] = int(line.split("=")[1], 16)
- elif line.startswith("YstatCAVS = "):
- data["y1"] = int(line.split("=")[1], 16)
- elif line.startswith("XstatIUT = "):
- data["x2"] = int(line.split("=")[1], 16)
- elif line.startswith("YstatIUT = "):
- data["y2"] = int(line.split("=")[1], 16)
- elif line.startswith("Result = "):
- result_str = line.split("=")[1].strip()
- match = result_rx.match(result_str)
- if match.group(1) == "F":
- if int(match.group(2)) in (5, 10):
- data["fail_z"] = True
- else:
- data["fail_agree"] = True
- vectors.append(data)
- data = {
- "p": data["p"],
- "q": data["q"],
- "g": data["g"],
- "fail_z": False,
- "fail_agree": False
- }
- return vectors
- def load_kasvs_ecdh_vectors(vector_data):
- """
- Loads data out of the KASVS key exchange vector data
- """
- curve_name_map = {
- "P-192": "secp192r1",
- "P-224": "secp224r1",
- "P-256": "secp256r1",
- "P-384": "secp384r1",
- "P-521": "secp521r1",
- }
- result_rx = re.compile(r"([FP]) \(([0-9]+) -")
- tags = []
- sets = {}
- vectors = []
- # find info in header
- for line in vector_data:
- line = line.strip()
- if line.startswith("#"):
- parm = line.split("Parameter set(s) supported:")
- if len(parm) == 2:
- names = parm[1].strip().split()
- for n in names:
- tags.append("[%s]" % n)
- break
- # Sets Metadata
- tag = None
- curve = None
- for line in vector_data:
- line = line.strip()
- if not line or line.startswith("#"):
- continue
- if line in tags:
- tag = line
- curve = None
- elif line.startswith("[Curve selected:"):
- curve = curve_name_map[line.split(':')[1].strip()[:-1]]
- if tag is not None and curve is not None:
- sets[tag.strip("[]")] = curve
- tag = None
- if len(tags) == len(sets):
- break
- # Data
- data = {
- "CAVS": {},
- "IUT": {},
- }
- tag = None
- for line in vector_data:
- line = line.strip()
- if not line or line.startswith("#"):
- continue
- if line.startswith("["):
- tag = line.split()[0][1:]
- elif line.startswith("COUNT = "):
- data["COUNT"] = int(line.split("=")[1])
- elif line.startswith("dsCAVS = "):
- data["CAVS"]["d"] = int(line.split("=")[1], 16)
- elif line.startswith("QsCAVSx = "):
- data["CAVS"]["x"] = int(line.split("=")[1], 16)
- elif line.startswith("QsCAVSy = "):
- data["CAVS"]["y"] = int(line.split("=")[1], 16)
- elif line.startswith("dsIUT = "):
- data["IUT"]["d"] = int(line.split("=")[1], 16)
- elif line.startswith("QsIUTx = "):
- data["IUT"]["x"] = int(line.split("=")[1], 16)
- elif line.startswith("QsIUTy = "):
- data["IUT"]["y"] = int(line.split("=")[1], 16)
- elif line.startswith("OI = "):
- data["OI"] = int(line.split("=")[1], 16)
- elif line.startswith("Z = "):
- data["Z"] = int(line.split("=")[1], 16)
- elif line.startswith("DKM = "):
- data["DKM"] = int(line.split("=")[1], 16)
- elif line.startswith("Result = "):
- result_str = line.split("=")[1].strip()
- match = result_rx.match(result_str)
- if match.group(1) == "F":
- data["fail"] = True
- else:
- data["fail"] = False
- data["errno"] = int(match.group(2))
- data["curve"] = sets[tag]
- vectors.append(data)
- data = {
- "CAVS": {},
- "IUT": {},
- }
- return vectors
- def load_x963_vectors(vector_data):
- """
- Loads data out of the X9.63 vector data
- """
- vectors = []
- # Sets Metadata
- hashname = None
- vector = {}
- for line in vector_data:
- line = line.strip()
- if line.startswith("[SHA"):
- hashname = line[1:-1]
- shared_secret_len = 0
- shared_info_len = 0
- key_data_len = 0
- elif line.startswith("[shared secret length"):
- shared_secret_len = int(line[1:-1].split("=")[1].strip())
- elif line.startswith("[SharedInfo length"):
- shared_info_len = int(line[1:-1].split("=")[1].strip())
- elif line.startswith("[key data length"):
- key_data_len = int(line[1:-1].split("=")[1].strip())
- elif line.startswith("COUNT"):
- count = int(line.split("=")[1].strip())
- vector["hash"] = hashname
- vector["count"] = count
- vector["shared_secret_length"] = shared_secret_len
- vector["sharedinfo_length"] = shared_info_len
- vector["key_data_length"] = key_data_len
- elif line.startswith("Z"):
- vector["Z"] = line.split("=")[1].strip()
- assert math.ceil(shared_secret_len / 8) * 2 == len(vector["Z"])
- elif line.startswith("SharedInfo"):
- if shared_info_len != 0:
- vector["sharedinfo"] = line.split("=")[1].strip()
- silen = len(vector["sharedinfo"])
- assert math.ceil(shared_info_len / 8) * 2 == silen
- elif line.startswith("key_data"):
- vector["key_data"] = line.split("=")[1].strip()
- assert math.ceil(key_data_len / 8) * 2 == len(vector["key_data"])
- vectors.append(vector)
- vector = {}
- return vectors
- def load_nist_kbkdf_vectors(vector_data):
- """
- Load NIST SP 800-108 KDF Vectors
- """
- vectors = []
- test_data = None
- tag = {}
- for line in vector_data:
- line = line.strip()
- if not line or line.startswith("#"):
- continue
- if line.startswith("[") and line.endswith("]"):
- tag_data = line[1:-1]
- name, value = [c.strip() for c in tag_data.split("=")]
- if value.endswith('_BITS'):
- value = int(value.split('_')[0])
- tag.update({name.lower(): value})
- continue
- tag.update({name.lower(): value.lower()})
- elif line.startswith("COUNT="):
- test_data = {}
- test_data.update(tag)
- vectors.append(test_data)
- elif line.startswith("L"):
- name, value = [c.strip() for c in line.split("=")]
- test_data[name.lower()] = int(value)
- else:
- name, value = [c.strip() for c in line.split("=")]
- test_data[name.lower()] = value.encode("ascii")
- return vectors
- def load_ed25519_vectors(vector_data):
- data = []
- for line in vector_data:
- secret_key, public_key, message, signature, _ = line.split(':')
- # In the vectors the first element is secret key + public key
- secret_key = secret_key[0:64]
- # In the vectors the signature section is signature + message
- signature = signature[0:128]
- data.append({
- "secret_key": secret_key,
- "public_key": public_key,
- "message": message,
- "signature": signature
- })
- return data
- def load_nist_ccm_vectors(vector_data):
- test_data = None
- section_data = None
- global_data = {}
- new_section = False
- data = []
- for line in vector_data:
- line = line.strip()
- # Blank lines and comments should be ignored
- if not line or line.startswith("#"):
- continue
- # Some of the CCM vectors have global values for this. They are always
- # at the top before the first section header (see: VADT, VNT, VPT)
- if line.startswith(("Alen", "Plen", "Nlen", "Tlen")):
- name, value = [c.strip() for c in line.split("=")]
- global_data[name.lower()] = int(value)
- continue
- # section headers contain length data we might care about
- if line.startswith("["):
- new_section = True
- section_data = {}
- section = line[1:-1]
- items = [c.strip() for c in section.split(",")]
- for item in items:
- name, value = [c.strip() for c in item.split("=")]
- section_data[name.lower()] = int(value)
- continue
- name, value = [c.strip() for c in line.split("=")]
- if name.lower() in ("key", "nonce") and new_section:
- section_data[name.lower()] = value.encode("ascii")
- continue
- new_section = False
- # Payload is sometimes special because these vectors are absurd. Each
- # example may or may not have a payload. If it does not then the
- # previous example's payload should be used. We accomplish this by
- # writing it into the section_data. Because we update each example
- # with the section data it will be overwritten if a new payload value
- # is present. NIST should be ashamed of their vector creation.
- if name.lower() == "payload":
- section_data[name.lower()] = value.encode("ascii")
- # Result is a special token telling us if the test should pass/fail.
- # This is only present in the DVPT CCM tests
- if name.lower() == "result":
- if value.lower() == "pass":
- test_data["fail"] = False
- else:
- test_data["fail"] = True
- continue
- # COUNT is a special token that indicates a new block of data
- if name.lower() == "count":
- test_data = {}
- test_data.update(global_data)
- test_data.update(section_data)
- data.append(test_data)
- continue
- # For all other tokens we simply want the name, value stored in
- # the dictionary
- else:
- test_data[name.lower()] = value.encode("ascii")
- return data
- class WycheproofTest(object):
- def __init__(self, testgroup, testcase):
- self.testgroup = testgroup
- self.testcase = testcase
- def __repr__(self):
- return "<WycheproofTest({!r}, {!r}, tcId={})>".format(
- self.testgroup, self.testcase, self.testcase["tcId"],
- )
- @property
- def valid(self):
- return self.testcase["result"] == "valid"
- @property
- def acceptable(self):
- return self.testcase["result"] == "acceptable"
- @property
- def invalid(self):
- return self.testcase["result"] == "invalid"
- def has_flag(self, flag):
- return flag in self.testcase["flags"]
- def skip_if_wycheproof_none(wycheproof):
- # This is factored into its own function so we can easily test both
- # branches
- if wycheproof is None:
- pytest.skip("--wycheproof-root not provided")
- def load_wycheproof_tests(wycheproof, test_file):
- path = os.path.join(wycheproof, "testvectors", test_file)
- with open(path) as f:
- data = json.load(f)
- for group in data["testGroups"]:
- cases = group.pop("tests")
- for c in cases:
- yield WycheproofTest(group, c)
|