Requantization of a Q5_K_M quant of a trending 70b model without better quant/fp16 available, this through a Q8_0 intermediary step. Model has a theta of 1,000,000, and not 10,000, like Llama 2 models usually have. So, no Alpha or Rope Base Frequency up to its base 32k context, if it works as intended.