Skip to content
Snippets Groups Projects
EncGOP.cpp 156 KiB
Newer Older
  • Learn to ignore specific revisions
  •                          std::list<PelUnitBuf*>&   rcListPicYuvRecOut,
                             int                       iNumPicRcvd,
                             int                       iTimeOffset,
                             Picture*&                 rpcPic,
                             int                       pocCurr,
                             bool                      isField )
    {
      int i;
      //  Rec. output
      std::list<PelUnitBuf*>::iterator     iterPicYuvRec = rcListPicYuvRecOut.end();
    
      if (isField && pocCurr > 1 && m_iGopSize!=1)
      {
        iTimeOffset--;
      }
    
    
      int multipleFactor = m_pcCfg->getUseCompositeRef() ? 2 : 1;
      for (i = 0; i < (iNumPicRcvd * multipleFactor - iTimeOffset + 1); i += multipleFactor)
    
      {
        iterPicYuvRec--;
      }
    
      //  Current pic.
      PicList::iterator        iterPic       = rcListPic.begin();
      while (iterPic != rcListPic.end())
      {
        rpcPic = *(iterPic);
        if (rpcPic->getPOC() == pocCurr)
        {
          break;
        }
        iterPic++;
      }
    
      CHECK(!(rpcPic != NULL), "Unspecified error");
      CHECK(!(rpcPic->getPOC() == pocCurr), "Unspecified error");
    
      (**iterPicYuvRec) = rpcPic->getRecoBuf();
      return;
    }
    
    #if ENABLE_QPA
    
    #ifndef BETA
    
      #define BETA 0.5 // value between 0.0 and 1; use 0.0 to obtain traditional PSNR
    
    static inline double calcWeightedSquaredError(const CPelBuf& org,        const CPelBuf& rec,
                                                  double &sumAct,            const uint32_t bitDepth,
                                                  const uint32_t imageWidth, const uint32_t imageHeight,
                                                  const uint32_t offsetX,    const uint32_t offsetY,
                                                  int blockWidth,            int blockHeight)
    
    {
      const int    O = org.stride;
      const int    R = rec.stride;
      const Pel   *o = org.bufAt(offsetX, offsetY);
      const Pel   *r = rec.bufAt(offsetX, offsetY);
      const int yAct = offsetY > 0 ? 0 : 1;
      const int xAct = offsetX > 0 ? 0 : 1;
    
      if (offsetY + (uint32_t)blockHeight > imageHeight) blockHeight = imageHeight - offsetY;
      if (offsetX + (uint32_t)blockWidth  > imageWidth ) blockWidth  = imageWidth  - offsetX;
    
      const int hAct = offsetY + (uint32_t)blockHeight < imageHeight ? blockHeight : blockHeight - 1;
      const int wAct = offsetX + (uint32_t)blockWidth  < imageWidth  ? blockWidth  : blockWidth  - 1;
    
      uint64_t ssErr = 0; // sum of squared diffs
      uint64_t saAct = 0; // sum of abs. activity
    
      double msAct;
      int x, y;
    
      // calculate image differences and activity
      for (y = 0; y < blockHeight; y++)  // error
      {
        for (x = 0; x < blockWidth; x++)
        {
    
          const     int64_t iDiff = (int64_t)o[y*O + x] - (int64_t)r[y*R + x];
    
          ssErr += uint64_t(iDiff * iDiff);
        }
      }
      if (wAct <= xAct || hAct <= yAct) return (double)ssErr;
    
      for (y = yAct; y < hAct; y++)   // activity
      {
        for (x = xAct; x < wAct; x++)
        {
    
          const int f = 12 * (int)o[y*O + x] - 2 * ((int)o[y*O + x-1] + (int)o[y*O + x+1] + (int)o[(y-1)*O + x] + (int)o[(y+1)*O + x])
                           - (int)o[(y-1)*O + x-1] - (int)o[(y-1)*O + x+1] - (int)o[(y+1)*O + x-1] - (int)o[(y+1)*O + x+1];
          saAct += abs(f);
    
        }
      }
    
      // calculate weight (mean squared activity)
      msAct = (double)saAct / (double(wAct - xAct) * double(hAct - yAct));
    
    
      // lower limit, accounts for high-pass gain
      if (msAct < double(1 << (bitDepth - 4))) msAct = double(1 << (bitDepth - 4));
    
    
      msAct *= msAct; // because ssErr is squared
    
      sumAct += msAct; // includes high-pass gain
    
      // calculate activity weighted error square
      return (double)ssErr * pow(msAct, -1.0 * BETA);
    }
    #endif // ENABLE_QPA
    
    uint64_t EncGOP::xFindDistortionPlane(const CPelBuf& pic0, const CPelBuf& pic1, const uint32_t rshift
    #if ENABLE_QPA
    
                                        , const uint32_t chromaShift /*= 0*/
    
    {
      uint64_t uiTotalDiff;
      const  Pel*  pSrc0 = pic0.bufAt(0, 0);
      const  Pel*  pSrc1 = pic1.bufAt(0, 0);
    
      CHECK(pic0.width  != pic1.width , "Unspecified error");
      CHECK(pic0.height != pic1.height, "Unspecified error");
    
      if( rshift > 0 )
      {
    #if ENABLE_QPA
        const   uint32_t  BD = rshift;      // image bit-depth
        if (BD >= 8)
        {
          const uint32_t   W = pic0.width;  // image width
          const uint32_t   H = pic0.height; // image height
    
          const double     R = double(W * H) / (1920.0 * 1080.0);
    
          const uint32_t   B = Clip3<uint32_t>(0, 128 >> chromaShift, 4 * uint32_t(16.0 * sqrt(R) + 0.5)); // WPSNR block size in integer multiple of 4 (for SIMD, = 64 at full-HD)
    
          uint32_t x, y;
    
          if (B < 4) // image is too small to use WPSNR, resort to traditional PSNR
          {
            uiTotalDiff = 0;
            for (y = 0; y < H; y++)
            {
              for (x = 0; x < W; x++)
              {
    
                const           int64_t iDiff = (int64_t)pSrc0[x] - (int64_t)pSrc1[x];
    
                uiTotalDiff += uint64_t(iDiff * iDiff);
              }
              pSrc0 += pic0.stride;
              pSrc1 += pic1.stride;
            }
            return uiTotalDiff;
          }
    
          double wmse = 0.0, sumAct = 0.0; // compute activity normalized SNR value
    
              wmse += calcWeightedSquaredError(pic1,   pic0,
                                               sumAct, BD,
                                               W,      H,
                                               x,      y,
                                               B,      B);
    
          sumAct = 16.0 * sqrt ((3840.0 * 2160.0) / double((W << chromaShift) * (H << chromaShift))) * double(1 << BD);
    
          return (wmse <= 0.0) ? 0 : uint64_t(wmse * pow(sumAct, BETA) + 0.5);
        }
    #endif // ENABLE_QPA
        uiTotalDiff = 0;
        for (int y = 0; y < pic0.height; y++)
        {
          for (int x = 0; x < pic0.width; x++)
          {
            Intermediate_Int iTemp = pSrc0[x] - pSrc1[x];
            uiTotalDiff += uint64_t((iTemp * iTemp) >> rshift);
          }
          pSrc0 += pic0.stride;
          pSrc1 += pic1.stride;
        }
      }
      else
      {
        uiTotalDiff = 0;
        for (int y = 0; y < pic0.height; y++)
        {
          for (int x = 0; x < pic0.width; x++)
          {
            Intermediate_Int iTemp = pSrc0[x] - pSrc1[x];
            uiTotalDiff += uint64_t(iTemp * iTemp);
          }
          pSrc0 += pic0.stride;
          pSrc1 += pic1.stride;
        }
      }
    
      return uiTotalDiff;
    }
    #if WCG_WPSNR
    
    double EncGOP::xFindDistortionPlaneWPSNR(const CPelBuf& pic0, const CPelBuf& pic1, const uint32_t rshift, const CPelBuf& picLuma0,
    
    Taoran Lu's avatar
    Taoran Lu committed
      const bool    useLumaWPSNR = m_pcEncLib->getLumaLevelToDeltaQPMapping().isEnabled() || (m_pcCfg->getReshaper() && m_pcCfg->getReshapeSignalType() == RESHAPE_SIGNAL_PQ);
    
      if (!useLumaWPSNR)
      {
        return 0;
      }
    
      double uiTotalDiffWPSNR;
      const  Pel*  pSrc0 = pic0.bufAt(0, 0);
      const  Pel*  pSrc1 = pic1.bufAt(0, 0);
      const  Pel*  pSrcLuma = picLuma0.bufAt(0, 0);
      CHECK(pic0.width  != pic1.width , "Unspecified error");
      CHECK(pic0.height != pic1.height, "Unspecified error");
    
      if( rshift > 0 )
      {
        uiTotalDiffWPSNR = 0;
        for (int y = 0; y < pic0.height; y++)
        {
          for (int x = 0; x < pic0.width; x++)
          {
            Intermediate_Int iTemp = pSrc0[x] - pSrc1[x];
            double dW = m_pcEncLib->getRdCost()->getWPSNRLumaLevelWeight(pSrcLuma[(x << getComponentScaleX(compID, chfmt))]);
            uiTotalDiffWPSNR += ((dW * (double)iTemp * (double)iTemp)) * (double)(1 >> rshift);
          }
          pSrc0 += pic0.stride;
          pSrc1 += pic1.stride;
          pSrcLuma += picLuma0.stride << getComponentScaleY(compID, chfmt);
        }
      }
      else
      {
        uiTotalDiffWPSNR = 0;
        for (int y = 0; y < pic0.height; y++)
        {
          for (int x = 0; x < pic0.width; x++)
          {
            Intermediate_Int iTemp = pSrc0[x] - pSrc1[x];
            double dW = m_pcEncLib->getRdCost()->getWPSNRLumaLevelWeight(pSrcLuma[x << getComponentScaleX(compID, chfmt)]);
            uiTotalDiffWPSNR += dW * (double)iTemp * (double)iTemp;
          }
          pSrc0 += pic0.stride;
          pSrc1 += pic1.stride;
          pSrcLuma += picLuma0.stride << getComponentScaleY(compID, chfmt);
        }
      }
    
      return uiTotalDiffWPSNR;
    }
    #endif
    
    
    void EncGOP::xCalculateAddPSNRs( const bool isField, const bool isFieldTopFieldFirst, const int iGOPid, Picture* pcPic, const AccessUnit&accessUnit, PicList &rcListPic, const int64_t dEncTime, const InputColourSpaceConversion snr_conversion, const bool printFrameMSE, double* PSNR_Y
                                   , bool isEncodeLtRef
    )
    
      xCalculateAddPSNR(pcPic, pcPic->getRecoBuf(), accessUnit, (double)dEncTime, snr_conversion, printFrameMSE, PSNR_Y
                      , isEncodeLtRef
      );
    
    
      //In case of field coding, compute the interlaced PSNR for both fields
      if(isField)
      {
        bool bothFieldsAreEncoded = false;
        int correspondingFieldPOC = pcPic->getPOC();
        int currentPicGOPPoc = m_pcCfg->getGOPEntry(iGOPid).m_POC;
        if(pcPic->getPOC() == 0)
        {
          // particular case for POC 0 and 1.
          // If they are not encoded first and separately from other pictures, we need to change this
          // POC 0 is always encoded first then POC 1 is encoded
          bothFieldsAreEncoded = false;
        }
        else if(pcPic->getPOC() == 1)
        {
          // if we are at POC 1, POC 0 has been encoded for sure
          correspondingFieldPOC = 0;
          bothFieldsAreEncoded = true;
        }
        else
        {
          if(pcPic->getPOC()%2 == 1)
          {
            correspondingFieldPOC -= 1; // all odd POC are associated with the preceding even POC (e.g poc 1 is associated to poc 0)
            currentPicGOPPoc      -= 1;
          }
          else
          {
            correspondingFieldPOC += 1; // all even POC are associated with the following odd POC (e.g poc 0 is associated to poc 1)
            currentPicGOPPoc      += 1;
          }
          for(int i = 0; i < m_iGopSize; i ++)
          {
            if(m_pcCfg->getGOPEntry(i).m_POC == currentPicGOPPoc)
            {
              bothFieldsAreEncoded = m_pcCfg->getGOPEntry(i).m_isEncoded;
              break;
            }
          }
        }
    
        if(bothFieldsAreEncoded)
        {
          //get complementary top field
          PicList::iterator   iterPic = rcListPic.begin();
          while ((*iterPic)->getPOC() != correspondingFieldPOC)
          {
            iterPic ++;
          }
          Picture* correspondingFieldPic = *(iterPic);
    
    
          if ((pcPic->topField && isFieldTopFieldFirst) || (!pcPic->topField && !isFieldTopFieldFirst))
    
            xCalculateInterlacedAddPSNR(pcPic, correspondingFieldPic, pcPic->getRecoBuf(), correspondingFieldPic->getRecoBuf(), snr_conversion, printFrameMSE, PSNR_Y
              , isEncodeLtRef
            );
    
            xCalculateInterlacedAddPSNR(correspondingFieldPic, pcPic, correspondingFieldPic->getRecoBuf(), pcPic->getRecoBuf(), snr_conversion, printFrameMSE, PSNR_Y
              , isEncodeLtRef
            );
    
    void EncGOP::xCalculateAddPSNR(Picture* pcPic, PelUnitBuf cPicD, const AccessUnit& accessUnit, double dEncTime, const InputColourSpaceConversion conversion, const bool printFrameMSE, double* PSNR_Y
                                  , bool isEncodeLtRef
    )
    
    {
      const SPS&         sps = *pcPic->cs->sps;
      const CPelUnitBuf& pic = cPicD;
      CHECK(!(conversion == IPCOLOURSPACE_UNCHANGED), "Unspecified error");
    //  const CPelUnitBuf& org = (conversion != IPCOLOURSPACE_UNCHANGED) ? pcPic->getPicYuvTrueOrg()->getBuf() : pcPic->getPicYuvOrg()->getBuf();
    
    Taoran Lu's avatar
    Taoran Lu committed
      const CPelUnitBuf& org = sps.getUseReshaper() ? pcPic->getTrueOrigBuf() : pcPic->getOrigBuf();
    
    #if ENABLE_QPA
      const bool    useWPSNR = m_pcEncLib->getUseWPSNR();
    #endif
      double  dPSNR[MAX_NUM_COMPONENT];
    #if WCG_WPSNR
    
    Taoran Lu's avatar
    Taoran Lu committed
      const bool    useLumaWPSNR = m_pcEncLib->getLumaLevelToDeltaQPMapping().isEnabled() || (m_pcCfg->getReshaper() && m_pcCfg->getReshapeSignalType() == RESHAPE_SIGNAL_PQ);
    
      double  dPSNRWeighted[MAX_NUM_COMPONENT];
      double  MSEyuvframeWeighted[MAX_NUM_COMPONENT];
    #endif
      for(int i=0; i<MAX_NUM_COMPONENT; i++)
      {
        dPSNR[i]=0.0;
    #if WCG_WPSNR
        dPSNRWeighted[i]=0.0;
        MSEyuvframeWeighted[i] = 0.0;
    #endif
      }
    
      #if JVET_O0756_CALCULATE_HDRMETRICS
      double dDeltaE[hdrtoolslib::NB_REF_WHITE];
      double dPSNRL[hdrtoolslib::NB_REF_WHITE];
      for (int i=0; i<hdrtoolslib::NB_REF_WHITE; i++) {
        dDeltaE[i] = 0.0;
        dPSNRL[i] = 0.0;
      }
      #endif
      
    
      PelStorage interm;
    
      if (conversion != IPCOLOURSPACE_UNCHANGED)
      {
        interm.create(pic.chromaFormat, Area(Position(), pic.Y()));
        VideoIOYuv::ColourSpaceConvert(pic, interm, conversion, false);
      }
    
      const CPelUnitBuf& picC = (conversion == IPCOLOURSPACE_UNCHANGED) ? pic : interm;
    
      //===== calculate PSNR =====
      double MSEyuvframe[MAX_NUM_COMPONENT] = {0, 0, 0};
      const ChromaFormat formatD = pic.chromaFormat;
      const ChromaFormat format  = sps.getChromaFormatIdc();
    
      const bool bPicIsField     = pcPic->fieldPic;
      const Slice*  pcSlice      = pcPic->slices[0];
    
      for (int comp = 0; comp < ::getNumberValidComponents(formatD); comp++)
      {
        const ComponentID compID = ComponentID(comp);
        const CPelBuf&    p = picC.get(compID);
        const CPelBuf&    o = org.get(compID);
    
        CHECK(!( p.width  == o.width), "Unspecified error");
        CHECK(!( p.height == o.height), "Unspecified error");
    
        const uint32_t   width  = p.width  - (m_pcEncLib->getPad(0) >> ::getComponentScaleX(compID, format));
        const uint32_t   height = p.height - (m_pcEncLib->getPad(1) >> (!!bPicIsField+::getComponentScaleY(compID,format)));
    
        // create new buffers with correct dimensions
        const CPelBuf recPB(p.bufAt(0, 0), p.stride, width, height);
        const CPelBuf orgPB(o.bufAt(0, 0), o.stride, width, height);
        const uint32_t    bitDepth = sps.getBitDepth(toChannelType(compID));
    #if ENABLE_QPA
        const uint64_t uiSSDtemp = xFindDistortionPlane(recPB, orgPB, useWPSNR ? bitDepth : 0, ::getComponentScaleX(compID, format));
    #else
        const uint64_t uiSSDtemp = xFindDistortionPlane(recPB, orgPB, 0);
    #endif
    
        const uint32_t size   = width * height;
        const double fRefValue = (double)maxval * maxval * size;
        dPSNR[comp]       = uiSSDtemp ? 10.0 * log10(fRefValue / (double)uiSSDtemp) : 999.99;
        MSEyuvframe[comp] = (double)uiSSDtemp / size;
    #if WCG_WPSNR
    
        const double uiSSDtempWeighted = xFindDistortionPlaneWPSNR(recPB, orgPB, 0, org.get(COMPONENT_Y), compID, format);
    
        if (useLumaWPSNR)
        {
          dPSNRWeighted[comp] = uiSSDtempWeighted ? 10.0 * log10(fRefValue / (double)uiSSDtempWeighted) : 999.99;
          MSEyuvframeWeighted[comp] = (double)uiSSDtempWeighted / size;
        }
    #endif
      }
    
    #if EXTENSION_360_VIDEO
      m_ext360.calculatePSNRs(pcPic);
    #endif
    
      
    #if JVET_O0756_CALCULATE_HDRMETRICS
      const bool calculateHdrMetrics = m_pcEncLib->getCalcluateHdrMetrics();
      if(calculateHdrMetrics)
      {
        auto beforeTime = std::chrono::steady_clock::now();
        xCalculateHDRMetrics(pcPic, dDeltaE, dPSNRL);
        auto elapsed = std::chrono::steady_clock::now() - beforeTime;
        m_metricTime += elapsed;
      }
    #endif
      
    
      /* calculate the size of the access unit, excluding:
       *  - any AnnexB contributions (start_code_prefix, zero_byte, etc.,)
       *  - SEI NAL units
       */
      uint32_t numRBSPBytes = 0;
      for (AccessUnit::const_iterator it = accessUnit.begin(); it != accessUnit.end(); it++)
      {
        uint32_t numRBSPBytes_nal = uint32_t((*it)->m_nalUnitData.str().size());
        if (m_pcCfg->getSummaryVerboseness() > 0)
        {
          msg( NOTICE, "*** %6s numBytesInNALunit: %u\n", nalUnitTypeToString((*it)->m_nalUnitType), numRBSPBytes_nal);
        }
        if( ( *it )->m_nalUnitType != NAL_UNIT_PREFIX_SEI && ( *it )->m_nalUnitType != NAL_UNIT_SUFFIX_SEI )
        {
          numRBSPBytes += numRBSPBytes_nal;
    
    Lidong Xu's avatar
    Lidong Xu committed
          if (it == accessUnit.begin() || (*it)->m_nalUnitType == NAL_UNIT_VPS || (*it)->m_nalUnitType == NAL_UNIT_DPS || (*it)->m_nalUnitType == NAL_UNIT_SPS || (*it)->m_nalUnitType == NAL_UNIT_PPS)
    
          {
            numRBSPBytes += 4;
          }
          else
          {
            numRBSPBytes += 3;
          }
        }
      }
    
      uint32_t uibits = numRBSPBytes * 8;
      m_vRVM_RP.push_back( uibits );
    
      //===== add PSNR =====
    
      m_gcAnalyzeAll.addResult(dPSNR, (double)uibits, MSEyuvframe
        , isEncodeLtRef
      );
    
    #if EXTENSION_360_VIDEO
      m_ext360.addResult(m_gcAnalyzeAll);
    #endif
    
      #if JVET_O0756_CALCULATE_HDRMETRICS
      if(calculateHdrMetrics){
          m_gcAnalyzeAll.addHDRMetricsResult(dDeltaE, dPSNRL);
      }
      #endif
    
        m_gcAnalyzeI.addResult(dPSNR, (double)uibits, MSEyuvframe
          , isEncodeLtRef
        );
    
        *PSNR_Y = dPSNR[COMPONENT_Y];
    #if EXTENSION_360_VIDEO
        m_ext360.addResult(m_gcAnalyzeI);
    #endif
    
        #if JVET_O0756_CALCULATE_HDRMETRICS
        if(calculateHdrMetrics){
              m_gcAnalyzeI.addHDRMetricsResult(dDeltaE, dPSNRL);
        }
        #endif
    
        m_gcAnalyzeP.addResult(dPSNR, (double)uibits, MSEyuvframe
          , isEncodeLtRef
        );
    
        *PSNR_Y = dPSNR[COMPONENT_Y];
    #if EXTENSION_360_VIDEO
        m_ext360.addResult(m_gcAnalyzeP);
    #endif
    
        #if JVET_O0756_CALCULATE_HDRMETRICS
        if(calculateHdrMetrics){
              m_gcAnalyzeP.addHDRMetricsResult(dDeltaE, dPSNRL);
        }
        #endif
    
        m_gcAnalyzeB.addResult(dPSNR, (double)uibits, MSEyuvframe
          , isEncodeLtRef
        );
    
        *PSNR_Y = dPSNR[COMPONENT_Y];
    #if EXTENSION_360_VIDEO
        m_ext360.addResult(m_gcAnalyzeB);
    #endif
    
        #if JVET_O0756_CALCULATE_HDRMETRICS
        if(calculateHdrMetrics){
              m_gcAnalyzeB.addHDRMetricsResult(dDeltaE, dPSNRL);
        }
        #endif
    
        m_gcAnalyzeWPSNR.addResult(dPSNRWeighted, (double)uibits, MSEyuvframeWeighted, isEncodeLtRef);
    
      }
    #endif
    
      char c = (pcSlice->isIntra() ? 'I' : pcSlice->isInterP() ? 'P' : 'B');
      if (! pcPic->referenced)
      {
        c += 32;
      }
    
      if( g_verbosity >= NOTICE )
      {
        msg( NOTICE, "POC %4d TId: %1d ( %c-SLICE, QP %d ) %10d bits",
    
             pcSlice->getPOC(),
    
             pcSlice->getTLayer(),
             c,
             pcSlice->getSliceQp(),
             uibits );
    
        msg( NOTICE, " [Y %6.4lf dB    U %6.4lf dB    V %6.4lf dB]", dPSNR[COMPONENT_Y], dPSNR[COMPONENT_Cb], dPSNR[COMPONENT_Cr] );
    
    
    #if EXTENSION_360_VIDEO
        m_ext360.printPerPOCInfo(NOTICE);
    #endif
    
    
        if (m_pcEncLib->getPrintHexPsnr())
        {
          uint64_t xPsnr[MAX_NUM_COMPONENT];
          for (int i = 0; i < MAX_NUM_COMPONENT; i++)
          {
            copy(reinterpret_cast<uint8_t *>(&dPSNR[i]),
                 reinterpret_cast<uint8_t *>(&dPSNR[i]) + sizeof(dPSNR[i]),
                 reinterpret_cast<uint8_t *>(&xPsnr[i]));
          }
          msg(NOTICE, " [xY %16" PRIx64 " xU %16" PRIx64 " xV %16" PRIx64 "]", xPsnr[COMPONENT_Y], xPsnr[COMPONENT_Cb], xPsnr[COMPONENT_Cr]);
    
    #if EXTENSION_360_VIDEO
    
          m_ext360.printPerPOCInfo(NOTICE, true);
    
    
        if( printFrameMSE )
        {
          msg( NOTICE, " [Y MSE %6.4lf  U MSE %6.4lf  V MSE %6.4lf]", MSEyuvframe[COMPONENT_Y], MSEyuvframe[COMPONENT_Cb], MSEyuvframe[COMPONENT_Cr] );
        }
    #if WCG_WPSNR
        if (useLumaWPSNR)
        {
          msg(NOTICE, " [WY %6.4lf dB    WU %6.4lf dB    WV %6.4lf dB]", dPSNRWeighted[COMPONENT_Y], dPSNRWeighted[COMPONENT_Cb], dPSNRWeighted[COMPONENT_Cr]);
    
    
          if (m_pcEncLib->getPrintHexPsnr())
          {
            uint64_t xPsnrWeighted[MAX_NUM_COMPONENT];
            for (int i = 0; i < MAX_NUM_COMPONENT; i++)
            {
              copy(reinterpret_cast<uint8_t *>(&dPSNRWeighted[i]),
                   reinterpret_cast<uint8_t *>(&dPSNRWeighted[i]) + sizeof(dPSNRWeighted[i]),
                   reinterpret_cast<uint8_t *>(&xPsnrWeighted[i]));
            }
            msg(NOTICE, " [xWY %16" PRIx64 " xWU %16" PRIx64 " xWV %16" PRIx64 "]", xPsnrWeighted[COMPONENT_Y], xPsnrWeighted[COMPONENT_Cb], xPsnrWeighted[COMPONENT_Cr]);
          }
    
    #if JVET_O0756_CALCULATE_HDRMETRICS
    
        if(calculateHdrMetrics){
        for (int i=0; i<1; i++) {
              msg(NOTICE, " [DeltaE%d %6.4lf dB]", (int)m_pcCfg->getWhitePointDeltaE(i), dDeltaE[i]);
        }
        for (int i=0; i<1; i++) {
              msg(NOTICE, " [PSNRL%d %6.4lf dB]", (int)m_pcCfg->getWhitePointDeltaE(i), dPSNRL[i]);
        }
        }
    
        msg( NOTICE, " [ET %5.0f ]", dEncTime );
    
        // msg( SOME, " [WP %d]", pcSlice->getUseWeightedPrediction());
    
        for( int iRefList = 0; iRefList < 2; iRefList++ )
        {
          msg( NOTICE, " [L%d ", iRefList );
          for( int iRefIndex = 0; iRefIndex < pcSlice->getNumRefIdx( RefPicList( iRefList ) ); iRefIndex++ )
          {
    
            msg( NOTICE, "%d ", pcSlice->getRefPOC( RefPicList( iRefList ), iRefIndex ) );
    
          }
          msg( NOTICE, "]" );
        }
      }
      else if( g_verbosity >= INFO )
      {
        std::cout << "\r\t" << pcSlice->getPOC();
        std::cout.flush();
      }
    }
    
    
    #if JVET_O0756_CALCULATE_HDRMETRICS
    void EncGOP::xCalculateHDRMetrics( Picture* pcPic, double dDeltaE[hdrtoolslib::NB_REF_WHITE], double dPSNRL[hdrtoolslib::NB_REF_WHITE])
    {
      copyBuftoFrame(pcPic);
      
      ChromaFormat chFmt =  pcPic->chromaFormat;
      
      if (chFmt != CHROMA_444) {
        m_pcConvertFormat->process(m_ppcFrameOrg[1], m_ppcFrameOrg[0]);
        m_pcConvertFormat->process(m_ppcFrameRec[1], m_ppcFrameRec[0]);
      }
      
      m_pcConvertIQuantize->process(m_ppcFrameOrg[2], m_ppcFrameOrg[1]);
      m_pcConvertIQuantize->process(m_ppcFrameRec[2], m_ppcFrameRec[1]);
      
      m_pcColorTransform->process(m_ppcFrameOrg[3], m_ppcFrameOrg[2]);
      m_pcColorTransform->process(m_ppcFrameRec[3], m_ppcFrameRec[2]);
      
      m_pcTransferFct->forward(m_ppcFrameOrg[4], m_ppcFrameOrg[3]);
      m_pcTransferFct->forward(m_ppcFrameRec[4], m_ppcFrameRec[3]);
      
      // Calculate the Metrics
      m_pcDistortionDeltaE->computeMetric(m_ppcFrameOrg[4], m_ppcFrameRec[4]);
      
      *dDeltaE = m_pcDistortionDeltaE->getDeltaE();
      *dPSNRL = m_pcDistortionDeltaE->getPsnrL();
      
    }
    
    void EncGOP::copyBuftoFrame( Picture* pcPic )
    {
      int cropOffsetLeft   = m_pcCfg->getCropOffsetLeft();
      int cropOffsetTop    = m_pcCfg->getCropOffsetTop();
      int cropOffsetRight  = m_pcCfg->getCropOffsetRight();
      int cropOffsetBottom = m_pcCfg->getCropOffsetBottom();
      
      int iHeight = pcPic->getOrigBuf(COMPONENT_Y).height - cropOffsetLeft + cropOffsetRight;
      int iWidth = pcPic->getOrigBuf(COMPONENT_Y).width - cropOffsetTop + cropOffsetBottom;
      
      ChromaFormat chFmt =  pcPic->chromaFormat;
      
      Pel* pOrg = pcPic->getOrigBuf(COMPONENT_Y).buf;
      Pel* pRec = pcPic->getRecoBuf(COMPONENT_Y).buf;
      
      uint16_t* yOrg = m_ppcFrameOrg[0]->m_ui16Comp[hdrtoolslib::Y_COMP];
      uint16_t* yRec = m_ppcFrameRec[0]->m_ui16Comp[hdrtoolslib::Y_COMP];
      uint16_t* uOrg = m_ppcFrameOrg[0]->m_ui16Comp[hdrtoolslib::Cb_COMP];
      uint16_t* uRec = m_ppcFrameRec[0]->m_ui16Comp[hdrtoolslib::Cb_COMP];
      uint16_t* vOrg = m_ppcFrameOrg[0]->m_ui16Comp[hdrtoolslib::Cr_COMP];
      uint16_t* vRec = m_ppcFrameRec[0]->m_ui16Comp[hdrtoolslib::Cr_COMP];
      
      if(chFmt == CHROMA_444){
        yOrg = m_ppcFrameOrg[1]->m_ui16Comp[hdrtoolslib::Y_COMP];
        yRec = m_ppcFrameRec[1]->m_ui16Comp[hdrtoolslib::Y_COMP];
        uOrg = m_ppcFrameOrg[1]->m_ui16Comp[hdrtoolslib::Cb_COMP];
        uRec = m_ppcFrameRec[1]->m_ui16Comp[hdrtoolslib::Cb_COMP];
        vOrg = m_ppcFrameOrg[1]->m_ui16Comp[hdrtoolslib::Cr_COMP];
        vRec = m_ppcFrameRec[1]->m_ui16Comp[hdrtoolslib::Cr_COMP];
      }
      
      for (int i = 0; i < iHeight; i++) {
        for (int j = 0; j < iWidth; j++) {
          yOrg[i*iWidth + j] = static_cast<uint16_t>(pOrg[(i + cropOffsetTop) * pcPic->getOrigBuf(COMPONENT_Y).stride + j + cropOffsetLeft]);
          yRec[i*iWidth + j] = static_cast<uint16_t>(pRec[(i + cropOffsetTop) * pcPic->getRecoBuf(COMPONENT_Y).stride + j + cropOffsetLeft]);
        }
      }
      
      if (chFmt != CHROMA_444) {
        iHeight >>= 1;
        iWidth  >>= 1;
        cropOffsetLeft >>= 1;
        cropOffsetTop >>= 1;
      }
      
      pOrg = pcPic->getOrigBuf(COMPONENT_Cb).buf;
      pRec = pcPic->getRecoBuf(COMPONENT_Cb).buf;
      
      for (int i = 0; i < iHeight; i++) {
        for (int j = 0; j < iWidth; j++) {
          uOrg[i*iWidth + j] = static_cast<uint16_t>(pOrg[(i + cropOffsetTop) * pcPic->getOrigBuf(COMPONENT_Cb).stride + j + cropOffsetLeft]);
          uRec[i*iWidth + j] = static_cast<uint16_t>(pRec[(i + cropOffsetTop) * pcPic->getRecoBuf(COMPONENT_Cb).stride + j + cropOffsetLeft]);
        }
      }
      
      pOrg = pcPic->getOrigBuf(COMPONENT_Cr).buf;
      pRec = pcPic->getRecoBuf(COMPONENT_Cr).buf;
      
      for (int i = 0; i < iHeight; i++) {
        for (int j = 0; j < iWidth; j++) {
          vOrg[i*iWidth + j] = static_cast<uint16_t>(pOrg[(i + cropOffsetTop) * pcPic->getOrigBuf(COMPONENT_Cr).stride + j + cropOffsetLeft]);
          vRec[i*iWidth + j] = static_cast<uint16_t>(pRec[(i + cropOffsetTop) * pcPic->getRecoBuf(COMPONENT_Cr).stride + j + cropOffsetLeft]);
        }
      }
    }
    #endif
    
    
    void EncGOP::xCalculateInterlacedAddPSNR( Picture* pcPicOrgFirstField, Picture* pcPicOrgSecondField,
                                              PelUnitBuf cPicRecFirstField, PelUnitBuf cPicRecSecondField,
    
                                              const InputColourSpaceConversion conversion, const bool printFrameMSE, double* PSNR_Y
                                            , bool isEncodeLtRef
    )
    
    {
      const SPS &sps = *pcPicOrgFirstField->cs->sps;
      const ChromaFormat format = sps.getChromaFormatIdc();
      double  dPSNR[MAX_NUM_COMPONENT];
      Picture    *apcPicOrgFields[2] = {pcPicOrgFirstField, pcPicOrgSecondField};
      PelUnitBuf acPicRecFields[2]   = {cPicRecFirstField, cPicRecSecondField};
    #if ENABLE_QPA
      const bool    useWPSNR = m_pcEncLib->getUseWPSNR();
    #endif
      for(int i=0; i<MAX_NUM_COMPONENT; i++)
      {
        dPSNR[i]=0.0;
      }
    
      PelStorage cscd[2 /* first/second field */];
      if (conversion!=IPCOLOURSPACE_UNCHANGED)
      {
        for(uint32_t fieldNum=0; fieldNum<2; fieldNum++)
        {
          PelUnitBuf& reconField= (acPicRecFields[fieldNum]);
          cscd[fieldNum].create( reconField.chromaFormat, Area( Position(), reconField.Y()) );
          VideoIOYuv::ColourSpaceConvert(reconField, cscd[fieldNum], conversion, false);
          acPicRecFields[fieldNum]=cscd[fieldNum];
        }
      }
    
      //===== calculate PSNR =====
      double MSEyuvframe[MAX_NUM_COMPONENT] = {0, 0, 0};
    
      CHECK(!(acPicRecFields[0].chromaFormat==acPicRecFields[1].chromaFormat), "Unspecified error");
      const uint32_t numValidComponents = ::getNumberValidComponents( acPicRecFields[0].chromaFormat );
    
      for (int chan = 0; chan < numValidComponents; chan++)
      {
        const ComponentID ch=ComponentID(chan);
        CHECK(!(acPicRecFields[0].get(ch).width==acPicRecFields[1].get(ch).width), "Unspecified error");
        CHECK(!(acPicRecFields[0].get(ch).height==acPicRecFields[0].get(ch).height), "Unspecified error");
    
        uint64_t uiSSDtemp=0;
        const uint32_t width    = acPicRecFields[0].get(ch).width - (m_pcEncLib->getPad(0) >> ::getComponentScaleX(ch, format));
        const uint32_t height   = acPicRecFields[0].get(ch).height - ((m_pcEncLib->getPad(1) >> 1) >> ::getComponentScaleY(ch, format));
        const uint32_t bitDepth = sps.getBitDepth(toChannelType(ch));
    
        for(uint32_t fieldNum=0; fieldNum<2; fieldNum++)
        {
          CHECK(!(conversion == IPCOLOURSPACE_UNCHANGED), "Unspecified error");
    #if ENABLE_QPA
          uiSSDtemp += xFindDistortionPlane( acPicRecFields[fieldNum].get(ch), apcPicOrgFields[fieldNum]->getOrigBuf().get(ch), useWPSNR ? bitDepth : 0, ::getComponentScaleX(ch, format) );
    #else
          uiSSDtemp += xFindDistortionPlane( acPicRecFields[fieldNum].get(ch), apcPicOrgFields[fieldNum]->getOrigBuf().get(ch), 0 );
    #endif
        }
        const uint32_t maxval = 255 << (bitDepth - 8);
        const uint32_t size   = width * height * 2;
        const double fRefValue = (double)maxval * maxval * size;
        dPSNR[ch]         = uiSSDtemp ? 10.0 * log10(fRefValue / (double)uiSSDtemp) : 999.99;
        MSEyuvframe[ch]   = (double)uiSSDtemp / size;
      }
    
      uint32_t uibits = 0; // the number of bits for the pair is not calculated here - instead the overall total is used elsewhere.
    
      //===== add PSNR =====
    
      m_gcAnalyzeAll_in.addResult (dPSNR, (double)uibits, MSEyuvframe
        , isEncodeLtRef
      );
    
    
      *PSNR_Y = dPSNR[COMPONENT_Y];
    
      msg( DETAILS, "\n                                      Interlaced frame %d: [Y %6.4lf dB    U %6.4lf dB    V %6.4lf dB]", pcPicOrgSecondField->getPOC()/2 , dPSNR[COMPONENT_Y], dPSNR[COMPONENT_Cb], dPSNR[COMPONENT_Cr] );
      if (printFrameMSE)
      {
        msg( DETAILS, " [Y MSE %6.4lf  U MSE %6.4lf  V MSE %6.4lf]", MSEyuvframe[COMPONENT_Y], MSEyuvframe[COMPONENT_Cb], MSEyuvframe[COMPONENT_Cr] );
      }
    
      for(uint32_t fieldNum=0; fieldNum<2; fieldNum++)
      {
        cscd[fieldNum].destroy();
      }
    }
    
    /** Function for deciding the nal_unit_type.
     * \param pocCurr POC of the current picture
     * \param lastIDR  POC of the last IDR picture
     * \param isField  true to indicate field coding
     * \returns the NAL unit type of the picture
     * This function checks the configuration and returns the appropriate nal_unit_type for the picture.
     */
    NalUnitType EncGOP::getNalUnitType(int pocCurr, int lastIDR, bool isField)
    {
      if (pocCurr == 0)
      {
        return NAL_UNIT_CODED_SLICE_IDR_W_RADL;
      }
    
    
      if (m_pcCfg->getEfficientFieldIRAPEnabled() && isField && pocCurr == (m_pcCfg->getUseCompositeRef() ? 2: 1))
    
        return NAL_UNIT_CODED_SLICE_TRAIL;
    
      if (m_pcCfg->getDecodingRefreshType() != 3 && (pocCurr - isField) % (m_pcCfg->getIntraPeriod() * (m_pcCfg->getUseCompositeRef() ? 2 : 1)) == 0)
    
      {
        if (m_pcCfg->getDecodingRefreshType() == 1)
        {
          return NAL_UNIT_CODED_SLICE_CRA;
        }
        else if (m_pcCfg->getDecodingRefreshType() == 2)
        {
          return NAL_UNIT_CODED_SLICE_IDR_W_RADL;
        }
      }
      if(m_pocCRA>0)
      {
        if(pocCurr<m_pocCRA)
        {
          // All leading pictures are being marked as TFD pictures here since current encoder uses all
          // reference pictures while encoding leading pictures. An encoder can ensure that a leading
          // picture can be still decodable when random accessing to a CRA/CRANT/BLA/BLANT picture by
          // controlling the reference pictures used for encoding that leading picture. Such a leading
          // picture need not be marked as a TFD picture.
    
          return NAL_UNIT_CODED_SLICE_RASL;
    
          return NAL_UNIT_CODED_SLICE_RADL;
    
      return NAL_UNIT_CODED_SLICE_TRAIL;
    
    }
    
    void EncGOP::xUpdateRasInit(Slice* slice)
    {
      slice->setPendingRasInit( false );
      if ( slice->getPOC() > m_lastRasPoc )
      {
        m_lastRasPoc = MAX_INT;
        slice->setPendingRasInit( true );
      }
      if ( slice->isIRAP() )
      {
        m_lastRasPoc = slice->getPOC();
      }
    }
    
    double EncGOP::xCalculateRVM()
    {
      double dRVM = 0;
    
      if( m_pcCfg->getGOPSize() == 1 && m_pcCfg->getIntraPeriod() != 1 && m_pcCfg->getFramesToBeEncoded() > RVM_VCEGAM10_M * 2 )
      {
        // calculate RVM only for lowdelay configurations
        std::vector<double> vRL , vB;
        size_t N = m_vRVM_RP.size();
        vRL.resize( N );
        vB.resize( N );
    
        int i;
        double dRavg = 0 , dBavg = 0;
        vB[RVM_VCEGAM10_M] = 0;
        for( i = RVM_VCEGAM10_M + 1 ; i < N - RVM_VCEGAM10_M + 1 ; i++ )
        {
          vRL[i] = 0;
          for( int j = i - RVM_VCEGAM10_M ; j <= i + RVM_VCEGAM10_M - 1 ; j++ )
          {
            vRL[i] += m_vRVM_RP[j];
          }
          vRL[i] /= ( 2 * RVM_VCEGAM10_M );
          vB[i] = vB[i-1] + m_vRVM_RP[i] - vRL[i];
          dRavg += m_vRVM_RP[i];
          dBavg += vB[i];
        }
    
        dRavg /= ( N - 2 * RVM_VCEGAM10_M );
        dBavg /= ( N - 2 * RVM_VCEGAM10_M );
    
        double dSigamB = 0;
        for( i = RVM_VCEGAM10_M + 1 ; i < N - RVM_VCEGAM10_M + 1 ; i++ )
        {
          double tmp = vB[i] - dBavg;
          dSigamB += tmp * tmp;
        }
        dSigamB = sqrt( dSigamB / ( N - 2 * RVM_VCEGAM10_M ) );
    
        double f = sqrt( 12.0 * ( RVM_VCEGAM10_M - 1 ) / ( RVM_VCEGAM10_M + 1 ) );
    
        dRVM = dSigamB / dRavg * f;
      }
    
      return( dRVM );
    }
    
    /** Attaches the input bitstream to the stream in the output NAL unit
        Updates rNalu to contain concatenated bitstream. rpcBitstreamRedirect is cleared at the end of this function call.
     *  \param codedSliceData contains the coded slice data (bitstream) to be concatenated to rNalu
     *  \param rNalu          target NAL unit
     */
    void EncGOP::xAttachSliceDataToNalUnit (OutputNALUnit& rNalu, OutputBitstream* codedSliceData)
    {
      // Byte-align
      rNalu.m_Bitstream.writeByteAlignment();   // Slice header byte-alignment
    
      // Perform bitstream concatenation
      if (codedSliceData->getNumberOfWrittenBits() > 0)
      {
        rNalu.m_Bitstream.addSubstream(codedSliceData);
      }
      codedSliceData->clear();
    }
    
    
    
    void EncGOP::arrangeCompositeReference(Slice* pcSlice, PicList& rcListPic, int pocCurr)
    {
      Picture* curPic = NULL;
      PicList::iterator  iterPic = rcListPic.begin();
      const PreCalcValues *pcv = pcSlice->getPPS()->pcv;
      m_bgPOC = pocCurr + 1;
      if (m_picBg->getSpliceFull())
      {
        return;
      }
      while (iterPic != rcListPic.end())
      {
        curPic = *(iterPic++);
        if (curPic->getPOC() == pocCurr)
        {
          break;
        }
      }
    
      if (pcSlice->isIRAP())
    
      {
        return;
      }
    
      int width = pcv->lumaWidth;
      int height = pcv->lumaHeight;
      int stride = curPic->getOrigBuf().get(COMPONENT_Y).stride;
      int cStride = curPic->getOrigBuf().get(COMPONENT_Cb).stride;
      Pel* curLumaAddr = curPic->getOrigBuf().get(COMPONENT_Y).buf;
      Pel* curCbAddr = curPic->getOrigBuf().get(COMPONENT_Cb).buf;
      Pel* curCrAddr = curPic->getOrigBuf().get(COMPONENT_Cr).buf;
      Pel* bgOrgLumaAddr = m_picOrig->getOrigBuf().get(COMPONENT_Y).buf;
      Pel* bgOrgCbAddr = m_picOrig->getOrigBuf().get(COMPONENT_Cb).buf;
      Pel* bgOrgCrAddr = m_picOrig->getOrigBuf().get(COMPONENT_Cr).buf;
      int cuMaxWidth = pcv->maxCUWidth;
      int cuMaxHeight = pcv->maxCUHeight;
      int maxReplace = (pcv->sizeInCtus) / 2;
      maxReplace = maxReplace < 1 ? 1 : maxReplace;
      typedef struct tagCostStr
      {
        double cost;
        int ctuIdx;
      }CostStr;
      CostStr* minCtuCost = new CostStr[maxReplace];
      for (int i = 0; i < maxReplace; i++)
      {
        minCtuCost[i].cost = 1e10;
        minCtuCost[i].ctuIdx = -1;
      }
      int bitIncrementY = pcSlice->getSPS()->getBitDepth(CHANNEL_TYPE_LUMA) - 8;
      int bitIncrementUV = pcSlice->getSPS()->getBitDepth(CHANNEL_TYPE_CHROMA) - 8;
      for (int y = 0; y < height; y += cuMaxHeight)
      {
        for (int x = 0; x < width; x += cuMaxWidth)
        {
          double lcuDist = 0.0;
          double lcuDistCb = 0.0;
          double lcuDistCr = 0.0;
          int    realPixelCnt = 0;
          double lcuCost = 1e10;
          int largeDist = 0;
    
          for (int tmpy = 0; tmpy < cuMaxHeight; tmpy++)
          {
            if (y + tmpy >= height)
            {
              break;
            }
            for (int tmpx = 0; tmpx < cuMaxWidth; tmpx++)
            {
              if (x + tmpx >= width)
              {
                break;
              }